From 125dfe775a4e16b93585e5cc609fdbf466fcbbbf Mon Sep 17 00:00:00 2001 From: Serge Bazanski Date: Tue, 3 Nov 2020 17:56:23 +0100 Subject: [PATCH 1/4] *: pass context to List/Get/Update --- cmd/delete.go | 2 +- cmd/diff.go | 2 +- cmd/update.go | 2 +- pkg/kubecfg/delete.go | 5 +++-- pkg/kubecfg/diff.go | 5 +++-- pkg/kubecfg/update.go | 31 ++++++++++++++++--------------- 6 files changed, 25 insertions(+), 22 deletions(-) diff --git a/cmd/delete.go b/cmd/delete.go index b1651024..480ce377 100644 --- a/cmd/delete.go +++ b/cmd/delete.go @@ -60,6 +60,6 @@ var deleteCmd = &cobra.Command{ return err } - return c.Run(objs) + return c.Run(cmd.Context(), objs) }, } diff --git a/cmd/diff.go b/cmd/diff.go index d741b37d..33476fc8 100644 --- a/cmd/diff.go +++ b/cmd/diff.go @@ -67,6 +67,6 @@ var diffCmd = &cobra.Command{ return err } - return c.Run(objs, cmd.OutOrStdout()) + return c.Run(cmd.Context(), objs, cmd.OutOrStdout()) }, } diff --git a/cmd/update.go b/cmd/update.go index 7a3521ec..99860627 100644 --- a/cmd/update.go +++ b/cmd/update.go @@ -104,6 +104,6 @@ var updateCmd = &cobra.Command{ } } - return c.Run(objs) + return c.Run(cmd.Context(), objs) }, } diff --git a/pkg/kubecfg/delete.go b/pkg/kubecfg/delete.go index ee85ed5e..334e1f95 100644 --- a/pkg/kubecfg/delete.go +++ b/pkg/kubecfg/delete.go @@ -16,6 +16,7 @@ package kubecfg import ( + "context" "fmt" "sort" @@ -40,7 +41,7 @@ type DeleteCmd struct { GracePeriod int64 } -func (c DeleteCmd) Run(apiObjects []*unstructured.Unstructured) error { +func (c DeleteCmd) Run(ctx context.Context, apiObjects []*unstructured.Unstructured) error { version, err := utils.FetchVersion(c.Discovery) if err != nil { version = utils.GetDefaultVersion() @@ -77,7 +78,7 @@ func (c DeleteCmd) Run(apiObjects []*unstructured.Unstructured) error { return err } - err = client.Delete(obj.GetName(), &deleteOpts) + err = client.Delete(ctx, obj.GetName(), deleteOpts) if err != nil && !errors.IsNotFound(err) { return fmt.Errorf("Error deleting %s: %s", desc, err) } diff --git a/pkg/kubecfg/diff.go b/pkg/kubecfg/diff.go index 9f87e807..a8fd018a 100644 --- a/pkg/kubecfg/diff.go +++ b/pkg/kubecfg/diff.go @@ -17,6 +17,7 @@ package kubecfg import ( "bytes" + "context" "encoding/json" "fmt" "io" @@ -53,7 +54,7 @@ type DiffCmd struct { DiffStrategy string } -func (c DiffCmd) Run(apiObjects []*unstructured.Unstructured, out io.Writer) error { +func (c DiffCmd) Run(ctx context.Context, apiObjects []*unstructured.Unstructured, out io.Writer) error { sort.Sort(utils.AlphabeticalOrder(apiObjects)) dmp := diffmatchpatch.New() @@ -71,7 +72,7 @@ func (c DiffCmd) Run(apiObjects []*unstructured.Unstructured, out io.Writer) err return fmt.Errorf("Error fetching one of the %s: it does not have a name set", utils.ResourceNameFor(c.Mapper, obj)) } - liveObj, err := client.Get(obj.GetName(), metav1.GetOptions{}) + liveObj, err := client.Get(ctx, obj.GetName(), metav1.GetOptions{}) if err != nil && errors.IsNotFound(err) { log.Debugf("%s doesn't exist on the server", desc) liveObj = nil diff --git a/pkg/kubecfg/update.go b/pkg/kubecfg/update.go index 218e29d6..c34134dd 100644 --- a/pkg/kubecfg/update.go +++ b/pkg/kubecfg/update.go @@ -1,6 +1,7 @@ package kubecfg import ( + "context" "fmt" "sort" "time" @@ -163,8 +164,8 @@ func patch(existing, new *unstructured.Unstructured, schema proto.Schema) (*unst return result.(*unstructured.Unstructured), nil } -func createOrUpdate(rc dynamic.ResourceInterface, obj *unstructured.Unstructured, create bool, dryRun bool, schema proto.Schema, desc, dryRunText string) (*unstructured.Unstructured, error) { - existing, err := rc.Get(obj.GetName(), metav1.GetOptions{}) +func createOrUpdate(ctx context.Context, rc dynamic.ResourceInterface, obj *unstructured.Unstructured, create bool, dryRun bool, schema proto.Schema, desc, dryRunText string) (*unstructured.Unstructured, error) { + existing, err := rc.Get(ctx, obj.GetName(), metav1.GetOptions{}) if create && errors.IsNotFound(err) { log.Info("Creating ", desc, dryRunText) @@ -177,7 +178,7 @@ func createOrUpdate(rc dynamic.ResourceInterface, obj *unstructured.Unstructured if dryRun { return obj, nil } - newobj, err := rc.Create(obj, metav1.CreateOptions{}) + newobj, err := rc.Create(ctx, obj, metav1.CreateOptions{}) log.Debugf("Create(%s) returned (%v, %v)", obj.GetName(), newobj, err) return newobj, err } @@ -208,7 +209,7 @@ func createOrUpdate(rc dynamic.ResourceInterface, obj *unstructured.Unstructured if dryRun { return mergedObj, nil } - newobj, err := rc.Update(mergedObj, metav1.UpdateOptions{}) + newobj, err := rc.Update(ctx, mergedObj, metav1.UpdateOptions{}) log.Debugf("Update(%s) returned (%v, %v)", mergedObj.GetName(), newobj, err) if err != nil { log.Debug("Updated object: ", diff.ObjectDiff(existing, newobj)) @@ -241,7 +242,7 @@ func isSchemaEstablished(obj *unstructured.Unstructured) bool { return false } -func waitForSchemaChange(disco discovery.DiscoveryInterface, rc dynamic.ResourceInterface, obj *unstructured.Unstructured) { +func waitForSchemaChange(ctx context.Context, disco discovery.DiscoveryInterface, rc dynamic.ResourceInterface, obj *unstructured.Unstructured) { if isSchemaEstablished(obj) { return } @@ -251,7 +252,7 @@ func waitForSchemaChange(disco discovery.DiscoveryInterface, rc dynamic.Resource utils.MaybeMarkStale(disco) var err error - obj, err = rc.Get(obj.GetName(), metav1.GetOptions{}) + obj, err = rc.Get(ctx, obj.GetName(), metav1.GetOptions{}) if err != nil { if errors.IsNotFound(err) { // continue polling @@ -268,7 +269,7 @@ func waitForSchemaChange(disco discovery.DiscoveryInterface, rc dynamic.Resource } // Run executes the update command -func (c UpdateCmd) Run(apiObjects []*unstructured.Unstructured) error { +func (c UpdateCmd) Run(ctx context.Context, apiObjects []*unstructured.Unstructured) error { dryRunText := "" if c.DryRun { dryRunText = " (dry-run)" @@ -318,7 +319,7 @@ func (c UpdateCmd) Run(apiObjects []*unstructured.Unstructured) error { var newobj *unstructured.Unstructured err = retry.RetryOnConflict(retry.DefaultBackoff, func() (err error) { - newobj, err = createOrUpdate(rc, obj, c.Create, c.DryRun, schema, desc, dryRunText) + newobj, err = createOrUpdate(ctx, rc, obj, c.Create, c.DryRun, schema, desc, dryRunText) return }) if err != nil { @@ -334,7 +335,7 @@ func (c UpdateCmd) Run(apiObjects []*unstructured.Unstructured) error { // Don't wait for CRDs to settle schema under DryRun if !c.DryRun { - waitForSchemaChange(c.Discovery, rc, newobj) + waitForSchemaChange(ctx, c.Discovery, rc, newobj) } } @@ -346,7 +347,7 @@ func (c UpdateCmd) Run(apiObjects []*unstructured.Unstructured) error { } // [gctag-migration]: Add LabelGcTag==c.GcTag to ListOptions.LabelSelector in phase2 - err = walkObjects(c.Client, c.Discovery, metav1.ListOptions{}, func(o runtime.Object) error { + err = walkObjects(ctx, c.Client, c.Discovery, metav1.ListOptions{}, func(o runtime.Object) error { meta, err := meta.Accessor(o) if err != nil { return err @@ -357,7 +358,7 @@ func (c UpdateCmd) Run(apiObjects []*unstructured.Unstructured) error { if eligibleForGc(meta, c.GcTag) && !seenUids.Has(string(meta.GetUID())) { log.Info("Garbage collecting ", desc, dryRunText) if !c.DryRun { - err := gcDelete(c.Client, c.Mapper, &version, o) + err := gcDelete(ctx, c.Client, c.Mapper, &version, o) if err != nil { return err } @@ -382,7 +383,7 @@ func stringListContains(list []string, value string) bool { return false } -func gcDelete(client dynamic.Interface, mapper meta.RESTMapper, version *utils.ServerVersion, o runtime.Object) error { +func gcDelete(ctx context.Context, client dynamic.Interface, mapper meta.RESTMapper, version *utils.ServerVersion, o runtime.Object) error { obj, err := meta.Accessor(o) if err != nil { return fmt.Errorf("Unexpected object type: %s", err) @@ -409,7 +410,7 @@ func gcDelete(client dynamic.Interface, mapper meta.RESTMapper, version *utils.S return err } - err = c.Delete(obj.GetName(), &deleteOpts) + err = c.Delete(ctx, obj.GetName(), deleteOpts) if err != nil && (errors.IsNotFound(err) || errors.IsConflict(err)) { // We lost a race with something else changing the object log.Debugf("Ignoring error while deleting %s: %s", desc, err) @@ -422,7 +423,7 @@ func gcDelete(client dynamic.Interface, mapper meta.RESTMapper, version *utils.S return nil } -func walkObjects(client dynamic.Interface, disco discovery.DiscoveryInterface, listopts metav1.ListOptions, callback func(runtime.Object) error) error { +func walkObjects(ctx context.Context, client dynamic.Interface, disco discovery.DiscoveryInterface, listopts metav1.ListOptions, callback func(runtime.Object) error) error { rsrclists, err := disco.ServerResources() if err != nil { return err @@ -455,7 +456,7 @@ func walkObjects(client dynamic.Interface, disco discovery.DiscoveryInterface, l } log.Debugf("Listing %s", gvr) - obj, err := rc.List(listopts) + obj, err := rc.List(ctx, listopts) if err != nil { return err } From 9280dc8464cffbca47118ab6f26fa54fdba3d6be Mon Sep 17 00:00:00 2001 From: Serge Bazanski Date: Tue, 3 Nov 2020 17:58:39 +0100 Subject: [PATCH 2/4] *: remove k8s.io/kubernetes dependency The last package to come from k8s.io/kubernetes was kubectl/cmd/util/openapi, which has since been moved to k8s.io/util. We update to Kubernetes 1.19, remove the direct dependency on k8s.io/kubernetes (now only using client code from other k8s.io repos), and unpin gnostic to the newest version. --- go.mod | 51 ++---- go.sum | 363 +++++++++++++++++++++---------------- pkg/kubecfg/update.go | 2 +- pkg/kubecfg/update_test.go | 4 +- utils/client.go | 2 +- utils/openapi.go | 2 +- utils/openapi_test.go | 2 +- utils/sort_test.go | 2 +- 8 files changed, 230 insertions(+), 198 deletions(-) diff --git a/go.mod b/go.mod index 7dec332d..7b990234 100644 --- a/go.mod +++ b/go.mod @@ -3,56 +3,35 @@ module github.com/bitnami/kubecfg require ( github.com/Azure/go-autorest/autorest v0.10.0 // indirect github.com/elazarl/go-bindata-assetfs v1.0.1-0.20180223160309-38087fe4dafb - github.com/evanphx/json-patch v4.5.0+incompatible + github.com/evanphx/json-patch v4.9.0+incompatible github.com/genuinetools/reg v0.16.1 github.com/ghodss/yaml v1.0.0 github.com/go-openapi/spec v0.19.7 // indirect github.com/go-openapi/swag v0.19.8 // indirect - github.com/gogo/protobuf v1.3.1 // indirect - github.com/golang/protobuf v1.3.5 - github.com/google/go-cmp v0.4.0 // indirect - github.com/google/go-jsonnet v0.17.0 - github.com/google/gofuzz v1.1.0 // indirect - github.com/googleapis/gnostic v0.4.0 + github.com/golang/protobuf v1.4.3 + github.com/google/go-jsonnet v0.15.0 + github.com/googleapis/gnostic v0.5.3 github.com/imdario/mergo v0.3.9 // indirect - github.com/json-iterator/go v1.1.9 // indirect - github.com/konsorten/go-windows-terminal-sequences v1.0.2 // indirect - github.com/kr/pretty v0.2.0 // indirect github.com/mailru/easyjson v0.7.1 // indirect github.com/mattn/go-isatty v0.0.11 github.com/onsi/ginkgo v1.11.0 github.com/onsi/gomega v1.7.0 - github.com/pkg/errors v0.9.1 // indirect - github.com/sergi/go-diff v1.1.0 - github.com/sirupsen/logrus v1.4.2 - github.com/spf13/cobra v0.0.7 + github.com/sergi/go-diff v1.0.0 + github.com/sirupsen/logrus v1.6.0 + github.com/spf13/cobra v1.0.0 github.com/spf13/pflag v1.0.5 - github.com/stretchr/testify v1.4.0 - golang.org/x/crypto v0.0.0-20200406173513-056763e48d71 - golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e // indirect + github.com/stretchr/testify v1.5.1 + golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9 golang.org/x/oauth2 v0.0.0-20200107190931-bf48bf16ab8d // indirect - golang.org/x/sys v0.0.0-20200409092240-59c9f1ba88fa // indirect - golang.org/x/time v0.0.0-20191024005414-555d28b269f0 // indirect - google.golang.org/appengine v1.6.5 // indirect - google.golang.org/genproto v0.0.0-20200409111301-baae70f3302d // indirect google.golang.org/grpc v1.28.1 // indirect gopkg.in/yaml.v2 v2.2.8 - k8s.io/api v0.17.4 - k8s.io/apiextensions-apiserver v0.17.4 - k8s.io/apimachinery v0.17.4 - k8s.io/client-go v0.17.4 + k8s.io/api v0.19.3 + k8s.io/apiextensions-apiserver v0.19.3 + k8s.io/apimachinery v0.19.3 + k8s.io/client-go v0.19.3 k8s.io/klog v1.0.0 - k8s.io/kube-openapi v0.0.0-20191107075043-30be4d16710a - k8s.io/kubernetes v1.13.4 - k8s.io/utils v0.0.0-20200327001022-6496210b90e8 // indirect - sigs.k8s.io/yaml v1.2.0 // indirect + k8s.io/kube-openapi v0.0.0-20200923155610-8b5066479488 + k8s.io/kubectl v0.19.3 ) -// NB: pinning gnostic to v0.4.0 as v0.4.1 renamed s/OpenAPIv2/openapiv2/ at -// https://github.com/googleapis/gnostic/pull/155, -// while k8s.io/client-go/discovery@v0.17 still uses OpenAPIv2, -// even as of 2020/04/09 there's no released k8s.io/client-go/discovery version -// (latest 0.18.1) fixed to use openapiv2 -replace github.com/googleapis/gnostic => github.com/googleapis/gnostic v0.4.0 - go 1.13 diff --git a/go.sum b/go.sum index 610fea28..4d02d49a 100644 --- a/go.sum +++ b/go.sum @@ -4,9 +4,21 @@ cloud.google.com/go v0.34.0 h1:eOI3/cP2VTU6uZLDYAoic+eyzzB9YyGmJ7eIjl8rOPg= cloud.google.com/go v0.34.0/go.mod h1:aQUYkXzVsufM+DwF1aE+0xfcU+56JwCaLick0ClmMTw= cloud.google.com/go v0.38.0 h1:ROfEUZz+Gh5pa62DJWXSaonyu3StP6EA6lPEXPI6mCo= cloud.google.com/go v0.38.0/go.mod h1:990N+gfupTy94rShfmMCWGDn0LpTmnzTp2qbd1dvSRU= +cloud.google.com/go v0.44.1/go.mod h1:iSa0KzasP4Uvy3f1mN/7PiObzGgflwredwwASm/v6AU= +cloud.google.com/go v0.44.2/go.mod h1:60680Gw3Yr4ikxnPRS/oxxkBccT6SA1yMk63TGekxKY= +cloud.google.com/go v0.45.1/go.mod h1:RpBamKRgapWJb87xiFSdk4g1CME7QZg3uwTez+TSTjc= +cloud.google.com/go v0.46.3/go.mod h1:a6bKKbmY7er1mI7TEI4lsAkts/mkhTSZK8w33B4RAg0= +cloud.google.com/go v0.51.0 h1:PvKAVQWCtlGUSlZkGW3QLelKaWq7KYv/MW1EboG8bfM= +cloud.google.com/go v0.51.0/go.mod h1:hWtGJ6gnXH+KgDv+V0zFGDvpi07n3z8ZNj3T1RW0Gcw= +cloud.google.com/go/bigquery v1.0.1/go.mod h1:i/xbL2UlR5RvWAURpBYZTtm/cXjCha9lbfbpx4poX+o= +cloud.google.com/go/datastore v1.0.0/go.mod h1:LXYbyblFSglQ5pkeyhO+Qmw7ukd3C+pD7TKLgZqpHYE= +cloud.google.com/go/pubsub v1.0.1/go.mod h1:R0Gpsv3s54REJCy4fxDixWD93lHJMoZTyQ2kNxGRt3I= +cloud.google.com/go/storage v1.0.0/go.mod h1:IhtSnM/ZTZV8YYJWCY8RULGVqBDmpoyjwiyrjsg+URw= +dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7gZCb22OMCxBHrMx7a5I7Hp++hsVxbQ4BYO7hU= github.com/Azure/go-ansiterm v0.0.0-20170929234023-d6e3b3328b78 h1:w+iIsaOQNcT7OZ575w+acHgRric5iCyQh+xv+KJ4HB8= github.com/Azure/go-ansiterm v0.0.0-20170929234023-d6e3b3328b78/go.mod h1:LmzpDX56iTiv29bbRTIsUNlaFfuhWRQBWjQdVyAevI8= github.com/Azure/go-autorest/autorest v0.9.0/go.mod h1:xyHB1BMZT0cuDHU7I0+g046+BFDTQ8rEZB0s4Yfa6bI= +github.com/Azure/go-autorest/autorest v0.9.6/go.mod h1:/FALq9T/kS7b5J5qsQ+RSTUdAmGFqi0vUdVNNx8q630= github.com/Azure/go-autorest/autorest v0.10.0 h1:mvdtztBqcL8se7MdrUweNieTNi4kfNG6GOJuurQJpuY= github.com/Azure/go-autorest/autorest v0.10.0/go.mod h1:/FALq9T/kS7b5J5qsQ+RSTUdAmGFqi0vUdVNNx8q630= github.com/Azure/go-autorest/autorest/adal v0.5.0/go.mod h1:8Z9fGy2MpX0PvDjB1pEgQTmVqjGhiHBW7RJJEciWzS0= @@ -25,6 +37,7 @@ github.com/Azure/go-autorest/tracing v0.5.0 h1:TRn4WjSnkcSy5AEG3pnbtFSwNtwzjr4VY github.com/Azure/go-autorest/tracing v0.5.0/go.mod h1:r/s2XiOKccPW3HrqB+W0TQzfbtp2fGCgRFtBroKn4Dk= github.com/BurntSushi/toml v0.3.1/go.mod h1:xHWCNGjB5oqiDr8zfno3MHue2Ht5sIBksp03qcyfWMU= github.com/BurntSushi/xgb v0.0.0-20160522181843-27f122750802/go.mod h1:IVnqGOEym/WlBOVXweHU+Q+/VP0lqqI8lqeDx9IjBqo= +github.com/MakeNowJust/heredoc v0.0.0-20170808103936-bb23615498cd/go.mod h1:64YHyfSL2R96J44Nlwm39UHepQbyR5q10x7iYa1ks2E= github.com/Microsoft/go-winio v0.4.14 h1:+hMXMk01us9KgxGb7ftKQt2Xpf5hH/yky+TDA+qxleU= github.com/Microsoft/go-winio v0.4.14/go.mod h1:qXqCSQ3Xa7+6tgxaGTIe4Kpcdsi+P8jBhyzoq1bpyYA= github.com/Microsoft/hcsshim v0.8.6 h1:ZfF0+zZeYdzMIVMZHKtDKJvLHj76XCuVae/jNkjj0IA= @@ -42,7 +55,9 @@ github.com/PuerkitoBio/urlesc v0.0.0-20170810143723-de5bf2ad4578 h1:d+Bc7a5rLufV github.com/PuerkitoBio/urlesc v0.0.0-20170810143723-de5bf2ad4578/go.mod h1:uGdkoq3SwY9Y+13GIhn11/XLaGBb4BfwItxLd5jeuXE= github.com/agnivade/levenshtein v1.0.1/go.mod h1:CURSv5d9Uaml+FovSIICkLbAUZ9S4RqaHDIsdSBg7lM= github.com/alecthomas/template v0.0.0-20160405071501-a0175ee3bccc/go.mod h1:LOuyumcjzFXgccqObfd/Ljyb9UuFJ6TxHnclSeseNhc= +github.com/alecthomas/template v0.0.0-20190718012654-fb15b899a751/go.mod h1:LOuyumcjzFXgccqObfd/Ljyb9UuFJ6TxHnclSeseNhc= github.com/alecthomas/units v0.0.0-20151022065526-2efee857e7cf/go.mod h1:ybxpYRFXyAe+OPACYpWeL0wqObRcbAqCMya13uyzqw0= +github.com/alecthomas/units v0.0.0-20190717042225-c3de453c63f4/go.mod h1:ybxpYRFXyAe+OPACYpWeL0wqObRcbAqCMya13uyzqw0= github.com/andreyvit/diff v0.0.0-20170406064948-c7f18ee00883/go.mod h1:rCTlJbsFo29Kk6CurOXKm700vrz8f0KW0JNfpkRJY/8= github.com/armon/consul-api v0.0.0-20180202201655-eb2c6b5be1b6/go.mod h1:grANhF5doyWs3UAsr3K4I6qtAmlQcZDesFNEHPZAzj8= github.com/asaskevich/govalidator v0.0.0-20180720115003-f9ffefc3facf/go.mod h1:lB+ZfQJz7igIIfQNfa7Ml4HSf2uFQQRzpGGRXenZAgY= @@ -51,12 +66,18 @@ github.com/asaskevich/govalidator v0.0.0-20190424111038-f61b66f89f4a/go.mod h1:l github.com/beorn7/perks v0.0.0-20180321164747-3a771d992973/go.mod h1:Dwedo/Wpr24TaqPxmxbtue+5NUziq4I4S80YR8gNf3Q= github.com/beorn7/perks v1.0.0 h1:HWo1m869IqiPhD389kmkxeTalrjNbbJTC8LXupb+sl0= github.com/beorn7/perks v1.0.0/go.mod h1:KWe93zE9D1o94FZ5RNwFwVgaQK1VOXiVxmqh+CedLV8= +github.com/beorn7/perks v1.0.1/go.mod h1:G2ZrVWU2WbWT9wwq4/hrbKbnv/1ERSJQ0ibhJ6rlkpw= github.com/bgentry/speakeasy v0.1.0/go.mod h1:+zsyZBPWlz7T6j88CTgSN5bM796AkVf0kBD4zp0CCIs= github.com/blang/semver v3.5.0+incompatible h1:CGxCgetQ64DKk7rdZ++Vfnb1+ogGNnB17OJKJXD2Cfs= github.com/blang/semver v3.5.0+incompatible/go.mod h1:kRBLl5iJ+tD4TcOOxsy/0fnwebNt5EWlYSAyrTnjyyk= github.com/census-instrumentation/opencensus-proto v0.2.1/go.mod h1:f6KPmirojxKA12rnyqOA5BBL4O983OfeGPqjHWSTneU= github.com/cespare/xxhash v1.1.0 h1:a6HrQnmkObjyL+Gs60czilIUGqrzKutQD6XZog3p+ko= github.com/cespare/xxhash v1.1.0/go.mod h1:XrSqR1VqqWfGrhpAt58auRo0WTKS1nRRg3ghfAqPWnc= +github.com/cespare/xxhash/v2 v2.1.1/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= +github.com/chai2010/gettext-go v0.0.0-20160711120539-c6fed771bfd5/go.mod h1:/iP1qXHoty45bqomnu2LM+VVyAEdWN+vtSHGlQgyxbw= +github.com/chzyer/logex v1.1.10/go.mod h1:+Ywpsq7O8HXn0nuIou7OrIPyXbp3wmkHB+jjWRnGsAI= +github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e/go.mod h1:nSuG5e5PlCu98SY8svDHJxuZscDgtXS6KTTbou5AhLI= +github.com/chzyer/test v0.0.0-20180213035817-a1ea475d72b1/go.mod h1:Q3SI9o4m/ZMnBNeIyt5eFwwo7qiLfzFZmjNmxjkiQlU= github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw= github.com/cncf/udpa/go v0.0.0-20191209042840-269d4d468f6f/go.mod h1:M8M6+tZqaGXZJjfX53e64911xZQV5JYwmTeXPW+k8Sc= github.com/cockroachdb/datadriven v0.0.0-20190809214429-80d97fb3cbaa/go.mod h1:zn76sxSg3SzpJ0PPJaLDCu+Bu0Lg3sKTORVIj19EIF8= @@ -69,7 +90,6 @@ github.com/coreos/clair v2.0.1-0.20190910143208-94150ab1f4ac+incompatible h1:V+K github.com/coreos/clair v2.0.1-0.20190910143208-94150ab1f4ac+incompatible/go.mod h1:uXhHPWAoRqw0jJc2f8RrPCwRhIo9otQ8OEWUFtpCiwA= github.com/coreos/etcd v3.3.10+incompatible h1:jFneRYjIvLMLhDLCzuTuU4rSJUjRplcJQ7pD7MnhC04= github.com/coreos/etcd v3.3.10+incompatible/go.mod h1:uF7uidLiAD3TWHmW31ZFd/JWoc32PjwdhPthX9715RE= -github.com/coreos/go-etcd v2.0.0+incompatible/go.mod h1:Jez6KQU2B/sWsbdaef3ED8NzMklzPG4d5KIOhIy30Tk= github.com/coreos/go-oidc v2.1.0+incompatible/go.mod h1:CgnwVTmzoESiwO9qyAFEMiHoZ1nMCKZlZ9V6mm3/LKc= github.com/coreos/go-semver v0.2.0/go.mod h1:nnelYz7RCh+5ahJtPPxZlU+153eP4D4r3EedlOD2RNk= github.com/coreos/go-semver v0.3.0 h1:wkHLiw0WNATZnSG7epLsujiMCgPAc9xhjJ4tgnAxmfM= @@ -78,17 +98,14 @@ github.com/coreos/go-systemd v0.0.0-20180511133405-39ca1b05acc7/go.mod h1:F5haX7 github.com/coreos/go-systemd v0.0.0-20190321100706-95778dfbb74e h1:Wf6HqHfScWJN9/ZjdUKyjop4mf3Qdd+1TvvltAvM3m8= github.com/coreos/go-systemd v0.0.0-20190321100706-95778dfbb74e/go.mod h1:F5haX7vjVVG0kc13fIWeqUViNPyEJxv/OmvnBo0Yme4= github.com/coreos/pkg v0.0.0-20160727233714-3ac0863d7acf/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA= -github.com/coreos/pkg v0.0.0-20180108230652-97fdf19511ea h1:n2Ltr3SrfQlf/9nOna1DoGKxLx3qTSI8Ttl6Xrqp6mw= -github.com/coreos/pkg v0.0.0-20180108230652-97fdf19511ea/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA= github.com/coreos/pkg v0.0.0-20180928190104-399ea9e2e55f h1:lBNOc5arjvs8E5mO2tbpBpLoyyu8B6e44T7hJy6potg= github.com/coreos/pkg v0.0.0-20180928190104-399ea9e2e55f/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA= -github.com/cpuguy83/go-md2man v1.0.10/go.mod h1:SmD6nW6nTyfqj6ABTjUi3V3JVMnlJmwcJI5acqYI6dE= github.com/cpuguy83/go-md2man/v2 v2.0.0/go.mod h1:maD7wRr/U5Z6m/iR4s+kqSMx2CaBsrgA7czyZG/E6dU= github.com/creack/pty v1.1.7/go.mod h1:lj5s0c3V2DBrqTV7llrYr5NG6My20zk30Fl46Y7DoTY= -github.com/davecgh/go-spew v0.0.0-20151105211317-5215b55f46b2/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c= github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= +github.com/daviddengcn/go-colortext v0.0.0-20160507010035-511bcaf42ccd/go.mod h1:dv4zxwHi5C/8AeI+4gX4dCWOIvNi7I6JCSX0HvlKPgE= github.com/deckarep/golang-set v1.7.1 h1:SCQV0S6gTtp6itiFrTqI+pfmJ4LN85S1YzhDf9rTHJQ= github.com/deckarep/golang-set v1.7.1/go.mod h1:93vsz/8Wt4joVM7c2AVqh+YRMiUSc14yDtF28KmMOgQ= github.com/dgrijalva/jwt-go v3.2.0+incompatible h1:7qlOGliEKZXTDg6OTjfoBKDXWrumCAMpl/TFQ4/5kLM= @@ -98,7 +115,6 @@ github.com/docker/cli v0.0.0-20190913211141-95327f4e6241 h1:btTBgRvrdoe+b7NfX/7P github.com/docker/cli v0.0.0-20190913211141-95327f4e6241/go.mod h1:JLrzqnKDaYBop7H2jaqPtU4hHvMKP+vjCwu2uszcLI8= github.com/docker/distribution v2.7.1+incompatible h1:a5mlkVzth6W5A4fOsS3D2EO5BUmsJpcB+cRlLU7cSug= github.com/docker/distribution v2.7.1+incompatible/go.mod h1:J2gT2udsDAN96Uj4KfcMRqY0/ypR+oyYUYmja8H+y+w= -github.com/docker/docker v0.7.3-0.20190327010347-be7ac8be2ae0/go.mod h1:eEKB0N0r5NX/I1kEveEz05bcu8tLC/8azJZsviup8Sk= github.com/docker/docker v1.4.2-0.20190916154449-92cc603036dd h1:kDIT0qjvLHbdL86aa+VteVpVZOR7coIyIejM/o3CwOo= github.com/docker/docker v1.4.2-0.20190916154449-92cc603036dd/go.mod h1:eEKB0N0r5NX/I1kEveEz05bcu8tLC/8azJZsviup8Sk= github.com/docker/docker-credential-helpers v0.6.3 h1:zI2p9+1NQYdnG6sMU26EX4aVGlqbInSQxQXLvzJ4RPQ= @@ -116,7 +132,7 @@ github.com/dustin/go-humanize v0.0.0-20171111073723-bb3d318650d4/go.mod h1:Htrtb github.com/dustin/go-humanize v1.0.0/go.mod h1:HtrtbFcZ19U5GC7JDqmcUSB87Iq5E25KnS6fMYU6eOk= github.com/elazarl/go-bindata-assetfs v1.0.1-0.20180223160309-38087fe4dafb h1:Dnxl6iOR/3QQRcCBDEOCpusGgsx7uDS+Pa/InwqCFfw= github.com/elazarl/go-bindata-assetfs v1.0.1-0.20180223160309-38087fe4dafb/go.mod h1:v+YaWX3bdea5J/mo8dSETolEo7R71Vk1u8bnjau5yw4= -github.com/elazarl/goproxy v0.0.0-20170405201442-c4fc26588b6e/go.mod h1:/Zj4wYkgs4iZTTu3o/KG3Itv/qCCa8VVMlb3i9OVuzc= +github.com/elazarl/goproxy v0.0.0-20180725130230-947c36da3153/go.mod h1:/Zj4wYkgs4iZTTu3o/KG3Itv/qCCa8VVMlb3i9OVuzc= github.com/emicklei/go-restful v0.0.0-20170410110728-ff4f55a20633/go.mod h1:otzb+WCGbkyDHkqmQmT5YD2WR4BBwUdeQoFo8l/7tVs= github.com/emicklei/go-restful v2.9.5+incompatible h1:spTtZBk5DYEvbxMVutUuTyh1Ao2r4iyvLdACqsl/Ljk= github.com/emicklei/go-restful v2.9.5+incompatible/go.mod h1:otzb+WCGbkyDHkqmQmT5YD2WR4BBwUdeQoFo8l/7tVs= @@ -124,16 +140,17 @@ github.com/envoyproxy/go-control-plane v0.9.0/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymF github.com/envoyproxy/go-control-plane v0.9.1-0.20191026205805-5f8ba28d4473/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.4/go.mod h1:6rpuAdCZL397s3pYoYcLgu1mIlRU8Am5FuJP05cCM98= github.com/envoyproxy/protoc-gen-validate v0.1.0/go.mod h1:iSmxcyjqTsJpI2R4NaDN7+kN2VEUnK/pcBlmesArF7c= -github.com/evanphx/json-patch v4.2.0+incompatible h1:fUDGZCv/7iAN7u0puUVhvKCcsR6vRfwrJatElLBEf0I= -github.com/evanphx/json-patch v4.2.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= -github.com/evanphx/json-patch v4.5.0+incompatible h1:ouOWdg56aJriqS0huScTkVXPC5IcNrDCXZ6OoTAWu7M= -github.com/evanphx/json-patch v4.5.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= +github.com/evanphx/json-patch v4.9.0+incompatible h1:kLcOMZeuLAJvL2BPWLMIj5oaZQobrkAqrL+WFZwQses= +github.com/evanphx/json-patch v4.9.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= +github.com/exponent-io/jsonpath v0.0.0-20151013193312-d6023ce2651d/go.mod h1:ZZMPRZwes7CROmyNKgQzC3XPs6L/G2EJLHddWejkmf4= +github.com/fatih/camelcase v1.0.0/go.mod h1:yN2Sb0lFhZJUdVvtELVWefmrXpuZESvPmqwoZc+/fpc= github.com/fatih/color v1.7.0/go.mod h1:Zm6kSWBoL9eyXnKyktHP6abPY2pDugNf5KwzbycvMj4= -github.com/fatih/color v1.9.0/go.mod h1:eQcE1qtQxscV5RaZvpXrrb8Drkc3/DdQ+uUYCNjL+zU= github.com/fernet/fernet-go v0.0.0-20180830025343-9eac43b88a5e h1:P10tZmVD2XclAaT9l7OduMH1OLFzTa1wUuUqHZnEdI0= github.com/fernet/fernet-go v0.0.0-20180830025343-9eac43b88a5e/go.mod h1:2H9hjfbpSMHwY503FclkV/lZTBh2YlOmLLSda12uL8c= github.com/fsnotify/fsnotify v1.4.7 h1:IXs+QLmnXW2CcXuY+8Mzv/fWEsPGWxqefPtCP5CnV9I= github.com/fsnotify/fsnotify v1.4.7/go.mod h1:jwhsz4b93w/PPRr/qN1Yymfu8t87LnFCMoQvtojpjFo= +github.com/fsnotify/fsnotify v1.4.9 h1:hsms1Qyu0jgnwNXIxa+/V/PDsU6CfLf6CNO8H7IWoS4= +github.com/fsnotify/fsnotify v1.4.9/go.mod h1:znqG4EE+3YCdAaPaxE2ZRY/06pZUdp0tY4IgpuI1SZQ= github.com/genuinetools/pkg v0.0.0-20181022210355-2fcf164d37cb h1:9MQ4N7zyYTtdjLGqE5McDbgjIjqR5TAPc6lytEOdndc= github.com/genuinetools/pkg v0.0.0-20181022210355-2fcf164d37cb/go.mod h1:XTcrCYlXPxnxL2UpnwuRn7tcaTn9HAhxFoFJucootk8= github.com/genuinetools/reg v0.16.1 h1:nZiceimcvxEVtWKSmPtRhEkl2TrxDhrbmYEAbTkKXvo= @@ -143,10 +160,14 @@ github.com/ghodss/yaml v1.0.0 h1:wQHKEahhL6wmXdzwWG11gIVCkOv05bNOh+Rxn0yngAk= github.com/ghodss/yaml v1.0.0/go.mod h1:4dBDuWmgqj2HViK6kFavaiC9ZROes6MMH2rRYeMEF04= github.com/globalsign/mgo v0.0.0-20180905125535-1ca0a4f7cbcb/go.mod h1:xkRDCp4j0OGD1HRkm4kmhM+pmpv3AKq5SU7GMg4oO/Q= github.com/globalsign/mgo v0.0.0-20181015135952-eeefdecb41b8/go.mod h1:xkRDCp4j0OGD1HRkm4kmhM+pmpv3AKq5SU7GMg4oO/Q= +github.com/go-gl/glfw/v3.3/glfw v0.0.0-20191125211704-12ad95a8df72/go.mod h1:tQ2UAYgL5IevRw8kRxooKSPJfGvJ9fJQFa0TUsXzTg8= github.com/go-kit/kit v0.8.0/go.mod h1:xBxKIO96dXMWWy0MnWVtmwkA9/13aqxPnvrjFYMA2as= +github.com/go-kit/kit v0.9.0/go.mod h1:xBxKIO96dXMWWy0MnWVtmwkA9/13aqxPnvrjFYMA2as= github.com/go-logfmt/logfmt v0.3.0/go.mod h1:Qt1PoO58o5twSAckw1HlFXLmHsOX5/0LbT9GBnD5lWE= github.com/go-logfmt/logfmt v0.4.0/go.mod h1:3RMwSq7FuexP4Kalkev3ejPJsZTpXXBr9+V4qmtdjCk= github.com/go-logr/logr v0.1.0/go.mod h1:ixOQHD9gLJUVQQ2ZOR7zLEifBX6tGkNJF4QyIY7sIas= +github.com/go-logr/logr v0.2.0 h1:QvGt2nLcHH0WK9orKa+ppBPAxREcH364nPUedEpK0TY= +github.com/go-logr/logr v0.2.0/go.mod h1:z6/tIYblkpsD+a4lm/fGIIU9mZ+XfAiaFtq7xTgseGU= github.com/go-openapi/analysis v0.0.0-20180825180245-b006789cd277/go.mod h1:k70tL6pCuVxPJOHXQ+wIac1FUrvNkHolPie/cLEU6hI= github.com/go-openapi/analysis v0.17.0/go.mod h1:IowGgpVeD0vNm45So8nr+IcQ3pxVtpRoBWb8PVZO0ik= github.com/go-openapi/analysis v0.18.0/go.mod h1:IowGgpVeD0vNm45So8nr+IcQ3pxVtpRoBWb8PVZO0ik= @@ -212,7 +233,6 @@ github.com/go-stack/stack v1.8.0 h1:5SgMzNM5HxrEjV0ww2lTmX6E2Izsfxas4+YHWRs3Lsk= github.com/go-stack/stack v1.8.0/go.mod h1:v0f6uXyyMGvRgIKkXu+yp6POWl0qKG85gN/melR3HDY= github.com/gogo/protobuf v1.1.1/go.mod h1:r8qH/GZQm5c6nD/R0oafs1akxWv10x8SbQlK7atdtwQ= github.com/gogo/protobuf v1.2.1/go.mod h1:hp+jE20tsWTFYpLwKvXlhS1hjn+gTNwPg2I6zVXpSg4= -github.com/gogo/protobuf v1.2.2-0.20190723190241-65acae22fc9d/go.mod h1:SlYgWuQ5SjCEi6WLHjHCa1yvBfUnHcTbrrZtXPKa29o= github.com/gogo/protobuf v1.3.0 h1:G8O7TerXerS4F6sx9OV7/nRfJdnXgHZu/S/7F2SN+UE= github.com/gogo/protobuf v1.3.0/go.mod h1:SlYgWuQ5SjCEi6WLHjHCa1yvBfUnHcTbrrZtXPKa29o= github.com/gogo/protobuf v1.3.1 h1:DqDEcV5aeaTmdFBePNpYsp3FlcVH/2ISVVM9Qf8PSls= @@ -222,17 +242,30 @@ github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfU github.com/golang/groupcache v0.0.0-20160516000752-02826c3e7903 h1:LbsanbbD6LieFkXbj9YNNBupiGHJgFeLpO0j0Fza1h8= github.com/golang/groupcache v0.0.0-20160516000752-02826c3e7903/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20190129154638-5b532d6fd5ef/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= +github.com/golang/groupcache v0.0.0-20190702054246-869f871628b6/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= +github.com/golang/groupcache v0.0.0-20191227052852-215e87163ea7/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/mock v1.1.1/go.mod h1:oTYuIxOrZwtPieC+H1uAHpcLFnEyAGVDL/k47Jfbm0A= github.com/golang/mock v1.2.0/go.mod h1:oTYuIxOrZwtPieC+H1uAHpcLFnEyAGVDL/k47Jfbm0A= -github.com/golang/protobuf v0.0.0-20161109072736-4bd1920723d7/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U= +github.com/golang/mock v1.3.1/go.mod h1:sBzyDLLjw3U8JLTeZvSv8jJB+tU5PVekmnlKIyFUx0Y= github.com/golang/protobuf v1.2.0 h1:P3YflyNX/ehuJFLhxviNdFxQPkGK5cDcApsge1SqnvM= github.com/golang/protobuf v1.2.0/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U= github.com/golang/protobuf v1.3.1/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U= github.com/golang/protobuf v1.3.2 h1:6nsPYzhq5kReh6QImI3k5qWzO4PEbvbIW2cwSfR/6xs= github.com/golang/protobuf v1.3.2/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U= github.com/golang/protobuf v1.3.3/go.mod h1:vzj43D7+SQXF/4pzW/hwtAqwc6iTitCiVSaWz5lYuqw= -github.com/golang/protobuf v1.3.5 h1:F768QJ1E9tib+q5Sc8MkdJi1RxLTbRcTf8LJV56aRls= -github.com/golang/protobuf v1.3.5/go.mod h1:6O5/vntMXwX2lRkT1hjjk0nAC1IDOTvTlVgjlRvqsdk= +github.com/golang/protobuf v1.4.0-rc.1/go.mod h1:ceaxUfeHdC40wWswd/P6IGgMaK3YpKi5j83Wpe3EHw8= +github.com/golang/protobuf v1.4.0-rc.1.0.20200221234624-67d41d38c208/go.mod h1:xKAWHe0F5eneWXFV3EuXVDTCmh+JuBKY0li0aMyXATA= +github.com/golang/protobuf v1.4.0-rc.2/go.mod h1:LlEzMj4AhA7rCAGe4KMBDvJI+AwstrUpVNzEA03Pprs= +github.com/golang/protobuf v1.4.0-rc.4.0.20200313231945-b860323f09d0/go.mod h1:WU3c8KckQ9AFe+yFwt9sWVRKCVIyN9cPHBJSNnbL67w= +github.com/golang/protobuf v1.4.0/go.mod h1:jodUvKwWbYaEsadDk5Fwe5c77LiNKVO9IDvqG2KuDX0= +github.com/golang/protobuf v1.4.1/go.mod h1:U8fpvMrcmy5pZrNK1lt4xCsGvpyWQ/VVv6QDs8UjoX8= +github.com/golang/protobuf v1.4.2 h1:+Z5KGCizgyZCbGh1KZqA0fcLLkwbsjIzS4aV2v7wJX0= +github.com/golang/protobuf v1.4.2/go.mod h1:oDoupMAO8OvCJWAcko0GGGIgR6R6ocIYbsSw735rRwI= +github.com/golang/protobuf v1.4.3 h1:JjCZWpVbqXDqFVmTfYWEVTMIYrL/NPdPSCHPJ0T/raM= +github.com/golang/protobuf v1.4.3/go.mod h1:oDoupMAO8OvCJWAcko0GGGIgR6R6ocIYbsSw735rRwI= +github.com/golangplus/bytes v0.0.0-20160111154220-45c989fe5450/go.mod h1:Bk6SMAONeMXrxql8uvOKuAZSu8aM5RUGv+1C6IJaEho= +github.com/golangplus/fmt v0.0.0-20150411045040-2a5d6d7d2995/go.mod h1:lJgMEyOkYFkPcDKwRXegd+iM6E7matEszMG5HhwytU8= +github.com/golangplus/testing v0.0.0-20180327235837-af21d9c3145e/go.mod h1:0AA//k/eakGydO4jKRoRL2j92ZKSzTgj9tclaCrvXHk= github.com/google/btree v0.0.0-20180813153112-4030bb1f1f0c/go.mod h1:lNA+9X1NB3Zf8V7Ke586lFgjr2dZNuvo3lPJSGZ5JPQ= github.com/google/btree v1.0.0 h1:0udJVsspx3VBr5FwtLhQQtuAsVc79tTq0ocGIPAU6qo= github.com/google/btree v1.0.0/go.mod h1:lNA+9X1NB3Zf8V7Ke586lFgjr2dZNuvo3lPJSGZ5JPQ= @@ -243,24 +276,23 @@ github.com/google/go-cmp v0.3.1 h1:Xye71clBPdm5HgqGwUkwhbynsUJZhDbS20FvLhQ2izg= github.com/google/go-cmp v0.3.1/go.mod h1:8QqcDgzrUqlUb/G2PQTWiueGozuR1884gddMywk6iLU= github.com/google/go-cmp v0.4.0 h1:xsAVV57WRhGj6kEIi8ReJzQlHHqcBYCElAvkovg3B/4= github.com/google/go-cmp v0.4.0/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= -github.com/google/go-jsonnet v0.17.0 h1:/9NIEfhK1NQRKl3sP2536b2+x5HnZMdql7x3yK/l8JY= -github.com/google/go-jsonnet v0.17.0/go.mod h1:sOcuej3UW1vpPTZOr8L7RQimqai1a57bt5j22LzGZCw= -github.com/google/gofuzz v0.0.0-20161122191042-44d81051d367 h1:ScAXWS+TR6MZKex+7Z8rneuSJH+FSDqd6ocQyl+ZHo4= -github.com/google/gofuzz v0.0.0-20161122191042-44d81051d367/go.mod h1:HP5RmnzzSNb993RKQDq4+1A4ia9nllfqcQFTQJedwGI= -github.com/google/gofuzz v1.0.0 h1:A8PeW59pxE9IoFRqBp37U+mSNaQoZ46F1f0f863XSXw= +github.com/google/go-jsonnet v0.15.0 h1:lEUXTDnVsHu+CLLzMeWAdWV4JpCgkJeDqdVNS8RtyuY= +github.com/google/go-jsonnet v0.15.0/go.mod h1:ex9QcU8vzXQUDeNe4gaN1uhGQbTYpOeZ6AbWdy6JbX4= github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/gofuzz v1.1.0 h1:Hsa8mG0dQ46ij8Sl2AYJDUv1oA9/d6Vk+3LG99Oe02g= github.com/google/gofuzz v1.1.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/martian v2.1.0+incompatible/go.mod h1:9I4somxYTbIHy5NJKHRl3wXiIaQGbYVAs8BPL6v8lEs= github.com/google/pprof v0.0.0-20181206194817-3ea8567a2e57/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc= +github.com/google/pprof v0.0.0-20190515194954-54271f7e092f/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc= +github.com/google/pprof v0.0.0-20191218002539-d4f498aebedc/go.mod h1:ZgVRPoUq/hfqzAqh7sHMqb3I9Rq5C59dIz2SbBwJ4eM= +github.com/google/renameio v0.1.0/go.mod h1:KWCgfxg9yswjAJkECMjeO8J8rahYeXnNhOm40UhjYkI= github.com/google/uuid v1.0.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= -github.com/google/uuid v1.1.1 h1:Gkbcsh/GbpXz7lPftLA3P6TYMwjCLYm83jiFQZF/3gY= github.com/google/uuid v1.1.1/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/googleapis/gax-go/v2 v2.0.4/go.mod h1:0Wqv26UfaUD9n4G6kQubkQ+KchISgw+vpHVxEJEs9eg= -github.com/googleapis/gnostic v0.4.0 h1:BXDUo8p/DaxC+4FJY/SSx3gvnx9C1VdHNgaUkiEL5mk= -github.com/googleapis/gnostic v0.4.0/go.mod h1:on+2t9HRStVgn95RSsFWFz+6Q0Snyqv1awfrALZdbtU= -github.com/gophercloud/gophercloud v0.1.0 h1:P/nh25+rzXouhytV2pUHBb65fnds26Ghl8/391+sT5o= -github.com/gophercloud/gophercloud v0.1.0/go.mod h1:vxM41WHh5uqHVBMZHzuwNOHh8XEoIEcSTewFxm1c5g8= +github.com/googleapis/gax-go/v2 v2.0.5/go.mod h1:DWXyrwAJ9X0FpwwEdw+IPEYBICEFu5mhpdKc/us6bOk= +github.com/googleapis/gnostic v0.4.1/go.mod h1:LRhVm6pbyptWbWbuZ38d1eyptfvIytN3ir6b65WBswg= +github.com/googleapis/gnostic v0.5.3 h1:2qsuRm+bzgwSIKikigPASa2GhW8H2Dn4Qq7UxD8K/48= +github.com/googleapis/gnostic v0.5.3/go.mod h1:TRWw1s4gxBGjSe301Dai3c7wXJAZy57+/6tawkOvqHQ= github.com/gorilla/mux v1.7.3 h1:gnP5JzjVOuiZD07fKKToCAOjS0yOpj/qPETTXCCS6hw= github.com/gorilla/mux v1.7.3/go.mod h1:1lud6UwP+6orDFRuTfBEV8e9/aOM/c4fVVCaMa2zaAs= github.com/gorilla/websocket v0.0.0-20170926233335-4201258b820c/go.mod h1:E7qHFY5m1UJ88s3WnNqhKjPHQ0heANvMoAMk2YaljkQ= @@ -268,43 +300,36 @@ github.com/gorilla/websocket v1.4.0/go.mod h1:E7qHFY5m1UJ88s3WnNqhKjPHQ0heANvMoA github.com/gregjones/httpcache v0.0.0-20180305231024-9cad4c3443a7/go.mod h1:FecbI9+v66THATjSRHfNgh1IVFe/9kFxbXtjV0ctIMA= github.com/grpc-ecosystem/go-grpc-middleware v1.0.0/go.mod h1:FiyG127CGDf3tlThmgyCl78X/SZQqEOJBCDaAfeWzPs= github.com/grpc-ecosystem/go-grpc-middleware v1.0.1-0.20190118093823-f849b5445de4/go.mod h1:FiyG127CGDf3tlThmgyCl78X/SZQqEOJBCDaAfeWzPs= -github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0 h1:Ovs26xHkKqVztRpIrF/92BcuyuQ/YW4NSIpoGtfXNho= github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0/go.mod h1:8NvIoxWQoOIhqOTXgfV/d3M/q6VIi02HzZEHgUlZvzk= github.com/grpc-ecosystem/grpc-gateway v1.9.0/go.mod h1:vNeuVxBJEsws4ogUvrchl83t/GYV9WGTSLVdBhOQFDY= github.com/grpc-ecosystem/grpc-gateway v1.9.5/go.mod h1:vNeuVxBJEsws4ogUvrchl83t/GYV9WGTSLVdBhOQFDY= -github.com/grpc-ecosystem/grpc-gateway v1.11.1 h1:/dBYI+n4xIL+Y9SKXQrjlKTmJJDwCSlNLRwZ5nBhIek= github.com/grpc-ecosystem/grpc-gateway v1.11.1/go.mod h1:vNeuVxBJEsws4ogUvrchl83t/GYV9WGTSLVdBhOQFDY= github.com/hashicorp/golang-lru v0.5.0/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8= -github.com/hashicorp/golang-lru v0.5.1 h1:0hERBMJE1eitiLkihrMvRVBYAkpHzc/J3QdDN+dAcgU= github.com/hashicorp/golang-lru v0.5.1/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8= github.com/hashicorp/hcl v1.0.0/go.mod h1:E5yfLk+7swimpb2L/Alb/PJmXilQ/rhwaUYs4T20WEQ= github.com/hpcloud/tail v1.0.0 h1:nfCOvKYfkgYP8hkirhJocXT2+zOD8yUNjXaWfTlyFKI= github.com/hpcloud/tail v1.0.0/go.mod h1:ab1qPbhIpdTxEkNHXyeSf5vhxWSCs/tWer42PpOxQnU= -github.com/imdario/mergo v0.3.5 h1:JboBksRwiiAJWvIYJVo46AfV+IAIKZpfrSzVKj42R4Q= +github.com/ianlancetaylor/demangle v0.0.0-20181102032728-5e5cf60278f6/go.mod h1:aSSvb/t6k1mPoxDqO4vJh6VOCGPwU4O0C2/Eqndh1Sc= github.com/imdario/mergo v0.3.5/go.mod h1:2EnlNZ0deacrJVfApfmtdGgDfMuh/nq6Ok1EcJh5FfA= github.com/imdario/mergo v0.3.9 h1:UauaLniWCFHWd+Jp9oCEkTBj8VO/9DKg3PV3VCNMDIg= github.com/imdario/mergo v0.3.9/go.mod h1:2EnlNZ0deacrJVfApfmtdGgDfMuh/nq6Ok1EcJh5FfA= github.com/inconshreveable/mousetrap v1.0.0 h1:Z8tu5sraLXCXIcARxBp/8cbvlwVa7Z1NHg9XEKhtSvM= github.com/inconshreveable/mousetrap v1.0.0/go.mod h1:PxqpIevigyE2G7u3NXJIT2ANytuPF1OarO4DADm73n8= github.com/jonboulle/clockwork v0.1.0/go.mod h1:Ii8DK3G1RaLaWxj9trq07+26W01tbo22gdxWY5EU2bo= -github.com/json-iterator/go v0.0.0-20180612202835-f2b4162afba3/go.mod h1:+SdeFBvtyEkXs7REEP0seUULqWtbJapLOCVDaaPEHmU= github.com/json-iterator/go v1.1.6/go.mod h1:+SdeFBvtyEkXs7REEP0seUULqWtbJapLOCVDaaPEHmU= github.com/json-iterator/go v1.1.7/go.mod h1:KdQUCv79m/52Kvf8AW2vK1V8akMuk1QjK/uOdHXbAo4= -github.com/json-iterator/go v1.1.8 h1:QiWkFLKq0T7mpzwOTu6BzNDbfTE8OLrYhVKYMLF46Ok= -github.com/json-iterator/go v1.1.8/go.mod h1:KdQUCv79m/52Kvf8AW2vK1V8akMuk1QjK/uOdHXbAo4= -github.com/json-iterator/go v1.1.9 h1:9yzud/Ht36ygwatGx56VwCZtlI/2AD15T1X2sjSuGns= -github.com/json-iterator/go v1.1.9/go.mod h1:KdQUCv79m/52Kvf8AW2vK1V8akMuk1QjK/uOdHXbAo4= +github.com/json-iterator/go v1.1.10 h1:Kz6Cvnvv2wGdaG/V8yMvfkmNiXq9Ya2KUv4rouJJr68= +github.com/json-iterator/go v1.1.10/go.mod h1:KdQUCv79m/52Kvf8AW2vK1V8akMuk1QjK/uOdHXbAo4= github.com/jstemmer/go-junit-report v0.0.0-20190106144839-af01ea7f8024/go.mod h1:6v2b51hI/fHJwM22ozAgKL4VKDeJcHhJFhtBdhmNjmU= +github.com/jstemmer/go-junit-report v0.9.1/go.mod h1:Brl9GWCQeLvo8nXZwPNNblvFj/XSXhF0NWZEnDohbsk= github.com/julienschmidt/httprouter v1.2.0/go.mod h1:SYymIcj16QtmaHHD7aYtjjsJG7VTCxuUUipMqKk8s4w= github.com/kisielk/errcheck v1.1.0/go.mod h1:EZBBE59ingxPouuu3KfxchcWSUPOHkagtvWXihfKN4Q= github.com/kisielk/errcheck v1.2.0/go.mod h1:/BMXB+zMLi60iA8Vv6Ksmxu/1UDYcXs4uQLJ+jE2L00= github.com/kisielk/gotool v1.0.0/go.mod h1:XhKaO+MFFWcvkIS/tQcRk01m1F5IRFswLeQ+oQHNcck= -github.com/konsorten/go-windows-terminal-sequences v1.0.1 h1:mweAR1A6xJ3oS2pRaGiHgQ4OO8tzTaLawm8vnODuwDk= github.com/konsorten/go-windows-terminal-sequences v1.0.1/go.mod h1:T0+1ngSBFLxvqU3pZ+m/2kptfBszLMUkC4ZK/EgS/cQ= -github.com/konsorten/go-windows-terminal-sequences v1.0.2 h1:DB17ag19krx9CFsz4o3enTrPXyIXCl+2iCXH/aMAp9s= -github.com/konsorten/go-windows-terminal-sequences v1.0.2/go.mod h1:T0+1ngSBFLxvqU3pZ+m/2kptfBszLMUkC4ZK/EgS/cQ= +github.com/konsorten/go-windows-terminal-sequences v1.0.3 h1:CE8S1cTafDpPvMhIxNJKvHsGVBgn1xWYf1NbHQhywc8= +github.com/konsorten/go-windows-terminal-sequences v1.0.3/go.mod h1:T0+1ngSBFLxvqU3pZ+m/2kptfBszLMUkC4ZK/EgS/cQ= github.com/kr/logfmt v0.0.0-20140226030751-b84e30acd515/go.mod h1:+0opPa2QZZtGFBFZlji/RkVcI2GknAs/DXo4wKdlNEc= -github.com/kr/pretty v0.1.0 h1:L/CwN0zerZDmRFUapSPitk6f+Q3+0za1rQkzVuMiMFI= github.com/kr/pretty v0.1.0/go.mod h1:dAy3ld7l9f0ibDNOQOHHMYYIIbhfbHSm3C4ZsoJORNo= github.com/kr/pretty v0.2.0 h1:s5hAObm+yFO5uHYt5dYjxi2rXrsnmRpJx4OYvIWUaQs= github.com/kr/pretty v0.2.0/go.mod h1:ipq/a2n7PKx3OHsz4KJII5eveXtPO4qwEXGdVfWzfnI= @@ -312,49 +337,46 @@ github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= github.com/kr/pty v1.1.5/go.mod h1:9r2w37qlBe7rQ6e1fg1S/9xpWHSnaqNdHD3WcMdbPDA= github.com/kr/text v0.1.0 h1:45sCR5RtlFHMR4UwH9sdQ5TC8v0qDQCHnXt+kaKSTVE= github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI= +github.com/liggitt/tabwriter v0.0.0-20181228230101-89fcab3d43de/go.mod h1:zAbeS9B/r2mtpb6U+EI2rYA5OAXxsYw6wTamcNW+zcE= +github.com/lithammer/dedent v1.1.0/go.mod h1:jrXYCQtgg0nJiN+StA2KgR7w6CiQNv9Fd/Z9BP0jIOc= github.com/magiconair/properties v1.8.0/go.mod h1:PppfXfuXeibc/6YijjN8zIbojt8czPbwD3XqdrwzmxQ= github.com/mailru/easyjson v0.0.0-20160728113105-d5b7844b561a/go.mod h1:C1wdFJiN94OJF2b5HbByQZoLdCWB1Yqtg26g4irojpc= -github.com/mailru/easyjson v0.0.0-20180823135443-60711f1a8329 h1:2gxZ0XQIU/5z3Z3bUBu+FXuk2pFbkN6tcwi/pjyaDic= github.com/mailru/easyjson v0.0.0-20180823135443-60711f1a8329/go.mod h1:C1wdFJiN94OJF2b5HbByQZoLdCWB1Yqtg26g4irojpc= github.com/mailru/easyjson v0.0.0-20190312143242-1de009706dbe/go.mod h1:C1wdFJiN94OJF2b5HbByQZoLdCWB1Yqtg26g4irojpc= github.com/mailru/easyjson v0.0.0-20190614124828-94de47d64c63/go.mod h1:C1wdFJiN94OJF2b5HbByQZoLdCWB1Yqtg26g4irojpc= github.com/mailru/easyjson v0.0.0-20190626092158-b2ccc519800e/go.mod h1:C1wdFJiN94OJF2b5HbByQZoLdCWB1Yqtg26g4irojpc= -github.com/mailru/easyjson v0.7.0 h1:aizVhC/NAAcKWb+5QsU1iNOZb4Yws5UO2I+aIprQITM= github.com/mailru/easyjson v0.7.0/go.mod h1:KAzv3t3aY1NaHWoQz1+4F1ccyAH66Jk7yos7ldAVICs= github.com/mailru/easyjson v0.7.1 h1:mdxE1MF9o53iCb2Ghj1VfWvh7ZOwHpnVG/xwXrV90U8= github.com/mailru/easyjson v0.7.1/go.mod h1:KAzv3t3aY1NaHWoQz1+4F1ccyAH66Jk7yos7ldAVICs= github.com/mattn/go-colorable v0.0.9/go.mod h1:9vuHe8Xs5qXnSaW/c/ABM9alt+Vo+STaOChaDxuIBZU= -github.com/mattn/go-colorable v0.1.4/go.mod h1:U0ppj6V5qS13XJ6of8GYAs25YV2eR4EVcfRqFIhoBtE= +github.com/mattn/go-colorable v0.1.1/go.mod h1:FuOcm+DKB9mbwrcAfNl7/TZVBZ6rcnceauSikq3lYCQ= github.com/mattn/go-isatty v0.0.4/go.mod h1:M+lRXTBqGeGNdLjl/ufCoiOlB5xdOkqRJdNxMWT7Zi4= -github.com/mattn/go-isatty v0.0.8/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= +github.com/mattn/go-isatty v0.0.5/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= +github.com/mattn/go-isatty v0.0.7/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= github.com/mattn/go-isatty v0.0.11 h1:FxPOTFNqGkuDUGi3H/qkUbQO4ZiBa2brKq5r0l8TGeM= github.com/mattn/go-isatty v0.0.11/go.mod h1:PhnuNfih5lzO57/f3n+odYbM4JtupLOxQOAqxQCu2WE= github.com/mattn/go-runewidth v0.0.2/go.mod h1:LwmH8dsx7+W8Uxz3IHJYH5QSwggIsqBzpuz5H//U1FU= -github.com/matttproud/golang_protobuf_extensions v1.0.1 h1:4hp9jkHxhMHkqkrB3Ix0jegS5sx/RkqARlsWZ6pIwiU= github.com/matttproud/golang_protobuf_extensions v1.0.1/go.mod h1:D8He9yQNgCq6Z5Ld7szi9bcBfOoFv/3dc6xSMkL2PC0= +github.com/matttproud/golang_protobuf_extensions v1.0.2-0.20181231171920-c182affec369/go.mod h1:BSXmuO+STAnVfrANrmjBb36TMTDstsz7MSK+HVaYKv4= github.com/mitchellh/go-homedir v1.1.0/go.mod h1:SfyaCUpYCn1Vlf4IUYiD9fPX4A5wJrkLzIz1N1q0pr0= github.com/mitchellh/go-wordwrap v1.0.0/go.mod h1:ZXFpozHsX6DPmq2I0TCekCxypsnAUbP2oI0UX1GXzOo= -github.com/mitchellh/mapstructure v1.1.2 h1:fmNYVwqnSfB9mZU6OS2O6GsXM+wcskZDuKQzvN1EDeE= github.com/mitchellh/mapstructure v1.1.2/go.mod h1:FVVH3fgwuzCH5S8UJGiWEs2h04kUh9fWfEaFds41c1Y= +github.com/moby/term v0.0.0-20200312100748-672ec06f55cd/go.mod h1:DdlQx2hp0Ss5/fLikoLlEeIYiATotOjgB//nb973jeo= github.com/modern-go/concurrent v0.0.0-20180228061459-e0a39a4cb421/go.mod h1:6dJC0mAP4ikYIbvyc7fijjWJddQyLn8Ig3JB5CqoB9Q= github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd h1:TRLaZ9cD/w8PVh93nsPXa1VrQ6jlwL5oN8l14QlcNfg= github.com/modern-go/concurrent v0.0.0-20180306012644-bacd9c7ef1dd/go.mod h1:6dJC0mAP4ikYIbvyc7fijjWJddQyLn8Ig3JB5CqoB9Q= -github.com/modern-go/reflect2 v0.0.0-20180320133207-05fbef0ca5da/go.mod h1:bx2lNnkwVCuqBIxFjflWJWanXIb3RllmbCylyMrvgv0= github.com/modern-go/reflect2 v0.0.0-20180701023420-4b7aa43c6742/go.mod h1:bx2lNnkwVCuqBIxFjflWJWanXIb3RllmbCylyMrvgv0= github.com/modern-go/reflect2 v1.0.1 h1:9f412s+6RmYXLWZSEzVVgPGK7C2PphHj5RJrvfx9AWI= github.com/modern-go/reflect2 v1.0.1/go.mod h1:bx2lNnkwVCuqBIxFjflWJWanXIb3RllmbCylyMrvgv0= github.com/morikuni/aec v0.0.0-20170113033406-39771216ff4c/go.mod h1:BbKIizmSmc5MMPqRYbxO4ZU0S0+P200+tUnFx7PXmsc= github.com/munnerz/goautoneg v0.0.0-20120707110453-a547fc61f48d/go.mod h1:+n7T8mK8HuQTcFwEeznm/DIxMOiR9yIdICNftLE1DvQ= -github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822 h1:C3w9PqII01/Oq1c1nUAm88MOHcQC9l5mIlSMApZMrHA= github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822/go.mod h1:+n7T8mK8HuQTcFwEeznm/DIxMOiR9yIdICNftLE1DvQ= github.com/mwitkow/go-conntrack v0.0.0-20161129095857-cc309e4a2223/go.mod h1:qRWi+5nqEBWmkhHvq77mSJWrCKwh8bxhgT7d/eI7P4U= github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= github.com/oklog/ulid v1.3.1/go.mod h1:CirwcVhetQ6Lv90oh/F+FBtV6XMibvdAFo93nm5qn4U= github.com/olekukonko/tablewriter v0.0.0-20170122224234-a0225b3f23b5/go.mod h1:vsDQFd/mU46D+Z4whnwzcISnGGzXWMclvtLoiIKAKIo= github.com/onsi/ginkgo v0.0.0-20170829012221-11459a886d9c/go.mod h1:lLunBs/Ym6LB5Z9jYTR76FiuTmxDTDusOGeTQH+WWjE= -github.com/onsi/ginkgo v1.6.0 h1:Ix8l273rp3QzYgXSR+c8d1fTG7UPgYkOSELPhiY/YGw= github.com/onsi/ginkgo v1.6.0/go.mod h1:lLunBs/Ym6LB5Z9jYTR76FiuTmxDTDusOGeTQH+WWjE= -github.com/onsi/ginkgo v1.10.1/go.mod h1:lLunBs/Ym6LB5Z9jYTR76FiuTmxDTDusOGeTQH+WWjE= github.com/onsi/ginkgo v1.11.0 h1:JAKSXpt1YjtLA7YpPiqO9ss6sNXEsPfSGdwN0UHqzrw= github.com/onsi/ginkgo v1.11.0/go.mod h1:lLunBs/Ym6LB5Z9jYTR76FiuTmxDTDusOGeTQH+WWjE= github.com/onsi/gomega v0.0.0-20170829124025-dcabb60a477c/go.mod h1:C1qb7wdrVGGVU+Z6iS04AVkA3Q65CEZX59MT0QO5uiA= @@ -368,127 +390,126 @@ github.com/opencontainers/runc v0.1.1 h1:GlxAyO6x8rfZYN9Tt0Kti5a/cP41iuiO2yYT0IJ github.com/opencontainers/runc v0.1.1/go.mod h1:qT5XzbpPznkRYVz/mWwUaVBUv2rmF59PVA73FjuZG0U= github.com/pborman/uuid v1.2.0/go.mod h1:X/NO0urCmaxf9VXbdlT7C2Yzkj2IKimNn4k+gtPdI/k= github.com/pelletier/go-toml v1.2.0/go.mod h1:5z9KED0ma1S8pY6P1sdut58dfprrGBbd/94hg7ilaic= -github.com/peterbourgon/diskv v2.0.1+incompatible h1:UBdAOUP5p4RWqPBg048CAvpKN+vxiaj6gdUUzhl4XmI= github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= github.com/peterhellberg/link v1.0.0 h1:mUWkiegowUXEcmlb+ybF75Q/8D2Y0BjZtR8cxoKhaQo= github.com/peterhellberg/link v1.0.0/go.mod h1:gtSlOT4jmkY8P47hbTc8PTgiDDWpdPbFYl75keYyBB8= github.com/pkg/errors v0.8.0/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= -github.com/pkg/errors v0.8.1 h1:iURUrRGxPUNPdy5/HRSm+Yj6okJ6UtLINN0Q9M4+h3I= github.com/pkg/errors v0.8.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= -github.com/pmezard/go-difflib v0.0.0-20151028094244-d8ed2627bdf0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= github.com/pquerna/cachecontrol v0.0.0-20171018203845-0dec1b30a021/go.mod h1:prYjPmNq4d1NPVmpShWobRqXY3q7Vp+80DqgxxUrUIA= github.com/prometheus/client_golang v0.9.1/go.mod h1:7SWBe2y4D6OKWSNQJUaRYU/AaXPKyh/dDVn+NZz0KFw= github.com/prometheus/client_golang v0.9.3/go.mod h1:/TN21ttK/J9q6uSwhBd54HahCDft0ttaMvbicHlPoso= -github.com/prometheus/client_golang v1.0.0 h1:vrDKnkGzuGvhNAL56c7DBz29ZL+KxnoR0x7enabFceM= github.com/prometheus/client_golang v1.0.0/go.mod h1:db9x61etRT2tGnBNRi70OPL5FsnadC4Ky3P0J6CfImo= +github.com/prometheus/client_golang v1.7.1/go.mod h1:PY5Wy2awLA44sXw4AOSfFBetzPP4j5+D6mVACh+pe2M= github.com/prometheus/client_model v0.0.0-20180712105110-5c3871d89910/go.mod h1:MbSGuTsp3dbXC40dX6PRTWyKYBIrTGTE9sqQNg2J8bo= -github.com/prometheus/client_model v0.0.0-20190129233127-fd36f4220a90 h1:S/YWwWx/RA8rT8tKFRuGUZhuA90OyIBpPCXkcbwU8DE= github.com/prometheus/client_model v0.0.0-20190129233127-fd36f4220a90/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA= github.com/prometheus/client_model v0.0.0-20190812154241-14fe0d1b01d4/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA= +github.com/prometheus/client_model v0.2.0/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA= github.com/prometheus/common v0.0.0-20181113130724-41aa239b4cce/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro= github.com/prometheus/common v0.4.0/go.mod h1:TNfzLD0ON7rHzMJeJkieUDPYmFC7Snx/y86RQel1bk4= -github.com/prometheus/common v0.4.1 h1:K0MGApIoQvMw27RTdJkPbr3JZ7DNbtxQNyi5STVM6Kw= github.com/prometheus/common v0.4.1/go.mod h1:TNfzLD0ON7rHzMJeJkieUDPYmFC7Snx/y86RQel1bk4= +github.com/prometheus/common v0.10.0/go.mod h1:Tlit/dnDKsSWFlCLTWaA1cyBgKHSMdTB80sz/V91rCo= github.com/prometheus/procfs v0.0.0-20181005140218-185b4288413d/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk= github.com/prometheus/procfs v0.0.0-20190507164030-5867b95ac084/go.mod h1:TjEm7ze935MbeOT/UhFTIMYKhuLP4wbCsTZCD3I8kEA= -github.com/prometheus/procfs v0.0.2 h1:6LJUbpNm42llc4HRCuvApCSWB/WfhuNo9K98Q9sNGfs= github.com/prometheus/procfs v0.0.2/go.mod h1:TjEm7ze935MbeOT/UhFTIMYKhuLP4wbCsTZCD3I8kEA= +github.com/prometheus/procfs v0.1.3/go.mod h1:lV6e/gmhEcM9IjHGsFOCxxuZ+z1YqCvr4OA4YeYWdaU= github.com/prometheus/tsdb v0.7.1/go.mod h1:qhTCs0VvXwvX/y3TZrWD7rabWM+ijKTux40TwIPHuXU= -github.com/remyoudompheng/bigfft v0.0.0-20170806203942-52369c62f446/go.mod h1:uYEyJGbgTkfkS4+E/PavXkNJcbFIpEtjt2B0KDQ5+9M= github.com/rogpeppe/fastuuid v0.0.0-20150106093220-6724a57986af/go.mod h1:XWv6SoW27p1b0cqNHllgS5HIMJraePCO15w5zCzIWYg= +github.com/rogpeppe/go-internal v1.3.0/go.mod h1:M8bDsm7K2OlrFYOpmOWEs/qY81heoFRclV5y23lUDJ4= github.com/russross/blackfriday v1.5.2/go.mod h1:JO/DiYxRf+HjHt06OyowR9PTA263kcR/rfWxYHBV53g= github.com/russross/blackfriday/v2 v2.0.1/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM= github.com/sergi/go-diff v1.0.0 h1:Kpca3qRNrduNnOQeazBd0ysaKrUJiIuISHxogkT9RPQ= github.com/sergi/go-diff v1.0.0/go.mod h1:0CfEIISq7TuYL3j771MWULgwwjU+GofnZX9QAmXWZgo= -github.com/sergi/go-diff v1.1.0 h1:we8PVUC3FE2uYfodKH/nBHMSetSfHDR6scGdBi+erh0= -github.com/sergi/go-diff v1.1.0/go.mod h1:STckp+ISIX8hZLjrqAeVduY0gWCT9IjLuqbuNXdaHfM= -github.com/shurcooL/httpfs v0.0.0-20190707220628-8d4bc4ba7749 h1:bUGsEnyNbVPw06Bs80sCeARAlK8lhwqGyi6UT8ymuGk= github.com/shurcooL/httpfs v0.0.0-20190707220628-8d4bc4ba7749/go.mod h1:ZY1cvUeJuFPAdZ/B6v7RHavJWZn2YPVFQ1OSXhCGOkg= github.com/shurcooL/sanitized_anchor_name v1.0.0/go.mod h1:1NzhyTcUVG4SuEtjjoZeVRXNmyL/1OwPU0+IJeTBvfc= github.com/sirupsen/logrus v1.2.0/go.mod h1:LxeOpSwHxABJmUn/MG1IvRgCAasNZTLOkJPxbbu5VWo= github.com/sirupsen/logrus v1.4.1/go.mod h1:ni0Sbl8bgC9z8RoU9G6nDWqqs/fq4eDPysMBDgk/93Q= -github.com/sirupsen/logrus v1.4.2 h1:SPIRibHv4MatM3XXNO2BJeFLZwZ2LvZgfQ5+UNI2im4= github.com/sirupsen/logrus v1.4.2/go.mod h1:tLMulIdttU9McNUspp0xgXVQah82FyeX6MwdIuYE2rE= +github.com/sirupsen/logrus v1.6.0 h1:UBcNElsrwanuuMsnGSlYmtmgbb23qDR5dG+6X6Oo89I= +github.com/sirupsen/logrus v1.6.0/go.mod h1:7uNnSEd1DgxDLC74fIahvMZmmYsHGZGEOFrfsX/uA88= github.com/soheilhy/cmux v0.1.4/go.mod h1:IM3LyeVVIOuxMH7sFAkER9+bJ4dT7Ms6E4xg4kGIyLM= github.com/spaolacci/murmur3 v0.0.0-20180118202830-f09979ecbc72/go.mod h1:JwIasOWyU6f++ZhiEuf87xNszmSA2myDM2Kzu9HwQUA= github.com/spf13/afero v1.1.2/go.mod h1:j4pytiNVoe2o6bmDsKpLACNPDBIoEAkihy7loJ1B0CQ= github.com/spf13/afero v1.2.2/go.mod h1:9ZxEEn6pIJ8Rxe320qSDBk6AsU0r9pR7Q4OcevTdifk= github.com/spf13/cast v1.3.0/go.mod h1:Qx5cxh0v+4UWYiBimWS+eyWzqEqokIECu5etghLkUJE= github.com/spf13/cobra v0.0.3/go.mod h1:1l0Ry5zgKvJasoi3XT1TypsSe7PqH0Sj9dhYf7v3XqQ= -github.com/spf13/cobra v0.0.5 h1:f0B+LkLX6DtmRH1isoNA9VTtNUK9K8xYd28JNNfOv/s= -github.com/spf13/cobra v0.0.5/go.mod h1:3K3wKZymM7VvHMDS9+Akkh4K60UwM26emMESw8tLCHU= -github.com/spf13/cobra v0.0.7 h1:FfTH+vuMXOas8jmfb5/M7dzEYx7LpcLb7a0LPe34uOU= -github.com/spf13/cobra v0.0.7/go.mod h1:/6GTrnGXV9HjY+aR4k0oJ5tcvakLuG6EuKReYlHNrgE= +github.com/spf13/cobra v1.0.0 h1:6m/oheQuQ13N9ks4hubMG6BnvwOeaJrqSPLahSnczz8= +github.com/spf13/cobra v1.0.0/go.mod h1:/6GTrnGXV9HjY+aR4k0oJ5tcvakLuG6EuKReYlHNrgE= github.com/spf13/jwalterweatherman v1.0.0/go.mod h1:cQK4TGJAtQXfYWX+Ddv3mKDzgVb68N+wFjFa4jdeBTo= github.com/spf13/pflag v0.0.0-20170130214245-9ff6c6923cff/go.mod h1:DYY7MBk1bdzusC3SYhjObp+wFpr4gzcvqqNjLnInEg4= github.com/spf13/pflag v1.0.1/go.mod h1:DYY7MBk1bdzusC3SYhjObp+wFpr4gzcvqqNjLnInEg4= github.com/spf13/pflag v1.0.3/go.mod h1:DYY7MBk1bdzusC3SYhjObp+wFpr4gzcvqqNjLnInEg4= github.com/spf13/pflag v1.0.5 h1:iy+VFUOCP1a+8yFto/drg2CJ5u0yRoB7fZw3DKv/JXA= github.com/spf13/pflag v1.0.5/go.mod h1:McXfInJRrz4CZXVZOBLb0bTZqETkiAhM9Iw0y3An2Bg= -github.com/spf13/viper v1.3.2/go.mod h1:ZiWeW+zYFKm7srdB9IoDzzZXaJaI5eL9QjNiN/DMA2s= github.com/spf13/viper v1.4.0/go.mod h1:PTJ7Z/lr49W6bUbkmS1V3by4uWynFiR9p7+dSq/yZzE= +github.com/stoewer/go-strcase v1.2.0/go.mod h1:IBiWB2sKIp3wVVQ3Y035++gc+knqhUQag1KpM8ahLw8= github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME= github.com/stretchr/objx v0.1.1/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME= github.com/stretchr/objx v0.2.0/go.mod h1:qt09Ya8vawLte6SNmTgCsAVtYtaKzEcn8ATUoHMkEqE= -github.com/stretchr/testify v0.0.0-20151208002404-e3a8ff8ce365/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= -github.com/stretchr/testify v1.2.2 h1:bSDNvY7ZPG5RlJ8otE/7V6gMiyenm9RtJ7IUVIAoJ1w= github.com/stretchr/testify v1.2.2/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= -github.com/stretchr/testify v1.3.0 h1:TivCn/peBQ7UY8ooIcPgZFpTNSz0Q2U6UrFlUfqbe0Q= github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI= -github.com/stretchr/testify v1.4.0 h1:2E4SXV/wtOkTonXsotYi4li6zVWxYlZuYNCXe9XRJyk= github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4= +github.com/stretchr/testify v1.5.1 h1:nOGnQDM7FYENwehXlg/kFVnos3rEvtKTjRvOWSzb6H4= +github.com/stretchr/testify v1.5.1/go.mod h1:5W2xD1RspED5o8YsWQXVCued0rvSQ+mT+I5cxcmMvtA= github.com/tidwall/pretty v1.0.0/go.mod h1:XNkn88O1ChpSDQmQeStsy+sBenx6DDtFZJxhVysOjyk= github.com/tmc/grpc-websocket-proxy v0.0.0-20170815181823-89b8d40f7ca8/go.mod h1:ncp9v5uamzpCO7NfCPTXjqaC+bZgJeR0sMTm6dMHP7U= github.com/tmc/grpc-websocket-proxy v0.0.0-20190109142713-0ad062ec5ee5/go.mod h1:ncp9v5uamzpCO7NfCPTXjqaC+bZgJeR0sMTm6dMHP7U= github.com/ugorji/go v1.1.4/go.mod h1:uQMGLiO92mf5W77hV/PUCpI3pbzQx3CRekS0kk+RGrc= -github.com/ugorji/go/codec v0.0.0-20181204163529-d75b2dcb6bc8/go.mod h1:VFNgLljTbGfSG7qAOspJ7OScBnGdDN/yBr0sguwnwf0= github.com/urfave/cli v1.20.0/go.mod h1:70zkFmudgCuE/ngEzBv17Jvp/497gISqfk5gWijbERA= github.com/vektah/gqlparser v1.1.2/go.mod h1:1ycwN7Ij5njmMkPPAOaRFY4rET2Enx7IkVv3vaXspKw= github.com/xiang90/probing v0.0.0-20190116061207-43a291ad63a2/go.mod h1:UETIi67q53MR2AWcXfiuqkDkRtnGDLqkBTpCHuJHxtU= +github.com/xlab/handysort v0.0.0-20150421192137-fb3537ed64a1/go.mod h1:QcJo0QPSfTONNIgpN5RA8prR7fF8nkF6cTWTcNerRO8= github.com/xordataexchange/crypt v0.0.3-0.20170626215501-b2862e3d0a77/go.mod h1:aYKd//L2LvnjZzWKhF00oedf4jCCReLcmhLdhm1A27Q= +github.com/yuin/goldmark v1.1.27/go.mod h1:3hX8gzYuyVAZsxl0MRgGTJEmQBFcNTphYh9decYSb74= go.etcd.io/bbolt v1.3.2/go.mod h1:IbVyRI1SCnLcuJnV2u8VeU0CEYM7e686BmAb1XKL+uU= go.etcd.io/bbolt v1.3.3/go.mod h1:IbVyRI1SCnLcuJnV2u8VeU0CEYM7e686BmAb1XKL+uU= -go.etcd.io/etcd v0.0.0-20191023171146-3cf2f69b5738 h1:VcrIfasaLFkyjk6KNlXQSzO+B0fZcnECiDrKJsfxka0= -go.etcd.io/etcd v0.0.0-20191023171146-3cf2f69b5738/go.mod h1:dnLIgRNXwCJa5e+c6mIZCrds/GIG4ncV9HhK5PX7jPg= +go.etcd.io/bbolt v1.3.5/go.mod h1:G5EMThwa9y8QZGBClrRx5EY+Yw9kAhnjy3bSjsnlVTQ= +go.etcd.io/etcd v0.5.0-alpha.5.0.20200819165624-17cef6e3e9d5/go.mod h1:skWido08r9w6Lq/w70DO5XYIKMu4QFu1+4VsqLQuJy8= go.mongodb.org/mongo-driver v1.0.3/go.mod h1:u7ryQJ+DOzQmeO7zB6MHyr8jkEQvC8vH7qLUO4lqsUM= go.mongodb.org/mongo-driver v1.1.1/go.mod h1:u7ryQJ+DOzQmeO7zB6MHyr8jkEQvC8vH7qLUO4lqsUM= -go.mongodb.org/mongo-driver v1.1.2 h1:jxcFYjlkl8xaERsgLo+RNquI0epW6zuy/ZRQs6jnrFA= go.mongodb.org/mongo-driver v1.1.2/go.mod h1:u7ryQJ+DOzQmeO7zB6MHyr8jkEQvC8vH7qLUO4lqsUM= go.opencensus.io v0.21.0/go.mod h1:mSImk1erAIZhrmZN+AvHh14ztQfjbGwt4TtuofqLduU= -go.uber.org/atomic v1.3.2 h1:2Oa65PReHzfn29GpvgsYwloV9AVFHPDk8tYxt2c2tr4= +go.opencensus.io v0.22.0/go.mod h1:+kGneAE2xo2IficOXnaByMWTGM9T73dGwxeWcUqIpI8= +go.opencensus.io v0.22.2/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw= go.uber.org/atomic v1.3.2/go.mod h1:gD2HeocX3+yG+ygLZcrzQJaqmWj9AIm7n08wl/qW/PE= go.uber.org/atomic v1.4.0/go.mod h1:gD2HeocX3+yG+ygLZcrzQJaqmWj9AIm7n08wl/qW/PE= -go.uber.org/multierr v1.1.0 h1:HoEmRHQPVSqub6w2z2d2EOVs2fjyFRGyofhKuyDq0QI= go.uber.org/multierr v1.1.0/go.mod h1:wR5kodmAFQ0UK8QlbwjlSNy0Z68gJhDJUG5sjR94q/0= -go.uber.org/zap v1.10.0 h1:ORx85nbTijNz8ljznvCMR1ZBIPKFn3jQrag10X2AsuM= go.uber.org/zap v1.10.0/go.mod h1:vwi/ZaCAaUcBkycHslxD9B2zi4UTXhF60s6SWpuDF0Q= golang.org/x/crypto v0.0.0-20180904163835-0709b304e793/go.mod h1:6SG95UA2DQfeDnfUPMdvaQW0Q7yPrPDi9nlGo2tz2b4= -golang.org/x/crypto v0.0.0-20181203042331-505ab145d0a9/go.mod h1:6SG95UA2DQfeDnfUPMdvaQW0Q7yPrPDi9nlGo2tz2b4= -golang.org/x/crypto v0.0.0-20190211182817-74369b46fc67/go.mod h1:6SG95UA2DQfeDnfUPMdvaQW0Q7yPrPDi9nlGo2tz2b4= golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w= golang.org/x/crypto v0.0.0-20190320223903-b7391e95e576/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w= +golang.org/x/crypto v0.0.0-20190510104115-cbcb75029529/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= +golang.org/x/crypto v0.0.0-20190605123033-f99c8df09eb5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190611184440-5c40567a22f8/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190617133340-57b3e21c3d56/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= -golang.org/x/crypto v0.0.0-20191206172530-e9b2fee46413 h1:ULYEB3JvPRE/IfO+9uO7vKV/xzVTO7XPAwm8xbf4w2g= +golang.org/x/crypto v0.0.0-20191011191535-87dc89f01550/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20191206172530-e9b2fee46413/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto= -golang.org/x/crypto v0.0.0-20200220183623-bac4c82f6975 h1:/Tl7pH94bvbAAHBdZJT947M/+gp0+CqQXDtMRC0fseo= -golang.org/x/crypto v0.0.0-20200220183623-bac4c82f6975/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto= -golang.org/x/crypto v0.0.0-20200406173513-056763e48d71 h1:DOmugCavvUtnUD114C1Wh+UgTgQZ4pMLzXxi1pSt+/Y= -golang.org/x/crypto v0.0.0-20200406173513-056763e48d71/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto= +golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9 h1:psW17arqaxU48Z5kZ0CQnkZWQJsqcURM6tKiBApRjXI= +golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto= golang.org/x/exp v0.0.0-20190121172915-509febef88a4/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA= -golang.org/x/exp v0.0.0-20190125153040-c74c464bbbf2/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA= -golang.org/x/exp v0.0.0-20190312203227-4b39c73a6495/go.mod h1:ZjyILWgesfNpC6sMxTJOJm9Kp84zZh5NQWvqDGG3Qr8= +golang.org/x/exp v0.0.0-20190306152737-a1d7652674e8/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA= +golang.org/x/exp v0.0.0-20190510132918-efd6b22b2522/go.mod h1:ZjyILWgesfNpC6sMxTJOJm9Kp84zZh5NQWvqDGG3Qr8= +golang.org/x/exp v0.0.0-20190829153037-c13cbed26979/go.mod h1:86+5VVa7VpoJ4kLfm080zCjGlMRFzhUhsZKEZO7MGek= +golang.org/x/exp v0.0.0-20191227195350-da58074b4299/go.mod h1:2RIsYlXP63K8oxa1u096TMicItID8zy7Y6sNkU49FU4= golang.org/x/image v0.0.0-20190227222117-0694c2d4d067/go.mod h1:kZ7UVZpmo3dzQBMxlp+ypCbDeSB+sBbTgSJuh5dn5js= +golang.org/x/image v0.0.0-20190802002840-cff245a6509b/go.mod h1:FeLwcggjj3mMvU+oOTbSwawSJRM1uh48EjtB4UJZlP0= golang.org/x/lint v0.0.0-20181026193005-c67002cb31c3/go.mod h1:UVdnD1Gm6xHRNCYTkRU2/jEulfH38KcIWyp/GAMgvoE= golang.org/x/lint v0.0.0-20190227174305-5b3e6a55c961/go.mod h1:wehouNa3lNwaWXcvxsM5YxQ5yQlVC4a0KAMCusXpPoU= golang.org/x/lint v0.0.0-20190301231843-5614ed5bae6f/go.mod h1:UVdnD1Gm6xHRNCYTkRU2/jEulfH38KcIWyp/GAMgvoE= golang.org/x/lint v0.0.0-20190313153728-d0100b6bd8b3/go.mod h1:6SW0HCj/g11FgYtHlgUYUwCkIfeOF89ocIRzGO/8vkc= +golang.org/x/lint v0.0.0-20190409202823-959b441ac422/go.mod h1:6SW0HCj/g11FgYtHlgUYUwCkIfeOF89ocIRzGO/8vkc= +golang.org/x/lint v0.0.0-20190909230951-414d861bb4ac/go.mod h1:6SW0HCj/g11FgYtHlgUYUwCkIfeOF89ocIRzGO/8vkc= +golang.org/x/lint v0.0.0-20191125180803-fdd1cda4f05f/go.mod h1:5qLYkcX4OjUUV8bRuDixDT3tpyyb+LUpUlRWLxfhWrs= golang.org/x/mobile v0.0.0-20190312151609-d3739f865fa6/go.mod h1:z+o9i4GpDbdi3rU15maQ/Ox0txvL9dWGYEHz965HBQE= -golang.org/x/net v0.0.0-20170114055629-f2499483f923/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= +golang.org/x/mobile v0.0.0-20190719004257-d2bd2a29d028/go.mod h1:E/iHnbuqvinMTCcRqshq8CkpyQDoeVncDDYHnLhea+o= +golang.org/x/mod v0.0.0-20190513183733-4bf6d317e70e/go.mod h1:mXi4GBBbnImb6dmsKGUJ2LatrhH/nqhxcFungHvyanc= +golang.org/x/mod v0.1.0/go.mod h1:0QHyrYULN0/3qlju5TqG8bIK38QM8yzMo5ekMj3DlcY= +golang.org/x/mod v0.1.1-0.20191105210325-c90efee705ee/go.mod h1:QqPTAvyqsEbceGzBzNggFXnrqF1CaUcvgkdR5Ot7KZg= +golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= +golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= @@ -497,65 +518,70 @@ golang.org/x/net v0.0.0-20181114220301-adae6a3d119a/go.mod h1:mL1N/T3taQHkDXs73r golang.org/x/net v0.0.0-20181220203305-927f97764cc3/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190213061140-3a22650c66bd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= -golang.org/x/net v0.0.0-20190311183353-d8887717615a h1:oWX7TPOiFAMXLq8o0ikBYfCJVlRHBcsciT5bXOrH628= golang.org/x/net v0.0.0-20190311183353-d8887717615a/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190320064053-1272bf9dcd53/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= -golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3 h1:0GoQqolDA55aaLxZyTzK/Y2ePZzZTUrRacwib7cNsYQ= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= +golang.org/x/net v0.0.0-20190501004415-9ce7a6920f09/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= +golang.org/x/net v0.0.0-20190503192946-f4e77d36d62c/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190522155817-f3200d17e092/go.mod h1:HSz+uSET+XFnRR8LxR5pz3Of3rY3CfYBVs4xY44aLks= golang.org/x/net v0.0.0-20190603091049-60506f45cf65/go.mod h1:HSz+uSET+XFnRR8LxR5pz3Of3rY3CfYBVs4xY44aLks= golang.org/x/net v0.0.0-20190613194153-d28f0bde5980/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190813141303-74dc4d7220e7/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190827160401-ba9fcec4b297/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= -golang.org/x/net v0.0.0-20191004110552-13f9640d40b9 h1:rjwSpXsdiK0dV8/Naq3kAw9ymfAeJIyd0upUIElB+lI= -golang.org/x/net v0.0.0-20191004110552-13f9640d40b9/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= -golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e h1:3G+cUijn7XD+S4eJFddp53Pv7+slrESplyjG25HgL+k= +golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= +golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= -golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be h1:vEDujvNQGv4jgYKudGeI/+DAX4Jffq6hpD55MmoEvKs= +golang.org/x/net v0.0.0-20200707034311-ab3426394381 h1:VXak5I6aEWmAXeQjA+QSZzlgNrpq9mjcfDemuexIKsU= +golang.org/x/net v0.0.0-20200707034311-ab3426394381/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= golang.org/x/oauth2 v0.0.0-20190226205417-e64efc72b421/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= -golang.org/x/oauth2 v0.0.0-20190604053449-0f29369cfe45 h1:SVwTIAaPC2U/AvvLNZ2a7OVsmBpC8L5BlwK1whH3hm0= golang.org/x/oauth2 v0.0.0-20190604053449-0f29369cfe45/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= +golang.org/x/oauth2 v0.0.0-20191202225959-858c2ad4c8b6/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= golang.org/x/oauth2 v0.0.0-20200107190931-bf48bf16ab8d h1:TzXSXBo42m9gQenoE3b9BGiEpg5IG2JkU5FkPIawgtw= golang.org/x/oauth2 v0.0.0-20200107190931-bf48bf16ab8d/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= -golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f h1:wMNYb4v58l5UBM7MYRLPG6ZhfOqbKu7X5eyFl8ZhKvA= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181108010431-42b317875d0f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20181221193216-37e7f081c4d4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190227155943-e225da77a7e6/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= -golang.org/x/sys v0.0.0-20170830134202-bb24a47a89ea/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= +golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sys v0.0.0-20180830151530-49385e6e1522/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20180905080454-ebe1bf3edb33/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20180909124046-d0be0721c37e/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20181107165924-66b7b1311ac8/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20181116152217-5ac8a444bdc5/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= -golang.org/x/sys v0.0.0-20181205085412-a5c9d58dba9a/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= -golang.org/x/sys v0.0.0-20190209173611-3b5209105503/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20190215142949-d0b11bdaac8a/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20190312061237-fead79001313/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190321052220-f7bb7a8bee54/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190412213103-97732733099d/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190422165155-953cdadca894/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20190502145724-3ef323f4f1fd/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190507160741-ecd444e8653b/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20190531175056-4c3a928424d2/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20190606165138-5da285871e9c/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190616124812-15dcb6c0061f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20190826190057-c7b8b68b1456 h1:ng0gs1AKnRRuEMZoTLLlbOd+C17zUDepwGQBb/n+JVg= +golang.org/x/sys v0.0.0-20190624142023-c5567b49c5d0/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190826190057-c7b8b68b1456/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20191005200804-aed5e4c7ecf9/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20191026070338-33540a1f6037/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20191204072324-ce4227a45e2e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20191228213918-04cbcbbfeed8/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20200106162015-b016eb3dc98e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20200202164722-d101bd2416d5/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20200302150141-5c8b2ff67527/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20200323222414-85ca7c5b95cd/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20200409092240-59c9f1ba88fa h1:mQTN3ECqfsViCNBgq+A40vdwhkGykrrQlYe3mPj6BoU= -golang.org/x/sys v0.0.0-20200409092240-59c9f1ba88fa/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/text v0.0.0-20160726164857-2910a502d2bf/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= -golang.org/x/text v0.3.0 h1:g61tztE5qeGQ89tm6NTjjM9VPIm088od1l6aSorWRWg= +golang.org/x/sys v0.0.0-20200615200032-f1bc736245b1/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= +golang.org/x/sys v0.0.0-20200622214017-ed371f2e16b4 h1:5/PjkGUjvEU5Gl6BxmvKRPpqo2uNMv4rcHBMwzk/st8= +golang.org/x/sys v0.0.0-20200622214017-ed371f2e16b4/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= golang.org/x/text v0.3.1-0.20180807135948-17ff2d5776d2/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= -golang.org/x/text v0.3.2 h1:tW2bmiBqwgJj/UpqtC8EpXEZVYOwU0yG4iWbprSVAcs= golang.org/x/text v0.3.2/go.mod h1:bEr9sfX3Q8Zfm5fL9x+3itogRgK3+ptLWKqgva+5dAk= +golang.org/x/text v0.3.3 h1:cokOdA+Jmi5PJGXLlLllQSgYigAEfHXJAERHVMaCc2k= +golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= golang.org/x/time v0.0.0-20180412165947-fbb02b2291d2/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= golang.org/x/time v0.0.0-20181108054448-85acf8d2951c/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= -golang.org/x/time v0.0.0-20190308202827-9d24e82272b4 h1:SvFZT6jyqRaOeXpc5h/JSfZenJ2O330aBsf7JfSUXmQ= golang.org/x/time v0.0.0-20190308202827-9d24e82272b4/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= golang.org/x/time v0.0.0-20191024005414-555d28b269f0 h1:/5xXl8Y5W96D+TtHSlonuFqGHIWVuyCkGJLwGh9JJFs= golang.org/x/time v0.0.0-20191024005414-555d28b269f0/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= @@ -565,113 +591,140 @@ golang.org/x/tools v0.0.0-20181011042414-1f849cf54d09/go.mod h1:n7NCudcB/nEzxVGm golang.org/x/tools v0.0.0-20181030221726-6c7e314b6563/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ= golang.org/x/tools v0.0.0-20190114222345-bf090417da8b/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ= golang.org/x/tools v0.0.0-20190125232054-d66bd3c5d5a6/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ= -golang.org/x/tools v0.0.0-20190206041539-40960b6deb8e/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ= golang.org/x/tools v0.0.0-20190226205152-f727befe758c/go.mod h1:9Yl7xja0Znq3iFh3HoIrodX9oNMXvdceNzlUR8zjMvY= golang.org/x/tools v0.0.0-20190311212946-11955173bddd/go.mod h1:LCzVGOaR6xXOjkQ3onu1FJEFr0SW1gC7cKk1uF8kGRs= golang.org/x/tools v0.0.0-20190312151545-0bb0c0a6e846/go.mod h1:LCzVGOaR6xXOjkQ3onu1FJEFr0SW1gC7cKk1uF8kGRs= golang.org/x/tools v0.0.0-20190312170243-e65039ee4138/go.mod h1:LCzVGOaR6xXOjkQ3onu1FJEFr0SW1gC7cKk1uF8kGRs= +golang.org/x/tools v0.0.0-20190425150028-36563e24a262/go.mod h1:RgjU9mgBXZiqYHBnxXauZ1Gv1EHHAz9KjViQ78xBX0Q= +golang.org/x/tools v0.0.0-20190506145303-2d16b83fe98c/go.mod h1:RgjU9mgBXZiqYHBnxXauZ1Gv1EHHAz9KjViQ78xBX0Q= golang.org/x/tools v0.0.0-20190524140312-2c0ae7006135/go.mod h1:RgjU9mgBXZiqYHBnxXauZ1Gv1EHHAz9KjViQ78xBX0Q= +golang.org/x/tools v0.0.0-20190606124116-d0a3d012864b/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc= golang.org/x/tools v0.0.0-20190614205625-5aca471b1d59/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc= golang.org/x/tools v0.0.0-20190617190820-da514acc4774/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc= -golang.org/x/tools v0.0.0-20190920225731-5eefd052ad72/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= +golang.org/x/tools v0.0.0-20190621195816-6e04913cbbac/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc= +golang.org/x/tools v0.0.0-20190624222133-a101b041ded4/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc= +golang.org/x/tools v0.0.0-20190628153133-6cdbf07be9d0/go.mod h1:/rFqwRUd4F7ZHNgwSSTFct+R/Kf4OFW1sUzUTQQTgfc= +golang.org/x/tools v0.0.0-20190816200558-6889da9d5479/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= +golang.org/x/tools v0.0.0-20190911174233-4f2ddba30aff/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= +golang.org/x/tools v0.0.0-20191012152004-8de300cfc20a/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= +golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= +golang.org/x/tools v0.0.0-20191125144606-a911d9008d1f/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= +golang.org/x/tools v0.0.0-20191227053925-7b8e75db28f4/go.mod h1:TB2adYChydJhpapKDTa4BR/hXlZSLoq2Wpct/0txZ28= +golang.org/x/tools v0.0.0-20200616133436-c1934b75d054/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= +golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543 h1:E7g+9GITq07hpfrRu66IVDexMakfv52eLZ2CXBWiKr4= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= -gonum.org/v1/gonum v0.0.0-20190331200053-3d26580ed485/go.mod h1:2ltnJ7xHfj0zHS40VVPYEAAMTa3ZGguvHGBSJeRWqE0= -gonum.org/v1/netlib v0.0.0-20190313105609-8cb42192e0e0/go.mod h1:wa6Ws7BG/ESfp6dHfk7C6KdzKA7wR7u/rKwOGE66zvw= -gonum.org/v1/netlib v0.0.0-20190331212654-76723241ea4e/go.mod h1:kS+toOQn6AQKjmKJ7gzohV1XkqsFehRA2FbsbkopSuQ= google.golang.org/api v0.4.0/go.mod h1:8k5glujaEP+g9n7WNsDg8QP6cUVNI86fCNMcbazEtwE= -google.golang.org/appengine v1.1.0 h1:igQkv0AAhEIvTEpD5LIpAfav2eeVO9HBTjvKHVJPRSs= +google.golang.org/api v0.7.0/go.mod h1:WtwebWUNSVBH/HAw79HIFXZNqEvBhG+Ra+ax0hx3E3M= +google.golang.org/api v0.8.0/go.mod h1:o4eAsZoiT+ibD93RtjEohWalFOjRDx6CVaqeizhEnKg= +google.golang.org/api v0.9.0/go.mod h1:o4eAsZoiT+ibD93RtjEohWalFOjRDx6CVaqeizhEnKg= +google.golang.org/api v0.15.0/go.mod h1:iLdEw5Ide6rF15KTC1Kkl0iskquN2gFfn9o9XIsbkAI= google.golang.org/appengine v1.1.0/go.mod h1:EbEs0AVv82hx2wNQdGPgUI5lhzA/G0D9YwlJXL52JkM= google.golang.org/appengine v1.4.0/go.mod h1:xpcJRLb0r/rnEns0DIKYYv+WjYCduHsrkT7/EB5XEv4= -google.golang.org/appengine v1.5.0 h1:KxkO13IPW4Lslp2bz+KHP2E3gtFlrIGNThxkZQ3g+4c= google.golang.org/appengine v1.5.0/go.mod h1:xpcJRLb0r/rnEns0DIKYYv+WjYCduHsrkT7/EB5XEv4= +google.golang.org/appengine v1.6.1/go.mod h1:i06prIuMbXzDqacNJfV5OdTW448YApPu5ww/cMBSeb0= google.golang.org/appengine v1.6.5 h1:tycE03LOZYQNhDpS27tcQdAzLCVMaj7QT2SXxebnpCM= google.golang.org/appengine v1.6.5/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20180817151627-c66870c02cf8/go.mod h1:JiN7NxoALGmiZfu7CAH4rXhgtRTLTxftemlI0sWmxmc= google.golang.org/genproto v0.0.0-20190307195333-5fe7a883aa19/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE= google.golang.org/genproto v0.0.0-20190418145605-e7d98fc518a7/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE= -google.golang.org/genproto v0.0.0-20190502173448-54afdca5d873 h1:nfPFGzJkUDX6uBmpN/pSw7MbOAWegH5QDQuoXFHedLg= +google.golang.org/genproto v0.0.0-20190425155659-357c62f0e4bb/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE= google.golang.org/genproto v0.0.0-20190502173448-54afdca5d873/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE= +google.golang.org/genproto v0.0.0-20190801165951-fa694d86fc64/go.mod h1:DMBHOl98Agz4BDEuKkezgsaosCRResVns1a3J2ZsMNc= google.golang.org/genproto v0.0.0-20190819201941-24fa4b261c55/go.mod h1:DMBHOl98Agz4BDEuKkezgsaosCRResVns1a3J2ZsMNc= -google.golang.org/genproto v0.0.0-20200409111301-baae70f3302d h1:I7Vuu5Ejagca+VcgfBINHke3xwjCTYnIG4Q57fv0wYY= -google.golang.org/genproto v0.0.0-20200409111301-baae70f3302d/go.mod h1:55QSHmfGQM9UVYDPBsyGGes0y52j32PQ3BqQfXhyH3c= +google.golang.org/genproto v0.0.0-20190911173649-1774047e7e51/go.mod h1:IbNlFCBrqXvoKpeg0TB2l7cyZUmoaFKYIwrEpbDKLA8= +google.golang.org/genproto v0.0.0-20191230161307-f3c370f40bfb/go.mod h1:n3cpQtvxv34hfy77yVDNjmbRyujviMdxYliBSkLhpCc= +google.golang.org/genproto v0.0.0-20200526211855-cb27e3aa2013/go.mod h1:NbSheEEYHJ7i3ixzK3sjbqSGDJWnxyFXZblF3eUsNvo= +google.golang.org/genproto v0.0.0-20201019141844-1ed22bb0c154 h1:bFFRpT+e8JJVY7lMMfvezL1ZIwqiwmPl2bsE2yx4HqM= +google.golang.org/genproto v0.0.0-20201019141844-1ed22bb0c154/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= google.golang.org/grpc v1.19.0/go.mod h1:mqu4LbDTu4XGKhr4mRzUsmM4RtVoemTSY81AxZiDr8c= +google.golang.org/grpc v1.20.1/go.mod h1:10oTOabMzJvdu6/UiuZezV6QK5dSlG84ov/aaiqXj38= google.golang.org/grpc v1.21.0/go.mod h1:oYelfM1adQP15Ek0mdvEgi9Df8B9CZIaU1084ijfRaM= +google.golang.org/grpc v1.21.1/go.mod h1:oYelfM1adQP15Ek0mdvEgi9Df8B9CZIaU1084ijfRaM= google.golang.org/grpc v1.23.0/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyacEbxg= -google.golang.org/grpc v1.23.1 h1:q4XQuHFC6I28BKZpo6IYyb3mNO+l7lSOxRuYTCiDfXk= google.golang.org/grpc v1.23.1/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyacEbxg= google.golang.org/grpc v1.25.1/go.mod h1:c3i+UQWmh7LiEpx4sFZnkU36qjEYZ0imhYfXVyQciAY= +google.golang.org/grpc v1.26.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk= google.golang.org/grpc v1.27.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk= google.golang.org/grpc v1.28.1 h1:C1QC6KzgSiLyBabDi87BbjaGreoRgGUF5nOyvfrAZ1k= google.golang.org/grpc v1.28.1/go.mod h1:rpkK4SK4GF4Ach/+MFLZUBavHOvF2JJB5uozKKal+60= +google.golang.org/protobuf v0.0.0-20200109180630-ec00e32a8dfd/go.mod h1:DFci5gLYBciE7Vtevhsrf46CRTquxDuWsQurQQe4oz8= +google.golang.org/protobuf v0.0.0-20200221191635-4d8936d0db64/go.mod h1:kwYJMbMJ01Woi6D6+Kah6886xMZcty6N08ah7+eCXa0= +google.golang.org/protobuf v0.0.0-20200228230310-ab0ca4ff8a60/go.mod h1:cfTl7dwQJ+fmap5saPgwCLgHXTUD7jkjRqWcaiX5VyM= +google.golang.org/protobuf v1.20.1-0.20200309200217-e05f789c0967/go.mod h1:A+miEFZTKqfCUM6K7xSMQL9OKL/b6hQv+e19PK+JZNE= +google.golang.org/protobuf v1.21.0/go.mod h1:47Nbq4nVaFHyn7ilMalzfO3qCViNmqZ2kzikPIcrTAo= +google.golang.org/protobuf v1.22.0/go.mod h1:EGpADcykh3NcUnDUJcl1+ZksZNG86OlYog2l/sGQquU= +google.golang.org/protobuf v1.23.0/go.mod h1:EGpADcykh3NcUnDUJcl1+ZksZNG86OlYog2l/sGQquU= +google.golang.org/protobuf v1.23.1-0.20200526195155-81db48ad09cc/go.mod h1:EGpADcykh3NcUnDUJcl1+ZksZNG86OlYog2l/sGQquU= +google.golang.org/protobuf v1.24.0 h1:UhZDfRO8JRQru4/+LlLE0BRKGF8L+PICnvYZmx/fEGA= +google.golang.org/protobuf v1.24.0/go.mod h1:r/3tXBNzIEhYS9I1OUVjXDlt8tc493IdKGjtUeSXeh4= gopkg.in/alecthomas/kingpin.v2 v2.2.6/go.mod h1:FMv+mEhP44yOT+4EoQTLFTRgOQ1FBLkstjWtayDeSgw= -gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= -gopkg.in/check.v1 v1.0.0-20180628173108-788fd7840127 h1:qIbj1fsPNlZgppZ+VLlY7N33q108Sa+fhmuc+sWQYwY= gopkg.in/check.v1 v1.0.0-20180628173108-788fd7840127/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= gopkg.in/check.v1 v1.0.0-20190902080502-41f04d3bba15 h1:YR8cESwS4TdDjEe65xsg0ogRM/Nc3DYOhEAlW+xobZo= gopkg.in/check.v1 v1.0.0-20190902080502-41f04d3bba15/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= gopkg.in/cheggaaa/pb.v1 v1.0.25/go.mod h1:V/YB90LKu/1FcN3WVnfiiE5oMCibMjukxqG/qStrOgw= +gopkg.in/errgo.v2 v2.1.0/go.mod h1:hNsd1EY+bozCKY1Ytp96fpM3vjJbqLJn88ws8XvfDNI= gopkg.in/fsnotify.v1 v1.4.7 h1:xOHLXZwVvI9hhs+cLKq5+I5onOuwQLhQwiu63xxlHs4= gopkg.in/fsnotify.v1 v1.4.7/go.mod h1:Tz8NjZHkW78fSQdbUxIjBTcgA1z1m8ZHf0WmKUhAMys= gopkg.in/inf.v0 v0.9.1 h1:73M5CoZyi3ZLMOyDlQh031Cx6N9NDJ2Vvfl76EDAgDc= gopkg.in/inf.v0 v0.9.1/go.mod h1:cWUDdTG/fYaXco+Dcufb5Vnc6Gp2YChqWtbxRZE0mXw= -gopkg.in/natefinch/lumberjack.v2 v2.0.0 h1:1Lc07Kr7qY4U2YPouBjpCLxpiyxIVoxqXgkXLknAOE8= gopkg.in/natefinch/lumberjack.v2 v2.0.0/go.mod h1:l0ndWWf7gzL7RNwBG7wST/UCcT4T24xpD6X8LsfU/+k= gopkg.in/resty.v1 v1.12.0/go.mod h1:mDo4pnntr5jdWRML875a/NmxYqAlA73dVijT2AXvQQo= gopkg.in/square/go-jose.v2 v2.2.2/go.mod h1:M9dMgbHiYLoDGQrXy7OpJDJWiKiU//h+vD76mk0e1AI= gopkg.in/tomb.v1 v1.0.0-20141024135613-dd632973f1e7 h1:uRGJdciOHaEIrze2W8Q3AKkepLTh2hOroT7a+7czfdQ= gopkg.in/tomb.v1 v1.0.0-20141024135613-dd632973f1e7/go.mod h1:dt/ZhP58zS4L8KSrWDmTeBkI65Dw0HsyUHuEVlX15mw= gopkg.in/yaml.v2 v2.0.0-20170812160011-eb3733d160e7/go.mod h1:JAlM8MvJe8wmxCU4Bli9HhUf9+ttbYbLASfIpnQbh74= -gopkg.in/yaml.v2 v2.2.1 h1:mUhvW9EsL+naU5Q3cakzfE91YhliOondGd6ZrsDBHQE= gopkg.in/yaml.v2 v2.2.1/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= -gopkg.in/yaml.v2 v2.2.2 h1:ZCJp+EgiOT7lHqUV2J862kp8Qj64Jo6az82+3Td9dZw= gopkg.in/yaml.v2 v2.2.2/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= gopkg.in/yaml.v2 v2.2.4/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= +gopkg.in/yaml.v2 v2.2.5/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= gopkg.in/yaml.v2 v2.2.8 h1:obN1ZagJSUGI0Ek/LBmuj4SNLPfIny3KsKFopxRdj10= gopkg.in/yaml.v2 v2.2.8/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= +gopkg.in/yaml.v3 v3.0.0-20200615113413-eeeca48fe776 h1:tQIYjPdBoyREyB9XMu+nnTclpTYkz2zFM+lzLJFO4gQ= +gopkg.in/yaml.v3 v3.0.0-20200615113413-eeeca48fe776/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM= gotest.tools v2.2.0+incompatible h1:VsBPFP1AI068pPrMxtb/S8Zkgf9xEmTLJjfM+P5UIEo= gotest.tools v2.2.0+incompatible/go.mod h1:DsYFclhRJ6vuDpmuTbkuFWG+y2sxOXAzmJt81HFBacw= +gotest.tools/v3 v3.0.2/go.mod h1:3SzNCllyD9/Y+b5r9JIKQ474KzkZyqLqEfYqMsX94Bk= honnef.co/go/tools v0.0.0-20190102054323-c2f93a96b099/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4= honnef.co/go/tools v0.0.0-20190106161140-3f1c8253044a/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4= +honnef.co/go/tools v0.0.0-20190418001031-e561f6794a2a/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4= honnef.co/go/tools v0.0.0-20190523083050-ea95bdfd59fc/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4= -k8s.io/api v0.17.4 h1:HbwOhDapkguO8lTAE8OX3hdF2qp8GtpC9CW/MQATXXo= -k8s.io/api v0.17.4/go.mod h1:5qxx6vjmwUVG2nHQTKGlLts8Tbok8PzHl4vHtVFuZCA= -k8s.io/apiextensions-apiserver v0.17.4 h1:ZKFnw3cJrGZ/9s6y+DerTF4FL+dmK0a04A++7JkmMho= -k8s.io/apiextensions-apiserver v0.17.4/go.mod h1:rCbbbaFS/s3Qau3/1HbPlHblrWpFivoaLYccCffvQGI= -k8s.io/apimachinery v0.17.4 h1:UzM+38cPUJnzqSQ+E1PY4YxMHIzQyCg29LOoGfo79Zw= -k8s.io/apimachinery v0.17.4/go.mod h1:gxLnyZcGNdZTCLnq3fgzyg2A5BVCHTNDFrw8AmuJ+0g= -k8s.io/apiserver v0.17.4 h1:bYc9LvDPEF9xAL3fhbDzqNOQOAnNF2ZYCrMW8v52/mE= -k8s.io/apiserver v0.17.4/go.mod h1:5ZDQ6Xr5MNBxyi3iUZXS84QOhZl+W7Oq2us/29c0j9I= -k8s.io/client-go v0.17.4 h1:VVdVbpTY70jiNHS1eiFkUt7ZIJX3txd29nDxxXH4en8= -k8s.io/client-go v0.17.4/go.mod h1:ouF6o5pz3is8qU0/qYL2RnoxOPqgfuidYLowytyLJmc= -k8s.io/code-generator v0.17.4/go.mod h1:l8BLVwASXQZTo2xamW5mQNFCe1XPiAesVq7Y1t7PiQQ= -k8s.io/component-base v0.17.4 h1:H9cdWZyiGVJfWmWIcHd66IsNBWTk1iEgU7D4kJksEnw= -k8s.io/component-base v0.17.4/go.mod h1:5BRqHMbbQPm2kKu35v3G+CpVq4K0RJKC7TRioF0I9lE= -k8s.io/gengo v0.0.0-20190128074634-0689ccc1d7d6/go.mod h1:ezvh/TsK7cY6rbqRK0oQQ8IAqLxYwwyPxAX1Pzy0ii0= -k8s.io/gengo v0.0.0-20190822140433-26a664648505/go.mod h1:ezvh/TsK7cY6rbqRK0oQQ8IAqLxYwwyPxAX1Pzy0ii0= -k8s.io/klog v0.0.0-20181102134211-b9b56d5dfc92/go.mod h1:Gq+BEi5rUBO/HRz0bTSXDUcqjScdoY3a9IHpCEIOOfk= -k8s.io/klog v0.3.0/go.mod h1:Gq+BEi5rUBO/HRz0bTSXDUcqjScdoY3a9IHpCEIOOfk= +honnef.co/go/tools v0.0.1-2019.2.3/go.mod h1:a3bituU0lyd329TUQxRnasdCoJDkEUEAqEt0JzvZhAg= +k8s.io/api v0.19.3 h1:GN6ntFnv44Vptj/b+OnMW7FmzkpDoIDLZRvKX3XH9aU= +k8s.io/api v0.19.3/go.mod h1:VF+5FT1B74Pw3KxMdKyinLo+zynBaMBiAfGMuldcNDs= +k8s.io/apiextensions-apiserver v0.19.3 h1:WZxBypSHW4SdXHbdPTS/Jy7L2la6Niggs8BuU5o+avo= +k8s.io/apiextensions-apiserver v0.19.3/go.mod h1:igVEkrE9TzInc1tYE7qSqxaLg/rEAp6B5+k9Q7+IC8Q= +k8s.io/apimachinery v0.19.3 h1:bpIQXlKjB4cB/oNpnNnV+BybGPR7iP5oYpsOTEJ4hgc= +k8s.io/apimachinery v0.19.3/go.mod h1:DnPGDnARWFvYa3pMHgSxtbZb7gpzzAZ1pTfaUNDVlmA= +k8s.io/apiserver v0.19.3/go.mod h1:bx6dMm+H6ifgKFpCQT/SAhPwhzoeIMlHIaibomUDec0= +k8s.io/cli-runtime v0.19.3/go.mod h1:q+l845i5/uWzcUpCrl+L4f3XLaJi8ZeLVQ/decwty0A= +k8s.io/client-go v0.19.3 h1:ctqR1nQ52NUs6LpI0w+a5U+xjYwflFwA13OJKcicMxg= +k8s.io/client-go v0.19.3/go.mod h1:+eEMktZM+MG0KO+PTkci8xnbCZHvj9TqR6Q1XDUIJOM= +k8s.io/code-generator v0.19.3/go.mod h1:moqLn7w0t9cMs4+5CQyxnfA/HV8MF6aAVENF+WZZhgk= +k8s.io/component-base v0.19.3/go.mod h1:WhLWSIefQn8W8jxSLl5WNiR6z8oyMe/8Zywg7alOkRc= +k8s.io/gengo v0.0.0-20200413195148-3a45101e95ac/go.mod h1:ezvh/TsK7cY6rbqRK0oQQ8IAqLxYwwyPxAX1Pzy0ii0= +k8s.io/gengo v0.0.0-20200428234225-8167cfdcfc14/go.mod h1:ezvh/TsK7cY6rbqRK0oQQ8IAqLxYwwyPxAX1Pzy0ii0= k8s.io/klog v1.0.0 h1:Pt+yjF5aB1xDSVbau4VsWe+dQNzA0qv1LlXdC2dF6Q8= k8s.io/klog v1.0.0/go.mod h1:4Bi6QPql/J/LkTDqv7R/cd3hPo4k2DG6Ptcz060Ez5I= -k8s.io/kube-openapi v0.0.0-20191107075043-30be4d16710a h1:UcxjrRMyNx/i/y8G7kPvLyy7rfbeuf1PYyBf973pgyU= -k8s.io/kube-openapi v0.0.0-20191107075043-30be4d16710a/go.mod h1:1TqjTSzOxsLGIKfj0lK8EeCP7K1iUG65v09OM0/WG5E= -k8s.io/kubernetes v1.13.4 h1:gQqFv/pH8hlbznLXQUsi8s5zqYnv0slmUDl/yVA0EWc= -k8s.io/kubernetes v1.13.4/go.mod h1:ocZa8+6APFNC2tX1DZASIbocyYT5jHzqFVsY5aoB7Jk= -k8s.io/utils v0.0.0-20191114184206-e782cd3c129f h1:GiPwtSzdP43eI1hpPCbROQCCIgCuiMMNF8YUVLF3vJo= -k8s.io/utils v0.0.0-20191114184206-e782cd3c129f/go.mod h1:sZAwmy6armz5eXlNoLmJcl4F1QuKu7sr+mFQ0byX7Ew= -k8s.io/utils v0.0.0-20200327001022-6496210b90e8 h1:6JFbaLjRyBz8K2Jvt+pcT+N3vvwMZfg8MfVENwe9aag= -k8s.io/utils v0.0.0-20200327001022-6496210b90e8/go.mod h1:sZAwmy6armz5eXlNoLmJcl4F1QuKu7sr+mFQ0byX7Ew= -modernc.org/cc v1.0.0/go.mod h1:1Sk4//wdnYJiUIxnW8ddKpaOJCF37yAdqYnkxUpaYxw= -modernc.org/golex v1.0.0/go.mod h1:b/QX9oBD/LhixY6NDh+IdGv17hgB+51fET1i2kPSmvk= -modernc.org/mathutil v1.0.0/go.mod h1:wU0vUrJsVWBZ4P6e7xtFJEhFSNsfRLJ8H458uRjg03k= -modernc.org/strutil v1.0.0/go.mod h1:lstksw84oURvj9y3tn8lGvRxyRC1S2+g5uuIzNfIOBs= -modernc.org/xc v1.0.0/go.mod h1:mRNCo0bvLjGhHO9WsyuKVU4q0ceiDDDoEeWDJHrNx8I= -sigs.k8s.io/structured-merge-diff v0.0.0-20190525122527-15d366b2352e/go.mod h1:wWxsB5ozmmv/SG7nM11ayaAW51xMvak/t1r0CSlcokI= -sigs.k8s.io/structured-merge-diff v1.0.1-0.20191108220359-b1b620dd3f06 h1:zD2IemQ4LmOcAumeiyDWXKUI2SO0NYDe3H6QGvPOVgU= -sigs.k8s.io/structured-merge-diff v1.0.1-0.20191108220359-b1b620dd3f06/go.mod h1:/ULNhyfzRopfcjskuui0cTITekDduZ7ycKN3oUT9R18= -sigs.k8s.io/yaml v1.1.0 h1:4A07+ZFc2wgJwo8YNlQpr1rVlgUDlxXHhPJciaPY5gs= +k8s.io/klog/v2 v2.0.0/go.mod h1:PBfzABfn139FHAV07az/IF9Wp1bkk3vpT2XSJ76fSDE= +k8s.io/klog/v2 v2.2.0 h1:XRvcwJozkgZ1UQJmfMGpvRthQHOvihEhYtDfAaxMz/A= +k8s.io/klog/v2 v2.2.0/go.mod h1:Od+F08eJP+W3HUb4pSrPpgp9DGU4GzlpG/TmITuYh/Y= +k8s.io/kube-openapi v0.0.0-20200805222855-6aeccd4b50c6/go.mod h1:UuqjUnNftUyPE5H64/qeyjQoUZhGpeFDVdxjTeEVN2o= +k8s.io/kube-openapi v0.0.0-20200923155610-8b5066479488 h1:mNpvQf4lkIHNOXCoM+Veu/UXwA56Yx1J7hY1Tvcs/oM= +k8s.io/kube-openapi v0.0.0-20200923155610-8b5066479488/go.mod h1:UuqjUnNftUyPE5H64/qeyjQoUZhGpeFDVdxjTeEVN2o= +k8s.io/kubectl v0.19.3 h1:T8IHHpg+uRIfn34wqJ8wHG5bbH+VV5FNPtJ+jKcho1U= +k8s.io/kubectl v0.19.3/go.mod h1:t5cscfrAuHUvEGNyNJjPKt+rGlaJzk8jrKYHXxEsANE= +k8s.io/metrics v0.19.3/go.mod h1:Eap/Lk1FiAIjkaArFuv41v+ph6dbDpVGwAg7jMI+4vg= +k8s.io/utils v0.0.0-20200729134348-d5654de09c73 h1:uJmqzgNWG7XyClnU/mLPBWwfKKF1K8Hf8whTseBgJcg= +k8s.io/utils v0.0.0-20200729134348-d5654de09c73/go.mod h1:jPW/WVKK9YHAvNhRxK0md/EJ228hCsBRufyofKtW8HA= +rsc.io/binaryregexp v0.2.0/go.mod h1:qTv7/COck+e2FymRvadv62gMdZztPaShugOCi3I+8D8= +sigs.k8s.io/apiserver-network-proxy/konnectivity-client v0.0.9/go.mod h1:dzAXnQbTRyDlZPJX2SUPEqvnB+j7AJjtlox7PEwigU0= +sigs.k8s.io/kustomize v2.0.3+incompatible/go.mod h1:MkjgH3RdOWrievjo6c9T245dYlB5QeXV4WCbnt/PEpU= +sigs.k8s.io/structured-merge-diff/v4 v4.0.1 h1:YXTMot5Qz/X1iBRJhAt+vI+HVttY0WkSqqhKxQ0xVbA= +sigs.k8s.io/structured-merge-diff/v4 v4.0.1/go.mod h1:bJZC9H9iH24zzfZ/41RGcq60oK1F7G282QMXDPYydCw= sigs.k8s.io/yaml v1.1.0/go.mod h1:UJmg0vDUVViEyp3mgSv9WPwZCDxu4rQW1olrI1uml+o= sigs.k8s.io/yaml v1.2.0 h1:kr/MCeFWJWTwyaHoR9c8EjH9OumOmoF9YGiZd7lFm/Q= sigs.k8s.io/yaml v1.2.0/go.mod h1:yfXDCHCao9+ENCvLSE62v9VSji2MKu5jeNfTrofGhJc= +vbom.ml/util v0.0.0-20160121211510-db5cfe13f5cc/go.mod h1:so/NYdZXCz+E3ZpW0uAoCj6uzU2+8OWDFv/HxUSs7kI= diff --git a/pkg/kubecfg/update.go b/pkg/kubecfg/update.go index c34134dd..275d48df 100644 --- a/pkg/kubecfg/update.go +++ b/pkg/kubecfg/update.go @@ -25,7 +25,7 @@ import ( "k8s.io/client-go/dynamic" "k8s.io/client-go/util/retry" "k8s.io/kube-openapi/pkg/util/proto" - "k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi" + "k8s.io/kubectl/pkg/util/openapi" "github.com/bitnami/kubecfg/utils" ) diff --git a/pkg/kubecfg/update_test.go b/pkg/kubecfg/update_test.go index 21b6da8d..d3198988 100644 --- a/pkg/kubecfg/update_test.go +++ b/pkg/kubecfg/update_test.go @@ -7,7 +7,7 @@ import ( "testing" pb_proto "github.com/golang/protobuf/proto" - "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" apiequality "k8s.io/apimachinery/pkg/api/equality" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" @@ -15,7 +15,7 @@ import ( "k8s.io/apimachinery/pkg/util/diff" "k8s.io/apimachinery/pkg/util/strategicpatch" "k8s.io/kube-openapi/pkg/util/proto" - "k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi" + "k8s.io/kubectl/pkg/util/openapi" "github.com/bitnami/kubecfg/utils" ) diff --git a/utils/client.go b/utils/client.go index af0747f8..07cf254d 100644 --- a/utils/client.go +++ b/utils/client.go @@ -27,7 +27,7 @@ import ( "sync" "syscall" - openapi_v2 "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" log "github.com/sirupsen/logrus" errorsutil "k8s.io/apimachinery/pkg/api/errors" diff --git a/utils/openapi.go b/utils/openapi.go index efd35dd9..4708bf19 100644 --- a/utils/openapi.go +++ b/utils/openapi.go @@ -25,7 +25,7 @@ import ( "k8s.io/client-go/discovery" "k8s.io/kube-openapi/pkg/util/proto" "k8s.io/kube-openapi/pkg/util/proto/validation" - "k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi" + "k8s.io/kubectl/pkg/util/openapi" ) // OpenAPISchema represents an OpenAPI schema for a given GroupVersionKind. diff --git a/utils/openapi_test.go b/utils/openapi_test.go index c424e263..3f0a5d05 100644 --- a/utils/openapi_test.go +++ b/utils/openapi_test.go @@ -22,7 +22,7 @@ import ( "testing" "github.com/golang/protobuf/proto" - "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apimachinery/pkg/runtime/schema" utilerrors "k8s.io/apimachinery/pkg/util/errors" diff --git a/utils/sort_test.go b/utils/sort_test.go index ff882068..784ba13c 100644 --- a/utils/sort_test.go +++ b/utils/sort_test.go @@ -21,7 +21,7 @@ import ( "sort" "testing" - "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" log "github.com/sirupsen/logrus" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" From 92ce2b247feba732a8bbaf54518c89f862d5fc71 Mon Sep 17 00:00:00 2001 From: Serge Bazanski Date: Tue, 3 Nov 2020 17:29:23 +0100 Subject: [PATCH 3/4] vendor: update --- .../go/compute/metadata/.repo-metadata.json | 12 + .../go/compute/metadata/metadata.go | 24 +- .../github.com/evanphx/json-patch/.travis.yml | 7 +- vendor/github.com/evanphx/json-patch/LICENSE | 2 +- .../github.com/evanphx/json-patch/README.md | 11 +- vendor/github.com/evanphx/json-patch/merge.go | 13 +- vendor/github.com/evanphx/json-patch/patch.go | 28 +- .../go-logr/logr}/LICENSE | 5 +- vendor/github.com/go-logr/logr/README.md | 181 + vendor/github.com/go-logr/logr/go.mod | 3 + vendor/github.com/go-logr/logr/logr.go | 178 + .../golang/protobuf/proto/buffer.go | 324 + .../github.com/golang/protobuf/proto/clone.go | 253 - .../golang/protobuf/proto/decode.go | 427 - .../golang/protobuf/proto/defaults.go | 63 + .../golang/protobuf/proto/deprecated.go | 126 +- .../golang/protobuf/proto/discard.go | 356 +- .../golang/protobuf/proto/encode.go | 203 - .../github.com/golang/protobuf/proto/equal.go | 301 - .../golang/protobuf/proto/extensions.go | 771 +- .../github.com/golang/protobuf/proto/lib.go | 965 - .../golang/protobuf/proto/message_set.go | 181 - .../golang/protobuf/proto/pointer_reflect.go | 360 - .../golang/protobuf/proto/pointer_unsafe.go | 313 - .../golang/protobuf/proto/properties.go | 648 +- .../github.com/golang/protobuf/proto/proto.go | 167 + .../golang/protobuf/proto/registry.go | 323 + .../golang/protobuf/proto/table_marshal.go | 2776 - .../golang/protobuf/proto/table_merge.go | 654 - .../golang/protobuf/proto/table_unmarshal.go | 2053 - .../github.com/golang/protobuf/proto/text.go | 845 - .../golang/protobuf/proto/text_decode.go | 801 + .../golang/protobuf/proto/text_encode.go | 560 + .../golang/protobuf/proto/text_parser.go | 880 - .../github.com/golang/protobuf/proto/wire.go | 78 + .../golang/protobuf/proto/wrappers.go | 34 + .../github.com/golang/protobuf/ptypes/any.go | 214 +- .../golang/protobuf/ptypes/any/any.pb.go | 233 +- .../golang/protobuf/ptypes/any/any.proto | 155 - .../github.com/golang/protobuf/ptypes/doc.go | 37 +- .../golang/protobuf/ptypes/duration.go | 116 +- .../protobuf/ptypes/duration/duration.pb.go | 194 +- .../protobuf/ptypes/duration/duration.proto | 116 - .../golang/protobuf/ptypes/timestamp.go | 109 +- .../protobuf/ptypes/timestamp/timestamp.pb.go | 217 +- .../protobuf/ptypes/timestamp/timestamp.proto | 138 - .../github.com/google/go-jsonnet/.gitignore | 3 - .../google/go-jsonnet/.golangci.yml | 15 - .../google/go-jsonnet/.goreleaser.yml | 87 - .../github.com/google/go-jsonnet/.travis.yml | 18 +- .../github.com/google/go-jsonnet/BUILD.bazel | 5 +- .../github.com/google/go-jsonnet/MANIFEST.in | 9 - vendor/github.com/google/go-jsonnet/Makefile | 4 +- vendor/github.com/google/go-jsonnet/README.md | 55 +- .../github.com/google/go-jsonnet/ast/ast.go | 76 +- .../github.com/google/go-jsonnet/ast/clone.go | 18 +- .../google/go-jsonnet/ast/fodder.go | 2 +- .../google/go-jsonnet/ast/location.go | 23 +- .../github.com/google/go-jsonnet/ast/util.go | 4 +- .../google/go-jsonnet/astgen/BUILD.bazel | 2 +- .../google/go-jsonnet/astgen/stdast.go | 63333 ++++++---------- .../github.com/google/go-jsonnet/builtins.go | 1014 +- .../google/go-jsonnet/error_formatter.go | 8 +- vendor/github.com/google/go-jsonnet/go.mod | 12 +- vendor/github.com/google/go-jsonnet/go.sum | 41 +- .../github.com/google/go-jsonnet/imports.go | 26 +- .../internal/errors/static_error.go | 47 +- .../go-jsonnet/internal/parser/BUILD.bazel | 1 - .../go-jsonnet/internal/parser/context.go | 26 +- .../go-jsonnet/internal/parser/lexer.go | 263 +- .../go-jsonnet/internal/parser/parser.go | 262 +- .../go-jsonnet/internal/parser/string_util.go | 134 - .../go-jsonnet/internal/program/desugarer.go | 139 +- .../go-jsonnet/internal/program/program.go | 4 +- .../internal/program/static_analyzer.go | 16 +- .../google/go-jsonnet/interpreter.go | 318 +- vendor/github.com/google/go-jsonnet/setup.py | 7 +- vendor/github.com/google/go-jsonnet/tests.sh | 23 +- vendor/github.com/google/go-jsonnet/thunks.go | 113 +- .../google/go-jsonnet/travisBazel.sh | 2 +- .../google/go-jsonnet/travisBuild.sh | 24 +- .../google/go-jsonnet/update_cpp_jsonnet.sh | 16 - vendor/github.com/google/go-jsonnet/util.go | 38 - vendor/github.com/google/go-jsonnet/value.go | 132 +- vendor/github.com/google/go-jsonnet/vm.go | 247 +- .../gnostic/OpenAPIv2/OpenAPIv2.pb.go | 5225 -- .../googleapis/gnostic/OpenAPIv2/README.md | 16 - .../googleapis/gnostic/compiler/README.md | 3 +- .../googleapis/gnostic/compiler/context.go | 20 +- .../googleapis/gnostic/compiler/error.go | 15 +- .../gnostic/compiler/extension-handler.go | 101 - .../googleapis/gnostic/compiler/extensions.go | 85 + .../googleapis/gnostic/compiler/helpers.go | 347 +- .../googleapis/gnostic/compiler/main.go | 2 +- .../googleapis/gnostic/compiler/reader.go | 133 +- .../googleapis/gnostic/extensions/README.md | 12 +- .../gnostic/extensions/extension.pb.go | 529 +- .../gnostic/extensions/extension.proto | 29 +- .../gnostic/extensions/extensions.go | 68 +- .../googleapis/gnostic/jsonschema/README.md | 4 + .../googleapis/gnostic/jsonschema/base.go | 84 + .../googleapis/gnostic/jsonschema/display.go | 229 + .../googleapis/gnostic/jsonschema/models.go | 228 + .../gnostic/jsonschema/operations.go | 394 + .../googleapis/gnostic/jsonschema/reader.go | 442 + .../googleapis/gnostic/jsonschema/schema.json | 150 + .../googleapis/gnostic/jsonschema/writer.go | 369 + .../{OpenAPIv2 => openapiv2}/OpenAPIv2.go | 3610 +- .../gnostic/openapiv2/OpenAPIv2.pb.go | 7347 ++ .../{OpenAPIv2 => openapiv2}/OpenAPIv2.proto | 7 +- .../googleapis/gnostic/openapiv2/README.md | 14 + .../googleapis/gnostic/openapiv2/document.go | 41 + .../{OpenAPIv2 => openapiv2}/openapi-2.0.json | 4 +- .../gophercloud/gophercloud/.gitignore | 3 - .../gophercloud/gophercloud/.travis.yml | 25 - .../gophercloud/gophercloud/.zuul.yaml | 114 - .../gophercloud/gophercloud/CHANGELOG.md | 0 .../gophercloud/gophercloud/README.md | 159 - .../gophercloud/gophercloud/auth_options.go | 437 - .../gophercloud/gophercloud/auth_result.go | 52 - .../github.com/gophercloud/gophercloud/doc.go | 110 - .../gophercloud/endpoint_search.go | 76 - .../gophercloud/gophercloud/errors.go | 471 - .../github.com/gophercloud/gophercloud/go.mod | 7 - .../github.com/gophercloud/gophercloud/go.sum | 8 - .../gophercloud/openstack/auth_env.go | 125 - .../gophercloud/openstack/client.go | 438 - .../gophercloud/gophercloud/openstack/doc.go | 14 - .../openstack/endpoint_location.go | 107 - .../gophercloud/openstack/errors.go | 71 - .../openstack/identity/v2/tenants/doc.go | 65 - .../openstack/identity/v2/tenants/requests.go | 116 - .../openstack/identity/v2/tenants/results.go | 91 - .../openstack/identity/v2/tenants/urls.go | 23 - .../openstack/identity/v2/tokens/doc.go | 46 - .../openstack/identity/v2/tokens/requests.go | 103 - .../openstack/identity/v2/tokens/results.go | 174 - .../openstack/identity/v2/tokens/urls.go | 13 - .../openstack/identity/v3/tokens/doc.go | 108 - .../openstack/identity/v3/tokens/requests.go | 162 - .../openstack/identity/v3/tokens/results.go | 178 - .../openstack/identity/v3/tokens/urls.go | 7 - .../openstack/utils/base_endpoint.go | 28 - .../openstack/utils/choose_version.go | 111 - .../gophercloud/pagination/http.go | 60 - .../gophercloud/pagination/linked.go | 92 - .../gophercloud/pagination/marker.go | 58 - .../gophercloud/pagination/pager.go | 251 - .../gophercloud/gophercloud/pagination/pkg.go | 4 - .../gophercloud/pagination/single.go | 33 - .../gophercloud/gophercloud/params.go | 491 - .../gophercloud/provider_client.go | 501 - .../gophercloud/gophercloud/results.go | 448 - .../gophercloud/gophercloud/service_client.go | 154 - .../gophercloud/gophercloud/util.go | 102 - vendor/github.com/json-iterator/go/README.md | 36 +- vendor/github.com/json-iterator/go/any_str.go | 4 +- vendor/github.com/json-iterator/go/config.go | 4 +- .../json-iterator/go/iter_object.go | 4 +- .../json-iterator/go/reflect_extension.go | 2 +- .../json-iterator/go/reflect_map.go | 80 +- .../json-iterator/go/reflect_optional.go | 4 - .../go/reflect_struct_decoder.go | 22 +- vendor/github.com/json-iterator/go/stream.go | 5 +- .../go-windows-terminal-sequences/README.md | 1 + .../sequences.go | 3 +- .../sergi/go-diff/diffmatchpatch/diff.go | 141 +- .../diffmatchpatch/operation_string.go | 17 - .../github.com/sirupsen/logrus/.golangci.yml | 40 + vendor/github.com/sirupsen/logrus/.travis.yml | 14 +- .../github.com/sirupsen/logrus/CHANGELOG.md | 27 +- vendor/github.com/sirupsen/logrus/README.md | 44 +- .../github.com/sirupsen/logrus/appveyor.yml | 28 +- vendor/github.com/sirupsen/logrus/entry.go | 47 +- vendor/github.com/sirupsen/logrus/exported.go | 2 +- vendor/github.com/sirupsen/logrus/go.mod | 5 +- vendor/github.com/sirupsen/logrus/go.sum | 8 +- .../sirupsen/logrus/json_formatter.go | 4 + vendor/github.com/sirupsen/logrus/logger.go | 11 +- vendor/github.com/sirupsen/logrus/logrus.go | 2 +- .../sirupsen/logrus/terminal_check_bsd.go | 2 +- .../sirupsen/logrus/terminal_check_js.go | 7 + .../sirupsen/logrus/terminal_check_unix.go | 2 +- .../sirupsen/logrus/text_formatter.go | 55 +- vendor/github.com/sirupsen/logrus/writer.go | 6 + vendor/github.com/spf13/cobra/README.md | 2 +- vendor/github.com/spf13/cobra/args.go | 10 +- .../spf13/cobra/bash_completions.go | 89 +- .../spf13/cobra/bash_completions.md | 239 +- vendor/github.com/spf13/cobra/command.go | 7 + .../spf13/cobra/custom_completions.go | 384 + .../spf13/cobra/fish_completions.go | 172 + .../spf13/cobra/fish_completions.md | 7 + .../testify/assert/assertion_format.go | 78 +- .../testify/assert/assertion_forward.go | 156 +- .../stretchr/testify/assert/assertions.go | 218 +- .../testify/assert/forward_assertions.go | 2 +- .../testify/require/forward_requirements.go | 2 +- .../stretchr/testify/require/require.go | 210 +- .../testify/require/require_forward.go | 156 +- .../stretchr/testify/require/requirements.go | 2 +- .../x/crypto/ssh/terminal/terminal.go | 8 + .../x/net/http2/client_conn_pool.go | 8 +- vendor/golang.org/x/net/http2/flow.go | 2 + .../golang.org/x/net/http2/hpack/huffman.go | 7 + vendor/golang.org/x/net/http2/http2.go | 7 + vendor/golang.org/x/net/http2/server.go | 8 +- vendor/golang.org/x/net/http2/transport.go | 78 +- .../sys/internal/unsafeheader/unsafeheader.go | 30 + vendor/golang.org/x/sys/unix/mkerrors.sh | 3 +- .../x/sys/unix/syscall_darwin.1_13.go | 21 +- .../golang.org/x/sys/unix/syscall_darwin.go | 1 + .../x/sys/unix/syscall_darwin_386.go | 11 - .../x/sys/unix/syscall_darwin_amd64.go | 11 - .../x/sys/unix/syscall_darwin_arm.go | 11 - .../x/sys/unix/syscall_darwin_arm64.go | 11 - vendor/golang.org/x/sys/unix/syscall_linux.go | 47 +- .../x/sys/unix/syscall_linux_386.go | 2 +- .../x/sys/unix/syscall_linux_amd64.go | 2 +- .../x/sys/unix/syscall_linux_arm.go | 2 +- .../x/sys/unix/syscall_linux_arm64.go | 28 +- .../x/sys/unix/syscall_linux_mips64x.go | 2 +- .../x/sys/unix/syscall_linux_mipsx.go | 2 +- .../x/sys/unix/syscall_linux_ppc64x.go | 2 +- .../x/sys/unix/syscall_linux_riscv64.go | 8 +- .../x/sys/unix/syscall_linux_s390x.go | 2 +- .../x/sys/unix/syscall_linux_sparc64.go | 2 +- vendor/golang.org/x/sys/unix/syscall_unix.go | 17 +- vendor/golang.org/x/sys/unix/zerrors_linux.go | 68 +- .../x/sys/unix/zerrors_linux_386.go | 3 + .../x/sys/unix/zerrors_linux_amd64.go | 3 + .../x/sys/unix/zerrors_linux_arm.go | 3 + .../x/sys/unix/zerrors_linux_arm64.go | 3 + .../x/sys/unix/zerrors_linux_mips.go | 3 + .../x/sys/unix/zerrors_linux_mips64.go | 3 + .../x/sys/unix/zerrors_linux_mips64le.go | 3 + .../x/sys/unix/zerrors_linux_mipsle.go | 3 + .../x/sys/unix/zerrors_linux_ppc64.go | 3 + .../x/sys/unix/zerrors_linux_ppc64le.go | 3 + .../x/sys/unix/zerrors_linux_riscv64.go | 3 + .../x/sys/unix/zerrors_linux_s390x.go | 3 + .../x/sys/unix/zerrors_linux_sparc64.go | 3 + .../x/sys/unix/zsyscall_darwin_386.1_11.go | 22 +- .../x/sys/unix/zsyscall_darwin_386.go | 32 +- .../x/sys/unix/zsyscall_darwin_amd64.1_11.go | 22 +- .../x/sys/unix/zsyscall_darwin_amd64.go | 32 +- .../x/sys/unix/zsyscall_darwin_arm.1_11.go | 22 +- .../x/sys/unix/zsyscall_darwin_arm.go | 32 +- .../x/sys/unix/zsyscall_darwin_arm64.1_11.go | 22 +- .../x/sys/unix/zsyscall_darwin_arm64.go | 32 +- .../golang.org/x/sys/unix/zsyscall_linux.go | 47 +- .../x/sys/unix/zsyscall_linux_386.go | 2 +- .../x/sys/unix/zsyscall_linux_amd64.go | 2 +- .../x/sys/unix/zsyscall_linux_arm.go | 2 +- .../x/sys/unix/zsyscall_linux_arm64.go | 4 +- .../x/sys/unix/zsyscall_linux_mips.go | 2 +- .../x/sys/unix/zsyscall_linux_mips64.go | 2 +- .../x/sys/unix/zsyscall_linux_mips64le.go | 2 +- .../x/sys/unix/zsyscall_linux_mipsle.go | 2 +- .../x/sys/unix/zsyscall_linux_ppc64.go | 2 +- .../x/sys/unix/zsyscall_linux_ppc64le.go | 2 +- .../x/sys/unix/zsyscall_linux_s390x.go | 2 +- .../x/sys/unix/zsyscall_linux_sparc64.go | 2 +- .../x/sys/unix/zsysctl_openbsd_386.go | 3 +- .../x/sys/unix/zsysctl_openbsd_amd64.go | 1 + .../x/sys/unix/zsysctl_openbsd_arm.go | 1 + .../x/sys/unix/ztypes_freebsd_arm.go | 12 +- vendor/golang.org/x/sys/unix/ztypes_linux.go | 437 +- .../golang.org/x/sys/windows/dll_windows.go | 29 + .../golang.org/x/sys/windows/env_windows.go | 11 +- .../x/sys/windows/memory_windows.go | 5 + .../x/sys/windows/security_windows.go | 20 +- .../x/sys/windows/syscall_windows.go | 39 +- .../x/sys/windows/zsyscall_windows.go | 19 + .../x/text/encoding/unicode/unicode.go | 94 +- .../golang.org/x/text/transform/transform.go | 6 +- vendor/golang.org/x/text/unicode/bidi/core.go | 8 +- .../x/text/unicode/bidi/tables11.0.0.go | 2 +- .../x/text/unicode/bidi/tables12.0.0.go | 1923 + .../x/text/unicode/norm/tables11.0.0.go | 2 +- .../x/text/unicode/norm/tables12.0.0.go | 7710 ++ .../golang.org/x/text/width/tables11.0.0.go | 2 +- .../golang.org/x/text/width/tables12.0.0.go | 1350 + .../googleapis/rpc/status/status.pb.go | 217 +- vendor/google.golang.org/protobuf/AUTHORS | 3 + .../google.golang.org/protobuf/CONTRIBUTORS | 3 + vendor/google.golang.org/protobuf/LICENSE | 27 + vendor/google.golang.org/protobuf/PATENTS | 22 + .../protobuf/encoding/prototext/decode.go | 796 + .../protobuf/encoding/prototext/doc.go | 7 + .../protobuf/encoding/prototext/encode.go | 433 + .../protobuf/encoding/protowire/wire.go | 538 + .../protobuf/internal/descfmt/stringer.go | 316 + .../protobuf/internal/descopts/options.go | 29 + .../protobuf/internal/detrand/rand.go | 61 + .../internal/encoding/defval/default.go | 213 + .../encoding/messageset/messageset.go | 258 + .../protobuf/internal/encoding/tag/tag.go | 207 + .../protobuf/internal/encoding/text/decode.go | 665 + .../internal/encoding/text/decode_number.go | 190 + .../internal/encoding/text/decode_string.go | 161 + .../internal/encoding/text/decode_token.go | 373 + .../protobuf/internal/encoding/text/doc.go | 29 + .../protobuf/internal/encoding/text/encode.go | 267 + .../protobuf/internal/errors/errors.go | 89 + .../protobuf/internal/errors/is_go112.go | 39 + .../protobuf/internal/errors/is_go113.go | 12 + .../protobuf/internal/fieldnum/any_gen.go | 13 + .../protobuf/internal/fieldnum/api_gen.go | 35 + .../internal/fieldnum/descriptor_gen.go | 240 + .../protobuf/internal/fieldnum/doc.go | 7 + .../internal/fieldnum/duration_gen.go | 13 + .../protobuf/internal/fieldnum/empty_gen.go | 10 + .../internal/fieldnum/field_mask_gen.go | 12 + .../internal/fieldnum/source_context_gen.go | 12 + .../protobuf/internal/fieldnum/struct_gen.go | 33 + .../internal/fieldnum/timestamp_gen.go | 13 + .../protobuf/internal/fieldnum/type_gen.go | 53 + .../internal/fieldnum/wrappers_gen.go | 52 + .../protobuf/internal/fieldsort/fieldsort.go | 40 + .../protobuf/internal/filedesc/build.go | 155 + .../protobuf/internal/filedesc/desc.go | 613 + .../protobuf/internal/filedesc/desc_init.go | 471 + .../protobuf/internal/filedesc/desc_lazy.go | 704 + .../protobuf/internal/filedesc/desc_list.go | 286 + .../internal/filedesc/desc_list_gen.go | 345 + .../protobuf/internal/filedesc/placeholder.go | 107 + .../protobuf/internal/filetype/build.go | 297 + .../protobuf/internal/flags/flags.go | 24 + .../internal/flags/proto_legacy_disable.go | 9 + .../internal/flags/proto_legacy_enable.go | 9 + .../protobuf/internal/genname/name.go | 25 + .../protobuf/internal/impl/api_export.go | 170 + .../protobuf/internal/impl/checkinit.go | 141 + .../protobuf/internal/impl/codec_extension.go | 223 + .../protobuf/internal/impl/codec_field.go | 828 + .../protobuf/internal/impl/codec_gen.go | 5637 ++ .../protobuf/internal/impl/codec_map.go | 388 + .../protobuf/internal/impl/codec_map_go111.go | 37 + .../protobuf/internal/impl/codec_map_go112.go | 11 + .../protobuf/internal/impl/codec_message.go | 159 + .../internal/impl/codec_messageset.go | 120 + .../protobuf/internal/impl/codec_reflect.go | 209 + .../protobuf/internal/impl/codec_tables.go | 557 + .../protobuf/internal/impl/codec_unsafe.go | 17 + .../protobuf/internal/impl/convert.go | 467 + .../protobuf/internal/impl/convert_list.go | 141 + .../protobuf/internal/impl/convert_map.go | 121 + .../protobuf/internal/impl/decode.go | 274 + .../protobuf/internal/impl/encode.go | 199 + .../protobuf/internal/impl/enum.go | 21 + .../protobuf/internal/impl/extension.go | 156 + .../protobuf/internal/impl/legacy_enum.go | 219 + .../protobuf/internal/impl/legacy_export.go | 92 + .../internal/impl/legacy_extension.go | 175 + .../protobuf/internal/impl/legacy_file.go | 81 + .../protobuf/internal/impl/legacy_message.go | 502 + .../protobuf/internal/impl/merge.go | 176 + .../protobuf/internal/impl/merge_gen.go | 209 + .../protobuf/internal/impl/message.go | 215 + .../protobuf/internal/impl/message_reflect.go | 364 + .../internal/impl/message_reflect_field.go | 466 + .../internal/impl/message_reflect_gen.go | 249 + .../protobuf/internal/impl/pointer_reflect.go | 177 + .../protobuf/internal/impl/pointer_unsafe.go | 173 + .../protobuf/internal/impl/validate.go | 575 + .../protobuf/internal/impl/weak.go | 74 + .../protobuf/internal/mapsort/mapsort.go | 43 + .../protobuf/internal/pragma/pragma.go | 29 + .../protobuf/internal/set/ints.go | 58 + .../protobuf/internal/strs/strings.go | 196 + .../protobuf/internal/strs/strings_pure.go | 27 + .../protobuf/internal/strs/strings_unsafe.go | 94 + .../protobuf/internal/version/version.go | 79 + .../protobuf/proto/checkinit.go | 71 + .../protobuf/proto/decode.go | 273 + .../protobuf/proto/decode_gen.go | 603 + .../google.golang.org/protobuf/proto/doc.go | 94 + .../protobuf/proto/encode.go | 346 + .../protobuf/proto/encode_gen.go | 97 + .../google.golang.org/protobuf/proto/equal.go | 154 + .../protobuf/proto/extension.go | 92 + .../google.golang.org/protobuf/proto/merge.go | 139 + .../protobuf/proto/messageset.go | 88 + .../google.golang.org/protobuf/proto/proto.go | 34 + .../protobuf/proto/proto_methods.go | 19 + .../protobuf/proto/proto_reflect.go | 19 + .../google.golang.org/protobuf/proto/reset.go | 43 + .../google.golang.org/protobuf/proto/size.go | 97 + .../protobuf/proto/size_gen.go | 55 + .../protobuf/proto/wrappers.go | 29 + .../protobuf/reflect/protoreflect/methods.go | 77 + .../protobuf/reflect/protoreflect/proto.go | 478 + .../protobuf/reflect/protoreflect/source.go | 52 + .../protobuf/reflect/protoreflect/type.go | 631 + .../protobuf/reflect/protoreflect/value.go | 285 + .../reflect/protoreflect/value_pure.go | 59 + .../reflect/protoreflect/value_union.go | 411 + .../reflect/protoreflect/value_unsafe.go | 98 + .../reflect/protoregistry/registry.go | 800 + .../protobuf/runtime/protoiface/legacy.go | 15 + .../protobuf/runtime/protoiface/methods.go | 167 + .../protobuf/runtime/protoimpl/impl.go | 44 + .../protobuf/runtime/protoimpl/version.go | 56 + .../protobuf/types/known/anypb/any.pb.go | 287 + .../types/known/durationpb/duration.pb.go | 249 + .../types/known/timestamppb/timestamp.pb.go | 271 + vendor/gopkg.in/yaml.v3/.travis.yml | 17 + vendor/gopkg.in/yaml.v3/LICENSE | 50 + vendor/gopkg.in/yaml.v3/NOTICE | 13 + vendor/gopkg.in/yaml.v3/README.md | 150 + vendor/gopkg.in/yaml.v3/apic.go | 747 + vendor/gopkg.in/yaml.v3/decode.go | 948 + vendor/gopkg.in/yaml.v3/emitterc.go | 2022 + vendor/gopkg.in/yaml.v3/encode.go | 572 + vendor/gopkg.in/yaml.v3/go.mod | 5 + vendor/gopkg.in/yaml.v3/parserc.go | 1249 + vendor/gopkg.in/yaml.v3/readerc.go | 434 + vendor/gopkg.in/yaml.v3/resolve.go | 326 + vendor/gopkg.in/yaml.v3/scannerc.go | 3028 + vendor/gopkg.in/yaml.v3/sorter.go | 134 + vendor/gopkg.in/yaml.v3/writerc.go | 48 + vendor/gopkg.in/yaml.v3/yaml.go | 693 + vendor/gopkg.in/yaml.v3/yamlh.go | 807 + vendor/gopkg.in/yaml.v3/yamlprivateh.go | 198 + .../admissionregistration/v1/generated.pb.go | 60 +- .../admissionregistration/v1/generated.proto | 4 +- .../api/admissionregistration/v1/types.go | 4 +- .../v1/types_swagger_doc_generated.go | 2 +- .../api/admissionregistration/v1beta1/doc.go | 1 + .../v1beta1/generated.pb.go | 60 +- .../v1beta1/generated.proto | 4 +- .../admissionregistration/v1beta1/types.go | 20 +- .../v1beta1/types_swagger_doc_generated.go | 2 +- .../zz_generated.prerelease-lifecycle.go | 121 + vendor/k8s.io/api/apps/v1/generated.pb.go | 60 +- vendor/k8s.io/api/apps/v1beta1/doc.go | 1 + .../k8s.io/api/apps/v1beta1/generated.pb.go | 60 +- vendor/k8s.io/api/apps/v1beta1/types.go | 32 + .../zz_generated.prerelease-lifecycle.go | 217 + vendor/k8s.io/api/apps/v1beta2/doc.go | 1 + .../k8s.io/api/apps/v1beta2/generated.pb.go | 60 +- vendor/k8s.io/api/apps/v1beta2/types.go | 44 + .../zz_generated.prerelease-lifecycle.go | 289 + .../v1alpha1/generated.pb.go | 2056 - .../v1alpha1/generated.proto | 162 - .../api/auditregistration/v1alpha1/types.go | 198 - .../v1alpha1/types_swagger_doc_generated.go | 111 - .../v1alpha1/zz_generated.deepcopy.go | 229 - .../api/authentication/v1/generated.pb.go | 60 +- vendor/k8s.io/api/authentication/v1/types.go | 2 +- .../k8s.io/api/authentication/v1beta1/doc.go | 1 + .../authentication/v1beta1/generated.pb.go | 60 +- .../api/authentication/v1beta1/types.go | 5 +- .../zz_generated.prerelease-lifecycle.go | 49 + .../api/authorization/v1/generated.pb.go | 60 +- vendor/k8s.io/api/authorization/v1/types.go | 8 +- .../k8s.io/api/authorization/v1beta1/doc.go | 1 + .../api/authorization/v1beta1/generated.pb.go | 60 +- .../k8s.io/api/authorization/v1beta1/types.go | 20 +- .../zz_generated.prerelease-lifecycle.go | 121 + .../k8s.io/api/autoscaling/v1/generated.pb.go | 60 +- vendor/k8s.io/api/autoscaling/v2beta1/doc.go | 1 + .../api/autoscaling/v2beta1/generated.pb.go | 60 +- .../k8s.io/api/autoscaling/v2beta1/types.go | 6 + .../zz_generated.prerelease-lifecycle.go | 73 + vendor/k8s.io/api/autoscaling/v2beta2/doc.go | 1 + .../api/autoscaling/v2beta2/generated.pb.go | 1298 +- .../api/autoscaling/v2beta2/generated.proto | 66 + .../k8s.io/api/autoscaling/v2beta2/types.go | 88 + .../v2beta2/types_swagger_doc_generated.go | 33 + .../v2beta2/zz_generated.deepcopy.go | 78 + .../zz_generated.prerelease-lifecycle.go | 57 + vendor/k8s.io/api/batch/v1/generated.pb.go | 60 +- vendor/k8s.io/api/batch/v1beta1/doc.go | 1 + .../k8s.io/api/batch/v1beta1/generated.pb.go | 60 +- vendor/k8s.io/api/batch/v1beta1/types.go | 8 +- .../zz_generated.prerelease-lifecycle.go | 75 + .../k8s.io/api/batch/v2alpha1/generated.pb.go | 60 +- .../v1alpha1 => certificates/v1}/doc.go | 6 +- .../api/certificates/v1/generated.pb.go | 2042 + .../api/certificates/v1/generated.proto | 226 + vendor/k8s.io/api/certificates/v1/register.go | 61 + vendor/k8s.io/api/certificates/v1/types.go | 284 + .../v1/types_swagger_doc_generated.go | 88 + .../certificates/v1/zz_generated.deepcopy.go | 198 + vendor/k8s.io/api/certificates/v1beta1/doc.go | 1 + .../api/certificates/v1beta1/generated.pb.go | 301 +- .../api/certificates/v1beta1/generated.proto | 59 +- .../k8s.io/api/certificates/v1beta1/types.go | 86 +- .../v1beta1/types_swagger_doc_generated.go | 25 +- .../v1beta1/zz_generated.deepcopy.go | 6 + .../zz_generated.prerelease-lifecycle.go | 73 + .../api/coordination/v1/generated.pb.go | 60 +- vendor/k8s.io/api/coordination/v1beta1/doc.go | 1 + .../api/coordination/v1beta1/generated.pb.go | 60 +- .../k8s.io/api/coordination/v1beta1/types.go | 6 + .../zz_generated.prerelease-lifecycle.go | 73 + .../api/core/v1/annotation_key_constants.go | 27 +- vendor/k8s.io/api/core/v1/generated.pb.go | 3455 +- vendor/k8s.io/api/core/v1/generated.proto | 307 +- vendor/k8s.io/api/core/v1/lifecycle.go | 37 + vendor/k8s.io/api/core/v1/resource.go | 8 + vendor/k8s.io/api/core/v1/types.go | 350 +- .../core/v1/types_swagger_doc_generated.go | 134 +- .../k8s.io/api/core/v1/well_known_taints.go | 7 - .../api/core/v1/zz_generated.deepcopy.go | 113 +- .../api/discovery/v1alpha1/generated.pb.go | 60 +- .../api/discovery/v1alpha1/generated.proto | 1 - vendor/k8s.io/api/discovery/v1alpha1/types.go | 1 - vendor/k8s.io/api/discovery/v1beta1/doc.go | 1 + .../api/discovery/v1beta1/generated.pb.go | 60 +- .../api/discovery/v1beta1/generated.proto | 6 +- vendor/k8s.io/api/discovery/v1beta1/types.go | 10 +- .../v1beta1/types_swagger_doc_generated.go | 2 +- .../discovery/v1beta1/well_known_labels.go | 4 + .../zz_generated.prerelease-lifecycle.go | 57 + vendor/k8s.io/api/events/v1/doc.go | 23 + vendor/k8s.io/api/events/v1/generated.pb.go | 1406 + vendor/k8s.io/api/events/v1/generated.proto | 125 + .../v1alpha1 => events/v1}/register.go | 25 +- vendor/k8s.io/api/events/v1/types.go | 119 + .../events/v1/types_swagger_doc_generated.go | 72 + .../api/events/v1/zz_generated.deepcopy.go | 117 + vendor/k8s.io/api/events/v1beta1/doc.go | 1 + .../k8s.io/api/events/v1beta1/generated.pb.go | 198 +- .../k8s.io/api/events/v1beta1/generated.proto | 47 +- vendor/k8s.io/api/events/v1beta1/types.go | 60 +- .../v1beta1/types_swagger_doc_generated.go | 35 +- .../zz_generated.prerelease-lifecycle.go | 57 + vendor/k8s.io/api/extensions/v1beta1/doc.go | 1 + .../api/extensions/v1beta1/generated.pb.go | 815 +- .../api/extensions/v1beta1/generated.proto | 109 +- .../k8s.io/api/extensions/v1beta1/register.go | 1 - vendor/k8s.io/api/extensions/v1beta1/types.go | 200 +- .../v1beta1/types_swagger_doc_generated.go | 49 +- .../v1beta1/zz_generated.deepcopy.go | 48 +- .../zz_generated.prerelease-lifecycle.go | 349 + .../api/flowcontrol/v1alpha1/generated.pb.go | 60 +- .../api/flowcontrol/v1alpha1/generated.proto | 30 +- .../k8s.io/api/flowcontrol/v1alpha1/types.go | 40 +- .../v1alpha1/types_swagger_doc_generated.go | 18 +- .../k8s.io/api/networking/v1/generated.pb.go | 4410 +- .../k8s.io/api/networking/v1/generated.proto | 280 +- vendor/k8s.io/api/networking/v1/register.go | 4 + vendor/k8s.io/api/networking/v1/types.go | 323 +- .../v1/types_swagger_doc_generated.go | 157 +- .../networking/v1/zz_generated.deepcopy.go | 362 + vendor/k8s.io/api/networking/v1beta1/doc.go | 1 + .../api/networking/v1beta1/generated.pb.go | 1165 +- .../api/networking/v1beta1/generated.proto | 133 +- .../k8s.io/api/networking/v1beta1/register.go | 2 + vendor/k8s.io/api/networking/v1beta1/types.go | 189 +- .../v1beta1/types_swagger_doc_generated.go | 51 +- .../v1beta1/well_known_annotations.go | 32 + .../v1beta1/zz_generated.deepcopy.go | 105 +- .../zz_generated.prerelease-lifecycle.go | 121 + .../k8s.io/api/node/v1alpha1/generated.pb.go | 60 +- vendor/k8s.io/api/node/v1beta1/doc.go | 1 + .../k8s.io/api/node/v1beta1/generated.pb.go | 60 +- vendor/k8s.io/api/node/v1beta1/types.go | 4 + .../zz_generated.prerelease-lifecycle.go | 57 + vendor/k8s.io/api/policy/v1beta1/doc.go | 1 + .../k8s.io/api/policy/v1beta1/generated.pb.go | 302 +- .../k8s.io/api/policy/v1beta1/generated.proto | 18 +- vendor/k8s.io/api/policy/v1beta1/types.go | 31 +- .../v1beta1/types_swagger_doc_generated.go | 16 +- .../zz_generated.prerelease-lifecycle.go | 111 + vendor/k8s.io/api/rbac/v1/generated.pb.go | 60 +- .../k8s.io/api/rbac/v1alpha1/generated.pb.go | 60 +- .../k8s.io/api/rbac/v1alpha1/generated.proto | 17 +- vendor/k8s.io/api/rbac/v1alpha1/types.go | 17 +- .../v1alpha1/types_swagger_doc_generated.go | 18 +- vendor/k8s.io/api/rbac/v1beta1/doc.go | 1 + .../k8s.io/api/rbac/v1beta1/generated.pb.go | 60 +- .../k8s.io/api/rbac/v1beta1/generated.proto | 16 +- vendor/k8s.io/api/rbac/v1beta1/types.go | 48 +- .../v1beta1/types_swagger_doc_generated.go | 16 +- .../zz_generated.prerelease-lifecycle.go | 217 + .../k8s.io/api/scheduling/v1/generated.pb.go | 60 +- .../k8s.io/api/scheduling/v1/generated.proto | 2 +- vendor/k8s.io/api/scheduling/v1/types.go | 2 +- .../v1/types_swagger_doc_generated.go | 2 +- .../api/scheduling/v1alpha1/generated.pb.go | 60 +- .../api/scheduling/v1alpha1/generated.proto | 2 +- .../k8s.io/api/scheduling/v1alpha1/types.go | 2 +- .../v1alpha1/types_swagger_doc_generated.go | 2 +- vendor/k8s.io/api/scheduling/v1beta1/doc.go | 1 + .../api/scheduling/v1beta1/generated.pb.go | 60 +- .../api/scheduling/v1beta1/generated.proto | 2 +- vendor/k8s.io/api/scheduling/v1beta1/types.go | 10 +- .../v1beta1/types_swagger_doc_generated.go | 2 +- .../zz_generated.prerelease-lifecycle.go | 73 + .../api/settings/v1alpha1/generated.pb.go | 60 +- vendor/k8s.io/api/storage/v1/generated.pb.go | 1089 +- vendor/k8s.io/api/storage/v1/generated.proto | 119 + vendor/k8s.io/api/storage/v1/register.go | 3 + vendor/k8s.io/api/storage/v1/types.go | 185 + .../storage/v1/types_swagger_doc_generated.go | 33 + .../api/storage/v1/zz_generated.deepcopy.go | 101 + .../api/storage/v1alpha1/generated.pb.go | 767 +- .../api/storage/v1alpha1/generated.proto | 75 + .../k8s.io/api/storage/v1alpha1/register.go | 2 + vendor/k8s.io/api/storage/v1alpha1/types.go | 82 + .../v1alpha1/types_swagger_doc_generated.go | 22 + .../storage/v1alpha1/zz_generated.deepcopy.go | 74 +- vendor/k8s.io/api/storage/v1beta1/doc.go | 1 + .../api/storage/v1beta1/generated.pb.go | 314 +- .../api/storage/v1beta1/generated.proto | 28 + vendor/k8s.io/api/storage/v1beta1/types.go | 80 + .../v1beta1/types_swagger_doc_generated.go | 2 + .../storage/v1beta1/zz_generated.deepcopy.go | 10 + .../zz_generated.prerelease-lifecycle.go | 217 + .../pkg/apis/apiextensions/types.go | 9 + .../v1beta1/.import-restrictions | 5 + .../pkg/apis/apiextensions/v1beta1/doc.go | 1 + .../apiextensions/v1beta1/generated.pb.go | 512 +- .../apiextensions/v1beta1/generated.proto | 20 +- .../pkg/apis/apiextensions/v1beta1/types.go | 28 +- .../apiextensions/v1beta1/types_jsonschema.go | 3 + .../v1beta1/zz_generated.conversion.go | 4 + .../v1beta1/zz_generated.deepcopy.go | 5 + .../zz_generated.prerelease-lifecycle.go | 97 + .../apiextensions/zz_generated.deepcopy.go | 5 + .../k8s.io/apimachinery/pkg/api/errors/OWNERS | 2 - .../apimachinery/pkg/api/errors/errors.go | 82 +- .../k8s.io/apimachinery/pkg/api/meta/OWNERS | 2 - .../apimachinery/pkg/api/meta/conditions.go | 101 + .../k8s.io/apimachinery/pkg/api/meta/meta.go | 2 +- .../apimachinery/pkg/api/resource/OWNERS | 3 - .../pkg/api/resource/generated.pb.go | 2 +- .../apimachinery/pkg/api/resource/math.go | 8 +- .../apimachinery/pkg/api/resource/quantity.go | 5 + .../apimachinery/pkg/apis/meta/v1/OWNERS | 1 - .../pkg/apis/meta/v1/conversion.go | 67 +- .../apimachinery/pkg/apis/meta/v1/duration.go | 5 + .../pkg/apis/meta/v1/generated.pb.go | 890 +- .../pkg/apis/meta/v1/generated.proto | 101 +- .../apimachinery/pkg/apis/meta/v1/helpers.go | 4 +- .../apimachinery/pkg/apis/meta/v1/meta.go | 4 +- .../apimachinery/pkg/apis/meta/v1/register.go | 17 +- .../apimachinery/pkg/apis/meta/v1/time.go | 10 + .../apimachinery/pkg/apis/meta/v1/types.go | 122 +- .../meta/v1/types_swagger_doc_generated.go | 37 +- .../pkg/apis/meta/v1/unstructured/helpers.go | 2 +- .../apis/meta/v1/zz_generated.conversion.go | 12 + .../pkg/apis/meta/v1/zz_generated.deepcopy.go | 17 + .../apimachinery/pkg/conversion/converter.go | 165 +- .../apimachinery/pkg/fields/selector.go | 16 +- .../k8s.io/apimachinery/pkg/labels/labels.go | 14 +- .../apimachinery/pkg/labels/selector.go | 41 +- .../k8s.io/apimachinery/pkg/runtime/codec.go | 2 +- .../apimachinery/pkg/runtime/codec_check.go | 10 +- .../apimachinery/pkg/runtime/conversion.go | 32 +- .../apimachinery/pkg/runtime/converter.go | 124 +- .../apimachinery/pkg/runtime/embedded.go | 15 +- .../apimachinery/pkg/runtime/generated.pb.go | 60 +- .../pkg/runtime/schema/generated.pb.go | 2 +- .../pkg/runtime/schema/group_version.go | 17 +- .../k8s.io/apimachinery/pkg/runtime/scheme.go | 58 +- .../pkg/runtime/serializer/json/json.go | 2 +- .../serializer/versioning/versioning.go | 2 +- .../apimachinery/pkg/util/clock/clock.go | 29 +- .../apimachinery/pkg/util/errors/errors.go | 32 +- .../pkg/util/intstr/generated.pb.go | 60 +- .../apimachinery/pkg/util/intstr/intstr.go | 5 +- .../k8s.io/apimachinery/pkg/util/json/json.go | 25 + .../k8s.io/apimachinery/pkg/util/net/http.go | 301 +- .../apimachinery/pkg/util/net/interface.go | 2 +- .../k8s.io/apimachinery/pkg/util/net/util.go | 31 +- .../apimachinery/pkg/util/runtime/runtime.go | 2 +- .../pkg/util/strategicpatch/patch.go | 4 +- .../pkg/util/validation/validation.go | 43 + .../k8s.io/apimachinery/pkg/util/wait/wait.go | 150 +- .../apimachinery/pkg/util/yaml/decoder.go | 6 +- .../apimachinery/pkg/watch/streamwatcher.go | 4 +- vendor/k8s.io/apimachinery/pkg/watch/watch.go | 18 +- .../forked/golang/reflect/deep_equal.go | 4 +- .../client-go/discovery/discovery_client.go | 15 +- .../client-go/discovery/fake/discovery.go | 2 +- vendor/k8s.io/client-go/dynamic/interface.go | 20 +- vendor/k8s.io/client-go/dynamic/simple.go | 47 +- .../client-go/kubernetes/scheme/register.go | 6 +- .../typed/apps/v1/controllerrevision.go | 64 +- .../kubernetes/typed/apps/v1/daemonset.go | 72 +- .../kubernetes/typed/apps/v1/deployment.go | 85 +- .../kubernetes/typed/apps/v1/replicaset.go | 85 +- .../kubernetes/typed/apps/v1/statefulset.go | 85 +- .../typed/core/v1/componentstatus.go | 64 +- .../kubernetes/typed/core/v1/configmap.go | 64 +- .../kubernetes/typed/core/v1/endpoints.go | 64 +- .../kubernetes/typed/core/v1/event.go | 64 +- .../typed/core/v1/event_expansion.go | 11 +- .../typed/core/v1/generated_expansion.go | 2 + .../kubernetes/typed/core/v1/limitrange.go | 64 +- .../kubernetes/typed/core/v1/namespace.go | 58 +- .../typed/core/v1/namespace_expansion.go | 14 +- .../kubernetes/typed/core/v1/node.go | 72 +- .../typed/core/v1/node_expansion.go | 10 +- .../typed/core/v1/persistentvolume.go | 72 +- .../typed/core/v1/persistentvolumeclaim.go | 72 +- .../client-go/kubernetes/typed/core/v1/pod.go | 85 +- .../kubernetes/typed/core/v1/pod_expansion.go | 33 +- .../kubernetes/typed/core/v1/podtemplate.go | 64 +- .../typed/core/v1/replicationcontroller.go | 85 +- .../kubernetes/typed/core/v1/resourcequota.go | 72 +- .../kubernetes/typed/core/v1/secret.go | 64 +- .../kubernetes/typed/core/v1/service.go | 58 +- .../typed/core/v1/serviceaccount.go | 82 +- .../typed/core/v1/serviceaccount_expansion.go | 41 - .../plugin/pkg/client/auth/azure/README.md | 8 +- .../plugin/pkg/client/auth/azure/azure.go | 189 +- .../plugin/pkg/client/auth/exec/exec.go | 131 +- .../plugin/pkg/client/auth/exec/metrics.go | 60 + .../plugin/pkg/client/auth/gcp/gcp.go | 2 +- .../plugin/pkg/client/auth/oidc/oidc.go | 29 +- .../pkg/client/auth/openstack/openstack.go | 193 - .../client/auth/openstack/openstack_stub.go | 36 + vendor/k8s.io/client-go/rest/client.go | 4 + vendor/k8s.io/client-go/rest/config.go | 30 +- vendor/k8s.io/client-go/rest/plugin.go | 2 +- vendor/k8s.io/client-go/rest/request.go | 238 +- vendor/k8s.io/client-go/rest/transport.go | 3 +- vendor/k8s.io/client-go/rest/urlbackoff.go | 2 +- vendor/k8s.io/client-go/rest/warnings.go | 144 + .../k8s.io/client-go/restmapper/discovery.go | 2 +- .../k8s.io/client-go/restmapper/shortcut.go | 2 +- vendor/k8s.io/client-go/testing/actions.go | 12 +- vendor/k8s.io/client-go/testing/fixture.go | 97 +- .../client-go/tools/clientcmd/api/helpers.go | 3 + .../client-go/tools/clientcmd/api/types.go | 21 +- .../client-go/tools/clientcmd/api/v1/types.go | 21 +- .../api/v1/zz_generated.conversion.go | 6 + .../tools/clientcmd/client_config.go | 99 +- .../client-go/tools/clientcmd/config.go | 29 +- .../client-go/tools/clientcmd/helpers.go | 15 + .../client-go/tools/clientcmd/loader.go | 2 +- .../tools/clientcmd/merged_client_builder.go | 37 +- .../client-go/tools/clientcmd/overrides.go | 4 + .../client-go/tools/clientcmd/validation.go | 61 +- vendor/k8s.io/client-go/tools/metrics/OWNERS | 1 - .../k8s.io/client-go/tools/metrics/metrics.go | 52 +- vendor/k8s.io/client-go/transport/cache.go | 42 +- .../client-go/transport/cert_rotation.go | 176 + vendor/k8s.io/client-go/transport/config.go | 15 +- .../client-go/transport/round_trippers.go | 10 +- .../client-go/transport/token_source.go | 2 +- .../k8s.io/client-go/transport/transport.go | 62 +- vendor/k8s.io/client-go/util/cert/cert.go | 6 +- .../client-go/util/jsonpath/jsonpath.go | 78 +- .../k8s.io/client-go/util/jsonpath/parser.go | 5 +- .../util/workqueue/default_rate_limiters.go | 211 + .../util/workqueue/delaying_queue.go | 280 + vendor/k8s.io/client-go/util/workqueue/doc.go | 26 + .../client-go/util/workqueue/metrics.go | 261 + .../client-go/util/workqueue/parallelizer.go | 101 + .../k8s.io/client-go/util/workqueue/queue.go | 212 + .../util/workqueue/rate_limiting_queue.go | 69 + vendor/k8s.io/klog/v2/.gitignore | 17 + vendor/k8s.io/klog/v2/CONTRIBUTING.md | 22 + vendor/k8s.io/klog/v2/LICENSE | 191 + vendor/k8s.io/klog/v2/OWNERS | 19 + vendor/k8s.io/klog/v2/README.md | 99 + vendor/k8s.io/klog/v2/RELEASE.md | 9 + vendor/k8s.io/klog/v2/SECURITY_CONTACTS | 20 + vendor/k8s.io/klog/v2/code-of-conduct.md | 3 + vendor/k8s.io/klog/v2/go.mod | 5 + vendor/k8s.io/klog/v2/go.sum | 2 + vendor/k8s.io/klog/v2/klog.go | 1525 + vendor/k8s.io/klog/v2/klog_file.go | 164 + .../kube-openapi/pkg/util/proto/document.go | 2 +- .../pkg/util/proto/validation/types.go | 2 +- .../gophercloud => k8s.io/kubectl}/LICENSE | 42 +- vendor/k8s.io/kubectl/pkg/util/openapi/OWNERS | 6 + .../cmd => kubectl/pkg}/util/openapi/doc.go | 2 +- .../pkg}/util/openapi/extensions.go | 1 + .../pkg}/util/openapi/openapi.go | 3 +- .../pkg}/util/openapi/openapi_getter.go | 0 .../pkg/kubectl/cmd/util/openapi/BUILD | 72 - .../pkg/kubectl/cmd/util/openapi/OWNERS | 4 - .../pkg/kubectl/cmd/util/openapi/dryrun.go | 65 - vendor/modules.txt | 103 +- .../structured-merge-diff/v4/LICENSE | 201 + .../v4/value/allocator.go | 203 + .../structured-merge-diff/v4/value/doc.go | 21 + .../structured-merge-diff/v4/value/fields.go | 97 + .../v4/value/jsontagutil.go | 91 + .../structured-merge-diff/v4/value/list.go | 139 + .../v4/value/listreflect.go | 98 + .../v4/value/listunstructured.go | 74 + .../structured-merge-diff/v4/value/map.go | 270 + .../v4/value/mapreflect.go | 209 + .../v4/value/mapunstructured.go | 190 + .../v4/value/reflectcache.go | 463 + .../structured-merge-diff/v4/value/scalar.go | 50 + .../v4/value/structreflect.go | 208 + .../structured-merge-diff/v4/value/value.go | 347 + .../v4/value/valuereflect.go | 294 + .../v4/value/valueunstructured.go | 178 + 799 files changed, 128464 insertions(+), 79282 deletions(-) create mode 100644 vendor/cloud.google.com/go/compute/metadata/.repo-metadata.json rename vendor/{k8s.io/kubernetes => github.com/go-logr/logr}/LICENSE (99%) create mode 100644 vendor/github.com/go-logr/logr/README.md create mode 100644 vendor/github.com/go-logr/logr/go.mod create mode 100644 vendor/github.com/go-logr/logr/logr.go create mode 100644 vendor/github.com/golang/protobuf/proto/buffer.go delete mode 100644 vendor/github.com/golang/protobuf/proto/clone.go delete mode 100644 vendor/github.com/golang/protobuf/proto/decode.go create mode 100644 vendor/github.com/golang/protobuf/proto/defaults.go delete mode 100644 vendor/github.com/golang/protobuf/proto/encode.go delete mode 100644 vendor/github.com/golang/protobuf/proto/equal.go delete mode 100644 vendor/github.com/golang/protobuf/proto/lib.go delete mode 100644 vendor/github.com/golang/protobuf/proto/message_set.go delete mode 100644 vendor/github.com/golang/protobuf/proto/pointer_reflect.go delete mode 100644 vendor/github.com/golang/protobuf/proto/pointer_unsafe.go create mode 100644 vendor/github.com/golang/protobuf/proto/proto.go create mode 100644 vendor/github.com/golang/protobuf/proto/registry.go delete mode 100644 vendor/github.com/golang/protobuf/proto/table_marshal.go delete mode 100644 vendor/github.com/golang/protobuf/proto/table_merge.go delete mode 100644 vendor/github.com/golang/protobuf/proto/table_unmarshal.go delete mode 100644 vendor/github.com/golang/protobuf/proto/text.go create mode 100644 vendor/github.com/golang/protobuf/proto/text_decode.go create mode 100644 vendor/github.com/golang/protobuf/proto/text_encode.go delete mode 100644 vendor/github.com/golang/protobuf/proto/text_parser.go create mode 100644 vendor/github.com/golang/protobuf/proto/wire.go create mode 100644 vendor/github.com/golang/protobuf/proto/wrappers.go delete mode 100644 vendor/github.com/golang/protobuf/ptypes/any/any.proto delete mode 100644 vendor/github.com/golang/protobuf/ptypes/duration/duration.proto delete mode 100644 vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.proto delete mode 100644 vendor/github.com/google/go-jsonnet/.golangci.yml delete mode 100644 vendor/github.com/google/go-jsonnet/.goreleaser.yml delete mode 100644 vendor/github.com/google/go-jsonnet/MANIFEST.in delete mode 100644 vendor/github.com/google/go-jsonnet/internal/parser/string_util.go delete mode 100644 vendor/github.com/google/go-jsonnet/update_cpp_jsonnet.sh delete mode 100644 vendor/github.com/google/go-jsonnet/util.go delete mode 100644 vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go delete mode 100644 vendor/github.com/googleapis/gnostic/OpenAPIv2/README.md delete mode 100644 vendor/github.com/googleapis/gnostic/compiler/extension-handler.go create mode 100644 vendor/github.com/googleapis/gnostic/compiler/extensions.go create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/README.md create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/base.go create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/display.go create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/models.go create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/operations.go create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/reader.go create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/schema.json create mode 100644 vendor/github.com/googleapis/gnostic/jsonschema/writer.go rename vendor/github.com/googleapis/gnostic/{OpenAPIv2 => openapiv2}/OpenAPIv2.go (65%) create mode 100644 vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.pb.go rename vendor/github.com/googleapis/gnostic/{OpenAPIv2 => openapiv2}/OpenAPIv2.proto (99%) create mode 100644 vendor/github.com/googleapis/gnostic/openapiv2/README.md create mode 100644 vendor/github.com/googleapis/gnostic/openapiv2/document.go rename vendor/github.com/googleapis/gnostic/{OpenAPIv2 => openapiv2}/openapi-2.0.json (99%) delete mode 100644 vendor/github.com/gophercloud/gophercloud/.gitignore delete mode 100644 vendor/github.com/gophercloud/gophercloud/.travis.yml delete mode 100644 vendor/github.com/gophercloud/gophercloud/.zuul.yaml delete mode 100644 vendor/github.com/gophercloud/gophercloud/CHANGELOG.md delete mode 100644 vendor/github.com/gophercloud/gophercloud/README.md delete mode 100644 vendor/github.com/gophercloud/gophercloud/auth_options.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/auth_result.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/doc.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/endpoint_search.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/errors.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/go.mod delete mode 100644 vendor/github.com/gophercloud/gophercloud/go.sum delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/auth_env.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/client.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/doc.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/endpoint_location.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/errors.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/doc.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/requests.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/results.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/urls.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/doc.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/requests.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/results.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/urls.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/doc.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/requests.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/results.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/urls.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/utils/base_endpoint.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/openstack/utils/choose_version.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/pagination/http.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/pagination/linked.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/pagination/marker.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/pagination/pager.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/pagination/pkg.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/pagination/single.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/params.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/provider_client.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/results.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/service_client.go delete mode 100644 vendor/github.com/gophercloud/gophercloud/util.go delete mode 100644 vendor/github.com/sergi/go-diff/diffmatchpatch/operation_string.go create mode 100644 vendor/github.com/sirupsen/logrus/.golangci.yml create mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_js.go create mode 100644 vendor/github.com/spf13/cobra/custom_completions.go create mode 100644 vendor/github.com/spf13/cobra/fish_completions.go create mode 100644 vendor/github.com/spf13/cobra/fish_completions.md create mode 100644 vendor/golang.org/x/sys/internal/unsafeheader/unsafeheader.go create mode 100644 vendor/golang.org/x/text/unicode/bidi/tables12.0.0.go create mode 100644 vendor/golang.org/x/text/unicode/norm/tables12.0.0.go create mode 100644 vendor/golang.org/x/text/width/tables12.0.0.go create mode 100644 vendor/google.golang.org/protobuf/AUTHORS create mode 100644 vendor/google.golang.org/protobuf/CONTRIBUTORS create mode 100644 vendor/google.golang.org/protobuf/LICENSE create mode 100644 vendor/google.golang.org/protobuf/PATENTS create mode 100644 vendor/google.golang.org/protobuf/encoding/prototext/decode.go create mode 100644 vendor/google.golang.org/protobuf/encoding/prototext/doc.go create mode 100644 vendor/google.golang.org/protobuf/encoding/prototext/encode.go create mode 100644 vendor/google.golang.org/protobuf/encoding/protowire/wire.go create mode 100644 vendor/google.golang.org/protobuf/internal/descfmt/stringer.go create mode 100644 vendor/google.golang.org/protobuf/internal/descopts/options.go create mode 100644 vendor/google.golang.org/protobuf/internal/detrand/rand.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/defval/default.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/messageset/messageset.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/tag/tag.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/text/decode.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/text/decode_number.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/text/decode_string.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/text/decode_token.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/text/doc.go create mode 100644 vendor/google.golang.org/protobuf/internal/encoding/text/encode.go create mode 100644 vendor/google.golang.org/protobuf/internal/errors/errors.go create mode 100644 vendor/google.golang.org/protobuf/internal/errors/is_go112.go create mode 100644 vendor/google.golang.org/protobuf/internal/errors/is_go113.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/any_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/api_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/descriptor_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/doc.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/duration_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/empty_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/field_mask_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/source_context_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/struct_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/timestamp_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/type_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldnum/wrappers_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/fieldsort/fieldsort.go create mode 100644 vendor/google.golang.org/protobuf/internal/filedesc/build.go create mode 100644 vendor/google.golang.org/protobuf/internal/filedesc/desc.go create mode 100644 vendor/google.golang.org/protobuf/internal/filedesc/desc_init.go create mode 100644 vendor/google.golang.org/protobuf/internal/filedesc/desc_lazy.go create mode 100644 vendor/google.golang.org/protobuf/internal/filedesc/desc_list.go create mode 100644 vendor/google.golang.org/protobuf/internal/filedesc/desc_list_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/filedesc/placeholder.go create mode 100644 vendor/google.golang.org/protobuf/internal/filetype/build.go create mode 100644 vendor/google.golang.org/protobuf/internal/flags/flags.go create mode 100644 vendor/google.golang.org/protobuf/internal/flags/proto_legacy_disable.go create mode 100644 vendor/google.golang.org/protobuf/internal/flags/proto_legacy_enable.go create mode 100644 vendor/google.golang.org/protobuf/internal/genname/name.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/api_export.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/checkinit.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_extension.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_field.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_map.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_map_go111.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_map_go112.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_message.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_messageset.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_reflect.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_tables.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/codec_unsafe.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/convert.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/convert_list.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/convert_map.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/decode.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/encode.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/enum.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/extension.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/legacy_enum.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/legacy_export.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/legacy_extension.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/legacy_file.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/legacy_message.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/merge.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/merge_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/message.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/message_reflect.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/message_reflect_field.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/message_reflect_gen.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/pointer_reflect.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/pointer_unsafe.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/validate.go create mode 100644 vendor/google.golang.org/protobuf/internal/impl/weak.go create mode 100644 vendor/google.golang.org/protobuf/internal/mapsort/mapsort.go create mode 100644 vendor/google.golang.org/protobuf/internal/pragma/pragma.go create mode 100644 vendor/google.golang.org/protobuf/internal/set/ints.go create mode 100644 vendor/google.golang.org/protobuf/internal/strs/strings.go create mode 100644 vendor/google.golang.org/protobuf/internal/strs/strings_pure.go create mode 100644 vendor/google.golang.org/protobuf/internal/strs/strings_unsafe.go create mode 100644 vendor/google.golang.org/protobuf/internal/version/version.go create mode 100644 vendor/google.golang.org/protobuf/proto/checkinit.go create mode 100644 vendor/google.golang.org/protobuf/proto/decode.go create mode 100644 vendor/google.golang.org/protobuf/proto/decode_gen.go create mode 100644 vendor/google.golang.org/protobuf/proto/doc.go create mode 100644 vendor/google.golang.org/protobuf/proto/encode.go create mode 100644 vendor/google.golang.org/protobuf/proto/encode_gen.go create mode 100644 vendor/google.golang.org/protobuf/proto/equal.go create mode 100644 vendor/google.golang.org/protobuf/proto/extension.go create mode 100644 vendor/google.golang.org/protobuf/proto/merge.go create mode 100644 vendor/google.golang.org/protobuf/proto/messageset.go create mode 100644 vendor/google.golang.org/protobuf/proto/proto.go create mode 100644 vendor/google.golang.org/protobuf/proto/proto_methods.go create mode 100644 vendor/google.golang.org/protobuf/proto/proto_reflect.go create mode 100644 vendor/google.golang.org/protobuf/proto/reset.go create mode 100644 vendor/google.golang.org/protobuf/proto/size.go create mode 100644 vendor/google.golang.org/protobuf/proto/size_gen.go create mode 100644 vendor/google.golang.org/protobuf/proto/wrappers.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/methods.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/proto.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/source.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/type.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/value.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/value_pure.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/value_union.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoreflect/value_unsafe.go create mode 100644 vendor/google.golang.org/protobuf/reflect/protoregistry/registry.go create mode 100644 vendor/google.golang.org/protobuf/runtime/protoiface/legacy.go create mode 100644 vendor/google.golang.org/protobuf/runtime/protoiface/methods.go create mode 100644 vendor/google.golang.org/protobuf/runtime/protoimpl/impl.go create mode 100644 vendor/google.golang.org/protobuf/runtime/protoimpl/version.go create mode 100644 vendor/google.golang.org/protobuf/types/known/anypb/any.pb.go create mode 100644 vendor/google.golang.org/protobuf/types/known/durationpb/duration.pb.go create mode 100644 vendor/google.golang.org/protobuf/types/known/timestamppb/timestamp.pb.go create mode 100644 vendor/gopkg.in/yaml.v3/.travis.yml create mode 100644 vendor/gopkg.in/yaml.v3/LICENSE create mode 100644 vendor/gopkg.in/yaml.v3/NOTICE create mode 100644 vendor/gopkg.in/yaml.v3/README.md create mode 100644 vendor/gopkg.in/yaml.v3/apic.go create mode 100644 vendor/gopkg.in/yaml.v3/decode.go create mode 100644 vendor/gopkg.in/yaml.v3/emitterc.go create mode 100644 vendor/gopkg.in/yaml.v3/encode.go create mode 100644 vendor/gopkg.in/yaml.v3/go.mod create mode 100644 vendor/gopkg.in/yaml.v3/parserc.go create mode 100644 vendor/gopkg.in/yaml.v3/readerc.go create mode 100644 vendor/gopkg.in/yaml.v3/resolve.go create mode 100644 vendor/gopkg.in/yaml.v3/scannerc.go create mode 100644 vendor/gopkg.in/yaml.v3/sorter.go create mode 100644 vendor/gopkg.in/yaml.v3/writerc.go create mode 100644 vendor/gopkg.in/yaml.v3/yaml.go create mode 100644 vendor/gopkg.in/yaml.v3/yamlh.go create mode 100644 vendor/gopkg.in/yaml.v3/yamlprivateh.go create mode 100644 vendor/k8s.io/api/admissionregistration/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/apps/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/apps/v1beta2/zz_generated.prerelease-lifecycle.go delete mode 100644 vendor/k8s.io/api/auditregistration/v1alpha1/generated.pb.go delete mode 100644 vendor/k8s.io/api/auditregistration/v1alpha1/generated.proto delete mode 100644 vendor/k8s.io/api/auditregistration/v1alpha1/types.go delete mode 100644 vendor/k8s.io/api/auditregistration/v1alpha1/types_swagger_doc_generated.go delete mode 100644 vendor/k8s.io/api/auditregistration/v1alpha1/zz_generated.deepcopy.go create mode 100644 vendor/k8s.io/api/authentication/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/authorization/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/autoscaling/v2beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/batch/v1beta1/zz_generated.prerelease-lifecycle.go rename vendor/k8s.io/api/{auditregistration/v1alpha1 => certificates/v1}/doc.go (80%) create mode 100644 vendor/k8s.io/api/certificates/v1/generated.pb.go create mode 100644 vendor/k8s.io/api/certificates/v1/generated.proto create mode 100644 vendor/k8s.io/api/certificates/v1/register.go create mode 100644 vendor/k8s.io/api/certificates/v1/types.go create mode 100644 vendor/k8s.io/api/certificates/v1/types_swagger_doc_generated.go create mode 100644 vendor/k8s.io/api/certificates/v1/zz_generated.deepcopy.go create mode 100644 vendor/k8s.io/api/certificates/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/coordination/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/core/v1/lifecycle.go create mode 100644 vendor/k8s.io/api/discovery/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/events/v1/doc.go create mode 100644 vendor/k8s.io/api/events/v1/generated.pb.go create mode 100644 vendor/k8s.io/api/events/v1/generated.proto rename vendor/k8s.io/api/{auditregistration/v1alpha1 => events/v1}/register.go (73%) create mode 100644 vendor/k8s.io/api/events/v1/types.go create mode 100644 vendor/k8s.io/api/events/v1/types_swagger_doc_generated.go create mode 100644 vendor/k8s.io/api/events/v1/zz_generated.deepcopy.go create mode 100644 vendor/k8s.io/api/events/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/extensions/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/networking/v1beta1/well_known_annotations.go create mode 100644 vendor/k8s.io/api/networking/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/node/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/policy/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/rbac/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/scheduling/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/api/storage/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/.import-restrictions create mode 100644 vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.prerelease-lifecycle.go create mode 100644 vendor/k8s.io/apimachinery/pkg/api/meta/conditions.go delete mode 100644 vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount_expansion.go create mode 100644 vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/metrics.go delete mode 100644 vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack.go create mode 100644 vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack_stub.go create mode 100644 vendor/k8s.io/client-go/rest/warnings.go create mode 100644 vendor/k8s.io/client-go/transport/cert_rotation.go create mode 100644 vendor/k8s.io/client-go/util/workqueue/default_rate_limiters.go create mode 100644 vendor/k8s.io/client-go/util/workqueue/delaying_queue.go create mode 100644 vendor/k8s.io/client-go/util/workqueue/doc.go create mode 100644 vendor/k8s.io/client-go/util/workqueue/metrics.go create mode 100644 vendor/k8s.io/client-go/util/workqueue/parallelizer.go create mode 100644 vendor/k8s.io/client-go/util/workqueue/queue.go create mode 100644 vendor/k8s.io/client-go/util/workqueue/rate_limiting_queue.go create mode 100644 vendor/k8s.io/klog/v2/.gitignore create mode 100644 vendor/k8s.io/klog/v2/CONTRIBUTING.md create mode 100644 vendor/k8s.io/klog/v2/LICENSE create mode 100644 vendor/k8s.io/klog/v2/OWNERS create mode 100644 vendor/k8s.io/klog/v2/README.md create mode 100644 vendor/k8s.io/klog/v2/RELEASE.md create mode 100644 vendor/k8s.io/klog/v2/SECURITY_CONTACTS create mode 100644 vendor/k8s.io/klog/v2/code-of-conduct.md create mode 100644 vendor/k8s.io/klog/v2/go.mod create mode 100644 vendor/k8s.io/klog/v2/go.sum create mode 100644 vendor/k8s.io/klog/v2/klog.go create mode 100644 vendor/k8s.io/klog/v2/klog_file.go rename vendor/{github.com/gophercloud/gophercloud => k8s.io/kubectl}/LICENSE (89%) create mode 100644 vendor/k8s.io/kubectl/pkg/util/openapi/OWNERS rename vendor/k8s.io/{kubernetes/pkg/kubectl/cmd => kubectl/pkg}/util/openapi/doc.go (92%) rename vendor/k8s.io/{kubernetes/pkg/kubectl/cmd => kubectl/pkg}/util/openapi/extensions.go (92%) rename vendor/k8s.io/{kubernetes/pkg/kubectl/cmd => kubectl/pkg}/util/openapi/openapi.go (96%) rename vendor/k8s.io/{kubernetes/pkg/kubectl/cmd => kubectl/pkg}/util/openapi/openapi_getter.go (100%) delete mode 100644 vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/BUILD delete mode 100644 vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/OWNERS delete mode 100644 vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/dryrun.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/LICENSE create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/allocator.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/doc.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/fields.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/jsontagutil.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/list.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/listreflect.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/listunstructured.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/map.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapreflect.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapunstructured.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/reflectcache.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/scalar.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/structreflect.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/value.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/valuereflect.go create mode 100644 vendor/sigs.k8s.io/structured-merge-diff/v4/value/valueunstructured.go diff --git a/vendor/cloud.google.com/go/compute/metadata/.repo-metadata.json b/vendor/cloud.google.com/go/compute/metadata/.repo-metadata.json new file mode 100644 index 00000000..ca022ccc --- /dev/null +++ b/vendor/cloud.google.com/go/compute/metadata/.repo-metadata.json @@ -0,0 +1,12 @@ +{ + "name": "metadata", + "name_pretty": "Google Compute Engine Metadata API", + "product_documentation": "https://cloud.google.com/compute/docs/storing-retrieving-metadata", + "client_documentation": "https://godoc.org/cloud.google.com/go/compute/metadata", + "release_level": "ga", + "language": "go", + "repo": "googleapis/google-cloud-go", + "distribution_name": "cloud.google.com/go/compute/metadata", + "api_id": "compute:metadata", + "requires_billing": false +} diff --git a/vendor/cloud.google.com/go/compute/metadata/metadata.go b/vendor/cloud.google.com/go/compute/metadata/metadata.go index 125b7033..0b50c2a7 100644 --- a/vendor/cloud.google.com/go/compute/metadata/metadata.go +++ b/vendor/cloud.google.com/go/compute/metadata/metadata.go @@ -227,6 +227,9 @@ func InternalIP() (string, error) { return defaultClient.InternalIP() } // ExternalIP returns the instance's primary external (public) IP address. func ExternalIP() (string, error) { return defaultClient.ExternalIP() } +// Email calls Client.Email on the default client. +func Email(serviceAccount string) (string, error) { return defaultClient.Email(serviceAccount) } + // Hostname returns the instance's hostname. This will be of the form // ".c..internal". func Hostname() (string, error) { return defaultClient.Hostname() } @@ -301,7 +304,10 @@ func (c *Client) getETag(suffix string) (value, etag string, err error) { host = metadataIP } u := "http://" + host + "/computeMetadata/v1/" + suffix - req, _ := http.NewRequest("GET", u, nil) + req, err := http.NewRequest("GET", u, nil) + if err != nil { + return "", "", err + } req.Header.Set("Metadata-Flavor", "Google") req.Header.Set("User-Agent", userAgent) res, err := c.hc.Do(req) @@ -367,6 +373,16 @@ func (c *Client) InternalIP() (string, error) { return c.getTrimmed("instance/network-interfaces/0/ip") } +// Email returns the email address associated with the service account. +// The account may be empty or the string "default" to use the instance's +// main account. +func (c *Client) Email(serviceAccount string) (string, error) { + if serviceAccount == "" { + serviceAccount = "default" + } + return c.getTrimmed("instance/service-accounts/" + serviceAccount + "/email") +} + // ExternalIP returns the instance's primary external (public) IP address. func (c *Client) ExternalIP() (string, error) { return c.getTrimmed("instance/network-interfaces/0/access-configs/0/external-ip") @@ -394,11 +410,7 @@ func (c *Client) InstanceTags() ([]string, error) { // InstanceName returns the current VM's instance ID string. func (c *Client) InstanceName() (string, error) { - host, err := c.Hostname() - if err != nil { - return "", err - } - return strings.Split(host, ".")[0], nil + return c.getTrimmed("instance/name") } // Zone returns the current VM's zone, such as "us-central1-b". diff --git a/vendor/github.com/evanphx/json-patch/.travis.yml b/vendor/github.com/evanphx/json-patch/.travis.yml index 2092c72c..50e4afd1 100644 --- a/vendor/github.com/evanphx/json-patch/.travis.yml +++ b/vendor/github.com/evanphx/json-patch/.travis.yml @@ -1,8 +1,8 @@ language: go go: - - 1.8 - - 1.7 + - 1.14 + - 1.13 install: - if ! go get code.google.com/p/go.tools/cmd/cover; then go get golang.org/x/tools/cmd/cover; fi @@ -11,6 +11,9 @@ install: script: - go get - go test -cover ./... + - cd ./v5 + - go get + - go test -cover ./... notifications: email: false diff --git a/vendor/github.com/evanphx/json-patch/LICENSE b/vendor/github.com/evanphx/json-patch/LICENSE index 0eb9b72d..df76d7d7 100644 --- a/vendor/github.com/evanphx/json-patch/LICENSE +++ b/vendor/github.com/evanphx/json-patch/LICENSE @@ -6,7 +6,7 @@ modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. -* Redistributions in binary form must reproduce the above copyright notice +* Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the Evan Phoenix nor the names of its contributors diff --git a/vendor/github.com/evanphx/json-patch/README.md b/vendor/github.com/evanphx/json-patch/README.md index 9c7f87f7..121b039d 100644 --- a/vendor/github.com/evanphx/json-patch/README.md +++ b/vendor/github.com/evanphx/json-patch/README.md @@ -1,5 +1,5 @@ # JSON-Patch -`jsonpatch` is a library which provides functionallity for both applying +`jsonpatch` is a library which provides functionality for both applying [RFC6902 JSON patches](http://tools.ietf.org/html/rfc6902) against documents, as well as for calculating & applying [RFC7396 JSON merge patches](https://tools.ietf.org/html/rfc7396). @@ -11,10 +11,11 @@ well as for calculating & applying [RFC7396 JSON merge patches](https://tools.ie **Latest and greatest**: ```bash -go get -u github.com/evanphx/json-patch +go get -u github.com/evanphx/json-patch/v5 ``` **Stable Versions**: +* Version 5: `go get -u gopkg.in/evanphx/json-patch.v5` * Version 4: `go get -u gopkg.in/evanphx/json-patch.v4` (previous versions below `v3` are unavailable) @@ -82,7 +83,7 @@ When ran, you get the following output: ```bash $ go run main.go patch document: {"height":null,"name":"Jane"} -updated tina doc: {"age":28,"name":"Jane"} +updated alternative doc: {"age":28,"name":"Jane"} ``` ## Create and apply a JSON Patch @@ -164,7 +165,7 @@ func main() { } if !jsonpatch.Equal(original, different) { - fmt.Println(`"original" is _not_ structurally equal to "similar"`) + fmt.Println(`"original" is _not_ structurally equal to "different"`) } } ``` @@ -173,7 +174,7 @@ When ran, you get the following output: ```bash $ go run main.go "original" is structurally equal to "similar" -"original" is _not_ structurally equal to "similar" +"original" is _not_ structurally equal to "different" ``` ## Combine merge patches diff --git a/vendor/github.com/evanphx/json-patch/merge.go b/vendor/github.com/evanphx/json-patch/merge.go index 6806c4c2..14e8bb5c 100644 --- a/vendor/github.com/evanphx/json-patch/merge.go +++ b/vendor/github.com/evanphx/json-patch/merge.go @@ -307,13 +307,16 @@ func matchesValue(av, bv interface{}) bool { return true case map[string]interface{}: bt := bv.(map[string]interface{}) - for key := range at { - if !matchesValue(at[key], bt[key]) { - return false - } + if len(bt) != len(at) { + return false } for key := range bt { - if !matchesValue(at[key], bt[key]) { + av, aOK := at[key] + bv, bOK := bt[key] + if aOK != bOK { + return false + } + if !matchesValue(av, bv) { return false } } diff --git a/vendor/github.com/evanphx/json-patch/patch.go b/vendor/github.com/evanphx/json-patch/patch.go index 1b5f95e6..f185a45b 100644 --- a/vendor/github.com/evanphx/json-patch/patch.go +++ b/vendor/github.com/evanphx/json-patch/patch.go @@ -202,6 +202,10 @@ func (n *lazyNode) equal(o *lazyNode) bool { return false } + if len(n.doc) != len(o.doc) { + return false + } + for k, v := range n.doc { ov, ok := o.doc[k] @@ -209,6 +213,10 @@ func (n *lazyNode) equal(o *lazyNode) bool { return false } + if (v == nil) != (ov == nil) { + return false + } + if v == nil && ov == nil { continue } @@ -429,14 +437,14 @@ func (d *partialArray) add(key string, val *lazyNode) error { return errors.Wrapf(ErrInvalidIndex, "Unable to access invalid index: %d", idx) } - if SupportNegativeIndices { - if idx < -len(ary) { + if idx < 0 { + if !SupportNegativeIndices { return errors.Wrapf(ErrInvalidIndex, "Unable to access invalid index: %d", idx) } - - if idx < 0 { - idx += len(ary) + if idx < -len(ary) { + return errors.Wrapf(ErrInvalidIndex, "Unable to access invalid index: %d", idx) } + idx += len(ary) } copy(ary[0:idx], cur[0:idx]) @@ -473,14 +481,14 @@ func (d *partialArray) remove(key string) error { return errors.Wrapf(ErrInvalidIndex, "Unable to access invalid index: %d", idx) } - if SupportNegativeIndices { - if idx < -len(cur) { + if idx < 0 { + if !SupportNegativeIndices { return errors.Wrapf(ErrInvalidIndex, "Unable to access invalid index: %d", idx) } - - if idx < 0 { - idx += len(cur) + if idx < -len(cur) { + return errors.Wrapf(ErrInvalidIndex, "Unable to access invalid index: %d", idx) } + idx += len(cur) } ary := make([]*lazyNode, len(cur)-1) diff --git a/vendor/k8s.io/kubernetes/LICENSE b/vendor/github.com/go-logr/logr/LICENSE similarity index 99% rename from vendor/k8s.io/kubernetes/LICENSE rename to vendor/github.com/go-logr/logr/LICENSE index d6456956..8dada3ed 100644 --- a/vendor/k8s.io/kubernetes/LICENSE +++ b/vendor/github.com/go-logr/logr/LICENSE @@ -1,4 +1,3 @@ - Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ @@ -179,7 +178,7 @@ APPENDIX: How to apply the Apache License to your work. To apply the Apache License to your work, attach the following - boilerplate notice, with the fields enclosed by brackets "[]" + boilerplate notice, with the fields enclosed by brackets "{}" replaced with your own identifying information. (Don't include the brackets!) The text should be enclosed in the appropriate comment syntax for the file format. We also recommend that a @@ -187,7 +186,7 @@ same "printed page" as the copyright notice for easier identification within third-party archives. - Copyright [yyyy] [name of copyright owner] + Copyright {yyyy} {name of copyright owner} Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. diff --git a/vendor/github.com/go-logr/logr/README.md b/vendor/github.com/go-logr/logr/README.md new file mode 100644 index 00000000..aca17f38 --- /dev/null +++ b/vendor/github.com/go-logr/logr/README.md @@ -0,0 +1,181 @@ +# A more minimal logging API for Go + +Before you consider this package, please read [this blog post by the +inimitable Dave Cheney][warning-makes-no-sense]. I really appreciate what +he has to say, and it largely aligns with my own experiences. Too many +choices of levels means inconsistent logs. + +This package offers a purely abstract interface, based on these ideas but with +a few twists. Code can depend on just this interface and have the actual +logging implementation be injected from callers. Ideally only `main()` knows +what logging implementation is being used. + +# Differences from Dave's ideas + +The main differences are: + +1) Dave basically proposes doing away with the notion of a logging API in favor +of `fmt.Printf()`. I disagree, especially when you consider things like output +locations, timestamps, file and line decorations, and structured logging. I +restrict the API to just 2 types of logs: info and error. + +Info logs are things you want to tell the user which are not errors. Error +logs are, well, errors. If your code receives an `error` from a subordinate +function call and is logging that `error` *and not returning it*, use error +logs. + +2) Verbosity-levels on info logs. This gives developers a chance to indicate +arbitrary grades of importance for info logs, without assigning names with +semantic meaning such as "warning", "trace", and "debug". Superficially this +may feel very similar, but the primary difference is the lack of semantics. +Because verbosity is a numerical value, it's safe to assume that an app running +with higher verbosity means more (and less important) logs will be generated. + +This is a BETA grade API. + +There are implementations for the following logging libraries: + +- **github.com/google/glog**: [glogr](https://github.com/go-logr/glogr) +- **k8s.io/klog**: [klogr](https://git.k8s.io/klog/klogr) +- **go.uber.org/zap**: [zapr](https://github.com/go-logr/zapr) +- **log** (the Go standard library logger): + [stdr](https://github.com/go-logr/stdr) +- **github.com/sirupsen/logrus**: [logrusr](https://github.com/bombsimon/logrusr) + +# FAQ + +## Conceptual + +## Why structured logging? + +- **Structured logs are more easily queriable**: Since you've got + key-value pairs, it's much easier to query your structured logs for + particular values by filtering on the contents of a particular key -- + think searching request logs for error codes, Kubernetes reconcilers for + the name and namespace of the reconciled object, etc + +- **Structured logging makes it easier to have cross-referencable logs**: + Similarly to searchability, if you maintain conventions around your + keys, it becomes easy to gather all log lines related to a particular + concept. + +- **Structured logs allow better dimensions of filtering**: if you have + structure to your logs, you've got more precise control over how much + information is logged -- you might choose in a particular configuration + to log certain keys but not others, only log lines where a certain key + matches a certain value, etc, instead of just having v-levels and names + to key off of. + +- **Structured logs better represent structured data**: sometimes, the + data that you want to log is inherently structured (think tuple-link + objects). Structured logs allow you to preserve that structure when + outputting. + +## Why V-levels? + +**V-levels give operators an easy way to control the chattiness of log +operations**. V-levels provide a way for a given package to distinguish +the relative importance or verbosity of a given log message. Then, if +a particular logger or package is logging too many messages, the user +of the package can simply change the v-levels for that library. + +## Why not more named levels, like Warning? + +Read [Dave Cheney's post][warning-makes-no-sense]. Then read [Differences +from Dave's ideas](#differences-from-daves-ideas). + +## Why not allow format strings, too? + +**Format strings negate many of the benefits of structured logs**: + +- They're not easily searchable without resorting to fuzzy searching, + regular expressions, etc + +- They don't store structured data well, since contents are flattened into + a string + +- They're not cross-referencable + +- They don't compress easily, since the message is not constant + +(unless you turn positional parameters into key-value pairs with numerical +keys, at which point you've gotten key-value logging with meaningless +keys) + +## Practical + +## Why key-value pairs, and not a map? + +Key-value pairs are *much* easier to optimize, especially around +allocations. Zap (a structured logger that inspired logr's interface) has +[performance measurements](https://github.com/uber-go/zap#performance) +that show this quite nicely. + +While the interface ends up being a little less obvious, you get +potentially better performance, plus avoid making users type +`map[string]string{}` every time they want to log. + +## What if my V-levels differ between libraries? + +That's fine. Control your V-levels on a per-logger basis, and use the +`WithName` function to pass different loggers to different libraries. + +Generally, you should take care to ensure that you have relatively +consistent V-levels within a given logger, however, as this makes deciding +on what verbosity of logs to request easier. + +## But I *really* want to use a format string! + +That's not actually a question. Assuming your question is "how do +I convert my mental model of logging with format strings to logging with +constant messages": + +1. figure out what the error actually is, as you'd write in a TL;DR style, + and use that as a message + +2. For every place you'd write a format specifier, look to the word before + it, and add that as a key value pair + +For instance, consider the following examples (all taken from spots in the +Kubernetes codebase): + +- `klog.V(4).Infof("Client is returning errors: code %v, error %v", + responseCode, err)` becomes `logger.Error(err, "client returned an + error", "code", responseCode)` + +- `klog.V(4).Infof("Got a Retry-After %ds response for attempt %d to %v", + seconds, retries, url)` becomes `logger.V(4).Info("got a retry-after + response when requesting url", "attempt", retries, "after + seconds", seconds, "url", url)` + +If you *really* must use a format string, place it as a key value, and +call `fmt.Sprintf` yourself -- for instance, `log.Printf("unable to +reflect over type %T")` becomes `logger.Info("unable to reflect over +type", "type", fmt.Sprintf("%T"))`. In general though, the cases where +this is necessary should be few and far between. + +## How do I choose my V-levels? + +This is basically the only hard constraint: increase V-levels to denote +more verbose or more debug-y logs. + +Otherwise, you can start out with `0` as "you always want to see this", +`1` as "common logging that you might *possibly* want to turn off", and +`10` as "I would like to performance-test your log collection stack". + +Then gradually choose levels in between as you need them, working your way +down from 10 (for debug and trace style logs) and up from 1 (for chattier +info-type logs). + +## How do I choose my keys + +- make your keys human-readable +- constant keys are generally a good idea +- be consistent across your codebase +- keys should naturally match parts of the message string + +While key names are mostly unrestricted (and spaces are acceptable), +it's generally a good idea to stick to printable ascii characters, or at +least match the general character set of your log lines. + +[warning-makes-no-sense]: http://dave.cheney.net/2015/11/05/lets-talk-about-logging diff --git a/vendor/github.com/go-logr/logr/go.mod b/vendor/github.com/go-logr/logr/go.mod new file mode 100644 index 00000000..591884e9 --- /dev/null +++ b/vendor/github.com/go-logr/logr/go.mod @@ -0,0 +1,3 @@ +module github.com/go-logr/logr + +go 1.14 diff --git a/vendor/github.com/go-logr/logr/logr.go b/vendor/github.com/go-logr/logr/logr.go new file mode 100644 index 00000000..520c4fe5 --- /dev/null +++ b/vendor/github.com/go-logr/logr/logr.go @@ -0,0 +1,178 @@ +/* +Copyright 2019 The logr Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Package logr defines abstract interfaces for logging. Packages can depend on +// these interfaces and callers can implement logging in whatever way is +// appropriate. +// +// This design derives from Dave Cheney's blog: +// http://dave.cheney.net/2015/11/05/lets-talk-about-logging +// +// This is a BETA grade API. Until there is a significant 2nd implementation, +// I don't really know how it will change. +// +// The logging specifically makes it non-trivial to use format strings, to encourage +// attaching structured information instead of unstructured format strings. +// +// Usage +// +// Logging is done using a Logger. Loggers can have name prefixes and named +// values attached, so that all log messages logged with that Logger have some +// base context associated. +// +// The term "key" is used to refer to the name associated with a particular +// value, to disambiguate it from the general Logger name. +// +// For instance, suppose we're trying to reconcile the state of an object, and +// we want to log that we've made some decision. +// +// With the traditional log package, we might write: +// log.Printf( +// "decided to set field foo to value %q for object %s/%s", +// targetValue, object.Namespace, object.Name) +// +// With logr's structured logging, we'd write: +// // elsewhere in the file, set up the logger to log with the prefix of "reconcilers", +// // and the named value target-type=Foo, for extra context. +// log := mainLogger.WithName("reconcilers").WithValues("target-type", "Foo") +// +// // later on... +// log.Info("setting field foo on object", "value", targetValue, "object", object) +// +// Depending on our logging implementation, we could then make logging decisions +// based on field values (like only logging such events for objects in a certain +// namespace), or copy the structured information into a structured log store. +// +// For logging errors, Logger has a method called Error. Suppose we wanted to +// log an error while reconciling. With the traditional log package, we might +// write: +// log.Errorf("unable to reconcile object %s/%s: %v", object.Namespace, object.Name, err) +// +// With logr, we'd instead write: +// // assuming the above setup for log +// log.Error(err, "unable to reconcile object", "object", object) +// +// This functions similarly to: +// log.Info("unable to reconcile object", "error", err, "object", object) +// +// However, it ensures that a standard key for the error value ("error") is used +// across all error logging. Furthermore, certain implementations may choose to +// attach additional information (such as stack traces) on calls to Error, so +// it's preferred to use Error to log errors. +// +// Parts of a log line +// +// Each log message from a Logger has four types of context: +// logger name, log verbosity, log message, and the named values. +// +// The Logger name constists of a series of name "segments" added by successive +// calls to WithName. These name segments will be joined in some way by the +// underlying implementation. It is strongly reccomended that name segements +// contain simple identifiers (letters, digits, and hyphen), and do not contain +// characters that could muddle the log output or confuse the joining operation +// (e.g. whitespace, commas, periods, slashes, brackets, quotes, etc). +// +// Log verbosity represents how little a log matters. Level zero, the default, +// matters most. Increasing levels matter less and less. Try to avoid lots of +// different verbosity levels, and instead provide useful keys, logger names, +// and log messages for users to filter on. It's illegal to pass a log level +// below zero. +// +// The log message consists of a constant message attached to the the log line. +// This should generally be a simple description of what's occuring, and should +// never be a format string. +// +// Variable information can then be attached using named values (key/value +// pairs). Keys are arbitrary strings, while values may be any Go value. +// +// Key Naming Conventions +// +// Keys are not strictly required to conform to any specification or regex, but +// it is recommended that they: +// * be human-readable and meaningful (not auto-generated or simple ordinals) +// * be constant (not dependent on input data) +// * contain only printable characters +// * not contain whitespace or punctuation +// +// These guidelines help ensure that log data is processed properly regardless +// of the log implementation. For example, log implementations will try to +// output JSON data or will store data for later database (e.g. SQL) queries. +// +// While users are generally free to use key names of their choice, it's +// generally best to avoid using the following keys, as they're frequently used +// by implementations: +// +// - `"caller"`: the calling information (file/line) of a particular log line. +// - `"error"`: the underlying error value in the `Error` method. +// - `"level"`: the log level. +// - `"logger"`: the name of the associated logger. +// - `"msg"`: the log message. +// - `"stacktrace"`: the stack trace associated with a particular log line or +// error (often from the `Error` message). +// - `"ts"`: the timestamp for a log line. +// +// Implementations are encouraged to make use of these keys to represent the +// above concepts, when neccessary (for example, in a pure-JSON output form, it +// would be necessary to represent at least message and timestamp as ordinary +// named values). +package logr + +// TODO: consider adding back in format strings if they're really needed +// TODO: consider other bits of zap/zapcore functionality like ObjectMarshaller (for arbitrary objects) +// TODO: consider other bits of glog functionality like Flush, InfoDepth, OutputStats + +// Logger represents the ability to log messages, both errors and not. +type Logger interface { + // Enabled tests whether this Logger is enabled. For example, commandline + // flags might be used to set the logging verbosity and disable some info + // logs. + Enabled() bool + + // Info logs a non-error message with the given key/value pairs as context. + // + // The msg argument should be used to add some constant description to + // the log line. The key/value pairs can then be used to add additional + // variable information. The key/value pairs should alternate string + // keys and arbitrary values. + Info(msg string, keysAndValues ...interface{}) + + // Error logs an error, with the given message and key/value pairs as context. + // It functions similarly to calling Info with the "error" named value, but may + // have unique behavior, and should be preferred for logging errors (see the + // package documentations for more information). + // + // The msg field should be used to add context to any underlying error, + // while the err field should be used to attach the actual error that + // triggered this log line, if present. + Error(err error, msg string, keysAndValues ...interface{}) + + // V returns an Logger value for a specific verbosity level, relative to + // this Logger. In other words, V values are additive. V higher verbosity + // level means a log message is less important. It's illegal to pass a log + // level less than zero. + V(level int) Logger + + // WithValues adds some key-value pairs of context to a logger. + // See Info for documentation on how key/value pairs work. + WithValues(keysAndValues ...interface{}) Logger + + // WithName adds a new element to the logger's name. + // Successive calls with WithName continue to append + // suffixes to the logger's name. It's strongly reccomended + // that name segments contain only letters, digits, and hyphens + // (see the package documentation for more information). + WithName(name string) Logger +} diff --git a/vendor/github.com/golang/protobuf/proto/buffer.go b/vendor/github.com/golang/protobuf/proto/buffer.go new file mode 100644 index 00000000..e810e6fe --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/buffer.go @@ -0,0 +1,324 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "errors" + "fmt" + + "google.golang.org/protobuf/encoding/prototext" + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/runtime/protoimpl" +) + +const ( + WireVarint = 0 + WireFixed32 = 5 + WireFixed64 = 1 + WireBytes = 2 + WireStartGroup = 3 + WireEndGroup = 4 +) + +// EncodeVarint returns the varint encoded bytes of v. +func EncodeVarint(v uint64) []byte { + return protowire.AppendVarint(nil, v) +} + +// SizeVarint returns the length of the varint encoded bytes of v. +// This is equal to len(EncodeVarint(v)). +func SizeVarint(v uint64) int { + return protowire.SizeVarint(v) +} + +// DecodeVarint parses a varint encoded integer from b, +// returning the integer value and the length of the varint. +// It returns (0, 0) if there is a parse error. +func DecodeVarint(b []byte) (uint64, int) { + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, 0 + } + return v, n +} + +// Buffer is a buffer for encoding and decoding the protobuf wire format. +// It may be reused between invocations to reduce memory usage. +type Buffer struct { + buf []byte + idx int + deterministic bool +} + +// NewBuffer allocates a new Buffer initialized with buf, +// where the contents of buf are considered the unread portion of the buffer. +func NewBuffer(buf []byte) *Buffer { + return &Buffer{buf: buf} +} + +// SetDeterministic specifies whether to use deterministic serialization. +// +// Deterministic serialization guarantees that for a given binary, equal +// messages will always be serialized to the same bytes. This implies: +// +// - Repeated serialization of a message will return the same bytes. +// - Different processes of the same binary (which may be executing on +// different machines) will serialize equal messages to the same bytes. +// +// Note that the deterministic serialization is NOT canonical across +// languages. It is not guaranteed to remain stable over time. It is unstable +// across different builds with schema changes due to unknown fields. +// Users who need canonical serialization (e.g., persistent storage in a +// canonical form, fingerprinting, etc.) should define their own +// canonicalization specification and implement their own serializer rather +// than relying on this API. +// +// If deterministic serialization is requested, map entries will be sorted +// by keys in lexographical order. This is an implementation detail and +// subject to change. +func (b *Buffer) SetDeterministic(deterministic bool) { + b.deterministic = deterministic +} + +// SetBuf sets buf as the internal buffer, +// where the contents of buf are considered the unread portion of the buffer. +func (b *Buffer) SetBuf(buf []byte) { + b.buf = buf + b.idx = 0 +} + +// Reset clears the internal buffer of all written and unread data. +func (b *Buffer) Reset() { + b.buf = b.buf[:0] + b.idx = 0 +} + +// Bytes returns the internal buffer. +func (b *Buffer) Bytes() []byte { + return b.buf +} + +// Unread returns the unread portion of the buffer. +func (b *Buffer) Unread() []byte { + return b.buf[b.idx:] +} + +// Marshal appends the wire-format encoding of m to the buffer. +func (b *Buffer) Marshal(m Message) error { + var err error + b.buf, err = marshalAppend(b.buf, m, b.deterministic) + return err +} + +// Unmarshal parses the wire-format message in the buffer and +// places the decoded results in m. +// It does not reset m before unmarshaling. +func (b *Buffer) Unmarshal(m Message) error { + err := UnmarshalMerge(b.Unread(), m) + b.idx = len(b.buf) + return err +} + +type unknownFields struct{ XXX_unrecognized protoimpl.UnknownFields } + +func (m *unknownFields) String() string { panic("not implemented") } +func (m *unknownFields) Reset() { panic("not implemented") } +func (m *unknownFields) ProtoMessage() { panic("not implemented") } + +// DebugPrint dumps the encoded bytes of b with a header and footer including s +// to stdout. This is only intended for debugging. +func (*Buffer) DebugPrint(s string, b []byte) { + m := MessageReflect(new(unknownFields)) + m.SetUnknown(b) + b, _ = prototext.MarshalOptions{AllowPartial: true, Indent: "\t"}.Marshal(m.Interface()) + fmt.Printf("==== %s ====\n%s==== %s ====\n", s, b, s) +} + +// EncodeVarint appends an unsigned varint encoding to the buffer. +func (b *Buffer) EncodeVarint(v uint64) error { + b.buf = protowire.AppendVarint(b.buf, v) + return nil +} + +// EncodeZigzag32 appends a 32-bit zig-zag varint encoding to the buffer. +func (b *Buffer) EncodeZigzag32(v uint64) error { + return b.EncodeVarint(uint64((uint32(v) << 1) ^ uint32((int32(v) >> 31)))) +} + +// EncodeZigzag64 appends a 64-bit zig-zag varint encoding to the buffer. +func (b *Buffer) EncodeZigzag64(v uint64) error { + return b.EncodeVarint(uint64((uint64(v) << 1) ^ uint64((int64(v) >> 63)))) +} + +// EncodeFixed32 appends a 32-bit little-endian integer to the buffer. +func (b *Buffer) EncodeFixed32(v uint64) error { + b.buf = protowire.AppendFixed32(b.buf, uint32(v)) + return nil +} + +// EncodeFixed64 appends a 64-bit little-endian integer to the buffer. +func (b *Buffer) EncodeFixed64(v uint64) error { + b.buf = protowire.AppendFixed64(b.buf, uint64(v)) + return nil +} + +// EncodeRawBytes appends a length-prefixed raw bytes to the buffer. +func (b *Buffer) EncodeRawBytes(v []byte) error { + b.buf = protowire.AppendBytes(b.buf, v) + return nil +} + +// EncodeStringBytes appends a length-prefixed raw bytes to the buffer. +// It does not validate whether v contains valid UTF-8. +func (b *Buffer) EncodeStringBytes(v string) error { + b.buf = protowire.AppendString(b.buf, v) + return nil +} + +// EncodeMessage appends a length-prefixed encoded message to the buffer. +func (b *Buffer) EncodeMessage(m Message) error { + var err error + b.buf = protowire.AppendVarint(b.buf, uint64(Size(m))) + b.buf, err = marshalAppend(b.buf, m, b.deterministic) + return err +} + +// DecodeVarint consumes an encoded unsigned varint from the buffer. +func (b *Buffer) DecodeVarint() (uint64, error) { + v, n := protowire.ConsumeVarint(b.buf[b.idx:]) + if n < 0 { + return 0, protowire.ParseError(n) + } + b.idx += n + return uint64(v), nil +} + +// DecodeZigzag32 consumes an encoded 32-bit zig-zag varint from the buffer. +func (b *Buffer) DecodeZigzag32() (uint64, error) { + v, err := b.DecodeVarint() + if err != nil { + return 0, err + } + return uint64((uint32(v) >> 1) ^ uint32((int32(v&1)<<31)>>31)), nil +} + +// DecodeZigzag64 consumes an encoded 64-bit zig-zag varint from the buffer. +func (b *Buffer) DecodeZigzag64() (uint64, error) { + v, err := b.DecodeVarint() + if err != nil { + return 0, err + } + return uint64((uint64(v) >> 1) ^ uint64((int64(v&1)<<63)>>63)), nil +} + +// DecodeFixed32 consumes a 32-bit little-endian integer from the buffer. +func (b *Buffer) DecodeFixed32() (uint64, error) { + v, n := protowire.ConsumeFixed32(b.buf[b.idx:]) + if n < 0 { + return 0, protowire.ParseError(n) + } + b.idx += n + return uint64(v), nil +} + +// DecodeFixed64 consumes a 64-bit little-endian integer from the buffer. +func (b *Buffer) DecodeFixed64() (uint64, error) { + v, n := protowire.ConsumeFixed64(b.buf[b.idx:]) + if n < 0 { + return 0, protowire.ParseError(n) + } + b.idx += n + return uint64(v), nil +} + +// DecodeRawBytes consumes a length-prefixed raw bytes from the buffer. +// If alloc is specified, it returns a copy the raw bytes +// rather than a sub-slice of the buffer. +func (b *Buffer) DecodeRawBytes(alloc bool) ([]byte, error) { + v, n := protowire.ConsumeBytes(b.buf[b.idx:]) + if n < 0 { + return nil, protowire.ParseError(n) + } + b.idx += n + if alloc { + v = append([]byte(nil), v...) + } + return v, nil +} + +// DecodeStringBytes consumes a length-prefixed raw bytes from the buffer. +// It does not validate whether the raw bytes contain valid UTF-8. +func (b *Buffer) DecodeStringBytes() (string, error) { + v, n := protowire.ConsumeString(b.buf[b.idx:]) + if n < 0 { + return "", protowire.ParseError(n) + } + b.idx += n + return v, nil +} + +// DecodeMessage consumes a length-prefixed message from the buffer. +// It does not reset m before unmarshaling. +func (b *Buffer) DecodeMessage(m Message) error { + v, err := b.DecodeRawBytes(false) + if err != nil { + return err + } + return UnmarshalMerge(v, m) +} + +// DecodeGroup consumes a message group from the buffer. +// It assumes that the start group marker has already been consumed and +// consumes all bytes until (and including the end group marker). +// It does not reset m before unmarshaling. +func (b *Buffer) DecodeGroup(m Message) error { + v, n, err := consumeGroup(b.buf[b.idx:]) + if err != nil { + return err + } + b.idx += n + return UnmarshalMerge(v, m) +} + +// consumeGroup parses b until it finds an end group marker, returning +// the raw bytes of the message (excluding the end group marker) and the +// the total length of the message (including the end group marker). +func consumeGroup(b []byte) ([]byte, int, error) { + b0 := b + depth := 1 // assume this follows a start group marker + for { + _, wtyp, tagLen := protowire.ConsumeTag(b) + if tagLen < 0 { + return nil, 0, protowire.ParseError(tagLen) + } + b = b[tagLen:] + + var valLen int + switch wtyp { + case protowire.VarintType: + _, valLen = protowire.ConsumeVarint(b) + case protowire.Fixed32Type: + _, valLen = protowire.ConsumeFixed32(b) + case protowire.Fixed64Type: + _, valLen = protowire.ConsumeFixed64(b) + case protowire.BytesType: + _, valLen = protowire.ConsumeBytes(b) + case protowire.StartGroupType: + depth++ + case protowire.EndGroupType: + depth-- + default: + return nil, 0, errors.New("proto: cannot parse reserved wire type") + } + if valLen < 0 { + return nil, 0, protowire.ParseError(valLen) + } + b = b[valLen:] + + if depth == 0 { + return b0[:len(b0)-len(b)-tagLen], len(b0) - len(b), nil + } + } +} diff --git a/vendor/github.com/golang/protobuf/proto/clone.go b/vendor/github.com/golang/protobuf/proto/clone.go deleted file mode 100644 index 3cd3249f..00000000 --- a/vendor/github.com/golang/protobuf/proto/clone.go +++ /dev/null @@ -1,253 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2011 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -// Protocol buffer deep copy and merge. -// TODO: RawMessage. - -package proto - -import ( - "fmt" - "log" - "reflect" - "strings" -) - -// Clone returns a deep copy of a protocol buffer. -func Clone(src Message) Message { - in := reflect.ValueOf(src) - if in.IsNil() { - return src - } - out := reflect.New(in.Type().Elem()) - dst := out.Interface().(Message) - Merge(dst, src) - return dst -} - -// Merger is the interface representing objects that can merge messages of the same type. -type Merger interface { - // Merge merges src into this message. - // Required and optional fields that are set in src will be set to that value in dst. - // Elements of repeated fields will be appended. - // - // Merge may panic if called with a different argument type than the receiver. - Merge(src Message) -} - -// generatedMerger is the custom merge method that generated protos will have. -// We must add this method since a generate Merge method will conflict with -// many existing protos that have a Merge data field already defined. -type generatedMerger interface { - XXX_Merge(src Message) -} - -// Merge merges src into dst. -// Required and optional fields that are set in src will be set to that value in dst. -// Elements of repeated fields will be appended. -// Merge panics if src and dst are not the same type, or if dst is nil. -func Merge(dst, src Message) { - if m, ok := dst.(Merger); ok { - m.Merge(src) - return - } - - in := reflect.ValueOf(src) - out := reflect.ValueOf(dst) - if out.IsNil() { - panic("proto: nil destination") - } - if in.Type() != out.Type() { - panic(fmt.Sprintf("proto.Merge(%T, %T) type mismatch", dst, src)) - } - if in.IsNil() { - return // Merge from nil src is a noop - } - if m, ok := dst.(generatedMerger); ok { - m.XXX_Merge(src) - return - } - mergeStruct(out.Elem(), in.Elem()) -} - -func mergeStruct(out, in reflect.Value) { - sprop := GetProperties(in.Type()) - for i := 0; i < in.NumField(); i++ { - f := in.Type().Field(i) - if strings.HasPrefix(f.Name, "XXX_") { - continue - } - mergeAny(out.Field(i), in.Field(i), false, sprop.Prop[i]) - } - - if emIn, err := extendable(in.Addr().Interface()); err == nil { - emOut, _ := extendable(out.Addr().Interface()) - mIn, muIn := emIn.extensionsRead() - if mIn != nil { - mOut := emOut.extensionsWrite() - muIn.Lock() - mergeExtension(mOut, mIn) - muIn.Unlock() - } - } - - uf := in.FieldByName("XXX_unrecognized") - if !uf.IsValid() { - return - } - uin := uf.Bytes() - if len(uin) > 0 { - out.FieldByName("XXX_unrecognized").SetBytes(append([]byte(nil), uin...)) - } -} - -// mergeAny performs a merge between two values of the same type. -// viaPtr indicates whether the values were indirected through a pointer (implying proto2). -// prop is set if this is a struct field (it may be nil). -func mergeAny(out, in reflect.Value, viaPtr bool, prop *Properties) { - if in.Type() == protoMessageType { - if !in.IsNil() { - if out.IsNil() { - out.Set(reflect.ValueOf(Clone(in.Interface().(Message)))) - } else { - Merge(out.Interface().(Message), in.Interface().(Message)) - } - } - return - } - switch in.Kind() { - case reflect.Bool, reflect.Float32, reflect.Float64, reflect.Int32, reflect.Int64, - reflect.String, reflect.Uint32, reflect.Uint64: - if !viaPtr && isProto3Zero(in) { - return - } - out.Set(in) - case reflect.Interface: - // Probably a oneof field; copy non-nil values. - if in.IsNil() { - return - } - // Allocate destination if it is not set, or set to a different type. - // Otherwise we will merge as normal. - if out.IsNil() || out.Elem().Type() != in.Elem().Type() { - out.Set(reflect.New(in.Elem().Elem().Type())) // interface -> *T -> T -> new(T) - } - mergeAny(out.Elem(), in.Elem(), false, nil) - case reflect.Map: - if in.Len() == 0 { - return - } - if out.IsNil() { - out.Set(reflect.MakeMap(in.Type())) - } - // For maps with value types of *T or []byte we need to deep copy each value. - elemKind := in.Type().Elem().Kind() - for _, key := range in.MapKeys() { - var val reflect.Value - switch elemKind { - case reflect.Ptr: - val = reflect.New(in.Type().Elem().Elem()) - mergeAny(val, in.MapIndex(key), false, nil) - case reflect.Slice: - val = in.MapIndex(key) - val = reflect.ValueOf(append([]byte{}, val.Bytes()...)) - default: - val = in.MapIndex(key) - } - out.SetMapIndex(key, val) - } - case reflect.Ptr: - if in.IsNil() { - return - } - if out.IsNil() { - out.Set(reflect.New(in.Elem().Type())) - } - mergeAny(out.Elem(), in.Elem(), true, nil) - case reflect.Slice: - if in.IsNil() { - return - } - if in.Type().Elem().Kind() == reflect.Uint8 { - // []byte is a scalar bytes field, not a repeated field. - - // Edge case: if this is in a proto3 message, a zero length - // bytes field is considered the zero value, and should not - // be merged. - if prop != nil && prop.proto3 && in.Len() == 0 { - return - } - - // Make a deep copy. - // Append to []byte{} instead of []byte(nil) so that we never end up - // with a nil result. - out.SetBytes(append([]byte{}, in.Bytes()...)) - return - } - n := in.Len() - if out.IsNil() { - out.Set(reflect.MakeSlice(in.Type(), 0, n)) - } - switch in.Type().Elem().Kind() { - case reflect.Bool, reflect.Float32, reflect.Float64, reflect.Int32, reflect.Int64, - reflect.String, reflect.Uint32, reflect.Uint64: - out.Set(reflect.AppendSlice(out, in)) - default: - for i := 0; i < n; i++ { - x := reflect.Indirect(reflect.New(in.Type().Elem())) - mergeAny(x, in.Index(i), false, nil) - out.Set(reflect.Append(out, x)) - } - } - case reflect.Struct: - mergeStruct(out, in) - default: - // unknown type, so not a protocol buffer - log.Printf("proto: don't know how to copy %v", in) - } -} - -func mergeExtension(out, in map[int32]Extension) { - for extNum, eIn := range in { - eOut := Extension{desc: eIn.desc} - if eIn.value != nil { - v := reflect.New(reflect.TypeOf(eIn.value)).Elem() - mergeAny(v, reflect.ValueOf(eIn.value), false, nil) - eOut.value = v.Interface() - } - if eIn.enc != nil { - eOut.enc = make([]byte, len(eIn.enc)) - copy(eOut.enc, eIn.enc) - } - - out[extNum] = eOut - } -} diff --git a/vendor/github.com/golang/protobuf/proto/decode.go b/vendor/github.com/golang/protobuf/proto/decode.go deleted file mode 100644 index 63b0f08b..00000000 --- a/vendor/github.com/golang/protobuf/proto/decode.go +++ /dev/null @@ -1,427 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -/* - * Routines for decoding protocol buffer data to construct in-memory representations. - */ - -import ( - "errors" - "fmt" - "io" -) - -// errOverflow is returned when an integer is too large to be represented. -var errOverflow = errors.New("proto: integer overflow") - -// ErrInternalBadWireType is returned by generated code when an incorrect -// wire type is encountered. It does not get returned to user code. -var ErrInternalBadWireType = errors.New("proto: internal error: bad wiretype for oneof") - -// DecodeVarint reads a varint-encoded integer from the slice. -// It returns the integer and the number of bytes consumed, or -// zero if there is not enough. -// This is the format for the -// int32, int64, uint32, uint64, bool, and enum -// protocol buffer types. -func DecodeVarint(buf []byte) (x uint64, n int) { - for shift := uint(0); shift < 64; shift += 7 { - if n >= len(buf) { - return 0, 0 - } - b := uint64(buf[n]) - n++ - x |= (b & 0x7F) << shift - if (b & 0x80) == 0 { - return x, n - } - } - - // The number is too large to represent in a 64-bit value. - return 0, 0 -} - -func (p *Buffer) decodeVarintSlow() (x uint64, err error) { - i := p.index - l := len(p.buf) - - for shift := uint(0); shift < 64; shift += 7 { - if i >= l { - err = io.ErrUnexpectedEOF - return - } - b := p.buf[i] - i++ - x |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - p.index = i - return - } - } - - // The number is too large to represent in a 64-bit value. - err = errOverflow - return -} - -// DecodeVarint reads a varint-encoded integer from the Buffer. -// This is the format for the -// int32, int64, uint32, uint64, bool, and enum -// protocol buffer types. -func (p *Buffer) DecodeVarint() (x uint64, err error) { - i := p.index - buf := p.buf - - if i >= len(buf) { - return 0, io.ErrUnexpectedEOF - } else if buf[i] < 0x80 { - p.index++ - return uint64(buf[i]), nil - } else if len(buf)-i < 10 { - return p.decodeVarintSlow() - } - - var b uint64 - // we already checked the first byte - x = uint64(buf[i]) - 0x80 - i++ - - b = uint64(buf[i]) - i++ - x += b << 7 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 7 - - b = uint64(buf[i]) - i++ - x += b << 14 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 14 - - b = uint64(buf[i]) - i++ - x += b << 21 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 21 - - b = uint64(buf[i]) - i++ - x += b << 28 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 28 - - b = uint64(buf[i]) - i++ - x += b << 35 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 35 - - b = uint64(buf[i]) - i++ - x += b << 42 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 42 - - b = uint64(buf[i]) - i++ - x += b << 49 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 49 - - b = uint64(buf[i]) - i++ - x += b << 56 - if b&0x80 == 0 { - goto done - } - x -= 0x80 << 56 - - b = uint64(buf[i]) - i++ - x += b << 63 - if b&0x80 == 0 { - goto done - } - - return 0, errOverflow - -done: - p.index = i - return x, nil -} - -// DecodeFixed64 reads a 64-bit integer from the Buffer. -// This is the format for the -// fixed64, sfixed64, and double protocol buffer types. -func (p *Buffer) DecodeFixed64() (x uint64, err error) { - // x, err already 0 - i := p.index + 8 - if i < 0 || i > len(p.buf) { - err = io.ErrUnexpectedEOF - return - } - p.index = i - - x = uint64(p.buf[i-8]) - x |= uint64(p.buf[i-7]) << 8 - x |= uint64(p.buf[i-6]) << 16 - x |= uint64(p.buf[i-5]) << 24 - x |= uint64(p.buf[i-4]) << 32 - x |= uint64(p.buf[i-3]) << 40 - x |= uint64(p.buf[i-2]) << 48 - x |= uint64(p.buf[i-1]) << 56 - return -} - -// DecodeFixed32 reads a 32-bit integer from the Buffer. -// This is the format for the -// fixed32, sfixed32, and float protocol buffer types. -func (p *Buffer) DecodeFixed32() (x uint64, err error) { - // x, err already 0 - i := p.index + 4 - if i < 0 || i > len(p.buf) { - err = io.ErrUnexpectedEOF - return - } - p.index = i - - x = uint64(p.buf[i-4]) - x |= uint64(p.buf[i-3]) << 8 - x |= uint64(p.buf[i-2]) << 16 - x |= uint64(p.buf[i-1]) << 24 - return -} - -// DecodeZigzag64 reads a zigzag-encoded 64-bit integer -// from the Buffer. -// This is the format used for the sint64 protocol buffer type. -func (p *Buffer) DecodeZigzag64() (x uint64, err error) { - x, err = p.DecodeVarint() - if err != nil { - return - } - x = (x >> 1) ^ uint64((int64(x&1)<<63)>>63) - return -} - -// DecodeZigzag32 reads a zigzag-encoded 32-bit integer -// from the Buffer. -// This is the format used for the sint32 protocol buffer type. -func (p *Buffer) DecodeZigzag32() (x uint64, err error) { - x, err = p.DecodeVarint() - if err != nil { - return - } - x = uint64((uint32(x) >> 1) ^ uint32((int32(x&1)<<31)>>31)) - return -} - -// DecodeRawBytes reads a count-delimited byte buffer from the Buffer. -// This is the format used for the bytes protocol buffer -// type and for embedded messages. -func (p *Buffer) DecodeRawBytes(alloc bool) (buf []byte, err error) { - n, err := p.DecodeVarint() - if err != nil { - return nil, err - } - - nb := int(n) - if nb < 0 { - return nil, fmt.Errorf("proto: bad byte length %d", nb) - } - end := p.index + nb - if end < p.index || end > len(p.buf) { - return nil, io.ErrUnexpectedEOF - } - - if !alloc { - // todo: check if can get more uses of alloc=false - buf = p.buf[p.index:end] - p.index += nb - return - } - - buf = make([]byte, nb) - copy(buf, p.buf[p.index:]) - p.index += nb - return -} - -// DecodeStringBytes reads an encoded string from the Buffer. -// This is the format used for the proto2 string type. -func (p *Buffer) DecodeStringBytes() (s string, err error) { - buf, err := p.DecodeRawBytes(false) - if err != nil { - return - } - return string(buf), nil -} - -// Unmarshaler is the interface representing objects that can -// unmarshal themselves. The argument points to data that may be -// overwritten, so implementations should not keep references to the -// buffer. -// Unmarshal implementations should not clear the receiver. -// Any unmarshaled data should be merged into the receiver. -// Callers of Unmarshal that do not want to retain existing data -// should Reset the receiver before calling Unmarshal. -type Unmarshaler interface { - Unmarshal([]byte) error -} - -// newUnmarshaler is the interface representing objects that can -// unmarshal themselves. The semantics are identical to Unmarshaler. -// -// This exists to support protoc-gen-go generated messages. -// The proto package will stop type-asserting to this interface in the future. -// -// DO NOT DEPEND ON THIS. -type newUnmarshaler interface { - XXX_Unmarshal([]byte) error -} - -// Unmarshal parses the protocol buffer representation in buf and places the -// decoded result in pb. If the struct underlying pb does not match -// the data in buf, the results can be unpredictable. -// -// Unmarshal resets pb before starting to unmarshal, so any -// existing data in pb is always removed. Use UnmarshalMerge -// to preserve and append to existing data. -func Unmarshal(buf []byte, pb Message) error { - pb.Reset() - if u, ok := pb.(newUnmarshaler); ok { - return u.XXX_Unmarshal(buf) - } - if u, ok := pb.(Unmarshaler); ok { - return u.Unmarshal(buf) - } - return NewBuffer(buf).Unmarshal(pb) -} - -// UnmarshalMerge parses the protocol buffer representation in buf and -// writes the decoded result to pb. If the struct underlying pb does not match -// the data in buf, the results can be unpredictable. -// -// UnmarshalMerge merges into existing data in pb. -// Most code should use Unmarshal instead. -func UnmarshalMerge(buf []byte, pb Message) error { - if u, ok := pb.(newUnmarshaler); ok { - return u.XXX_Unmarshal(buf) - } - if u, ok := pb.(Unmarshaler); ok { - // NOTE: The history of proto have unfortunately been inconsistent - // whether Unmarshaler should or should not implicitly clear itself. - // Some implementations do, most do not. - // Thus, calling this here may or may not do what people want. - // - // See https://github.com/golang/protobuf/issues/424 - return u.Unmarshal(buf) - } - return NewBuffer(buf).Unmarshal(pb) -} - -// DecodeMessage reads a count-delimited message from the Buffer. -func (p *Buffer) DecodeMessage(pb Message) error { - enc, err := p.DecodeRawBytes(false) - if err != nil { - return err - } - return NewBuffer(enc).Unmarshal(pb) -} - -// DecodeGroup reads a tag-delimited group from the Buffer. -// StartGroup tag is already consumed. This function consumes -// EndGroup tag. -func (p *Buffer) DecodeGroup(pb Message) error { - b := p.buf[p.index:] - x, y := findEndGroup(b) - if x < 0 { - return io.ErrUnexpectedEOF - } - err := Unmarshal(b[:x], pb) - p.index += y - return err -} - -// Unmarshal parses the protocol buffer representation in the -// Buffer and places the decoded result in pb. If the struct -// underlying pb does not match the data in the buffer, the results can be -// unpredictable. -// -// Unlike proto.Unmarshal, this does not reset pb before starting to unmarshal. -func (p *Buffer) Unmarshal(pb Message) error { - // If the object can unmarshal itself, let it. - if u, ok := pb.(newUnmarshaler); ok { - err := u.XXX_Unmarshal(p.buf[p.index:]) - p.index = len(p.buf) - return err - } - if u, ok := pb.(Unmarshaler); ok { - // NOTE: The history of proto have unfortunately been inconsistent - // whether Unmarshaler should or should not implicitly clear itself. - // Some implementations do, most do not. - // Thus, calling this here may or may not do what people want. - // - // See https://github.com/golang/protobuf/issues/424 - err := u.Unmarshal(p.buf[p.index:]) - p.index = len(p.buf) - return err - } - - // Slow workaround for messages that aren't Unmarshalers. - // This includes some hand-coded .pb.go files and - // bootstrap protos. - // TODO: fix all of those and then add Unmarshal to - // the Message interface. Then: - // The cast above and code below can be deleted. - // The old unmarshaler can be deleted. - // Clients can call Unmarshal directly (can already do that, actually). - var info InternalMessageInfo - err := info.Unmarshal(pb, p.buf[p.index:]) - p.index = len(p.buf) - return err -} diff --git a/vendor/github.com/golang/protobuf/proto/defaults.go b/vendor/github.com/golang/protobuf/proto/defaults.go new file mode 100644 index 00000000..d399bf06 --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/defaults.go @@ -0,0 +1,63 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "google.golang.org/protobuf/reflect/protoreflect" +) + +// SetDefaults sets unpopulated scalar fields to their default values. +// Fields within a oneof are not set even if they have a default value. +// SetDefaults is recursively called upon any populated message fields. +func SetDefaults(m Message) { + if m != nil { + setDefaults(MessageReflect(m)) + } +} + +func setDefaults(m protoreflect.Message) { + fds := m.Descriptor().Fields() + for i := 0; i < fds.Len(); i++ { + fd := fds.Get(i) + if !m.Has(fd) { + if fd.HasDefault() && fd.ContainingOneof() == nil { + v := fd.Default() + if fd.Kind() == protoreflect.BytesKind { + v = protoreflect.ValueOf(append([]byte(nil), v.Bytes()...)) // copy the default bytes + } + m.Set(fd, v) + } + continue + } + } + + m.Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + switch { + // Handle singular message. + case fd.Cardinality() != protoreflect.Repeated: + if fd.Message() != nil { + setDefaults(m.Get(fd).Message()) + } + // Handle list of messages. + case fd.IsList(): + if fd.Message() != nil { + ls := m.Get(fd).List() + for i := 0; i < ls.Len(); i++ { + setDefaults(ls.Get(i).Message()) + } + } + // Handle map of messages. + case fd.IsMap(): + if fd.MapValue().Message() != nil { + ms := m.Get(fd).Map() + ms.Range(func(_ protoreflect.MapKey, v protoreflect.Value) bool { + setDefaults(v.Message()) + return true + }) + } + } + return true + }) +} diff --git a/vendor/github.com/golang/protobuf/proto/deprecated.go b/vendor/github.com/golang/protobuf/proto/deprecated.go index 35b882c0..e8db57e0 100644 --- a/vendor/github.com/golang/protobuf/proto/deprecated.go +++ b/vendor/github.com/golang/protobuf/proto/deprecated.go @@ -1,63 +1,113 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2018 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. package proto -import "errors" +import ( + "encoding/json" + "errors" + "fmt" + "strconv" -// Deprecated: do not use. + protoV2 "google.golang.org/protobuf/proto" +) + +var ( + // Deprecated: No longer returned. + ErrNil = errors.New("proto: Marshal called with nil") + + // Deprecated: No longer returned. + ErrTooLarge = errors.New("proto: message encodes to over 2 GB") + + // Deprecated: No longer returned. + ErrInternalBadWireType = errors.New("proto: internal error: bad wiretype for oneof") +) + +// Deprecated: Do not use. type Stats struct{ Emalloc, Dmalloc, Encode, Decode, Chit, Cmiss, Size uint64 } -// Deprecated: do not use. +// Deprecated: Do not use. func GetStats() Stats { return Stats{} } -// Deprecated: do not use. +// Deprecated: Do not use. func MarshalMessageSet(interface{}) ([]byte, error) { return nil, errors.New("proto: not implemented") } -// Deprecated: do not use. +// Deprecated: Do not use. func UnmarshalMessageSet([]byte, interface{}) error { return errors.New("proto: not implemented") } -// Deprecated: do not use. +// Deprecated: Do not use. func MarshalMessageSetJSON(interface{}) ([]byte, error) { return nil, errors.New("proto: not implemented") } -// Deprecated: do not use. +// Deprecated: Do not use. func UnmarshalMessageSetJSON([]byte, interface{}) error { return errors.New("proto: not implemented") } -// Deprecated: do not use. +// Deprecated: Do not use. func RegisterMessageSetType(Message, int32, string) {} + +// Deprecated: Do not use. +func EnumName(m map[int32]string, v int32) string { + s, ok := m[v] + if ok { + return s + } + return strconv.Itoa(int(v)) +} + +// Deprecated: Do not use. +func UnmarshalJSONEnum(m map[string]int32, data []byte, enumName string) (int32, error) { + if data[0] == '"' { + // New style: enums are strings. + var repr string + if err := json.Unmarshal(data, &repr); err != nil { + return -1, err + } + val, ok := m[repr] + if !ok { + return 0, fmt.Errorf("unrecognized enum %s value %q", enumName, repr) + } + return val, nil + } + // Old style: enums are ints. + var val int32 + if err := json.Unmarshal(data, &val); err != nil { + return 0, fmt.Errorf("cannot unmarshal %#q into enum %s", data, enumName) + } + return val, nil +} + +// Deprecated: Do not use; this type existed for intenal-use only. +type InternalMessageInfo struct{} + +// Deprecated: Do not use; this method existed for intenal-use only. +func (*InternalMessageInfo) DiscardUnknown(m Message) { + DiscardUnknown(m) +} + +// Deprecated: Do not use; this method existed for intenal-use only. +func (*InternalMessageInfo) Marshal(b []byte, m Message, deterministic bool) ([]byte, error) { + return protoV2.MarshalOptions{Deterministic: deterministic}.MarshalAppend(b, MessageV2(m)) +} + +// Deprecated: Do not use; this method existed for intenal-use only. +func (*InternalMessageInfo) Merge(dst, src Message) { + protoV2.Merge(MessageV2(dst), MessageV2(src)) +} + +// Deprecated: Do not use; this method existed for intenal-use only. +func (*InternalMessageInfo) Size(m Message) int { + return protoV2.Size(MessageV2(m)) +} + +// Deprecated: Do not use; this method existed for intenal-use only. +func (*InternalMessageInfo) Unmarshal(m Message, b []byte) error { + return protoV2.UnmarshalOptions{Merge: true}.Unmarshal(b, MessageV2(m)) +} diff --git a/vendor/github.com/golang/protobuf/proto/discard.go b/vendor/github.com/golang/protobuf/proto/discard.go index dea2617c..2187e877 100644 --- a/vendor/github.com/golang/protobuf/proto/discard.go +++ b/vendor/github.com/golang/protobuf/proto/discard.go @@ -1,48 +1,13 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2017 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. package proto import ( - "fmt" - "reflect" - "strings" - "sync" - "sync/atomic" + "google.golang.org/protobuf/reflect/protoreflect" ) -type generatedDiscarder interface { - XXX_DiscardUnknown() -} - // DiscardUnknown recursively discards all unknown fields from this message // and all embedded messages. // @@ -51,300 +16,43 @@ type generatedDiscarder interface { // marshal to be able to produce a message that continues to have those // unrecognized fields. To avoid this, DiscardUnknown is used to // explicitly clear the unknown fields after unmarshaling. -// -// For proto2 messages, the unknown fields of message extensions are only -// discarded from messages that have been accessed via GetExtension. func DiscardUnknown(m Message) { - if m, ok := m.(generatedDiscarder); ok { - m.XXX_DiscardUnknown() - return - } - // TODO: Dynamically populate a InternalMessageInfo for legacy messages, - // but the master branch has no implementation for InternalMessageInfo, - // so it would be more work to replicate that approach. - discardLegacy(m) -} - -// DiscardUnknown recursively discards all unknown fields. -func (a *InternalMessageInfo) DiscardUnknown(m Message) { - di := atomicLoadDiscardInfo(&a.discard) - if di == nil { - di = getDiscardInfo(reflect.TypeOf(m).Elem()) - atomicStoreDiscardInfo(&a.discard, di) - } - di.discard(toPointer(&m)) -} - -type discardInfo struct { - typ reflect.Type - - initialized int32 // 0: only typ is valid, 1: everything is valid - lock sync.Mutex - - fields []discardFieldInfo - unrecognized field -} - -type discardFieldInfo struct { - field field // Offset of field, guaranteed to be valid - discard func(src pointer) -} - -var ( - discardInfoMap = map[reflect.Type]*discardInfo{} - discardInfoLock sync.Mutex -) - -func getDiscardInfo(t reflect.Type) *discardInfo { - discardInfoLock.Lock() - defer discardInfoLock.Unlock() - di := discardInfoMap[t] - if di == nil { - di = &discardInfo{typ: t} - discardInfoMap[t] = di + if m != nil { + discardUnknown(MessageReflect(m)) } - return di } -func (di *discardInfo) discard(src pointer) { - if src.isNil() { - return // Nothing to do. - } - - if atomic.LoadInt32(&di.initialized) == 0 { - di.computeDiscardInfo() - } - - for _, fi := range di.fields { - sfp := src.offset(fi.field) - fi.discard(sfp) - } - - // For proto2 messages, only discard unknown fields in message extensions - // that have been accessed via GetExtension. - if em, err := extendable(src.asPointerTo(di.typ).Interface()); err == nil { - // Ignore lock since DiscardUnknown is not concurrency safe. - emm, _ := em.extensionsRead() - for _, mx := range emm { - if m, ok := mx.value.(Message); ok { - DiscardUnknown(m) +func discardUnknown(m protoreflect.Message) { + m.Range(func(fd protoreflect.FieldDescriptor, val protoreflect.Value) bool { + switch { + // Handle singular message. + case fd.Cardinality() != protoreflect.Repeated: + if fd.Message() != nil { + discardUnknown(m.Get(fd).Message()) } - } - } - - if di.unrecognized.IsValid() { - *src.offset(di.unrecognized).toBytes() = nil - } -} - -func (di *discardInfo) computeDiscardInfo() { - di.lock.Lock() - defer di.lock.Unlock() - if di.initialized != 0 { - return - } - t := di.typ - n := t.NumField() - - for i := 0; i < n; i++ { - f := t.Field(i) - if strings.HasPrefix(f.Name, "XXX_") { - continue - } - - dfi := discardFieldInfo{field: toField(&f)} - tf := f.Type - - // Unwrap tf to get its most basic type. - var isPointer, isSlice bool - if tf.Kind() == reflect.Slice && tf.Elem().Kind() != reflect.Uint8 { - isSlice = true - tf = tf.Elem() - } - if tf.Kind() == reflect.Ptr { - isPointer = true - tf = tf.Elem() - } - if isPointer && isSlice && tf.Kind() != reflect.Struct { - panic(fmt.Sprintf("%v.%s cannot be a slice of pointers to primitive types", t, f.Name)) - } - - switch tf.Kind() { - case reflect.Struct: - switch { - case !isPointer: - panic(fmt.Sprintf("%v.%s cannot be a direct struct value", t, f.Name)) - case isSlice: // E.g., []*pb.T - di := getDiscardInfo(tf) - dfi.discard = func(src pointer) { - sps := src.getPointerSlice() - for _, sp := range sps { - if !sp.isNil() { - di.discard(sp) - } - } - } - default: // E.g., *pb.T - di := getDiscardInfo(tf) - dfi.discard = func(src pointer) { - sp := src.getPointer() - if !sp.isNil() { - di.discard(sp) - } + // Handle list of messages. + case fd.IsList(): + if fd.Message() != nil { + ls := m.Get(fd).List() + for i := 0; i < ls.Len(); i++ { + discardUnknown(ls.Get(i).Message()) } } - case reflect.Map: - switch { - case isPointer || isSlice: - panic(fmt.Sprintf("%v.%s cannot be a pointer to a map or a slice of map values", t, f.Name)) - default: // E.g., map[K]V - if tf.Elem().Kind() == reflect.Ptr { // Proto struct (e.g., *T) - dfi.discard = func(src pointer) { - sm := src.asPointerTo(tf).Elem() - if sm.Len() == 0 { - return - } - for _, key := range sm.MapKeys() { - val := sm.MapIndex(key) - DiscardUnknown(val.Interface().(Message)) - } - } - } else { - dfi.discard = func(pointer) {} // Noop - } - } - case reflect.Interface: - // Must be oneof field. - switch { - case isPointer || isSlice: - panic(fmt.Sprintf("%v.%s cannot be a pointer to a interface or a slice of interface values", t, f.Name)) - default: // E.g., interface{} - // TODO: Make this faster? - dfi.discard = func(src pointer) { - su := src.asPointerTo(tf).Elem() - if !su.IsNil() { - sv := su.Elem().Elem().Field(0) - if sv.Kind() == reflect.Ptr && sv.IsNil() { - return - } - switch sv.Type().Kind() { - case reflect.Ptr: // Proto struct (e.g., *T) - DiscardUnknown(sv.Interface().(Message)) - } - } - } + // Handle map of messages. + case fd.IsMap(): + if fd.MapValue().Message() != nil { + ms := m.Get(fd).Map() + ms.Range(func(_ protoreflect.MapKey, v protoreflect.Value) bool { + discardUnknown(v.Message()) + return true + }) } - default: - continue - } - di.fields = append(di.fields, dfi) - } - - di.unrecognized = invalidField - if f, ok := t.FieldByName("XXX_unrecognized"); ok { - if f.Type != reflect.TypeOf([]byte{}) { - panic("expected XXX_unrecognized to be of type []byte") - } - di.unrecognized = toField(&f) - } - - atomic.StoreInt32(&di.initialized, 1) -} - -func discardLegacy(m Message) { - v := reflect.ValueOf(m) - if v.Kind() != reflect.Ptr || v.IsNil() { - return - } - v = v.Elem() - if v.Kind() != reflect.Struct { - return - } - t := v.Type() - - for i := 0; i < v.NumField(); i++ { - f := t.Field(i) - if strings.HasPrefix(f.Name, "XXX_") { - continue } - vf := v.Field(i) - tf := f.Type + return true + }) - // Unwrap tf to get its most basic type. - var isPointer, isSlice bool - if tf.Kind() == reflect.Slice && tf.Elem().Kind() != reflect.Uint8 { - isSlice = true - tf = tf.Elem() - } - if tf.Kind() == reflect.Ptr { - isPointer = true - tf = tf.Elem() - } - if isPointer && isSlice && tf.Kind() != reflect.Struct { - panic(fmt.Sprintf("%T.%s cannot be a slice of pointers to primitive types", m, f.Name)) - } - - switch tf.Kind() { - case reflect.Struct: - switch { - case !isPointer: - panic(fmt.Sprintf("%T.%s cannot be a direct struct value", m, f.Name)) - case isSlice: // E.g., []*pb.T - for j := 0; j < vf.Len(); j++ { - discardLegacy(vf.Index(j).Interface().(Message)) - } - default: // E.g., *pb.T - discardLegacy(vf.Interface().(Message)) - } - case reflect.Map: - switch { - case isPointer || isSlice: - panic(fmt.Sprintf("%T.%s cannot be a pointer to a map or a slice of map values", m, f.Name)) - default: // E.g., map[K]V - tv := vf.Type().Elem() - if tv.Kind() == reflect.Ptr && tv.Implements(protoMessageType) { // Proto struct (e.g., *T) - for _, key := range vf.MapKeys() { - val := vf.MapIndex(key) - discardLegacy(val.Interface().(Message)) - } - } - } - case reflect.Interface: - // Must be oneof field. - switch { - case isPointer || isSlice: - panic(fmt.Sprintf("%T.%s cannot be a pointer to a interface or a slice of interface values", m, f.Name)) - default: // E.g., test_proto.isCommunique_Union interface - if !vf.IsNil() && f.Tag.Get("protobuf_oneof") != "" { - vf = vf.Elem() // E.g., *test_proto.Communique_Msg - if !vf.IsNil() { - vf = vf.Elem() // E.g., test_proto.Communique_Msg - vf = vf.Field(0) // E.g., Proto struct (e.g., *T) or primitive value - if vf.Kind() == reflect.Ptr { - discardLegacy(vf.Interface().(Message)) - } - } - } - } - } - } - - if vf := v.FieldByName("XXX_unrecognized"); vf.IsValid() { - if vf.Type() != reflect.TypeOf([]byte{}) { - panic("expected XXX_unrecognized to be of type []byte") - } - vf.Set(reflect.ValueOf([]byte(nil))) - } - - // For proto2 messages, only discard unknown fields in message extensions - // that have been accessed via GetExtension. - if em, err := extendable(m); err == nil { - // Ignore lock since discardLegacy is not concurrency safe. - emm, _ := em.extensionsRead() - for _, mx := range emm { - if m, ok := mx.value.(Message); ok { - discardLegacy(m) - } - } + // Discard unknown fields. + if len(m.GetUnknown()) > 0 { + m.SetUnknown(nil) } } diff --git a/vendor/github.com/golang/protobuf/proto/encode.go b/vendor/github.com/golang/protobuf/proto/encode.go deleted file mode 100644 index 3abfed2c..00000000 --- a/vendor/github.com/golang/protobuf/proto/encode.go +++ /dev/null @@ -1,203 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -/* - * Routines for encoding data into the wire format for protocol buffers. - */ - -import ( - "errors" - "reflect" -) - -var ( - // errRepeatedHasNil is the error returned if Marshal is called with - // a struct with a repeated field containing a nil element. - errRepeatedHasNil = errors.New("proto: repeated field has nil element") - - // errOneofHasNil is the error returned if Marshal is called with - // a struct with a oneof field containing a nil element. - errOneofHasNil = errors.New("proto: oneof field has nil value") - - // ErrNil is the error returned if Marshal is called with nil. - ErrNil = errors.New("proto: Marshal called with nil") - - // ErrTooLarge is the error returned if Marshal is called with a - // message that encodes to >2GB. - ErrTooLarge = errors.New("proto: message encodes to over 2 GB") -) - -// The fundamental encoders that put bytes on the wire. -// Those that take integer types all accept uint64 and are -// therefore of type valueEncoder. - -const maxVarintBytes = 10 // maximum length of a varint - -// EncodeVarint returns the varint encoding of x. -// This is the format for the -// int32, int64, uint32, uint64, bool, and enum -// protocol buffer types. -// Not used by the package itself, but helpful to clients -// wishing to use the same encoding. -func EncodeVarint(x uint64) []byte { - var buf [maxVarintBytes]byte - var n int - for n = 0; x > 127; n++ { - buf[n] = 0x80 | uint8(x&0x7F) - x >>= 7 - } - buf[n] = uint8(x) - n++ - return buf[0:n] -} - -// EncodeVarint writes a varint-encoded integer to the Buffer. -// This is the format for the -// int32, int64, uint32, uint64, bool, and enum -// protocol buffer types. -func (p *Buffer) EncodeVarint(x uint64) error { - for x >= 1<<7 { - p.buf = append(p.buf, uint8(x&0x7f|0x80)) - x >>= 7 - } - p.buf = append(p.buf, uint8(x)) - return nil -} - -// SizeVarint returns the varint encoding size of an integer. -func SizeVarint(x uint64) int { - switch { - case x < 1<<7: - return 1 - case x < 1<<14: - return 2 - case x < 1<<21: - return 3 - case x < 1<<28: - return 4 - case x < 1<<35: - return 5 - case x < 1<<42: - return 6 - case x < 1<<49: - return 7 - case x < 1<<56: - return 8 - case x < 1<<63: - return 9 - } - return 10 -} - -// EncodeFixed64 writes a 64-bit integer to the Buffer. -// This is the format for the -// fixed64, sfixed64, and double protocol buffer types. -func (p *Buffer) EncodeFixed64(x uint64) error { - p.buf = append(p.buf, - uint8(x), - uint8(x>>8), - uint8(x>>16), - uint8(x>>24), - uint8(x>>32), - uint8(x>>40), - uint8(x>>48), - uint8(x>>56)) - return nil -} - -// EncodeFixed32 writes a 32-bit integer to the Buffer. -// This is the format for the -// fixed32, sfixed32, and float protocol buffer types. -func (p *Buffer) EncodeFixed32(x uint64) error { - p.buf = append(p.buf, - uint8(x), - uint8(x>>8), - uint8(x>>16), - uint8(x>>24)) - return nil -} - -// EncodeZigzag64 writes a zigzag-encoded 64-bit integer -// to the Buffer. -// This is the format used for the sint64 protocol buffer type. -func (p *Buffer) EncodeZigzag64(x uint64) error { - // use signed number to get arithmetic right shift. - return p.EncodeVarint(uint64((x << 1) ^ uint64((int64(x) >> 63)))) -} - -// EncodeZigzag32 writes a zigzag-encoded 32-bit integer -// to the Buffer. -// This is the format used for the sint32 protocol buffer type. -func (p *Buffer) EncodeZigzag32(x uint64) error { - // use signed number to get arithmetic right shift. - return p.EncodeVarint(uint64((uint32(x) << 1) ^ uint32((int32(x) >> 31)))) -} - -// EncodeRawBytes writes a count-delimited byte buffer to the Buffer. -// This is the format used for the bytes protocol buffer -// type and for embedded messages. -func (p *Buffer) EncodeRawBytes(b []byte) error { - p.EncodeVarint(uint64(len(b))) - p.buf = append(p.buf, b...) - return nil -} - -// EncodeStringBytes writes an encoded string to the Buffer. -// This is the format used for the proto2 string type. -func (p *Buffer) EncodeStringBytes(s string) error { - p.EncodeVarint(uint64(len(s))) - p.buf = append(p.buf, s...) - return nil -} - -// Marshaler is the interface representing objects that can marshal themselves. -type Marshaler interface { - Marshal() ([]byte, error) -} - -// EncodeMessage writes the protocol buffer to the Buffer, -// prefixed by a varint-encoded length. -func (p *Buffer) EncodeMessage(pb Message) error { - siz := Size(pb) - p.EncodeVarint(uint64(siz)) - return p.Marshal(pb) -} - -// All protocol buffer fields are nillable, but be careful. -func isNil(v reflect.Value) bool { - switch v.Kind() { - case reflect.Interface, reflect.Map, reflect.Ptr, reflect.Slice: - return v.IsNil() - } - return false -} diff --git a/vendor/github.com/golang/protobuf/proto/equal.go b/vendor/github.com/golang/protobuf/proto/equal.go deleted file mode 100644 index f9b6e41b..00000000 --- a/vendor/github.com/golang/protobuf/proto/equal.go +++ /dev/null @@ -1,301 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2011 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -// Protocol buffer comparison. - -package proto - -import ( - "bytes" - "log" - "reflect" - "strings" -) - -/* -Equal returns true iff protocol buffers a and b are equal. -The arguments must both be pointers to protocol buffer structs. - -Equality is defined in this way: - - Two messages are equal iff they are the same type, - corresponding fields are equal, unknown field sets - are equal, and extensions sets are equal. - - Two set scalar fields are equal iff their values are equal. - If the fields are of a floating-point type, remember that - NaN != x for all x, including NaN. If the message is defined - in a proto3 .proto file, fields are not "set"; specifically, - zero length proto3 "bytes" fields are equal (nil == {}). - - Two repeated fields are equal iff their lengths are the same, - and their corresponding elements are equal. Note a "bytes" field, - although represented by []byte, is not a repeated field and the - rule for the scalar fields described above applies. - - Two unset fields are equal. - - Two unknown field sets are equal if their current - encoded state is equal. - - Two extension sets are equal iff they have corresponding - elements that are pairwise equal. - - Two map fields are equal iff their lengths are the same, - and they contain the same set of elements. Zero-length map - fields are equal. - - Every other combination of things are not equal. - -The return value is undefined if a and b are not protocol buffers. -*/ -func Equal(a, b Message) bool { - if a == nil || b == nil { - return a == b - } - v1, v2 := reflect.ValueOf(a), reflect.ValueOf(b) - if v1.Type() != v2.Type() { - return false - } - if v1.Kind() == reflect.Ptr { - if v1.IsNil() { - return v2.IsNil() - } - if v2.IsNil() { - return false - } - v1, v2 = v1.Elem(), v2.Elem() - } - if v1.Kind() != reflect.Struct { - return false - } - return equalStruct(v1, v2) -} - -// v1 and v2 are known to have the same type. -func equalStruct(v1, v2 reflect.Value) bool { - sprop := GetProperties(v1.Type()) - for i := 0; i < v1.NumField(); i++ { - f := v1.Type().Field(i) - if strings.HasPrefix(f.Name, "XXX_") { - continue - } - f1, f2 := v1.Field(i), v2.Field(i) - if f.Type.Kind() == reflect.Ptr { - if n1, n2 := f1.IsNil(), f2.IsNil(); n1 && n2 { - // both unset - continue - } else if n1 != n2 { - // set/unset mismatch - return false - } - f1, f2 = f1.Elem(), f2.Elem() - } - if !equalAny(f1, f2, sprop.Prop[i]) { - return false - } - } - - if em1 := v1.FieldByName("XXX_InternalExtensions"); em1.IsValid() { - em2 := v2.FieldByName("XXX_InternalExtensions") - if !equalExtensions(v1.Type(), em1.Interface().(XXX_InternalExtensions), em2.Interface().(XXX_InternalExtensions)) { - return false - } - } - - if em1 := v1.FieldByName("XXX_extensions"); em1.IsValid() { - em2 := v2.FieldByName("XXX_extensions") - if !equalExtMap(v1.Type(), em1.Interface().(map[int32]Extension), em2.Interface().(map[int32]Extension)) { - return false - } - } - - uf := v1.FieldByName("XXX_unrecognized") - if !uf.IsValid() { - return true - } - - u1 := uf.Bytes() - u2 := v2.FieldByName("XXX_unrecognized").Bytes() - return bytes.Equal(u1, u2) -} - -// v1 and v2 are known to have the same type. -// prop may be nil. -func equalAny(v1, v2 reflect.Value, prop *Properties) bool { - if v1.Type() == protoMessageType { - m1, _ := v1.Interface().(Message) - m2, _ := v2.Interface().(Message) - return Equal(m1, m2) - } - switch v1.Kind() { - case reflect.Bool: - return v1.Bool() == v2.Bool() - case reflect.Float32, reflect.Float64: - return v1.Float() == v2.Float() - case reflect.Int32, reflect.Int64: - return v1.Int() == v2.Int() - case reflect.Interface: - // Probably a oneof field; compare the inner values. - n1, n2 := v1.IsNil(), v2.IsNil() - if n1 || n2 { - return n1 == n2 - } - e1, e2 := v1.Elem(), v2.Elem() - if e1.Type() != e2.Type() { - return false - } - return equalAny(e1, e2, nil) - case reflect.Map: - if v1.Len() != v2.Len() { - return false - } - for _, key := range v1.MapKeys() { - val2 := v2.MapIndex(key) - if !val2.IsValid() { - // This key was not found in the second map. - return false - } - if !equalAny(v1.MapIndex(key), val2, nil) { - return false - } - } - return true - case reflect.Ptr: - // Maps may have nil values in them, so check for nil. - if v1.IsNil() && v2.IsNil() { - return true - } - if v1.IsNil() != v2.IsNil() { - return false - } - return equalAny(v1.Elem(), v2.Elem(), prop) - case reflect.Slice: - if v1.Type().Elem().Kind() == reflect.Uint8 { - // short circuit: []byte - - // Edge case: if this is in a proto3 message, a zero length - // bytes field is considered the zero value. - if prop != nil && prop.proto3 && v1.Len() == 0 && v2.Len() == 0 { - return true - } - if v1.IsNil() != v2.IsNil() { - return false - } - return bytes.Equal(v1.Interface().([]byte), v2.Interface().([]byte)) - } - - if v1.Len() != v2.Len() { - return false - } - for i := 0; i < v1.Len(); i++ { - if !equalAny(v1.Index(i), v2.Index(i), prop) { - return false - } - } - return true - case reflect.String: - return v1.Interface().(string) == v2.Interface().(string) - case reflect.Struct: - return equalStruct(v1, v2) - case reflect.Uint32, reflect.Uint64: - return v1.Uint() == v2.Uint() - } - - // unknown type, so not a protocol buffer - log.Printf("proto: don't know how to compare %v", v1) - return false -} - -// base is the struct type that the extensions are based on. -// x1 and x2 are InternalExtensions. -func equalExtensions(base reflect.Type, x1, x2 XXX_InternalExtensions) bool { - em1, _ := x1.extensionsRead() - em2, _ := x2.extensionsRead() - return equalExtMap(base, em1, em2) -} - -func equalExtMap(base reflect.Type, em1, em2 map[int32]Extension) bool { - if len(em1) != len(em2) { - return false - } - - for extNum, e1 := range em1 { - e2, ok := em2[extNum] - if !ok { - return false - } - - m1 := extensionAsLegacyType(e1.value) - m2 := extensionAsLegacyType(e2.value) - - if m1 == nil && m2 == nil { - // Both have only encoded form. - if bytes.Equal(e1.enc, e2.enc) { - continue - } - // The bytes are different, but the extensions might still be - // equal. We need to decode them to compare. - } - - if m1 != nil && m2 != nil { - // Both are unencoded. - if !equalAny(reflect.ValueOf(m1), reflect.ValueOf(m2), nil) { - return false - } - continue - } - - // At least one is encoded. To do a semantically correct comparison - // we need to unmarshal them first. - var desc *ExtensionDesc - if m := extensionMaps[base]; m != nil { - desc = m[extNum] - } - if desc == nil { - // If both have only encoded form and the bytes are the same, - // it is handled above. We get here when the bytes are different. - // We don't know how to decode it, so just compare them as byte - // slices. - log.Printf("proto: don't know how to compare extension %d of %v", extNum, base) - return false - } - var err error - if m1 == nil { - m1, err = decodeExtension(e1.enc, desc) - } - if m2 == nil && err == nil { - m2, err = decodeExtension(e2.enc, desc) - } - if err != nil { - // The encoded form is invalid. - log.Printf("proto: badly encoded extension %d of %v: %v", extNum, base, err) - return false - } - if !equalAny(reflect.ValueOf(m1), reflect.ValueOf(m2), nil) { - return false - } - } - - return true -} diff --git a/vendor/github.com/golang/protobuf/proto/extensions.go b/vendor/github.com/golang/protobuf/proto/extensions.go index fa88add3..42fc120c 100644 --- a/vendor/github.com/golang/protobuf/proto/extensions.go +++ b/vendor/github.com/golang/protobuf/proto/extensions.go @@ -1,607 +1,356 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2010 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. package proto -/* - * Types and routines for supporting protocol buffer extensions. - */ - import ( "errors" "fmt" - "io" "reflect" - "strconv" - "sync" -) - -// ErrMissingExtension is the error returned by GetExtension if the named extension is not in the message. -var ErrMissingExtension = errors.New("proto: missing extension") - -// ExtensionRange represents a range of message extensions for a protocol buffer. -// Used in code generated by the protocol compiler. -type ExtensionRange struct { - Start, End int32 // both inclusive -} - -// extendableProto is an interface implemented by any protocol buffer generated by the current -// proto compiler that may be extended. -type extendableProto interface { - Message - ExtensionRangeArray() []ExtensionRange - extensionsWrite() map[int32]Extension - extensionsRead() (map[int32]Extension, sync.Locker) -} - -// extendableProtoV1 is an interface implemented by a protocol buffer generated by the previous -// version of the proto compiler that may be extended. -type extendableProtoV1 interface { - Message - ExtensionRangeArray() []ExtensionRange - ExtensionMap() map[int32]Extension -} -// extensionAdapter is a wrapper around extendableProtoV1 that implements extendableProto. -type extensionAdapter struct { - extendableProtoV1 -} + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/proto" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" + "google.golang.org/protobuf/runtime/protoiface" + "google.golang.org/protobuf/runtime/protoimpl" +) -func (e extensionAdapter) extensionsWrite() map[int32]Extension { - return e.ExtensionMap() -} +type ( + // ExtensionDesc represents an extension descriptor and + // is used to interact with an extension field in a message. + // + // Variables of this type are generated in code by protoc-gen-go. + ExtensionDesc = protoimpl.ExtensionInfo -func (e extensionAdapter) extensionsRead() (map[int32]Extension, sync.Locker) { - return e.ExtensionMap(), notLocker{} -} + // ExtensionRange represents a range of message extensions. + // Used in code generated by protoc-gen-go. + ExtensionRange = protoiface.ExtensionRangeV1 -// notLocker is a sync.Locker whose Lock and Unlock methods are nops. -type notLocker struct{} + // Deprecated: Do not use; this is an internal type. + Extension = protoimpl.ExtensionFieldV1 -func (n notLocker) Lock() {} -func (n notLocker) Unlock() {} + // Deprecated: Do not use; this is an internal type. + XXX_InternalExtensions = protoimpl.ExtensionFields +) -// extendable returns the extendableProto interface for the given generated proto message. -// If the proto message has the old extension format, it returns a wrapper that implements -// the extendableProto interface. -func extendable(p interface{}) (extendableProto, error) { - switch p := p.(type) { - case extendableProto: - if isNilPtr(p) { - return nil, fmt.Errorf("proto: nil %T is not extendable", p) - } - return p, nil - case extendableProtoV1: - if isNilPtr(p) { - return nil, fmt.Errorf("proto: nil %T is not extendable", p) - } - return extensionAdapter{p}, nil - } - // Don't allocate a specific error containing %T: - // this is the hot path for Clone and MarshalText. - return nil, errNotExtendable -} +// ErrMissingExtension reports whether the extension was not present. +var ErrMissingExtension = errors.New("proto: missing extension") var errNotExtendable = errors.New("proto: not an extendable proto.Message") -func isNilPtr(x interface{}) bool { - v := reflect.ValueOf(x) - return v.Kind() == reflect.Ptr && v.IsNil() -} - -// XXX_InternalExtensions is an internal representation of proto extensions. -// -// Each generated message struct type embeds an anonymous XXX_InternalExtensions field, -// thus gaining the unexported 'extensions' method, which can be called only from the proto package. -// -// The methods of XXX_InternalExtensions are not concurrency safe in general, -// but calls to logically read-only methods such as has and get may be executed concurrently. -type XXX_InternalExtensions struct { - // The struct must be indirect so that if a user inadvertently copies a - // generated message and its embedded XXX_InternalExtensions, they - // avoid the mayhem of a copied mutex. - // - // The mutex serializes all logically read-only operations to p.extensionMap. - // It is up to the client to ensure that write operations to p.extensionMap are - // mutually exclusive with other accesses. - p *struct { - mu sync.Mutex - extensionMap map[int32]Extension +// HasExtension reports whether the extension field is present in m +// either as an explicitly populated field or as an unknown field. +func HasExtension(m Message, xt *ExtensionDesc) (has bool) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() { + return false } -} -// extensionsWrite returns the extension map, creating it on first use. -func (e *XXX_InternalExtensions) extensionsWrite() map[int32]Extension { - if e.p == nil { - e.p = new(struct { - mu sync.Mutex - extensionMap map[int32]Extension + // Check whether any populated known field matches the field number. + xtd := xt.TypeDescriptor() + if isValidExtension(mr.Descriptor(), xtd) { + has = mr.Has(xtd) + } else { + mr.Range(func(fd protoreflect.FieldDescriptor, _ protoreflect.Value) bool { + has = int32(fd.Number()) == xt.Field + return !has }) - e.p.extensionMap = make(map[int32]Extension) } - return e.p.extensionMap -} -// extensionsRead returns the extensions map for read-only use. It may be nil. -// The caller must hold the returned mutex's lock when accessing Elements within the map. -func (e *XXX_InternalExtensions) extensionsRead() (map[int32]Extension, sync.Locker) { - if e.p == nil { - return nil, nil + // Check whether any unknown field matches the field number. + for b := mr.GetUnknown(); !has && len(b) > 0; { + num, _, n := protowire.ConsumeField(b) + has = int32(num) == xt.Field + b = b[n:] } - return e.p.extensionMap, &e.p.mu -} - -// ExtensionDesc represents an extension specification. -// Used in generated code from the protocol compiler. -type ExtensionDesc struct { - ExtendedType Message // nil pointer to the type that is being extended - ExtensionType interface{} // nil pointer to the extension type - Field int32 // field number - Name string // fully-qualified name of extension, for text formatting - Tag string // protobuf tag style - Filename string // name of the file in which the extension is defined -} - -func (ed *ExtensionDesc) repeated() bool { - t := reflect.TypeOf(ed.ExtensionType) - return t.Kind() == reflect.Slice && t.Elem().Kind() != reflect.Uint8 -} - -// Extension represents an extension in a message. -type Extension struct { - // When an extension is stored in a message using SetExtension - // only desc and value are set. When the message is marshaled - // enc will be set to the encoded form of the message. - // - // When a message is unmarshaled and contains extensions, each - // extension will have only enc set. When such an extension is - // accessed using GetExtension (or GetExtensions) desc and value - // will be set. - desc *ExtensionDesc - - // value is a concrete value for the extension field. Let the type of - // desc.ExtensionType be the "API type" and the type of Extension.value - // be the "storage type". The API type and storage type are the same except: - // * For scalars (except []byte), the API type uses *T, - // while the storage type uses T. - // * For repeated fields, the API type uses []T, while the storage type - // uses *[]T. - // - // The reason for the divergence is so that the storage type more naturally - // matches what is expected of when retrieving the values through the - // protobuf reflection APIs. - // - // The value may only be populated if desc is also populated. - value interface{} - - // enc is the raw bytes for the extension field. - enc []byte + return has } -// SetRawExtension is for testing only. -func SetRawExtension(base Message, id int32, b []byte) { - epb, err := extendable(base) - if err != nil { +// ClearExtension removes the extension field from m +// either as an explicitly populated field or as an unknown field. +func ClearExtension(m Message, xt *ExtensionDesc) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() { return } - extmap := epb.extensionsWrite() - extmap[id] = Extension{enc: b} -} -// isExtensionField returns true iff the given field number is in an extension range. -func isExtensionField(pb extendableProto, field int32) bool { - for _, er := range pb.ExtensionRangeArray() { - if er.Start <= field && field <= er.End { + xtd := xt.TypeDescriptor() + if isValidExtension(mr.Descriptor(), xtd) { + mr.Clear(xtd) + } else { + mr.Range(func(fd protoreflect.FieldDescriptor, _ protoreflect.Value) bool { + if int32(fd.Number()) == xt.Field { + mr.Clear(fd) + return false + } return true - } - } - return false -} - -// checkExtensionTypes checks that the given extension is valid for pb. -func checkExtensionTypes(pb extendableProto, extension *ExtensionDesc) error { - var pbi interface{} = pb - // Check the extended type. - if ea, ok := pbi.(extensionAdapter); ok { - pbi = ea.extendableProtoV1 - } - if a, b := reflect.TypeOf(pbi), reflect.TypeOf(extension.ExtendedType); a != b { - return fmt.Errorf("proto: bad extended type; %v does not extend %v", b, a) - } - // Check the range. - if !isExtensionField(pb, extension.Field) { - return errors.New("proto: bad extension number; not in declared ranges") - } - return nil -} - -// extPropKey is sufficient to uniquely identify an extension. -type extPropKey struct { - base reflect.Type - field int32 -} - -var extProp = struct { - sync.RWMutex - m map[extPropKey]*Properties -}{ - m: make(map[extPropKey]*Properties), -} - -func extensionProperties(ed *ExtensionDesc) *Properties { - key := extPropKey{base: reflect.TypeOf(ed.ExtendedType), field: ed.Field} - - extProp.RLock() - if prop, ok := extProp.m[key]; ok { - extProp.RUnlock() - return prop - } - extProp.RUnlock() - - extProp.Lock() - defer extProp.Unlock() - // Check again. - if prop, ok := extProp.m[key]; ok { - return prop - } - - prop := new(Properties) - prop.Init(reflect.TypeOf(ed.ExtensionType), "unknown_name", ed.Tag, nil) - extProp.m[key] = prop - return prop -} - -// HasExtension returns whether the given extension is present in pb. -func HasExtension(pb Message, extension *ExtensionDesc) bool { - // TODO: Check types, field numbers, etc.? - epb, err := extendable(pb) - if err != nil { - return false - } - extmap, mu := epb.extensionsRead() - if extmap == nil { - return false + }) } - mu.Lock() - _, ok := extmap[extension.Field] - mu.Unlock() - return ok + clearUnknown(mr, fieldNum(xt.Field)) } -// ClearExtension removes the given extension from pb. -func ClearExtension(pb Message, extension *ExtensionDesc) { - epb, err := extendable(pb) - if err != nil { +// ClearAllExtensions clears all extensions from m. +// This includes populated fields and unknown fields in the extension range. +func ClearAllExtensions(m Message) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() { return } - // TODO: Check types, field numbers, etc.? - extmap := epb.extensionsWrite() - delete(extmap, extension.Field) + + mr.Range(func(fd protoreflect.FieldDescriptor, _ protoreflect.Value) bool { + if fd.IsExtension() { + mr.Clear(fd) + } + return true + }) + clearUnknown(mr, mr.Descriptor().ExtensionRanges()) } -// GetExtension retrieves a proto2 extended field from pb. +// GetExtension retrieves a proto2 extended field from m. // // If the descriptor is type complete (i.e., ExtensionDesc.ExtensionType is non-nil), // then GetExtension parses the encoded field and returns a Go value of the specified type. // If the field is not present, then the default value is returned (if one is specified), // otherwise ErrMissingExtension is reported. // -// If the descriptor is not type complete (i.e., ExtensionDesc.ExtensionType is nil), -// then GetExtension returns the raw encoded bytes of the field extension. -func GetExtension(pb Message, extension *ExtensionDesc) (interface{}, error) { - epb, err := extendable(pb) - if err != nil { - return nil, err - } - - if extension.ExtendedType != nil { - // can only check type if this is a complete descriptor - if err := checkExtensionTypes(epb, extension); err != nil { - return nil, err +// If the descriptor is type incomplete (i.e., ExtensionDesc.ExtensionType is nil), +// then GetExtension returns the raw encoded bytes for the extension field. +func GetExtension(m Message, xt *ExtensionDesc) (interface{}, error) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() || mr.Descriptor().ExtensionRanges().Len() == 0 { + return nil, errNotExtendable + } + + // Retrieve the unknown fields for this extension field. + var bo protoreflect.RawFields + for bi := mr.GetUnknown(); len(bi) > 0; { + num, _, n := protowire.ConsumeField(bi) + if int32(num) == xt.Field { + bo = append(bo, bi[:n]...) } + bi = bi[n:] } - emap, mu := epb.extensionsRead() - if emap == nil { - return defaultExtensionValue(extension) - } - mu.Lock() - defer mu.Unlock() - e, ok := emap[extension.Field] - if !ok { - // defaultExtensionValue returns the default value or - // ErrMissingExtension if there is no default. - return defaultExtensionValue(extension) - } - - if e.value != nil { - // Already decoded. Check the descriptor, though. - if e.desc != extension { - // This shouldn't happen. If it does, it means that - // GetExtension was called twice with two different - // descriptors with the same field number. - return nil, errors.New("proto: descriptor conflict") - } - return extensionAsLegacyType(e.value), nil + // For type incomplete descriptors, only retrieve the unknown fields. + if xt.ExtensionType == nil { + return []byte(bo), nil } - if extension.ExtensionType == nil { - // incomplete descriptor - return e.enc, nil + // If the extension field only exists as unknown fields, unmarshal it. + // This is rarely done since proto.Unmarshal eagerly unmarshals extensions. + xtd := xt.TypeDescriptor() + if !isValidExtension(mr.Descriptor(), xtd) { + return nil, fmt.Errorf("proto: bad extended type; %T does not extend %T", xt.ExtendedType, m) } - - v, err := decodeExtension(e.enc, extension) - if err != nil { - return nil, err + if !mr.Has(xtd) && len(bo) > 0 { + m2 := mr.New() + if err := (proto.UnmarshalOptions{ + Resolver: extensionResolver{xt}, + }.Unmarshal(bo, m2.Interface())); err != nil { + return nil, err + } + if m2.Has(xtd) { + mr.Set(xtd, m2.Get(xtd)) + clearUnknown(mr, fieldNum(xt.Field)) + } } - // Remember the decoded version and drop the encoded version. - // That way it is safe to mutate what we return. - e.value = extensionAsStorageType(v) - e.desc = extension - e.enc = nil - emap[extension.Field] = e - return extensionAsLegacyType(e.value), nil -} - -// defaultExtensionValue returns the default value for extension. -// If no default for an extension is defined ErrMissingExtension is returned. -func defaultExtensionValue(extension *ExtensionDesc) (interface{}, error) { - if extension.ExtensionType == nil { - // incomplete descriptor, so no default + // Check whether the message has the extension field set or a default. + var pv protoreflect.Value + switch { + case mr.Has(xtd): + pv = mr.Get(xtd) + case xtd.HasDefault(): + pv = xtd.Default() + default: return nil, ErrMissingExtension } - t := reflect.TypeOf(extension.ExtensionType) - props := extensionProperties(extension) - - sf, _, err := fieldDefault(t, props) - if err != nil { - return nil, err - } - - if sf == nil || sf.value == nil { - // There is no default value. - return nil, ErrMissingExtension + v := xt.InterfaceOf(pv) + rv := reflect.ValueOf(v) + if isScalarKind(rv.Kind()) { + rv2 := reflect.New(rv.Type()) + rv2.Elem().Set(rv) + v = rv2.Interface() } + return v, nil +} - if t.Kind() != reflect.Ptr { - // We do not need to return a Ptr, we can directly return sf.value. - return sf.value, nil - } +// extensionResolver is a custom extension resolver that stores a single +// extension type that takes precedence over the global registry. +type extensionResolver struct{ xt protoreflect.ExtensionType } - // We need to return an interface{} that is a pointer to sf.value. - value := reflect.New(t).Elem() - value.Set(reflect.New(value.Type().Elem())) - if sf.kind == reflect.Int32 { - // We may have an int32 or an enum, but the underlying data is int32. - // Since we can't set an int32 into a non int32 reflect.value directly - // set it as a int32. - value.Elem().SetInt(int64(sf.value.(int32))) - } else { - value.Elem().Set(reflect.ValueOf(sf.value)) +func (r extensionResolver) FindExtensionByName(field protoreflect.FullName) (protoreflect.ExtensionType, error) { + if xtd := r.xt.TypeDescriptor(); xtd.FullName() == field { + return r.xt, nil } - return value.Interface(), nil + return protoregistry.GlobalTypes.FindExtensionByName(field) } -// decodeExtension decodes an extension encoded in b. -func decodeExtension(b []byte, extension *ExtensionDesc) (interface{}, error) { - t := reflect.TypeOf(extension.ExtensionType) - unmarshal := typeUnmarshaler(t, extension.Tag) - - // t is a pointer to a struct, pointer to basic type or a slice. - // Allocate space to store the pointer/slice. - value := reflect.New(t).Elem() - - var err error - for { - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - wire := int(x) & 7 - - b, err = unmarshal(b, valToPointer(value.Addr()), wire) - if err != nil { - return nil, err - } - - if len(b) == 0 { - break - } +func (r extensionResolver) FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) { + if xtd := r.xt.TypeDescriptor(); xtd.ContainingMessage().FullName() == message && xtd.Number() == field { + return r.xt, nil } - return value.Interface(), nil + return protoregistry.GlobalTypes.FindExtensionByNumber(message, field) } -// GetExtensions returns a slice of the extensions present in pb that are also listed in es. -// The returned slice has the same length as es; missing extensions will appear as nil elements. -func GetExtensions(pb Message, es []*ExtensionDesc) (extensions []interface{}, err error) { - epb, err := extendable(pb) - if err != nil { - return nil, err +// GetExtensions returns a list of the extensions values present in m, +// corresponding with the provided list of extension descriptors, xts. +// If an extension is missing in m, the corresponding value is nil. +func GetExtensions(m Message, xts []*ExtensionDesc) ([]interface{}, error) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() { + return nil, errNotExtendable } - extensions = make([]interface{}, len(es)) - for i, e := range es { - extensions[i], err = GetExtension(epb, e) - if err == ErrMissingExtension { - err = nil - } + + vs := make([]interface{}, len(xts)) + for i, xt := range xts { + v, err := GetExtension(m, xt) if err != nil { - return + if err == ErrMissingExtension { + continue + } + return vs, err } + vs[i] = v } - return + return vs, nil } -// ExtensionDescs returns a new slice containing pb's extension descriptors, in undefined order. -// For non-registered extensions, ExtensionDescs returns an incomplete descriptor containing -// just the Field field, which defines the extension's field number. -func ExtensionDescs(pb Message) ([]*ExtensionDesc, error) { - epb, err := extendable(pb) - if err != nil { - return nil, err +// SetExtension sets an extension field in m to the provided value. +func SetExtension(m Message, xt *ExtensionDesc, v interface{}) error { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() || mr.Descriptor().ExtensionRanges().Len() == 0 { + return errNotExtendable } - registeredExtensions := RegisteredExtensions(pb) - emap, mu := epb.extensionsRead() - if emap == nil { - return nil, nil + rv := reflect.ValueOf(v) + if reflect.TypeOf(v) != reflect.TypeOf(xt.ExtensionType) { + return fmt.Errorf("proto: bad extension value type. got: %T, want: %T", v, xt.ExtensionType) } - mu.Lock() - defer mu.Unlock() - extensions := make([]*ExtensionDesc, 0, len(emap)) - for extid, e := range emap { - desc := e.desc - if desc == nil { - desc = registeredExtensions[extid] - if desc == nil { - desc = &ExtensionDesc{Field: extid} - } + if rv.Kind() == reflect.Ptr { + if rv.IsNil() { + return fmt.Errorf("proto: SetExtension called with nil value of type %T", v) + } + if isScalarKind(rv.Elem().Kind()) { + v = rv.Elem().Interface() } - - extensions = append(extensions, desc) } - return extensions, nil -} -// SetExtension sets the specified extension of pb to the specified value. -func SetExtension(pb Message, extension *ExtensionDesc, value interface{}) error { - epb, err := extendable(pb) - if err != nil { - return err - } - if err := checkExtensionTypes(epb, extension); err != nil { - return err - } - typ := reflect.TypeOf(extension.ExtensionType) - if typ != reflect.TypeOf(value) { - return fmt.Errorf("proto: bad extension value type. got: %T, want: %T", value, extension.ExtensionType) + xtd := xt.TypeDescriptor() + if !isValidExtension(mr.Descriptor(), xtd) { + return fmt.Errorf("proto: bad extended type; %T does not extend %T", xt.ExtendedType, m) } - // nil extension values need to be caught early, because the - // encoder can't distinguish an ErrNil due to a nil extension - // from an ErrNil due to a missing field. Extensions are - // always optional, so the encoder would just swallow the error - // and drop all the extensions from the encoded message. - if reflect.ValueOf(value).IsNil() { - return fmt.Errorf("proto: SetExtension called with nil value of type %T", value) - } - - extmap := epb.extensionsWrite() - extmap[extension.Field] = Extension{desc: extension, value: extensionAsStorageType(value)} + mr.Set(xtd, xt.ValueOf(v)) + clearUnknown(mr, fieldNum(xt.Field)) return nil } -// ClearAllExtensions clears all extensions from pb. -func ClearAllExtensions(pb Message) { - epb, err := extendable(pb) - if err != nil { +// SetRawExtension inserts b into the unknown fields of m. +// +// Deprecated: Use Message.ProtoReflect.SetUnknown instead. +func SetRawExtension(m Message, fnum int32, b []byte) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() { return } - m := epb.extensionsWrite() - for k := range m { - delete(m, k) + + // Verify that the raw field is valid. + for b0 := b; len(b0) > 0; { + num, _, n := protowire.ConsumeField(b0) + if int32(num) != fnum { + panic(fmt.Sprintf("mismatching field number: got %d, want %d", num, fnum)) + } + b0 = b0[n:] } -} -// A global registry of extensions. -// The generated code will register the generated descriptors by calling RegisterExtension. + ClearExtension(m, &ExtensionDesc{Field: fnum}) + mr.SetUnknown(append(mr.GetUnknown(), b...)) +} -var extensionMaps = make(map[reflect.Type]map[int32]*ExtensionDesc) +// ExtensionDescs returns a list of extension descriptors found in m, +// containing descriptors for both populated extension fields in m and +// also unknown fields of m that are in the extension range. +// For the later case, an type incomplete descriptor is provided where only +// the ExtensionDesc.Field field is populated. +// The order of the extension descriptors is undefined. +func ExtensionDescs(m Message) ([]*ExtensionDesc, error) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() || mr.Descriptor().ExtensionRanges().Len() == 0 { + return nil, errNotExtendable + } -// RegisterExtension is called from the generated code. -func RegisterExtension(desc *ExtensionDesc) { - st := reflect.TypeOf(desc.ExtendedType).Elem() - m := extensionMaps[st] - if m == nil { - m = make(map[int32]*ExtensionDesc) - extensionMaps[st] = m + // Collect a set of known extension descriptors. + extDescs := make(map[protoreflect.FieldNumber]*ExtensionDesc) + mr.Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + if fd.IsExtension() { + xt := fd.(protoreflect.ExtensionTypeDescriptor) + if xd, ok := xt.Type().(*ExtensionDesc); ok { + extDescs[fd.Number()] = xd + } + } + return true + }) + + // Collect a set of unknown extension descriptors. + extRanges := mr.Descriptor().ExtensionRanges() + for b := mr.GetUnknown(); len(b) > 0; { + num, _, n := protowire.ConsumeField(b) + if extRanges.Has(num) && extDescs[num] == nil { + extDescs[num] = nil + } + b = b[n:] } - if _, ok := m[desc.Field]; ok { - panic("proto: duplicate extension registered: " + st.String() + " " + strconv.Itoa(int(desc.Field))) + + // Transpose the set of descriptors into a list. + var xts []*ExtensionDesc + for num, xt := range extDescs { + if xt == nil { + xt = &ExtensionDesc{Field: int32(num)} + } + xts = append(xts, xt) } - m[desc.Field] = desc + return xts, nil } -// RegisteredExtensions returns a map of the registered extensions of a -// protocol buffer struct, indexed by the extension number. -// The argument pb should be a nil pointer to the struct type. -func RegisteredExtensions(pb Message) map[int32]*ExtensionDesc { - return extensionMaps[reflect.TypeOf(pb).Elem()] +// isValidExtension reports whether xtd is a valid extension descriptor for md. +func isValidExtension(md protoreflect.MessageDescriptor, xtd protoreflect.ExtensionTypeDescriptor) bool { + return xtd.ContainingMessage() == md && md.ExtensionRanges().Has(xtd.Number()) } -// extensionAsLegacyType converts an value in the storage type as the API type. -// See Extension.value. -func extensionAsLegacyType(v interface{}) interface{} { - switch rv := reflect.ValueOf(v); rv.Kind() { +// isScalarKind reports whether k is a protobuf scalar kind (except bytes). +// This function exists for historical reasons since the representation of +// scalars differs between v1 and v2, where v1 uses *T and v2 uses T. +func isScalarKind(k reflect.Kind) bool { + switch k { case reflect.Bool, reflect.Int32, reflect.Int64, reflect.Uint32, reflect.Uint64, reflect.Float32, reflect.Float64, reflect.String: - // Represent primitive types as a pointer to the value. - rv2 := reflect.New(rv.Type()) - rv2.Elem().Set(rv) - v = rv2.Interface() - case reflect.Ptr: - // Represent slice types as the value itself. - switch rv.Type().Elem().Kind() { - case reflect.Slice: - if rv.IsNil() { - v = reflect.Zero(rv.Type().Elem()).Interface() - } else { - v = rv.Elem().Interface() - } - } + return true + default: + return false } - return v } -// extensionAsStorageType converts an value in the API type as the storage type. -// See Extension.value. -func extensionAsStorageType(v interface{}) interface{} { - switch rv := reflect.ValueOf(v); rv.Kind() { - case reflect.Ptr: - // Represent slice types as the value itself. - switch rv.Type().Elem().Kind() { - case reflect.Bool, reflect.Int32, reflect.Int64, reflect.Uint32, reflect.Uint64, reflect.Float32, reflect.Float64, reflect.String: - if rv.IsNil() { - v = reflect.Zero(rv.Type().Elem()).Interface() - } else { - v = rv.Elem().Interface() - } - } - case reflect.Slice: - // Represent slice types as a pointer to the value. - if rv.Type().Elem().Kind() != reflect.Uint8 { - rv2 := reflect.New(rv.Type()) - rv2.Elem().Set(rv) - v = rv2.Interface() +// clearUnknown removes unknown fields from m where remover.Has reports true. +func clearUnknown(m protoreflect.Message, remover interface { + Has(protoreflect.FieldNumber) bool +}) { + var bo protoreflect.RawFields + for bi := m.GetUnknown(); len(bi) > 0; { + num, _, n := protowire.ConsumeField(bi) + if !remover.Has(num) { + bo = append(bo, bi[:n]...) } + bi = bi[n:] } - return v + if bi := m.GetUnknown(); len(bi) != len(bo) { + m.SetUnknown(bo) + } +} + +type fieldNum protoreflect.FieldNumber + +func (n1 fieldNum) Has(n2 protoreflect.FieldNumber) bool { + return protoreflect.FieldNumber(n1) == n2 } diff --git a/vendor/github.com/golang/protobuf/proto/lib.go b/vendor/github.com/golang/protobuf/proto/lib.go deleted file mode 100644 index 70fbda53..00000000 --- a/vendor/github.com/golang/protobuf/proto/lib.go +++ /dev/null @@ -1,965 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -/* -Package proto converts data structures to and from the wire format of -protocol buffers. It works in concert with the Go source code generated -for .proto files by the protocol compiler. - -A summary of the properties of the protocol buffer interface -for a protocol buffer variable v: - - - Names are turned from camel_case to CamelCase for export. - - There are no methods on v to set fields; just treat - them as structure fields. - - There are getters that return a field's value if set, - and return the field's default value if unset. - The getters work even if the receiver is a nil message. - - The zero value for a struct is its correct initialization state. - All desired fields must be set before marshaling. - - A Reset() method will restore a protobuf struct to its zero state. - - Non-repeated fields are pointers to the values; nil means unset. - That is, optional or required field int32 f becomes F *int32. - - Repeated fields are slices. - - Helper functions are available to aid the setting of fields. - msg.Foo = proto.String("hello") // set field - - Constants are defined to hold the default values of all fields that - have them. They have the form Default_StructName_FieldName. - Because the getter methods handle defaulted values, - direct use of these constants should be rare. - - Enums are given type names and maps from names to values. - Enum values are prefixed by the enclosing message's name, or by the - enum's type name if it is a top-level enum. Enum types have a String - method, and a Enum method to assist in message construction. - - Nested messages, groups and enums have type names prefixed with the name of - the surrounding message type. - - Extensions are given descriptor names that start with E_, - followed by an underscore-delimited list of the nested messages - that contain it (if any) followed by the CamelCased name of the - extension field itself. HasExtension, ClearExtension, GetExtension - and SetExtension are functions for manipulating extensions. - - Oneof field sets are given a single field in their message, - with distinguished wrapper types for each possible field value. - - Marshal and Unmarshal are functions to encode and decode the wire format. - -When the .proto file specifies `syntax="proto3"`, there are some differences: - - - Non-repeated fields of non-message type are values instead of pointers. - - Enum types do not get an Enum method. - -The simplest way to describe this is to see an example. -Given file test.proto, containing - - package example; - - enum FOO { X = 17; } - - message Test { - required string label = 1; - optional int32 type = 2 [default=77]; - repeated int64 reps = 3; - optional group OptionalGroup = 4 { - required string RequiredField = 5; - } - oneof union { - int32 number = 6; - string name = 7; - } - } - -The resulting file, test.pb.go, is: - - package example - - import proto "github.com/golang/protobuf/proto" - import math "math" - - type FOO int32 - const ( - FOO_X FOO = 17 - ) - var FOO_name = map[int32]string{ - 17: "X", - } - var FOO_value = map[string]int32{ - "X": 17, - } - - func (x FOO) Enum() *FOO { - p := new(FOO) - *p = x - return p - } - func (x FOO) String() string { - return proto.EnumName(FOO_name, int32(x)) - } - func (x *FOO) UnmarshalJSON(data []byte) error { - value, err := proto.UnmarshalJSONEnum(FOO_value, data) - if err != nil { - return err - } - *x = FOO(value) - return nil - } - - type Test struct { - Label *string `protobuf:"bytes,1,req,name=label" json:"label,omitempty"` - Type *int32 `protobuf:"varint,2,opt,name=type,def=77" json:"type,omitempty"` - Reps []int64 `protobuf:"varint,3,rep,name=reps" json:"reps,omitempty"` - Optionalgroup *Test_OptionalGroup `protobuf:"group,4,opt,name=OptionalGroup" json:"optionalgroup,omitempty"` - // Types that are valid to be assigned to Union: - // *Test_Number - // *Test_Name - Union isTest_Union `protobuf_oneof:"union"` - XXX_unrecognized []byte `json:"-"` - } - func (m *Test) Reset() { *m = Test{} } - func (m *Test) String() string { return proto.CompactTextString(m) } - func (*Test) ProtoMessage() {} - - type isTest_Union interface { - isTest_Union() - } - - type Test_Number struct { - Number int32 `protobuf:"varint,6,opt,name=number"` - } - type Test_Name struct { - Name string `protobuf:"bytes,7,opt,name=name"` - } - - func (*Test_Number) isTest_Union() {} - func (*Test_Name) isTest_Union() {} - - func (m *Test) GetUnion() isTest_Union { - if m != nil { - return m.Union - } - return nil - } - const Default_Test_Type int32 = 77 - - func (m *Test) GetLabel() string { - if m != nil && m.Label != nil { - return *m.Label - } - return "" - } - - func (m *Test) GetType() int32 { - if m != nil && m.Type != nil { - return *m.Type - } - return Default_Test_Type - } - - func (m *Test) GetOptionalgroup() *Test_OptionalGroup { - if m != nil { - return m.Optionalgroup - } - return nil - } - - type Test_OptionalGroup struct { - RequiredField *string `protobuf:"bytes,5,req" json:"RequiredField,omitempty"` - } - func (m *Test_OptionalGroup) Reset() { *m = Test_OptionalGroup{} } - func (m *Test_OptionalGroup) String() string { return proto.CompactTextString(m) } - - func (m *Test_OptionalGroup) GetRequiredField() string { - if m != nil && m.RequiredField != nil { - return *m.RequiredField - } - return "" - } - - func (m *Test) GetNumber() int32 { - if x, ok := m.GetUnion().(*Test_Number); ok { - return x.Number - } - return 0 - } - - func (m *Test) GetName() string { - if x, ok := m.GetUnion().(*Test_Name); ok { - return x.Name - } - return "" - } - - func init() { - proto.RegisterEnum("example.FOO", FOO_name, FOO_value) - } - -To create and play with a Test object: - - package main - - import ( - "log" - - "github.com/golang/protobuf/proto" - pb "./example.pb" - ) - - func main() { - test := &pb.Test{ - Label: proto.String("hello"), - Type: proto.Int32(17), - Reps: []int64{1, 2, 3}, - Optionalgroup: &pb.Test_OptionalGroup{ - RequiredField: proto.String("good bye"), - }, - Union: &pb.Test_Name{"fred"}, - } - data, err := proto.Marshal(test) - if err != nil { - log.Fatal("marshaling error: ", err) - } - newTest := &pb.Test{} - err = proto.Unmarshal(data, newTest) - if err != nil { - log.Fatal("unmarshaling error: ", err) - } - // Now test and newTest contain the same data. - if test.GetLabel() != newTest.GetLabel() { - log.Fatalf("data mismatch %q != %q", test.GetLabel(), newTest.GetLabel()) - } - // Use a type switch to determine which oneof was set. - switch u := test.Union.(type) { - case *pb.Test_Number: // u.Number contains the number. - case *pb.Test_Name: // u.Name contains the string. - } - // etc. - } -*/ -package proto - -import ( - "encoding/json" - "fmt" - "log" - "reflect" - "sort" - "strconv" - "sync" -) - -// RequiredNotSetError is an error type returned by either Marshal or Unmarshal. -// Marshal reports this when a required field is not initialized. -// Unmarshal reports this when a required field is missing from the wire data. -type RequiredNotSetError struct{ field string } - -func (e *RequiredNotSetError) Error() string { - if e.field == "" { - return fmt.Sprintf("proto: required field not set") - } - return fmt.Sprintf("proto: required field %q not set", e.field) -} -func (e *RequiredNotSetError) RequiredNotSet() bool { - return true -} - -type invalidUTF8Error struct{ field string } - -func (e *invalidUTF8Error) Error() string { - if e.field == "" { - return "proto: invalid UTF-8 detected" - } - return fmt.Sprintf("proto: field %q contains invalid UTF-8", e.field) -} -func (e *invalidUTF8Error) InvalidUTF8() bool { - return true -} - -// errInvalidUTF8 is a sentinel error to identify fields with invalid UTF-8. -// This error should not be exposed to the external API as such errors should -// be recreated with the field information. -var errInvalidUTF8 = &invalidUTF8Error{} - -// isNonFatal reports whether the error is either a RequiredNotSet error -// or a InvalidUTF8 error. -func isNonFatal(err error) bool { - if re, ok := err.(interface{ RequiredNotSet() bool }); ok && re.RequiredNotSet() { - return true - } - if re, ok := err.(interface{ InvalidUTF8() bool }); ok && re.InvalidUTF8() { - return true - } - return false -} - -type nonFatal struct{ E error } - -// Merge merges err into nf and reports whether it was successful. -// Otherwise it returns false for any fatal non-nil errors. -func (nf *nonFatal) Merge(err error) (ok bool) { - if err == nil { - return true // not an error - } - if !isNonFatal(err) { - return false // fatal error - } - if nf.E == nil { - nf.E = err // store first instance of non-fatal error - } - return true -} - -// Message is implemented by generated protocol buffer messages. -type Message interface { - Reset() - String() string - ProtoMessage() -} - -// A Buffer is a buffer manager for marshaling and unmarshaling -// protocol buffers. It may be reused between invocations to -// reduce memory usage. It is not necessary to use a Buffer; -// the global functions Marshal and Unmarshal create a -// temporary Buffer and are fine for most applications. -type Buffer struct { - buf []byte // encode/decode byte stream - index int // read point - - deterministic bool -} - -// NewBuffer allocates a new Buffer and initializes its internal data to -// the contents of the argument slice. -func NewBuffer(e []byte) *Buffer { - return &Buffer{buf: e} -} - -// Reset resets the Buffer, ready for marshaling a new protocol buffer. -func (p *Buffer) Reset() { - p.buf = p.buf[0:0] // for reading/writing - p.index = 0 // for reading -} - -// SetBuf replaces the internal buffer with the slice, -// ready for unmarshaling the contents of the slice. -func (p *Buffer) SetBuf(s []byte) { - p.buf = s - p.index = 0 -} - -// Bytes returns the contents of the Buffer. -func (p *Buffer) Bytes() []byte { return p.buf } - -// SetDeterministic sets whether to use deterministic serialization. -// -// Deterministic serialization guarantees that for a given binary, equal -// messages will always be serialized to the same bytes. This implies: -// -// - Repeated serialization of a message will return the same bytes. -// - Different processes of the same binary (which may be executing on -// different machines) will serialize equal messages to the same bytes. -// -// Note that the deterministic serialization is NOT canonical across -// languages. It is not guaranteed to remain stable over time. It is unstable -// across different builds with schema changes due to unknown fields. -// Users who need canonical serialization (e.g., persistent storage in a -// canonical form, fingerprinting, etc.) should define their own -// canonicalization specification and implement their own serializer rather -// than relying on this API. -// -// If deterministic serialization is requested, map entries will be sorted -// by keys in lexicographical order. This is an implementation detail and -// subject to change. -func (p *Buffer) SetDeterministic(deterministic bool) { - p.deterministic = deterministic -} - -/* - * Helper routines for simplifying the creation of optional fields of basic type. - */ - -// Bool is a helper routine that allocates a new bool value -// to store v and returns a pointer to it. -func Bool(v bool) *bool { - return &v -} - -// Int32 is a helper routine that allocates a new int32 value -// to store v and returns a pointer to it. -func Int32(v int32) *int32 { - return &v -} - -// Int is a helper routine that allocates a new int32 value -// to store v and returns a pointer to it, but unlike Int32 -// its argument value is an int. -func Int(v int) *int32 { - p := new(int32) - *p = int32(v) - return p -} - -// Int64 is a helper routine that allocates a new int64 value -// to store v and returns a pointer to it. -func Int64(v int64) *int64 { - return &v -} - -// Float32 is a helper routine that allocates a new float32 value -// to store v and returns a pointer to it. -func Float32(v float32) *float32 { - return &v -} - -// Float64 is a helper routine that allocates a new float64 value -// to store v and returns a pointer to it. -func Float64(v float64) *float64 { - return &v -} - -// Uint32 is a helper routine that allocates a new uint32 value -// to store v and returns a pointer to it. -func Uint32(v uint32) *uint32 { - return &v -} - -// Uint64 is a helper routine that allocates a new uint64 value -// to store v and returns a pointer to it. -func Uint64(v uint64) *uint64 { - return &v -} - -// String is a helper routine that allocates a new string value -// to store v and returns a pointer to it. -func String(v string) *string { - return &v -} - -// EnumName is a helper function to simplify printing protocol buffer enums -// by name. Given an enum map and a value, it returns a useful string. -func EnumName(m map[int32]string, v int32) string { - s, ok := m[v] - if ok { - return s - } - return strconv.Itoa(int(v)) -} - -// UnmarshalJSONEnum is a helper function to simplify recovering enum int values -// from their JSON-encoded representation. Given a map from the enum's symbolic -// names to its int values, and a byte buffer containing the JSON-encoded -// value, it returns an int32 that can be cast to the enum type by the caller. -// -// The function can deal with both JSON representations, numeric and symbolic. -func UnmarshalJSONEnum(m map[string]int32, data []byte, enumName string) (int32, error) { - if data[0] == '"' { - // New style: enums are strings. - var repr string - if err := json.Unmarshal(data, &repr); err != nil { - return -1, err - } - val, ok := m[repr] - if !ok { - return 0, fmt.Errorf("unrecognized enum %s value %q", enumName, repr) - } - return val, nil - } - // Old style: enums are ints. - var val int32 - if err := json.Unmarshal(data, &val); err != nil { - return 0, fmt.Errorf("cannot unmarshal %#q into enum %s", data, enumName) - } - return val, nil -} - -// DebugPrint dumps the encoded data in b in a debugging format with a header -// including the string s. Used in testing but made available for general debugging. -func (p *Buffer) DebugPrint(s string, b []byte) { - var u uint64 - - obuf := p.buf - index := p.index - p.buf = b - p.index = 0 - depth := 0 - - fmt.Printf("\n--- %s ---\n", s) - -out: - for { - for i := 0; i < depth; i++ { - fmt.Print(" ") - } - - index := p.index - if index == len(p.buf) { - break - } - - op, err := p.DecodeVarint() - if err != nil { - fmt.Printf("%3d: fetching op err %v\n", index, err) - break out - } - tag := op >> 3 - wire := op & 7 - - switch wire { - default: - fmt.Printf("%3d: t=%3d unknown wire=%d\n", - index, tag, wire) - break out - - case WireBytes: - var r []byte - - r, err = p.DecodeRawBytes(false) - if err != nil { - break out - } - fmt.Printf("%3d: t=%3d bytes [%d]", index, tag, len(r)) - if len(r) <= 6 { - for i := 0; i < len(r); i++ { - fmt.Printf(" %.2x", r[i]) - } - } else { - for i := 0; i < 3; i++ { - fmt.Printf(" %.2x", r[i]) - } - fmt.Printf(" ..") - for i := len(r) - 3; i < len(r); i++ { - fmt.Printf(" %.2x", r[i]) - } - } - fmt.Printf("\n") - - case WireFixed32: - u, err = p.DecodeFixed32() - if err != nil { - fmt.Printf("%3d: t=%3d fix32 err %v\n", index, tag, err) - break out - } - fmt.Printf("%3d: t=%3d fix32 %d\n", index, tag, u) - - case WireFixed64: - u, err = p.DecodeFixed64() - if err != nil { - fmt.Printf("%3d: t=%3d fix64 err %v\n", index, tag, err) - break out - } - fmt.Printf("%3d: t=%3d fix64 %d\n", index, tag, u) - - case WireVarint: - u, err = p.DecodeVarint() - if err != nil { - fmt.Printf("%3d: t=%3d varint err %v\n", index, tag, err) - break out - } - fmt.Printf("%3d: t=%3d varint %d\n", index, tag, u) - - case WireStartGroup: - fmt.Printf("%3d: t=%3d start\n", index, tag) - depth++ - - case WireEndGroup: - depth-- - fmt.Printf("%3d: t=%3d end\n", index, tag) - } - } - - if depth != 0 { - fmt.Printf("%3d: start-end not balanced %d\n", p.index, depth) - } - fmt.Printf("\n") - - p.buf = obuf - p.index = index -} - -// SetDefaults sets unset protocol buffer fields to their default values. -// It only modifies fields that are both unset and have defined defaults. -// It recursively sets default values in any non-nil sub-messages. -func SetDefaults(pb Message) { - setDefaults(reflect.ValueOf(pb), true, false) -} - -// v is a pointer to a struct. -func setDefaults(v reflect.Value, recur, zeros bool) { - v = v.Elem() - - defaultMu.RLock() - dm, ok := defaults[v.Type()] - defaultMu.RUnlock() - if !ok { - dm = buildDefaultMessage(v.Type()) - defaultMu.Lock() - defaults[v.Type()] = dm - defaultMu.Unlock() - } - - for _, sf := range dm.scalars { - f := v.Field(sf.index) - if !f.IsNil() { - // field already set - continue - } - dv := sf.value - if dv == nil && !zeros { - // no explicit default, and don't want to set zeros - continue - } - fptr := f.Addr().Interface() // **T - // TODO: Consider batching the allocations we do here. - switch sf.kind { - case reflect.Bool: - b := new(bool) - if dv != nil { - *b = dv.(bool) - } - *(fptr.(**bool)) = b - case reflect.Float32: - f := new(float32) - if dv != nil { - *f = dv.(float32) - } - *(fptr.(**float32)) = f - case reflect.Float64: - f := new(float64) - if dv != nil { - *f = dv.(float64) - } - *(fptr.(**float64)) = f - case reflect.Int32: - // might be an enum - if ft := f.Type(); ft != int32PtrType { - // enum - f.Set(reflect.New(ft.Elem())) - if dv != nil { - f.Elem().SetInt(int64(dv.(int32))) - } - } else { - // int32 field - i := new(int32) - if dv != nil { - *i = dv.(int32) - } - *(fptr.(**int32)) = i - } - case reflect.Int64: - i := new(int64) - if dv != nil { - *i = dv.(int64) - } - *(fptr.(**int64)) = i - case reflect.String: - s := new(string) - if dv != nil { - *s = dv.(string) - } - *(fptr.(**string)) = s - case reflect.Uint8: - // exceptional case: []byte - var b []byte - if dv != nil { - db := dv.([]byte) - b = make([]byte, len(db)) - copy(b, db) - } else { - b = []byte{} - } - *(fptr.(*[]byte)) = b - case reflect.Uint32: - u := new(uint32) - if dv != nil { - *u = dv.(uint32) - } - *(fptr.(**uint32)) = u - case reflect.Uint64: - u := new(uint64) - if dv != nil { - *u = dv.(uint64) - } - *(fptr.(**uint64)) = u - default: - log.Printf("proto: can't set default for field %v (sf.kind=%v)", f, sf.kind) - } - } - - for _, ni := range dm.nested { - f := v.Field(ni) - // f is *T or []*T or map[T]*T - switch f.Kind() { - case reflect.Ptr: - if f.IsNil() { - continue - } - setDefaults(f, recur, zeros) - - case reflect.Slice: - for i := 0; i < f.Len(); i++ { - e := f.Index(i) - if e.IsNil() { - continue - } - setDefaults(e, recur, zeros) - } - - case reflect.Map: - for _, k := range f.MapKeys() { - e := f.MapIndex(k) - if e.IsNil() { - continue - } - setDefaults(e, recur, zeros) - } - } - } -} - -var ( - // defaults maps a protocol buffer struct type to a slice of the fields, - // with its scalar fields set to their proto-declared non-zero default values. - defaultMu sync.RWMutex - defaults = make(map[reflect.Type]defaultMessage) - - int32PtrType = reflect.TypeOf((*int32)(nil)) -) - -// defaultMessage represents information about the default values of a message. -type defaultMessage struct { - scalars []scalarField - nested []int // struct field index of nested messages -} - -type scalarField struct { - index int // struct field index - kind reflect.Kind // element type (the T in *T or []T) - value interface{} // the proto-declared default value, or nil -} - -// t is a struct type. -func buildDefaultMessage(t reflect.Type) (dm defaultMessage) { - sprop := GetProperties(t) - for _, prop := range sprop.Prop { - fi, ok := sprop.decoderTags.get(prop.Tag) - if !ok { - // XXX_unrecognized - continue - } - ft := t.Field(fi).Type - - sf, nested, err := fieldDefault(ft, prop) - switch { - case err != nil: - log.Print(err) - case nested: - dm.nested = append(dm.nested, fi) - case sf != nil: - sf.index = fi - dm.scalars = append(dm.scalars, *sf) - } - } - - return dm -} - -// fieldDefault returns the scalarField for field type ft. -// sf will be nil if the field can not have a default. -// nestedMessage will be true if this is a nested message. -// Note that sf.index is not set on return. -func fieldDefault(ft reflect.Type, prop *Properties) (sf *scalarField, nestedMessage bool, err error) { - var canHaveDefault bool - switch ft.Kind() { - case reflect.Ptr: - if ft.Elem().Kind() == reflect.Struct { - nestedMessage = true - } else { - canHaveDefault = true // proto2 scalar field - } - - case reflect.Slice: - switch ft.Elem().Kind() { - case reflect.Ptr: - nestedMessage = true // repeated message - case reflect.Uint8: - canHaveDefault = true // bytes field - } - - case reflect.Map: - if ft.Elem().Kind() == reflect.Ptr { - nestedMessage = true // map with message values - } - } - - if !canHaveDefault { - if nestedMessage { - return nil, true, nil - } - return nil, false, nil - } - - // We now know that ft is a pointer or slice. - sf = &scalarField{kind: ft.Elem().Kind()} - - // scalar fields without defaults - if !prop.HasDefault { - return sf, false, nil - } - - // a scalar field: either *T or []byte - switch ft.Elem().Kind() { - case reflect.Bool: - x, err := strconv.ParseBool(prop.Default) - if err != nil { - return nil, false, fmt.Errorf("proto: bad default bool %q: %v", prop.Default, err) - } - sf.value = x - case reflect.Float32: - x, err := strconv.ParseFloat(prop.Default, 32) - if err != nil { - return nil, false, fmt.Errorf("proto: bad default float32 %q: %v", prop.Default, err) - } - sf.value = float32(x) - case reflect.Float64: - x, err := strconv.ParseFloat(prop.Default, 64) - if err != nil { - return nil, false, fmt.Errorf("proto: bad default float64 %q: %v", prop.Default, err) - } - sf.value = x - case reflect.Int32: - x, err := strconv.ParseInt(prop.Default, 10, 32) - if err != nil { - return nil, false, fmt.Errorf("proto: bad default int32 %q: %v", prop.Default, err) - } - sf.value = int32(x) - case reflect.Int64: - x, err := strconv.ParseInt(prop.Default, 10, 64) - if err != nil { - return nil, false, fmt.Errorf("proto: bad default int64 %q: %v", prop.Default, err) - } - sf.value = x - case reflect.String: - sf.value = prop.Default - case reflect.Uint8: - // []byte (not *uint8) - sf.value = []byte(prop.Default) - case reflect.Uint32: - x, err := strconv.ParseUint(prop.Default, 10, 32) - if err != nil { - return nil, false, fmt.Errorf("proto: bad default uint32 %q: %v", prop.Default, err) - } - sf.value = uint32(x) - case reflect.Uint64: - x, err := strconv.ParseUint(prop.Default, 10, 64) - if err != nil { - return nil, false, fmt.Errorf("proto: bad default uint64 %q: %v", prop.Default, err) - } - sf.value = x - default: - return nil, false, fmt.Errorf("proto: unhandled def kind %v", ft.Elem().Kind()) - } - - return sf, false, nil -} - -// mapKeys returns a sort.Interface to be used for sorting the map keys. -// Map fields may have key types of non-float scalars, strings and enums. -func mapKeys(vs []reflect.Value) sort.Interface { - s := mapKeySorter{vs: vs} - - // Type specialization per https://developers.google.com/protocol-buffers/docs/proto#maps. - if len(vs) == 0 { - return s - } - switch vs[0].Kind() { - case reflect.Int32, reflect.Int64: - s.less = func(a, b reflect.Value) bool { return a.Int() < b.Int() } - case reflect.Uint32, reflect.Uint64: - s.less = func(a, b reflect.Value) bool { return a.Uint() < b.Uint() } - case reflect.Bool: - s.less = func(a, b reflect.Value) bool { return !a.Bool() && b.Bool() } // false < true - case reflect.String: - s.less = func(a, b reflect.Value) bool { return a.String() < b.String() } - default: - panic(fmt.Sprintf("unsupported map key type: %v", vs[0].Kind())) - } - - return s -} - -type mapKeySorter struct { - vs []reflect.Value - less func(a, b reflect.Value) bool -} - -func (s mapKeySorter) Len() int { return len(s.vs) } -func (s mapKeySorter) Swap(i, j int) { s.vs[i], s.vs[j] = s.vs[j], s.vs[i] } -func (s mapKeySorter) Less(i, j int) bool { - return s.less(s.vs[i], s.vs[j]) -} - -// isProto3Zero reports whether v is a zero proto3 value. -func isProto3Zero(v reflect.Value) bool { - switch v.Kind() { - case reflect.Bool: - return !v.Bool() - case reflect.Int32, reflect.Int64: - return v.Int() == 0 - case reflect.Uint32, reflect.Uint64: - return v.Uint() == 0 - case reflect.Float32, reflect.Float64: - return v.Float() == 0 - case reflect.String: - return v.String() == "" - } - return false -} - -const ( - // ProtoPackageIsVersion3 is referenced from generated protocol buffer files - // to assert that that code is compatible with this version of the proto package. - ProtoPackageIsVersion3 = true - - // ProtoPackageIsVersion2 is referenced from generated protocol buffer files - // to assert that that code is compatible with this version of the proto package. - ProtoPackageIsVersion2 = true - - // ProtoPackageIsVersion1 is referenced from generated protocol buffer files - // to assert that that code is compatible with this version of the proto package. - ProtoPackageIsVersion1 = true -) - -// InternalMessageInfo is a type used internally by generated .pb.go files. -// This type is not intended to be used by non-generated code. -// This type is not subject to any compatibility guarantee. -type InternalMessageInfo struct { - marshal *marshalInfo - unmarshal *unmarshalInfo - merge *mergeInfo - discard *discardInfo -} diff --git a/vendor/github.com/golang/protobuf/proto/message_set.go b/vendor/github.com/golang/protobuf/proto/message_set.go deleted file mode 100644 index f48a7567..00000000 --- a/vendor/github.com/golang/protobuf/proto/message_set.go +++ /dev/null @@ -1,181 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -/* - * Support for message sets. - */ - -import ( - "errors" -) - -// errNoMessageTypeID occurs when a protocol buffer does not have a message type ID. -// A message type ID is required for storing a protocol buffer in a message set. -var errNoMessageTypeID = errors.New("proto does not have a message type ID") - -// The first two types (_MessageSet_Item and messageSet) -// model what the protocol compiler produces for the following protocol message: -// message MessageSet { -// repeated group Item = 1 { -// required int32 type_id = 2; -// required string message = 3; -// }; -// } -// That is the MessageSet wire format. We can't use a proto to generate these -// because that would introduce a circular dependency between it and this package. - -type _MessageSet_Item struct { - TypeId *int32 `protobuf:"varint,2,req,name=type_id"` - Message []byte `protobuf:"bytes,3,req,name=message"` -} - -type messageSet struct { - Item []*_MessageSet_Item `protobuf:"group,1,rep"` - XXX_unrecognized []byte - // TODO: caching? -} - -// Make sure messageSet is a Message. -var _ Message = (*messageSet)(nil) - -// messageTypeIder is an interface satisfied by a protocol buffer type -// that may be stored in a MessageSet. -type messageTypeIder interface { - MessageTypeId() int32 -} - -func (ms *messageSet) find(pb Message) *_MessageSet_Item { - mti, ok := pb.(messageTypeIder) - if !ok { - return nil - } - id := mti.MessageTypeId() - for _, item := range ms.Item { - if *item.TypeId == id { - return item - } - } - return nil -} - -func (ms *messageSet) Has(pb Message) bool { - return ms.find(pb) != nil -} - -func (ms *messageSet) Unmarshal(pb Message) error { - if item := ms.find(pb); item != nil { - return Unmarshal(item.Message, pb) - } - if _, ok := pb.(messageTypeIder); !ok { - return errNoMessageTypeID - } - return nil // TODO: return error instead? -} - -func (ms *messageSet) Marshal(pb Message) error { - msg, err := Marshal(pb) - if err != nil { - return err - } - if item := ms.find(pb); item != nil { - // reuse existing item - item.Message = msg - return nil - } - - mti, ok := pb.(messageTypeIder) - if !ok { - return errNoMessageTypeID - } - - mtid := mti.MessageTypeId() - ms.Item = append(ms.Item, &_MessageSet_Item{ - TypeId: &mtid, - Message: msg, - }) - return nil -} - -func (ms *messageSet) Reset() { *ms = messageSet{} } -func (ms *messageSet) String() string { return CompactTextString(ms) } -func (*messageSet) ProtoMessage() {} - -// Support for the message_set_wire_format message option. - -func skipVarint(buf []byte) []byte { - i := 0 - for ; buf[i]&0x80 != 0; i++ { - } - return buf[i+1:] -} - -// unmarshalMessageSet decodes the extension map encoded in buf in the message set wire format. -// It is called by Unmarshal methods on protocol buffer messages with the message_set_wire_format option. -func unmarshalMessageSet(buf []byte, exts interface{}) error { - var m map[int32]Extension - switch exts := exts.(type) { - case *XXX_InternalExtensions: - m = exts.extensionsWrite() - case map[int32]Extension: - m = exts - default: - return errors.New("proto: not an extension map") - } - - ms := new(messageSet) - if err := Unmarshal(buf, ms); err != nil { - return err - } - for _, item := range ms.Item { - id := *item.TypeId - msg := item.Message - - // Restore wire type and field number varint, plus length varint. - // Be careful to preserve duplicate items. - b := EncodeVarint(uint64(id)<<3 | WireBytes) - if ext, ok := m[id]; ok { - // Existing data; rip off the tag and length varint - // so we join the new data correctly. - // We can assume that ext.enc is set because we are unmarshaling. - o := ext.enc[len(b):] // skip wire type and field number - _, n := DecodeVarint(o) // calculate length of length varint - o = o[n:] // skip length varint - msg = append(o, msg...) // join old data and new data - } - b = append(b, EncodeVarint(uint64(len(msg)))...) - b = append(b, msg...) - - m[id] = Extension{enc: b} - } - return nil -} diff --git a/vendor/github.com/golang/protobuf/proto/pointer_reflect.go b/vendor/github.com/golang/protobuf/proto/pointer_reflect.go deleted file mode 100644 index 94fa9194..00000000 --- a/vendor/github.com/golang/protobuf/proto/pointer_reflect.go +++ /dev/null @@ -1,360 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2012 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -// +build purego appengine js - -// This file contains an implementation of proto field accesses using package reflect. -// It is slower than the code in pointer_unsafe.go but it avoids package unsafe and can -// be used on App Engine. - -package proto - -import ( - "reflect" - "sync" -) - -const unsafeAllowed = false - -// A field identifies a field in a struct, accessible from a pointer. -// In this implementation, a field is identified by the sequence of field indices -// passed to reflect's FieldByIndex. -type field []int - -// toField returns a field equivalent to the given reflect field. -func toField(f *reflect.StructField) field { - return f.Index -} - -// invalidField is an invalid field identifier. -var invalidField = field(nil) - -// zeroField is a noop when calling pointer.offset. -var zeroField = field([]int{}) - -// IsValid reports whether the field identifier is valid. -func (f field) IsValid() bool { return f != nil } - -// The pointer type is for the table-driven decoder. -// The implementation here uses a reflect.Value of pointer type to -// create a generic pointer. In pointer_unsafe.go we use unsafe -// instead of reflect to implement the same (but faster) interface. -type pointer struct { - v reflect.Value -} - -// toPointer converts an interface of pointer type to a pointer -// that points to the same target. -func toPointer(i *Message) pointer { - return pointer{v: reflect.ValueOf(*i)} -} - -// toAddrPointer converts an interface to a pointer that points to -// the interface data. -func toAddrPointer(i *interface{}, isptr, deref bool) pointer { - v := reflect.ValueOf(*i) - u := reflect.New(v.Type()) - u.Elem().Set(v) - if deref { - u = u.Elem() - } - return pointer{v: u} -} - -// valToPointer converts v to a pointer. v must be of pointer type. -func valToPointer(v reflect.Value) pointer { - return pointer{v: v} -} - -// offset converts from a pointer to a structure to a pointer to -// one of its fields. -func (p pointer) offset(f field) pointer { - return pointer{v: p.v.Elem().FieldByIndex(f).Addr()} -} - -func (p pointer) isNil() bool { - return p.v.IsNil() -} - -// grow updates the slice s in place to make it one element longer. -// s must be addressable. -// Returns the (addressable) new element. -func grow(s reflect.Value) reflect.Value { - n, m := s.Len(), s.Cap() - if n < m { - s.SetLen(n + 1) - } else { - s.Set(reflect.Append(s, reflect.Zero(s.Type().Elem()))) - } - return s.Index(n) -} - -func (p pointer) toInt64() *int64 { - return p.v.Interface().(*int64) -} -func (p pointer) toInt64Ptr() **int64 { - return p.v.Interface().(**int64) -} -func (p pointer) toInt64Slice() *[]int64 { - return p.v.Interface().(*[]int64) -} - -var int32ptr = reflect.TypeOf((*int32)(nil)) - -func (p pointer) toInt32() *int32 { - return p.v.Convert(int32ptr).Interface().(*int32) -} - -// The toInt32Ptr/Slice methods don't work because of enums. -// Instead, we must use set/get methods for the int32ptr/slice case. -/* - func (p pointer) toInt32Ptr() **int32 { - return p.v.Interface().(**int32) -} - func (p pointer) toInt32Slice() *[]int32 { - return p.v.Interface().(*[]int32) -} -*/ -func (p pointer) getInt32Ptr() *int32 { - if p.v.Type().Elem().Elem() == reflect.TypeOf(int32(0)) { - // raw int32 type - return p.v.Elem().Interface().(*int32) - } - // an enum - return p.v.Elem().Convert(int32PtrType).Interface().(*int32) -} -func (p pointer) setInt32Ptr(v int32) { - // Allocate value in a *int32. Possibly convert that to a *enum. - // Then assign it to a **int32 or **enum. - // Note: we can convert *int32 to *enum, but we can't convert - // **int32 to **enum! - p.v.Elem().Set(reflect.ValueOf(&v).Convert(p.v.Type().Elem())) -} - -// getInt32Slice copies []int32 from p as a new slice. -// This behavior differs from the implementation in pointer_unsafe.go. -func (p pointer) getInt32Slice() []int32 { - if p.v.Type().Elem().Elem() == reflect.TypeOf(int32(0)) { - // raw int32 type - return p.v.Elem().Interface().([]int32) - } - // an enum - // Allocate a []int32, then assign []enum's values into it. - // Note: we can't convert []enum to []int32. - slice := p.v.Elem() - s := make([]int32, slice.Len()) - for i := 0; i < slice.Len(); i++ { - s[i] = int32(slice.Index(i).Int()) - } - return s -} - -// setInt32Slice copies []int32 into p as a new slice. -// This behavior differs from the implementation in pointer_unsafe.go. -func (p pointer) setInt32Slice(v []int32) { - if p.v.Type().Elem().Elem() == reflect.TypeOf(int32(0)) { - // raw int32 type - p.v.Elem().Set(reflect.ValueOf(v)) - return - } - // an enum - // Allocate a []enum, then assign []int32's values into it. - // Note: we can't convert []enum to []int32. - slice := reflect.MakeSlice(p.v.Type().Elem(), len(v), cap(v)) - for i, x := range v { - slice.Index(i).SetInt(int64(x)) - } - p.v.Elem().Set(slice) -} -func (p pointer) appendInt32Slice(v int32) { - grow(p.v.Elem()).SetInt(int64(v)) -} - -func (p pointer) toUint64() *uint64 { - return p.v.Interface().(*uint64) -} -func (p pointer) toUint64Ptr() **uint64 { - return p.v.Interface().(**uint64) -} -func (p pointer) toUint64Slice() *[]uint64 { - return p.v.Interface().(*[]uint64) -} -func (p pointer) toUint32() *uint32 { - return p.v.Interface().(*uint32) -} -func (p pointer) toUint32Ptr() **uint32 { - return p.v.Interface().(**uint32) -} -func (p pointer) toUint32Slice() *[]uint32 { - return p.v.Interface().(*[]uint32) -} -func (p pointer) toBool() *bool { - return p.v.Interface().(*bool) -} -func (p pointer) toBoolPtr() **bool { - return p.v.Interface().(**bool) -} -func (p pointer) toBoolSlice() *[]bool { - return p.v.Interface().(*[]bool) -} -func (p pointer) toFloat64() *float64 { - return p.v.Interface().(*float64) -} -func (p pointer) toFloat64Ptr() **float64 { - return p.v.Interface().(**float64) -} -func (p pointer) toFloat64Slice() *[]float64 { - return p.v.Interface().(*[]float64) -} -func (p pointer) toFloat32() *float32 { - return p.v.Interface().(*float32) -} -func (p pointer) toFloat32Ptr() **float32 { - return p.v.Interface().(**float32) -} -func (p pointer) toFloat32Slice() *[]float32 { - return p.v.Interface().(*[]float32) -} -func (p pointer) toString() *string { - return p.v.Interface().(*string) -} -func (p pointer) toStringPtr() **string { - return p.v.Interface().(**string) -} -func (p pointer) toStringSlice() *[]string { - return p.v.Interface().(*[]string) -} -func (p pointer) toBytes() *[]byte { - return p.v.Interface().(*[]byte) -} -func (p pointer) toBytesSlice() *[][]byte { - return p.v.Interface().(*[][]byte) -} -func (p pointer) toExtensions() *XXX_InternalExtensions { - return p.v.Interface().(*XXX_InternalExtensions) -} -func (p pointer) toOldExtensions() *map[int32]Extension { - return p.v.Interface().(*map[int32]Extension) -} -func (p pointer) getPointer() pointer { - return pointer{v: p.v.Elem()} -} -func (p pointer) setPointer(q pointer) { - p.v.Elem().Set(q.v) -} -func (p pointer) appendPointer(q pointer) { - grow(p.v.Elem()).Set(q.v) -} - -// getPointerSlice copies []*T from p as a new []pointer. -// This behavior differs from the implementation in pointer_unsafe.go. -func (p pointer) getPointerSlice() []pointer { - if p.v.IsNil() { - return nil - } - n := p.v.Elem().Len() - s := make([]pointer, n) - for i := 0; i < n; i++ { - s[i] = pointer{v: p.v.Elem().Index(i)} - } - return s -} - -// setPointerSlice copies []pointer into p as a new []*T. -// This behavior differs from the implementation in pointer_unsafe.go. -func (p pointer) setPointerSlice(v []pointer) { - if v == nil { - p.v.Elem().Set(reflect.New(p.v.Elem().Type()).Elem()) - return - } - s := reflect.MakeSlice(p.v.Elem().Type(), 0, len(v)) - for _, p := range v { - s = reflect.Append(s, p.v) - } - p.v.Elem().Set(s) -} - -// getInterfacePointer returns a pointer that points to the -// interface data of the interface pointed by p. -func (p pointer) getInterfacePointer() pointer { - if p.v.Elem().IsNil() { - return pointer{v: p.v.Elem()} - } - return pointer{v: p.v.Elem().Elem().Elem().Field(0).Addr()} // *interface -> interface -> *struct -> struct -} - -func (p pointer) asPointerTo(t reflect.Type) reflect.Value { - // TODO: check that p.v.Type().Elem() == t? - return p.v -} - -func atomicLoadUnmarshalInfo(p **unmarshalInfo) *unmarshalInfo { - atomicLock.Lock() - defer atomicLock.Unlock() - return *p -} -func atomicStoreUnmarshalInfo(p **unmarshalInfo, v *unmarshalInfo) { - atomicLock.Lock() - defer atomicLock.Unlock() - *p = v -} -func atomicLoadMarshalInfo(p **marshalInfo) *marshalInfo { - atomicLock.Lock() - defer atomicLock.Unlock() - return *p -} -func atomicStoreMarshalInfo(p **marshalInfo, v *marshalInfo) { - atomicLock.Lock() - defer atomicLock.Unlock() - *p = v -} -func atomicLoadMergeInfo(p **mergeInfo) *mergeInfo { - atomicLock.Lock() - defer atomicLock.Unlock() - return *p -} -func atomicStoreMergeInfo(p **mergeInfo, v *mergeInfo) { - atomicLock.Lock() - defer atomicLock.Unlock() - *p = v -} -func atomicLoadDiscardInfo(p **discardInfo) *discardInfo { - atomicLock.Lock() - defer atomicLock.Unlock() - return *p -} -func atomicStoreDiscardInfo(p **discardInfo, v *discardInfo) { - atomicLock.Lock() - defer atomicLock.Unlock() - *p = v -} - -var atomicLock sync.Mutex diff --git a/vendor/github.com/golang/protobuf/proto/pointer_unsafe.go b/vendor/github.com/golang/protobuf/proto/pointer_unsafe.go deleted file mode 100644 index dbfffe07..00000000 --- a/vendor/github.com/golang/protobuf/proto/pointer_unsafe.go +++ /dev/null @@ -1,313 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2012 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -// +build !purego,!appengine,!js - -// This file contains the implementation of the proto field accesses using package unsafe. - -package proto - -import ( - "reflect" - "sync/atomic" - "unsafe" -) - -const unsafeAllowed = true - -// A field identifies a field in a struct, accessible from a pointer. -// In this implementation, a field is identified by its byte offset from the start of the struct. -type field uintptr - -// toField returns a field equivalent to the given reflect field. -func toField(f *reflect.StructField) field { - return field(f.Offset) -} - -// invalidField is an invalid field identifier. -const invalidField = ^field(0) - -// zeroField is a noop when calling pointer.offset. -const zeroField = field(0) - -// IsValid reports whether the field identifier is valid. -func (f field) IsValid() bool { - return f != invalidField -} - -// The pointer type below is for the new table-driven encoder/decoder. -// The implementation here uses unsafe.Pointer to create a generic pointer. -// In pointer_reflect.go we use reflect instead of unsafe to implement -// the same (but slower) interface. -type pointer struct { - p unsafe.Pointer -} - -// size of pointer -var ptrSize = unsafe.Sizeof(uintptr(0)) - -// toPointer converts an interface of pointer type to a pointer -// that points to the same target. -func toPointer(i *Message) pointer { - // Super-tricky - read pointer out of data word of interface value. - // Saves ~25ns over the equivalent: - // return valToPointer(reflect.ValueOf(*i)) - return pointer{p: (*[2]unsafe.Pointer)(unsafe.Pointer(i))[1]} -} - -// toAddrPointer converts an interface to a pointer that points to -// the interface data. -func toAddrPointer(i *interface{}, isptr, deref bool) (p pointer) { - // Super-tricky - read or get the address of data word of interface value. - if isptr { - // The interface is of pointer type, thus it is a direct interface. - // The data word is the pointer data itself. We take its address. - p = pointer{p: unsafe.Pointer(uintptr(unsafe.Pointer(i)) + ptrSize)} - } else { - // The interface is not of pointer type. The data word is the pointer - // to the data. - p = pointer{p: (*[2]unsafe.Pointer)(unsafe.Pointer(i))[1]} - } - if deref { - p.p = *(*unsafe.Pointer)(p.p) - } - return p -} - -// valToPointer converts v to a pointer. v must be of pointer type. -func valToPointer(v reflect.Value) pointer { - return pointer{p: unsafe.Pointer(v.Pointer())} -} - -// offset converts from a pointer to a structure to a pointer to -// one of its fields. -func (p pointer) offset(f field) pointer { - // For safety, we should panic if !f.IsValid, however calling panic causes - // this to no longer be inlineable, which is a serious performance cost. - /* - if !f.IsValid() { - panic("invalid field") - } - */ - return pointer{p: unsafe.Pointer(uintptr(p.p) + uintptr(f))} -} - -func (p pointer) isNil() bool { - return p.p == nil -} - -func (p pointer) toInt64() *int64 { - return (*int64)(p.p) -} -func (p pointer) toInt64Ptr() **int64 { - return (**int64)(p.p) -} -func (p pointer) toInt64Slice() *[]int64 { - return (*[]int64)(p.p) -} -func (p pointer) toInt32() *int32 { - return (*int32)(p.p) -} - -// See pointer_reflect.go for why toInt32Ptr/Slice doesn't exist. -/* - func (p pointer) toInt32Ptr() **int32 { - return (**int32)(p.p) - } - func (p pointer) toInt32Slice() *[]int32 { - return (*[]int32)(p.p) - } -*/ -func (p pointer) getInt32Ptr() *int32 { - return *(**int32)(p.p) -} -func (p pointer) setInt32Ptr(v int32) { - *(**int32)(p.p) = &v -} - -// getInt32Slice loads a []int32 from p. -// The value returned is aliased with the original slice. -// This behavior differs from the implementation in pointer_reflect.go. -func (p pointer) getInt32Slice() []int32 { - return *(*[]int32)(p.p) -} - -// setInt32Slice stores a []int32 to p. -// The value set is aliased with the input slice. -// This behavior differs from the implementation in pointer_reflect.go. -func (p pointer) setInt32Slice(v []int32) { - *(*[]int32)(p.p) = v -} - -// TODO: Can we get rid of appendInt32Slice and use setInt32Slice instead? -func (p pointer) appendInt32Slice(v int32) { - s := (*[]int32)(p.p) - *s = append(*s, v) -} - -func (p pointer) toUint64() *uint64 { - return (*uint64)(p.p) -} -func (p pointer) toUint64Ptr() **uint64 { - return (**uint64)(p.p) -} -func (p pointer) toUint64Slice() *[]uint64 { - return (*[]uint64)(p.p) -} -func (p pointer) toUint32() *uint32 { - return (*uint32)(p.p) -} -func (p pointer) toUint32Ptr() **uint32 { - return (**uint32)(p.p) -} -func (p pointer) toUint32Slice() *[]uint32 { - return (*[]uint32)(p.p) -} -func (p pointer) toBool() *bool { - return (*bool)(p.p) -} -func (p pointer) toBoolPtr() **bool { - return (**bool)(p.p) -} -func (p pointer) toBoolSlice() *[]bool { - return (*[]bool)(p.p) -} -func (p pointer) toFloat64() *float64 { - return (*float64)(p.p) -} -func (p pointer) toFloat64Ptr() **float64 { - return (**float64)(p.p) -} -func (p pointer) toFloat64Slice() *[]float64 { - return (*[]float64)(p.p) -} -func (p pointer) toFloat32() *float32 { - return (*float32)(p.p) -} -func (p pointer) toFloat32Ptr() **float32 { - return (**float32)(p.p) -} -func (p pointer) toFloat32Slice() *[]float32 { - return (*[]float32)(p.p) -} -func (p pointer) toString() *string { - return (*string)(p.p) -} -func (p pointer) toStringPtr() **string { - return (**string)(p.p) -} -func (p pointer) toStringSlice() *[]string { - return (*[]string)(p.p) -} -func (p pointer) toBytes() *[]byte { - return (*[]byte)(p.p) -} -func (p pointer) toBytesSlice() *[][]byte { - return (*[][]byte)(p.p) -} -func (p pointer) toExtensions() *XXX_InternalExtensions { - return (*XXX_InternalExtensions)(p.p) -} -func (p pointer) toOldExtensions() *map[int32]Extension { - return (*map[int32]Extension)(p.p) -} - -// getPointerSlice loads []*T from p as a []pointer. -// The value returned is aliased with the original slice. -// This behavior differs from the implementation in pointer_reflect.go. -func (p pointer) getPointerSlice() []pointer { - // Super-tricky - p should point to a []*T where T is a - // message type. We load it as []pointer. - return *(*[]pointer)(p.p) -} - -// setPointerSlice stores []pointer into p as a []*T. -// The value set is aliased with the input slice. -// This behavior differs from the implementation in pointer_reflect.go. -func (p pointer) setPointerSlice(v []pointer) { - // Super-tricky - p should point to a []*T where T is a - // message type. We store it as []pointer. - *(*[]pointer)(p.p) = v -} - -// getPointer loads the pointer at p and returns it. -func (p pointer) getPointer() pointer { - return pointer{p: *(*unsafe.Pointer)(p.p)} -} - -// setPointer stores the pointer q at p. -func (p pointer) setPointer(q pointer) { - *(*unsafe.Pointer)(p.p) = q.p -} - -// append q to the slice pointed to by p. -func (p pointer) appendPointer(q pointer) { - s := (*[]unsafe.Pointer)(p.p) - *s = append(*s, q.p) -} - -// getInterfacePointer returns a pointer that points to the -// interface data of the interface pointed by p. -func (p pointer) getInterfacePointer() pointer { - // Super-tricky - read pointer out of data word of interface value. - return pointer{p: (*(*[2]unsafe.Pointer)(p.p))[1]} -} - -// asPointerTo returns a reflect.Value that is a pointer to an -// object of type t stored at p. -func (p pointer) asPointerTo(t reflect.Type) reflect.Value { - return reflect.NewAt(t, p.p) -} - -func atomicLoadUnmarshalInfo(p **unmarshalInfo) *unmarshalInfo { - return (*unmarshalInfo)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(p)))) -} -func atomicStoreUnmarshalInfo(p **unmarshalInfo, v *unmarshalInfo) { - atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(p)), unsafe.Pointer(v)) -} -func atomicLoadMarshalInfo(p **marshalInfo) *marshalInfo { - return (*marshalInfo)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(p)))) -} -func atomicStoreMarshalInfo(p **marshalInfo, v *marshalInfo) { - atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(p)), unsafe.Pointer(v)) -} -func atomicLoadMergeInfo(p **mergeInfo) *mergeInfo { - return (*mergeInfo)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(p)))) -} -func atomicStoreMergeInfo(p **mergeInfo, v *mergeInfo) { - atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(p)), unsafe.Pointer(v)) -} -func atomicLoadDiscardInfo(p **discardInfo) *discardInfo { - return (*discardInfo)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(p)))) -} -func atomicStoreDiscardInfo(p **discardInfo, v *discardInfo) { - atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(p)), unsafe.Pointer(v)) -} diff --git a/vendor/github.com/golang/protobuf/proto/properties.go b/vendor/github.com/golang/protobuf/proto/properties.go index a4b8c0cd..dcdc2202 100644 --- a/vendor/github.com/golang/protobuf/proto/properties.go +++ b/vendor/github.com/golang/protobuf/proto/properties.go @@ -1,162 +1,104 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2010 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. package proto -/* - * Routines for encoding data into the wire format for protocol buffers. - */ - import ( "fmt" - "log" "reflect" - "sort" "strconv" "strings" "sync" -) - -const debug bool = false -// Constants that identify the encoding of a value on the wire. -const ( - WireVarint = 0 - WireFixed64 = 1 - WireBytes = 2 - WireStartGroup = 3 - WireEndGroup = 4 - WireFixed32 = 5 + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoimpl" ) -// tagMap is an optimization over map[int]int for typical protocol buffer -// use-cases. Encoded protocol buffers are often in tag order with small tag -// numbers. -type tagMap struct { - fastTags []int - slowTags map[int]int -} - -// tagMapFastLimit is the upper bound on the tag number that will be stored in -// the tagMap slice rather than its map. -const tagMapFastLimit = 1024 - -func (p *tagMap) get(t int) (int, bool) { - if t > 0 && t < tagMapFastLimit { - if t >= len(p.fastTags) { - return 0, false - } - fi := p.fastTags[t] - return fi, fi >= 0 - } - fi, ok := p.slowTags[t] - return fi, ok -} - -func (p *tagMap) put(t int, fi int) { - if t > 0 && t < tagMapFastLimit { - for len(p.fastTags) < t+1 { - p.fastTags = append(p.fastTags, -1) - } - p.fastTags[t] = fi - return - } - if p.slowTags == nil { - p.slowTags = make(map[int]int) - } - p.slowTags[t] = fi -} - -// StructProperties represents properties for all the fields of a struct. -// decoderTags and decoderOrigNames should only be used by the decoder. +// StructProperties represents protocol buffer type information for a +// generated protobuf message in the open-struct API. +// +// Deprecated: Do not use. type StructProperties struct { - Prop []*Properties // properties for each field - reqCount int // required count - decoderTags tagMap // map from proto tag to struct field number - decoderOrigNames map[string]int // map from original name to struct field number - order []int // list of struct field numbers in tag order + // Prop are the properties for each field. + // + // Fields belonging to a oneof are stored in OneofTypes instead, with a + // single Properties representing the parent oneof held here. + // + // The order of Prop matches the order of fields in the Go struct. + // Struct fields that are not related to protobufs have a "XXX_" prefix + // in the Properties.Name and must be ignored by the user. + Prop []*Properties // OneofTypes contains information about the oneof fields in this message. - // It is keyed by the original name of a field. + // It is keyed by the protobuf field name. OneofTypes map[string]*OneofProperties } -// OneofProperties represents information about a specific field in a oneof. -type OneofProperties struct { - Type reflect.Type // pointer to generated struct type for this oneof field - Field int // struct field number of the containing oneof in the message - Prop *Properties -} - -// Implement the sorting interface so we can sort the fields in tag order, as recommended by the spec. -// See encode.go, (*Buffer).enc_struct. - -func (sp *StructProperties) Len() int { return len(sp.order) } -func (sp *StructProperties) Less(i, j int) bool { - return sp.Prop[sp.order[i]].Tag < sp.Prop[sp.order[j]].Tag -} -func (sp *StructProperties) Swap(i, j int) { sp.order[i], sp.order[j] = sp.order[j], sp.order[i] } - -// Properties represents the protocol-specific behavior of a single struct field. +// Properties represents the type information for a protobuf message field. +// +// Deprecated: Do not use. type Properties struct { - Name string // name of the field, for error messages - OrigName string // original name before protocol compiler (always set) - JSONName string // name to use for JSON; determined by protoc - Wire string + // Name is a placeholder name with little meaningful semantic value. + // If the name has an "XXX_" prefix, the entire Properties must be ignored. + Name string + // OrigName is the protobuf field name or oneof name. + OrigName string + // JSONName is the JSON name for the protobuf field. + JSONName string + // Enum is a placeholder name for enums. + // For historical reasons, this is neither the Go name for the enum, + // nor the protobuf name for the enum. + Enum string // Deprecated: Do not use. + // Weak contains the full name of the weakly referenced message. + Weak string + // Wire is a string representation of the wire type. + Wire string + // WireType is the protobuf wire type for the field. WireType int - Tag int + // Tag is the protobuf field number. + Tag int + // Required reports whether this is a required field. Required bool + // Optional reports whether this is a optional field. Optional bool + // Repeated reports whether this is a repeated field. Repeated bool - Packed bool // relevant for repeated primitives only - Enum string // set for enum types only - proto3 bool // whether this is known to be a proto3 field - oneof bool // whether this is a oneof field - - Default string // default value - HasDefault bool // whether an explicit default was provided - - stype reflect.Type // set for struct types only - sprop *StructProperties // set for struct types only + // Packed reports whether this is a packed repeated field of scalars. + Packed bool + // Proto3 reports whether this field operates under the proto3 syntax. + Proto3 bool + // Oneof reports whether this field belongs within a oneof. + Oneof bool + + // Default is the default value in string form. + Default string + // HasDefault reports whether the field has a default value. + HasDefault bool + + // MapKeyProp is the properties for the key field for a map field. + MapKeyProp *Properties + // MapValProp is the properties for the value field for a map field. + MapValProp *Properties +} - mtype reflect.Type // set for map types only - MapKeyProp *Properties // set for map types only - MapValProp *Properties // set for map types only +// OneofProperties represents the type information for a protobuf oneof. +// +// Deprecated: Do not use. +type OneofProperties struct { + // Type is a pointer to the generated wrapper type for the field value. + // This is nil for messages that are not in the open-struct API. + Type reflect.Type + // Field is the index into StructProperties.Prop for the containing oneof. + Field int + // Prop is the properties for the field. + Prop *Properties } // String formats the properties in the protobuf struct field tag style. func (p *Properties) String() string { s := p.Wire - s += "," - s += strconv.Itoa(p.Tag) + s += "," + strconv.Itoa(p.Tag) if p.Required { s += ",req" } @@ -170,18 +112,21 @@ func (p *Properties) String() string { s += ",packed" } s += ",name=" + p.OrigName - if p.JSONName != p.OrigName { + if p.JSONName != "" { s += ",json=" + p.JSONName } - if p.proto3 { + if len(p.Enum) > 0 { + s += ",enum=" + p.Enum + } + if len(p.Weak) > 0 { + s += ",weak=" + p.Weak + } + if p.Proto3 { s += ",proto3" } - if p.oneof { + if p.Oneof { s += ",oneof" } - if len(p.Enum) > 0 { - s += ",enum=" + p.Enum - } if p.HasDefault { s += ",def=" + p.Default } @@ -189,356 +134,173 @@ func (p *Properties) String() string { } // Parse populates p by parsing a string in the protobuf struct field tag style. -func (p *Properties) Parse(s string) { - // "bytes,49,opt,name=foo,def=hello!" - fields := strings.Split(s, ",") // breaks def=, but handled below. - if len(fields) < 2 { - log.Printf("proto: tag has too few fields: %q", s) - return - } - - p.Wire = fields[0] - switch p.Wire { - case "varint": - p.WireType = WireVarint - case "fixed32": - p.WireType = WireFixed32 - case "fixed64": - p.WireType = WireFixed64 - case "zigzag32": - p.WireType = WireVarint - case "zigzag64": - p.WireType = WireVarint - case "bytes", "group": - p.WireType = WireBytes - // no numeric converter for non-numeric types - default: - log.Printf("proto: tag has unknown wire type: %q", s) - return - } - - var err error - p.Tag, err = strconv.Atoi(fields[1]) - if err != nil { - return - } - -outer: - for i := 2; i < len(fields); i++ { - f := fields[i] - switch { - case f == "req": - p.Required = true - case f == "opt": +func (p *Properties) Parse(tag string) { + // For example: "bytes,49,opt,name=foo,def=hello!" + for len(tag) > 0 { + i := strings.IndexByte(tag, ',') + if i < 0 { + i = len(tag) + } + switch s := tag[:i]; { + case strings.HasPrefix(s, "name="): + p.OrigName = s[len("name="):] + case strings.HasPrefix(s, "json="): + p.JSONName = s[len("json="):] + case strings.HasPrefix(s, "enum="): + p.Enum = s[len("enum="):] + case strings.HasPrefix(s, "weak="): + p.Weak = s[len("weak="):] + case strings.Trim(s, "0123456789") == "": + n, _ := strconv.ParseUint(s, 10, 32) + p.Tag = int(n) + case s == "opt": p.Optional = true - case f == "rep": + case s == "req": + p.Required = true + case s == "rep": p.Repeated = true - case f == "packed": + case s == "varint" || s == "zigzag32" || s == "zigzag64": + p.Wire = s + p.WireType = WireVarint + case s == "fixed32": + p.Wire = s + p.WireType = WireFixed32 + case s == "fixed64": + p.Wire = s + p.WireType = WireFixed64 + case s == "bytes": + p.Wire = s + p.WireType = WireBytes + case s == "group": + p.Wire = s + p.WireType = WireStartGroup + case s == "packed": p.Packed = true - case strings.HasPrefix(f, "name="): - p.OrigName = f[5:] - case strings.HasPrefix(f, "json="): - p.JSONName = f[5:] - case strings.HasPrefix(f, "enum="): - p.Enum = f[5:] - case f == "proto3": - p.proto3 = true - case f == "oneof": - p.oneof = true - case strings.HasPrefix(f, "def="): + case s == "proto3": + p.Proto3 = true + case s == "oneof": + p.Oneof = true + case strings.HasPrefix(s, "def="): + // The default tag is special in that everything afterwards is the + // default regardless of the presence of commas. p.HasDefault = true - p.Default = f[4:] // rest of string - if i+1 < len(fields) { - // Commas aren't escaped, and def is always last. - p.Default += "," + strings.Join(fields[i+1:], ",") - break outer - } - } - } -} - -var protoMessageType = reflect.TypeOf((*Message)(nil)).Elem() - -// setFieldProps initializes the field properties for submessages and maps. -func (p *Properties) setFieldProps(typ reflect.Type, f *reflect.StructField, lockGetProp bool) { - switch t1 := typ; t1.Kind() { - case reflect.Ptr: - if t1.Elem().Kind() == reflect.Struct { - p.stype = t1.Elem() - } - - case reflect.Slice: - if t2 := t1.Elem(); t2.Kind() == reflect.Ptr && t2.Elem().Kind() == reflect.Struct { - p.stype = t2.Elem() - } - - case reflect.Map: - p.mtype = t1 - p.MapKeyProp = &Properties{} - p.MapKeyProp.init(reflect.PtrTo(p.mtype.Key()), "Key", f.Tag.Get("protobuf_key"), nil, lockGetProp) - p.MapValProp = &Properties{} - vtype := p.mtype.Elem() - if vtype.Kind() != reflect.Ptr && vtype.Kind() != reflect.Slice { - // The value type is not a message (*T) or bytes ([]byte), - // so we need encoders for the pointer to this type. - vtype = reflect.PtrTo(vtype) - } - p.MapValProp.init(vtype, "Value", f.Tag.Get("protobuf_val"), nil, lockGetProp) - } - - if p.stype != nil { - if lockGetProp { - p.sprop = GetProperties(p.stype) - } else { - p.sprop = getPropertiesLocked(p.stype) + p.Default, i = tag[len("def="):], len(tag) } + tag = strings.TrimPrefix(tag[i:], ",") } } -var ( - marshalerType = reflect.TypeOf((*Marshaler)(nil)).Elem() -) - // Init populates the properties from a protocol buffer struct tag. +// +// Deprecated: Do not use. func (p *Properties) Init(typ reflect.Type, name, tag string, f *reflect.StructField) { - p.init(typ, name, tag, f, true) -} - -func (p *Properties) init(typ reflect.Type, name, tag string, f *reflect.StructField, lockGetProp bool) { - // "bytes,49,opt,def=hello!" p.Name = name p.OrigName = name if tag == "" { return } p.Parse(tag) - p.setFieldProps(typ, f, lockGetProp) + + if typ != nil && typ.Kind() == reflect.Map { + p.MapKeyProp = new(Properties) + p.MapKeyProp.Init(nil, "Key", f.Tag.Get("protobuf_key"), nil) + p.MapValProp = new(Properties) + p.MapValProp.Init(nil, "Value", f.Tag.Get("protobuf_val"), nil) + } } -var ( - propertiesMu sync.RWMutex - propertiesMap = make(map[reflect.Type]*StructProperties) -) +var propertiesCache sync.Map // map[reflect.Type]*StructProperties -// GetProperties returns the list of properties for the type represented by t. -// t must represent a generated struct type of a protocol message. +// GetProperties returns the list of properties for the type represented by t, +// which must be a generated protocol buffer message in the open-struct API, +// where protobuf message fields are represented by exported Go struct fields. +// +// Deprecated: Use protobuf reflection instead. func GetProperties(t reflect.Type) *StructProperties { - if t.Kind() != reflect.Struct { - panic("proto: type must have kind struct") - } - - // Most calls to GetProperties in a long-running program will be - // retrieving details for types we have seen before. - propertiesMu.RLock() - sprop, ok := propertiesMap[t] - propertiesMu.RUnlock() - if ok { - return sprop + if p, ok := propertiesCache.Load(t); ok { + return p.(*StructProperties) } - - propertiesMu.Lock() - sprop = getPropertiesLocked(t) - propertiesMu.Unlock() - return sprop + p, _ := propertiesCache.LoadOrStore(t, newProperties(t)) + return p.(*StructProperties) } -type ( - oneofFuncsIface interface { - XXX_OneofFuncs() (func(Message, *Buffer) error, func(Message, int, int, *Buffer) (bool, error), func(Message) int, []interface{}) - } - oneofWrappersIface interface { - XXX_OneofWrappers() []interface{} - } -) - -// getPropertiesLocked requires that propertiesMu is held. -func getPropertiesLocked(t reflect.Type) *StructProperties { - if prop, ok := propertiesMap[t]; ok { - return prop +func newProperties(t reflect.Type) *StructProperties { + if t.Kind() != reflect.Struct { + panic(fmt.Sprintf("%v is not a generated message in the open-struct API", t)) } + var hasOneof bool prop := new(StructProperties) - // in case of recursive protos, fill this in now. - propertiesMap[t] = prop - - // build properties - prop.Prop = make([]*Properties, t.NumField()) - prop.order = make([]int, t.NumField()) + // Construct a list of properties for each field in the struct. for i := 0; i < t.NumField(); i++ { - f := t.Field(i) p := new(Properties) - name := f.Name - p.init(f.Type, name, f.Tag.Get("protobuf"), &f, false) + f := t.Field(i) + tagField := f.Tag.Get("protobuf") + p.Init(f.Type, f.Name, tagField, &f) - oneof := f.Tag.Get("protobuf_oneof") // special case - if oneof != "" { - // Oneof fields don't use the traditional protobuf tag. - p.OrigName = oneof + tagOneof := f.Tag.Get("protobuf_oneof") + if tagOneof != "" { + hasOneof = true + p.OrigName = tagOneof } - prop.Prop[i] = p - prop.order[i] = i - if debug { - print(i, " ", f.Name, " ", t.String(), " ") - if p.Tag > 0 { - print(p.String()) - } - print("\n") + + // Rename unrelated struct fields with the "XXX_" prefix since so much + // user code simply checks for this to exclude special fields. + if tagField == "" && tagOneof == "" && !strings.HasPrefix(p.Name, "XXX_") { + p.Name = "XXX_" + p.Name + p.OrigName = "XXX_" + p.OrigName + } else if p.Weak != "" { + p.Name = p.OrigName // avoid possible "XXX_" prefix on weak field } + + prop.Prop = append(prop.Prop, p) } - // Re-order prop.order. - sort.Sort(prop) + // Construct a mapping of oneof field names to properties. + if hasOneof { + var oneofWrappers []interface{} + if fn, ok := reflect.PtrTo(t).MethodByName("XXX_OneofFuncs"); ok { + oneofWrappers = fn.Func.Call([]reflect.Value{reflect.Zero(fn.Type.In(0))})[3].Interface().([]interface{}) + } + if fn, ok := reflect.PtrTo(t).MethodByName("XXX_OneofWrappers"); ok { + oneofWrappers = fn.Func.Call([]reflect.Value{reflect.Zero(fn.Type.In(0))})[0].Interface().([]interface{}) + } + if m, ok := reflect.Zero(reflect.PtrTo(t)).Interface().(protoreflect.ProtoMessage); ok { + if m, ok := m.ProtoReflect().(interface{ ProtoMessageInfo() *protoimpl.MessageInfo }); ok { + oneofWrappers = m.ProtoMessageInfo().OneofWrappers + } + } - var oots []interface{} - switch m := reflect.Zero(reflect.PtrTo(t)).Interface().(type) { - case oneofFuncsIface: - _, _, _, oots = m.XXX_OneofFuncs() - case oneofWrappersIface: - oots = m.XXX_OneofWrappers() - } - if len(oots) > 0 { - // Interpret oneof metadata. prop.OneofTypes = make(map[string]*OneofProperties) - for _, oot := range oots { - oop := &OneofProperties{ - Type: reflect.ValueOf(oot).Type(), // *T + for _, wrapper := range oneofWrappers { + p := &OneofProperties{ + Type: reflect.ValueOf(wrapper).Type(), // *T Prop: new(Properties), } - sft := oop.Type.Elem().Field(0) - oop.Prop.Name = sft.Name - oop.Prop.Parse(sft.Tag.Get("protobuf")) - // There will be exactly one interface field that - // this new value is assignable to. - for i := 0; i < t.NumField(); i++ { - f := t.Field(i) - if f.Type.Kind() != reflect.Interface { - continue + f := p.Type.Elem().Field(0) + p.Prop.Name = f.Name + p.Prop.Parse(f.Tag.Get("protobuf")) + + // Determine the struct field that contains this oneof. + // Each wrapper is assignable to exactly one parent field. + var foundOneof bool + for i := 0; i < t.NumField() && !foundOneof; i++ { + if p.Type.AssignableTo(t.Field(i).Type) { + p.Field = i + foundOneof = true } - if !oop.Type.AssignableTo(f.Type) { - continue - } - oop.Field = i - break } - prop.OneofTypes[oop.Prop.OrigName] = oop - } - } - - // build required counts - // build tags - reqCount := 0 - prop.decoderOrigNames = make(map[string]int) - for i, p := range prop.Prop { - if strings.HasPrefix(p.Name, "XXX_") { - // Internal fields should not appear in tags/origNames maps. - // They are handled specially when encoding and decoding. - continue - } - if p.Required { - reqCount++ + if !foundOneof { + panic(fmt.Sprintf("%v is not a generated message in the open-struct API", t)) + } + prop.OneofTypes[p.Prop.OrigName] = p } - prop.decoderTags.put(p.Tag, i) - prop.decoderOrigNames[p.OrigName] = i } - prop.reqCount = reqCount return prop } -// A global registry of enum types. -// The generated code will register the generated maps by calling RegisterEnum. - -var enumValueMaps = make(map[string]map[string]int32) - -// RegisterEnum is called from the generated code to install the enum descriptor -// maps into the global table to aid parsing text format protocol buffers. -func RegisterEnum(typeName string, unusedNameMap map[int32]string, valueMap map[string]int32) { - if _, ok := enumValueMaps[typeName]; ok { - panic("proto: duplicate enum registered: " + typeName) - } - enumValueMaps[typeName] = valueMap -} - -// EnumValueMap returns the mapping from names to integers of the -// enum type enumType, or a nil if not found. -func EnumValueMap(enumType string) map[string]int32 { - return enumValueMaps[enumType] -} - -// A registry of all linked message types. -// The string is a fully-qualified proto name ("pkg.Message"). -var ( - protoTypedNils = make(map[string]Message) // a map from proto names to typed nil pointers - protoMapTypes = make(map[string]reflect.Type) // a map from proto names to map types - revProtoTypes = make(map[reflect.Type]string) -) - -// RegisterType is called from generated code and maps from the fully qualified -// proto name to the type (pointer to struct) of the protocol buffer. -func RegisterType(x Message, name string) { - if _, ok := protoTypedNils[name]; ok { - // TODO: Some day, make this a panic. - log.Printf("proto: duplicate proto type registered: %s", name) - return - } - t := reflect.TypeOf(x) - if v := reflect.ValueOf(x); v.Kind() == reflect.Ptr && v.Pointer() == 0 { - // Generated code always calls RegisterType with nil x. - // This check is just for extra safety. - protoTypedNils[name] = x - } else { - protoTypedNils[name] = reflect.Zero(t).Interface().(Message) - } - revProtoTypes[t] = name -} - -// RegisterMapType is called from generated code and maps from the fully qualified -// proto name to the native map type of the proto map definition. -func RegisterMapType(x interface{}, name string) { - if reflect.TypeOf(x).Kind() != reflect.Map { - panic(fmt.Sprintf("RegisterMapType(%T, %q); want map", x, name)) - } - if _, ok := protoMapTypes[name]; ok { - log.Printf("proto: duplicate proto type registered: %s", name) - return - } - t := reflect.TypeOf(x) - protoMapTypes[name] = t - revProtoTypes[t] = name -} - -// MessageName returns the fully-qualified proto name for the given message type. -func MessageName(x Message) string { - type xname interface { - XXX_MessageName() string - } - if m, ok := x.(xname); ok { - return m.XXX_MessageName() - } - return revProtoTypes[reflect.TypeOf(x)] -} - -// MessageType returns the message type (pointer to struct) for a named message. -// The type is not guaranteed to implement proto.Message if the name refers to a -// map entry. -func MessageType(name string) reflect.Type { - if t, ok := protoTypedNils[name]; ok { - return reflect.TypeOf(t) - } - return protoMapTypes[name] -} - -// A registry of all linked proto files. -var ( - protoFiles = make(map[string][]byte) // file name => fileDescriptor -) - -// RegisterFile is called from generated code and maps from the -// full file name of a .proto file to its compressed FileDescriptorProto. -func RegisterFile(filename string, fileDescriptor []byte) { - protoFiles[filename] = fileDescriptor -} - -// FileDescriptor returns the compressed FileDescriptorProto for a .proto file. -func FileDescriptor(filename string) []byte { return protoFiles[filename] } +func (sp *StructProperties) Len() int { return len(sp.Prop) } +func (sp *StructProperties) Less(i, j int) bool { return false } +func (sp *StructProperties) Swap(i, j int) { return } diff --git a/vendor/github.com/golang/protobuf/proto/proto.go b/vendor/github.com/golang/protobuf/proto/proto.go new file mode 100644 index 00000000..5aee89c3 --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/proto.go @@ -0,0 +1,167 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package proto provides functionality for handling protocol buffer messages. +// In particular, it provides marshaling and unmarshaling between a protobuf +// message and the binary wire format. +// +// See https://developers.google.com/protocol-buffers/docs/gotutorial for +// more information. +// +// Deprecated: Use the "google.golang.org/protobuf/proto" package instead. +package proto + +import ( + protoV2 "google.golang.org/protobuf/proto" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" + "google.golang.org/protobuf/runtime/protoimpl" +) + +const ( + ProtoPackageIsVersion1 = true + ProtoPackageIsVersion2 = true + ProtoPackageIsVersion3 = true + ProtoPackageIsVersion4 = true +) + +// GeneratedEnum is any enum type generated by protoc-gen-go +// which is a named int32 kind. +// This type exists for documentation purposes. +type GeneratedEnum interface{} + +// GeneratedMessage is any message type generated by protoc-gen-go +// which is a pointer to a named struct kind. +// This type exists for documentation purposes. +type GeneratedMessage interface{} + +// Message is a protocol buffer message. +// +// This is the v1 version of the message interface and is marginally better +// than an empty interface as it lacks any method to programatically interact +// with the contents of the message. +// +// A v2 message is declared in "google.golang.org/protobuf/proto".Message and +// exposes protobuf reflection as a first-class feature of the interface. +// +// To convert a v1 message to a v2 message, use the MessageV2 function. +// To convert a v2 message to a v1 message, use the MessageV1 function. +type Message = protoiface.MessageV1 + +// MessageV1 converts either a v1 or v2 message to a v1 message. +// It returns nil if m is nil. +func MessageV1(m GeneratedMessage) protoiface.MessageV1 { + return protoimpl.X.ProtoMessageV1Of(m) +} + +// MessageV2 converts either a v1 or v2 message to a v2 message. +// It returns nil if m is nil. +func MessageV2(m GeneratedMessage) protoV2.Message { + return protoimpl.X.ProtoMessageV2Of(m) +} + +// MessageReflect returns a reflective view for a message. +// It returns nil if m is nil. +func MessageReflect(m Message) protoreflect.Message { + return protoimpl.X.MessageOf(m) +} + +// Marshaler is implemented by messages that can marshal themselves. +// This interface is used by the following functions: Size, Marshal, +// Buffer.Marshal, and Buffer.EncodeMessage. +// +// Deprecated: Do not implement. +type Marshaler interface { + // Marshal formats the encoded bytes of the message. + // It should be deterministic and emit valid protobuf wire data. + // The caller takes ownership of the returned buffer. + Marshal() ([]byte, error) +} + +// Unmarshaler is implemented by messages that can unmarshal themselves. +// This interface is used by the following functions: Unmarshal, UnmarshalMerge, +// Buffer.Unmarshal, Buffer.DecodeMessage, and Buffer.DecodeGroup. +// +// Deprecated: Do not implement. +type Unmarshaler interface { + // Unmarshal parses the encoded bytes of the protobuf wire input. + // The provided buffer is only valid for during method call. + // It should not reset the receiver message. + Unmarshal([]byte) error +} + +// Merger is implemented by messages that can merge themselves. +// This interface is used by the following functions: Clone and Merge. +// +// Deprecated: Do not implement. +type Merger interface { + // Merge merges the contents of src into the receiver message. + // It clones all data structures in src such that it aliases no mutable + // memory referenced by src. + Merge(src Message) +} + +// RequiredNotSetError is an error type returned when +// marshaling or unmarshaling a message with missing required fields. +type RequiredNotSetError struct { + err error +} + +func (e *RequiredNotSetError) Error() string { + if e.err != nil { + return e.err.Error() + } + return "proto: required field not set" +} +func (e *RequiredNotSetError) RequiredNotSet() bool { + return true +} + +func checkRequiredNotSet(m protoV2.Message) error { + if err := protoV2.CheckInitialized(m); err != nil { + return &RequiredNotSetError{err: err} + } + return nil +} + +// Clone returns a deep copy of src. +func Clone(src Message) Message { + return MessageV1(protoV2.Clone(MessageV2(src))) +} + +// Merge merges src into dst, which must be messages of the same type. +// +// Populated scalar fields in src are copied to dst, while populated +// singular messages in src are merged into dst by recursively calling Merge. +// The elements of every list field in src is appended to the corresponded +// list fields in dst. The entries of every map field in src is copied into +// the corresponding map field in dst, possibly replacing existing entries. +// The unknown fields of src are appended to the unknown fields of dst. +func Merge(dst, src Message) { + protoV2.Merge(MessageV2(dst), MessageV2(src)) +} + +// Equal reports whether two messages are equal. +// If two messages marshal to the same bytes under deterministic serialization, +// then Equal is guaranteed to report true. +// +// Two messages are equal if they are the same protobuf message type, +// have the same set of populated known and extension field values, +// and the same set of unknown fields values. +// +// Scalar values are compared with the equivalent of the == operator in Go, +// except bytes values which are compared using bytes.Equal and +// floating point values which specially treat NaNs as equal. +// Message values are compared by recursively calling Equal. +// Lists are equal if each element value is also equal. +// Maps are equal if they have the same set of keys, where the pair of values +// for each key is also equal. +func Equal(x, y Message) bool { + return protoV2.Equal(MessageV2(x), MessageV2(y)) +} + +func isMessageSet(md protoreflect.MessageDescriptor) bool { + ms, ok := md.(interface{ IsMessageSet() bool }) + return ok && ms.IsMessageSet() +} diff --git a/vendor/github.com/golang/protobuf/proto/registry.go b/vendor/github.com/golang/protobuf/proto/registry.go new file mode 100644 index 00000000..1e7ff642 --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/registry.go @@ -0,0 +1,323 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "bytes" + "compress/gzip" + "fmt" + "io/ioutil" + "reflect" + "strings" + "sync" + + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" + "google.golang.org/protobuf/runtime/protoimpl" +) + +// filePath is the path to the proto source file. +type filePath = string // e.g., "google/protobuf/descriptor.proto" + +// fileDescGZIP is the compressed contents of the encoded FileDescriptorProto. +type fileDescGZIP = []byte + +var fileCache sync.Map // map[filePath]fileDescGZIP + +// RegisterFile is called from generated code to register the compressed +// FileDescriptorProto with the file path for a proto source file. +// +// Deprecated: Use protoregistry.GlobalFiles.RegisterFile instead. +func RegisterFile(s filePath, d fileDescGZIP) { + // Decompress the descriptor. + zr, err := gzip.NewReader(bytes.NewReader(d)) + if err != nil { + panic(fmt.Sprintf("proto: invalid compressed file descriptor: %v", err)) + } + b, err := ioutil.ReadAll(zr) + if err != nil { + panic(fmt.Sprintf("proto: invalid compressed file descriptor: %v", err)) + } + + // Construct a protoreflect.FileDescriptor from the raw descriptor. + // Note that DescBuilder.Build automatically registers the constructed + // file descriptor with the v2 registry. + protoimpl.DescBuilder{RawDescriptor: b}.Build() + + // Locally cache the raw descriptor form for the file. + fileCache.Store(s, d) +} + +// FileDescriptor returns the compressed FileDescriptorProto given the file path +// for a proto source file. It returns nil if not found. +// +// Deprecated: Use protoregistry.GlobalFiles.FindFileByPath instead. +func FileDescriptor(s filePath) fileDescGZIP { + if v, ok := fileCache.Load(s); ok { + return v.(fileDescGZIP) + } + + // Find the descriptor in the v2 registry. + var b []byte + if fd, _ := protoregistry.GlobalFiles.FindFileByPath(s); fd != nil { + if fd, ok := fd.(interface{ ProtoLegacyRawDesc() []byte }); ok { + b = fd.ProtoLegacyRawDesc() + } else { + // TODO: Use protodesc.ToFileDescriptorProto to construct + // a descriptorpb.FileDescriptorProto and marshal it. + // However, doing so causes the proto package to have a dependency + // on descriptorpb, leading to cyclic dependency issues. + } + } + + // Locally cache the raw descriptor form for the file. + if len(b) > 0 { + v, _ := fileCache.LoadOrStore(s, protoimpl.X.CompressGZIP(b)) + return v.(fileDescGZIP) + } + return nil +} + +// enumName is the name of an enum. For historical reasons, the enum name is +// neither the full Go name nor the full protobuf name of the enum. +// The name is the dot-separated combination of just the proto package that the +// enum is declared within followed by the Go type name of the generated enum. +type enumName = string // e.g., "my.proto.package.GoMessage_GoEnum" + +// enumsByName maps enum values by name to their numeric counterpart. +type enumsByName = map[string]int32 + +// enumsByNumber maps enum values by number to their name counterpart. +type enumsByNumber = map[int32]string + +var enumCache sync.Map // map[enumName]enumsByName +var numFilesCache sync.Map // map[protoreflect.FullName]int + +// RegisterEnum is called from the generated code to register the mapping of +// enum value names to enum numbers for the enum identified by s. +// +// Deprecated: Use protoregistry.GlobalTypes.RegisterEnum instead. +func RegisterEnum(s enumName, _ enumsByNumber, m enumsByName) { + if _, ok := enumCache.Load(s); ok { + panic("proto: duplicate enum registered: " + s) + } + enumCache.Store(s, m) + + // This does not forward registration to the v2 registry since this API + // lacks sufficient information to construct a complete v2 enum descriptor. +} + +// EnumValueMap returns the mapping from enum value names to enum numbers for +// the enum of the given name. It returns nil if not found. +// +// Deprecated: Use protoregistry.GlobalTypes.FindEnumByName instead. +func EnumValueMap(s enumName) enumsByName { + if v, ok := enumCache.Load(s); ok { + return v.(enumsByName) + } + + // Check whether the cache is stale. If the number of files in the current + // package differs, then it means that some enums may have been recently + // registered upstream that we do not know about. + var protoPkg protoreflect.FullName + if i := strings.LastIndexByte(s, '.'); i >= 0 { + protoPkg = protoreflect.FullName(s[:i]) + } + v, _ := numFilesCache.Load(protoPkg) + numFiles, _ := v.(int) + if protoregistry.GlobalFiles.NumFilesByPackage(protoPkg) == numFiles { + return nil // cache is up-to-date; was not found earlier + } + + // Update the enum cache for all enums declared in the given proto package. + numFiles = 0 + protoregistry.GlobalFiles.RangeFilesByPackage(protoPkg, func(fd protoreflect.FileDescriptor) bool { + walkEnums(fd, func(ed protoreflect.EnumDescriptor) { + name := protoimpl.X.LegacyEnumName(ed) + if _, ok := enumCache.Load(name); !ok { + m := make(enumsByName) + evs := ed.Values() + for i := evs.Len() - 1; i >= 0; i-- { + ev := evs.Get(i) + m[string(ev.Name())] = int32(ev.Number()) + } + enumCache.LoadOrStore(name, m) + } + }) + numFiles++ + return true + }) + numFilesCache.Store(protoPkg, numFiles) + + // Check cache again for enum map. + if v, ok := enumCache.Load(s); ok { + return v.(enumsByName) + } + return nil +} + +// walkEnums recursively walks all enums declared in d. +func walkEnums(d interface { + Enums() protoreflect.EnumDescriptors + Messages() protoreflect.MessageDescriptors +}, f func(protoreflect.EnumDescriptor)) { + eds := d.Enums() + for i := eds.Len() - 1; i >= 0; i-- { + f(eds.Get(i)) + } + mds := d.Messages() + for i := mds.Len() - 1; i >= 0; i-- { + walkEnums(mds.Get(i), f) + } +} + +// messageName is the full name of protobuf message. +type messageName = string + +var messageTypeCache sync.Map // map[messageName]reflect.Type + +// RegisterType is called from generated code to register the message Go type +// for a message of the given name. +// +// Deprecated: Use protoregistry.GlobalTypes.RegisterMessage instead. +func RegisterType(m Message, s messageName) { + mt := protoimpl.X.LegacyMessageTypeOf(m, protoreflect.FullName(s)) + if err := protoregistry.GlobalTypes.RegisterMessage(mt); err != nil { + panic(err) + } + messageTypeCache.Store(s, reflect.TypeOf(m)) +} + +// RegisterMapType is called from generated code to register the Go map type +// for a protobuf message representing a map entry. +// +// Deprecated: Do not use. +func RegisterMapType(m interface{}, s messageName) { + t := reflect.TypeOf(m) + if t.Kind() != reflect.Map { + panic(fmt.Sprintf("invalid map kind: %v", t)) + } + if _, ok := messageTypeCache.Load(s); ok { + panic(fmt.Errorf("proto: duplicate proto message registered: %s", s)) + } + messageTypeCache.Store(s, t) +} + +// MessageType returns the message type for a named message. +// It returns nil if not found. +// +// Deprecated: Use protoregistry.GlobalTypes.FindMessageByName instead. +func MessageType(s messageName) reflect.Type { + if v, ok := messageTypeCache.Load(s); ok { + return v.(reflect.Type) + } + + // Derive the message type from the v2 registry. + var t reflect.Type + if mt, _ := protoregistry.GlobalTypes.FindMessageByName(protoreflect.FullName(s)); mt != nil { + t = messageGoType(mt) + } + + // If we could not get a concrete type, it is possible that it is a + // pseudo-message for a map entry. + if t == nil { + d, _ := protoregistry.GlobalFiles.FindDescriptorByName(protoreflect.FullName(s)) + if md, _ := d.(protoreflect.MessageDescriptor); md != nil && md.IsMapEntry() { + kt := goTypeForField(md.Fields().ByNumber(1)) + vt := goTypeForField(md.Fields().ByNumber(2)) + t = reflect.MapOf(kt, vt) + } + } + + // Locally cache the message type for the given name. + if t != nil { + v, _ := messageTypeCache.LoadOrStore(s, t) + return v.(reflect.Type) + } + return nil +} + +func goTypeForField(fd protoreflect.FieldDescriptor) reflect.Type { + switch k := fd.Kind(); k { + case protoreflect.EnumKind: + if et, _ := protoregistry.GlobalTypes.FindEnumByName(fd.Enum().FullName()); et != nil { + return enumGoType(et) + } + return reflect.TypeOf(protoreflect.EnumNumber(0)) + case protoreflect.MessageKind, protoreflect.GroupKind: + if mt, _ := protoregistry.GlobalTypes.FindMessageByName(fd.Message().FullName()); mt != nil { + return messageGoType(mt) + } + return reflect.TypeOf((*protoreflect.Message)(nil)).Elem() + default: + return reflect.TypeOf(fd.Default().Interface()) + } +} + +func enumGoType(et protoreflect.EnumType) reflect.Type { + return reflect.TypeOf(et.New(0)) +} + +func messageGoType(mt protoreflect.MessageType) reflect.Type { + return reflect.TypeOf(MessageV1(mt.Zero().Interface())) +} + +// MessageName returns the full protobuf name for the given message type. +// +// Deprecated: Use protoreflect.MessageDescriptor.FullName instead. +func MessageName(m Message) messageName { + if m == nil { + return "" + } + if m, ok := m.(interface{ XXX_MessageName() messageName }); ok { + return m.XXX_MessageName() + } + return messageName(protoimpl.X.MessageDescriptorOf(m).FullName()) +} + +// RegisterExtension is called from the generated code to register +// the extension descriptor. +// +// Deprecated: Use protoregistry.GlobalTypes.RegisterExtension instead. +func RegisterExtension(d *ExtensionDesc) { + if err := protoregistry.GlobalTypes.RegisterExtension(d); err != nil { + panic(err) + } +} + +type extensionsByNumber = map[int32]*ExtensionDesc + +var extensionCache sync.Map // map[messageName]extensionsByNumber + +// RegisteredExtensions returns a map of the registered extensions for the +// provided protobuf message, indexed by the extension field number. +// +// Deprecated: Use protoregistry.GlobalTypes.RangeExtensionsByMessage instead. +func RegisteredExtensions(m Message) extensionsByNumber { + // Check whether the cache is stale. If the number of extensions for + // the given message differs, then it means that some extensions were + // recently registered upstream that we do not know about. + s := MessageName(m) + v, _ := extensionCache.Load(s) + xs, _ := v.(extensionsByNumber) + if protoregistry.GlobalTypes.NumExtensionsByMessage(protoreflect.FullName(s)) == len(xs) { + return xs // cache is up-to-date + } + + // Cache is stale, re-compute the extensions map. + xs = make(extensionsByNumber) + protoregistry.GlobalTypes.RangeExtensionsByMessage(protoreflect.FullName(s), func(xt protoreflect.ExtensionType) bool { + if xd, ok := xt.(*ExtensionDesc); ok { + xs[int32(xt.TypeDescriptor().Number())] = xd + } else { + // TODO: This implies that the protoreflect.ExtensionType is a + // custom type not generated by protoc-gen-go. We could try and + // convert the type to an ExtensionDesc. + } + return true + }) + extensionCache.Store(s, xs) + return xs +} diff --git a/vendor/github.com/golang/protobuf/proto/table_marshal.go b/vendor/github.com/golang/protobuf/proto/table_marshal.go deleted file mode 100644 index 5cb11fa9..00000000 --- a/vendor/github.com/golang/protobuf/proto/table_marshal.go +++ /dev/null @@ -1,2776 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2016 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -import ( - "errors" - "fmt" - "math" - "reflect" - "sort" - "strconv" - "strings" - "sync" - "sync/atomic" - "unicode/utf8" -) - -// a sizer takes a pointer to a field and the size of its tag, computes the size of -// the encoded data. -type sizer func(pointer, int) int - -// a marshaler takes a byte slice, a pointer to a field, and its tag (in wire format), -// marshals the field to the end of the slice, returns the slice and error (if any). -type marshaler func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) - -// marshalInfo is the information used for marshaling a message. -type marshalInfo struct { - typ reflect.Type - fields []*marshalFieldInfo - unrecognized field // offset of XXX_unrecognized - extensions field // offset of XXX_InternalExtensions - v1extensions field // offset of XXX_extensions - sizecache field // offset of XXX_sizecache - initialized int32 // 0 -- only typ is set, 1 -- fully initialized - messageset bool // uses message set wire format - hasmarshaler bool // has custom marshaler - sync.RWMutex // protect extElems map, also for initialization - extElems map[int32]*marshalElemInfo // info of extension elements -} - -// marshalFieldInfo is the information used for marshaling a field of a message. -type marshalFieldInfo struct { - field field - wiretag uint64 // tag in wire format - tagsize int // size of tag in wire format - sizer sizer - marshaler marshaler - isPointer bool - required bool // field is required - name string // name of the field, for error reporting - oneofElems map[reflect.Type]*marshalElemInfo // info of oneof elements -} - -// marshalElemInfo is the information used for marshaling an extension or oneof element. -type marshalElemInfo struct { - wiretag uint64 // tag in wire format - tagsize int // size of tag in wire format - sizer sizer - marshaler marshaler - isptr bool // elem is pointer typed, thus interface of this type is a direct interface (extension only) - deref bool // dereference the pointer before operating on it; implies isptr -} - -var ( - marshalInfoMap = map[reflect.Type]*marshalInfo{} - marshalInfoLock sync.Mutex -) - -// getMarshalInfo returns the information to marshal a given type of message. -// The info it returns may not necessarily initialized. -// t is the type of the message (NOT the pointer to it). -func getMarshalInfo(t reflect.Type) *marshalInfo { - marshalInfoLock.Lock() - u, ok := marshalInfoMap[t] - if !ok { - u = &marshalInfo{typ: t} - marshalInfoMap[t] = u - } - marshalInfoLock.Unlock() - return u -} - -// Size is the entry point from generated code, -// and should be ONLY called by generated code. -// It computes the size of encoded data of msg. -// a is a pointer to a place to store cached marshal info. -func (a *InternalMessageInfo) Size(msg Message) int { - u := getMessageMarshalInfo(msg, a) - ptr := toPointer(&msg) - if ptr.isNil() { - // We get here if msg is a typed nil ((*SomeMessage)(nil)), - // so it satisfies the interface, and msg == nil wouldn't - // catch it. We don't want crash in this case. - return 0 - } - return u.size(ptr) -} - -// Marshal is the entry point from generated code, -// and should be ONLY called by generated code. -// It marshals msg to the end of b. -// a is a pointer to a place to store cached marshal info. -func (a *InternalMessageInfo) Marshal(b []byte, msg Message, deterministic bool) ([]byte, error) { - u := getMessageMarshalInfo(msg, a) - ptr := toPointer(&msg) - if ptr.isNil() { - // We get here if msg is a typed nil ((*SomeMessage)(nil)), - // so it satisfies the interface, and msg == nil wouldn't - // catch it. We don't want crash in this case. - return b, ErrNil - } - return u.marshal(b, ptr, deterministic) -} - -func getMessageMarshalInfo(msg interface{}, a *InternalMessageInfo) *marshalInfo { - // u := a.marshal, but atomically. - // We use an atomic here to ensure memory consistency. - u := atomicLoadMarshalInfo(&a.marshal) - if u == nil { - // Get marshal information from type of message. - t := reflect.ValueOf(msg).Type() - if t.Kind() != reflect.Ptr { - panic(fmt.Sprintf("cannot handle non-pointer message type %v", t)) - } - u = getMarshalInfo(t.Elem()) - // Store it in the cache for later users. - // a.marshal = u, but atomically. - atomicStoreMarshalInfo(&a.marshal, u) - } - return u -} - -// size is the main function to compute the size of the encoded data of a message. -// ptr is the pointer to the message. -func (u *marshalInfo) size(ptr pointer) int { - if atomic.LoadInt32(&u.initialized) == 0 { - u.computeMarshalInfo() - } - - // If the message can marshal itself, let it do it, for compatibility. - // NOTE: This is not efficient. - if u.hasmarshaler { - m := ptr.asPointerTo(u.typ).Interface().(Marshaler) - b, _ := m.Marshal() - return len(b) - } - - n := 0 - for _, f := range u.fields { - if f.isPointer && ptr.offset(f.field).getPointer().isNil() { - // nil pointer always marshals to nothing - continue - } - n += f.sizer(ptr.offset(f.field), f.tagsize) - } - if u.extensions.IsValid() { - e := ptr.offset(u.extensions).toExtensions() - if u.messageset { - n += u.sizeMessageSet(e) - } else { - n += u.sizeExtensions(e) - } - } - if u.v1extensions.IsValid() { - m := *ptr.offset(u.v1extensions).toOldExtensions() - n += u.sizeV1Extensions(m) - } - if u.unrecognized.IsValid() { - s := *ptr.offset(u.unrecognized).toBytes() - n += len(s) - } - // cache the result for use in marshal - if u.sizecache.IsValid() { - atomic.StoreInt32(ptr.offset(u.sizecache).toInt32(), int32(n)) - } - return n -} - -// cachedsize gets the size from cache. If there is no cache (i.e. message is not generated), -// fall back to compute the size. -func (u *marshalInfo) cachedsize(ptr pointer) int { - if u.sizecache.IsValid() { - return int(atomic.LoadInt32(ptr.offset(u.sizecache).toInt32())) - } - return u.size(ptr) -} - -// marshal is the main function to marshal a message. It takes a byte slice and appends -// the encoded data to the end of the slice, returns the slice and error (if any). -// ptr is the pointer to the message. -// If deterministic is true, map is marshaled in deterministic order. -func (u *marshalInfo) marshal(b []byte, ptr pointer, deterministic bool) ([]byte, error) { - if atomic.LoadInt32(&u.initialized) == 0 { - u.computeMarshalInfo() - } - - // If the message can marshal itself, let it do it, for compatibility. - // NOTE: This is not efficient. - if u.hasmarshaler { - m := ptr.asPointerTo(u.typ).Interface().(Marshaler) - b1, err := m.Marshal() - b = append(b, b1...) - return b, err - } - - var err, errLater error - // The old marshaler encodes extensions at beginning. - if u.extensions.IsValid() { - e := ptr.offset(u.extensions).toExtensions() - if u.messageset { - b, err = u.appendMessageSet(b, e, deterministic) - } else { - b, err = u.appendExtensions(b, e, deterministic) - } - if err != nil { - return b, err - } - } - if u.v1extensions.IsValid() { - m := *ptr.offset(u.v1extensions).toOldExtensions() - b, err = u.appendV1Extensions(b, m, deterministic) - if err != nil { - return b, err - } - } - for _, f := range u.fields { - if f.required { - if ptr.offset(f.field).getPointer().isNil() { - // Required field is not set. - // We record the error but keep going, to give a complete marshaling. - if errLater == nil { - errLater = &RequiredNotSetError{f.name} - } - continue - } - } - if f.isPointer && ptr.offset(f.field).getPointer().isNil() { - // nil pointer always marshals to nothing - continue - } - b, err = f.marshaler(b, ptr.offset(f.field), f.wiretag, deterministic) - if err != nil { - if err1, ok := err.(*RequiredNotSetError); ok { - // Required field in submessage is not set. - // We record the error but keep going, to give a complete marshaling. - if errLater == nil { - errLater = &RequiredNotSetError{f.name + "." + err1.field} - } - continue - } - if err == errRepeatedHasNil { - err = errors.New("proto: repeated field " + f.name + " has nil element") - } - if err == errInvalidUTF8 { - if errLater == nil { - fullName := revProtoTypes[reflect.PtrTo(u.typ)] + "." + f.name - errLater = &invalidUTF8Error{fullName} - } - continue - } - return b, err - } - } - if u.unrecognized.IsValid() { - s := *ptr.offset(u.unrecognized).toBytes() - b = append(b, s...) - } - return b, errLater -} - -// computeMarshalInfo initializes the marshal info. -func (u *marshalInfo) computeMarshalInfo() { - u.Lock() - defer u.Unlock() - if u.initialized != 0 { // non-atomic read is ok as it is protected by the lock - return - } - - t := u.typ - u.unrecognized = invalidField - u.extensions = invalidField - u.v1extensions = invalidField - u.sizecache = invalidField - - // If the message can marshal itself, let it do it, for compatibility. - // NOTE: This is not efficient. - if reflect.PtrTo(t).Implements(marshalerType) { - u.hasmarshaler = true - atomic.StoreInt32(&u.initialized, 1) - return - } - - // get oneof implementers - var oneofImplementers []interface{} - switch m := reflect.Zero(reflect.PtrTo(t)).Interface().(type) { - case oneofFuncsIface: - _, _, _, oneofImplementers = m.XXX_OneofFuncs() - case oneofWrappersIface: - oneofImplementers = m.XXX_OneofWrappers() - } - - n := t.NumField() - - // deal with XXX fields first - for i := 0; i < t.NumField(); i++ { - f := t.Field(i) - if !strings.HasPrefix(f.Name, "XXX_") { - continue - } - switch f.Name { - case "XXX_sizecache": - u.sizecache = toField(&f) - case "XXX_unrecognized": - u.unrecognized = toField(&f) - case "XXX_InternalExtensions": - u.extensions = toField(&f) - u.messageset = f.Tag.Get("protobuf_messageset") == "1" - case "XXX_extensions": - u.v1extensions = toField(&f) - case "XXX_NoUnkeyedLiteral": - // nothing to do - default: - panic("unknown XXX field: " + f.Name) - } - n-- - } - - // normal fields - fields := make([]marshalFieldInfo, n) // batch allocation - u.fields = make([]*marshalFieldInfo, 0, n) - for i, j := 0, 0; i < t.NumField(); i++ { - f := t.Field(i) - - if strings.HasPrefix(f.Name, "XXX_") { - continue - } - field := &fields[j] - j++ - field.name = f.Name - u.fields = append(u.fields, field) - if f.Tag.Get("protobuf_oneof") != "" { - field.computeOneofFieldInfo(&f, oneofImplementers) - continue - } - if f.Tag.Get("protobuf") == "" { - // field has no tag (not in generated message), ignore it - u.fields = u.fields[:len(u.fields)-1] - j-- - continue - } - field.computeMarshalFieldInfo(&f) - } - - // fields are marshaled in tag order on the wire. - sort.Sort(byTag(u.fields)) - - atomic.StoreInt32(&u.initialized, 1) -} - -// helper for sorting fields by tag -type byTag []*marshalFieldInfo - -func (a byTag) Len() int { return len(a) } -func (a byTag) Swap(i, j int) { a[i], a[j] = a[j], a[i] } -func (a byTag) Less(i, j int) bool { return a[i].wiretag < a[j].wiretag } - -// getExtElemInfo returns the information to marshal an extension element. -// The info it returns is initialized. -func (u *marshalInfo) getExtElemInfo(desc *ExtensionDesc) *marshalElemInfo { - // get from cache first - u.RLock() - e, ok := u.extElems[desc.Field] - u.RUnlock() - if ok { - return e - } - - t := reflect.TypeOf(desc.ExtensionType) // pointer or slice to basic type or struct - tags := strings.Split(desc.Tag, ",") - tag, err := strconv.Atoi(tags[1]) - if err != nil { - panic("tag is not an integer") - } - wt := wiretype(tags[0]) - if t.Kind() == reflect.Ptr && t.Elem().Kind() != reflect.Struct { - t = t.Elem() - } - sizer, marshaler := typeMarshaler(t, tags, false, false) - var deref bool - if t.Kind() == reflect.Slice && t.Elem().Kind() != reflect.Uint8 { - t = reflect.PtrTo(t) - deref = true - } - e = &marshalElemInfo{ - wiretag: uint64(tag)<<3 | wt, - tagsize: SizeVarint(uint64(tag) << 3), - sizer: sizer, - marshaler: marshaler, - isptr: t.Kind() == reflect.Ptr, - deref: deref, - } - - // update cache - u.Lock() - if u.extElems == nil { - u.extElems = make(map[int32]*marshalElemInfo) - } - u.extElems[desc.Field] = e - u.Unlock() - return e -} - -// computeMarshalFieldInfo fills up the information to marshal a field. -func (fi *marshalFieldInfo) computeMarshalFieldInfo(f *reflect.StructField) { - // parse protobuf tag of the field. - // tag has format of "bytes,49,opt,name=foo,def=hello!" - tags := strings.Split(f.Tag.Get("protobuf"), ",") - if tags[0] == "" { - return - } - tag, err := strconv.Atoi(tags[1]) - if err != nil { - panic("tag is not an integer") - } - wt := wiretype(tags[0]) - if tags[2] == "req" { - fi.required = true - } - fi.setTag(f, tag, wt) - fi.setMarshaler(f, tags) -} - -func (fi *marshalFieldInfo) computeOneofFieldInfo(f *reflect.StructField, oneofImplementers []interface{}) { - fi.field = toField(f) - fi.wiretag = math.MaxInt32 // Use a large tag number, make oneofs sorted at the end. This tag will not appear on the wire. - fi.isPointer = true - fi.sizer, fi.marshaler = makeOneOfMarshaler(fi, f) - fi.oneofElems = make(map[reflect.Type]*marshalElemInfo) - - ityp := f.Type // interface type - for _, o := range oneofImplementers { - t := reflect.TypeOf(o) - if !t.Implements(ityp) { - continue - } - sf := t.Elem().Field(0) // oneof implementer is a struct with a single field - tags := strings.Split(sf.Tag.Get("protobuf"), ",") - tag, err := strconv.Atoi(tags[1]) - if err != nil { - panic("tag is not an integer") - } - wt := wiretype(tags[0]) - sizer, marshaler := typeMarshaler(sf.Type, tags, false, true) // oneof should not omit any zero value - fi.oneofElems[t.Elem()] = &marshalElemInfo{ - wiretag: uint64(tag)<<3 | wt, - tagsize: SizeVarint(uint64(tag) << 3), - sizer: sizer, - marshaler: marshaler, - } - } -} - -// wiretype returns the wire encoding of the type. -func wiretype(encoding string) uint64 { - switch encoding { - case "fixed32": - return WireFixed32 - case "fixed64": - return WireFixed64 - case "varint", "zigzag32", "zigzag64": - return WireVarint - case "bytes": - return WireBytes - case "group": - return WireStartGroup - } - panic("unknown wire type " + encoding) -} - -// setTag fills up the tag (in wire format) and its size in the info of a field. -func (fi *marshalFieldInfo) setTag(f *reflect.StructField, tag int, wt uint64) { - fi.field = toField(f) - fi.wiretag = uint64(tag)<<3 | wt - fi.tagsize = SizeVarint(uint64(tag) << 3) -} - -// setMarshaler fills up the sizer and marshaler in the info of a field. -func (fi *marshalFieldInfo) setMarshaler(f *reflect.StructField, tags []string) { - switch f.Type.Kind() { - case reflect.Map: - // map field - fi.isPointer = true - fi.sizer, fi.marshaler = makeMapMarshaler(f) - return - case reflect.Ptr, reflect.Slice: - fi.isPointer = true - } - fi.sizer, fi.marshaler = typeMarshaler(f.Type, tags, true, false) -} - -// typeMarshaler returns the sizer and marshaler of a given field. -// t is the type of the field. -// tags is the generated "protobuf" tag of the field. -// If nozero is true, zero value is not marshaled to the wire. -// If oneof is true, it is a oneof field. -func typeMarshaler(t reflect.Type, tags []string, nozero, oneof bool) (sizer, marshaler) { - encoding := tags[0] - - pointer := false - slice := false - if t.Kind() == reflect.Slice && t.Elem().Kind() != reflect.Uint8 { - slice = true - t = t.Elem() - } - if t.Kind() == reflect.Ptr { - pointer = true - t = t.Elem() - } - - packed := false - proto3 := false - validateUTF8 := true - for i := 2; i < len(tags); i++ { - if tags[i] == "packed" { - packed = true - } - if tags[i] == "proto3" { - proto3 = true - } - } - validateUTF8 = validateUTF8 && proto3 - - switch t.Kind() { - case reflect.Bool: - if pointer { - return sizeBoolPtr, appendBoolPtr - } - if slice { - if packed { - return sizeBoolPackedSlice, appendBoolPackedSlice - } - return sizeBoolSlice, appendBoolSlice - } - if nozero { - return sizeBoolValueNoZero, appendBoolValueNoZero - } - return sizeBoolValue, appendBoolValue - case reflect.Uint32: - switch encoding { - case "fixed32": - if pointer { - return sizeFixed32Ptr, appendFixed32Ptr - } - if slice { - if packed { - return sizeFixed32PackedSlice, appendFixed32PackedSlice - } - return sizeFixed32Slice, appendFixed32Slice - } - if nozero { - return sizeFixed32ValueNoZero, appendFixed32ValueNoZero - } - return sizeFixed32Value, appendFixed32Value - case "varint": - if pointer { - return sizeVarint32Ptr, appendVarint32Ptr - } - if slice { - if packed { - return sizeVarint32PackedSlice, appendVarint32PackedSlice - } - return sizeVarint32Slice, appendVarint32Slice - } - if nozero { - return sizeVarint32ValueNoZero, appendVarint32ValueNoZero - } - return sizeVarint32Value, appendVarint32Value - } - case reflect.Int32: - switch encoding { - case "fixed32": - if pointer { - return sizeFixedS32Ptr, appendFixedS32Ptr - } - if slice { - if packed { - return sizeFixedS32PackedSlice, appendFixedS32PackedSlice - } - return sizeFixedS32Slice, appendFixedS32Slice - } - if nozero { - return sizeFixedS32ValueNoZero, appendFixedS32ValueNoZero - } - return sizeFixedS32Value, appendFixedS32Value - case "varint": - if pointer { - return sizeVarintS32Ptr, appendVarintS32Ptr - } - if slice { - if packed { - return sizeVarintS32PackedSlice, appendVarintS32PackedSlice - } - return sizeVarintS32Slice, appendVarintS32Slice - } - if nozero { - return sizeVarintS32ValueNoZero, appendVarintS32ValueNoZero - } - return sizeVarintS32Value, appendVarintS32Value - case "zigzag32": - if pointer { - return sizeZigzag32Ptr, appendZigzag32Ptr - } - if slice { - if packed { - return sizeZigzag32PackedSlice, appendZigzag32PackedSlice - } - return sizeZigzag32Slice, appendZigzag32Slice - } - if nozero { - return sizeZigzag32ValueNoZero, appendZigzag32ValueNoZero - } - return sizeZigzag32Value, appendZigzag32Value - } - case reflect.Uint64: - switch encoding { - case "fixed64": - if pointer { - return sizeFixed64Ptr, appendFixed64Ptr - } - if slice { - if packed { - return sizeFixed64PackedSlice, appendFixed64PackedSlice - } - return sizeFixed64Slice, appendFixed64Slice - } - if nozero { - return sizeFixed64ValueNoZero, appendFixed64ValueNoZero - } - return sizeFixed64Value, appendFixed64Value - case "varint": - if pointer { - return sizeVarint64Ptr, appendVarint64Ptr - } - if slice { - if packed { - return sizeVarint64PackedSlice, appendVarint64PackedSlice - } - return sizeVarint64Slice, appendVarint64Slice - } - if nozero { - return sizeVarint64ValueNoZero, appendVarint64ValueNoZero - } - return sizeVarint64Value, appendVarint64Value - } - case reflect.Int64: - switch encoding { - case "fixed64": - if pointer { - return sizeFixedS64Ptr, appendFixedS64Ptr - } - if slice { - if packed { - return sizeFixedS64PackedSlice, appendFixedS64PackedSlice - } - return sizeFixedS64Slice, appendFixedS64Slice - } - if nozero { - return sizeFixedS64ValueNoZero, appendFixedS64ValueNoZero - } - return sizeFixedS64Value, appendFixedS64Value - case "varint": - if pointer { - return sizeVarintS64Ptr, appendVarintS64Ptr - } - if slice { - if packed { - return sizeVarintS64PackedSlice, appendVarintS64PackedSlice - } - return sizeVarintS64Slice, appendVarintS64Slice - } - if nozero { - return sizeVarintS64ValueNoZero, appendVarintS64ValueNoZero - } - return sizeVarintS64Value, appendVarintS64Value - case "zigzag64": - if pointer { - return sizeZigzag64Ptr, appendZigzag64Ptr - } - if slice { - if packed { - return sizeZigzag64PackedSlice, appendZigzag64PackedSlice - } - return sizeZigzag64Slice, appendZigzag64Slice - } - if nozero { - return sizeZigzag64ValueNoZero, appendZigzag64ValueNoZero - } - return sizeZigzag64Value, appendZigzag64Value - } - case reflect.Float32: - if pointer { - return sizeFloat32Ptr, appendFloat32Ptr - } - if slice { - if packed { - return sizeFloat32PackedSlice, appendFloat32PackedSlice - } - return sizeFloat32Slice, appendFloat32Slice - } - if nozero { - return sizeFloat32ValueNoZero, appendFloat32ValueNoZero - } - return sizeFloat32Value, appendFloat32Value - case reflect.Float64: - if pointer { - return sizeFloat64Ptr, appendFloat64Ptr - } - if slice { - if packed { - return sizeFloat64PackedSlice, appendFloat64PackedSlice - } - return sizeFloat64Slice, appendFloat64Slice - } - if nozero { - return sizeFloat64ValueNoZero, appendFloat64ValueNoZero - } - return sizeFloat64Value, appendFloat64Value - case reflect.String: - if validateUTF8 { - if pointer { - return sizeStringPtr, appendUTF8StringPtr - } - if slice { - return sizeStringSlice, appendUTF8StringSlice - } - if nozero { - return sizeStringValueNoZero, appendUTF8StringValueNoZero - } - return sizeStringValue, appendUTF8StringValue - } - if pointer { - return sizeStringPtr, appendStringPtr - } - if slice { - return sizeStringSlice, appendStringSlice - } - if nozero { - return sizeStringValueNoZero, appendStringValueNoZero - } - return sizeStringValue, appendStringValue - case reflect.Slice: - if slice { - return sizeBytesSlice, appendBytesSlice - } - if oneof { - // Oneof bytes field may also have "proto3" tag. - // We want to marshal it as a oneof field. Do this - // check before the proto3 check. - return sizeBytesOneof, appendBytesOneof - } - if proto3 { - return sizeBytes3, appendBytes3 - } - return sizeBytes, appendBytes - case reflect.Struct: - switch encoding { - case "group": - if slice { - return makeGroupSliceMarshaler(getMarshalInfo(t)) - } - return makeGroupMarshaler(getMarshalInfo(t)) - case "bytes": - if slice { - return makeMessageSliceMarshaler(getMarshalInfo(t)) - } - return makeMessageMarshaler(getMarshalInfo(t)) - } - } - panic(fmt.Sprintf("unknown or mismatched type: type: %v, wire type: %v", t, encoding)) -} - -// Below are functions to size/marshal a specific type of a field. -// They are stored in the field's info, and called by function pointers. -// They have type sizer or marshaler. - -func sizeFixed32Value(_ pointer, tagsize int) int { - return 4 + tagsize -} -func sizeFixed32ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toUint32() - if v == 0 { - return 0 - } - return 4 + tagsize -} -func sizeFixed32Ptr(ptr pointer, tagsize int) int { - p := *ptr.toUint32Ptr() - if p == nil { - return 0 - } - return 4 + tagsize -} -func sizeFixed32Slice(ptr pointer, tagsize int) int { - s := *ptr.toUint32Slice() - return (4 + tagsize) * len(s) -} -func sizeFixed32PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toUint32Slice() - if len(s) == 0 { - return 0 - } - return 4*len(s) + SizeVarint(uint64(4*len(s))) + tagsize -} -func sizeFixedS32Value(_ pointer, tagsize int) int { - return 4 + tagsize -} -func sizeFixedS32ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toInt32() - if v == 0 { - return 0 - } - return 4 + tagsize -} -func sizeFixedS32Ptr(ptr pointer, tagsize int) int { - p := ptr.getInt32Ptr() - if p == nil { - return 0 - } - return 4 + tagsize -} -func sizeFixedS32Slice(ptr pointer, tagsize int) int { - s := ptr.getInt32Slice() - return (4 + tagsize) * len(s) -} -func sizeFixedS32PackedSlice(ptr pointer, tagsize int) int { - s := ptr.getInt32Slice() - if len(s) == 0 { - return 0 - } - return 4*len(s) + SizeVarint(uint64(4*len(s))) + tagsize -} -func sizeFloat32Value(_ pointer, tagsize int) int { - return 4 + tagsize -} -func sizeFloat32ValueNoZero(ptr pointer, tagsize int) int { - v := math.Float32bits(*ptr.toFloat32()) - if v == 0 { - return 0 - } - return 4 + tagsize -} -func sizeFloat32Ptr(ptr pointer, tagsize int) int { - p := *ptr.toFloat32Ptr() - if p == nil { - return 0 - } - return 4 + tagsize -} -func sizeFloat32Slice(ptr pointer, tagsize int) int { - s := *ptr.toFloat32Slice() - return (4 + tagsize) * len(s) -} -func sizeFloat32PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toFloat32Slice() - if len(s) == 0 { - return 0 - } - return 4*len(s) + SizeVarint(uint64(4*len(s))) + tagsize -} -func sizeFixed64Value(_ pointer, tagsize int) int { - return 8 + tagsize -} -func sizeFixed64ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toUint64() - if v == 0 { - return 0 - } - return 8 + tagsize -} -func sizeFixed64Ptr(ptr pointer, tagsize int) int { - p := *ptr.toUint64Ptr() - if p == nil { - return 0 - } - return 8 + tagsize -} -func sizeFixed64Slice(ptr pointer, tagsize int) int { - s := *ptr.toUint64Slice() - return (8 + tagsize) * len(s) -} -func sizeFixed64PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toUint64Slice() - if len(s) == 0 { - return 0 - } - return 8*len(s) + SizeVarint(uint64(8*len(s))) + tagsize -} -func sizeFixedS64Value(_ pointer, tagsize int) int { - return 8 + tagsize -} -func sizeFixedS64ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toInt64() - if v == 0 { - return 0 - } - return 8 + tagsize -} -func sizeFixedS64Ptr(ptr pointer, tagsize int) int { - p := *ptr.toInt64Ptr() - if p == nil { - return 0 - } - return 8 + tagsize -} -func sizeFixedS64Slice(ptr pointer, tagsize int) int { - s := *ptr.toInt64Slice() - return (8 + tagsize) * len(s) -} -func sizeFixedS64PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toInt64Slice() - if len(s) == 0 { - return 0 - } - return 8*len(s) + SizeVarint(uint64(8*len(s))) + tagsize -} -func sizeFloat64Value(_ pointer, tagsize int) int { - return 8 + tagsize -} -func sizeFloat64ValueNoZero(ptr pointer, tagsize int) int { - v := math.Float64bits(*ptr.toFloat64()) - if v == 0 { - return 0 - } - return 8 + tagsize -} -func sizeFloat64Ptr(ptr pointer, tagsize int) int { - p := *ptr.toFloat64Ptr() - if p == nil { - return 0 - } - return 8 + tagsize -} -func sizeFloat64Slice(ptr pointer, tagsize int) int { - s := *ptr.toFloat64Slice() - return (8 + tagsize) * len(s) -} -func sizeFloat64PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toFloat64Slice() - if len(s) == 0 { - return 0 - } - return 8*len(s) + SizeVarint(uint64(8*len(s))) + tagsize -} -func sizeVarint32Value(ptr pointer, tagsize int) int { - v := *ptr.toUint32() - return SizeVarint(uint64(v)) + tagsize -} -func sizeVarint32ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toUint32() - if v == 0 { - return 0 - } - return SizeVarint(uint64(v)) + tagsize -} -func sizeVarint32Ptr(ptr pointer, tagsize int) int { - p := *ptr.toUint32Ptr() - if p == nil { - return 0 - } - return SizeVarint(uint64(*p)) + tagsize -} -func sizeVarint32Slice(ptr pointer, tagsize int) int { - s := *ptr.toUint32Slice() - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) + tagsize - } - return n -} -func sizeVarint32PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toUint32Slice() - if len(s) == 0 { - return 0 - } - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) - } - return n + SizeVarint(uint64(n)) + tagsize -} -func sizeVarintS32Value(ptr pointer, tagsize int) int { - v := *ptr.toInt32() - return SizeVarint(uint64(v)) + tagsize -} -func sizeVarintS32ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toInt32() - if v == 0 { - return 0 - } - return SizeVarint(uint64(v)) + tagsize -} -func sizeVarintS32Ptr(ptr pointer, tagsize int) int { - p := ptr.getInt32Ptr() - if p == nil { - return 0 - } - return SizeVarint(uint64(*p)) + tagsize -} -func sizeVarintS32Slice(ptr pointer, tagsize int) int { - s := ptr.getInt32Slice() - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) + tagsize - } - return n -} -func sizeVarintS32PackedSlice(ptr pointer, tagsize int) int { - s := ptr.getInt32Slice() - if len(s) == 0 { - return 0 - } - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) - } - return n + SizeVarint(uint64(n)) + tagsize -} -func sizeVarint64Value(ptr pointer, tagsize int) int { - v := *ptr.toUint64() - return SizeVarint(v) + tagsize -} -func sizeVarint64ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toUint64() - if v == 0 { - return 0 - } - return SizeVarint(v) + tagsize -} -func sizeVarint64Ptr(ptr pointer, tagsize int) int { - p := *ptr.toUint64Ptr() - if p == nil { - return 0 - } - return SizeVarint(*p) + tagsize -} -func sizeVarint64Slice(ptr pointer, tagsize int) int { - s := *ptr.toUint64Slice() - n := 0 - for _, v := range s { - n += SizeVarint(v) + tagsize - } - return n -} -func sizeVarint64PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toUint64Slice() - if len(s) == 0 { - return 0 - } - n := 0 - for _, v := range s { - n += SizeVarint(v) - } - return n + SizeVarint(uint64(n)) + tagsize -} -func sizeVarintS64Value(ptr pointer, tagsize int) int { - v := *ptr.toInt64() - return SizeVarint(uint64(v)) + tagsize -} -func sizeVarintS64ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toInt64() - if v == 0 { - return 0 - } - return SizeVarint(uint64(v)) + tagsize -} -func sizeVarintS64Ptr(ptr pointer, tagsize int) int { - p := *ptr.toInt64Ptr() - if p == nil { - return 0 - } - return SizeVarint(uint64(*p)) + tagsize -} -func sizeVarintS64Slice(ptr pointer, tagsize int) int { - s := *ptr.toInt64Slice() - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) + tagsize - } - return n -} -func sizeVarintS64PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toInt64Slice() - if len(s) == 0 { - return 0 - } - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) - } - return n + SizeVarint(uint64(n)) + tagsize -} -func sizeZigzag32Value(ptr pointer, tagsize int) int { - v := *ptr.toInt32() - return SizeVarint(uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) + tagsize -} -func sizeZigzag32ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toInt32() - if v == 0 { - return 0 - } - return SizeVarint(uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) + tagsize -} -func sizeZigzag32Ptr(ptr pointer, tagsize int) int { - p := ptr.getInt32Ptr() - if p == nil { - return 0 - } - v := *p - return SizeVarint(uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) + tagsize -} -func sizeZigzag32Slice(ptr pointer, tagsize int) int { - s := ptr.getInt32Slice() - n := 0 - for _, v := range s { - n += SizeVarint(uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) + tagsize - } - return n -} -func sizeZigzag32PackedSlice(ptr pointer, tagsize int) int { - s := ptr.getInt32Slice() - if len(s) == 0 { - return 0 - } - n := 0 - for _, v := range s { - n += SizeVarint(uint64((uint32(v) << 1) ^ uint32((int32(v) >> 31)))) - } - return n + SizeVarint(uint64(n)) + tagsize -} -func sizeZigzag64Value(ptr pointer, tagsize int) int { - v := *ptr.toInt64() - return SizeVarint(uint64(v<<1)^uint64((int64(v)>>63))) + tagsize -} -func sizeZigzag64ValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toInt64() - if v == 0 { - return 0 - } - return SizeVarint(uint64(v<<1)^uint64((int64(v)>>63))) + tagsize -} -func sizeZigzag64Ptr(ptr pointer, tagsize int) int { - p := *ptr.toInt64Ptr() - if p == nil { - return 0 - } - v := *p - return SizeVarint(uint64(v<<1)^uint64((int64(v)>>63))) + tagsize -} -func sizeZigzag64Slice(ptr pointer, tagsize int) int { - s := *ptr.toInt64Slice() - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v<<1)^uint64((int64(v)>>63))) + tagsize - } - return n -} -func sizeZigzag64PackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toInt64Slice() - if len(s) == 0 { - return 0 - } - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v<<1) ^ uint64((int64(v) >> 63))) - } - return n + SizeVarint(uint64(n)) + tagsize -} -func sizeBoolValue(_ pointer, tagsize int) int { - return 1 + tagsize -} -func sizeBoolValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toBool() - if !v { - return 0 - } - return 1 + tagsize -} -func sizeBoolPtr(ptr pointer, tagsize int) int { - p := *ptr.toBoolPtr() - if p == nil { - return 0 - } - return 1 + tagsize -} -func sizeBoolSlice(ptr pointer, tagsize int) int { - s := *ptr.toBoolSlice() - return (1 + tagsize) * len(s) -} -func sizeBoolPackedSlice(ptr pointer, tagsize int) int { - s := *ptr.toBoolSlice() - if len(s) == 0 { - return 0 - } - return len(s) + SizeVarint(uint64(len(s))) + tagsize -} -func sizeStringValue(ptr pointer, tagsize int) int { - v := *ptr.toString() - return len(v) + SizeVarint(uint64(len(v))) + tagsize -} -func sizeStringValueNoZero(ptr pointer, tagsize int) int { - v := *ptr.toString() - if v == "" { - return 0 - } - return len(v) + SizeVarint(uint64(len(v))) + tagsize -} -func sizeStringPtr(ptr pointer, tagsize int) int { - p := *ptr.toStringPtr() - if p == nil { - return 0 - } - v := *p - return len(v) + SizeVarint(uint64(len(v))) + tagsize -} -func sizeStringSlice(ptr pointer, tagsize int) int { - s := *ptr.toStringSlice() - n := 0 - for _, v := range s { - n += len(v) + SizeVarint(uint64(len(v))) + tagsize - } - return n -} -func sizeBytes(ptr pointer, tagsize int) int { - v := *ptr.toBytes() - if v == nil { - return 0 - } - return len(v) + SizeVarint(uint64(len(v))) + tagsize -} -func sizeBytes3(ptr pointer, tagsize int) int { - v := *ptr.toBytes() - if len(v) == 0 { - return 0 - } - return len(v) + SizeVarint(uint64(len(v))) + tagsize -} -func sizeBytesOneof(ptr pointer, tagsize int) int { - v := *ptr.toBytes() - return len(v) + SizeVarint(uint64(len(v))) + tagsize -} -func sizeBytesSlice(ptr pointer, tagsize int) int { - s := *ptr.toBytesSlice() - n := 0 - for _, v := range s { - n += len(v) + SizeVarint(uint64(len(v))) + tagsize - } - return n -} - -// appendFixed32 appends an encoded fixed32 to b. -func appendFixed32(b []byte, v uint32) []byte { - b = append(b, - byte(v), - byte(v>>8), - byte(v>>16), - byte(v>>24)) - return b -} - -// appendFixed64 appends an encoded fixed64 to b. -func appendFixed64(b []byte, v uint64) []byte { - b = append(b, - byte(v), - byte(v>>8), - byte(v>>16), - byte(v>>24), - byte(v>>32), - byte(v>>40), - byte(v>>48), - byte(v>>56)) - return b -} - -// appendVarint appends an encoded varint to b. -func appendVarint(b []byte, v uint64) []byte { - // TODO: make 1-byte (maybe 2-byte) case inline-able, once we - // have non-leaf inliner. - switch { - case v < 1<<7: - b = append(b, byte(v)) - case v < 1<<14: - b = append(b, - byte(v&0x7f|0x80), - byte(v>>7)) - case v < 1<<21: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte(v>>14)) - case v < 1<<28: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte((v>>14)&0x7f|0x80), - byte(v>>21)) - case v < 1<<35: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte((v>>14)&0x7f|0x80), - byte((v>>21)&0x7f|0x80), - byte(v>>28)) - case v < 1<<42: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte((v>>14)&0x7f|0x80), - byte((v>>21)&0x7f|0x80), - byte((v>>28)&0x7f|0x80), - byte(v>>35)) - case v < 1<<49: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte((v>>14)&0x7f|0x80), - byte((v>>21)&0x7f|0x80), - byte((v>>28)&0x7f|0x80), - byte((v>>35)&0x7f|0x80), - byte(v>>42)) - case v < 1<<56: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte((v>>14)&0x7f|0x80), - byte((v>>21)&0x7f|0x80), - byte((v>>28)&0x7f|0x80), - byte((v>>35)&0x7f|0x80), - byte((v>>42)&0x7f|0x80), - byte(v>>49)) - case v < 1<<63: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte((v>>14)&0x7f|0x80), - byte((v>>21)&0x7f|0x80), - byte((v>>28)&0x7f|0x80), - byte((v>>35)&0x7f|0x80), - byte((v>>42)&0x7f|0x80), - byte((v>>49)&0x7f|0x80), - byte(v>>56)) - default: - b = append(b, - byte(v&0x7f|0x80), - byte((v>>7)&0x7f|0x80), - byte((v>>14)&0x7f|0x80), - byte((v>>21)&0x7f|0x80), - byte((v>>28)&0x7f|0x80), - byte((v>>35)&0x7f|0x80), - byte((v>>42)&0x7f|0x80), - byte((v>>49)&0x7f|0x80), - byte((v>>56)&0x7f|0x80), - 1) - } - return b -} - -func appendFixed32Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint32() - b = appendVarint(b, wiretag) - b = appendFixed32(b, v) - return b, nil -} -func appendFixed32ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint32() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed32(b, v) - return b, nil -} -func appendFixed32Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toUint32Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed32(b, *p) - return b, nil -} -func appendFixed32Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint32Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendFixed32(b, v) - } - return b, nil -} -func appendFixed32PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint32Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - b = appendVarint(b, uint64(4*len(s))) - for _, v := range s { - b = appendFixed32(b, v) - } - return b, nil -} -func appendFixedS32Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt32() - b = appendVarint(b, wiretag) - b = appendFixed32(b, uint32(v)) - return b, nil -} -func appendFixedS32ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt32() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed32(b, uint32(v)) - return b, nil -} -func appendFixedS32Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := ptr.getInt32Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed32(b, uint32(*p)) - return b, nil -} -func appendFixedS32Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := ptr.getInt32Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendFixed32(b, uint32(v)) - } - return b, nil -} -func appendFixedS32PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := ptr.getInt32Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - b = appendVarint(b, uint64(4*len(s))) - for _, v := range s { - b = appendFixed32(b, uint32(v)) - } - return b, nil -} -func appendFloat32Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := math.Float32bits(*ptr.toFloat32()) - b = appendVarint(b, wiretag) - b = appendFixed32(b, v) - return b, nil -} -func appendFloat32ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := math.Float32bits(*ptr.toFloat32()) - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed32(b, v) - return b, nil -} -func appendFloat32Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toFloat32Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed32(b, math.Float32bits(*p)) - return b, nil -} -func appendFloat32Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toFloat32Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendFixed32(b, math.Float32bits(v)) - } - return b, nil -} -func appendFloat32PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toFloat32Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - b = appendVarint(b, uint64(4*len(s))) - for _, v := range s { - b = appendFixed32(b, math.Float32bits(v)) - } - return b, nil -} -func appendFixed64Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint64() - b = appendVarint(b, wiretag) - b = appendFixed64(b, v) - return b, nil -} -func appendFixed64ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint64() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed64(b, v) - return b, nil -} -func appendFixed64Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toUint64Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed64(b, *p) - return b, nil -} -func appendFixed64Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint64Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendFixed64(b, v) - } - return b, nil -} -func appendFixed64PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint64Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - b = appendVarint(b, uint64(8*len(s))) - for _, v := range s { - b = appendFixed64(b, v) - } - return b, nil -} -func appendFixedS64Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt64() - b = appendVarint(b, wiretag) - b = appendFixed64(b, uint64(v)) - return b, nil -} -func appendFixedS64ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt64() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed64(b, uint64(v)) - return b, nil -} -func appendFixedS64Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toInt64Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed64(b, uint64(*p)) - return b, nil -} -func appendFixedS64Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toInt64Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendFixed64(b, uint64(v)) - } - return b, nil -} -func appendFixedS64PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toInt64Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - b = appendVarint(b, uint64(8*len(s))) - for _, v := range s { - b = appendFixed64(b, uint64(v)) - } - return b, nil -} -func appendFloat64Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := math.Float64bits(*ptr.toFloat64()) - b = appendVarint(b, wiretag) - b = appendFixed64(b, v) - return b, nil -} -func appendFloat64ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := math.Float64bits(*ptr.toFloat64()) - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed64(b, v) - return b, nil -} -func appendFloat64Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toFloat64Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendFixed64(b, math.Float64bits(*p)) - return b, nil -} -func appendFloat64Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toFloat64Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendFixed64(b, math.Float64bits(v)) - } - return b, nil -} -func appendFloat64PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toFloat64Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - b = appendVarint(b, uint64(8*len(s))) - for _, v := range s { - b = appendFixed64(b, math.Float64bits(v)) - } - return b, nil -} -func appendVarint32Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint32() - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - return b, nil -} -func appendVarint32ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint32() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - return b, nil -} -func appendVarint32Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toUint32Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(*p)) - return b, nil -} -func appendVarint32Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint32Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - } - return b, nil -} -func appendVarint32PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint32Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - // compute size - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) - } - b = appendVarint(b, uint64(n)) - for _, v := range s { - b = appendVarint(b, uint64(v)) - } - return b, nil -} -func appendVarintS32Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt32() - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - return b, nil -} -func appendVarintS32ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt32() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - return b, nil -} -func appendVarintS32Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := ptr.getInt32Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(*p)) - return b, nil -} -func appendVarintS32Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := ptr.getInt32Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - } - return b, nil -} -func appendVarintS32PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := ptr.getInt32Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - // compute size - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) - } - b = appendVarint(b, uint64(n)) - for _, v := range s { - b = appendVarint(b, uint64(v)) - } - return b, nil -} -func appendVarint64Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint64() - b = appendVarint(b, wiretag) - b = appendVarint(b, v) - return b, nil -} -func appendVarint64ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toUint64() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, v) - return b, nil -} -func appendVarint64Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toUint64Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, *p) - return b, nil -} -func appendVarint64Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint64Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, v) - } - return b, nil -} -func appendVarint64PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toUint64Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - // compute size - n := 0 - for _, v := range s { - n += SizeVarint(v) - } - b = appendVarint(b, uint64(n)) - for _, v := range s { - b = appendVarint(b, v) - } - return b, nil -} -func appendVarintS64Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt64() - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - return b, nil -} -func appendVarintS64ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt64() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - return b, nil -} -func appendVarintS64Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toInt64Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(*p)) - return b, nil -} -func appendVarintS64Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toInt64Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v)) - } - return b, nil -} -func appendVarintS64PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toInt64Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - // compute size - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v)) - } - b = appendVarint(b, uint64(n)) - for _, v := range s { - b = appendVarint(b, uint64(v)) - } - return b, nil -} -func appendZigzag32Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt32() - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) - return b, nil -} -func appendZigzag32ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt32() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) - return b, nil -} -func appendZigzag32Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := ptr.getInt32Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - v := *p - b = appendVarint(b, uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) - return b, nil -} -func appendZigzag32Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := ptr.getInt32Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) - } - return b, nil -} -func appendZigzag32PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := ptr.getInt32Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - // compute size - n := 0 - for _, v := range s { - n += SizeVarint(uint64((uint32(v) << 1) ^ uint32((int32(v) >> 31)))) - } - b = appendVarint(b, uint64(n)) - for _, v := range s { - b = appendVarint(b, uint64((uint32(v)<<1)^uint32((int32(v)>>31)))) - } - return b, nil -} -func appendZigzag64Value(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt64() - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v<<1)^uint64((int64(v)>>63))) - return b, nil -} -func appendZigzag64ValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toInt64() - if v == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v<<1)^uint64((int64(v)>>63))) - return b, nil -} -func appendZigzag64Ptr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toInt64Ptr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - v := *p - b = appendVarint(b, uint64(v<<1)^uint64((int64(v)>>63))) - return b, nil -} -func appendZigzag64Slice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toInt64Slice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(v<<1)^uint64((int64(v)>>63))) - } - return b, nil -} -func appendZigzag64PackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toInt64Slice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - // compute size - n := 0 - for _, v := range s { - n += SizeVarint(uint64(v<<1) ^ uint64((int64(v) >> 63))) - } - b = appendVarint(b, uint64(n)) - for _, v := range s { - b = appendVarint(b, uint64(v<<1)^uint64((int64(v)>>63))) - } - return b, nil -} -func appendBoolValue(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toBool() - b = appendVarint(b, wiretag) - if v { - b = append(b, 1) - } else { - b = append(b, 0) - } - return b, nil -} -func appendBoolValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toBool() - if !v { - return b, nil - } - b = appendVarint(b, wiretag) - b = append(b, 1) - return b, nil -} - -func appendBoolPtr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toBoolPtr() - if p == nil { - return b, nil - } - b = appendVarint(b, wiretag) - if *p { - b = append(b, 1) - } else { - b = append(b, 0) - } - return b, nil -} -func appendBoolSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toBoolSlice() - for _, v := range s { - b = appendVarint(b, wiretag) - if v { - b = append(b, 1) - } else { - b = append(b, 0) - } - } - return b, nil -} -func appendBoolPackedSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toBoolSlice() - if len(s) == 0 { - return b, nil - } - b = appendVarint(b, wiretag&^7|WireBytes) - b = appendVarint(b, uint64(len(s))) - for _, v := range s { - if v { - b = append(b, 1) - } else { - b = append(b, 0) - } - } - return b, nil -} -func appendStringValue(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toString() - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - return b, nil -} -func appendStringValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toString() - if v == "" { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - return b, nil -} -func appendStringPtr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - p := *ptr.toStringPtr() - if p == nil { - return b, nil - } - v := *p - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - return b, nil -} -func appendStringSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toStringSlice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - } - return b, nil -} -func appendUTF8StringValue(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - var invalidUTF8 bool - v := *ptr.toString() - if !utf8.ValidString(v) { - invalidUTF8 = true - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - if invalidUTF8 { - return b, errInvalidUTF8 - } - return b, nil -} -func appendUTF8StringValueNoZero(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - var invalidUTF8 bool - v := *ptr.toString() - if v == "" { - return b, nil - } - if !utf8.ValidString(v) { - invalidUTF8 = true - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - if invalidUTF8 { - return b, errInvalidUTF8 - } - return b, nil -} -func appendUTF8StringPtr(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - var invalidUTF8 bool - p := *ptr.toStringPtr() - if p == nil { - return b, nil - } - v := *p - if !utf8.ValidString(v) { - invalidUTF8 = true - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - if invalidUTF8 { - return b, errInvalidUTF8 - } - return b, nil -} -func appendUTF8StringSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - var invalidUTF8 bool - s := *ptr.toStringSlice() - for _, v := range s { - if !utf8.ValidString(v) { - invalidUTF8 = true - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - } - if invalidUTF8 { - return b, errInvalidUTF8 - } - return b, nil -} -func appendBytes(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toBytes() - if v == nil { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - return b, nil -} -func appendBytes3(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toBytes() - if len(v) == 0 { - return b, nil - } - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - return b, nil -} -func appendBytesOneof(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - v := *ptr.toBytes() - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - return b, nil -} -func appendBytesSlice(b []byte, ptr pointer, wiretag uint64, _ bool) ([]byte, error) { - s := *ptr.toBytesSlice() - for _, v := range s { - b = appendVarint(b, wiretag) - b = appendVarint(b, uint64(len(v))) - b = append(b, v...) - } - return b, nil -} - -// makeGroupMarshaler returns the sizer and marshaler for a group. -// u is the marshal info of the underlying message. -func makeGroupMarshaler(u *marshalInfo) (sizer, marshaler) { - return func(ptr pointer, tagsize int) int { - p := ptr.getPointer() - if p.isNil() { - return 0 - } - return u.size(p) + 2*tagsize - }, - func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { - p := ptr.getPointer() - if p.isNil() { - return b, nil - } - var err error - b = appendVarint(b, wiretag) // start group - b, err = u.marshal(b, p, deterministic) - b = appendVarint(b, wiretag+(WireEndGroup-WireStartGroup)) // end group - return b, err - } -} - -// makeGroupSliceMarshaler returns the sizer and marshaler for a group slice. -// u is the marshal info of the underlying message. -func makeGroupSliceMarshaler(u *marshalInfo) (sizer, marshaler) { - return func(ptr pointer, tagsize int) int { - s := ptr.getPointerSlice() - n := 0 - for _, v := range s { - if v.isNil() { - continue - } - n += u.size(v) + 2*tagsize - } - return n - }, - func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { - s := ptr.getPointerSlice() - var err error - var nerr nonFatal - for _, v := range s { - if v.isNil() { - return b, errRepeatedHasNil - } - b = appendVarint(b, wiretag) // start group - b, err = u.marshal(b, v, deterministic) - b = appendVarint(b, wiretag+(WireEndGroup-WireStartGroup)) // end group - if !nerr.Merge(err) { - if err == ErrNil { - err = errRepeatedHasNil - } - return b, err - } - } - return b, nerr.E - } -} - -// makeMessageMarshaler returns the sizer and marshaler for a message field. -// u is the marshal info of the message. -func makeMessageMarshaler(u *marshalInfo) (sizer, marshaler) { - return func(ptr pointer, tagsize int) int { - p := ptr.getPointer() - if p.isNil() { - return 0 - } - siz := u.size(p) - return siz + SizeVarint(uint64(siz)) + tagsize - }, - func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { - p := ptr.getPointer() - if p.isNil() { - return b, nil - } - b = appendVarint(b, wiretag) - siz := u.cachedsize(p) - b = appendVarint(b, uint64(siz)) - return u.marshal(b, p, deterministic) - } -} - -// makeMessageSliceMarshaler returns the sizer and marshaler for a message slice. -// u is the marshal info of the message. -func makeMessageSliceMarshaler(u *marshalInfo) (sizer, marshaler) { - return func(ptr pointer, tagsize int) int { - s := ptr.getPointerSlice() - n := 0 - for _, v := range s { - if v.isNil() { - continue - } - siz := u.size(v) - n += siz + SizeVarint(uint64(siz)) + tagsize - } - return n - }, - func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { - s := ptr.getPointerSlice() - var err error - var nerr nonFatal - for _, v := range s { - if v.isNil() { - return b, errRepeatedHasNil - } - b = appendVarint(b, wiretag) - siz := u.cachedsize(v) - b = appendVarint(b, uint64(siz)) - b, err = u.marshal(b, v, deterministic) - - if !nerr.Merge(err) { - if err == ErrNil { - err = errRepeatedHasNil - } - return b, err - } - } - return b, nerr.E - } -} - -// makeMapMarshaler returns the sizer and marshaler for a map field. -// f is the pointer to the reflect data structure of the field. -func makeMapMarshaler(f *reflect.StructField) (sizer, marshaler) { - // figure out key and value type - t := f.Type - keyType := t.Key() - valType := t.Elem() - keyTags := strings.Split(f.Tag.Get("protobuf_key"), ",") - valTags := strings.Split(f.Tag.Get("protobuf_val"), ",") - keySizer, keyMarshaler := typeMarshaler(keyType, keyTags, false, false) // don't omit zero value in map - valSizer, valMarshaler := typeMarshaler(valType, valTags, false, false) // don't omit zero value in map - keyWireTag := 1<<3 | wiretype(keyTags[0]) - valWireTag := 2<<3 | wiretype(valTags[0]) - - // We create an interface to get the addresses of the map key and value. - // If value is pointer-typed, the interface is a direct interface, the - // idata itself is the value. Otherwise, the idata is the pointer to the - // value. - // Key cannot be pointer-typed. - valIsPtr := valType.Kind() == reflect.Ptr - - // If value is a message with nested maps, calling - // valSizer in marshal may be quadratic. We should use - // cached version in marshal (but not in size). - // If value is not message type, we don't have size cache, - // but it cannot be nested either. Just use valSizer. - valCachedSizer := valSizer - if valIsPtr && valType.Elem().Kind() == reflect.Struct { - u := getMarshalInfo(valType.Elem()) - valCachedSizer = func(ptr pointer, tagsize int) int { - // Same as message sizer, but use cache. - p := ptr.getPointer() - if p.isNil() { - return 0 - } - siz := u.cachedsize(p) - return siz + SizeVarint(uint64(siz)) + tagsize - } - } - return func(ptr pointer, tagsize int) int { - m := ptr.asPointerTo(t).Elem() // the map - n := 0 - for _, k := range m.MapKeys() { - ki := k.Interface() - vi := m.MapIndex(k).Interface() - kaddr := toAddrPointer(&ki, false, false) // pointer to key - vaddr := toAddrPointer(&vi, valIsPtr, false) // pointer to value - siz := keySizer(kaddr, 1) + valSizer(vaddr, 1) // tag of key = 1 (size=1), tag of val = 2 (size=1) - n += siz + SizeVarint(uint64(siz)) + tagsize - } - return n - }, - func(b []byte, ptr pointer, tag uint64, deterministic bool) ([]byte, error) { - m := ptr.asPointerTo(t).Elem() // the map - var err error - keys := m.MapKeys() - if len(keys) > 1 && deterministic { - sort.Sort(mapKeys(keys)) - } - - var nerr nonFatal - for _, k := range keys { - ki := k.Interface() - vi := m.MapIndex(k).Interface() - kaddr := toAddrPointer(&ki, false, false) // pointer to key - vaddr := toAddrPointer(&vi, valIsPtr, false) // pointer to value - b = appendVarint(b, tag) - siz := keySizer(kaddr, 1) + valCachedSizer(vaddr, 1) // tag of key = 1 (size=1), tag of val = 2 (size=1) - b = appendVarint(b, uint64(siz)) - b, err = keyMarshaler(b, kaddr, keyWireTag, deterministic) - if !nerr.Merge(err) { - return b, err - } - b, err = valMarshaler(b, vaddr, valWireTag, deterministic) - if err != ErrNil && !nerr.Merge(err) { // allow nil value in map - return b, err - } - } - return b, nerr.E - } -} - -// makeOneOfMarshaler returns the sizer and marshaler for a oneof field. -// fi is the marshal info of the field. -// f is the pointer to the reflect data structure of the field. -func makeOneOfMarshaler(fi *marshalFieldInfo, f *reflect.StructField) (sizer, marshaler) { - // Oneof field is an interface. We need to get the actual data type on the fly. - t := f.Type - return func(ptr pointer, _ int) int { - p := ptr.getInterfacePointer() - if p.isNil() { - return 0 - } - v := ptr.asPointerTo(t).Elem().Elem().Elem() // *interface -> interface -> *struct -> struct - telem := v.Type() - e := fi.oneofElems[telem] - return e.sizer(p, e.tagsize) - }, - func(b []byte, ptr pointer, _ uint64, deterministic bool) ([]byte, error) { - p := ptr.getInterfacePointer() - if p.isNil() { - return b, nil - } - v := ptr.asPointerTo(t).Elem().Elem().Elem() // *interface -> interface -> *struct -> struct - telem := v.Type() - if telem.Field(0).Type.Kind() == reflect.Ptr && p.getPointer().isNil() { - return b, errOneofHasNil - } - e := fi.oneofElems[telem] - return e.marshaler(b, p, e.wiretag, deterministic) - } -} - -// sizeExtensions computes the size of encoded data for a XXX_InternalExtensions field. -func (u *marshalInfo) sizeExtensions(ext *XXX_InternalExtensions) int { - m, mu := ext.extensionsRead() - if m == nil { - return 0 - } - mu.Lock() - - n := 0 - for _, e := range m { - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - n += len(e.enc) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - n += ei.sizer(p, ei.tagsize) - } - mu.Unlock() - return n -} - -// appendExtensions marshals a XXX_InternalExtensions field to the end of byte slice b. -func (u *marshalInfo) appendExtensions(b []byte, ext *XXX_InternalExtensions, deterministic bool) ([]byte, error) { - m, mu := ext.extensionsRead() - if m == nil { - return b, nil - } - mu.Lock() - defer mu.Unlock() - - var err error - var nerr nonFatal - - // Fast-path for common cases: zero or one extensions. - // Don't bother sorting the keys. - if len(m) <= 1 { - for _, e := range m { - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - b = append(b, e.enc...) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - b, err = ei.marshaler(b, p, ei.wiretag, deterministic) - if !nerr.Merge(err) { - return b, err - } - } - return b, nerr.E - } - - // Sort the keys to provide a deterministic encoding. - // Not sure this is required, but the old code does it. - keys := make([]int, 0, len(m)) - for k := range m { - keys = append(keys, int(k)) - } - sort.Ints(keys) - - for _, k := range keys { - e := m[int32(k)] - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - b = append(b, e.enc...) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - b, err = ei.marshaler(b, p, ei.wiretag, deterministic) - if !nerr.Merge(err) { - return b, err - } - } - return b, nerr.E -} - -// message set format is: -// message MessageSet { -// repeated group Item = 1 { -// required int32 type_id = 2; -// required string message = 3; -// }; -// } - -// sizeMessageSet computes the size of encoded data for a XXX_InternalExtensions field -// in message set format (above). -func (u *marshalInfo) sizeMessageSet(ext *XXX_InternalExtensions) int { - m, mu := ext.extensionsRead() - if m == nil { - return 0 - } - mu.Lock() - - n := 0 - for id, e := range m { - n += 2 // start group, end group. tag = 1 (size=1) - n += SizeVarint(uint64(id)) + 1 // type_id, tag = 2 (size=1) - - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - msgWithLen := skipVarint(e.enc) // skip old tag, but leave the length varint - siz := len(msgWithLen) - n += siz + 1 // message, tag = 3 (size=1) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - n += ei.sizer(p, 1) // message, tag = 3 (size=1) - } - mu.Unlock() - return n -} - -// appendMessageSet marshals a XXX_InternalExtensions field in message set format (above) -// to the end of byte slice b. -func (u *marshalInfo) appendMessageSet(b []byte, ext *XXX_InternalExtensions, deterministic bool) ([]byte, error) { - m, mu := ext.extensionsRead() - if m == nil { - return b, nil - } - mu.Lock() - defer mu.Unlock() - - var err error - var nerr nonFatal - - // Fast-path for common cases: zero or one extensions. - // Don't bother sorting the keys. - if len(m) <= 1 { - for id, e := range m { - b = append(b, 1<<3|WireStartGroup) - b = append(b, 2<<3|WireVarint) - b = appendVarint(b, uint64(id)) - - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - msgWithLen := skipVarint(e.enc) // skip old tag, but leave the length varint - b = append(b, 3<<3|WireBytes) - b = append(b, msgWithLen...) - b = append(b, 1<<3|WireEndGroup) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - b, err = ei.marshaler(b, p, 3<<3|WireBytes, deterministic) - if !nerr.Merge(err) { - return b, err - } - b = append(b, 1<<3|WireEndGroup) - } - return b, nerr.E - } - - // Sort the keys to provide a deterministic encoding. - keys := make([]int, 0, len(m)) - for k := range m { - keys = append(keys, int(k)) - } - sort.Ints(keys) - - for _, id := range keys { - e := m[int32(id)] - b = append(b, 1<<3|WireStartGroup) - b = append(b, 2<<3|WireVarint) - b = appendVarint(b, uint64(id)) - - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - msgWithLen := skipVarint(e.enc) // skip old tag, but leave the length varint - b = append(b, 3<<3|WireBytes) - b = append(b, msgWithLen...) - b = append(b, 1<<3|WireEndGroup) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - b, err = ei.marshaler(b, p, 3<<3|WireBytes, deterministic) - b = append(b, 1<<3|WireEndGroup) - if !nerr.Merge(err) { - return b, err - } - } - return b, nerr.E -} - -// sizeV1Extensions computes the size of encoded data for a V1-API extension field. -func (u *marshalInfo) sizeV1Extensions(m map[int32]Extension) int { - if m == nil { - return 0 - } - - n := 0 - for _, e := range m { - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - n += len(e.enc) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - n += ei.sizer(p, ei.tagsize) - } - return n -} - -// appendV1Extensions marshals a V1-API extension field to the end of byte slice b. -func (u *marshalInfo) appendV1Extensions(b []byte, m map[int32]Extension, deterministic bool) ([]byte, error) { - if m == nil { - return b, nil - } - - // Sort the keys to provide a deterministic encoding. - keys := make([]int, 0, len(m)) - for k := range m { - keys = append(keys, int(k)) - } - sort.Ints(keys) - - var err error - var nerr nonFatal - for _, k := range keys { - e := m[int32(k)] - if e.value == nil || e.desc == nil { - // Extension is only in its encoded form. - b = append(b, e.enc...) - continue - } - - // We don't skip extensions that have an encoded form set, - // because the extension value may have been mutated after - // the last time this function was called. - - ei := u.getExtElemInfo(e.desc) - v := e.value - p := toAddrPointer(&v, ei.isptr, ei.deref) - b, err = ei.marshaler(b, p, ei.wiretag, deterministic) - if !nerr.Merge(err) { - return b, err - } - } - return b, nerr.E -} - -// newMarshaler is the interface representing objects that can marshal themselves. -// -// This exists to support protoc-gen-go generated messages. -// The proto package will stop type-asserting to this interface in the future. -// -// DO NOT DEPEND ON THIS. -type newMarshaler interface { - XXX_Size() int - XXX_Marshal(b []byte, deterministic bool) ([]byte, error) -} - -// Size returns the encoded size of a protocol buffer message. -// This is the main entry point. -func Size(pb Message) int { - if m, ok := pb.(newMarshaler); ok { - return m.XXX_Size() - } - if m, ok := pb.(Marshaler); ok { - // If the message can marshal itself, let it do it, for compatibility. - // NOTE: This is not efficient. - b, _ := m.Marshal() - return len(b) - } - // in case somehow we didn't generate the wrapper - if pb == nil { - return 0 - } - var info InternalMessageInfo - return info.Size(pb) -} - -// Marshal takes a protocol buffer message -// and encodes it into the wire format, returning the data. -// This is the main entry point. -func Marshal(pb Message) ([]byte, error) { - if m, ok := pb.(newMarshaler); ok { - siz := m.XXX_Size() - b := make([]byte, 0, siz) - return m.XXX_Marshal(b, false) - } - if m, ok := pb.(Marshaler); ok { - // If the message can marshal itself, let it do it, for compatibility. - // NOTE: This is not efficient. - return m.Marshal() - } - // in case somehow we didn't generate the wrapper - if pb == nil { - return nil, ErrNil - } - var info InternalMessageInfo - siz := info.Size(pb) - b := make([]byte, 0, siz) - return info.Marshal(b, pb, false) -} - -// Marshal takes a protocol buffer message -// and encodes it into the wire format, writing the result to the -// Buffer. -// This is an alternative entry point. It is not necessary to use -// a Buffer for most applications. -func (p *Buffer) Marshal(pb Message) error { - var err error - if m, ok := pb.(newMarshaler); ok { - siz := m.XXX_Size() - p.grow(siz) // make sure buf has enough capacity - p.buf, err = m.XXX_Marshal(p.buf, p.deterministic) - return err - } - if m, ok := pb.(Marshaler); ok { - // If the message can marshal itself, let it do it, for compatibility. - // NOTE: This is not efficient. - b, err := m.Marshal() - p.buf = append(p.buf, b...) - return err - } - // in case somehow we didn't generate the wrapper - if pb == nil { - return ErrNil - } - var info InternalMessageInfo - siz := info.Size(pb) - p.grow(siz) // make sure buf has enough capacity - p.buf, err = info.Marshal(p.buf, pb, p.deterministic) - return err -} - -// grow grows the buffer's capacity, if necessary, to guarantee space for -// another n bytes. After grow(n), at least n bytes can be written to the -// buffer without another allocation. -func (p *Buffer) grow(n int) { - need := len(p.buf) + n - if need <= cap(p.buf) { - return - } - newCap := len(p.buf) * 2 - if newCap < need { - newCap = need - } - p.buf = append(make([]byte, 0, newCap), p.buf...) -} diff --git a/vendor/github.com/golang/protobuf/proto/table_merge.go b/vendor/github.com/golang/protobuf/proto/table_merge.go deleted file mode 100644 index 5525def6..00000000 --- a/vendor/github.com/golang/protobuf/proto/table_merge.go +++ /dev/null @@ -1,654 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2016 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -import ( - "fmt" - "reflect" - "strings" - "sync" - "sync/atomic" -) - -// Merge merges the src message into dst. -// This assumes that dst and src of the same type and are non-nil. -func (a *InternalMessageInfo) Merge(dst, src Message) { - mi := atomicLoadMergeInfo(&a.merge) - if mi == nil { - mi = getMergeInfo(reflect.TypeOf(dst).Elem()) - atomicStoreMergeInfo(&a.merge, mi) - } - mi.merge(toPointer(&dst), toPointer(&src)) -} - -type mergeInfo struct { - typ reflect.Type - - initialized int32 // 0: only typ is valid, 1: everything is valid - lock sync.Mutex - - fields []mergeFieldInfo - unrecognized field // Offset of XXX_unrecognized -} - -type mergeFieldInfo struct { - field field // Offset of field, guaranteed to be valid - - // isPointer reports whether the value in the field is a pointer. - // This is true for the following situations: - // * Pointer to struct - // * Pointer to basic type (proto2 only) - // * Slice (first value in slice header is a pointer) - // * String (first value in string header is a pointer) - isPointer bool - - // basicWidth reports the width of the field assuming that it is directly - // embedded in the struct (as is the case for basic types in proto3). - // The possible values are: - // 0: invalid - // 1: bool - // 4: int32, uint32, float32 - // 8: int64, uint64, float64 - basicWidth int - - // Where dst and src are pointers to the types being merged. - merge func(dst, src pointer) -} - -var ( - mergeInfoMap = map[reflect.Type]*mergeInfo{} - mergeInfoLock sync.Mutex -) - -func getMergeInfo(t reflect.Type) *mergeInfo { - mergeInfoLock.Lock() - defer mergeInfoLock.Unlock() - mi := mergeInfoMap[t] - if mi == nil { - mi = &mergeInfo{typ: t} - mergeInfoMap[t] = mi - } - return mi -} - -// merge merges src into dst assuming they are both of type *mi.typ. -func (mi *mergeInfo) merge(dst, src pointer) { - if dst.isNil() { - panic("proto: nil destination") - } - if src.isNil() { - return // Nothing to do. - } - - if atomic.LoadInt32(&mi.initialized) == 0 { - mi.computeMergeInfo() - } - - for _, fi := range mi.fields { - sfp := src.offset(fi.field) - - // As an optimization, we can avoid the merge function call cost - // if we know for sure that the source will have no effect - // by checking if it is the zero value. - if unsafeAllowed { - if fi.isPointer && sfp.getPointer().isNil() { // Could be slice or string - continue - } - if fi.basicWidth > 0 { - switch { - case fi.basicWidth == 1 && !*sfp.toBool(): - continue - case fi.basicWidth == 4 && *sfp.toUint32() == 0: - continue - case fi.basicWidth == 8 && *sfp.toUint64() == 0: - continue - } - } - } - - dfp := dst.offset(fi.field) - fi.merge(dfp, sfp) - } - - // TODO: Make this faster? - out := dst.asPointerTo(mi.typ).Elem() - in := src.asPointerTo(mi.typ).Elem() - if emIn, err := extendable(in.Addr().Interface()); err == nil { - emOut, _ := extendable(out.Addr().Interface()) - mIn, muIn := emIn.extensionsRead() - if mIn != nil { - mOut := emOut.extensionsWrite() - muIn.Lock() - mergeExtension(mOut, mIn) - muIn.Unlock() - } - } - - if mi.unrecognized.IsValid() { - if b := *src.offset(mi.unrecognized).toBytes(); len(b) > 0 { - *dst.offset(mi.unrecognized).toBytes() = append([]byte(nil), b...) - } - } -} - -func (mi *mergeInfo) computeMergeInfo() { - mi.lock.Lock() - defer mi.lock.Unlock() - if mi.initialized != 0 { - return - } - t := mi.typ - n := t.NumField() - - props := GetProperties(t) - for i := 0; i < n; i++ { - f := t.Field(i) - if strings.HasPrefix(f.Name, "XXX_") { - continue - } - - mfi := mergeFieldInfo{field: toField(&f)} - tf := f.Type - - // As an optimization, we can avoid the merge function call cost - // if we know for sure that the source will have no effect - // by checking if it is the zero value. - if unsafeAllowed { - switch tf.Kind() { - case reflect.Ptr, reflect.Slice, reflect.String: - // As a special case, we assume slices and strings are pointers - // since we know that the first field in the SliceSlice or - // StringHeader is a data pointer. - mfi.isPointer = true - case reflect.Bool: - mfi.basicWidth = 1 - case reflect.Int32, reflect.Uint32, reflect.Float32: - mfi.basicWidth = 4 - case reflect.Int64, reflect.Uint64, reflect.Float64: - mfi.basicWidth = 8 - } - } - - // Unwrap tf to get at its most basic type. - var isPointer, isSlice bool - if tf.Kind() == reflect.Slice && tf.Elem().Kind() != reflect.Uint8 { - isSlice = true - tf = tf.Elem() - } - if tf.Kind() == reflect.Ptr { - isPointer = true - tf = tf.Elem() - } - if isPointer && isSlice && tf.Kind() != reflect.Struct { - panic("both pointer and slice for basic type in " + tf.Name()) - } - - switch tf.Kind() { - case reflect.Int32: - switch { - case isSlice: // E.g., []int32 - mfi.merge = func(dst, src pointer) { - // NOTE: toInt32Slice is not defined (see pointer_reflect.go). - /* - sfsp := src.toInt32Slice() - if *sfsp != nil { - dfsp := dst.toInt32Slice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []int64{} - } - } - */ - sfs := src.getInt32Slice() - if sfs != nil { - dfs := dst.getInt32Slice() - dfs = append(dfs, sfs...) - if dfs == nil { - dfs = []int32{} - } - dst.setInt32Slice(dfs) - } - } - case isPointer: // E.g., *int32 - mfi.merge = func(dst, src pointer) { - // NOTE: toInt32Ptr is not defined (see pointer_reflect.go). - /* - sfpp := src.toInt32Ptr() - if *sfpp != nil { - dfpp := dst.toInt32Ptr() - if *dfpp == nil { - *dfpp = Int32(**sfpp) - } else { - **dfpp = **sfpp - } - } - */ - sfp := src.getInt32Ptr() - if sfp != nil { - dfp := dst.getInt32Ptr() - if dfp == nil { - dst.setInt32Ptr(*sfp) - } else { - *dfp = *sfp - } - } - } - default: // E.g., int32 - mfi.merge = func(dst, src pointer) { - if v := *src.toInt32(); v != 0 { - *dst.toInt32() = v - } - } - } - case reflect.Int64: - switch { - case isSlice: // E.g., []int64 - mfi.merge = func(dst, src pointer) { - sfsp := src.toInt64Slice() - if *sfsp != nil { - dfsp := dst.toInt64Slice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []int64{} - } - } - } - case isPointer: // E.g., *int64 - mfi.merge = func(dst, src pointer) { - sfpp := src.toInt64Ptr() - if *sfpp != nil { - dfpp := dst.toInt64Ptr() - if *dfpp == nil { - *dfpp = Int64(**sfpp) - } else { - **dfpp = **sfpp - } - } - } - default: // E.g., int64 - mfi.merge = func(dst, src pointer) { - if v := *src.toInt64(); v != 0 { - *dst.toInt64() = v - } - } - } - case reflect.Uint32: - switch { - case isSlice: // E.g., []uint32 - mfi.merge = func(dst, src pointer) { - sfsp := src.toUint32Slice() - if *sfsp != nil { - dfsp := dst.toUint32Slice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []uint32{} - } - } - } - case isPointer: // E.g., *uint32 - mfi.merge = func(dst, src pointer) { - sfpp := src.toUint32Ptr() - if *sfpp != nil { - dfpp := dst.toUint32Ptr() - if *dfpp == nil { - *dfpp = Uint32(**sfpp) - } else { - **dfpp = **sfpp - } - } - } - default: // E.g., uint32 - mfi.merge = func(dst, src pointer) { - if v := *src.toUint32(); v != 0 { - *dst.toUint32() = v - } - } - } - case reflect.Uint64: - switch { - case isSlice: // E.g., []uint64 - mfi.merge = func(dst, src pointer) { - sfsp := src.toUint64Slice() - if *sfsp != nil { - dfsp := dst.toUint64Slice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []uint64{} - } - } - } - case isPointer: // E.g., *uint64 - mfi.merge = func(dst, src pointer) { - sfpp := src.toUint64Ptr() - if *sfpp != nil { - dfpp := dst.toUint64Ptr() - if *dfpp == nil { - *dfpp = Uint64(**sfpp) - } else { - **dfpp = **sfpp - } - } - } - default: // E.g., uint64 - mfi.merge = func(dst, src pointer) { - if v := *src.toUint64(); v != 0 { - *dst.toUint64() = v - } - } - } - case reflect.Float32: - switch { - case isSlice: // E.g., []float32 - mfi.merge = func(dst, src pointer) { - sfsp := src.toFloat32Slice() - if *sfsp != nil { - dfsp := dst.toFloat32Slice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []float32{} - } - } - } - case isPointer: // E.g., *float32 - mfi.merge = func(dst, src pointer) { - sfpp := src.toFloat32Ptr() - if *sfpp != nil { - dfpp := dst.toFloat32Ptr() - if *dfpp == nil { - *dfpp = Float32(**sfpp) - } else { - **dfpp = **sfpp - } - } - } - default: // E.g., float32 - mfi.merge = func(dst, src pointer) { - if v := *src.toFloat32(); v != 0 { - *dst.toFloat32() = v - } - } - } - case reflect.Float64: - switch { - case isSlice: // E.g., []float64 - mfi.merge = func(dst, src pointer) { - sfsp := src.toFloat64Slice() - if *sfsp != nil { - dfsp := dst.toFloat64Slice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []float64{} - } - } - } - case isPointer: // E.g., *float64 - mfi.merge = func(dst, src pointer) { - sfpp := src.toFloat64Ptr() - if *sfpp != nil { - dfpp := dst.toFloat64Ptr() - if *dfpp == nil { - *dfpp = Float64(**sfpp) - } else { - **dfpp = **sfpp - } - } - } - default: // E.g., float64 - mfi.merge = func(dst, src pointer) { - if v := *src.toFloat64(); v != 0 { - *dst.toFloat64() = v - } - } - } - case reflect.Bool: - switch { - case isSlice: // E.g., []bool - mfi.merge = func(dst, src pointer) { - sfsp := src.toBoolSlice() - if *sfsp != nil { - dfsp := dst.toBoolSlice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []bool{} - } - } - } - case isPointer: // E.g., *bool - mfi.merge = func(dst, src pointer) { - sfpp := src.toBoolPtr() - if *sfpp != nil { - dfpp := dst.toBoolPtr() - if *dfpp == nil { - *dfpp = Bool(**sfpp) - } else { - **dfpp = **sfpp - } - } - } - default: // E.g., bool - mfi.merge = func(dst, src pointer) { - if v := *src.toBool(); v { - *dst.toBool() = v - } - } - } - case reflect.String: - switch { - case isSlice: // E.g., []string - mfi.merge = func(dst, src pointer) { - sfsp := src.toStringSlice() - if *sfsp != nil { - dfsp := dst.toStringSlice() - *dfsp = append(*dfsp, *sfsp...) - if *dfsp == nil { - *dfsp = []string{} - } - } - } - case isPointer: // E.g., *string - mfi.merge = func(dst, src pointer) { - sfpp := src.toStringPtr() - if *sfpp != nil { - dfpp := dst.toStringPtr() - if *dfpp == nil { - *dfpp = String(**sfpp) - } else { - **dfpp = **sfpp - } - } - } - default: // E.g., string - mfi.merge = func(dst, src pointer) { - if v := *src.toString(); v != "" { - *dst.toString() = v - } - } - } - case reflect.Slice: - isProto3 := props.Prop[i].proto3 - switch { - case isPointer: - panic("bad pointer in byte slice case in " + tf.Name()) - case tf.Elem().Kind() != reflect.Uint8: - panic("bad element kind in byte slice case in " + tf.Name()) - case isSlice: // E.g., [][]byte - mfi.merge = func(dst, src pointer) { - sbsp := src.toBytesSlice() - if *sbsp != nil { - dbsp := dst.toBytesSlice() - for _, sb := range *sbsp { - if sb == nil { - *dbsp = append(*dbsp, nil) - } else { - *dbsp = append(*dbsp, append([]byte{}, sb...)) - } - } - if *dbsp == nil { - *dbsp = [][]byte{} - } - } - } - default: // E.g., []byte - mfi.merge = func(dst, src pointer) { - sbp := src.toBytes() - if *sbp != nil { - dbp := dst.toBytes() - if !isProto3 || len(*sbp) > 0 { - *dbp = append([]byte{}, *sbp...) - } - } - } - } - case reflect.Struct: - switch { - case !isPointer: - panic(fmt.Sprintf("message field %s without pointer", tf)) - case isSlice: // E.g., []*pb.T - mi := getMergeInfo(tf) - mfi.merge = func(dst, src pointer) { - sps := src.getPointerSlice() - if sps != nil { - dps := dst.getPointerSlice() - for _, sp := range sps { - var dp pointer - if !sp.isNil() { - dp = valToPointer(reflect.New(tf)) - mi.merge(dp, sp) - } - dps = append(dps, dp) - } - if dps == nil { - dps = []pointer{} - } - dst.setPointerSlice(dps) - } - } - default: // E.g., *pb.T - mi := getMergeInfo(tf) - mfi.merge = func(dst, src pointer) { - sp := src.getPointer() - if !sp.isNil() { - dp := dst.getPointer() - if dp.isNil() { - dp = valToPointer(reflect.New(tf)) - dst.setPointer(dp) - } - mi.merge(dp, sp) - } - } - } - case reflect.Map: - switch { - case isPointer || isSlice: - panic("bad pointer or slice in map case in " + tf.Name()) - default: // E.g., map[K]V - mfi.merge = func(dst, src pointer) { - sm := src.asPointerTo(tf).Elem() - if sm.Len() == 0 { - return - } - dm := dst.asPointerTo(tf).Elem() - if dm.IsNil() { - dm.Set(reflect.MakeMap(tf)) - } - - switch tf.Elem().Kind() { - case reflect.Ptr: // Proto struct (e.g., *T) - for _, key := range sm.MapKeys() { - val := sm.MapIndex(key) - val = reflect.ValueOf(Clone(val.Interface().(Message))) - dm.SetMapIndex(key, val) - } - case reflect.Slice: // E.g. Bytes type (e.g., []byte) - for _, key := range sm.MapKeys() { - val := sm.MapIndex(key) - val = reflect.ValueOf(append([]byte{}, val.Bytes()...)) - dm.SetMapIndex(key, val) - } - default: // Basic type (e.g., string) - for _, key := range sm.MapKeys() { - val := sm.MapIndex(key) - dm.SetMapIndex(key, val) - } - } - } - } - case reflect.Interface: - // Must be oneof field. - switch { - case isPointer || isSlice: - panic("bad pointer or slice in interface case in " + tf.Name()) - default: // E.g., interface{} - // TODO: Make this faster? - mfi.merge = func(dst, src pointer) { - su := src.asPointerTo(tf).Elem() - if !su.IsNil() { - du := dst.asPointerTo(tf).Elem() - typ := su.Elem().Type() - if du.IsNil() || du.Elem().Type() != typ { - du.Set(reflect.New(typ.Elem())) // Initialize interface if empty - } - sv := su.Elem().Elem().Field(0) - if sv.Kind() == reflect.Ptr && sv.IsNil() { - return - } - dv := du.Elem().Elem().Field(0) - if dv.Kind() == reflect.Ptr && dv.IsNil() { - dv.Set(reflect.New(sv.Type().Elem())) // Initialize proto message if empty - } - switch sv.Type().Kind() { - case reflect.Ptr: // Proto struct (e.g., *T) - Merge(dv.Interface().(Message), sv.Interface().(Message)) - case reflect.Slice: // E.g. Bytes type (e.g., []byte) - dv.Set(reflect.ValueOf(append([]byte{}, sv.Bytes()...))) - default: // Basic type (e.g., string) - dv.Set(sv) - } - } - } - } - default: - panic(fmt.Sprintf("merger not found for type:%s", tf)) - } - mi.fields = append(mi.fields, mfi) - } - - mi.unrecognized = invalidField - if f, ok := t.FieldByName("XXX_unrecognized"); ok { - if f.Type != reflect.TypeOf([]byte{}) { - panic("expected XXX_unrecognized to be of type []byte") - } - mi.unrecognized = toField(&f) - } - - atomic.StoreInt32(&mi.initialized, 1) -} diff --git a/vendor/github.com/golang/protobuf/proto/table_unmarshal.go b/vendor/github.com/golang/protobuf/proto/table_unmarshal.go deleted file mode 100644 index acee2fc5..00000000 --- a/vendor/github.com/golang/protobuf/proto/table_unmarshal.go +++ /dev/null @@ -1,2053 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2016 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -import ( - "errors" - "fmt" - "io" - "math" - "reflect" - "strconv" - "strings" - "sync" - "sync/atomic" - "unicode/utf8" -) - -// Unmarshal is the entry point from the generated .pb.go files. -// This function is not intended to be used by non-generated code. -// This function is not subject to any compatibility guarantee. -// msg contains a pointer to a protocol buffer struct. -// b is the data to be unmarshaled into the protocol buffer. -// a is a pointer to a place to store cached unmarshal information. -func (a *InternalMessageInfo) Unmarshal(msg Message, b []byte) error { - // Load the unmarshal information for this message type. - // The atomic load ensures memory consistency. - u := atomicLoadUnmarshalInfo(&a.unmarshal) - if u == nil { - // Slow path: find unmarshal info for msg, update a with it. - u = getUnmarshalInfo(reflect.TypeOf(msg).Elem()) - atomicStoreUnmarshalInfo(&a.unmarshal, u) - } - // Then do the unmarshaling. - err := u.unmarshal(toPointer(&msg), b) - return err -} - -type unmarshalInfo struct { - typ reflect.Type // type of the protobuf struct - - // 0 = only typ field is initialized - // 1 = completely initialized - initialized int32 - lock sync.Mutex // prevents double initialization - dense []unmarshalFieldInfo // fields indexed by tag # - sparse map[uint64]unmarshalFieldInfo // fields indexed by tag # - reqFields []string // names of required fields - reqMask uint64 // 1< 0 { - // Read tag and wire type. - // Special case 1 and 2 byte varints. - var x uint64 - if b[0] < 128 { - x = uint64(b[0]) - b = b[1:] - } else if len(b) >= 2 && b[1] < 128 { - x = uint64(b[0]&0x7f) + uint64(b[1])<<7 - b = b[2:] - } else { - var n int - x, n = decodeVarint(b) - if n == 0 { - return io.ErrUnexpectedEOF - } - b = b[n:] - } - tag := x >> 3 - wire := int(x) & 7 - - // Dispatch on the tag to one of the unmarshal* functions below. - var f unmarshalFieldInfo - if tag < uint64(len(u.dense)) { - f = u.dense[tag] - } else { - f = u.sparse[tag] - } - if fn := f.unmarshal; fn != nil { - var err error - b, err = fn(b, m.offset(f.field), wire) - if err == nil { - reqMask |= f.reqMask - continue - } - if r, ok := err.(*RequiredNotSetError); ok { - // Remember this error, but keep parsing. We need to produce - // a full parse even if a required field is missing. - if errLater == nil { - errLater = r - } - reqMask |= f.reqMask - continue - } - if err != errInternalBadWireType { - if err == errInvalidUTF8 { - if errLater == nil { - fullName := revProtoTypes[reflect.PtrTo(u.typ)] + "." + f.name - errLater = &invalidUTF8Error{fullName} - } - continue - } - return err - } - // Fragments with bad wire type are treated as unknown fields. - } - - // Unknown tag. - if !u.unrecognized.IsValid() { - // Don't keep unrecognized data; just skip it. - var err error - b, err = skipField(b, wire) - if err != nil { - return err - } - continue - } - // Keep unrecognized data around. - // maybe in extensions, maybe in the unrecognized field. - z := m.offset(u.unrecognized).toBytes() - var emap map[int32]Extension - var e Extension - for _, r := range u.extensionRanges { - if uint64(r.Start) <= tag && tag <= uint64(r.End) { - if u.extensions.IsValid() { - mp := m.offset(u.extensions).toExtensions() - emap = mp.extensionsWrite() - e = emap[int32(tag)] - z = &e.enc - break - } - if u.oldExtensions.IsValid() { - p := m.offset(u.oldExtensions).toOldExtensions() - emap = *p - if emap == nil { - emap = map[int32]Extension{} - *p = emap - } - e = emap[int32(tag)] - z = &e.enc - break - } - panic("no extensions field available") - } - } - - // Use wire type to skip data. - var err error - b0 := b - b, err = skipField(b, wire) - if err != nil { - return err - } - *z = encodeVarint(*z, tag<<3|uint64(wire)) - *z = append(*z, b0[:len(b0)-len(b)]...) - - if emap != nil { - emap[int32(tag)] = e - } - } - if reqMask != u.reqMask && errLater == nil { - // A required field of this message is missing. - for _, n := range u.reqFields { - if reqMask&1 == 0 { - errLater = &RequiredNotSetError{n} - } - reqMask >>= 1 - } - } - return errLater -} - -// computeUnmarshalInfo fills in u with information for use -// in unmarshaling protocol buffers of type u.typ. -func (u *unmarshalInfo) computeUnmarshalInfo() { - u.lock.Lock() - defer u.lock.Unlock() - if u.initialized != 0 { - return - } - t := u.typ - n := t.NumField() - - // Set up the "not found" value for the unrecognized byte buffer. - // This is the default for proto3. - u.unrecognized = invalidField - u.extensions = invalidField - u.oldExtensions = invalidField - - // List of the generated type and offset for each oneof field. - type oneofField struct { - ityp reflect.Type // interface type of oneof field - field field // offset in containing message - } - var oneofFields []oneofField - - for i := 0; i < n; i++ { - f := t.Field(i) - if f.Name == "XXX_unrecognized" { - // The byte slice used to hold unrecognized input is special. - if f.Type != reflect.TypeOf(([]byte)(nil)) { - panic("bad type for XXX_unrecognized field: " + f.Type.Name()) - } - u.unrecognized = toField(&f) - continue - } - if f.Name == "XXX_InternalExtensions" { - // Ditto here. - if f.Type != reflect.TypeOf(XXX_InternalExtensions{}) { - panic("bad type for XXX_InternalExtensions field: " + f.Type.Name()) - } - u.extensions = toField(&f) - if f.Tag.Get("protobuf_messageset") == "1" { - u.isMessageSet = true - } - continue - } - if f.Name == "XXX_extensions" { - // An older form of the extensions field. - if f.Type != reflect.TypeOf((map[int32]Extension)(nil)) { - panic("bad type for XXX_extensions field: " + f.Type.Name()) - } - u.oldExtensions = toField(&f) - continue - } - if f.Name == "XXX_NoUnkeyedLiteral" || f.Name == "XXX_sizecache" { - continue - } - - oneof := f.Tag.Get("protobuf_oneof") - if oneof != "" { - oneofFields = append(oneofFields, oneofField{f.Type, toField(&f)}) - // The rest of oneof processing happens below. - continue - } - - tags := f.Tag.Get("protobuf") - tagArray := strings.Split(tags, ",") - if len(tagArray) < 2 { - panic("protobuf tag not enough fields in " + t.Name() + "." + f.Name + ": " + tags) - } - tag, err := strconv.Atoi(tagArray[1]) - if err != nil { - panic("protobuf tag field not an integer: " + tagArray[1]) - } - - name := "" - for _, tag := range tagArray[3:] { - if strings.HasPrefix(tag, "name=") { - name = tag[5:] - } - } - - // Extract unmarshaling function from the field (its type and tags). - unmarshal := fieldUnmarshaler(&f) - - // Required field? - var reqMask uint64 - if tagArray[2] == "req" { - bit := len(u.reqFields) - u.reqFields = append(u.reqFields, name) - reqMask = uint64(1) << uint(bit) - // TODO: if we have more than 64 required fields, we end up - // not verifying that all required fields are present. - // Fix this, perhaps using a count of required fields? - } - - // Store the info in the correct slot in the message. - u.setTag(tag, toField(&f), unmarshal, reqMask, name) - } - - // Find any types associated with oneof fields. - var oneofImplementers []interface{} - switch m := reflect.Zero(reflect.PtrTo(t)).Interface().(type) { - case oneofFuncsIface: - _, _, _, oneofImplementers = m.XXX_OneofFuncs() - case oneofWrappersIface: - oneofImplementers = m.XXX_OneofWrappers() - } - for _, v := range oneofImplementers { - tptr := reflect.TypeOf(v) // *Msg_X - typ := tptr.Elem() // Msg_X - - f := typ.Field(0) // oneof implementers have one field - baseUnmarshal := fieldUnmarshaler(&f) - tags := strings.Split(f.Tag.Get("protobuf"), ",") - fieldNum, err := strconv.Atoi(tags[1]) - if err != nil { - panic("protobuf tag field not an integer: " + tags[1]) - } - var name string - for _, tag := range tags { - if strings.HasPrefix(tag, "name=") { - name = strings.TrimPrefix(tag, "name=") - break - } - } - - // Find the oneof field that this struct implements. - // Might take O(n^2) to process all of the oneofs, but who cares. - for _, of := range oneofFields { - if tptr.Implements(of.ityp) { - // We have found the corresponding interface for this struct. - // That lets us know where this struct should be stored - // when we encounter it during unmarshaling. - unmarshal := makeUnmarshalOneof(typ, of.ityp, baseUnmarshal) - u.setTag(fieldNum, of.field, unmarshal, 0, name) - } - } - - } - - // Get extension ranges, if any. - fn := reflect.Zero(reflect.PtrTo(t)).MethodByName("ExtensionRangeArray") - if fn.IsValid() { - if !u.extensions.IsValid() && !u.oldExtensions.IsValid() { - panic("a message with extensions, but no extensions field in " + t.Name()) - } - u.extensionRanges = fn.Call(nil)[0].Interface().([]ExtensionRange) - } - - // Explicitly disallow tag 0. This will ensure we flag an error - // when decoding a buffer of all zeros. Without this code, we - // would decode and skip an all-zero buffer of even length. - // [0 0] is [tag=0/wiretype=varint varint-encoded-0]. - u.setTag(0, zeroField, func(b []byte, f pointer, w int) ([]byte, error) { - return nil, fmt.Errorf("proto: %s: illegal tag 0 (wire type %d)", t, w) - }, 0, "") - - // Set mask for required field check. - u.reqMask = uint64(1)<= 0 && (tag < 16 || tag < 2*n) { // TODO: what are the right numbers here? - for len(u.dense) <= tag { - u.dense = append(u.dense, unmarshalFieldInfo{}) - } - u.dense[tag] = i - return - } - if u.sparse == nil { - u.sparse = map[uint64]unmarshalFieldInfo{} - } - u.sparse[uint64(tag)] = i -} - -// fieldUnmarshaler returns an unmarshaler for the given field. -func fieldUnmarshaler(f *reflect.StructField) unmarshaler { - if f.Type.Kind() == reflect.Map { - return makeUnmarshalMap(f) - } - return typeUnmarshaler(f.Type, f.Tag.Get("protobuf")) -} - -// typeUnmarshaler returns an unmarshaler for the given field type / field tag pair. -func typeUnmarshaler(t reflect.Type, tags string) unmarshaler { - tagArray := strings.Split(tags, ",") - encoding := tagArray[0] - name := "unknown" - proto3 := false - validateUTF8 := true - for _, tag := range tagArray[3:] { - if strings.HasPrefix(tag, "name=") { - name = tag[5:] - } - if tag == "proto3" { - proto3 = true - } - } - validateUTF8 = validateUTF8 && proto3 - - // Figure out packaging (pointer, slice, or both) - slice := false - pointer := false - if t.Kind() == reflect.Slice && t.Elem().Kind() != reflect.Uint8 { - slice = true - t = t.Elem() - } - if t.Kind() == reflect.Ptr { - pointer = true - t = t.Elem() - } - - // We'll never have both pointer and slice for basic types. - if pointer && slice && t.Kind() != reflect.Struct { - panic("both pointer and slice for basic type in " + t.Name()) - } - - switch t.Kind() { - case reflect.Bool: - if pointer { - return unmarshalBoolPtr - } - if slice { - return unmarshalBoolSlice - } - return unmarshalBoolValue - case reflect.Int32: - switch encoding { - case "fixed32": - if pointer { - return unmarshalFixedS32Ptr - } - if slice { - return unmarshalFixedS32Slice - } - return unmarshalFixedS32Value - case "varint": - // this could be int32 or enum - if pointer { - return unmarshalInt32Ptr - } - if slice { - return unmarshalInt32Slice - } - return unmarshalInt32Value - case "zigzag32": - if pointer { - return unmarshalSint32Ptr - } - if slice { - return unmarshalSint32Slice - } - return unmarshalSint32Value - } - case reflect.Int64: - switch encoding { - case "fixed64": - if pointer { - return unmarshalFixedS64Ptr - } - if slice { - return unmarshalFixedS64Slice - } - return unmarshalFixedS64Value - case "varint": - if pointer { - return unmarshalInt64Ptr - } - if slice { - return unmarshalInt64Slice - } - return unmarshalInt64Value - case "zigzag64": - if pointer { - return unmarshalSint64Ptr - } - if slice { - return unmarshalSint64Slice - } - return unmarshalSint64Value - } - case reflect.Uint32: - switch encoding { - case "fixed32": - if pointer { - return unmarshalFixed32Ptr - } - if slice { - return unmarshalFixed32Slice - } - return unmarshalFixed32Value - case "varint": - if pointer { - return unmarshalUint32Ptr - } - if slice { - return unmarshalUint32Slice - } - return unmarshalUint32Value - } - case reflect.Uint64: - switch encoding { - case "fixed64": - if pointer { - return unmarshalFixed64Ptr - } - if slice { - return unmarshalFixed64Slice - } - return unmarshalFixed64Value - case "varint": - if pointer { - return unmarshalUint64Ptr - } - if slice { - return unmarshalUint64Slice - } - return unmarshalUint64Value - } - case reflect.Float32: - if pointer { - return unmarshalFloat32Ptr - } - if slice { - return unmarshalFloat32Slice - } - return unmarshalFloat32Value - case reflect.Float64: - if pointer { - return unmarshalFloat64Ptr - } - if slice { - return unmarshalFloat64Slice - } - return unmarshalFloat64Value - case reflect.Map: - panic("map type in typeUnmarshaler in " + t.Name()) - case reflect.Slice: - if pointer { - panic("bad pointer in slice case in " + t.Name()) - } - if slice { - return unmarshalBytesSlice - } - return unmarshalBytesValue - case reflect.String: - if validateUTF8 { - if pointer { - return unmarshalUTF8StringPtr - } - if slice { - return unmarshalUTF8StringSlice - } - return unmarshalUTF8StringValue - } - if pointer { - return unmarshalStringPtr - } - if slice { - return unmarshalStringSlice - } - return unmarshalStringValue - case reflect.Struct: - // message or group field - if !pointer { - panic(fmt.Sprintf("message/group field %s:%s without pointer", t, encoding)) - } - switch encoding { - case "bytes": - if slice { - return makeUnmarshalMessageSlicePtr(getUnmarshalInfo(t), name) - } - return makeUnmarshalMessagePtr(getUnmarshalInfo(t), name) - case "group": - if slice { - return makeUnmarshalGroupSlicePtr(getUnmarshalInfo(t), name) - } - return makeUnmarshalGroupPtr(getUnmarshalInfo(t), name) - } - } - panic(fmt.Sprintf("unmarshaler not found type:%s encoding:%s", t, encoding)) -} - -// Below are all the unmarshalers for individual fields of various types. - -func unmarshalInt64Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x) - *f.toInt64() = v - return b, nil -} - -func unmarshalInt64Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x) - *f.toInt64Ptr() = &v - return b, nil -} - -func unmarshalInt64Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - x, n = decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x) - s := f.toInt64Slice() - *s = append(*s, v) - } - return res, nil - } - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x) - s := f.toInt64Slice() - *s = append(*s, v) - return b, nil -} - -func unmarshalSint64Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x>>1) ^ int64(x)<<63>>63 - *f.toInt64() = v - return b, nil -} - -func unmarshalSint64Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x>>1) ^ int64(x)<<63>>63 - *f.toInt64Ptr() = &v - return b, nil -} - -func unmarshalSint64Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - x, n = decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x>>1) ^ int64(x)<<63>>63 - s := f.toInt64Slice() - *s = append(*s, v) - } - return res, nil - } - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int64(x>>1) ^ int64(x)<<63>>63 - s := f.toInt64Slice() - *s = append(*s, v) - return b, nil -} - -func unmarshalUint64Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint64(x) - *f.toUint64() = v - return b, nil -} - -func unmarshalUint64Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint64(x) - *f.toUint64Ptr() = &v - return b, nil -} - -func unmarshalUint64Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - x, n = decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint64(x) - s := f.toUint64Slice() - *s = append(*s, v) - } - return res, nil - } - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint64(x) - s := f.toUint64Slice() - *s = append(*s, v) - return b, nil -} - -func unmarshalInt32Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x) - *f.toInt32() = v - return b, nil -} - -func unmarshalInt32Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x) - f.setInt32Ptr(v) - return b, nil -} - -func unmarshalInt32Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - x, n = decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x) - f.appendInt32Slice(v) - } - return res, nil - } - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x) - f.appendInt32Slice(v) - return b, nil -} - -func unmarshalSint32Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x>>1) ^ int32(x)<<31>>31 - *f.toInt32() = v - return b, nil -} - -func unmarshalSint32Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x>>1) ^ int32(x)<<31>>31 - f.setInt32Ptr(v) - return b, nil -} - -func unmarshalSint32Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - x, n = decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x>>1) ^ int32(x)<<31>>31 - f.appendInt32Slice(v) - } - return res, nil - } - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := int32(x>>1) ^ int32(x)<<31>>31 - f.appendInt32Slice(v) - return b, nil -} - -func unmarshalUint32Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint32(x) - *f.toUint32() = v - return b, nil -} - -func unmarshalUint32Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint32(x) - *f.toUint32Ptr() = &v - return b, nil -} - -func unmarshalUint32Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - x, n = decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint32(x) - s := f.toUint32Slice() - *s = append(*s, v) - } - return res, nil - } - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - v := uint32(x) - s := f.toUint32Slice() - *s = append(*s, v) - return b, nil -} - -func unmarshalFixed64Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56 - *f.toUint64() = v - return b[8:], nil -} - -func unmarshalFixed64Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56 - *f.toUint64Ptr() = &v - return b[8:], nil -} - -func unmarshalFixed64Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56 - s := f.toUint64Slice() - *s = append(*s, v) - b = b[8:] - } - return res, nil - } - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56 - s := f.toUint64Slice() - *s = append(*s, v) - return b[8:], nil -} - -func unmarshalFixedS64Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := int64(b[0]) | int64(b[1])<<8 | int64(b[2])<<16 | int64(b[3])<<24 | int64(b[4])<<32 | int64(b[5])<<40 | int64(b[6])<<48 | int64(b[7])<<56 - *f.toInt64() = v - return b[8:], nil -} - -func unmarshalFixedS64Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := int64(b[0]) | int64(b[1])<<8 | int64(b[2])<<16 | int64(b[3])<<24 | int64(b[4])<<32 | int64(b[5])<<40 | int64(b[6])<<48 | int64(b[7])<<56 - *f.toInt64Ptr() = &v - return b[8:], nil -} - -func unmarshalFixedS64Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := int64(b[0]) | int64(b[1])<<8 | int64(b[2])<<16 | int64(b[3])<<24 | int64(b[4])<<32 | int64(b[5])<<40 | int64(b[6])<<48 | int64(b[7])<<56 - s := f.toInt64Slice() - *s = append(*s, v) - b = b[8:] - } - return res, nil - } - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := int64(b[0]) | int64(b[1])<<8 | int64(b[2])<<16 | int64(b[3])<<24 | int64(b[4])<<32 | int64(b[5])<<40 | int64(b[6])<<48 | int64(b[7])<<56 - s := f.toInt64Slice() - *s = append(*s, v) - return b[8:], nil -} - -func unmarshalFixed32Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24 - *f.toUint32() = v - return b[4:], nil -} - -func unmarshalFixed32Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24 - *f.toUint32Ptr() = &v - return b[4:], nil -} - -func unmarshalFixed32Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24 - s := f.toUint32Slice() - *s = append(*s, v) - b = b[4:] - } - return res, nil - } - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24 - s := f.toUint32Slice() - *s = append(*s, v) - return b[4:], nil -} - -func unmarshalFixedS32Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := int32(b[0]) | int32(b[1])<<8 | int32(b[2])<<16 | int32(b[3])<<24 - *f.toInt32() = v - return b[4:], nil -} - -func unmarshalFixedS32Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := int32(b[0]) | int32(b[1])<<8 | int32(b[2])<<16 | int32(b[3])<<24 - f.setInt32Ptr(v) - return b[4:], nil -} - -func unmarshalFixedS32Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := int32(b[0]) | int32(b[1])<<8 | int32(b[2])<<16 | int32(b[3])<<24 - f.appendInt32Slice(v) - b = b[4:] - } - return res, nil - } - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := int32(b[0]) | int32(b[1])<<8 | int32(b[2])<<16 | int32(b[3])<<24 - f.appendInt32Slice(v) - return b[4:], nil -} - -func unmarshalBoolValue(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - // Note: any length varint is allowed, even though any sane - // encoder will use one byte. - // See https://github.com/golang/protobuf/issues/76 - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - // TODO: check if x>1? Tests seem to indicate no. - v := x != 0 - *f.toBool() = v - return b[n:], nil -} - -func unmarshalBoolPtr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - v := x != 0 - *f.toBoolPtr() = &v - return b[n:], nil -} - -func unmarshalBoolSlice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - x, n = decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - v := x != 0 - s := f.toBoolSlice() - *s = append(*s, v) - b = b[n:] - } - return res, nil - } - if w != WireVarint { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - v := x != 0 - s := f.toBoolSlice() - *s = append(*s, v) - return b[n:], nil -} - -func unmarshalFloat64Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float64frombits(uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56) - *f.toFloat64() = v - return b[8:], nil -} - -func unmarshalFloat64Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float64frombits(uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56) - *f.toFloat64Ptr() = &v - return b[8:], nil -} - -func unmarshalFloat64Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float64frombits(uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56) - s := f.toFloat64Slice() - *s = append(*s, v) - b = b[8:] - } - return res, nil - } - if w != WireFixed64 { - return b, errInternalBadWireType - } - if len(b) < 8 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float64frombits(uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56) - s := f.toFloat64Slice() - *s = append(*s, v) - return b[8:], nil -} - -func unmarshalFloat32Value(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float32frombits(uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24) - *f.toFloat32() = v - return b[4:], nil -} - -func unmarshalFloat32Ptr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float32frombits(uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24) - *f.toFloat32Ptr() = &v - return b[4:], nil -} - -func unmarshalFloat32Slice(b []byte, f pointer, w int) ([]byte, error) { - if w == WireBytes { // packed - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - res := b[x:] - b = b[:x] - for len(b) > 0 { - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float32frombits(uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24) - s := f.toFloat32Slice() - *s = append(*s, v) - b = b[4:] - } - return res, nil - } - if w != WireFixed32 { - return b, errInternalBadWireType - } - if len(b) < 4 { - return nil, io.ErrUnexpectedEOF - } - v := math.Float32frombits(uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24) - s := f.toFloat32Slice() - *s = append(*s, v) - return b[4:], nil -} - -func unmarshalStringValue(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := string(b[:x]) - *f.toString() = v - return b[x:], nil -} - -func unmarshalStringPtr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := string(b[:x]) - *f.toStringPtr() = &v - return b[x:], nil -} - -func unmarshalStringSlice(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := string(b[:x]) - s := f.toStringSlice() - *s = append(*s, v) - return b[x:], nil -} - -func unmarshalUTF8StringValue(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := string(b[:x]) - *f.toString() = v - if !utf8.ValidString(v) { - return b[x:], errInvalidUTF8 - } - return b[x:], nil -} - -func unmarshalUTF8StringPtr(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := string(b[:x]) - *f.toStringPtr() = &v - if !utf8.ValidString(v) { - return b[x:], errInvalidUTF8 - } - return b[x:], nil -} - -func unmarshalUTF8StringSlice(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := string(b[:x]) - s := f.toStringSlice() - *s = append(*s, v) - if !utf8.ValidString(v) { - return b[x:], errInvalidUTF8 - } - return b[x:], nil -} - -var emptyBuf [0]byte - -func unmarshalBytesValue(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - // The use of append here is a trick which avoids the zeroing - // that would be required if we used a make/copy pair. - // We append to emptyBuf instead of nil because we want - // a non-nil result even when the length is 0. - v := append(emptyBuf[:], b[:x]...) - *f.toBytes() = v - return b[x:], nil -} - -func unmarshalBytesSlice(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := append(emptyBuf[:], b[:x]...) - s := f.toBytesSlice() - *s = append(*s, v) - return b[x:], nil -} - -func makeUnmarshalMessagePtr(sub *unmarshalInfo, name string) unmarshaler { - return func(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - // First read the message field to see if something is there. - // The semantics of multiple submessages are weird. Instead of - // the last one winning (as it is for all other fields), multiple - // submessages are merged. - v := f.getPointer() - if v.isNil() { - v = valToPointer(reflect.New(sub.typ)) - f.setPointer(v) - } - err := sub.unmarshal(v, b[:x]) - if err != nil { - if r, ok := err.(*RequiredNotSetError); ok { - r.field = name + "." + r.field - } else { - return nil, err - } - } - return b[x:], err - } -} - -func makeUnmarshalMessageSlicePtr(sub *unmarshalInfo, name string) unmarshaler { - return func(b []byte, f pointer, w int) ([]byte, error) { - if w != WireBytes { - return b, errInternalBadWireType - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - v := valToPointer(reflect.New(sub.typ)) - err := sub.unmarshal(v, b[:x]) - if err != nil { - if r, ok := err.(*RequiredNotSetError); ok { - r.field = name + "." + r.field - } else { - return nil, err - } - } - f.appendPointer(v) - return b[x:], err - } -} - -func makeUnmarshalGroupPtr(sub *unmarshalInfo, name string) unmarshaler { - return func(b []byte, f pointer, w int) ([]byte, error) { - if w != WireStartGroup { - return b, errInternalBadWireType - } - x, y := findEndGroup(b) - if x < 0 { - return nil, io.ErrUnexpectedEOF - } - v := f.getPointer() - if v.isNil() { - v = valToPointer(reflect.New(sub.typ)) - f.setPointer(v) - } - err := sub.unmarshal(v, b[:x]) - if err != nil { - if r, ok := err.(*RequiredNotSetError); ok { - r.field = name + "." + r.field - } else { - return nil, err - } - } - return b[y:], err - } -} - -func makeUnmarshalGroupSlicePtr(sub *unmarshalInfo, name string) unmarshaler { - return func(b []byte, f pointer, w int) ([]byte, error) { - if w != WireStartGroup { - return b, errInternalBadWireType - } - x, y := findEndGroup(b) - if x < 0 { - return nil, io.ErrUnexpectedEOF - } - v := valToPointer(reflect.New(sub.typ)) - err := sub.unmarshal(v, b[:x]) - if err != nil { - if r, ok := err.(*RequiredNotSetError); ok { - r.field = name + "." + r.field - } else { - return nil, err - } - } - f.appendPointer(v) - return b[y:], err - } -} - -func makeUnmarshalMap(f *reflect.StructField) unmarshaler { - t := f.Type - kt := t.Key() - vt := t.Elem() - unmarshalKey := typeUnmarshaler(kt, f.Tag.Get("protobuf_key")) - unmarshalVal := typeUnmarshaler(vt, f.Tag.Get("protobuf_val")) - return func(b []byte, f pointer, w int) ([]byte, error) { - // The map entry is a submessage. Figure out how big it is. - if w != WireBytes { - return nil, fmt.Errorf("proto: bad wiretype for map field: got %d want %d", w, WireBytes) - } - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - b = b[n:] - if x > uint64(len(b)) { - return nil, io.ErrUnexpectedEOF - } - r := b[x:] // unused data to return - b = b[:x] // data for map entry - - // Note: we could use #keys * #values ~= 200 functions - // to do map decoding without reflection. Probably not worth it. - // Maps will be somewhat slow. Oh well. - - // Read key and value from data. - var nerr nonFatal - k := reflect.New(kt) - v := reflect.New(vt) - for len(b) > 0 { - x, n := decodeVarint(b) - if n == 0 { - return nil, io.ErrUnexpectedEOF - } - wire := int(x) & 7 - b = b[n:] - - var err error - switch x >> 3 { - case 1: - b, err = unmarshalKey(b, valToPointer(k), wire) - case 2: - b, err = unmarshalVal(b, valToPointer(v), wire) - default: - err = errInternalBadWireType // skip unknown tag - } - - if nerr.Merge(err) { - continue - } - if err != errInternalBadWireType { - return nil, err - } - - // Skip past unknown fields. - b, err = skipField(b, wire) - if err != nil { - return nil, err - } - } - - // Get map, allocate if needed. - m := f.asPointerTo(t).Elem() // an addressable map[K]T - if m.IsNil() { - m.Set(reflect.MakeMap(t)) - } - - // Insert into map. - m.SetMapIndex(k.Elem(), v.Elem()) - - return r, nerr.E - } -} - -// makeUnmarshalOneof makes an unmarshaler for oneof fields. -// for: -// message Msg { -// oneof F { -// int64 X = 1; -// float64 Y = 2; -// } -// } -// typ is the type of the concrete entry for a oneof case (e.g. Msg_X). -// ityp is the interface type of the oneof field (e.g. isMsg_F). -// unmarshal is the unmarshaler for the base type of the oneof case (e.g. int64). -// Note that this function will be called once for each case in the oneof. -func makeUnmarshalOneof(typ, ityp reflect.Type, unmarshal unmarshaler) unmarshaler { - sf := typ.Field(0) - field0 := toField(&sf) - return func(b []byte, f pointer, w int) ([]byte, error) { - // Allocate holder for value. - v := reflect.New(typ) - - // Unmarshal data into holder. - // We unmarshal into the first field of the holder object. - var err error - var nerr nonFatal - b, err = unmarshal(b, valToPointer(v).offset(field0), w) - if !nerr.Merge(err) { - return nil, err - } - - // Write pointer to holder into target field. - f.asPointerTo(ityp).Elem().Set(v) - - return b, nerr.E - } -} - -// Error used by decode internally. -var errInternalBadWireType = errors.New("proto: internal error: bad wiretype") - -// skipField skips past a field of type wire and returns the remaining bytes. -func skipField(b []byte, wire int) ([]byte, error) { - switch wire { - case WireVarint: - _, k := decodeVarint(b) - if k == 0 { - return b, io.ErrUnexpectedEOF - } - b = b[k:] - case WireFixed32: - if len(b) < 4 { - return b, io.ErrUnexpectedEOF - } - b = b[4:] - case WireFixed64: - if len(b) < 8 { - return b, io.ErrUnexpectedEOF - } - b = b[8:] - case WireBytes: - m, k := decodeVarint(b) - if k == 0 || uint64(len(b)-k) < m { - return b, io.ErrUnexpectedEOF - } - b = b[uint64(k)+m:] - case WireStartGroup: - _, i := findEndGroup(b) - if i == -1 { - return b, io.ErrUnexpectedEOF - } - b = b[i:] - default: - return b, fmt.Errorf("proto: can't skip unknown wire type %d", wire) - } - return b, nil -} - -// findEndGroup finds the index of the next EndGroup tag. -// Groups may be nested, so the "next" EndGroup tag is the first -// unpaired EndGroup. -// findEndGroup returns the indexes of the start and end of the EndGroup tag. -// Returns (-1,-1) if it can't find one. -func findEndGroup(b []byte) (int, int) { - depth := 1 - i := 0 - for { - x, n := decodeVarint(b[i:]) - if n == 0 { - return -1, -1 - } - j := i - i += n - switch x & 7 { - case WireVarint: - _, k := decodeVarint(b[i:]) - if k == 0 { - return -1, -1 - } - i += k - case WireFixed32: - if len(b)-4 < i { - return -1, -1 - } - i += 4 - case WireFixed64: - if len(b)-8 < i { - return -1, -1 - } - i += 8 - case WireBytes: - m, k := decodeVarint(b[i:]) - if k == 0 { - return -1, -1 - } - i += k - if uint64(len(b)-i) < m { - return -1, -1 - } - i += int(m) - case WireStartGroup: - depth++ - case WireEndGroup: - depth-- - if depth == 0 { - return j, i - } - default: - return -1, -1 - } - } -} - -// encodeVarint appends a varint-encoded integer to b and returns the result. -func encodeVarint(b []byte, x uint64) []byte { - for x >= 1<<7 { - b = append(b, byte(x&0x7f|0x80)) - x >>= 7 - } - return append(b, byte(x)) -} - -// decodeVarint reads a varint-encoded integer from b. -// Returns the decoded integer and the number of bytes read. -// If there is an error, it returns 0,0. -func decodeVarint(b []byte) (uint64, int) { - var x, y uint64 - if len(b) == 0 { - goto bad - } - x = uint64(b[0]) - if x < 0x80 { - return x, 1 - } - x -= 0x80 - - if len(b) <= 1 { - goto bad - } - y = uint64(b[1]) - x += y << 7 - if y < 0x80 { - return x, 2 - } - x -= 0x80 << 7 - - if len(b) <= 2 { - goto bad - } - y = uint64(b[2]) - x += y << 14 - if y < 0x80 { - return x, 3 - } - x -= 0x80 << 14 - - if len(b) <= 3 { - goto bad - } - y = uint64(b[3]) - x += y << 21 - if y < 0x80 { - return x, 4 - } - x -= 0x80 << 21 - - if len(b) <= 4 { - goto bad - } - y = uint64(b[4]) - x += y << 28 - if y < 0x80 { - return x, 5 - } - x -= 0x80 << 28 - - if len(b) <= 5 { - goto bad - } - y = uint64(b[5]) - x += y << 35 - if y < 0x80 { - return x, 6 - } - x -= 0x80 << 35 - - if len(b) <= 6 { - goto bad - } - y = uint64(b[6]) - x += y << 42 - if y < 0x80 { - return x, 7 - } - x -= 0x80 << 42 - - if len(b) <= 7 { - goto bad - } - y = uint64(b[7]) - x += y << 49 - if y < 0x80 { - return x, 8 - } - x -= 0x80 << 49 - - if len(b) <= 8 { - goto bad - } - y = uint64(b[8]) - x += y << 56 - if y < 0x80 { - return x, 9 - } - x -= 0x80 << 56 - - if len(b) <= 9 { - goto bad - } - y = uint64(b[9]) - x += y << 63 - if y < 2 { - return x, 10 - } - -bad: - return 0, 0 -} diff --git a/vendor/github.com/golang/protobuf/proto/text.go b/vendor/github.com/golang/protobuf/proto/text.go deleted file mode 100644 index d97f9b35..00000000 --- a/vendor/github.com/golang/protobuf/proto/text.go +++ /dev/null @@ -1,845 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -// Functions for writing the text protocol buffer format. - -import ( - "bufio" - "bytes" - "encoding" - "errors" - "fmt" - "io" - "log" - "math" - "reflect" - "sort" - "strings" -) - -var ( - newline = []byte("\n") - spaces = []byte(" ") - endBraceNewline = []byte("}\n") - backslashN = []byte{'\\', 'n'} - backslashR = []byte{'\\', 'r'} - backslashT = []byte{'\\', 't'} - backslashDQ = []byte{'\\', '"'} - backslashBS = []byte{'\\', '\\'} - posInf = []byte("inf") - negInf = []byte("-inf") - nan = []byte("nan") -) - -type writer interface { - io.Writer - WriteByte(byte) error -} - -// textWriter is an io.Writer that tracks its indentation level. -type textWriter struct { - ind int - complete bool // if the current position is a complete line - compact bool // whether to write out as a one-liner - w writer -} - -func (w *textWriter) WriteString(s string) (n int, err error) { - if !strings.Contains(s, "\n") { - if !w.compact && w.complete { - w.writeIndent() - } - w.complete = false - return io.WriteString(w.w, s) - } - // WriteString is typically called without newlines, so this - // codepath and its copy are rare. We copy to avoid - // duplicating all of Write's logic here. - return w.Write([]byte(s)) -} - -func (w *textWriter) Write(p []byte) (n int, err error) { - newlines := bytes.Count(p, newline) - if newlines == 0 { - if !w.compact && w.complete { - w.writeIndent() - } - n, err = w.w.Write(p) - w.complete = false - return n, err - } - - frags := bytes.SplitN(p, newline, newlines+1) - if w.compact { - for i, frag := range frags { - if i > 0 { - if err := w.w.WriteByte(' '); err != nil { - return n, err - } - n++ - } - nn, err := w.w.Write(frag) - n += nn - if err != nil { - return n, err - } - } - return n, nil - } - - for i, frag := range frags { - if w.complete { - w.writeIndent() - } - nn, err := w.w.Write(frag) - n += nn - if err != nil { - return n, err - } - if i+1 < len(frags) { - if err := w.w.WriteByte('\n'); err != nil { - return n, err - } - n++ - } - } - w.complete = len(frags[len(frags)-1]) == 0 - return n, nil -} - -func (w *textWriter) WriteByte(c byte) error { - if w.compact && c == '\n' { - c = ' ' - } - if !w.compact && w.complete { - w.writeIndent() - } - err := w.w.WriteByte(c) - w.complete = c == '\n' - return err -} - -func (w *textWriter) indent() { w.ind++ } - -func (w *textWriter) unindent() { - if w.ind == 0 { - log.Print("proto: textWriter unindented too far") - return - } - w.ind-- -} - -func writeName(w *textWriter, props *Properties) error { - if _, err := w.WriteString(props.OrigName); err != nil { - return err - } - if props.Wire != "group" { - return w.WriteByte(':') - } - return nil -} - -func requiresQuotes(u string) bool { - // When type URL contains any characters except [0-9A-Za-z./\-]*, it must be quoted. - for _, ch := range u { - switch { - case ch == '.' || ch == '/' || ch == '_': - continue - case '0' <= ch && ch <= '9': - continue - case 'A' <= ch && ch <= 'Z': - continue - case 'a' <= ch && ch <= 'z': - continue - default: - return true - } - } - return false -} - -// isAny reports whether sv is a google.protobuf.Any message -func isAny(sv reflect.Value) bool { - type wkt interface { - XXX_WellKnownType() string - } - t, ok := sv.Addr().Interface().(wkt) - return ok && t.XXX_WellKnownType() == "Any" -} - -// writeProto3Any writes an expanded google.protobuf.Any message. -// -// It returns (false, nil) if sv value can't be unmarshaled (e.g. because -// required messages are not linked in). -// -// It returns (true, error) when sv was written in expanded format or an error -// was encountered. -func (tm *TextMarshaler) writeProto3Any(w *textWriter, sv reflect.Value) (bool, error) { - turl := sv.FieldByName("TypeUrl") - val := sv.FieldByName("Value") - if !turl.IsValid() || !val.IsValid() { - return true, errors.New("proto: invalid google.protobuf.Any message") - } - - b, ok := val.Interface().([]byte) - if !ok { - return true, errors.New("proto: invalid google.protobuf.Any message") - } - - parts := strings.Split(turl.String(), "/") - mt := MessageType(parts[len(parts)-1]) - if mt == nil { - return false, nil - } - m := reflect.New(mt.Elem()) - if err := Unmarshal(b, m.Interface().(Message)); err != nil { - return false, nil - } - w.Write([]byte("[")) - u := turl.String() - if requiresQuotes(u) { - writeString(w, u) - } else { - w.Write([]byte(u)) - } - if w.compact { - w.Write([]byte("]:<")) - } else { - w.Write([]byte("]: <\n")) - w.ind++ - } - if err := tm.writeStruct(w, m.Elem()); err != nil { - return true, err - } - if w.compact { - w.Write([]byte("> ")) - } else { - w.ind-- - w.Write([]byte(">\n")) - } - return true, nil -} - -func (tm *TextMarshaler) writeStruct(w *textWriter, sv reflect.Value) error { - if tm.ExpandAny && isAny(sv) { - if canExpand, err := tm.writeProto3Any(w, sv); canExpand { - return err - } - } - st := sv.Type() - sprops := GetProperties(st) - for i := 0; i < sv.NumField(); i++ { - fv := sv.Field(i) - props := sprops.Prop[i] - name := st.Field(i).Name - - if name == "XXX_NoUnkeyedLiteral" { - continue - } - - if strings.HasPrefix(name, "XXX_") { - // There are two XXX_ fields: - // XXX_unrecognized []byte - // XXX_extensions map[int32]proto.Extension - // The first is handled here; - // the second is handled at the bottom of this function. - if name == "XXX_unrecognized" && !fv.IsNil() { - if err := writeUnknownStruct(w, fv.Interface().([]byte)); err != nil { - return err - } - } - continue - } - if fv.Kind() == reflect.Ptr && fv.IsNil() { - // Field not filled in. This could be an optional field or - // a required field that wasn't filled in. Either way, there - // isn't anything we can show for it. - continue - } - if fv.Kind() == reflect.Slice && fv.IsNil() { - // Repeated field that is empty, or a bytes field that is unused. - continue - } - - if props.Repeated && fv.Kind() == reflect.Slice { - // Repeated field. - for j := 0; j < fv.Len(); j++ { - if err := writeName(w, props); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte(' '); err != nil { - return err - } - } - v := fv.Index(j) - if v.Kind() == reflect.Ptr && v.IsNil() { - // A nil message in a repeated field is not valid, - // but we can handle that more gracefully than panicking. - if _, err := w.Write([]byte("\n")); err != nil { - return err - } - continue - } - if err := tm.writeAny(w, v, props); err != nil { - return err - } - if err := w.WriteByte('\n'); err != nil { - return err - } - } - continue - } - if fv.Kind() == reflect.Map { - // Map fields are rendered as a repeated struct with key/value fields. - keys := fv.MapKeys() - sort.Sort(mapKeys(keys)) - for _, key := range keys { - val := fv.MapIndex(key) - if err := writeName(w, props); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte(' '); err != nil { - return err - } - } - // open struct - if err := w.WriteByte('<'); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte('\n'); err != nil { - return err - } - } - w.indent() - // key - if _, err := w.WriteString("key:"); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte(' '); err != nil { - return err - } - } - if err := tm.writeAny(w, key, props.MapKeyProp); err != nil { - return err - } - if err := w.WriteByte('\n'); err != nil { - return err - } - // nil values aren't legal, but we can avoid panicking because of them. - if val.Kind() != reflect.Ptr || !val.IsNil() { - // value - if _, err := w.WriteString("value:"); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte(' '); err != nil { - return err - } - } - if err := tm.writeAny(w, val, props.MapValProp); err != nil { - return err - } - if err := w.WriteByte('\n'); err != nil { - return err - } - } - // close struct - w.unindent() - if err := w.WriteByte('>'); err != nil { - return err - } - if err := w.WriteByte('\n'); err != nil { - return err - } - } - continue - } - if props.proto3 && fv.Kind() == reflect.Slice && fv.Len() == 0 { - // empty bytes field - continue - } - if fv.Kind() != reflect.Ptr && fv.Kind() != reflect.Slice { - // proto3 non-repeated scalar field; skip if zero value - if isProto3Zero(fv) { - continue - } - } - - if fv.Kind() == reflect.Interface { - // Check if it is a oneof. - if st.Field(i).Tag.Get("protobuf_oneof") != "" { - // fv is nil, or holds a pointer to generated struct. - // That generated struct has exactly one field, - // which has a protobuf struct tag. - if fv.IsNil() { - continue - } - inner := fv.Elem().Elem() // interface -> *T -> T - tag := inner.Type().Field(0).Tag.Get("protobuf") - props = new(Properties) // Overwrite the outer props var, but not its pointee. - props.Parse(tag) - // Write the value in the oneof, not the oneof itself. - fv = inner.Field(0) - - // Special case to cope with malformed messages gracefully: - // If the value in the oneof is a nil pointer, don't panic - // in writeAny. - if fv.Kind() == reflect.Ptr && fv.IsNil() { - // Use errors.New so writeAny won't render quotes. - msg := errors.New("/* nil */") - fv = reflect.ValueOf(&msg).Elem() - } - } - } - - if err := writeName(w, props); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte(' '); err != nil { - return err - } - } - - // Enums have a String method, so writeAny will work fine. - if err := tm.writeAny(w, fv, props); err != nil { - return err - } - - if err := w.WriteByte('\n'); err != nil { - return err - } - } - - // Extensions (the XXX_extensions field). - pv := sv.Addr() - if _, err := extendable(pv.Interface()); err == nil { - if err := tm.writeExtensions(w, pv); err != nil { - return err - } - } - - return nil -} - -var textMarshalerType = reflect.TypeOf((*encoding.TextMarshaler)(nil)).Elem() - -// writeAny writes an arbitrary field. -func (tm *TextMarshaler) writeAny(w *textWriter, v reflect.Value, props *Properties) error { - v = reflect.Indirect(v) - - // Floats have special cases. - if v.Kind() == reflect.Float32 || v.Kind() == reflect.Float64 { - x := v.Float() - var b []byte - switch { - case math.IsInf(x, 1): - b = posInf - case math.IsInf(x, -1): - b = negInf - case math.IsNaN(x): - b = nan - } - if b != nil { - _, err := w.Write(b) - return err - } - // Other values are handled below. - } - - // We don't attempt to serialise every possible value type; only those - // that can occur in protocol buffers. - switch v.Kind() { - case reflect.Slice: - // Should only be a []byte; repeated fields are handled in writeStruct. - if err := writeString(w, string(v.Bytes())); err != nil { - return err - } - case reflect.String: - if err := writeString(w, v.String()); err != nil { - return err - } - case reflect.Struct: - // Required/optional group/message. - var bra, ket byte = '<', '>' - if props != nil && props.Wire == "group" { - bra, ket = '{', '}' - } - if err := w.WriteByte(bra); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte('\n'); err != nil { - return err - } - } - w.indent() - if v.CanAddr() { - // Calling v.Interface on a struct causes the reflect package to - // copy the entire struct. This is racy with the new Marshaler - // since we atomically update the XXX_sizecache. - // - // Thus, we retrieve a pointer to the struct if possible to avoid - // a race since v.Interface on the pointer doesn't copy the struct. - // - // If v is not addressable, then we are not worried about a race - // since it implies that the binary Marshaler cannot possibly be - // mutating this value. - v = v.Addr() - } - if v.Type().Implements(textMarshalerType) { - text, err := v.Interface().(encoding.TextMarshaler).MarshalText() - if err != nil { - return err - } - if _, err = w.Write(text); err != nil { - return err - } - } else { - if v.Kind() == reflect.Ptr { - v = v.Elem() - } - if err := tm.writeStruct(w, v); err != nil { - return err - } - } - w.unindent() - if err := w.WriteByte(ket); err != nil { - return err - } - default: - _, err := fmt.Fprint(w, v.Interface()) - return err - } - return nil -} - -// equivalent to C's isprint. -func isprint(c byte) bool { - return c >= 0x20 && c < 0x7f -} - -// writeString writes a string in the protocol buffer text format. -// It is similar to strconv.Quote except we don't use Go escape sequences, -// we treat the string as a byte sequence, and we use octal escapes. -// These differences are to maintain interoperability with the other -// languages' implementations of the text format. -func writeString(w *textWriter, s string) error { - // use WriteByte here to get any needed indent - if err := w.WriteByte('"'); err != nil { - return err - } - // Loop over the bytes, not the runes. - for i := 0; i < len(s); i++ { - var err error - // Divergence from C++: we don't escape apostrophes. - // There's no need to escape them, and the C++ parser - // copes with a naked apostrophe. - switch c := s[i]; c { - case '\n': - _, err = w.w.Write(backslashN) - case '\r': - _, err = w.w.Write(backslashR) - case '\t': - _, err = w.w.Write(backslashT) - case '"': - _, err = w.w.Write(backslashDQ) - case '\\': - _, err = w.w.Write(backslashBS) - default: - if isprint(c) { - err = w.w.WriteByte(c) - } else { - _, err = fmt.Fprintf(w.w, "\\%03o", c) - } - } - if err != nil { - return err - } - } - return w.WriteByte('"') -} - -func writeUnknownStruct(w *textWriter, data []byte) (err error) { - if !w.compact { - if _, err := fmt.Fprintf(w, "/* %d unknown bytes */\n", len(data)); err != nil { - return err - } - } - b := NewBuffer(data) - for b.index < len(b.buf) { - x, err := b.DecodeVarint() - if err != nil { - _, err := fmt.Fprintf(w, "/* %v */\n", err) - return err - } - wire, tag := x&7, x>>3 - if wire == WireEndGroup { - w.unindent() - if _, err := w.Write(endBraceNewline); err != nil { - return err - } - continue - } - if _, err := fmt.Fprint(w, tag); err != nil { - return err - } - if wire != WireStartGroup { - if err := w.WriteByte(':'); err != nil { - return err - } - } - if !w.compact || wire == WireStartGroup { - if err := w.WriteByte(' '); err != nil { - return err - } - } - switch wire { - case WireBytes: - buf, e := b.DecodeRawBytes(false) - if e == nil { - _, err = fmt.Fprintf(w, "%q", buf) - } else { - _, err = fmt.Fprintf(w, "/* %v */", e) - } - case WireFixed32: - x, err = b.DecodeFixed32() - err = writeUnknownInt(w, x, err) - case WireFixed64: - x, err = b.DecodeFixed64() - err = writeUnknownInt(w, x, err) - case WireStartGroup: - err = w.WriteByte('{') - w.indent() - case WireVarint: - x, err = b.DecodeVarint() - err = writeUnknownInt(w, x, err) - default: - _, err = fmt.Fprintf(w, "/* unknown wire type %d */", wire) - } - if err != nil { - return err - } - if err = w.WriteByte('\n'); err != nil { - return err - } - } - return nil -} - -func writeUnknownInt(w *textWriter, x uint64, err error) error { - if err == nil { - _, err = fmt.Fprint(w, x) - } else { - _, err = fmt.Fprintf(w, "/* %v */", err) - } - return err -} - -type int32Slice []int32 - -func (s int32Slice) Len() int { return len(s) } -func (s int32Slice) Less(i, j int) bool { return s[i] < s[j] } -func (s int32Slice) Swap(i, j int) { s[i], s[j] = s[j], s[i] } - -// writeExtensions writes all the extensions in pv. -// pv is assumed to be a pointer to a protocol message struct that is extendable. -func (tm *TextMarshaler) writeExtensions(w *textWriter, pv reflect.Value) error { - emap := extensionMaps[pv.Type().Elem()] - ep, _ := extendable(pv.Interface()) - - // Order the extensions by ID. - // This isn't strictly necessary, but it will give us - // canonical output, which will also make testing easier. - m, mu := ep.extensionsRead() - if m == nil { - return nil - } - mu.Lock() - ids := make([]int32, 0, len(m)) - for id := range m { - ids = append(ids, id) - } - sort.Sort(int32Slice(ids)) - mu.Unlock() - - for _, extNum := range ids { - ext := m[extNum] - var desc *ExtensionDesc - if emap != nil { - desc = emap[extNum] - } - if desc == nil { - // Unknown extension. - if err := writeUnknownStruct(w, ext.enc); err != nil { - return err - } - continue - } - - pb, err := GetExtension(ep, desc) - if err != nil { - return fmt.Errorf("failed getting extension: %v", err) - } - - // Repeated extensions will appear as a slice. - if !desc.repeated() { - if err := tm.writeExtension(w, desc.Name, pb); err != nil { - return err - } - } else { - v := reflect.ValueOf(pb) - for i := 0; i < v.Len(); i++ { - if err := tm.writeExtension(w, desc.Name, v.Index(i).Interface()); err != nil { - return err - } - } - } - } - return nil -} - -func (tm *TextMarshaler) writeExtension(w *textWriter, name string, pb interface{}) error { - if _, err := fmt.Fprintf(w, "[%s]:", name); err != nil { - return err - } - if !w.compact { - if err := w.WriteByte(' '); err != nil { - return err - } - } - if err := tm.writeAny(w, reflect.ValueOf(pb), nil); err != nil { - return err - } - if err := w.WriteByte('\n'); err != nil { - return err - } - return nil -} - -func (w *textWriter) writeIndent() { - if !w.complete { - return - } - remain := w.ind * 2 - for remain > 0 { - n := remain - if n > len(spaces) { - n = len(spaces) - } - w.w.Write(spaces[:n]) - remain -= n - } - w.complete = false -} - -// TextMarshaler is a configurable text format marshaler. -type TextMarshaler struct { - Compact bool // use compact text format (one line). - ExpandAny bool // expand google.protobuf.Any messages of known types -} - -// Marshal writes a given protocol buffer in text format. -// The only errors returned are from w. -func (tm *TextMarshaler) Marshal(w io.Writer, pb Message) error { - val := reflect.ValueOf(pb) - if pb == nil || val.IsNil() { - w.Write([]byte("")) - return nil - } - var bw *bufio.Writer - ww, ok := w.(writer) - if !ok { - bw = bufio.NewWriter(w) - ww = bw - } - aw := &textWriter{ - w: ww, - complete: true, - compact: tm.Compact, - } - - if etm, ok := pb.(encoding.TextMarshaler); ok { - text, err := etm.MarshalText() - if err != nil { - return err - } - if _, err = aw.Write(text); err != nil { - return err - } - if bw != nil { - return bw.Flush() - } - return nil - } - // Dereference the received pointer so we don't have outer < and >. - v := reflect.Indirect(val) - if err := tm.writeStruct(aw, v); err != nil { - return err - } - if bw != nil { - return bw.Flush() - } - return nil -} - -// Text is the same as Marshal, but returns the string directly. -func (tm *TextMarshaler) Text(pb Message) string { - var buf bytes.Buffer - tm.Marshal(&buf, pb) - return buf.String() -} - -var ( - defaultTextMarshaler = TextMarshaler{} - compactTextMarshaler = TextMarshaler{Compact: true} -) - -// TODO: consider removing some of the Marshal functions below. - -// MarshalText writes a given protocol buffer in text format. -// The only errors returned are from w. -func MarshalText(w io.Writer, pb Message) error { return defaultTextMarshaler.Marshal(w, pb) } - -// MarshalTextString is the same as MarshalText, but returns the string directly. -func MarshalTextString(pb Message) string { return defaultTextMarshaler.Text(pb) } - -// CompactText writes a given protocol buffer in compact text format (one line). -func CompactText(w io.Writer, pb Message) error { return compactTextMarshaler.Marshal(w, pb) } - -// CompactTextString is the same as CompactText, but returns the string directly. -func CompactTextString(pb Message) string { return compactTextMarshaler.Text(pb) } diff --git a/vendor/github.com/golang/protobuf/proto/text_decode.go b/vendor/github.com/golang/protobuf/proto/text_decode.go new file mode 100644 index 00000000..47eb3e44 --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/text_decode.go @@ -0,0 +1,801 @@ +// Copyright 2010 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "encoding" + "errors" + "fmt" + "reflect" + "strconv" + "strings" + "unicode/utf8" + + "google.golang.org/protobuf/encoding/prototext" + protoV2 "google.golang.org/protobuf/proto" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +const wrapTextUnmarshalV2 = false + +// ParseError is returned by UnmarshalText. +type ParseError struct { + Message string + + // Deprecated: Do not use. + Line, Offset int +} + +func (e *ParseError) Error() string { + if wrapTextUnmarshalV2 { + return e.Message + } + if e.Line == 1 { + return fmt.Sprintf("line 1.%d: %v", e.Offset, e.Message) + } + return fmt.Sprintf("line %d: %v", e.Line, e.Message) +} + +// UnmarshalText parses a proto text formatted string into m. +func UnmarshalText(s string, m Message) error { + if u, ok := m.(encoding.TextUnmarshaler); ok { + return u.UnmarshalText([]byte(s)) + } + + m.Reset() + mi := MessageV2(m) + + if wrapTextUnmarshalV2 { + err := prototext.UnmarshalOptions{ + AllowPartial: true, + }.Unmarshal([]byte(s), mi) + if err != nil { + return &ParseError{Message: err.Error()} + } + return checkRequiredNotSet(mi) + } else { + if err := newTextParser(s).unmarshalMessage(mi.ProtoReflect(), ""); err != nil { + return err + } + return checkRequiredNotSet(mi) + } +} + +type textParser struct { + s string // remaining input + done bool // whether the parsing is finished (success or error) + backed bool // whether back() was called + offset, line int + cur token +} + +type token struct { + value string + err *ParseError + line int // line number + offset int // byte number from start of input, not start of line + unquoted string // the unquoted version of value, if it was a quoted string +} + +func newTextParser(s string) *textParser { + p := new(textParser) + p.s = s + p.line = 1 + p.cur.line = 1 + return p +} + +func (p *textParser) unmarshalMessage(m protoreflect.Message, terminator string) (err error) { + md := m.Descriptor() + fds := md.Fields() + + // A struct is a sequence of "name: value", terminated by one of + // '>' or '}', or the end of the input. A name may also be + // "[extension]" or "[type/url]". + // + // The whole struct can also be an expanded Any message, like: + // [type/url] < ... struct contents ... > + seen := make(map[protoreflect.FieldNumber]bool) + for { + tok := p.next() + if tok.err != nil { + return tok.err + } + if tok.value == terminator { + break + } + if tok.value == "[" { + if err := p.unmarshalExtensionOrAny(m, seen); err != nil { + return err + } + continue + } + + // This is a normal, non-extension field. + name := protoreflect.Name(tok.value) + fd := fds.ByName(name) + switch { + case fd == nil: + gd := fds.ByName(protoreflect.Name(strings.ToLower(string(name)))) + if gd != nil && gd.Kind() == protoreflect.GroupKind && gd.Message().Name() == name { + fd = gd + } + case fd.Kind() == protoreflect.GroupKind && fd.Message().Name() != name: + fd = nil + case fd.IsWeak() && fd.Message().IsPlaceholder(): + fd = nil + } + if fd == nil { + typeName := string(md.FullName()) + if m, ok := m.Interface().(Message); ok { + t := reflect.TypeOf(m) + if t.Kind() == reflect.Ptr { + typeName = t.Elem().String() + } + } + return p.errorf("unknown field name %q in %v", name, typeName) + } + if od := fd.ContainingOneof(); od != nil && m.WhichOneof(od) != nil { + return p.errorf("field '%s' would overwrite already parsed oneof '%s'", name, od.Name()) + } + if fd.Cardinality() != protoreflect.Repeated && seen[fd.Number()] { + return p.errorf("non-repeated field %q was repeated", fd.Name()) + } + seen[fd.Number()] = true + + // Consume any colon. + if err := p.checkForColon(fd); err != nil { + return err + } + + // Parse into the field. + v := m.Get(fd) + if !m.Has(fd) && (fd.IsList() || fd.IsMap() || fd.Message() != nil) { + v = m.Mutable(fd) + } + if v, err = p.unmarshalValue(v, fd); err != nil { + return err + } + m.Set(fd, v) + + if err := p.consumeOptionalSeparator(); err != nil { + return err + } + } + return nil +} + +func (p *textParser) unmarshalExtensionOrAny(m protoreflect.Message, seen map[protoreflect.FieldNumber]bool) error { + name, err := p.consumeExtensionOrAnyName() + if err != nil { + return err + } + + // If it contains a slash, it's an Any type URL. + if slashIdx := strings.LastIndex(name, "/"); slashIdx >= 0 { + tok := p.next() + if tok.err != nil { + return tok.err + } + // consume an optional colon + if tok.value == ":" { + tok = p.next() + if tok.err != nil { + return tok.err + } + } + + var terminator string + switch tok.value { + case "<": + terminator = ">" + case "{": + terminator = "}" + default: + return p.errorf("expected '{' or '<', found %q", tok.value) + } + + mt, err := protoregistry.GlobalTypes.FindMessageByURL(name) + if err != nil { + return p.errorf("unrecognized message %q in google.protobuf.Any", name[slashIdx+len("/"):]) + } + m2 := mt.New() + if err := p.unmarshalMessage(m2, terminator); err != nil { + return err + } + b, err := protoV2.Marshal(m2.Interface()) + if err != nil { + return p.errorf("failed to marshal message of type %q: %v", name[slashIdx+len("/"):], err) + } + + urlFD := m.Descriptor().Fields().ByName("type_url") + valFD := m.Descriptor().Fields().ByName("value") + if seen[urlFD.Number()] { + return p.errorf("Any message unpacked multiple times, or %q already set", urlFD.Name()) + } + if seen[valFD.Number()] { + return p.errorf("Any message unpacked multiple times, or %q already set", valFD.Name()) + } + m.Set(urlFD, protoreflect.ValueOfString(name)) + m.Set(valFD, protoreflect.ValueOfBytes(b)) + seen[urlFD.Number()] = true + seen[valFD.Number()] = true + return nil + } + + xname := protoreflect.FullName(name) + xt, _ := protoregistry.GlobalTypes.FindExtensionByName(xname) + if xt == nil && isMessageSet(m.Descriptor()) { + xt, _ = protoregistry.GlobalTypes.FindExtensionByName(xname.Append("message_set_extension")) + } + if xt == nil { + return p.errorf("unrecognized extension %q", name) + } + fd := xt.TypeDescriptor() + if fd.ContainingMessage().FullName() != m.Descriptor().FullName() { + return p.errorf("extension field %q does not extend message %q", name, m.Descriptor().FullName()) + } + + if err := p.checkForColon(fd); err != nil { + return err + } + + v := m.Get(fd) + if !m.Has(fd) && (fd.IsList() || fd.IsMap() || fd.Message() != nil) { + v = m.Mutable(fd) + } + v, err = p.unmarshalValue(v, fd) + if err != nil { + return err + } + m.Set(fd, v) + return p.consumeOptionalSeparator() +} + +func (p *textParser) unmarshalValue(v protoreflect.Value, fd protoreflect.FieldDescriptor) (protoreflect.Value, error) { + tok := p.next() + if tok.err != nil { + return v, tok.err + } + if tok.value == "" { + return v, p.errorf("unexpected EOF") + } + + switch { + case fd.IsList(): + lv := v.List() + var err error + if tok.value == "[" { + // Repeated field with list notation, like [1,2,3]. + for { + vv := lv.NewElement() + vv, err = p.unmarshalSingularValue(vv, fd) + if err != nil { + return v, err + } + lv.Append(vv) + + tok := p.next() + if tok.err != nil { + return v, tok.err + } + if tok.value == "]" { + break + } + if tok.value != "," { + return v, p.errorf("Expected ']' or ',' found %q", tok.value) + } + } + return v, nil + } + + // One value of the repeated field. + p.back() + vv := lv.NewElement() + vv, err = p.unmarshalSingularValue(vv, fd) + if err != nil { + return v, err + } + lv.Append(vv) + return v, nil + case fd.IsMap(): + // The map entry should be this sequence of tokens: + // < key : KEY value : VALUE > + // However, implementations may omit key or value, and technically + // we should support them in any order. + var terminator string + switch tok.value { + case "<": + terminator = ">" + case "{": + terminator = "}" + default: + return v, p.errorf("expected '{' or '<', found %q", tok.value) + } + + keyFD := fd.MapKey() + valFD := fd.MapValue() + + mv := v.Map() + kv := keyFD.Default() + vv := mv.NewValue() + for { + tok := p.next() + if tok.err != nil { + return v, tok.err + } + if tok.value == terminator { + break + } + var err error + switch tok.value { + case "key": + if err := p.consumeToken(":"); err != nil { + return v, err + } + if kv, err = p.unmarshalSingularValue(kv, keyFD); err != nil { + return v, err + } + if err := p.consumeOptionalSeparator(); err != nil { + return v, err + } + case "value": + if err := p.checkForColon(valFD); err != nil { + return v, err + } + if vv, err = p.unmarshalSingularValue(vv, valFD); err != nil { + return v, err + } + if err := p.consumeOptionalSeparator(); err != nil { + return v, err + } + default: + p.back() + return v, p.errorf(`expected "key", "value", or %q, found %q`, terminator, tok.value) + } + } + mv.Set(kv.MapKey(), vv) + return v, nil + default: + p.back() + return p.unmarshalSingularValue(v, fd) + } +} + +func (p *textParser) unmarshalSingularValue(v protoreflect.Value, fd protoreflect.FieldDescriptor) (protoreflect.Value, error) { + tok := p.next() + if tok.err != nil { + return v, tok.err + } + if tok.value == "" { + return v, p.errorf("unexpected EOF") + } + + switch fd.Kind() { + case protoreflect.BoolKind: + switch tok.value { + case "true", "1", "t", "True": + return protoreflect.ValueOfBool(true), nil + case "false", "0", "f", "False": + return protoreflect.ValueOfBool(false), nil + } + case protoreflect.Int32Kind, protoreflect.Sint32Kind, protoreflect.Sfixed32Kind: + if x, err := strconv.ParseInt(tok.value, 0, 32); err == nil { + return protoreflect.ValueOfInt32(int32(x)), nil + } + + // The C++ parser accepts large positive hex numbers that uses + // two's complement arithmetic to represent negative numbers. + // This feature is here for backwards compatibility with C++. + if strings.HasPrefix(tok.value, "0x") { + if x, err := strconv.ParseUint(tok.value, 0, 32); err == nil { + return protoreflect.ValueOfInt32(int32(-(int64(^x) + 1))), nil + } + } + case protoreflect.Int64Kind, protoreflect.Sint64Kind, protoreflect.Sfixed64Kind: + if x, err := strconv.ParseInt(tok.value, 0, 64); err == nil { + return protoreflect.ValueOfInt64(int64(x)), nil + } + + // The C++ parser accepts large positive hex numbers that uses + // two's complement arithmetic to represent negative numbers. + // This feature is here for backwards compatibility with C++. + if strings.HasPrefix(tok.value, "0x") { + if x, err := strconv.ParseUint(tok.value, 0, 64); err == nil { + return protoreflect.ValueOfInt64(int64(-(int64(^x) + 1))), nil + } + } + case protoreflect.Uint32Kind, protoreflect.Fixed32Kind: + if x, err := strconv.ParseUint(tok.value, 0, 32); err == nil { + return protoreflect.ValueOfUint32(uint32(x)), nil + } + case protoreflect.Uint64Kind, protoreflect.Fixed64Kind: + if x, err := strconv.ParseUint(tok.value, 0, 64); err == nil { + return protoreflect.ValueOfUint64(uint64(x)), nil + } + case protoreflect.FloatKind: + // Ignore 'f' for compatibility with output generated by C++, + // but don't remove 'f' when the value is "-inf" or "inf". + v := tok.value + if strings.HasSuffix(v, "f") && v != "-inf" && v != "inf" { + v = v[:len(v)-len("f")] + } + if x, err := strconv.ParseFloat(v, 32); err == nil { + return protoreflect.ValueOfFloat32(float32(x)), nil + } + case protoreflect.DoubleKind: + // Ignore 'f' for compatibility with output generated by C++, + // but don't remove 'f' when the value is "-inf" or "inf". + v := tok.value + if strings.HasSuffix(v, "f") && v != "-inf" && v != "inf" { + v = v[:len(v)-len("f")] + } + if x, err := strconv.ParseFloat(v, 64); err == nil { + return protoreflect.ValueOfFloat64(float64(x)), nil + } + case protoreflect.StringKind: + if isQuote(tok.value[0]) { + return protoreflect.ValueOfString(tok.unquoted), nil + } + case protoreflect.BytesKind: + if isQuote(tok.value[0]) { + return protoreflect.ValueOfBytes([]byte(tok.unquoted)), nil + } + case protoreflect.EnumKind: + if x, err := strconv.ParseInt(tok.value, 0, 32); err == nil { + return protoreflect.ValueOfEnum(protoreflect.EnumNumber(x)), nil + } + vd := fd.Enum().Values().ByName(protoreflect.Name(tok.value)) + if vd != nil { + return protoreflect.ValueOfEnum(vd.Number()), nil + } + case protoreflect.MessageKind, protoreflect.GroupKind: + var terminator string + switch tok.value { + case "{": + terminator = "}" + case "<": + terminator = ">" + default: + return v, p.errorf("expected '{' or '<', found %q", tok.value) + } + err := p.unmarshalMessage(v.Message(), terminator) + return v, err + default: + panic(fmt.Sprintf("invalid kind %v", fd.Kind())) + } + return v, p.errorf("invalid %v: %v", fd.Kind(), tok.value) +} + +// Consume a ':' from the input stream (if the next token is a colon), +// returning an error if a colon is needed but not present. +func (p *textParser) checkForColon(fd protoreflect.FieldDescriptor) *ParseError { + tok := p.next() + if tok.err != nil { + return tok.err + } + if tok.value != ":" { + if fd.Message() == nil { + return p.errorf("expected ':', found %q", tok.value) + } + p.back() + } + return nil +} + +// consumeExtensionOrAnyName consumes an extension name or an Any type URL and +// the following ']'. It returns the name or URL consumed. +func (p *textParser) consumeExtensionOrAnyName() (string, error) { + tok := p.next() + if tok.err != nil { + return "", tok.err + } + + // If extension name or type url is quoted, it's a single token. + if len(tok.value) > 2 && isQuote(tok.value[0]) && tok.value[len(tok.value)-1] == tok.value[0] { + name, err := unquoteC(tok.value[1:len(tok.value)-1], rune(tok.value[0])) + if err != nil { + return "", err + } + return name, p.consumeToken("]") + } + + // Consume everything up to "]" + var parts []string + for tok.value != "]" { + parts = append(parts, tok.value) + tok = p.next() + if tok.err != nil { + return "", p.errorf("unrecognized type_url or extension name: %s", tok.err) + } + if p.done && tok.value != "]" { + return "", p.errorf("unclosed type_url or extension name") + } + } + return strings.Join(parts, ""), nil +} + +// consumeOptionalSeparator consumes an optional semicolon or comma. +// It is used in unmarshalMessage to provide backward compatibility. +func (p *textParser) consumeOptionalSeparator() error { + tok := p.next() + if tok.err != nil { + return tok.err + } + if tok.value != ";" && tok.value != "," { + p.back() + } + return nil +} + +func (p *textParser) errorf(format string, a ...interface{}) *ParseError { + pe := &ParseError{fmt.Sprintf(format, a...), p.cur.line, p.cur.offset} + p.cur.err = pe + p.done = true + return pe +} + +func (p *textParser) skipWhitespace() { + i := 0 + for i < len(p.s) && (isWhitespace(p.s[i]) || p.s[i] == '#') { + if p.s[i] == '#' { + // comment; skip to end of line or input + for i < len(p.s) && p.s[i] != '\n' { + i++ + } + if i == len(p.s) { + break + } + } + if p.s[i] == '\n' { + p.line++ + } + i++ + } + p.offset += i + p.s = p.s[i:len(p.s)] + if len(p.s) == 0 { + p.done = true + } +} + +func (p *textParser) advance() { + // Skip whitespace + p.skipWhitespace() + if p.done { + return + } + + // Start of non-whitespace + p.cur.err = nil + p.cur.offset, p.cur.line = p.offset, p.line + p.cur.unquoted = "" + switch p.s[0] { + case '<', '>', '{', '}', ':', '[', ']', ';', ',', '/': + // Single symbol + p.cur.value, p.s = p.s[0:1], p.s[1:len(p.s)] + case '"', '\'': + // Quoted string + i := 1 + for i < len(p.s) && p.s[i] != p.s[0] && p.s[i] != '\n' { + if p.s[i] == '\\' && i+1 < len(p.s) { + // skip escaped char + i++ + } + i++ + } + if i >= len(p.s) || p.s[i] != p.s[0] { + p.errorf("unmatched quote") + return + } + unq, err := unquoteC(p.s[1:i], rune(p.s[0])) + if err != nil { + p.errorf("invalid quoted string %s: %v", p.s[0:i+1], err) + return + } + p.cur.value, p.s = p.s[0:i+1], p.s[i+1:len(p.s)] + p.cur.unquoted = unq + default: + i := 0 + for i < len(p.s) && isIdentOrNumberChar(p.s[i]) { + i++ + } + if i == 0 { + p.errorf("unexpected byte %#x", p.s[0]) + return + } + p.cur.value, p.s = p.s[0:i], p.s[i:len(p.s)] + } + p.offset += len(p.cur.value) +} + +// Back off the parser by one token. Can only be done between calls to next(). +// It makes the next advance() a no-op. +func (p *textParser) back() { p.backed = true } + +// Advances the parser and returns the new current token. +func (p *textParser) next() *token { + if p.backed || p.done { + p.backed = false + return &p.cur + } + p.advance() + if p.done { + p.cur.value = "" + } else if len(p.cur.value) > 0 && isQuote(p.cur.value[0]) { + // Look for multiple quoted strings separated by whitespace, + // and concatenate them. + cat := p.cur + for { + p.skipWhitespace() + if p.done || !isQuote(p.s[0]) { + break + } + p.advance() + if p.cur.err != nil { + return &p.cur + } + cat.value += " " + p.cur.value + cat.unquoted += p.cur.unquoted + } + p.done = false // parser may have seen EOF, but we want to return cat + p.cur = cat + } + return &p.cur +} + +func (p *textParser) consumeToken(s string) error { + tok := p.next() + if tok.err != nil { + return tok.err + } + if tok.value != s { + p.back() + return p.errorf("expected %q, found %q", s, tok.value) + } + return nil +} + +var errBadUTF8 = errors.New("proto: bad UTF-8") + +func unquoteC(s string, quote rune) (string, error) { + // This is based on C++'s tokenizer.cc. + // Despite its name, this is *not* parsing C syntax. + // For instance, "\0" is an invalid quoted string. + + // Avoid allocation in trivial cases. + simple := true + for _, r := range s { + if r == '\\' || r == quote { + simple = false + break + } + } + if simple { + return s, nil + } + + buf := make([]byte, 0, 3*len(s)/2) + for len(s) > 0 { + r, n := utf8.DecodeRuneInString(s) + if r == utf8.RuneError && n == 1 { + return "", errBadUTF8 + } + s = s[n:] + if r != '\\' { + if r < utf8.RuneSelf { + buf = append(buf, byte(r)) + } else { + buf = append(buf, string(r)...) + } + continue + } + + ch, tail, err := unescape(s) + if err != nil { + return "", err + } + buf = append(buf, ch...) + s = tail + } + return string(buf), nil +} + +func unescape(s string) (ch string, tail string, err error) { + r, n := utf8.DecodeRuneInString(s) + if r == utf8.RuneError && n == 1 { + return "", "", errBadUTF8 + } + s = s[n:] + switch r { + case 'a': + return "\a", s, nil + case 'b': + return "\b", s, nil + case 'f': + return "\f", s, nil + case 'n': + return "\n", s, nil + case 'r': + return "\r", s, nil + case 't': + return "\t", s, nil + case 'v': + return "\v", s, nil + case '?': + return "?", s, nil // trigraph workaround + case '\'', '"', '\\': + return string(r), s, nil + case '0', '1', '2', '3', '4', '5', '6', '7': + if len(s) < 2 { + return "", "", fmt.Errorf(`\%c requires 2 following digits`, r) + } + ss := string(r) + s[:2] + s = s[2:] + i, err := strconv.ParseUint(ss, 8, 8) + if err != nil { + return "", "", fmt.Errorf(`\%s contains non-octal digits`, ss) + } + return string([]byte{byte(i)}), s, nil + case 'x', 'X', 'u', 'U': + var n int + switch r { + case 'x', 'X': + n = 2 + case 'u': + n = 4 + case 'U': + n = 8 + } + if len(s) < n { + return "", "", fmt.Errorf(`\%c requires %d following digits`, r, n) + } + ss := s[:n] + s = s[n:] + i, err := strconv.ParseUint(ss, 16, 64) + if err != nil { + return "", "", fmt.Errorf(`\%c%s contains non-hexadecimal digits`, r, ss) + } + if r == 'x' || r == 'X' { + return string([]byte{byte(i)}), s, nil + } + if i > utf8.MaxRune { + return "", "", fmt.Errorf(`\%c%s is not a valid Unicode code point`, r, ss) + } + return string(rune(i)), s, nil + } + return "", "", fmt.Errorf(`unknown escape \%c`, r) +} + +func isIdentOrNumberChar(c byte) bool { + switch { + case 'A' <= c && c <= 'Z', 'a' <= c && c <= 'z': + return true + case '0' <= c && c <= '9': + return true + } + switch c { + case '-', '+', '.', '_': + return true + } + return false +} + +func isWhitespace(c byte) bool { + switch c { + case ' ', '\t', '\n', '\r': + return true + } + return false +} + +func isQuote(c byte) bool { + switch c { + case '"', '\'': + return true + } + return false +} diff --git a/vendor/github.com/golang/protobuf/proto/text_encode.go b/vendor/github.com/golang/protobuf/proto/text_encode.go new file mode 100644 index 00000000..a31134ee --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/text_encode.go @@ -0,0 +1,560 @@ +// Copyright 2010 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "bytes" + "encoding" + "fmt" + "io" + "math" + "sort" + "strings" + + "google.golang.org/protobuf/encoding/prototext" + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/proto" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +const wrapTextMarshalV2 = false + +// TextMarshaler is a configurable text format marshaler. +type TextMarshaler struct { + Compact bool // use compact text format (one line) + ExpandAny bool // expand google.protobuf.Any messages of known types +} + +// Marshal writes the proto text format of m to w. +func (tm *TextMarshaler) Marshal(w io.Writer, m Message) error { + b, err := tm.marshal(m) + if len(b) > 0 { + if _, err := w.Write(b); err != nil { + return err + } + } + return err +} + +// Text returns a proto text formatted string of m. +func (tm *TextMarshaler) Text(m Message) string { + b, _ := tm.marshal(m) + return string(b) +} + +func (tm *TextMarshaler) marshal(m Message) ([]byte, error) { + mr := MessageReflect(m) + if mr == nil || !mr.IsValid() { + return []byte(""), nil + } + + if wrapTextMarshalV2 { + if m, ok := m.(encoding.TextMarshaler); ok { + return m.MarshalText() + } + + opts := prototext.MarshalOptions{ + AllowPartial: true, + EmitUnknown: true, + } + if !tm.Compact { + opts.Indent = " " + } + if !tm.ExpandAny { + opts.Resolver = (*protoregistry.Types)(nil) + } + return opts.Marshal(mr.Interface()) + } else { + w := &textWriter{ + compact: tm.Compact, + expandAny: tm.ExpandAny, + complete: true, + } + + if m, ok := m.(encoding.TextMarshaler); ok { + b, err := m.MarshalText() + if err != nil { + return nil, err + } + w.Write(b) + return w.buf, nil + } + + err := w.writeMessage(mr) + return w.buf, err + } +} + +var ( + defaultTextMarshaler = TextMarshaler{} + compactTextMarshaler = TextMarshaler{Compact: true} +) + +// MarshalText writes the proto text format of m to w. +func MarshalText(w io.Writer, m Message) error { return defaultTextMarshaler.Marshal(w, m) } + +// MarshalTextString returns a proto text formatted string of m. +func MarshalTextString(m Message) string { return defaultTextMarshaler.Text(m) } + +// CompactText writes the compact proto text format of m to w. +func CompactText(w io.Writer, m Message) error { return compactTextMarshaler.Marshal(w, m) } + +// CompactTextString returns a compact proto text formatted string of m. +func CompactTextString(m Message) string { return compactTextMarshaler.Text(m) } + +var ( + newline = []byte("\n") + endBraceNewline = []byte("}\n") + posInf = []byte("inf") + negInf = []byte("-inf") + nan = []byte("nan") +) + +// textWriter is an io.Writer that tracks its indentation level. +type textWriter struct { + compact bool // same as TextMarshaler.Compact + expandAny bool // same as TextMarshaler.ExpandAny + complete bool // whether the current position is a complete line + indent int // indentation level; never negative + buf []byte +} + +func (w *textWriter) Write(p []byte) (n int, _ error) { + newlines := bytes.Count(p, newline) + if newlines == 0 { + if !w.compact && w.complete { + w.writeIndent() + } + w.buf = append(w.buf, p...) + w.complete = false + return len(p), nil + } + + frags := bytes.SplitN(p, newline, newlines+1) + if w.compact { + for i, frag := range frags { + if i > 0 { + w.buf = append(w.buf, ' ') + n++ + } + w.buf = append(w.buf, frag...) + n += len(frag) + } + return n, nil + } + + for i, frag := range frags { + if w.complete { + w.writeIndent() + } + w.buf = append(w.buf, frag...) + n += len(frag) + if i+1 < len(frags) { + w.buf = append(w.buf, '\n') + n++ + } + } + w.complete = len(frags[len(frags)-1]) == 0 + return n, nil +} + +func (w *textWriter) WriteByte(c byte) error { + if w.compact && c == '\n' { + c = ' ' + } + if !w.compact && w.complete { + w.writeIndent() + } + w.buf = append(w.buf, c) + w.complete = c == '\n' + return nil +} + +func (w *textWriter) writeName(fd protoreflect.FieldDescriptor) { + if !w.compact && w.complete { + w.writeIndent() + } + w.complete = false + + if fd.Kind() != protoreflect.GroupKind { + w.buf = append(w.buf, fd.Name()...) + w.WriteByte(':') + } else { + // Use message type name for group field name. + w.buf = append(w.buf, fd.Message().Name()...) + } + + if !w.compact { + w.WriteByte(' ') + } +} + +func requiresQuotes(u string) bool { + // When type URL contains any characters except [0-9A-Za-z./\-]*, it must be quoted. + for _, ch := range u { + switch { + case ch == '.' || ch == '/' || ch == '_': + continue + case '0' <= ch && ch <= '9': + continue + case 'A' <= ch && ch <= 'Z': + continue + case 'a' <= ch && ch <= 'z': + continue + default: + return true + } + } + return false +} + +// writeProto3Any writes an expanded google.protobuf.Any message. +// +// It returns (false, nil) if sv value can't be unmarshaled (e.g. because +// required messages are not linked in). +// +// It returns (true, error) when sv was written in expanded format or an error +// was encountered. +func (w *textWriter) writeProto3Any(m protoreflect.Message) (bool, error) { + md := m.Descriptor() + fdURL := md.Fields().ByName("type_url") + fdVal := md.Fields().ByName("value") + + url := m.Get(fdURL).String() + mt, err := protoregistry.GlobalTypes.FindMessageByURL(url) + if err != nil { + return false, nil + } + + b := m.Get(fdVal).Bytes() + m2 := mt.New() + if err := proto.Unmarshal(b, m2.Interface()); err != nil { + return false, nil + } + w.Write([]byte("[")) + if requiresQuotes(url) { + w.writeQuotedString(url) + } else { + w.Write([]byte(url)) + } + if w.compact { + w.Write([]byte("]:<")) + } else { + w.Write([]byte("]: <\n")) + w.indent++ + } + if err := w.writeMessage(m2); err != nil { + return true, err + } + if w.compact { + w.Write([]byte("> ")) + } else { + w.indent-- + w.Write([]byte(">\n")) + } + return true, nil +} + +func (w *textWriter) writeMessage(m protoreflect.Message) error { + md := m.Descriptor() + if w.expandAny && md.FullName() == "google.protobuf.Any" { + if canExpand, err := w.writeProto3Any(m); canExpand { + return err + } + } + + fds := md.Fields() + for i := 0; i < fds.Len(); { + fd := fds.Get(i) + if od := fd.ContainingOneof(); od != nil { + fd = m.WhichOneof(od) + i += od.Fields().Len() + } else { + i++ + } + if fd == nil || !m.Has(fd) { + continue + } + + switch { + case fd.IsList(): + lv := m.Get(fd).List() + for j := 0; j < lv.Len(); j++ { + w.writeName(fd) + v := lv.Get(j) + if err := w.writeSingularValue(v, fd); err != nil { + return err + } + w.WriteByte('\n') + } + case fd.IsMap(): + kfd := fd.MapKey() + vfd := fd.MapValue() + mv := m.Get(fd).Map() + + type entry struct{ key, val protoreflect.Value } + var entries []entry + mv.Range(func(k protoreflect.MapKey, v protoreflect.Value) bool { + entries = append(entries, entry{k.Value(), v}) + return true + }) + sort.Slice(entries, func(i, j int) bool { + switch kfd.Kind() { + case protoreflect.BoolKind: + return !entries[i].key.Bool() && entries[j].key.Bool() + case protoreflect.Int32Kind, protoreflect.Sint32Kind, protoreflect.Sfixed32Kind, protoreflect.Int64Kind, protoreflect.Sint64Kind, protoreflect.Sfixed64Kind: + return entries[i].key.Int() < entries[j].key.Int() + case protoreflect.Uint32Kind, protoreflect.Fixed32Kind, protoreflect.Uint64Kind, protoreflect.Fixed64Kind: + return entries[i].key.Uint() < entries[j].key.Uint() + case protoreflect.StringKind: + return entries[i].key.String() < entries[j].key.String() + default: + panic("invalid kind") + } + }) + for _, entry := range entries { + w.writeName(fd) + w.WriteByte('<') + if !w.compact { + w.WriteByte('\n') + } + w.indent++ + w.writeName(kfd) + if err := w.writeSingularValue(entry.key, kfd); err != nil { + return err + } + w.WriteByte('\n') + w.writeName(vfd) + if err := w.writeSingularValue(entry.val, vfd); err != nil { + return err + } + w.WriteByte('\n') + w.indent-- + w.WriteByte('>') + w.WriteByte('\n') + } + default: + w.writeName(fd) + if err := w.writeSingularValue(m.Get(fd), fd); err != nil { + return err + } + w.WriteByte('\n') + } + } + + if b := m.GetUnknown(); len(b) > 0 { + w.writeUnknownFields(b) + } + return w.writeExtensions(m) +} + +func (w *textWriter) writeSingularValue(v protoreflect.Value, fd protoreflect.FieldDescriptor) error { + switch fd.Kind() { + case protoreflect.FloatKind, protoreflect.DoubleKind: + switch vf := v.Float(); { + case math.IsInf(vf, +1): + w.Write(posInf) + case math.IsInf(vf, -1): + w.Write(negInf) + case math.IsNaN(vf): + w.Write(nan) + default: + fmt.Fprint(w, v.Interface()) + } + case protoreflect.StringKind: + // NOTE: This does not validate UTF-8 for historical reasons. + w.writeQuotedString(string(v.String())) + case protoreflect.BytesKind: + w.writeQuotedString(string(v.Bytes())) + case protoreflect.MessageKind, protoreflect.GroupKind: + var bra, ket byte = '<', '>' + if fd.Kind() == protoreflect.GroupKind { + bra, ket = '{', '}' + } + w.WriteByte(bra) + if !w.compact { + w.WriteByte('\n') + } + w.indent++ + m := v.Message() + if m2, ok := m.Interface().(encoding.TextMarshaler); ok { + b, err := m2.MarshalText() + if err != nil { + return err + } + w.Write(b) + } else { + w.writeMessage(m) + } + w.indent-- + w.WriteByte(ket) + case protoreflect.EnumKind: + if ev := fd.Enum().Values().ByNumber(v.Enum()); ev != nil { + fmt.Fprint(w, ev.Name()) + } else { + fmt.Fprint(w, v.Enum()) + } + default: + fmt.Fprint(w, v.Interface()) + } + return nil +} + +// writeQuotedString writes a quoted string in the protocol buffer text format. +func (w *textWriter) writeQuotedString(s string) { + w.WriteByte('"') + for i := 0; i < len(s); i++ { + switch c := s[i]; c { + case '\n': + w.buf = append(w.buf, `\n`...) + case '\r': + w.buf = append(w.buf, `\r`...) + case '\t': + w.buf = append(w.buf, `\t`...) + case '"': + w.buf = append(w.buf, `\"`...) + case '\\': + w.buf = append(w.buf, `\\`...) + default: + if isPrint := c >= 0x20 && c < 0x7f; isPrint { + w.buf = append(w.buf, c) + } else { + w.buf = append(w.buf, fmt.Sprintf(`\%03o`, c)...) + } + } + } + w.WriteByte('"') +} + +func (w *textWriter) writeUnknownFields(b []byte) { + if !w.compact { + fmt.Fprintf(w, "/* %d unknown bytes */\n", len(b)) + } + + for len(b) > 0 { + num, wtyp, n := protowire.ConsumeTag(b) + if n < 0 { + return + } + b = b[n:] + + if wtyp == protowire.EndGroupType { + w.indent-- + w.Write(endBraceNewline) + continue + } + fmt.Fprint(w, num) + if wtyp != protowire.StartGroupType { + w.WriteByte(':') + } + if !w.compact || wtyp == protowire.StartGroupType { + w.WriteByte(' ') + } + switch wtyp { + case protowire.VarintType: + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return + } + b = b[n:] + fmt.Fprint(w, v) + case protowire.Fixed32Type: + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return + } + b = b[n:] + fmt.Fprint(w, v) + case protowire.Fixed64Type: + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return + } + b = b[n:] + fmt.Fprint(w, v) + case protowire.BytesType: + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return + } + b = b[n:] + fmt.Fprintf(w, "%q", v) + case protowire.StartGroupType: + w.WriteByte('{') + w.indent++ + default: + fmt.Fprintf(w, "/* unknown wire type %d */", wtyp) + } + w.WriteByte('\n') + } +} + +// writeExtensions writes all the extensions in m. +func (w *textWriter) writeExtensions(m protoreflect.Message) error { + md := m.Descriptor() + if md.ExtensionRanges().Len() == 0 { + return nil + } + + type ext struct { + desc protoreflect.FieldDescriptor + val protoreflect.Value + } + var exts []ext + m.Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + if fd.IsExtension() { + exts = append(exts, ext{fd, v}) + } + return true + }) + sort.Slice(exts, func(i, j int) bool { + return exts[i].desc.Number() < exts[j].desc.Number() + }) + + for _, ext := range exts { + // For message set, use the name of the message as the extension name. + name := string(ext.desc.FullName()) + if isMessageSet(ext.desc.ContainingMessage()) { + name = strings.TrimSuffix(name, ".message_set_extension") + } + + if !ext.desc.IsList() { + if err := w.writeSingularExtension(name, ext.val, ext.desc); err != nil { + return err + } + } else { + lv := ext.val.List() + for i := 0; i < lv.Len(); i++ { + if err := w.writeSingularExtension(name, lv.Get(i), ext.desc); err != nil { + return err + } + } + } + } + return nil +} + +func (w *textWriter) writeSingularExtension(name string, v protoreflect.Value, fd protoreflect.FieldDescriptor) error { + fmt.Fprintf(w, "[%s]:", name) + if !w.compact { + w.WriteByte(' ') + } + if err := w.writeSingularValue(v, fd); err != nil { + return err + } + w.WriteByte('\n') + return nil +} + +func (w *textWriter) writeIndent() { + if !w.complete { + return + } + for i := 0; i < w.indent*2; i++ { + w.buf = append(w.buf, ' ') + } + w.complete = false +} diff --git a/vendor/github.com/golang/protobuf/proto/text_parser.go b/vendor/github.com/golang/protobuf/proto/text_parser.go deleted file mode 100644 index bb55a3af..00000000 --- a/vendor/github.com/golang/protobuf/proto/text_parser.go +++ /dev/null @@ -1,880 +0,0 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2010 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package proto - -// Functions for parsing the Text protocol buffer format. -// TODO: message sets. - -import ( - "encoding" - "errors" - "fmt" - "reflect" - "strconv" - "strings" - "unicode/utf8" -) - -// Error string emitted when deserializing Any and fields are already set -const anyRepeatedlyUnpacked = "Any message unpacked multiple times, or %q already set" - -type ParseError struct { - Message string - Line int // 1-based line number - Offset int // 0-based byte offset from start of input -} - -func (p *ParseError) Error() string { - if p.Line == 1 { - // show offset only for first line - return fmt.Sprintf("line 1.%d: %v", p.Offset, p.Message) - } - return fmt.Sprintf("line %d: %v", p.Line, p.Message) -} - -type token struct { - value string - err *ParseError - line int // line number - offset int // byte number from start of input, not start of line - unquoted string // the unquoted version of value, if it was a quoted string -} - -func (t *token) String() string { - if t.err == nil { - return fmt.Sprintf("%q (line=%d, offset=%d)", t.value, t.line, t.offset) - } - return fmt.Sprintf("parse error: %v", t.err) -} - -type textParser struct { - s string // remaining input - done bool // whether the parsing is finished (success or error) - backed bool // whether back() was called - offset, line int - cur token -} - -func newTextParser(s string) *textParser { - p := new(textParser) - p.s = s - p.line = 1 - p.cur.line = 1 - return p -} - -func (p *textParser) errorf(format string, a ...interface{}) *ParseError { - pe := &ParseError{fmt.Sprintf(format, a...), p.cur.line, p.cur.offset} - p.cur.err = pe - p.done = true - return pe -} - -// Numbers and identifiers are matched by [-+._A-Za-z0-9] -func isIdentOrNumberChar(c byte) bool { - switch { - case 'A' <= c && c <= 'Z', 'a' <= c && c <= 'z': - return true - case '0' <= c && c <= '9': - return true - } - switch c { - case '-', '+', '.', '_': - return true - } - return false -} - -func isWhitespace(c byte) bool { - switch c { - case ' ', '\t', '\n', '\r': - return true - } - return false -} - -func isQuote(c byte) bool { - switch c { - case '"', '\'': - return true - } - return false -} - -func (p *textParser) skipWhitespace() { - i := 0 - for i < len(p.s) && (isWhitespace(p.s[i]) || p.s[i] == '#') { - if p.s[i] == '#' { - // comment; skip to end of line or input - for i < len(p.s) && p.s[i] != '\n' { - i++ - } - if i == len(p.s) { - break - } - } - if p.s[i] == '\n' { - p.line++ - } - i++ - } - p.offset += i - p.s = p.s[i:len(p.s)] - if len(p.s) == 0 { - p.done = true - } -} - -func (p *textParser) advance() { - // Skip whitespace - p.skipWhitespace() - if p.done { - return - } - - // Start of non-whitespace - p.cur.err = nil - p.cur.offset, p.cur.line = p.offset, p.line - p.cur.unquoted = "" - switch p.s[0] { - case '<', '>', '{', '}', ':', '[', ']', ';', ',', '/': - // Single symbol - p.cur.value, p.s = p.s[0:1], p.s[1:len(p.s)] - case '"', '\'': - // Quoted string - i := 1 - for i < len(p.s) && p.s[i] != p.s[0] && p.s[i] != '\n' { - if p.s[i] == '\\' && i+1 < len(p.s) { - // skip escaped char - i++ - } - i++ - } - if i >= len(p.s) || p.s[i] != p.s[0] { - p.errorf("unmatched quote") - return - } - unq, err := unquoteC(p.s[1:i], rune(p.s[0])) - if err != nil { - p.errorf("invalid quoted string %s: %v", p.s[0:i+1], err) - return - } - p.cur.value, p.s = p.s[0:i+1], p.s[i+1:len(p.s)] - p.cur.unquoted = unq - default: - i := 0 - for i < len(p.s) && isIdentOrNumberChar(p.s[i]) { - i++ - } - if i == 0 { - p.errorf("unexpected byte %#x", p.s[0]) - return - } - p.cur.value, p.s = p.s[0:i], p.s[i:len(p.s)] - } - p.offset += len(p.cur.value) -} - -var ( - errBadUTF8 = errors.New("proto: bad UTF-8") -) - -func unquoteC(s string, quote rune) (string, error) { - // This is based on C++'s tokenizer.cc. - // Despite its name, this is *not* parsing C syntax. - // For instance, "\0" is an invalid quoted string. - - // Avoid allocation in trivial cases. - simple := true - for _, r := range s { - if r == '\\' || r == quote { - simple = false - break - } - } - if simple { - return s, nil - } - - buf := make([]byte, 0, 3*len(s)/2) - for len(s) > 0 { - r, n := utf8.DecodeRuneInString(s) - if r == utf8.RuneError && n == 1 { - return "", errBadUTF8 - } - s = s[n:] - if r != '\\' { - if r < utf8.RuneSelf { - buf = append(buf, byte(r)) - } else { - buf = append(buf, string(r)...) - } - continue - } - - ch, tail, err := unescape(s) - if err != nil { - return "", err - } - buf = append(buf, ch...) - s = tail - } - return string(buf), nil -} - -func unescape(s string) (ch string, tail string, err error) { - r, n := utf8.DecodeRuneInString(s) - if r == utf8.RuneError && n == 1 { - return "", "", errBadUTF8 - } - s = s[n:] - switch r { - case 'a': - return "\a", s, nil - case 'b': - return "\b", s, nil - case 'f': - return "\f", s, nil - case 'n': - return "\n", s, nil - case 'r': - return "\r", s, nil - case 't': - return "\t", s, nil - case 'v': - return "\v", s, nil - case '?': - return "?", s, nil // trigraph workaround - case '\'', '"', '\\': - return string(r), s, nil - case '0', '1', '2', '3', '4', '5', '6', '7': - if len(s) < 2 { - return "", "", fmt.Errorf(`\%c requires 2 following digits`, r) - } - ss := string(r) + s[:2] - s = s[2:] - i, err := strconv.ParseUint(ss, 8, 8) - if err != nil { - return "", "", fmt.Errorf(`\%s contains non-octal digits`, ss) - } - return string([]byte{byte(i)}), s, nil - case 'x', 'X', 'u', 'U': - var n int - switch r { - case 'x', 'X': - n = 2 - case 'u': - n = 4 - case 'U': - n = 8 - } - if len(s) < n { - return "", "", fmt.Errorf(`\%c requires %d following digits`, r, n) - } - ss := s[:n] - s = s[n:] - i, err := strconv.ParseUint(ss, 16, 64) - if err != nil { - return "", "", fmt.Errorf(`\%c%s contains non-hexadecimal digits`, r, ss) - } - if r == 'x' || r == 'X' { - return string([]byte{byte(i)}), s, nil - } - if i > utf8.MaxRune { - return "", "", fmt.Errorf(`\%c%s is not a valid Unicode code point`, r, ss) - } - return string(i), s, nil - } - return "", "", fmt.Errorf(`unknown escape \%c`, r) -} - -// Back off the parser by one token. Can only be done between calls to next(). -// It makes the next advance() a no-op. -func (p *textParser) back() { p.backed = true } - -// Advances the parser and returns the new current token. -func (p *textParser) next() *token { - if p.backed || p.done { - p.backed = false - return &p.cur - } - p.advance() - if p.done { - p.cur.value = "" - } else if len(p.cur.value) > 0 && isQuote(p.cur.value[0]) { - // Look for multiple quoted strings separated by whitespace, - // and concatenate them. - cat := p.cur - for { - p.skipWhitespace() - if p.done || !isQuote(p.s[0]) { - break - } - p.advance() - if p.cur.err != nil { - return &p.cur - } - cat.value += " " + p.cur.value - cat.unquoted += p.cur.unquoted - } - p.done = false // parser may have seen EOF, but we want to return cat - p.cur = cat - } - return &p.cur -} - -func (p *textParser) consumeToken(s string) error { - tok := p.next() - if tok.err != nil { - return tok.err - } - if tok.value != s { - p.back() - return p.errorf("expected %q, found %q", s, tok.value) - } - return nil -} - -// Return a RequiredNotSetError indicating which required field was not set. -func (p *textParser) missingRequiredFieldError(sv reflect.Value) *RequiredNotSetError { - st := sv.Type() - sprops := GetProperties(st) - for i := 0; i < st.NumField(); i++ { - if !isNil(sv.Field(i)) { - continue - } - - props := sprops.Prop[i] - if props.Required { - return &RequiredNotSetError{fmt.Sprintf("%v.%v", st, props.OrigName)} - } - } - return &RequiredNotSetError{fmt.Sprintf("%v.", st)} // should not happen -} - -// Returns the index in the struct for the named field, as well as the parsed tag properties. -func structFieldByName(sprops *StructProperties, name string) (int, *Properties, bool) { - i, ok := sprops.decoderOrigNames[name] - if ok { - return i, sprops.Prop[i], true - } - return -1, nil, false -} - -// Consume a ':' from the input stream (if the next token is a colon), -// returning an error if a colon is needed but not present. -func (p *textParser) checkForColon(props *Properties, typ reflect.Type) *ParseError { - tok := p.next() - if tok.err != nil { - return tok.err - } - if tok.value != ":" { - // Colon is optional when the field is a group or message. - needColon := true - switch props.Wire { - case "group": - needColon = false - case "bytes": - // A "bytes" field is either a message, a string, or a repeated field; - // those three become *T, *string and []T respectively, so we can check for - // this field being a pointer to a non-string. - if typ.Kind() == reflect.Ptr { - // *T or *string - if typ.Elem().Kind() == reflect.String { - break - } - } else if typ.Kind() == reflect.Slice { - // []T or []*T - if typ.Elem().Kind() != reflect.Ptr { - break - } - } else if typ.Kind() == reflect.String { - // The proto3 exception is for a string field, - // which requires a colon. - break - } - needColon = false - } - if needColon { - return p.errorf("expected ':', found %q", tok.value) - } - p.back() - } - return nil -} - -func (p *textParser) readStruct(sv reflect.Value, terminator string) error { - st := sv.Type() - sprops := GetProperties(st) - reqCount := sprops.reqCount - var reqFieldErr error - fieldSet := make(map[string]bool) - // A struct is a sequence of "name: value", terminated by one of - // '>' or '}', or the end of the input. A name may also be - // "[extension]" or "[type/url]". - // - // The whole struct can also be an expanded Any message, like: - // [type/url] < ... struct contents ... > - for { - tok := p.next() - if tok.err != nil { - return tok.err - } - if tok.value == terminator { - break - } - if tok.value == "[" { - // Looks like an extension or an Any. - // - // TODO: Check whether we need to handle - // namespace rooted names (e.g. ".something.Foo"). - extName, err := p.consumeExtName() - if err != nil { - return err - } - - if s := strings.LastIndex(extName, "/"); s >= 0 { - // If it contains a slash, it's an Any type URL. - messageName := extName[s+1:] - mt := MessageType(messageName) - if mt == nil { - return p.errorf("unrecognized message %q in google.protobuf.Any", messageName) - } - tok = p.next() - if tok.err != nil { - return tok.err - } - // consume an optional colon - if tok.value == ":" { - tok = p.next() - if tok.err != nil { - return tok.err - } - } - var terminator string - switch tok.value { - case "<": - terminator = ">" - case "{": - terminator = "}" - default: - return p.errorf("expected '{' or '<', found %q", tok.value) - } - v := reflect.New(mt.Elem()) - if pe := p.readStruct(v.Elem(), terminator); pe != nil { - return pe - } - b, err := Marshal(v.Interface().(Message)) - if err != nil { - return p.errorf("failed to marshal message of type %q: %v", messageName, err) - } - if fieldSet["type_url"] { - return p.errorf(anyRepeatedlyUnpacked, "type_url") - } - if fieldSet["value"] { - return p.errorf(anyRepeatedlyUnpacked, "value") - } - sv.FieldByName("TypeUrl").SetString(extName) - sv.FieldByName("Value").SetBytes(b) - fieldSet["type_url"] = true - fieldSet["value"] = true - continue - } - - var desc *ExtensionDesc - // This could be faster, but it's functional. - // TODO: Do something smarter than a linear scan. - for _, d := range RegisteredExtensions(reflect.New(st).Interface().(Message)) { - if d.Name == extName { - desc = d - break - } - } - if desc == nil { - return p.errorf("unrecognized extension %q", extName) - } - - props := &Properties{} - props.Parse(desc.Tag) - - typ := reflect.TypeOf(desc.ExtensionType) - if err := p.checkForColon(props, typ); err != nil { - return err - } - - rep := desc.repeated() - - // Read the extension structure, and set it in - // the value we're constructing. - var ext reflect.Value - if !rep { - ext = reflect.New(typ).Elem() - } else { - ext = reflect.New(typ.Elem()).Elem() - } - if err := p.readAny(ext, props); err != nil { - if _, ok := err.(*RequiredNotSetError); !ok { - return err - } - reqFieldErr = err - } - ep := sv.Addr().Interface().(Message) - if !rep { - SetExtension(ep, desc, ext.Interface()) - } else { - old, err := GetExtension(ep, desc) - var sl reflect.Value - if err == nil { - sl = reflect.ValueOf(old) // existing slice - } else { - sl = reflect.MakeSlice(typ, 0, 1) - } - sl = reflect.Append(sl, ext) - SetExtension(ep, desc, sl.Interface()) - } - if err := p.consumeOptionalSeparator(); err != nil { - return err - } - continue - } - - // This is a normal, non-extension field. - name := tok.value - var dst reflect.Value - fi, props, ok := structFieldByName(sprops, name) - if ok { - dst = sv.Field(fi) - } else if oop, ok := sprops.OneofTypes[name]; ok { - // It is a oneof. - props = oop.Prop - nv := reflect.New(oop.Type.Elem()) - dst = nv.Elem().Field(0) - field := sv.Field(oop.Field) - if !field.IsNil() { - return p.errorf("field '%s' would overwrite already parsed oneof '%s'", name, sv.Type().Field(oop.Field).Name) - } - field.Set(nv) - } - if !dst.IsValid() { - return p.errorf("unknown field name %q in %v", name, st) - } - - if dst.Kind() == reflect.Map { - // Consume any colon. - if err := p.checkForColon(props, dst.Type()); err != nil { - return err - } - - // Construct the map if it doesn't already exist. - if dst.IsNil() { - dst.Set(reflect.MakeMap(dst.Type())) - } - key := reflect.New(dst.Type().Key()).Elem() - val := reflect.New(dst.Type().Elem()).Elem() - - // The map entry should be this sequence of tokens: - // < key : KEY value : VALUE > - // However, implementations may omit key or value, and technically - // we should support them in any order. See b/28924776 for a time - // this went wrong. - - tok := p.next() - var terminator string - switch tok.value { - case "<": - terminator = ">" - case "{": - terminator = "}" - default: - return p.errorf("expected '{' or '<', found %q", tok.value) - } - for { - tok := p.next() - if tok.err != nil { - return tok.err - } - if tok.value == terminator { - break - } - switch tok.value { - case "key": - if err := p.consumeToken(":"); err != nil { - return err - } - if err := p.readAny(key, props.MapKeyProp); err != nil { - return err - } - if err := p.consumeOptionalSeparator(); err != nil { - return err - } - case "value": - if err := p.checkForColon(props.MapValProp, dst.Type().Elem()); err != nil { - return err - } - if err := p.readAny(val, props.MapValProp); err != nil { - return err - } - if err := p.consumeOptionalSeparator(); err != nil { - return err - } - default: - p.back() - return p.errorf(`expected "key", "value", or %q, found %q`, terminator, tok.value) - } - } - - dst.SetMapIndex(key, val) - continue - } - - // Check that it's not already set if it's not a repeated field. - if !props.Repeated && fieldSet[name] { - return p.errorf("non-repeated field %q was repeated", name) - } - - if err := p.checkForColon(props, dst.Type()); err != nil { - return err - } - - // Parse into the field. - fieldSet[name] = true - if err := p.readAny(dst, props); err != nil { - if _, ok := err.(*RequiredNotSetError); !ok { - return err - } - reqFieldErr = err - } - if props.Required { - reqCount-- - } - - if err := p.consumeOptionalSeparator(); err != nil { - return err - } - - } - - if reqCount > 0 { - return p.missingRequiredFieldError(sv) - } - return reqFieldErr -} - -// consumeExtName consumes extension name or expanded Any type URL and the -// following ']'. It returns the name or URL consumed. -func (p *textParser) consumeExtName() (string, error) { - tok := p.next() - if tok.err != nil { - return "", tok.err - } - - // If extension name or type url is quoted, it's a single token. - if len(tok.value) > 2 && isQuote(tok.value[0]) && tok.value[len(tok.value)-1] == tok.value[0] { - name, err := unquoteC(tok.value[1:len(tok.value)-1], rune(tok.value[0])) - if err != nil { - return "", err - } - return name, p.consumeToken("]") - } - - // Consume everything up to "]" - var parts []string - for tok.value != "]" { - parts = append(parts, tok.value) - tok = p.next() - if tok.err != nil { - return "", p.errorf("unrecognized type_url or extension name: %s", tok.err) - } - if p.done && tok.value != "]" { - return "", p.errorf("unclosed type_url or extension name") - } - } - return strings.Join(parts, ""), nil -} - -// consumeOptionalSeparator consumes an optional semicolon or comma. -// It is used in readStruct to provide backward compatibility. -func (p *textParser) consumeOptionalSeparator() error { - tok := p.next() - if tok.err != nil { - return tok.err - } - if tok.value != ";" && tok.value != "," { - p.back() - } - return nil -} - -func (p *textParser) readAny(v reflect.Value, props *Properties) error { - tok := p.next() - if tok.err != nil { - return tok.err - } - if tok.value == "" { - return p.errorf("unexpected EOF") - } - - switch fv := v; fv.Kind() { - case reflect.Slice: - at := v.Type() - if at.Elem().Kind() == reflect.Uint8 { - // Special case for []byte - if tok.value[0] != '"' && tok.value[0] != '\'' { - // Deliberately written out here, as the error after - // this switch statement would write "invalid []byte: ...", - // which is not as user-friendly. - return p.errorf("invalid string: %v", tok.value) - } - bytes := []byte(tok.unquoted) - fv.Set(reflect.ValueOf(bytes)) - return nil - } - // Repeated field. - if tok.value == "[" { - // Repeated field with list notation, like [1,2,3]. - for { - fv.Set(reflect.Append(fv, reflect.New(at.Elem()).Elem())) - err := p.readAny(fv.Index(fv.Len()-1), props) - if err != nil { - return err - } - tok := p.next() - if tok.err != nil { - return tok.err - } - if tok.value == "]" { - break - } - if tok.value != "," { - return p.errorf("Expected ']' or ',' found %q", tok.value) - } - } - return nil - } - // One value of the repeated field. - p.back() - fv.Set(reflect.Append(fv, reflect.New(at.Elem()).Elem())) - return p.readAny(fv.Index(fv.Len()-1), props) - case reflect.Bool: - // true/1/t/True or false/f/0/False. - switch tok.value { - case "true", "1", "t", "True": - fv.SetBool(true) - return nil - case "false", "0", "f", "False": - fv.SetBool(false) - return nil - } - case reflect.Float32, reflect.Float64: - v := tok.value - // Ignore 'f' for compatibility with output generated by C++, but don't - // remove 'f' when the value is "-inf" or "inf". - if strings.HasSuffix(v, "f") && tok.value != "-inf" && tok.value != "inf" { - v = v[:len(v)-1] - } - if f, err := strconv.ParseFloat(v, fv.Type().Bits()); err == nil { - fv.SetFloat(f) - return nil - } - case reflect.Int32: - if x, err := strconv.ParseInt(tok.value, 0, 32); err == nil { - fv.SetInt(x) - return nil - } - - if len(props.Enum) == 0 { - break - } - m, ok := enumValueMaps[props.Enum] - if !ok { - break - } - x, ok := m[tok.value] - if !ok { - break - } - fv.SetInt(int64(x)) - return nil - case reflect.Int64: - if x, err := strconv.ParseInt(tok.value, 0, 64); err == nil { - fv.SetInt(x) - return nil - } - - case reflect.Ptr: - // A basic field (indirected through pointer), or a repeated message/group - p.back() - fv.Set(reflect.New(fv.Type().Elem())) - return p.readAny(fv.Elem(), props) - case reflect.String: - if tok.value[0] == '"' || tok.value[0] == '\'' { - fv.SetString(tok.unquoted) - return nil - } - case reflect.Struct: - var terminator string - switch tok.value { - case "{": - terminator = "}" - case "<": - terminator = ">" - default: - return p.errorf("expected '{' or '<', found %q", tok.value) - } - // TODO: Handle nested messages which implement encoding.TextUnmarshaler. - return p.readStruct(fv, terminator) - case reflect.Uint32: - if x, err := strconv.ParseUint(tok.value, 0, 32); err == nil { - fv.SetUint(uint64(x)) - return nil - } - case reflect.Uint64: - if x, err := strconv.ParseUint(tok.value, 0, 64); err == nil { - fv.SetUint(x) - return nil - } - } - return p.errorf("invalid %v: %v", v.Type(), tok.value) -} - -// UnmarshalText reads a protocol buffer in Text format. UnmarshalText resets pb -// before starting to unmarshal, so any existing data in pb is always removed. -// If a required field is not set and no other error occurs, -// UnmarshalText returns *RequiredNotSetError. -func UnmarshalText(s string, pb Message) error { - if um, ok := pb.(encoding.TextUnmarshaler); ok { - return um.UnmarshalText([]byte(s)) - } - pb.Reset() - v := reflect.ValueOf(pb) - return newTextParser(s).readStruct(v.Elem(), "") -} diff --git a/vendor/github.com/golang/protobuf/proto/wire.go b/vendor/github.com/golang/protobuf/proto/wire.go new file mode 100644 index 00000000..d7c28da5 --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/wire.go @@ -0,0 +1,78 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + protoV2 "google.golang.org/protobuf/proto" + "google.golang.org/protobuf/runtime/protoiface" +) + +// Size returns the size in bytes of the wire-format encoding of m. +func Size(m Message) int { + if m == nil { + return 0 + } + mi := MessageV2(m) + return protoV2.Size(mi) +} + +// Marshal returns the wire-format encoding of m. +func Marshal(m Message) ([]byte, error) { + b, err := marshalAppend(nil, m, false) + if b == nil { + b = zeroBytes + } + return b, err +} + +var zeroBytes = make([]byte, 0, 0) + +func marshalAppend(buf []byte, m Message, deterministic bool) ([]byte, error) { + if m == nil { + return nil, ErrNil + } + mi := MessageV2(m) + nbuf, err := protoV2.MarshalOptions{ + Deterministic: deterministic, + AllowPartial: true, + }.MarshalAppend(buf, mi) + if err != nil { + return buf, err + } + if len(buf) == len(nbuf) { + if !mi.ProtoReflect().IsValid() { + return buf, ErrNil + } + } + return nbuf, checkRequiredNotSet(mi) +} + +// Unmarshal parses a wire-format message in b and places the decoded results in m. +// +// Unmarshal resets m before starting to unmarshal, so any existing data in m is always +// removed. Use UnmarshalMerge to preserve and append to existing data. +func Unmarshal(b []byte, m Message) error { + m.Reset() + return UnmarshalMerge(b, m) +} + +// UnmarshalMerge parses a wire-format message in b and places the decoded results in m. +func UnmarshalMerge(b []byte, m Message) error { + mi := MessageV2(m) + out, err := protoV2.UnmarshalOptions{ + AllowPartial: true, + Merge: true, + }.UnmarshalState(protoiface.UnmarshalInput{ + Buf: b, + Message: mi.ProtoReflect(), + }) + if err != nil { + return err + } + if out.Flags&protoiface.UnmarshalInitialized > 0 { + return nil + } + return checkRequiredNotSet(mi) +} diff --git a/vendor/github.com/golang/protobuf/proto/wrappers.go b/vendor/github.com/golang/protobuf/proto/wrappers.go new file mode 100644 index 00000000..398e3485 --- /dev/null +++ b/vendor/github.com/golang/protobuf/proto/wrappers.go @@ -0,0 +1,34 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +// Bool stores v in a new bool value and returns a pointer to it. +func Bool(v bool) *bool { return &v } + +// Int stores v in a new int32 value and returns a pointer to it. +// +// Deprecated: Use Int32 instead. +func Int(v int) *int32 { return Int32(int32(v)) } + +// Int32 stores v in a new int32 value and returns a pointer to it. +func Int32(v int32) *int32 { return &v } + +// Int64 stores v in a new int64 value and returns a pointer to it. +func Int64(v int64) *int64 { return &v } + +// Uint32 stores v in a new uint32 value and returns a pointer to it. +func Uint32(v uint32) *uint32 { return &v } + +// Uint64 stores v in a new uint64 value and returns a pointer to it. +func Uint64(v uint64) *uint64 { return &v } + +// Float32 stores v in a new float32 value and returns a pointer to it. +func Float32(v float32) *float32 { return &v } + +// Float64 stores v in a new float64 value and returns a pointer to it. +func Float64(v float64) *float64 { return &v } + +// String stores v in a new string value and returns a pointer to it. +func String(v string) *string { return &v } diff --git a/vendor/github.com/golang/protobuf/ptypes/any.go b/vendor/github.com/golang/protobuf/ptypes/any.go index 70276e8f..e729dcff 100644 --- a/vendor/github.com/golang/protobuf/ptypes/any.go +++ b/vendor/github.com/golang/protobuf/ptypes/any.go @@ -1,141 +1,165 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2016 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2016 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. package ptypes -// This file implements functions to marshal proto.Message to/from -// google.protobuf.Any message. - import ( "fmt" - "reflect" "strings" "github.com/golang/protobuf/proto" - "github.com/golang/protobuf/ptypes/any" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" + + anypb "github.com/golang/protobuf/ptypes/any" ) -const googleApis = "type.googleapis.com/" +const urlPrefix = "type.googleapis.com/" -// AnyMessageName returns the name of the message contained in a google.protobuf.Any message. -// -// Note that regular type assertions should be done using the Is -// function. AnyMessageName is provided for less common use cases like filtering a -// sequence of Any messages based on a set of allowed message type names. -func AnyMessageName(any *any.Any) (string, error) { +// AnyMessageName returns the message name contained in an anypb.Any message. +// Most type assertions should use the Is function instead. +func AnyMessageName(any *anypb.Any) (string, error) { + name, err := anyMessageName(any) + return string(name), err +} +func anyMessageName(any *anypb.Any) (protoreflect.FullName, error) { if any == nil { return "", fmt.Errorf("message is nil") } - slash := strings.LastIndex(any.TypeUrl, "/") - if slash < 0 { + name := protoreflect.FullName(any.TypeUrl) + if i := strings.LastIndex(any.TypeUrl, "/"); i >= 0 { + name = name[i+len("/"):] + } + if !name.IsValid() { return "", fmt.Errorf("message type url %q is invalid", any.TypeUrl) } - return any.TypeUrl[slash+1:], nil + return name, nil } -// MarshalAny takes the protocol buffer and encodes it into google.protobuf.Any. -func MarshalAny(pb proto.Message) (*any.Any, error) { - value, err := proto.Marshal(pb) +// MarshalAny marshals the given message m into an anypb.Any message. +func MarshalAny(m proto.Message) (*anypb.Any, error) { + switch dm := m.(type) { + case DynamicAny: + m = dm.Message + case *DynamicAny: + if dm == nil { + return nil, proto.ErrNil + } + m = dm.Message + } + b, err := proto.Marshal(m) if err != nil { return nil, err } - return &any.Any{TypeUrl: googleApis + proto.MessageName(pb), Value: value}, nil -} - -// DynamicAny is a value that can be passed to UnmarshalAny to automatically -// allocate a proto.Message for the type specified in a google.protobuf.Any -// message. The allocated message is stored in the embedded proto.Message. -// -// Example: -// -// var x ptypes.DynamicAny -// if err := ptypes.UnmarshalAny(a, &x); err != nil { ... } -// fmt.Printf("unmarshaled message: %v", x.Message) -type DynamicAny struct { - proto.Message + return &anypb.Any{TypeUrl: urlPrefix + proto.MessageName(m), Value: b}, nil } -// Empty returns a new proto.Message of the type specified in a -// google.protobuf.Any message. It returns an error if corresponding message -// type isn't linked in. -func Empty(any *any.Any) (proto.Message, error) { - aname, err := AnyMessageName(any) +// Empty returns a new message of the type specified in an anypb.Any message. +// It returns protoregistry.NotFound if the corresponding message type could not +// be resolved in the global registry. +func Empty(any *anypb.Any) (proto.Message, error) { + name, err := anyMessageName(any) if err != nil { return nil, err } - - t := proto.MessageType(aname) - if t == nil { - return nil, fmt.Errorf("any: message type %q isn't linked in", aname) + mt, err := protoregistry.GlobalTypes.FindMessageByName(name) + if err != nil { + return nil, err } - return reflect.New(t.Elem()).Interface().(proto.Message), nil + return proto.MessageV1(mt.New().Interface()), nil } -// UnmarshalAny parses the protocol buffer representation in a google.protobuf.Any -// message and places the decoded result in pb. It returns an error if type of -// contents of Any message does not match type of pb message. +// UnmarshalAny unmarshals the encoded value contained in the anypb.Any message +// into the provided message m. It returns an error if the target message +// does not match the type in the Any message or if an unmarshal error occurs. // -// pb can be a proto.Message, or a *DynamicAny. -func UnmarshalAny(any *any.Any, pb proto.Message) error { - if d, ok := pb.(*DynamicAny); ok { - if d.Message == nil { +// The target message m may be a *DynamicAny message. If the underlying message +// type could not be resolved, then this returns protoregistry.NotFound. +func UnmarshalAny(any *anypb.Any, m proto.Message) error { + if dm, ok := m.(*DynamicAny); ok { + if dm.Message == nil { var err error - d.Message, err = Empty(any) + dm.Message, err = Empty(any) if err != nil { return err } } - return UnmarshalAny(any, d.Message) + m = dm.Message } - aname, err := AnyMessageName(any) + anyName, err := AnyMessageName(any) if err != nil { return err } - - mname := proto.MessageName(pb) - if aname != mname { - return fmt.Errorf("mismatched message type: got %q want %q", aname, mname) + msgName := proto.MessageName(m) + if anyName != msgName { + return fmt.Errorf("mismatched message type: got %q want %q", anyName, msgName) } - return proto.Unmarshal(any.Value, pb) + return proto.Unmarshal(any.Value, m) } -// Is returns true if any value contains a given message type. -func Is(any *any.Any, pb proto.Message) bool { - // The following is equivalent to AnyMessageName(any) == proto.MessageName(pb), - // but it avoids scanning TypeUrl for the slash. - if any == nil { +// Is reports whether the Any message contains a message of the specified type. +func Is(any *anypb.Any, m proto.Message) bool { + if any == nil || m == nil { return false } - name := proto.MessageName(pb) - prefix := len(any.TypeUrl) - len(name) - return prefix >= 1 && any.TypeUrl[prefix-1] == '/' && any.TypeUrl[prefix:] == name + name := proto.MessageName(m) + if !strings.HasSuffix(any.TypeUrl, name) { + return false + } + return len(any.TypeUrl) == len(name) || any.TypeUrl[len(any.TypeUrl)-len(name)-1] == '/' +} + +// DynamicAny is a value that can be passed to UnmarshalAny to automatically +// allocate a proto.Message for the type specified in an anypb.Any message. +// The allocated message is stored in the embedded proto.Message. +// +// Example: +// var x ptypes.DynamicAny +// if err := ptypes.UnmarshalAny(a, &x); err != nil { ... } +// fmt.Printf("unmarshaled message: %v", x.Message) +type DynamicAny struct{ proto.Message } + +func (m DynamicAny) String() string { + if m.Message == nil { + return "" + } + return m.Message.String() +} +func (m DynamicAny) Reset() { + if m.Message == nil { + return + } + m.Message.Reset() +} +func (m DynamicAny) ProtoMessage() { + return +} +func (m DynamicAny) ProtoReflect() protoreflect.Message { + if m.Message == nil { + return nil + } + return dynamicAny{proto.MessageReflect(m.Message)} +} + +type dynamicAny struct{ protoreflect.Message } + +func (m dynamicAny) Type() protoreflect.MessageType { + return dynamicAnyType{m.Message.Type()} +} +func (m dynamicAny) New() protoreflect.Message { + return dynamicAnyType{m.Message.Type()}.New() +} +func (m dynamicAny) Interface() protoreflect.ProtoMessage { + return DynamicAny{proto.MessageV1(m.Message.Interface())} +} + +type dynamicAnyType struct{ protoreflect.MessageType } + +func (t dynamicAnyType) New() protoreflect.Message { + return dynamicAny{t.MessageType.New()} +} +func (t dynamicAnyType) Zero() protoreflect.Message { + return dynamicAny{t.MessageType.Zero()} } diff --git a/vendor/github.com/golang/protobuf/ptypes/any/any.pb.go b/vendor/github.com/golang/protobuf/ptypes/any/any.pb.go index 7b0ad1ad..0ef27d33 100644 --- a/vendor/github.com/golang/protobuf/ptypes/any/any.pb.go +++ b/vendor/github.com/golang/protobuf/ptypes/any/any.pb.go @@ -1,203 +1,62 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: google/protobuf/any.proto +// source: github.com/golang/protobuf/ptypes/any/any.proto package any import ( - fmt "fmt" - proto "github.com/golang/protobuf/proto" - math "math" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + anypb "google.golang.org/protobuf/types/known/anypb" + reflect "reflect" ) -// Reference imports to suppress errors if they are not otherwise used. -var _ = proto.Marshal -var _ = fmt.Errorf -var _ = math.Inf +// Symbols defined in public import of google/protobuf/any.proto. -// This is a compile-time assertion to ensure that this generated file -// is compatible with the proto package it is being compiled against. -// A compilation error at this line likely means your copy of the -// proto package needs to be updated. -const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package +type Any = anypb.Any -// `Any` contains an arbitrary serialized protocol buffer message along with a -// URL that describes the type of the serialized message. -// -// Protobuf library provides support to pack/unpack Any values in the form -// of utility functions or additional generated methods of the Any type. -// -// Example 1: Pack and unpack a message in C++. -// -// Foo foo = ...; -// Any any; -// any.PackFrom(foo); -// ... -// if (any.UnpackTo(&foo)) { -// ... -// } -// -// Example 2: Pack and unpack a message in Java. -// -// Foo foo = ...; -// Any any = Any.pack(foo); -// ... -// if (any.is(Foo.class)) { -// foo = any.unpack(Foo.class); -// } -// -// Example 3: Pack and unpack a message in Python. -// -// foo = Foo(...) -// any = Any() -// any.Pack(foo) -// ... -// if any.Is(Foo.DESCRIPTOR): -// any.Unpack(foo) -// ... -// -// Example 4: Pack and unpack a message in Go -// -// foo := &pb.Foo{...} -// any, err := ptypes.MarshalAny(foo) -// ... -// foo := &pb.Foo{} -// if err := ptypes.UnmarshalAny(any, foo); err != nil { -// ... -// } -// -// The pack methods provided by protobuf library will by default use -// 'type.googleapis.com/full.type.name' as the type URL and the unpack -// methods only use the fully qualified type name after the last '/' -// in the type URL, for example "foo.bar.com/x/y.z" will yield type -// name "y.z". -// -// -// JSON -// ==== -// The JSON representation of an `Any` value uses the regular -// representation of the deserialized, embedded message, with an -// additional field `@type` which contains the type URL. Example: -// -// package google.profile; -// message Person { -// string first_name = 1; -// string last_name = 2; -// } -// -// { -// "@type": "type.googleapis.com/google.profile.Person", -// "firstName": , -// "lastName": -// } -// -// If the embedded message type is well-known and has a custom JSON -// representation, that representation will be embedded adding a field -// `value` which holds the custom JSON in addition to the `@type` -// field. Example (for message [google.protobuf.Duration][]): -// -// { -// "@type": "type.googleapis.com/google.protobuf.Duration", -// "value": "1.212s" -// } -// -type Any struct { - // A URL/resource name that uniquely identifies the type of the serialized - // protocol buffer message. This string must contain at least - // one "/" character. The last segment of the URL's path must represent - // the fully qualified name of the type (as in - // `path/google.protobuf.Duration`). The name should be in a canonical form - // (e.g., leading "." is not accepted). - // - // In practice, teams usually precompile into the binary all types that they - // expect it to use in the context of Any. However, for URLs which use the - // scheme `http`, `https`, or no scheme, one can optionally set up a type - // server that maps type URLs to message definitions as follows: - // - // * If no scheme is provided, `https` is assumed. - // * An HTTP GET on the URL must yield a [google.protobuf.Type][] - // value in binary format, or produce an error. - // * Applications are allowed to cache lookup results based on the - // URL, or have them precompiled into a binary to avoid any - // lookup. Therefore, binary compatibility needs to be preserved - // on changes to types. (Use versioned type names to manage - // breaking changes.) - // - // Note: this functionality is not currently available in the official - // protobuf release, and it is not used for type URLs beginning with - // type.googleapis.com. - // - // Schemes other than `http`, `https` (or the empty scheme) might be - // used with implementation specific semantics. - // - TypeUrl string `protobuf:"bytes,1,opt,name=type_url,json=typeUrl,proto3" json:"type_url,omitempty"` - // Must be a valid serialized protocol buffer of the above specified type. - Value []byte `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Any) Reset() { *m = Any{} } -func (m *Any) String() string { return proto.CompactTextString(m) } -func (*Any) ProtoMessage() {} -func (*Any) Descriptor() ([]byte, []int) { - return fileDescriptor_b53526c13ae22eb4, []int{0} -} - -func (*Any) XXX_WellKnownType() string { return "Any" } +var File_github_com_golang_protobuf_ptypes_any_any_proto protoreflect.FileDescriptor -func (m *Any) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Any.Unmarshal(m, b) -} -func (m *Any) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Any.Marshal(b, m, deterministic) -} -func (m *Any) XXX_Merge(src proto.Message) { - xxx_messageInfo_Any.Merge(m, src) -} -func (m *Any) XXX_Size() int { - return xxx_messageInfo_Any.Size(m) -} -func (m *Any) XXX_DiscardUnknown() { - xxx_messageInfo_Any.DiscardUnknown(m) +var file_github_com_golang_protobuf_ptypes_any_any_proto_rawDesc = []byte{ + 0x0a, 0x2f, 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, + 0x61, 0x6e, 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, + 0x70, 0x65, 0x73, 0x2f, 0x61, 0x6e, 0x79, 0x2f, 0x61, 0x6e, 0x79, 0x2e, 0x70, 0x72, 0x6f, 0x74, + 0x6f, 0x1a, 0x19, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, + 0x75, 0x66, 0x2f, 0x61, 0x6e, 0x79, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x42, 0x2b, 0x5a, 0x29, + 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, 0x61, 0x6e, + 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, 0x70, 0x65, + 0x73, 0x2f, 0x61, 0x6e, 0x79, 0x3b, 0x61, 0x6e, 0x79, 0x50, 0x00, 0x62, 0x06, 0x70, 0x72, 0x6f, + 0x74, 0x6f, 0x33, } -var xxx_messageInfo_Any proto.InternalMessageInfo - -func (m *Any) GetTypeUrl() string { - if m != nil { - return m.TypeUrl - } - return "" +var file_github_com_golang_protobuf_ptypes_any_any_proto_goTypes = []interface{}{} +var file_github_com_golang_protobuf_ptypes_any_any_proto_depIdxs = []int32{ + 0, // [0:0] is the sub-list for method output_type + 0, // [0:0] is the sub-list for method input_type + 0, // [0:0] is the sub-list for extension type_name + 0, // [0:0] is the sub-list for extension extendee + 0, // [0:0] is the sub-list for field type_name } -func (m *Any) GetValue() []byte { - if m != nil { - return m.Value +func init() { file_github_com_golang_protobuf_ptypes_any_any_proto_init() } +func file_github_com_golang_protobuf_ptypes_any_any_proto_init() { + if File_github_com_golang_protobuf_ptypes_any_any_proto != nil { + return } - return nil -} - -func init() { - proto.RegisterType((*Any)(nil), "google.protobuf.Any") -} - -func init() { - proto.RegisterFile("google/protobuf/any.proto", fileDescriptor_b53526c13ae22eb4) -} - -var fileDescriptor_b53526c13ae22eb4 = []byte{ - // 185 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x92, 0x4c, 0xcf, 0xcf, 0x4f, - 0xcf, 0x49, 0xd5, 0x2f, 0x28, 0xca, 0x2f, 0xc9, 0x4f, 0x2a, 0x4d, 0xd3, 0x4f, 0xcc, 0xab, 0xd4, - 0x03, 0x73, 0x84, 0xf8, 0x21, 0x52, 0x7a, 0x30, 0x29, 0x25, 0x33, 0x2e, 0x66, 0xc7, 0xbc, 0x4a, - 0x21, 0x49, 0x2e, 0x8e, 0x92, 0xca, 0x82, 0xd4, 0xf8, 0xd2, 0xa2, 0x1c, 0x09, 0x46, 0x05, 0x46, - 0x0d, 0xce, 0x20, 0x76, 0x10, 0x3f, 0xb4, 0x28, 0x47, 0x48, 0x84, 0x8b, 0xb5, 0x2c, 0x31, 0xa7, - 0x34, 0x55, 0x82, 0x49, 0x81, 0x51, 0x83, 0x27, 0x08, 0xc2, 0x71, 0xca, 0xe7, 0x12, 0x4e, 0xce, - 0xcf, 0xd5, 0x43, 0x33, 0xce, 0x89, 0xc3, 0x31, 0xaf, 0x32, 0x00, 0xc4, 0x09, 0x60, 0x8c, 0x52, - 0x4d, 0xcf, 0x2c, 0xc9, 0x28, 0x4d, 0xd2, 0x4b, 0xce, 0xcf, 0xd5, 0x4f, 0xcf, 0xcf, 0x49, 0xcc, - 0x4b, 0x47, 0xb8, 0xa8, 0x00, 0x64, 0x7a, 0x31, 0xc8, 0x61, 0x8b, 0x98, 0x98, 0xdd, 0x03, 0x9c, - 0x56, 0x31, 0xc9, 0xb9, 0x43, 0x8c, 0x0a, 0x80, 0x2a, 0xd1, 0x0b, 0x4f, 0xcd, 0xc9, 0xf1, 0xce, - 0xcb, 0x2f, 0xcf, 0x0b, 0x01, 0x29, 0x4d, 0x62, 0x03, 0xeb, 0x35, 0x06, 0x04, 0x00, 0x00, 0xff, - 0xff, 0x13, 0xf8, 0xe8, 0x42, 0xdd, 0x00, 0x00, 0x00, + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_github_com_golang_protobuf_ptypes_any_any_proto_rawDesc, + NumEnums: 0, + NumMessages: 0, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_github_com_golang_protobuf_ptypes_any_any_proto_goTypes, + DependencyIndexes: file_github_com_golang_protobuf_ptypes_any_any_proto_depIdxs, + }.Build() + File_github_com_golang_protobuf_ptypes_any_any_proto = out.File + file_github_com_golang_protobuf_ptypes_any_any_proto_rawDesc = nil + file_github_com_golang_protobuf_ptypes_any_any_proto_goTypes = nil + file_github_com_golang_protobuf_ptypes_any_any_proto_depIdxs = nil } diff --git a/vendor/github.com/golang/protobuf/ptypes/any/any.proto b/vendor/github.com/golang/protobuf/ptypes/any/any.proto deleted file mode 100644 index c9be8541..00000000 --- a/vendor/github.com/golang/protobuf/ptypes/any/any.proto +++ /dev/null @@ -1,155 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -syntax = "proto3"; - -package google.protobuf; - -option csharp_namespace = "Google.Protobuf.WellKnownTypes"; -option go_package = "github.com/golang/protobuf/ptypes/any"; -option java_package = "com.google.protobuf"; -option java_outer_classname = "AnyProto"; -option java_multiple_files = true; -option objc_class_prefix = "GPB"; - -// `Any` contains an arbitrary serialized protocol buffer message along with a -// URL that describes the type of the serialized message. -// -// Protobuf library provides support to pack/unpack Any values in the form -// of utility functions or additional generated methods of the Any type. -// -// Example 1: Pack and unpack a message in C++. -// -// Foo foo = ...; -// Any any; -// any.PackFrom(foo); -// ... -// if (any.UnpackTo(&foo)) { -// ... -// } -// -// Example 2: Pack and unpack a message in Java. -// -// Foo foo = ...; -// Any any = Any.pack(foo); -// ... -// if (any.is(Foo.class)) { -// foo = any.unpack(Foo.class); -// } -// -// Example 3: Pack and unpack a message in Python. -// -// foo = Foo(...) -// any = Any() -// any.Pack(foo) -// ... -// if any.Is(Foo.DESCRIPTOR): -// any.Unpack(foo) -// ... -// -// Example 4: Pack and unpack a message in Go -// -// foo := &pb.Foo{...} -// any, err := ptypes.MarshalAny(foo) -// ... -// foo := &pb.Foo{} -// if err := ptypes.UnmarshalAny(any, foo); err != nil { -// ... -// } -// -// The pack methods provided by protobuf library will by default use -// 'type.googleapis.com/full.type.name' as the type URL and the unpack -// methods only use the fully qualified type name after the last '/' -// in the type URL, for example "foo.bar.com/x/y.z" will yield type -// name "y.z". -// -// -// JSON -// ==== -// The JSON representation of an `Any` value uses the regular -// representation of the deserialized, embedded message, with an -// additional field `@type` which contains the type URL. Example: -// -// package google.profile; -// message Person { -// string first_name = 1; -// string last_name = 2; -// } -// -// { -// "@type": "type.googleapis.com/google.profile.Person", -// "firstName": , -// "lastName": -// } -// -// If the embedded message type is well-known and has a custom JSON -// representation, that representation will be embedded adding a field -// `value` which holds the custom JSON in addition to the `@type` -// field. Example (for message [google.protobuf.Duration][]): -// -// { -// "@type": "type.googleapis.com/google.protobuf.Duration", -// "value": "1.212s" -// } -// -message Any { - // A URL/resource name that uniquely identifies the type of the serialized - // protocol buffer message. This string must contain at least - // one "/" character. The last segment of the URL's path must represent - // the fully qualified name of the type (as in - // `path/google.protobuf.Duration`). The name should be in a canonical form - // (e.g., leading "." is not accepted). - // - // In practice, teams usually precompile into the binary all types that they - // expect it to use in the context of Any. However, for URLs which use the - // scheme `http`, `https`, or no scheme, one can optionally set up a type - // server that maps type URLs to message definitions as follows: - // - // * If no scheme is provided, `https` is assumed. - // * An HTTP GET on the URL must yield a [google.protobuf.Type][] - // value in binary format, or produce an error. - // * Applications are allowed to cache lookup results based on the - // URL, or have them precompiled into a binary to avoid any - // lookup. Therefore, binary compatibility needs to be preserved - // on changes to types. (Use versioned type names to manage - // breaking changes.) - // - // Note: this functionality is not currently available in the official - // protobuf release, and it is not used for type URLs beginning with - // type.googleapis.com. - // - // Schemes other than `http`, `https` (or the empty scheme) might be - // used with implementation specific semantics. - // - string type_url = 1; - - // Must be a valid serialized protocol buffer of the above specified type. - bytes value = 2; -} diff --git a/vendor/github.com/golang/protobuf/ptypes/doc.go b/vendor/github.com/golang/protobuf/ptypes/doc.go index c0d595da..fb9edd5c 100644 --- a/vendor/github.com/golang/protobuf/ptypes/doc.go +++ b/vendor/github.com/golang/protobuf/ptypes/doc.go @@ -1,35 +1,6 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2016 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2016 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. -/* -Package ptypes contains code for interacting with well-known types. -*/ +// Package ptypes provides functionality for interacting with well-known types. package ptypes diff --git a/vendor/github.com/golang/protobuf/ptypes/duration.go b/vendor/github.com/golang/protobuf/ptypes/duration.go index 26d1ca2f..6110ae8a 100644 --- a/vendor/github.com/golang/protobuf/ptypes/duration.go +++ b/vendor/github.com/golang/protobuf/ptypes/duration.go @@ -1,102 +1,72 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2016 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2016 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. package ptypes -// This file implements conversions between google.protobuf.Duration -// and time.Duration. - import ( "errors" "fmt" "time" - durpb "github.com/golang/protobuf/ptypes/duration" + durationpb "github.com/golang/protobuf/ptypes/duration" ) +// Range of google.protobuf.Duration as specified in duration.proto. +// This is about 10,000 years in seconds. const ( - // Range of a durpb.Duration in seconds, as specified in - // google/protobuf/duration.proto. This is about 10,000 years in seconds. maxSeconds = int64(10000 * 365.25 * 24 * 60 * 60) minSeconds = -maxSeconds ) -// validateDuration determines whether the durpb.Duration is valid according to the -// definition in google/protobuf/duration.proto. A valid durpb.Duration -// may still be too large to fit into a time.Duration (the range of durpb.Duration -// is about 10,000 years, and the range of time.Duration is about 290). -func validateDuration(d *durpb.Duration) error { - if d == nil { - return errors.New("duration: nil Duration") - } - if d.Seconds < minSeconds || d.Seconds > maxSeconds { - return fmt.Errorf("duration: %v: seconds out of range", d) - } - if d.Nanos <= -1e9 || d.Nanos >= 1e9 { - return fmt.Errorf("duration: %v: nanos out of range", d) - } - // Seconds and Nanos must have the same sign, unless d.Nanos is zero. - if (d.Seconds < 0 && d.Nanos > 0) || (d.Seconds > 0 && d.Nanos < 0) { - return fmt.Errorf("duration: %v: seconds and nanos have different signs", d) - } - return nil -} - -// Duration converts a durpb.Duration to a time.Duration. Duration -// returns an error if the durpb.Duration is invalid or is too large to be -// represented in a time.Duration. -func Duration(p *durpb.Duration) (time.Duration, error) { - if err := validateDuration(p); err != nil { +// Duration converts a durationpb.Duration to a time.Duration. +// Duration returns an error if dur is invalid or overflows a time.Duration. +func Duration(dur *durationpb.Duration) (time.Duration, error) { + if err := validateDuration(dur); err != nil { return 0, err } - d := time.Duration(p.Seconds) * time.Second - if int64(d/time.Second) != p.Seconds { - return 0, fmt.Errorf("duration: %v is out of range for time.Duration", p) + d := time.Duration(dur.Seconds) * time.Second + if int64(d/time.Second) != dur.Seconds { + return 0, fmt.Errorf("duration: %v is out of range for time.Duration", dur) } - if p.Nanos != 0 { - d += time.Duration(p.Nanos) * time.Nanosecond - if (d < 0) != (p.Nanos < 0) { - return 0, fmt.Errorf("duration: %v is out of range for time.Duration", p) + if dur.Nanos != 0 { + d += time.Duration(dur.Nanos) * time.Nanosecond + if (d < 0) != (dur.Nanos < 0) { + return 0, fmt.Errorf("duration: %v is out of range for time.Duration", dur) } } return d, nil } -// DurationProto converts a time.Duration to a durpb.Duration. -func DurationProto(d time.Duration) *durpb.Duration { +// DurationProto converts a time.Duration to a durationpb.Duration. +func DurationProto(d time.Duration) *durationpb.Duration { nanos := d.Nanoseconds() secs := nanos / 1e9 nanos -= secs * 1e9 - return &durpb.Duration{ - Seconds: secs, + return &durationpb.Duration{ + Seconds: int64(secs), Nanos: int32(nanos), } } + +// validateDuration determines whether the durationpb.Duration is valid +// according to the definition in google/protobuf/duration.proto. +// A valid durpb.Duration may still be too large to fit into a time.Duration +// Note that the range of durationpb.Duration is about 10,000 years, +// while the range of time.Duration is about 290 years. +func validateDuration(dur *durationpb.Duration) error { + if dur == nil { + return errors.New("duration: nil Duration") + } + if dur.Seconds < minSeconds || dur.Seconds > maxSeconds { + return fmt.Errorf("duration: %v: seconds out of range", dur) + } + if dur.Nanos <= -1e9 || dur.Nanos >= 1e9 { + return fmt.Errorf("duration: %v: nanos out of range", dur) + } + // Seconds and Nanos must have the same sign, unless d.Nanos is zero. + if (dur.Seconds < 0 && dur.Nanos > 0) || (dur.Seconds > 0 && dur.Nanos < 0) { + return fmt.Errorf("duration: %v: seconds and nanos have different signs", dur) + } + return nil +} diff --git a/vendor/github.com/golang/protobuf/ptypes/duration/duration.pb.go b/vendor/github.com/golang/protobuf/ptypes/duration/duration.pb.go index 58b07869..d0079ee3 100644 --- a/vendor/github.com/golang/protobuf/ptypes/duration/duration.pb.go +++ b/vendor/github.com/golang/protobuf/ptypes/duration/duration.pb.go @@ -1,163 +1,63 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: google/protobuf/duration.proto +// source: github.com/golang/protobuf/ptypes/duration/duration.proto package duration import ( - fmt "fmt" - proto "github.com/golang/protobuf/proto" - math "math" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + durationpb "google.golang.org/protobuf/types/known/durationpb" + reflect "reflect" ) -// Reference imports to suppress errors if they are not otherwise used. -var _ = proto.Marshal -var _ = fmt.Errorf -var _ = math.Inf +// Symbols defined in public import of google/protobuf/duration.proto. -// This is a compile-time assertion to ensure that this generated file -// is compatible with the proto package it is being compiled against. -// A compilation error at this line likely means your copy of the -// proto package needs to be updated. -const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package +type Duration = durationpb.Duration -// A Duration represents a signed, fixed-length span of time represented -// as a count of seconds and fractions of seconds at nanosecond -// resolution. It is independent of any calendar and concepts like "day" -// or "month". It is related to Timestamp in that the difference between -// two Timestamp values is a Duration and it can be added or subtracted -// from a Timestamp. Range is approximately +-10,000 years. -// -// # Examples -// -// Example 1: Compute Duration from two Timestamps in pseudo code. -// -// Timestamp start = ...; -// Timestamp end = ...; -// Duration duration = ...; -// -// duration.seconds = end.seconds - start.seconds; -// duration.nanos = end.nanos - start.nanos; -// -// if (duration.seconds < 0 && duration.nanos > 0) { -// duration.seconds += 1; -// duration.nanos -= 1000000000; -// } else if (duration.seconds > 0 && duration.nanos < 0) { -// duration.seconds -= 1; -// duration.nanos += 1000000000; -// } -// -// Example 2: Compute Timestamp from Timestamp + Duration in pseudo code. -// -// Timestamp start = ...; -// Duration duration = ...; -// Timestamp end = ...; -// -// end.seconds = start.seconds + duration.seconds; -// end.nanos = start.nanos + duration.nanos; -// -// if (end.nanos < 0) { -// end.seconds -= 1; -// end.nanos += 1000000000; -// } else if (end.nanos >= 1000000000) { -// end.seconds += 1; -// end.nanos -= 1000000000; -// } -// -// Example 3: Compute Duration from datetime.timedelta in Python. -// -// td = datetime.timedelta(days=3, minutes=10) -// duration = Duration() -// duration.FromTimedelta(td) -// -// # JSON Mapping -// -// In JSON format, the Duration type is encoded as a string rather than an -// object, where the string ends in the suffix "s" (indicating seconds) and -// is preceded by the number of seconds, with nanoseconds expressed as -// fractional seconds. For example, 3 seconds with 0 nanoseconds should be -// encoded in JSON format as "3s", while 3 seconds and 1 nanosecond should -// be expressed in JSON format as "3.000000001s", and 3 seconds and 1 -// microsecond should be expressed in JSON format as "3.000001s". -// -// -type Duration struct { - // Signed seconds of the span of time. Must be from -315,576,000,000 - // to +315,576,000,000 inclusive. Note: these bounds are computed from: - // 60 sec/min * 60 min/hr * 24 hr/day * 365.25 days/year * 10000 years - Seconds int64 `protobuf:"varint,1,opt,name=seconds,proto3" json:"seconds,omitempty"` - // Signed fractions of a second at nanosecond resolution of the span - // of time. Durations less than one second are represented with a 0 - // `seconds` field and a positive or negative `nanos` field. For durations - // of one second or more, a non-zero value for the `nanos` field must be - // of the same sign as the `seconds` field. Must be from -999,999,999 - // to +999,999,999 inclusive. - Nanos int32 `protobuf:"varint,2,opt,name=nanos,proto3" json:"nanos,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Duration) Reset() { *m = Duration{} } -func (m *Duration) String() string { return proto.CompactTextString(m) } -func (*Duration) ProtoMessage() {} -func (*Duration) Descriptor() ([]byte, []int) { - return fileDescriptor_23597b2ebd7ac6c5, []int{0} -} - -func (*Duration) XXX_WellKnownType() string { return "Duration" } +var File_github_com_golang_protobuf_ptypes_duration_duration_proto protoreflect.FileDescriptor -func (m *Duration) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Duration.Unmarshal(m, b) -} -func (m *Duration) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Duration.Marshal(b, m, deterministic) -} -func (m *Duration) XXX_Merge(src proto.Message) { - xxx_messageInfo_Duration.Merge(m, src) -} -func (m *Duration) XXX_Size() int { - return xxx_messageInfo_Duration.Size(m) -} -func (m *Duration) XXX_DiscardUnknown() { - xxx_messageInfo_Duration.DiscardUnknown(m) +var file_github_com_golang_protobuf_ptypes_duration_duration_proto_rawDesc = []byte{ + 0x0a, 0x39, 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, + 0x61, 0x6e, 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, + 0x70, 0x65, 0x73, 0x2f, 0x64, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x2f, 0x64, 0x75, 0x72, + 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x1a, 0x1e, 0x67, 0x6f, 0x6f, + 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x64, 0x75, 0x72, + 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x42, 0x35, 0x5a, 0x33, 0x67, + 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, + 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, 0x70, 0x65, 0x73, + 0x2f, 0x64, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x3b, 0x64, 0x75, 0x72, 0x61, 0x74, 0x69, + 0x6f, 0x6e, 0x50, 0x00, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33, } -var xxx_messageInfo_Duration proto.InternalMessageInfo - -func (m *Duration) GetSeconds() int64 { - if m != nil { - return m.Seconds - } - return 0 +var file_github_com_golang_protobuf_ptypes_duration_duration_proto_goTypes = []interface{}{} +var file_github_com_golang_protobuf_ptypes_duration_duration_proto_depIdxs = []int32{ + 0, // [0:0] is the sub-list for method output_type + 0, // [0:0] is the sub-list for method input_type + 0, // [0:0] is the sub-list for extension type_name + 0, // [0:0] is the sub-list for extension extendee + 0, // [0:0] is the sub-list for field type_name } -func (m *Duration) GetNanos() int32 { - if m != nil { - return m.Nanos +func init() { file_github_com_golang_protobuf_ptypes_duration_duration_proto_init() } +func file_github_com_golang_protobuf_ptypes_duration_duration_proto_init() { + if File_github_com_golang_protobuf_ptypes_duration_duration_proto != nil { + return } - return 0 -} - -func init() { - proto.RegisterType((*Duration)(nil), "google.protobuf.Duration") -} - -func init() { - proto.RegisterFile("google/protobuf/duration.proto", fileDescriptor_23597b2ebd7ac6c5) -} - -var fileDescriptor_23597b2ebd7ac6c5 = []byte{ - // 190 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x92, 0x4b, 0xcf, 0xcf, 0x4f, - 0xcf, 0x49, 0xd5, 0x2f, 0x28, 0xca, 0x2f, 0xc9, 0x4f, 0x2a, 0x4d, 0xd3, 0x4f, 0x29, 0x2d, 0x4a, - 0x2c, 0xc9, 0xcc, 0xcf, 0xd3, 0x03, 0x8b, 0x08, 0xf1, 0x43, 0xe4, 0xf5, 0x60, 0xf2, 0x4a, 0x56, - 0x5c, 0x1c, 0x2e, 0x50, 0x25, 0x42, 0x12, 0x5c, 0xec, 0xc5, 0xa9, 0xc9, 0xf9, 0x79, 0x29, 0xc5, - 0x12, 0x8c, 0x0a, 0x8c, 0x1a, 0xcc, 0x41, 0x30, 0xae, 0x90, 0x08, 0x17, 0x6b, 0x5e, 0x62, 0x5e, - 0x7e, 0xb1, 0x04, 0x93, 0x02, 0xa3, 0x06, 0x6b, 0x10, 0x84, 0xe3, 0x54, 0xc3, 0x25, 0x9c, 0x9c, - 0x9f, 0xab, 0x87, 0x66, 0xa4, 0x13, 0x2f, 0xcc, 0xc0, 0x00, 0x90, 0x48, 0x00, 0x63, 0x94, 0x56, - 0x7a, 0x66, 0x49, 0x46, 0x69, 0x92, 0x5e, 0x72, 0x7e, 0xae, 0x7e, 0x7a, 0x7e, 0x4e, 0x62, 0x5e, - 0x3a, 0xc2, 0x7d, 0x05, 0x25, 0x95, 0x05, 0xa9, 0xc5, 0x70, 0x67, 0xfe, 0x60, 0x64, 0x5c, 0xc4, - 0xc4, 0xec, 0x1e, 0xe0, 0xb4, 0x8a, 0x49, 0xce, 0x1d, 0x62, 0x6e, 0x00, 0x54, 0xa9, 0x5e, 0x78, - 0x6a, 0x4e, 0x8e, 0x77, 0x5e, 0x7e, 0x79, 0x5e, 0x08, 0x48, 0x4b, 0x12, 0x1b, 0xd8, 0x0c, 0x63, - 0x40, 0x00, 0x00, 0x00, 0xff, 0xff, 0xdc, 0x84, 0x30, 0xff, 0xf3, 0x00, 0x00, 0x00, + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_github_com_golang_protobuf_ptypes_duration_duration_proto_rawDesc, + NumEnums: 0, + NumMessages: 0, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_github_com_golang_protobuf_ptypes_duration_duration_proto_goTypes, + DependencyIndexes: file_github_com_golang_protobuf_ptypes_duration_duration_proto_depIdxs, + }.Build() + File_github_com_golang_protobuf_ptypes_duration_duration_proto = out.File + file_github_com_golang_protobuf_ptypes_duration_duration_proto_rawDesc = nil + file_github_com_golang_protobuf_ptypes_duration_duration_proto_goTypes = nil + file_github_com_golang_protobuf_ptypes_duration_duration_proto_depIdxs = nil } diff --git a/vendor/github.com/golang/protobuf/ptypes/duration/duration.proto b/vendor/github.com/golang/protobuf/ptypes/duration/duration.proto deleted file mode 100644 index 99cb102c..00000000 --- a/vendor/github.com/golang/protobuf/ptypes/duration/duration.proto +++ /dev/null @@ -1,116 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -syntax = "proto3"; - -package google.protobuf; - -option csharp_namespace = "Google.Protobuf.WellKnownTypes"; -option cc_enable_arenas = true; -option go_package = "github.com/golang/protobuf/ptypes/duration"; -option java_package = "com.google.protobuf"; -option java_outer_classname = "DurationProto"; -option java_multiple_files = true; -option objc_class_prefix = "GPB"; - -// A Duration represents a signed, fixed-length span of time represented -// as a count of seconds and fractions of seconds at nanosecond -// resolution. It is independent of any calendar and concepts like "day" -// or "month". It is related to Timestamp in that the difference between -// two Timestamp values is a Duration and it can be added or subtracted -// from a Timestamp. Range is approximately +-10,000 years. -// -// # Examples -// -// Example 1: Compute Duration from two Timestamps in pseudo code. -// -// Timestamp start = ...; -// Timestamp end = ...; -// Duration duration = ...; -// -// duration.seconds = end.seconds - start.seconds; -// duration.nanos = end.nanos - start.nanos; -// -// if (duration.seconds < 0 && duration.nanos > 0) { -// duration.seconds += 1; -// duration.nanos -= 1000000000; -// } else if (duration.seconds > 0 && duration.nanos < 0) { -// duration.seconds -= 1; -// duration.nanos += 1000000000; -// } -// -// Example 2: Compute Timestamp from Timestamp + Duration in pseudo code. -// -// Timestamp start = ...; -// Duration duration = ...; -// Timestamp end = ...; -// -// end.seconds = start.seconds + duration.seconds; -// end.nanos = start.nanos + duration.nanos; -// -// if (end.nanos < 0) { -// end.seconds -= 1; -// end.nanos += 1000000000; -// } else if (end.nanos >= 1000000000) { -// end.seconds += 1; -// end.nanos -= 1000000000; -// } -// -// Example 3: Compute Duration from datetime.timedelta in Python. -// -// td = datetime.timedelta(days=3, minutes=10) -// duration = Duration() -// duration.FromTimedelta(td) -// -// # JSON Mapping -// -// In JSON format, the Duration type is encoded as a string rather than an -// object, where the string ends in the suffix "s" (indicating seconds) and -// is preceded by the number of seconds, with nanoseconds expressed as -// fractional seconds. For example, 3 seconds with 0 nanoseconds should be -// encoded in JSON format as "3s", while 3 seconds and 1 nanosecond should -// be expressed in JSON format as "3.000000001s", and 3 seconds and 1 -// microsecond should be expressed in JSON format as "3.000001s". -// -// -message Duration { - // Signed seconds of the span of time. Must be from -315,576,000,000 - // to +315,576,000,000 inclusive. Note: these bounds are computed from: - // 60 sec/min * 60 min/hr * 24 hr/day * 365.25 days/year * 10000 years - int64 seconds = 1; - - // Signed fractions of a second at nanosecond resolution of the span - // of time. Durations less than one second are represented with a 0 - // `seconds` field and a positive or negative `nanos` field. For durations - // of one second or more, a non-zero value for the `nanos` field must be - // of the same sign as the `seconds` field. Must be from -999,999,999 - // to +999,999,999 inclusive. - int32 nanos = 2; -} diff --git a/vendor/github.com/golang/protobuf/ptypes/timestamp.go b/vendor/github.com/golang/protobuf/ptypes/timestamp.go index 8da0df01..026d0d49 100644 --- a/vendor/github.com/golang/protobuf/ptypes/timestamp.go +++ b/vendor/github.com/golang/protobuf/ptypes/timestamp.go @@ -1,46 +1,18 @@ -// Go support for Protocol Buffers - Google's data interchange format -// -// Copyright 2016 The Go Authors. All rights reserved. -// https://github.com/golang/protobuf -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +// Copyright 2016 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. package ptypes -// This file implements operations on google.protobuf.Timestamp. - import ( "errors" "fmt" "time" - tspb "github.com/golang/protobuf/ptypes/timestamp" + timestamppb "github.com/golang/protobuf/ptypes/timestamp" ) +// Range of google.protobuf.Duration as specified in timestamp.proto. const ( // Seconds field of the earliest valid Timestamp. // This is time.Date(1, 1, 1, 0, 0, 0, 0, time.UTC).Unix(). @@ -50,44 +22,18 @@ const ( maxValidSeconds = 253402300800 ) -// validateTimestamp determines whether a Timestamp is valid. -// A valid timestamp represents a time in the range -// [0001-01-01, 10000-01-01) and has a Nanos field -// in the range [0, 1e9). -// -// If the Timestamp is valid, validateTimestamp returns nil. -// Otherwise, it returns an error that describes -// the problem. -// -// Every valid Timestamp can be represented by a time.Time, but the converse is not true. -func validateTimestamp(ts *tspb.Timestamp) error { - if ts == nil { - return errors.New("timestamp: nil Timestamp") - } - if ts.Seconds < minValidSeconds { - return fmt.Errorf("timestamp: %v before 0001-01-01", ts) - } - if ts.Seconds >= maxValidSeconds { - return fmt.Errorf("timestamp: %v after 10000-01-01", ts) - } - if ts.Nanos < 0 || ts.Nanos >= 1e9 { - return fmt.Errorf("timestamp: %v: nanos not in range [0, 1e9)", ts) - } - return nil -} - -// Timestamp converts a google.protobuf.Timestamp proto to a time.Time. +// Timestamp converts a timestamppb.Timestamp to a time.Time. // It returns an error if the argument is invalid. // -// Unlike most Go functions, if Timestamp returns an error, the first return value -// is not the zero time.Time. Instead, it is the value obtained from the +// Unlike most Go functions, if Timestamp returns an error, the first return +// value is not the zero time.Time. Instead, it is the value obtained from the // time.Unix function when passed the contents of the Timestamp, in the UTC // locale. This may or may not be a meaningful time; many invalid Timestamps // do map to valid time.Times. // // A nil Timestamp returns an error. The first return value in that case is // undefined. -func Timestamp(ts *tspb.Timestamp) (time.Time, error) { +func Timestamp(ts *timestamppb.Timestamp) (time.Time, error) { // Don't return the zero value on error, because corresponds to a valid // timestamp. Instead return whatever time.Unix gives us. var t time.Time @@ -100,7 +46,7 @@ func Timestamp(ts *tspb.Timestamp) (time.Time, error) { } // TimestampNow returns a google.protobuf.Timestamp for the current time. -func TimestampNow() *tspb.Timestamp { +func TimestampNow() *timestamppb.Timestamp { ts, err := TimestampProto(time.Now()) if err != nil { panic("ptypes: time.Now() out of Timestamp range") @@ -110,8 +56,8 @@ func TimestampNow() *tspb.Timestamp { // TimestampProto converts the time.Time to a google.protobuf.Timestamp proto. // It returns an error if the resulting Timestamp is invalid. -func TimestampProto(t time.Time) (*tspb.Timestamp, error) { - ts := &tspb.Timestamp{ +func TimestampProto(t time.Time) (*timestamppb.Timestamp, error) { + ts := ×tamppb.Timestamp{ Seconds: t.Unix(), Nanos: int32(t.Nanosecond()), } @@ -121,12 +67,37 @@ func TimestampProto(t time.Time) (*tspb.Timestamp, error) { return ts, nil } -// TimestampString returns the RFC 3339 string for valid Timestamps. For invalid -// Timestamps, it returns an error message in parentheses. -func TimestampString(ts *tspb.Timestamp) string { +// TimestampString returns the RFC 3339 string for valid Timestamps. +// For invalid Timestamps, it returns an error message in parentheses. +func TimestampString(ts *timestamppb.Timestamp) string { t, err := Timestamp(ts) if err != nil { return fmt.Sprintf("(%v)", err) } return t.Format(time.RFC3339Nano) } + +// validateTimestamp determines whether a Timestamp is valid. +// A valid timestamp represents a time in the range [0001-01-01, 10000-01-01) +// and has a Nanos field in the range [0, 1e9). +// +// If the Timestamp is valid, validateTimestamp returns nil. +// Otherwise, it returns an error that describes the problem. +// +// Every valid Timestamp can be represented by a time.Time, +// but the converse is not true. +func validateTimestamp(ts *timestamppb.Timestamp) error { + if ts == nil { + return errors.New("timestamp: nil Timestamp") + } + if ts.Seconds < minValidSeconds { + return fmt.Errorf("timestamp: %v before 0001-01-01", ts) + } + if ts.Seconds >= maxValidSeconds { + return fmt.Errorf("timestamp: %v after 10000-01-01", ts) + } + if ts.Nanos < 0 || ts.Nanos >= 1e9 { + return fmt.Errorf("timestamp: %v: nanos not in range [0, 1e9)", ts) + } + return nil +} diff --git a/vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.pb.go b/vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.pb.go index 7a3b1e40..a76f8076 100644 --- a/vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.pb.go +++ b/vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.pb.go @@ -1,185 +1,64 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: google/protobuf/timestamp.proto +// source: github.com/golang/protobuf/ptypes/timestamp/timestamp.proto package timestamp import ( - fmt "fmt" - proto "github.com/golang/protobuf/proto" - math "math" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + timestamppb "google.golang.org/protobuf/types/known/timestamppb" + reflect "reflect" ) -// Reference imports to suppress errors if they are not otherwise used. -var _ = proto.Marshal -var _ = fmt.Errorf -var _ = math.Inf +// Symbols defined in public import of google/protobuf/timestamp.proto. -// This is a compile-time assertion to ensure that this generated file -// is compatible with the proto package it is being compiled against. -// A compilation error at this line likely means your copy of the -// proto package needs to be updated. -const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package +type Timestamp = timestamppb.Timestamp -// A Timestamp represents a point in time independent of any time zone or local -// calendar, encoded as a count of seconds and fractions of seconds at -// nanosecond resolution. The count is relative to an epoch at UTC midnight on -// January 1, 1970, in the proleptic Gregorian calendar which extends the -// Gregorian calendar backwards to year one. -// -// All minutes are 60 seconds long. Leap seconds are "smeared" so that no leap -// second table is needed for interpretation, using a [24-hour linear -// smear](https://developers.google.com/time/smear). -// -// The range is from 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z. By -// restricting to that range, we ensure that we can convert to and from [RFC -// 3339](https://www.ietf.org/rfc/rfc3339.txt) date strings. -// -// # Examples -// -// Example 1: Compute Timestamp from POSIX `time()`. -// -// Timestamp timestamp; -// timestamp.set_seconds(time(NULL)); -// timestamp.set_nanos(0); -// -// Example 2: Compute Timestamp from POSIX `gettimeofday()`. -// -// struct timeval tv; -// gettimeofday(&tv, NULL); -// -// Timestamp timestamp; -// timestamp.set_seconds(tv.tv_sec); -// timestamp.set_nanos(tv.tv_usec * 1000); -// -// Example 3: Compute Timestamp from Win32 `GetSystemTimeAsFileTime()`. -// -// FILETIME ft; -// GetSystemTimeAsFileTime(&ft); -// UINT64 ticks = (((UINT64)ft.dwHighDateTime) << 32) | ft.dwLowDateTime; -// -// // A Windows tick is 100 nanoseconds. Windows epoch 1601-01-01T00:00:00Z -// // is 11644473600 seconds before Unix epoch 1970-01-01T00:00:00Z. -// Timestamp timestamp; -// timestamp.set_seconds((INT64) ((ticks / 10000000) - 11644473600LL)); -// timestamp.set_nanos((INT32) ((ticks % 10000000) * 100)); -// -// Example 4: Compute Timestamp from Java `System.currentTimeMillis()`. -// -// long millis = System.currentTimeMillis(); -// -// Timestamp timestamp = Timestamp.newBuilder().setSeconds(millis / 1000) -// .setNanos((int) ((millis % 1000) * 1000000)).build(); -// -// -// Example 5: Compute Timestamp from current time in Python. -// -// timestamp = Timestamp() -// timestamp.GetCurrentTime() -// -// # JSON Mapping -// -// In JSON format, the Timestamp type is encoded as a string in the -// [RFC 3339](https://www.ietf.org/rfc/rfc3339.txt) format. That is, the -// format is "{year}-{month}-{day}T{hour}:{min}:{sec}[.{frac_sec}]Z" -// where {year} is always expressed using four digits while {month}, {day}, -// {hour}, {min}, and {sec} are zero-padded to two digits each. The fractional -// seconds, which can go up to 9 digits (i.e. up to 1 nanosecond resolution), -// are optional. The "Z" suffix indicates the timezone ("UTC"); the timezone -// is required. A proto3 JSON serializer should always use UTC (as indicated by -// "Z") when printing the Timestamp type and a proto3 JSON parser should be -// able to accept both UTC and other timezones (as indicated by an offset). -// -// For example, "2017-01-15T01:30:15.01Z" encodes 15.01 seconds past -// 01:30 UTC on January 15, 2017. -// -// In JavaScript, one can convert a Date object to this format using the -// standard -// [toISOString()](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Date/toISOString) -// method. In Python, a standard `datetime.datetime` object can be converted -// to this format using -// [`strftime`](https://docs.python.org/2/library/time.html#time.strftime) with -// the time format spec '%Y-%m-%dT%H:%M:%S.%fZ'. Likewise, in Java, one can use -// the Joda Time's [`ISODateTimeFormat.dateTime()`]( -// http://www.joda.org/joda-time/apidocs/org/joda/time/format/ISODateTimeFormat.html#dateTime%2D%2D -// ) to obtain a formatter capable of generating timestamps in this format. -// -// -type Timestamp struct { - // Represents seconds of UTC time since Unix epoch - // 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to - // 9999-12-31T23:59:59Z inclusive. - Seconds int64 `protobuf:"varint,1,opt,name=seconds,proto3" json:"seconds,omitempty"` - // Non-negative fractions of a second at nanosecond resolution. Negative - // second values with fractions must still have non-negative nanos values - // that count forward in time. Must be from 0 to 999,999,999 - // inclusive. - Nanos int32 `protobuf:"varint,2,opt,name=nanos,proto3" json:"nanos,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Timestamp) Reset() { *m = Timestamp{} } -func (m *Timestamp) String() string { return proto.CompactTextString(m) } -func (*Timestamp) ProtoMessage() {} -func (*Timestamp) Descriptor() ([]byte, []int) { - return fileDescriptor_292007bbfe81227e, []int{0} -} - -func (*Timestamp) XXX_WellKnownType() string { return "Timestamp" } +var File_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto protoreflect.FileDescriptor -func (m *Timestamp) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Timestamp.Unmarshal(m, b) -} -func (m *Timestamp) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Timestamp.Marshal(b, m, deterministic) -} -func (m *Timestamp) XXX_Merge(src proto.Message) { - xxx_messageInfo_Timestamp.Merge(m, src) -} -func (m *Timestamp) XXX_Size() int { - return xxx_messageInfo_Timestamp.Size(m) -} -func (m *Timestamp) XXX_DiscardUnknown() { - xxx_messageInfo_Timestamp.DiscardUnknown(m) +var file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_rawDesc = []byte{ + 0x0a, 0x3b, 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, + 0x61, 0x6e, 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, + 0x70, 0x65, 0x73, 0x2f, 0x74, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x2f, 0x74, 0x69, + 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x1a, 0x1f, 0x67, + 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x74, + 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x42, 0x37, + 0x5a, 0x35, 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, + 0x61, 0x6e, 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, + 0x70, 0x65, 0x73, 0x2f, 0x74, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x3b, 0x74, 0x69, + 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x50, 0x00, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, + 0x33, } -var xxx_messageInfo_Timestamp proto.InternalMessageInfo - -func (m *Timestamp) GetSeconds() int64 { - if m != nil { - return m.Seconds - } - return 0 +var file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_goTypes = []interface{}{} +var file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_depIdxs = []int32{ + 0, // [0:0] is the sub-list for method output_type + 0, // [0:0] is the sub-list for method input_type + 0, // [0:0] is the sub-list for extension type_name + 0, // [0:0] is the sub-list for extension extendee + 0, // [0:0] is the sub-list for field type_name } -func (m *Timestamp) GetNanos() int32 { - if m != nil { - return m.Nanos +func init() { file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_init() } +func file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_init() { + if File_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto != nil { + return } - return 0 -} - -func init() { - proto.RegisterType((*Timestamp)(nil), "google.protobuf.Timestamp") -} - -func init() { - proto.RegisterFile("google/protobuf/timestamp.proto", fileDescriptor_292007bbfe81227e) -} - -var fileDescriptor_292007bbfe81227e = []byte{ - // 191 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x92, 0x4f, 0xcf, 0xcf, 0x4f, - 0xcf, 0x49, 0xd5, 0x2f, 0x28, 0xca, 0x2f, 0xc9, 0x4f, 0x2a, 0x4d, 0xd3, 0x2f, 0xc9, 0xcc, 0x4d, - 0x2d, 0x2e, 0x49, 0xcc, 0x2d, 0xd0, 0x03, 0x0b, 0x09, 0xf1, 0x43, 0x14, 0xe8, 0xc1, 0x14, 0x28, - 0x59, 0x73, 0x71, 0x86, 0xc0, 0xd4, 0x08, 0x49, 0x70, 0xb1, 0x17, 0xa7, 0x26, 0xe7, 0xe7, 0xa5, - 0x14, 0x4b, 0x30, 0x2a, 0x30, 0x6a, 0x30, 0x07, 0xc1, 0xb8, 0x42, 0x22, 0x5c, 0xac, 0x79, 0x89, - 0x79, 0xf9, 0xc5, 0x12, 0x4c, 0x0a, 0x8c, 0x1a, 0xac, 0x41, 0x10, 0x8e, 0x53, 0x1d, 0x97, 0x70, - 0x72, 0x7e, 0xae, 0x1e, 0x9a, 0x99, 0x4e, 0x7c, 0x70, 0x13, 0x03, 0x40, 0x42, 0x01, 0x8c, 0x51, - 0xda, 0xe9, 0x99, 0x25, 0x19, 0xa5, 0x49, 0x7a, 0xc9, 0xf9, 0xb9, 0xfa, 0xe9, 0xf9, 0x39, 0x89, - 0x79, 0xe9, 0x08, 0x27, 0x16, 0x94, 0x54, 0x16, 0xa4, 0x16, 0x23, 0x5c, 0xfa, 0x83, 0x91, 0x71, - 0x11, 0x13, 0xb3, 0x7b, 0x80, 0xd3, 0x2a, 0x26, 0x39, 0x77, 0x88, 0xc9, 0x01, 0x50, 0xb5, 0x7a, - 0xe1, 0xa9, 0x39, 0x39, 0xde, 0x79, 0xf9, 0xe5, 0x79, 0x21, 0x20, 0x3d, 0x49, 0x6c, 0x60, 0x43, - 0x8c, 0x01, 0x01, 0x00, 0x00, 0xff, 0xff, 0xbc, 0x77, 0x4a, 0x07, 0xf7, 0x00, 0x00, 0x00, + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_rawDesc, + NumEnums: 0, + NumMessages: 0, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_goTypes, + DependencyIndexes: file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_depIdxs, + }.Build() + File_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto = out.File + file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_rawDesc = nil + file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_goTypes = nil + file_github_com_golang_protobuf_ptypes_timestamp_timestamp_proto_depIdxs = nil } diff --git a/vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.proto b/vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.proto deleted file mode 100644 index cd357864..00000000 --- a/vendor/github.com/golang/protobuf/ptypes/timestamp/timestamp.proto +++ /dev/null @@ -1,138 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -syntax = "proto3"; - -package google.protobuf; - -option csharp_namespace = "Google.Protobuf.WellKnownTypes"; -option cc_enable_arenas = true; -option go_package = "github.com/golang/protobuf/ptypes/timestamp"; -option java_package = "com.google.protobuf"; -option java_outer_classname = "TimestampProto"; -option java_multiple_files = true; -option objc_class_prefix = "GPB"; - -// A Timestamp represents a point in time independent of any time zone or local -// calendar, encoded as a count of seconds and fractions of seconds at -// nanosecond resolution. The count is relative to an epoch at UTC midnight on -// January 1, 1970, in the proleptic Gregorian calendar which extends the -// Gregorian calendar backwards to year one. -// -// All minutes are 60 seconds long. Leap seconds are "smeared" so that no leap -// second table is needed for interpretation, using a [24-hour linear -// smear](https://developers.google.com/time/smear). -// -// The range is from 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z. By -// restricting to that range, we ensure that we can convert to and from [RFC -// 3339](https://www.ietf.org/rfc/rfc3339.txt) date strings. -// -// # Examples -// -// Example 1: Compute Timestamp from POSIX `time()`. -// -// Timestamp timestamp; -// timestamp.set_seconds(time(NULL)); -// timestamp.set_nanos(0); -// -// Example 2: Compute Timestamp from POSIX `gettimeofday()`. -// -// struct timeval tv; -// gettimeofday(&tv, NULL); -// -// Timestamp timestamp; -// timestamp.set_seconds(tv.tv_sec); -// timestamp.set_nanos(tv.tv_usec * 1000); -// -// Example 3: Compute Timestamp from Win32 `GetSystemTimeAsFileTime()`. -// -// FILETIME ft; -// GetSystemTimeAsFileTime(&ft); -// UINT64 ticks = (((UINT64)ft.dwHighDateTime) << 32) | ft.dwLowDateTime; -// -// // A Windows tick is 100 nanoseconds. Windows epoch 1601-01-01T00:00:00Z -// // is 11644473600 seconds before Unix epoch 1970-01-01T00:00:00Z. -// Timestamp timestamp; -// timestamp.set_seconds((INT64) ((ticks / 10000000) - 11644473600LL)); -// timestamp.set_nanos((INT32) ((ticks % 10000000) * 100)); -// -// Example 4: Compute Timestamp from Java `System.currentTimeMillis()`. -// -// long millis = System.currentTimeMillis(); -// -// Timestamp timestamp = Timestamp.newBuilder().setSeconds(millis / 1000) -// .setNanos((int) ((millis % 1000) * 1000000)).build(); -// -// -// Example 5: Compute Timestamp from current time in Python. -// -// timestamp = Timestamp() -// timestamp.GetCurrentTime() -// -// # JSON Mapping -// -// In JSON format, the Timestamp type is encoded as a string in the -// [RFC 3339](https://www.ietf.org/rfc/rfc3339.txt) format. That is, the -// format is "{year}-{month}-{day}T{hour}:{min}:{sec}[.{frac_sec}]Z" -// where {year} is always expressed using four digits while {month}, {day}, -// {hour}, {min}, and {sec} are zero-padded to two digits each. The fractional -// seconds, which can go up to 9 digits (i.e. up to 1 nanosecond resolution), -// are optional. The "Z" suffix indicates the timezone ("UTC"); the timezone -// is required. A proto3 JSON serializer should always use UTC (as indicated by -// "Z") when printing the Timestamp type and a proto3 JSON parser should be -// able to accept both UTC and other timezones (as indicated by an offset). -// -// For example, "2017-01-15T01:30:15.01Z" encodes 15.01 seconds past -// 01:30 UTC on January 15, 2017. -// -// In JavaScript, one can convert a Date object to this format using the -// standard -// [toISOString()](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Date/toISOString) -// method. In Python, a standard `datetime.datetime` object can be converted -// to this format using -// [`strftime`](https://docs.python.org/2/library/time.html#time.strftime) with -// the time format spec '%Y-%m-%dT%H:%M:%S.%fZ'. Likewise, in Java, one can use -// the Joda Time's [`ISODateTimeFormat.dateTime()`]( -// http://www.joda.org/joda-time/apidocs/org/joda/time/format/ISODateTimeFormat.html#dateTime%2D%2D -// ) to obtain a formatter capable of generating timestamps in this format. -// -// -message Timestamp { - // Represents seconds of UTC time since Unix epoch - // 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to - // 9999-12-31T23:59:59Z inclusive. - int64 seconds = 1; - - // Non-negative fractions of a second at nanosecond resolution. Negative - // second values with fractions must still have non-negative nanos values - // that count forward in time. Must be from 0 to 999,999,999 - // inclusive. - int32 nanos = 2; -} diff --git a/vendor/github.com/google/go-jsonnet/.gitignore b/vendor/github.com/google/go-jsonnet/.gitignore index 7712fea0..efa31694 100644 --- a/vendor/github.com/google/go-jsonnet/.gitignore +++ b/vendor/github.com/google/go-jsonnet/.gitignore @@ -23,10 +23,7 @@ gojsonnet.egg-info/ /jsonnet-old /jsonnet-old.exe -/jsonnetfmt /linter/jsonnet-lint/jsonnet-lint /tests_path.source -/jsonnet-lint - /builtin-benchmark-results diff --git a/vendor/github.com/google/go-jsonnet/.golangci.yml b/vendor/github.com/google/go-jsonnet/.golangci.yml deleted file mode 100644 index 3e5a3cf6..00000000 --- a/vendor/github.com/google/go-jsonnet/.golangci.yml +++ /dev/null @@ -1,15 +0,0 @@ -run: - skip-files: ast/identifier_set.go -linters: - enable: - - stylecheck - - gochecknoinits - - golint -issues: - exclude-use-default: false - exclude: - - "should have a package comment, unless it's in another file for this package" - - "the surrounding loop is unconditionally terminated" -linters-settings: - golint: - min-confidence: 0 diff --git a/vendor/github.com/google/go-jsonnet/.goreleaser.yml b/vendor/github.com/google/go-jsonnet/.goreleaser.yml deleted file mode 100644 index 596edb7e..00000000 --- a/vendor/github.com/google/go-jsonnet/.goreleaser.yml +++ /dev/null @@ -1,87 +0,0 @@ -# This is an example goreleaser.yaml file with some sane defaults. -# Make sure to check the documentation at http://goreleaser.com - -builds: - - env: - - CGO_ENABLED=0 - goos: - - linux - - windows - - darwin - goarch: - - 386 - - amd64 - - arm - - arm64 - ignore: - - goos: darwin - - goarch: 386 - - id: jsonnet - main: ./cmd/jsonnet - binary: jsonnet - - # goreleaser complains about unexpected keys, so there's nowhere to hang an - # anchor, so we have to repeat the common elements :( - - env: - - CGO_ENABLED=0 - goos: - - linux - - windows - - darwin - goarch: - - 386 - - amd64 - - arm - - arm64 - ignore: - - goos: darwin - - goarch: 386 - - id: jsonnetfmt - main: ./cmd/jsonnetfmt - binary: jsonnetfmt - -archives: - - replacements: - darwin: Darwin - linux: Linux - windows: Windows - 386: i386 - amd64: x86_64 -checksum: - name_template: 'checksums.txt' - -nfpms: - - id: jsonnet - package_name: jsonnet-go - builds: - - jsonnet - description: A data templating language for app and tool developers - homepage: https://github.com/google/go-jsonnet - license: Apache 2.0 - formats: - - deb - bindir: /usr/bin - maintainer: David Cunningham - file_name_template: "jsonnet-go_{{ .Version }}_{{ .Os }}_{{ .Arch }}" - overrides: - deb: - conflicts: - # See: https://packages.ubuntu.com/jsonnet - - jsonnet - - id: jsonnetfmt - package_name: jsonnetfmt-go - builds: - - jsonnetfmt - homepage: https://github.com/google/go-jsonnet - license: Apache 2.0 - formats: - - deb - bindir: /usr/bin - file_name_template: "jsonnetfmt-go_{{ .Version }}_{{ .Os }}_{{ .Arch }}" - overrides: - deb: - conflicts: - # See: https://packages.ubuntu.com/jsonnet - - jsonnetfmt diff --git a/vendor/github.com/google/go-jsonnet/.travis.yml b/vendor/github.com/google/go-jsonnet/.travis.yml index 0f4516dc..ebbfd131 100644 --- a/vendor/github.com/google/go-jsonnet/.travis.yml +++ b/vendor/github.com/google/go-jsonnet/.travis.yml @@ -6,26 +6,18 @@ matrix: - go: 1.11.x - go: 1.12.x - go: 1.13.x - - go: 1.x arch: amd64 - name: "arch: arm64" - go: 1.x + go: 1.13.x arch: arm64 env: - PYTHON_COMMAND=python3 - - name: "arch: i686" - go: 1.x - arch: amd64 - env: - - PYTHON_COMMAND=python3 - - GOARCH=386 - - CGO_ENABLED=1 - - SKIP_PYTHON_BINDINGS_TESTS=1 - name: "arch: ppc64le" - go: 1.x + go: 1.13.x arch: ppc64le env: - PYTHON_COMMAND=python3 + - go: tip - name: "Bazel Check" go: 1.x script: ./travisBazel.sh @@ -39,17 +31,13 @@ matrix: - echo "deb [arch=amd64] https://storage.googleapis.com/bazel-apt stable jdk1.8" | sudo tee /etc/apt/sources.list.d/bazel.list - curl https://bazel.build/bazel-release.pub.gpg | sudo apt-key add - - sudo apt-get update && sudo apt-get install bazel make - - sudo apt install python3-dev - - pip install -U pytest --user script: make all before_install: - sudo apt install python3-dev - - pip install -U pytest --user - go get github.com/axw/gocov/gocov - go get github.com/mattn/goveralls - go get github.com/fatih/color - - curl -sSfL https://raw.githubusercontent.com/golangci/golangci-lint/master/install.sh | sh -s -- -b $(go env GOPATH)/bin v1.27.0 - if ! go get github.com/golang/tools/cmd/cover; then go get golang.org/x/tools/cmd/cover; fi - go get github.com/sergi/go-diff/diffmatchpatch diff --git a/vendor/github.com/google/go-jsonnet/BUILD.bazel b/vendor/github.com/google/go-jsonnet/BUILD.bazel index 1410047d..6bb698d4 100644 --- a/vendor/github.com/google/go-jsonnet/BUILD.bazel +++ b/vendor/github.com/google/go-jsonnet/BUILD.bazel @@ -19,7 +19,6 @@ go_library( "interpreter.go", "runtime_error.go", "thunks.go", - "util.go", "value.go", "vm.go", ], @@ -29,7 +28,6 @@ go_library( "//ast:go_default_library", "//astgen:go_default_library", "//internal/errors:go_default_library", - "//internal/parser:go_default_library", "//internal/program:go_default_library", ], ) @@ -37,7 +35,6 @@ go_library( go_test( name = "go_default_test", srcs = [ - "builtins_benchmark_test.go", "interpreter_test.go", "jsonnet_test.go", "main_test.go", @@ -47,6 +44,6 @@ go_test( deps = [ "//ast:go_default_library", "//internal/parser:go_default_library", - "//internal/testutils:go_default_library", + "@com_github_sergi_go_diff//diffmatchpatch:go_default_library", ], ) diff --git a/vendor/github.com/google/go-jsonnet/MANIFEST.in b/vendor/github.com/google/go-jsonnet/MANIFEST.in deleted file mode 100644 index a2859e0f..00000000 --- a/vendor/github.com/google/go-jsonnet/MANIFEST.in +++ /dev/null @@ -1,9 +0,0 @@ -include *.go -graft internal -graft ast -graft toolutils -graft astgen -include cpp-jsonnet/include/libjsonnet.h -include go.mod -include go.sum -graft c-bindings diff --git a/vendor/github.com/google/go-jsonnet/Makefile b/vendor/github.com/google/go-jsonnet/Makefile index 201696a2..3072cf54 100644 --- a/vendor/github.com/google/go-jsonnet/Makefile +++ b/vendor/github.com/google/go-jsonnet/Makefile @@ -43,14 +43,14 @@ build: .PHONY: build build.old: - go build -o jsonnet-old ./cmd/jsonnet + go build ./cmd/jsonnet -o jsonnet-old .PHONY: build.old test: ./tests.sh .PHONY: test -benchmark : FILTER ?= Builtin +benchmark : FILTER="Builtin" benchmark: build ./benchmark.sh ${FILTER} .PHONY: benchmark diff --git a/vendor/github.com/google/go-jsonnet/README.md b/vendor/github.com/google/go-jsonnet/README.md index 8a5124c3..61c30626 100644 --- a/vendor/github.com/google/go-jsonnet/README.md +++ b/vendor/github.com/google/go-jsonnet/README.md @@ -9,9 +9,9 @@ [Coverage Status Widget]: https://coveralls.io/repos/github/google/go-jsonnet/badge.svg?branch=master [Coverage Status]: https://coveralls.io/github/google/go-jsonnet?branch=master -This an implementation of [Jsonnet](http://jsonnet.org/) in pure Go. It is a feature complete, production-ready implementation. It is compatible with the original [Jsonnet C++ implementation](https://github.com/google/jsonnet). Bindings to C and Python are available (but not battle-tested yet). +This an implementation of [Jsonnet](http://jsonnet.org/) in pure Go. It is feature complete but is not as heavily exercised as the [Jsonnet C++ implementation](https://github.com/google/jsonnet). Please try it out and give feedback. -This code is known to work on Go 1.11 and above. We recommend always using the newest stable release of Go. +This code is known to work on Go 1.8 and above. We recommend always using the newest stable release of Go. ## Installation instructions @@ -19,20 +19,12 @@ This code is known to work on Go 1.11 and above. We recommend always using the n go get github.com/google/go-jsonnet/cmd/jsonnet ``` -It's also available on Homebrew: - -``` -brew install go-jsonnet -``` - ## Build instructions (go 1.11+) ```bash git clone git@github.com:google/go-jsonnet.git cd go-jsonnet go build ./cmd/jsonnet -go build ./cmd/jsonnetfmt -go build ./cmd/jsonnet-deps ``` To build with [Bazel](https://bazel.build/) instead: ```bash @@ -41,8 +33,6 @@ cd go-jsonnet git submodule init git submodule update bazel build //cmd/jsonnet -bazel build //cmd/jsonnetfmt -bazel build //cmd/jsonnet-deps ``` The resulting _jsonnet_ program will then be available at a platform-specific path, such as _bazel-bin/cmd/jsonnet/darwin_amd64_stripped/jsonnet_ for macOS. @@ -67,7 +57,7 @@ Additionally if any files were moved around, see the section [Keeping the Bazel ## Running Benchmarks -### Method 1 +Setup ```bash go get golang.org/x/tools/cmd/benchcmp @@ -76,7 +66,7 @@ go get golang.org/x/tools/cmd/benchcmp 1. Make sure you build a jsonnet binary _prior_ to making changes. ```bash -go build -o jsonnet-old ./cmd/jsonnet +go build ./cmd/jsonnet -o jsonnet-old ``` 2. Make changes (iterate as needed), and rebuild new binary @@ -92,28 +82,6 @@ go build ./cmd/jsonnet ./benchmark.sh ``` -### Method 2 - -1. get `benchcmp` - -```bash -go get golang.org/x/tools/cmd/benchcmp -``` - -2. Make sure you build a jsonnet binary _prior_ to making changes. - -```bash -make build-old -``` - -3. iterate with (which will also automatically rebuild the new binary `./jsonnet`) - -_replace the FILTER with the name of the test you are working on_ - -```bash -FILTER=Builtin_manifestJsonEx make benchmark -``` - ## Implementation Notes We are generating some helper classes on types by using http://clipperhouse.github.io/gen/. Do the following to regenerate these if necessary: @@ -125,17 +93,6 @@ export PATH=$PATH:$GOPATH/bin # If you haven't already go generate ``` -## Update cpp-jsonnet sub-repo - -This repo depends on [the original Jsonnet repo](https://github.com/google/jsonnet). Shared parts include the standard library, headers files for C API and some tests. - -You can update the submodule and regenerate dependent files with one command: -``` -./update_cpp_jsonnet.sh -``` - -Note: It needs to be run from repo root. - ## Updating and modifying the standard library Standard library source code is kept in `cpp-jsonnet` submodule, because it is shared with [Jsonnet C++ @@ -144,11 +101,11 @@ implementation](https://github.com/google/jsonnet). For performance reasons we perform preprocessing on the standard library, so for the changes to be visible, regeneration is necessary: ```bash +git submodule init +git submodule update go run cmd/dumpstdlibast/dumpstdlibast.go cpp-jsonnet/stdlib/std.jsonnet > astgen/stdast.go ``` -**The - The above command creates the _astgen/stdast.go_ file which puts the desugared standard library into the right data structures, which lets us avoid the parsing overhead during execution. Note that this step is not necessary to perform manually when building with Bazel; the Bazel target regenerates the _astgen/stdast.go_ (writing it into Bazel's build sandbox directory tree) file when necessary. ## Keeping the Bazel files up to date diff --git a/vendor/github.com/google/go-jsonnet/ast/ast.go b/vendor/github.com/google/go-jsonnet/ast/ast.go index 087e52f3..6d207f5f 100644 --- a/vendor/github.com/google/go-jsonnet/ast/ast.go +++ b/vendor/github.com/google/go-jsonnet/ast/ast.go @@ -43,13 +43,6 @@ type Node interface { FreeVariables() Identifiers SetFreeVariables(Identifiers) SetContext(Context) - // OpenFodder returns the fodder before the first token of an AST node. - // Since every AST node has opening fodder, it is defined here. - // If the AST node is left recursive (e.g. BinaryOp) then it is ambiguous - // where the fodder should be stored. This is resolved by storing it as - // far inside the tree as possible. OpenFodder returns a pointer to allow - // the caller to modify the fodder. - OpenFodder() *Fodder } // Nodes represents a Node slice. @@ -89,8 +82,8 @@ func (n *NodeBase) Loc() *LocationRange { } // OpenFodder returns a NodeBase's opening fodder. -func (n *NodeBase) OpenFodder() *Fodder { - return &n.Fodder +func (n *NodeBase) OpenFodder() Fodder { + return n.Fodder } // FreeVariables returns a NodeBase's freeVariables. @@ -378,23 +371,20 @@ type Error struct { type Function struct { NodeBase ParenLeftFodder Fodder - Parameters []Parameter + Parameters Parameters // Always false if there were no parameters. TrailingComma bool ParenRightFodder Fodder Body Node } -// Parameter represents a parameter of function. -// If DefaultArg is set, it's an optional named parameter. -// Otherwise, it's a positional parameter and EqFodder is not used. -type Parameter struct { +// NamedParameter represents an optional named parameter of a function. +type NamedParameter struct { NameFodder Fodder Name Identifier EqFodder Fodder DefaultArg Node CommaFodder Fodder - LocRange LocationRange } // CommaSeparatedID represents an expression that is an element of a @@ -405,6 +395,13 @@ type CommaSeparatedID struct { CommaFodder Fodder } +// Parameters represents the required positional parameters and optional named +// parameters to a function definition. +type Parameters struct { + Required []CommaSeparatedID + Optional []NamedParameter +} + // --------------------------------------------------------------------------- // Import represents import "file". @@ -435,10 +432,9 @@ type Index struct { LeftBracketFodder Fodder Index Node // When Index is being used, this is the fodder before the ']'. - // When Id is being used, this is the fodder before the id. + // When Id is being used, this is always empty. RightBracketFodder Fodder - //nolint: golint,stylecheck // keeping Id instead of ID for now to avoid breaking 3rd parties - Id *Identifier + Id *Identifier } // Slice represents an array slice a[begin:end:step]. @@ -469,8 +465,6 @@ type LocalBind struct { Fun *Function // The fodder before the closing ',' or ';' (whichever it is) CloseFodder Fodder - - LocRange LocationRange } // LocalBinds represents a LocalBind slice. @@ -501,6 +495,7 @@ type LiteralNull struct{ NodeBase } // LiteralNumber represents a JSON number type LiteralNumber struct { NodeBase + Value float64 OriginalString string } @@ -533,10 +528,9 @@ func (k LiteralStringKind) FullyEscaped() bool { // LiteralString represents a JSON string type LiteralString struct { NodeBase - Value string - Kind LiteralStringKind - BlockIndent string - BlockTermIndent string + Value string + Kind LiteralStringKind + BlockIndent string } // --------------------------------------------------------------------------- @@ -588,26 +582,23 @@ type ObjectField struct { // If Method is set then Expr2 == Method.Body. // There is no base fodder in Method because there was no `function` // keyword. - Method *Function - Fodder1 Fodder - Expr1 Node // Not in scope of the object - //nolint: golint,stylecheck // keeping Id instead of ID for now to avoid breaking 3rd parties + Method *Function + Fodder1 Fodder + Expr1 Node // Not in scope of the object Id *Identifier Fodder2 Fodder OpFodder Fodder Expr2, Expr3 Node // In scope of the object (can see self). CommaFodder Fodder - LocRange LocationRange } // ObjectFieldLocalNoMethod creates a non-method local object field. -func ObjectFieldLocalNoMethod(id *Identifier, body Node, loc LocationRange) ObjectField { +func ObjectFieldLocalNoMethod(id *Identifier, body Node) ObjectField { return ObjectField{ - Kind: ObjectLocal, - Hide: ObjectFieldVisible, - Id: id, - Expr2: body, - LocRange: loc, + Kind: ObjectLocal, + Hide: ObjectFieldVisible, + Id: id, + Expr2: body, } } @@ -633,8 +624,6 @@ type DesugaredObjectField struct { Name Node Body Node PlusSuper bool - - LocRange LocationRange } // DesugaredObjectFields represents a DesugaredObjectField slice. @@ -657,11 +646,10 @@ type DesugaredObject struct { // { [e]: e for x in e for.. if... }. type ObjectComp struct { NodeBase - Fields ObjectFields - TrailingCommaFodder Fodder - TrailingComma bool - Spec ForSpec - CloseFodder Fodder + Fields ObjectFields + TrailingComma bool + Spec ForSpec + CloseFodder Fodder } // --------------------------------------------------------------------------- @@ -694,8 +682,7 @@ type SuperIndex struct { // If super.f, the fodder before the 'f' // If super[e], the fodder before the ']'. IDFodder Fodder - //nolint: golint,stylecheck // keeping Id instead of ID for now to avoid breaking 3rd parties - Id *Identifier + Id *Identifier } // InSuper represents the e in super construct. @@ -753,7 +740,6 @@ type Unary struct { // Var represents variables. type Var struct { NodeBase - //nolint: golint,stylecheck // keeping Id instead of ID for now to avoid breaking 3rd parties Id Identifier } diff --git a/vendor/github.com/google/go-jsonnet/ast/clone.go b/vendor/github.com/google/go-jsonnet/ast/clone.go index ce877b0b..65ab146d 100644 --- a/vendor/github.com/google/go-jsonnet/ast/clone.go +++ b/vendor/github.com/google/go-jsonnet/ast/clone.go @@ -35,6 +35,17 @@ func cloneForSpec(specPtr *ForSpec) { } } +// Updates fields of params to point to deep clones. +func cloneParameters(params *Parameters) { + if params == nil { + return + } + params.Optional = append(make([]NamedParameter, 0), params.Optional...) + for i := range params.Optional { + clone(¶ms.Optional[i].DefaultArg) + } +} + // Updates fields of field to point to deep clones. func cloneField(field *ObjectField) { if field.Method != nil { @@ -147,12 +158,7 @@ func clone(astPtr *Node) { r := new(Function) *astPtr = r *r = *node - if r.Parameters != nil { - r.Parameters = append(make([]Parameter, 0), r.Parameters...) - for i := range r.Parameters { - clone(&r.Parameters[i].DefaultArg) - } - } + cloneParameters(&r.Parameters) clone(&r.Body) case *Import: diff --git a/vendor/github.com/google/go-jsonnet/ast/fodder.go b/vendor/github.com/google/go-jsonnet/ast/fodder.go index ab6c2fe5..22eaa13a 100644 --- a/vendor/github.com/google/go-jsonnet/ast/fodder.go +++ b/vendor/github.com/google/go-jsonnet/ast/fodder.go @@ -87,7 +87,7 @@ func MakeFodderElement(kind FodderKind, blanks int, indent int, comment []string panic(fmt.Sprintf("FodderInterstitial but comment == %v.", comment)) } if kind == FodderParagraph && len(comment) == 0 { - panic("FodderParagraph but comment was empty") + panic(fmt.Sprintf("FodderParagraph but comment was empty")) } return FodderElement{Kind: kind, Blanks: blanks, Indent: indent, Comment: comment} } diff --git a/vendor/github.com/google/go-jsonnet/ast/location.go b/vendor/github.com/google/go-jsonnet/ast/location.go index e7d08f86..2ac88751 100644 --- a/vendor/github.com/google/go-jsonnet/ast/location.go +++ b/vendor/github.com/google/go-jsonnet/ast/location.go @@ -21,17 +21,9 @@ import ( "fmt" ) -// DiagnosticFileName is a file name used for diagnostics. -// It might be a dummy value, such as or . -// It should never be passed to an importer. -type DiagnosticFileName string - // Source represents a source file. type Source struct { Lines []string - // DiagnosticFileName is the imported path or a special string - // for indicating stdin, extvars and other non-imported sources. - DiagnosticFileName DiagnosticFileName } ////////////////////////////////////////////////////////////////////////////// @@ -53,10 +45,7 @@ func (l *Location) String() string { return fmt.Sprintf("%v:%v", l.Line, l.Column) } -// LocationBefore returns whether one code location -// refers to the location closer to the beginning -// of the file than the other one. -func LocationBefore(a Location, b Location) bool { +func locationBefore(a Location, b Location) bool { if a.Line != b.Line { return a.Line < b.Line } @@ -68,7 +57,6 @@ func LocationBefore(a Location, b Location) bool { // LocationRange represents a range of a source file. type LocationRange struct { - // FileName should be the imported path or "" for snippets etc. FileName string Begin Location End Location // TODO(sbarzowski) inclusive? exclusive? a gap? @@ -90,13 +78,12 @@ func (lr *LocationRange) IsSet() bool { func (lr *LocationRange) String() string { if !lr.IsSet() { - // TODO(sbarzowski) when could this happen? return lr.FileName } var filePrefix string - if len(lr.File.DiagnosticFileName) > 0 { - filePrefix = string(lr.File.DiagnosticFileName) + ":" + if len(lr.FileName) > 0 { + filePrefix = lr.FileName + ":" } if lr.Begin.Line == lr.End.Line { if lr.Begin.Column == lr.End.Column { @@ -151,7 +138,7 @@ func (sp *SourceProvider) GetSnippet(loc LocationRange) string { // BuildSource transforms a source file string into a Source struct. // TODO: This seems like a job for strings.Split() with a final \n touch-up. -func BuildSource(dFilename DiagnosticFileName, s string) *Source { +func BuildSource(s string) *Source { var result []string var lineBuf bytes.Buffer for _, runeValue := range s { @@ -164,7 +151,7 @@ func BuildSource(dFilename DiagnosticFileName, s string) *Source { rest := lineBuf.String() // Stuff after last end-of-line (EOF or some more code) result = append(result, rest+"\n") - return &Source{result, dFilename} + return &Source{result} } func trimToLine(loc LocationRange, line int) LocationRange { diff --git a/vendor/github.com/google/go-jsonnet/ast/util.go b/vendor/github.com/google/go-jsonnet/ast/util.go index e105a365..f086b015 100644 --- a/vendor/github.com/google/go-jsonnet/ast/util.go +++ b/vendor/github.com/google/go-jsonnet/ast/util.go @@ -16,9 +16,7 @@ limitations under the License. package ast -import ( - "sort" -) +import "sort" // AddIdentifiers adds a slice of identifiers to an identifier set. func (i IdentifierSet) AddIdentifiers(idents Identifiers) { diff --git a/vendor/github.com/google/go-jsonnet/astgen/BUILD.bazel b/vendor/github.com/google/go-jsonnet/astgen/BUILD.bazel index 2d0a4b57..0bec2f1d 100644 --- a/vendor/github.com/google/go-jsonnet/astgen/BUILD.bazel +++ b/vendor/github.com/google/go-jsonnet/astgen/BUILD.bazel @@ -2,7 +2,7 @@ load("@io_bazel_rules_go//go:def.bzl", "go_library") genrule( name = "dumpstdlibast", - srcs = ["@cpp_jsonnet//stdlib"], + srcs = ["//cpp-jsonnet/stdlib"], outs = ["stdast.go"], cmd = "./$(location //cmd/dumpstdlibast) \"$<\" > \"$@\"", tools = ["//cmd/dumpstdlibast"], diff --git a/vendor/github.com/google/go-jsonnet/astgen/stdast.go b/vendor/github.com/google/go-jsonnet/astgen/stdast.go index f538fc72..be80e0f6 100644 --- a/vendor/github.com/google/go-jsonnet/astgen/stdast.go +++ b/vendor/github.com/google/go-jsonnet/astgen/stdast.go @@ -11,8 +11,8 @@ import ( "github.com/google/go-jsonnet/ast" ) -var p7Var = "$" -var p7 = &p7Var +var p6Var = "$" +var p6 = &p6Var var p11Var = "object " var p11 = &p11Var var p15Var = "function " @@ -775,1608 +775,1540 @@ var p5262Var = "thunk from >" var p5262 = &p5262Var var p5275Var = "thunk from >>" var p5275 = &p5275Var -var p5297Var = "thunk from >" -var p5297 = &p5297Var -var p5303Var = "function " +var p5285Var = "thunk from >" +var p5285 = &p5285Var +var p5290Var = "function " +var p5290 = &p5290Var +var p5294Var = "thunk from >" +var p5294 = &p5294Var +var p5303Var = "thunk from from >>" var p5303 = &p5303Var -var p5308Var = "thunk from >" -var p5308 = &p5308Var -var p5323Var = "thunk from from >>" -var p5323 = &p5323Var -var p5328Var = "function " -var p5328 = &p5328Var -var p5348Var = "thunk from >" -var p5348 = &p5348Var -var p5357Var = "thunk from >>" -var p5357 = &p5357Var -var p5382Var = "thunk from from >>" -var p5382 = &p5382Var -var p5389Var = "thunk from >" -var p5389 = &p5389Var -var p5412Var = "thunk from >" -var p5412 = &p5412Var -var p5421Var = "thunk from from >>" -var p5421 = &p5421Var -var p5430Var = "thunk from >" -var p5430 = &p5430Var -var p5436Var = "thunk from from >>" -var p5436 = &p5436Var -var p5467Var = "thunk from >" -var p5467 = &p5467Var -var p5472Var = "function " -var p5472 = &p5472Var -var p5476Var = "thunk from >" -var p5476 = &p5476Var -var p5481Var = "thunk from from >>" -var p5481 = &p5481Var -var p5499Var = "thunk from from >>" +var p5310Var = "thunk from >" +var p5310 = &p5310Var +var p5315Var = "function " +var p5315 = &p5315Var +var p5335Var = "thunk from >" +var p5335 = &p5335Var +var p5344Var = "thunk from >>" +var p5344 = &p5344Var +var p5367Var = "thunk from >" +var p5367 = &p5367Var +var p5380Var = "thunk from from >>" +var p5380 = &p5380Var +var p5390Var = "thunk from from >>" +var p5390 = &p5390Var +var p5399Var = "thunk from from >>>" +var p5399 = &p5399Var +var p5406Var = "thunk from >" +var p5406 = &p5406Var +var p5416Var = "thunk from >" +var p5416 = &p5416Var +var p5439Var = "thunk from >" +var p5439 = &p5439Var +var p5448Var = "thunk from from >>" +var p5448 = &p5448Var +var p5457Var = "thunk from >" +var p5457 = &p5457Var +var p5463Var = "thunk from from >>" +var p5463 = &p5463Var +var p5494Var = "thunk from >" +var p5494 = &p5494Var +var p5499Var = "function " var p5499 = &p5499Var -var p5509Var = "thunk from from >>" -var p5509 = &p5509Var -var p5520Var = "thunk from >" -var p5520 = &p5520Var -var p5529Var = "thunk from from >>" -var p5529 = &p5529Var -var p5536Var = "thunk from >" +var p5503Var = "thunk from >" +var p5503 = &p5503Var +var p5508Var = "thunk from from >>" +var p5508 = &p5508Var +var p5526Var = "thunk from from >>" +var p5526 = &p5526Var +var p5536Var = "thunk from from >>" var p5536 = &p5536Var -var p5541Var = "function " -var p5541 = &p5541Var -var p5560Var = "thunk from >" -var p5560 = &p5560Var -var p5569Var = "thunk from >>" -var p5569 = &p5569Var -var p5594Var = "thunk from >" -var p5594 = &p5594Var -var p5607Var = "thunk from from >>" -var p5607 = &p5607Var -var p5617Var = "thunk from from >>" -var p5617 = &p5617Var -var p5626Var = "thunk from from >>>" -var p5626 = &p5626Var -var p5633Var = "thunk from >" -var p5633 = &p5633Var -var p5643Var = "thunk from >" -var p5643 = &p5643Var -var p5675Var = "thunk from >" -var p5675 = &p5675Var -var p5684Var = "thunk from from >>" -var p5684 = &p5684Var -var p5693Var = "thunk from >" -var p5693 = &p5693Var -var p5713Var = "thunk from from >>" -var p5713 = &p5713Var -var p5743Var = "thunk from >" -var p5743 = &p5743Var -var p5748Var = "function " -var p5748 = &p5748Var -var p5752Var = "thunk from >" -var p5752 = &p5752Var -var p5757Var = "function " -var p5757 = &p5757Var -var p5786Var = "thunk from >" -var p5786 = &p5786Var -var p5804Var = "thunk from >" -var p5804 = &p5804Var -var p5819Var = "thunk from >" -var p5819 = &p5819Var -var p5832Var = "thunk from >>" -var p5832 = &p5832Var -var p5841Var = "thunk from >" -var p5841 = &p5841Var -var p5846Var = "function " +var p5547Var = "thunk from >" +var p5547 = &p5547Var +var p5556Var = "thunk from from >>" +var p5556 = &p5556Var +var p5563Var = "thunk from >" +var p5563 = &p5563Var +var p5568Var = "function " +var p5568 = &p5568Var +var p5587Var = "thunk from >" +var p5587 = &p5587Var +var p5596Var = "thunk from >>" +var p5596 = &p5596Var +var p5621Var = "thunk from >" +var p5621 = &p5621Var +var p5634Var = "thunk from from >>" +var p5634 = &p5634Var +var p5644Var = "thunk from from >>" +var p5644 = &p5644Var +var p5653Var = "thunk from from >>>" +var p5653 = &p5653Var +var p5660Var = "thunk from >" +var p5660 = &p5660Var +var p5670Var = "thunk from >" +var p5670 = &p5670Var +var p5702Var = "thunk from >" +var p5702 = &p5702Var +var p5711Var = "thunk from from >>" +var p5711 = &p5711Var +var p5720Var = "thunk from >" +var p5720 = &p5720Var +var p5740Var = "thunk from from >>" +var p5740 = &p5740Var +var p5770Var = "thunk from >" +var p5770 = &p5770Var +var p5775Var = "function " +var p5775 = &p5775Var +var p5779Var = "thunk from >" +var p5779 = &p5779Var +var p5784Var = "function " +var p5784 = &p5784Var +var p5813Var = "thunk from >" +var p5813 = &p5813Var +var p5831Var = "thunk from >" +var p5831 = &p5831Var +var p5846Var = "thunk from >" var p5846 = &p5846Var -var p5850Var = "thunk from >" -var p5850 = &p5850Var -var p5859Var = "thunk from from >>" +var p5859Var = "thunk from >>" var p5859 = &p5859Var -var p5866Var = "thunk from >" -var p5866 = &p5866Var -var p5875Var = "thunk from from >>" -var p5875 = &p5875Var -var p5882Var = "thunk from >" -var p5882 = &p5882Var -var p5902Var = "thunk from >" +var p5868Var = "thunk from >" +var p5868 = &p5868Var +var p5873Var = "function " +var p5873 = &p5873Var +var p5877Var = "thunk from >" +var p5877 = &p5877Var +var p5886Var = "thunk from from >>" +var p5886 = &p5886Var +var p5893Var = "thunk from >" +var p5893 = &p5893Var +var p5902Var = "thunk from from >>" var p5902 = &p5902Var -var p5917Var = "thunk from >" -var p5917 = &p5917Var -var p5923Var = "thunk from from >>" -var p5923 = &p5923Var -var p5964Var = "thunk from >" -var p5964 = &p5964Var -var p5973Var = "thunk from from >>" -var p5973 = &p5973Var -var p5992Var = "thunk from from >>>" -var p5992 = &p5992Var -var p6013Var = "thunk from >" -var p6013 = &p6013Var -var p6019Var = "thunk from from >>" -var p6019 = &p6019Var -var p6049Var = "thunk from >" -var p6049 = &p6049Var -var p6063Var = "thunk from >" -var p6063 = &p6063Var -var p6068Var = "function " -var p6068 = &p6068Var -var p6072Var = "thunk from >" -var p6072 = &p6072Var -var p6089Var = "thunk from from >>" -var p6089 = &p6089Var -var p6100Var = "thunk from from >>>" -var p6100 = &p6100Var -var p6109Var = "thunk from from >>>>" -var p6109 = &p6109Var -var p6120Var = "thunk from from >>>" -var p6120 = &p6120Var -var p6126Var = "thunk from >" -var p6126 = &p6126Var -var p6141Var = "thunk from from >>" -var p6141 = &p6141Var -var p6155Var = "thunk from from >>>" -var p6155 = &p6155Var -var p6168Var = "thunk from >" -var p6168 = &p6168Var -var p6195Var = "thunk from from >>" -var p6195 = &p6195Var -var p6216Var = "thunk from from >>" -var p6216 = &p6216Var -var p6224Var = "thunk from >" -var p6224 = &p6224Var -var p6237Var = "thunk from from >>" -var p6237 = &p6237Var -var p6248Var = "thunk from >" -var p6248 = &p6248Var -var p6270Var = "thunk from >" -var p6270 = &p6270Var -var p6275Var = "function " -var p6275 = &p6275Var -var p6279Var = "thunk from >" -var p6279 = &p6279Var -var p6289Var = "thunk from >" -var p6289 = &p6289Var -var p6304Var = "thunk from >" -var p6304 = &p6304Var -var p6319Var = "thunk from >" -var p6319 = &p6319Var -var p6359Var = "thunk from >" -var p6359 = &p6359Var -var p6386Var = "thunk from >" -var p6386 = &p6386Var -var p6412Var = "thunk from >" -var p6412 = &p6412Var -var p6422Var = "thunk from >" -var p6422 = &p6422Var -var p6437Var = "thunk from >>" -var p6437 = &p6437Var -var p6446Var = "thunk from >>>" -var p6446 = &p6446Var -var p6489Var = "thunk from >" -var p6489 = &p6489Var -var p6515Var = "thunk from >" -var p6515 = &p6515Var -var p6523Var = "thunk from >" -var p6523 = &p6523Var -var p6539Var = "thunk from >" -var p6539 = &p6539Var -var p6554Var = "thunk from >>" -var p6554 = &p6554Var -var p6563Var = "thunk from >>>" -var p6563 = &p6563Var -var p6607Var = "thunk from >" -var p6607 = &p6607Var -var p6633Var = "thunk from >" -var p6633 = &p6633Var -var p6643Var = "thunk from >" -var p6643 = &p6643Var -var p6652Var = "thunk from >>" -var p6652 = &p6652Var -var p6709Var = "thunk from >" -var p6709 = &p6709Var -var p6735Var = "thunk from >" -var p6735 = &p6735Var -var p6745Var = "thunk from >" -var p6745 = &p6745Var -var p6798Var = "thunk from >" -var p6798 = &p6798Var -var p6824Var = "thunk from >" -var p6824 = &p6824Var -var p6834Var = "thunk from >" -var p6834 = &p6834Var -var p6893Var = "thunk from >" -var p6893 = &p6893Var -var p6919Var = "thunk from >" +var p5909Var = "thunk from >" +var p5909 = &p5909Var +var p5929Var = "thunk from >" +var p5929 = &p5929Var +var p5944Var = "thunk from >" +var p5944 = &p5944Var +var p5950Var = "thunk from from >>" +var p5950 = &p5950Var +var p5961Var = "thunk from from >>>" +var p5961 = &p5961Var +var p5999Var = "thunk from >" +var p5999 = &p5999Var +var p6008Var = "thunk from from >>" +var p6008 = &p6008Var +var p6027Var = "thunk from from >>>" +var p6027 = &p6027Var +var p6048Var = "thunk from >" +var p6048 = &p6048Var +var p6054Var = "thunk from from >>" +var p6054 = &p6054Var +var p6083Var = "thunk from >" +var p6083 = &p6083Var +var p6097Var = "thunk from >" +var p6097 = &p6097Var +var p6102Var = "function " +var p6102 = &p6102Var +var p6106Var = "thunk from >" +var p6106 = &p6106Var +var p6123Var = "thunk from from >>" +var p6123 = &p6123Var +var p6134Var = "thunk from from >>>" +var p6134 = &p6134Var +var p6143Var = "thunk from from >>>>" +var p6143 = &p6143Var +var p6154Var = "thunk from from >>>" +var p6154 = &p6154Var +var p6160Var = "thunk from >" +var p6160 = &p6160Var +var p6175Var = "thunk from from >>" +var p6175 = &p6175Var +var p6188Var = "thunk from >" +var p6188 = &p6188Var +var p6215Var = "thunk from from >>" +var p6215 = &p6215Var +var p6236Var = "thunk from from >>" +var p6236 = &p6236Var +var p6244Var = "thunk from >" +var p6244 = &p6244Var +var p6257Var = "thunk from from >>" +var p6257 = &p6257Var +var p6268Var = "thunk from >" +var p6268 = &p6268Var +var p6290Var = "thunk from >" +var p6290 = &p6290Var +var p6295Var = "function " +var p6295 = &p6295Var +var p6299Var = "thunk from >" +var p6299 = &p6299Var +var p6309Var = "thunk from >" +var p6309 = &p6309Var +var p6324Var = "thunk from >" +var p6324 = &p6324Var +var p6339Var = "thunk from >" +var p6339 = &p6339Var +var p6379Var = "thunk from >" +var p6379 = &p6379Var +var p6406Var = "thunk from >" +var p6406 = &p6406Var +var p6432Var = "thunk from >" +var p6432 = &p6432Var +var p6442Var = "thunk from >" +var p6442 = &p6442Var +var p6485Var = "thunk from >" +var p6485 = &p6485Var +var p6511Var = "thunk from >" +var p6511 = &p6511Var +var p6519Var = "thunk from >" +var p6519 = &p6519Var +var p6535Var = "thunk from >" +var p6535 = &p6535Var +var p6579Var = "thunk from >" +var p6579 = &p6579Var +var p6605Var = "thunk from >" +var p6605 = &p6605Var +var p6615Var = "thunk from >" +var p6615 = &p6615Var +var p6672Var = "thunk from >" +var p6672 = &p6672Var +var p6698Var = "thunk from >" +var p6698 = &p6698Var +var p6708Var = "thunk from >" +var p6708 = &p6708Var +var p6761Var = "thunk from >" +var p6761 = &p6761Var +var p6787Var = "thunk from >" +var p6787 = &p6787Var +var p6797Var = "thunk from >" +var p6797 = &p6797Var +var p6856Var = "thunk from >" +var p6856 = &p6856Var +var p6882Var = "thunk from >" +var p6882 = &p6882Var +var p6890Var = "thunk from >" +var p6890 = &p6890Var +var p6899Var = "thunk from from >>" +var p6899 = &p6899Var +var p6910Var = "thunk from from >>>" +var p6910 = &p6910Var +var p6919Var = "thunk from from >>>>" var p6919 = &p6919Var -var p6927Var = "thunk from >" -var p6927 = &p6927Var -var p6936Var = "thunk from from >>" -var p6936 = &p6936Var -var p6947Var = "thunk from from >>>" -var p6947 = &p6947Var -var p6956Var = "thunk from from >>>>" -var p6956 = &p6956Var -var p6967Var = "thunk from from >>>" -var p6967 = &p6967Var -var p6992Var = "thunk from >" -var p6992 = &p6992Var -var p7039Var = "thunk from >" -var p7039 = &p7039Var -var p7048Var = "thunk from from >>" -var p7048 = &p7048Var -var p7061Var = "thunk from >" -var p7061 = &p7061Var -var p7122Var = "thunk from >" -var p7122 = &p7122Var -var p7135Var = "thunk from >" -var p7135 = &p7135Var -var p7151Var = "thunk from >" -var p7151 = &p7151Var -var p7168Var = "thunk from >" -var p7168 = &p7168Var -var p7190Var = "thunk from >" -var p7190 = &p7190Var -var p7208Var = "thunk from >" -var p7208 = &p7208Var -var p7228Var = "thunk from >" -var p7228 = &p7228Var -var p7233Var = "function " -var p7233 = &p7233Var -var p7248Var = "thunk from >" -var p7248 = &p7248Var -var p7266Var = "thunk from >" -var p7266 = &p7266Var -var p7287Var = "thunk from >" -var p7287 = &p7287Var -var p7302Var = "thunk from >" -var p7302 = &p7302Var -var p7322Var = "thunk from >" -var p7322 = &p7322Var -var p7332Var = "thunk from >" -var p7332 = &p7332Var -var p7355Var = "thunk from >" -var p7355 = &p7355Var -var p7370Var = "object " -var p7370 = &p7370Var -var p7391Var = "thunk from >" -var p7391 = &p7391Var -var p7412Var = "thunk from >" -var p7412 = &p7412Var -var p7430Var = "object " -var p7430 = &p7430Var -var p7443Var = "thunk from >" -var p7443 = &p7443Var -var p7458Var = "object " -var p7458 = &p7458Var -var p7485Var = "thunk from >" -var p7485 = &p7485Var -var p7506Var = "thunk from >" -var p7506 = &p7506Var -var p7530Var = "object " -var p7530 = &p7530Var -var p7546Var = "thunk from >" -var p7546 = &p7546Var -var p7557Var = "thunk from >" -var p7557 = &p7557Var -var p7572Var = "thunk from from >>" -var p7572 = &p7572Var -var p7601Var = "thunk from from >>" -var p7601 = &p7601Var -var p7612Var = "thunk from >" -var p7612 = &p7612Var -var p7632Var = "thunk from from >>" -var p7632 = &p7632Var -var p7654Var = "thunk from >" -var p7654 = &p7654Var -var p7671Var = "thunk from from >>" -var p7671 = &p7671Var -var p7687Var = "thunk from from >>" -var p7687 = &p7687Var -var p7701Var = "thunk from >" -var p7701 = &p7701Var -var p7728Var = "thunk from >" -var p7728 = &p7728Var -var p7751Var = "thunk from >" -var p7751 = &p7751Var -var p7756Var = "function " -var p7756 = &p7756Var -var p7771Var = "thunk from >" -var p7771 = &p7771Var -var p7782Var = "thunk from >" -var p7782 = &p7782Var -var p7802Var = "thunk from >" -var p7802 = &p7802Var -var p7812Var = "thunk from >" -var p7812 = &p7812Var -var p7834Var = "thunk from >" -var p7834 = &p7834Var -var p7860Var = "thunk from >" -var p7860 = &p7860Var -var p7886Var = "thunk from >" +var p6930Var = "thunk from from >>>" +var p6930 = &p6930Var +var p6955Var = "thunk from >" +var p6955 = &p6955Var +var p7002Var = "thunk from >" +var p7002 = &p7002Var +var p7011Var = "thunk from from >>" +var p7011 = &p7011Var +var p7024Var = "thunk from >" +var p7024 = &p7024Var +var p7085Var = "thunk from >" +var p7085 = &p7085Var +var p7098Var = "thunk from >" +var p7098 = &p7098Var +var p7114Var = "thunk from >" +var p7114 = &p7114Var +var p7131Var = "thunk from >" +var p7131 = &p7131Var +var p7153Var = "thunk from >" +var p7153 = &p7153Var +var p7171Var = "thunk from >" +var p7171 = &p7171Var +var p7191Var = "thunk from >" +var p7191 = &p7191Var +var p7196Var = "function " +var p7196 = &p7196Var +var p7211Var = "thunk from >" +var p7211 = &p7211Var +var p7229Var = "thunk from >" +var p7229 = &p7229Var +var p7250Var = "thunk from >" +var p7250 = &p7250Var +var p7265Var = "thunk from >" +var p7265 = &p7265Var +var p7285Var = "thunk from >" +var p7285 = &p7285Var +var p7295Var = "thunk from >" +var p7295 = &p7295Var +var p7318Var = "thunk from >" +var p7318 = &p7318Var +var p7333Var = "object " +var p7333 = &p7333Var +var p7354Var = "thunk from >" +var p7354 = &p7354Var +var p7375Var = "thunk from >" +var p7375 = &p7375Var +var p7393Var = "object " +var p7393 = &p7393Var +var p7406Var = "thunk from >" +var p7406 = &p7406Var +var p7421Var = "object " +var p7421 = &p7421Var +var p7448Var = "thunk from >" +var p7448 = &p7448Var +var p7469Var = "thunk from >" +var p7469 = &p7469Var +var p7493Var = "object " +var p7493 = &p7493Var +var p7509Var = "thunk from >" +var p7509 = &p7509Var +var p7520Var = "thunk from >" +var p7520 = &p7520Var +var p7535Var = "thunk from from >>" +var p7535 = &p7535Var +var p7564Var = "thunk from from >>" +var p7564 = &p7564Var +var p7575Var = "thunk from >" +var p7575 = &p7575Var +var p7595Var = "thunk from from >>" +var p7595 = &p7595Var +var p7617Var = "thunk from >" +var p7617 = &p7617Var +var p7634Var = "thunk from from >>" +var p7634 = &p7634Var +var p7650Var = "thunk from from >>" +var p7650 = &p7650Var +var p7664Var = "thunk from >" +var p7664 = &p7664Var +var p7691Var = "thunk from >" +var p7691 = &p7691Var +var p7714Var = "thunk from >" +var p7714 = &p7714Var +var p7719Var = "function " +var p7719 = &p7719Var +var p7734Var = "thunk from >" +var p7734 = &p7734Var +var p7745Var = "thunk from >" +var p7745 = &p7745Var +var p7765Var = "thunk from >" +var p7765 = &p7765Var +var p7775Var = "thunk from >" +var p7775 = &p7775Var +var p7797Var = "thunk from >" +var p7797 = &p7797Var +var p7823Var = "thunk from >" +var p7823 = &p7823Var +var p7849Var = "thunk from >" +var p7849 = &p7849Var +var p7875Var = "thunk from >" +var p7875 = &p7875Var +var p7886Var = "thunk from from >>" var p7886 = &p7886Var -var p7912Var = "thunk from >" +var p7912Var = "thunk from >" var p7912 = &p7912Var -var p7923Var = "thunk from from >>" -var p7923 = &p7923Var -var p7949Var = "thunk from >" -var p7949 = &p7949Var -var p7969Var = "thunk from from >>" -var p7969 = &p7969Var -var p7985Var = "thunk from >" -var p7985 = &p7985Var -var p8002Var = "thunk from from >>" -var p8002 = &p8002Var -var p8015Var = "thunk from from >>" -var p8015 = &p8015Var -var p8027Var = "thunk from >" -var p8027 = &p8027Var +var p7932Var = "thunk from from >>" +var p7932 = &p7932Var +var p7948Var = "thunk from >" +var p7948 = &p7948Var +var p7965Var = "thunk from from >>" +var p7965 = &p7965Var +var p7978Var = "thunk from from >>" +var p7978 = &p7978Var +var p7990Var = "thunk from >" +var p7990 = &p7990Var +var p8017Var = "thunk from >" +var p8017 = &p8017Var +var p8026Var = "thunk from >" +var p8026 = &p8026Var +var p8045Var = "thunk from >" +var p8045 = &p8045Var var p8054Var = "thunk from >" var p8054 = &p8054Var -var p8063Var = "thunk from >" -var p8063 = &p8063Var -var p8082Var = "thunk from >" -var p8082 = &p8082Var -var p8091Var = "thunk from >" -var p8091 = &p8091Var -var p8105Var = "thunk from >" -var p8105 = &p8105Var -var p8111Var = "thunk from >>" -var p8111 = &p8111Var -var p8122Var = "function " +var p8068Var = "thunk from >" +var p8068 = &p8068Var +var p8074Var = "thunk from >>" +var p8074 = &p8074Var +var p8085Var = "function " +var p8085 = &p8085Var +var p8089Var = "thunk from >" +var p8089 = &p8089Var +var p8094Var = "function " +var p8094 = &p8094Var +var p8112Var = "thunk from >" +var p8112 = &p8112Var +var p8122Var = "thunk from >>" var p8122 = &p8122Var -var p8126Var = "thunk from >" -var p8126 = &p8126Var -var p8131Var = "function " -var p8131 = &p8131Var -var p8149Var = "thunk from >" -var p8149 = &p8149Var -var p8159Var = "thunk from >>" +var p8142Var = "thunk from >" +var p8142 = &p8142Var +var p8159Var = "thunk from >>" var p8159 = &p8159Var -var p8179Var = "thunk from >" -var p8179 = &p8179Var -var p8196Var = "thunk from >>" -var p8196 = &p8196Var -var p8205Var = "function " +var p8168Var = "function " +var p8168 = &p8168Var +var p8172Var = "thunk from >" +var p8172 = &p8172Var +var p8177Var = "function " +var p8177 = &p8177Var +var p8192Var = "thunk from >" +var p8192 = &p8192Var +var p8205Var = "thunk from >" var p8205 = &p8205Var -var p8209Var = "thunk from >" -var p8209 = &p8209Var -var p8214Var = "function " -var p8214 = &p8214Var -var p8229Var = "thunk from >" -var p8229 = &p8229Var -var p8242Var = "thunk from >" -var p8242 = &p8242Var -var p8252Var = "thunk from >>" -var p8252 = &p8252Var -var p8272Var = "thunk from >" -var p8272 = &p8272Var -var p8285Var = "function " -var p8285 = &p8285Var -var p8298Var = "thunk from >" -var p8298 = &p8298Var -var p8315Var = "thunk from >" -var p8315 = &p8315Var -var p8331Var = "thunk from >" -var p8331 = &p8331Var -var p8348Var = "thunk from >" -var p8348 = &p8348Var -var p8364Var = "thunk from >" -var p8364 = &p8364Var -var p8381Var = "thunk from >" -var p8381 = &p8381Var -var p8394Var = "thunk from >" -var p8394 = &p8394Var -var p8405Var = "thunk from >>" -var p8405 = &p8405Var -var p8414Var = "function " -var p8414 = &p8414Var -var p8446Var = "function " -var p8446 = &p8446Var -var p8459Var = "thunk from >" -var p8459 = &p8459Var -var p8476Var = "thunk from >" -var p8476 = &p8476Var -var p8499Var = "function " -var p8499 = &p8499Var -var p8512Var = "thunk from >" -var p8512 = &p8512Var -var p8529Var = "thunk from >" -var p8529 = &p8529Var -var p8561Var = "function " -var p8561 = &p8561Var -var p8574Var = "thunk from >" -var p8574 = &p8574Var -var p8591Var = "thunk from >" -var p8591 = &p8591Var -var p8607Var = "thunk from >" -var p8607 = &p8607Var -var p8624Var = "thunk from >" -var p8624 = &p8624Var -var p8646Var = "function " -var p8646 = &p8646Var -var p8659Var = "thunk from >" -var p8659 = &p8659Var -var p8676Var = "thunk from >" -var p8676 = &p8676Var -var p8692Var = "thunk from >" -var p8692 = &p8692Var -var p8709Var = "thunk from >" -var p8709 = &p8709Var -var p8730Var = "function " -var p8730 = &p8730Var -var p8759Var = "function " -var p8759 = &p8759Var -var p8769Var = "thunk from >" -var p8769 = &p8769Var -var p8772Var = "function " -var p8772 = &p8772Var -var p8787Var = "function " -var p8787 = &p8787Var -var p8791Var = "thunk from >" -var p8791 = &p8791Var -var p8795Var = "function " -var p8795 = &p8795Var -var p8805Var = "thunk from >" -var p8805 = &p8805Var -var p8822Var = "thunk from >>" -var p8822 = &p8822Var -var p8826Var = "thunk from >>>" -var p8826 = &p8826Var -var p8844Var = "thunk from >>>" -var p8844 = &p8844Var -var p8869Var = "thunk from >>>" -var p8869 = &p8869Var -var p8874Var = "thunk from >>>>" -var p8874 = &p8874Var -var p8894Var = "thunk from >>>" -var p8894 = &p8894Var -var p8899Var = "thunk from >>>>" -var p8899 = &p8899Var -var p8912Var = "thunk from >>" -var p8912 = &p8912Var -var p8919Var = "thunk from >" -var p8919 = &p8919Var -var p8923Var = "function " -var p8923 = &p8923Var -var p8937Var = "thunk from >" +var p8215Var = "thunk from >>" +var p8215 = &p8215Var +var p8235Var = "thunk from >" +var p8235 = &p8235Var +var p8248Var = "function " +var p8248 = &p8248Var +var p8261Var = "thunk from >" +var p8261 = &p8261Var +var p8278Var = "thunk from >" +var p8278 = &p8278Var +var p8294Var = "thunk from >" +var p8294 = &p8294Var +var p8311Var = "thunk from >" +var p8311 = &p8311Var +var p8327Var = "thunk from >" +var p8327 = &p8327Var +var p8344Var = "thunk from >" +var p8344 = &p8344Var +var p8357Var = "thunk from >" +var p8357 = &p8357Var +var p8368Var = "thunk from >>" +var p8368 = &p8368Var +var p8377Var = "function " +var p8377 = &p8377Var +var p8409Var = "function " +var p8409 = &p8409Var +var p8422Var = "thunk from >" +var p8422 = &p8422Var +var p8439Var = "thunk from >" +var p8439 = &p8439Var +var p8462Var = "function " +var p8462 = &p8462Var +var p8475Var = "thunk from >" +var p8475 = &p8475Var +var p8492Var = "thunk from >" +var p8492 = &p8492Var +var p8524Var = "function " +var p8524 = &p8524Var +var p8537Var = "thunk from >" +var p8537 = &p8537Var +var p8554Var = "thunk from >" +var p8554 = &p8554Var +var p8570Var = "thunk from >" +var p8570 = &p8570Var +var p8587Var = "thunk from >" +var p8587 = &p8587Var +var p8609Var = "function " +var p8609 = &p8609Var +var p8622Var = "thunk from >" +var p8622 = &p8622Var +var p8639Var = "thunk from >" +var p8639 = &p8639Var +var p8655Var = "thunk from >" +var p8655 = &p8655Var +var p8672Var = "thunk from >" +var p8672 = &p8672Var +var p8693Var = "function " +var p8693 = &p8693Var +var p8722Var = "function " +var p8722 = &p8722Var +var p8732Var = "thunk from >" +var p8732 = &p8732Var +var p8735Var = "function " +var p8735 = &p8735Var +var p8750Var = "function " +var p8750 = &p8750Var +var p8754Var = "thunk from >" +var p8754 = &p8754Var +var p8758Var = "function " +var p8758 = &p8758Var +var p8768Var = "thunk from >" +var p8768 = &p8768Var +var p8785Var = "thunk from >>" +var p8785 = &p8785Var +var p8789Var = "thunk from >>>" +var p8789 = &p8789Var +var p8807Var = "thunk from >>>" +var p8807 = &p8807Var +var p8832Var = "thunk from >>>" +var p8832 = &p8832Var +var p8837Var = "thunk from >>>>" +var p8837 = &p8837Var +var p8857Var = "thunk from >>>" +var p8857 = &p8857Var +var p8862Var = "thunk from >>>>" +var p8862 = &p8862Var +var p8875Var = "thunk from >>" +var p8875 = &p8875Var +var p8882Var = "thunk from >" +var p8882 = &p8882Var +var p8886Var = "function " +var p8886 = &p8886Var +var p8900Var = "thunk from >" +var p8900 = &p8900Var +var p8905Var = "thunk from >>" +var p8905 = &p8905Var +var p8913Var = "thunk from >" +var p8913 = &p8913Var +var p8917Var = "thunk from >" +var p8917 = &p8917Var +var p8928Var = "thunk from from >>" +var p8928 = &p8928Var +var p8937Var = "thunk from from >>" var p8937 = &p8937Var -var p8942Var = "thunk from >>" -var p8942 = &p8942Var -var p8950Var = "thunk from >" -var p8950 = &p8950Var -var p8954Var = "thunk from >" -var p8954 = &p8954Var -var p8965Var = "thunk from from >>" -var p8965 = &p8965Var -var p8974Var = "thunk from from >>" -var p8974 = &p8974Var -var p8996Var = "thunk from from >>" -var p8996 = &p8996Var -var p9003Var = "thunk from from >>>" -var p9003 = &p9003Var -var p9015Var = "thunk from >" -var p9015 = &p9015Var -var p9024Var = "thunk from from >>" -var p9024 = &p9024Var -var p9039Var = "thunk from >" -var p9039 = &p9039Var -var p9055Var = "thunk from >>" +var p8959Var = "thunk from from >>" +var p8959 = &p8959Var +var p8966Var = "thunk from from >>>" +var p8966 = &p8966Var +var p8978Var = "thunk from >" +var p8978 = &p8978Var +var p8987Var = "thunk from from >>" +var p8987 = &p8987Var +var p9002Var = "thunk from >" +var p9002 = &p9002Var +var p9018Var = "thunk from >>" +var p9018 = &p9018Var +var p9023Var = "thunk from >>" +var p9023 = &p9023Var +var p9030Var = "function " +var p9030 = &p9030Var +var p9034Var = "thunk from >" +var p9034 = &p9034Var +var p9043Var = "thunk from from >>" +var p9043 = &p9043Var +var p9050Var = "thunk from >" +var p9050 = &p9050Var +var p9055Var = "function " var p9055 = &p9055Var -var p9060Var = "thunk from >>" -var p9060 = &p9060Var -var p9067Var = "function " -var p9067 = &p9067Var -var p9071Var = "thunk from >" -var p9071 = &p9071Var -var p9080Var = "thunk from from >>" -var p9080 = &p9080Var -var p9087Var = "thunk from >" -var p9087 = &p9087Var -var p9092Var = "function " -var p9092 = &p9092Var -var p9167Var = "thunk from >" -var p9167 = &p9167Var -var p9176Var = "thunk from from >>" -var p9176 = &p9176Var -var p9214Var = "thunk from >" -var p9214 = &p9214Var -var p9239Var = "thunk from >" -var p9239 = &p9239Var -var p9255Var = "thunk from >>" -var p9255 = &p9255Var -var p9261Var = "thunk from >>>" -var p9261 = &p9261Var -var p9272Var = "thunk from >>" -var p9272 = &p9272Var -var p9279Var = "function " -var p9279 = &p9279Var -var p9289Var = "thunk from >" -var p9289 = &p9289Var -var p9297Var = "function " -var p9297 = &p9297Var -var p9301Var = "thunk from >" -var p9301 = &p9301Var -var p9310Var = "thunk from from >>" -var p9310 = &p9310Var -var p9317Var = "thunk from >" -var p9317 = &p9317Var -var p9321Var = "function " -var p9321 = &p9321Var -var p9353Var = "thunk from >" -var p9353 = &p9353Var -var p9369Var = "thunk from >>" -var p9369 = &p9369Var -var p9375Var = "thunk from >>>" -var p9375 = &p9375Var -var p9386Var = "thunk from >>" -var p9386 = &p9386Var -var p9394Var = "function " -var p9394 = &p9394Var -var p9398Var = "thunk from >" -var p9398 = &p9398Var -var p9407Var = "thunk from from >>" -var p9407 = &p9407Var -var p9414Var = "thunk from >" -var p9414 = &p9414Var -var p9418Var = "function " +var p9130Var = "thunk from >" +var p9130 = &p9130Var +var p9139Var = "thunk from from >>" +var p9139 = &p9139Var +var p9177Var = "thunk from >" +var p9177 = &p9177Var +var p9202Var = "thunk from >" +var p9202 = &p9202Var +var p9218Var = "thunk from >>" +var p9218 = &p9218Var +var p9224Var = "thunk from >>>" +var p9224 = &p9224Var +var p9235Var = "thunk from >>" +var p9235 = &p9235Var +var p9242Var = "function " +var p9242 = &p9242Var +var p9252Var = "thunk from >" +var p9252 = &p9252Var +var p9260Var = "function " +var p9260 = &p9260Var +var p9264Var = "thunk from >" +var p9264 = &p9264Var +var p9273Var = "thunk from from >>" +var p9273 = &p9273Var +var p9280Var = "thunk from >" +var p9280 = &p9280Var +var p9284Var = "function " +var p9284 = &p9284Var +var p9316Var = "thunk from >" +var p9316 = &p9316Var +var p9332Var = "thunk from >>" +var p9332 = &p9332Var +var p9338Var = "thunk from >>>" +var p9338 = &p9338Var +var p9349Var = "thunk from >>" +var p9349 = &p9349Var +var p9357Var = "function " +var p9357 = &p9357Var +var p9361Var = "thunk from >" +var p9361 = &p9361Var +var p9370Var = "thunk from from >>" +var p9370 = &p9370Var +var p9377Var = "thunk from >" +var p9377 = &p9377Var +var p9381Var = "function " +var p9381 = &p9381Var +var p9404Var = "thunk from >" +var p9404 = &p9404Var +var p9408Var = "function " +var p9408 = &p9408Var +var p9418Var = "thunk from >" var p9418 = &p9418Var -var p9441Var = "thunk from >" -var p9441 = &p9441Var -var p9445Var = "function " -var p9445 = &p9445Var -var p9455Var = "thunk from >" +var p9429Var = "thunk from >>" +var p9429 = &p9429Var +var p9437Var = "function " +var p9437 = &p9437Var +var p9446Var = "thunk from >" +var p9446 = &p9446Var +var p9455Var = "function " var p9455 = &p9455Var -var p9466Var = "thunk from >>" -var p9466 = &p9466Var -var p9474Var = "function " -var p9474 = &p9474Var -var p9483Var = "thunk from >" -var p9483 = &p9483Var -var p9492Var = "function " -var p9492 = &p9492Var -var p9496Var = "thunk from >" -var p9496 = &p9496Var -var p9501Var = "function " -var p9501 = &p9501Var -var p9542Var = "thunk from >" -var p9542 = &p9542Var -var p9562Var = "thunk from >" -var p9562 = &p9562Var -var p9574Var = "thunk from >" -var p9574 = &p9574Var -var p9588Var = "thunk from >" -var p9588 = &p9588Var -var p9610Var = "thunk from >" -var p9610 = &p9610Var -var p9617Var = "thunk from >" -var p9617 = &p9617Var -var p9626Var = "thunk from from >>" -var p9626 = &p9626Var -var p9638Var = "thunk from from >>>" +var p9459Var = "thunk from >" +var p9459 = &p9459Var +var p9464Var = "function " +var p9464 = &p9464Var +var p9505Var = "thunk from >" +var p9505 = &p9505Var +var p9525Var = "thunk from >" +var p9525 = &p9525Var +var p9537Var = "thunk from >" +var p9537 = &p9537Var +var p9551Var = "thunk from >" +var p9551 = &p9551Var +var p9573Var = "thunk from >" +var p9573 = &p9573Var +var p9580Var = "thunk from >" +var p9580 = &p9580Var +var p9589Var = "thunk from from >>" +var p9589 = &p9589Var +var p9601Var = "thunk from from >>>" +var p9601 = &p9601Var +var p9609Var = "thunk from >" +var p9609 = &p9609Var +var p9620Var = "thunk from >" +var p9620 = &p9620Var +var p9627Var = "thunk from from >>" +var p9627 = &p9627Var +var p9638Var = "thunk from from >>" var p9638 = &p9638Var -var p9646Var = "thunk from >" -var p9646 = &p9646Var -var p9657Var = "thunk from >" -var p9657 = &p9657Var -var p9664Var = "thunk from from >>" -var p9664 = &p9664Var -var p9675Var = "thunk from from >>" -var p9675 = &p9675Var -var p9678Var = "thunk from from >>>" -var p9678 = &p9678Var -var p9695Var = "thunk from from >>>" -var p9695 = &p9695Var -var p9699Var = "thunk from from >>>>" -var p9699 = &p9699Var -var p9709Var = "thunk from from >>>>>" -var p9709 = &p9709Var -var p9723Var = "thunk from from >>>>>>" -var p9723 = &p9723Var -var p9734Var = "thunk from from >>" -var p9734 = &p9734Var -var p9752Var = "thunk from >" -var p9752 = &p9752Var -var p9767Var = "thunk from >" -var p9767 = &p9767Var -var p9774Var = "thunk from >" -var p9774 = &p9774Var -var p9781Var = "thunk from from >>" -var p9781 = &p9781Var -var p9792Var = "thunk from from >>" -var p9792 = &p9792Var -var p9795Var = "thunk from from >>>" -var p9795 = &p9795Var -var p9812Var = "thunk from from >>>" -var p9812 = &p9812Var -var p9816Var = "thunk from from >>>>" -var p9816 = &p9816Var -var p9837Var = "thunk from from >>>>>" -var p9837 = &p9837Var -var p9847Var = "thunk from from >>>>>" -var p9847 = &p9847Var -var p9861Var = "thunk from from >>>>>>" -var p9861 = &p9861Var -var p9878Var = "thunk from from >>>" -var p9878 = &p9878Var -var p9885Var = "thunk from from >>" -var p9885 = &p9885Var -var p9903Var = "thunk from >" -var p9903 = &p9903Var -var p9914Var = "thunk from >" -var p9914 = &p9914Var -var p9923Var = "function " -var p9923 = &p9923Var -var p9929Var = "thunk from >" -var p9929 = &p9929Var -var p9934Var = "function " -var p9934 = &p9934Var -var p9975Var = "thunk from >" +var p9641Var = "thunk from from >>>" +var p9641 = &p9641Var +var p9658Var = "thunk from from >>>" +var p9658 = &p9658Var +var p9662Var = "thunk from from >>>>" +var p9662 = &p9662Var +var p9672Var = "thunk from from >>>>>" +var p9672 = &p9672Var +var p9686Var = "thunk from from >>>>>>" +var p9686 = &p9686Var +var p9697Var = "thunk from from >>" +var p9697 = &p9697Var +var p9715Var = "thunk from >" +var p9715 = &p9715Var +var p9730Var = "thunk from >" +var p9730 = &p9730Var +var p9737Var = "thunk from >" +var p9737 = &p9737Var +var p9744Var = "thunk from from >>" +var p9744 = &p9744Var +var p9755Var = "thunk from from >>" +var p9755 = &p9755Var +var p9758Var = "thunk from from >>>" +var p9758 = &p9758Var +var p9775Var = "thunk from from >>>" +var p9775 = &p9775Var +var p9779Var = "thunk from from >>>>" +var p9779 = &p9779Var +var p9800Var = "thunk from from >>>>>" +var p9800 = &p9800Var +var p9810Var = "thunk from from >>>>>" +var p9810 = &p9810Var +var p9824Var = "thunk from from >>>>>>" +var p9824 = &p9824Var +var p9841Var = "thunk from from >>>" +var p9841 = &p9841Var +var p9848Var = "thunk from from >>" +var p9848 = &p9848Var +var p9866Var = "thunk from >" +var p9866 = &p9866Var +var p9877Var = "thunk from >" +var p9877 = &p9877Var +var p9887Var = "function " +var p9887 = &p9887Var +var p9893Var = "thunk from >" +var p9893 = &p9893Var +var p9898Var = "function " +var p9898 = &p9898Var +var p9939Var = "thunk from >" +var p9939 = &p9939Var +var p9959Var = "thunk from >" +var p9959 = &p9959Var +var p9966Var = "thunk from >" +var p9966 = &p9966Var +var p9975Var = "thunk from from >>" var p9975 = &p9975Var -var p9995Var = "thunk from >" -var p9995 = &p9995Var -var p10002Var = "thunk from >" -var p10002 = &p10002Var -var p10011Var = "thunk from from >>" -var p10011 = &p10011Var -var p10043Var = "thunk from >" -var p10043 = &p10043Var -var p10052Var = "thunk from from >>" -var p10052 = &p10052Var -var p10065Var = "thunk from >" +var p10007Var = "thunk from >" +var p10007 = &p10007Var +var p10016Var = "thunk from from >>" +var p10016 = &p10016Var +var p10029Var = "thunk from >" +var p10029 = &p10029Var +var p10042Var = "thunk from >>" +var p10042 = &p10042Var +var p10065Var = "thunk from >>" var p10065 = &p10065Var -var p10078Var = "thunk from >>" -var p10078 = &p10078Var -var p10101Var = "thunk from >>" -var p10101 = &p10101Var +var p10080Var = "thunk from >" +var p10080 = &p10080Var +var p10094Var = "thunk from >" +var p10094 = &p10094Var var p10116Var = "thunk from >" var p10116 = &p10116Var -var p10130Var = "thunk from >" -var p10130 = &p10130Var -var p10152Var = "thunk from >" -var p10152 = &p10152Var -var p10168Var = "thunk from >" -var p10168 = &p10168Var -var p10179Var = "thunk from >" -var p10179 = &p10179Var -var p10184Var = "function " -var p10184 = &p10184Var -var p10197Var = "thunk from >" -var p10197 = &p10197Var -var p10210Var = "thunk from >" -var p10210 = &p10210Var -var p10218Var = "object " -var p10218 = &p10218Var -var p10242Var = "thunk from >" -var p10242 = &p10242Var -var p10255Var = "thunk from >" -var p10255 = &p10255Var -var p10263Var = "object " -var p10263 = &p10263Var -var p10275Var = "object " -var p10275 = &p10275Var -var p10284Var = "thunk from >" -var p10284 = &p10284Var -var p10293Var = "thunk from from >>" -var p10293 = &p10293Var -var p10305Var = "thunk from from >>>" -var p10305 = &p10305Var -var p10338Var = "thunk from from >>" -var p10338 = &p10338Var -var p10355Var = "thunk from from >>>" -var p10355 = &p10355Var -var p10369Var = "thunk from from >>>>" +var p10132Var = "thunk from >" +var p10132 = &p10132Var +var p10143Var = "thunk from >" +var p10143 = &p10143Var +var p10148Var = "function " +var p10148 = &p10148Var +var p10161Var = "thunk from >" +var p10161 = &p10161Var +var p10174Var = "thunk from >" +var p10174 = &p10174Var +var p10182Var = "object " +var p10182 = &p10182Var +var p10206Var = "thunk from >" +var p10206 = &p10206Var +var p10219Var = "thunk from >" +var p10219 = &p10219Var +var p10227Var = "object " +var p10227 = &p10227Var +var p10239Var = "object " +var p10239 = &p10239Var +var p10248Var = "thunk from >" +var p10248 = &p10248Var +var p10257Var = "thunk from from >>" +var p10257 = &p10257Var +var p10269Var = "thunk from from >>>" +var p10269 = &p10269Var +var p10302Var = "thunk from from >>" +var p10302 = &p10302Var +var p10319Var = "thunk from from >>>" +var p10319 = &p10319Var +var p10333Var = "thunk from from >>>>" +var p10333 = &p10333Var +var p10341Var = "thunk from >" +var p10341 = &p10341Var +var p10345Var = "thunk from from >>" +var p10345 = &p10345Var +var p10351Var = "thunk from from >>>" +var p10351 = &p10351Var +var p10369Var = "thunk from >" var p10369 = &p10369Var -var p10377Var = "thunk from >" -var p10377 = &p10377Var -var p10381Var = "thunk from from >>" -var p10381 = &p10381Var -var p10387Var = "thunk from from >>>" -var p10387 = &p10387Var -var p10405Var = "thunk from >" -var p10405 = &p10405Var -var p10424Var = "thunk from >" -var p10424 = &p10424Var -var p10440Var = "thunk from >" -var p10440 = &p10440Var -var p10451Var = "thunk from >" -var p10451 = &p10451Var -var p10456Var = "function " -var p10456 = &p10456Var -var p10469Var = "thunk from >" -var p10469 = &p10469Var -var p10482Var = "thunk from >" -var p10482 = &p10482Var -var p10490Var = "object " -var p10490 = &p10490Var -var p10520Var = "thunk from >" -var p10520 = &p10520Var -var p10533Var = "thunk from >" -var p10533 = &p10533Var -var p10541Var = "object " -var p10541 = &p10541Var -var p10557Var = "object " -var p10557 = &p10557Var -var p10591Var = "thunk from from >>" -var p10591 = &p10591Var -var p10607Var = "thunk from from >>>" +var p10388Var = "thunk from >" +var p10388 = &p10388Var +var p10404Var = "thunk from >" +var p10404 = &p10404Var +var p10415Var = "thunk from >" +var p10415 = &p10415Var +var p10420Var = "function " +var p10420 = &p10420Var +var p10433Var = "thunk from >" +var p10433 = &p10433Var +var p10446Var = "thunk from >" +var p10446 = &p10446Var +var p10454Var = "object " +var p10454 = &p10454Var +var p10484Var = "thunk from >" +var p10484 = &p10484Var +var p10497Var = "thunk from >" +var p10497 = &p10497Var +var p10505Var = "object " +var p10505 = &p10505Var +var p10521Var = "object " +var p10521 = &p10521Var +var p10555Var = "thunk from from >>" +var p10555 = &p10555Var +var p10571Var = "thunk from from >>>" +var p10571 = &p10571Var +var p10585Var = "thunk from from >>>" +var p10585 = &p10585Var +var p10599Var = "thunk from from >>>>" +var p10599 = &p10599Var +var p10607Var = "thunk from >" var p10607 = &p10607Var -var p10621Var = "thunk from from >>>" -var p10621 = &p10621Var -var p10635Var = "thunk from from >>>>" -var p10635 = &p10635Var -var p10643Var = "thunk from >" -var p10643 = &p10643Var -var p10647Var = "thunk from from >>" -var p10647 = &p10647Var -var p10653Var = "thunk from from >>>" -var p10653 = &p10653Var -var p10668Var = "thunk from from >>" -var p10668 = &p10668Var -var p10680Var = "thunk from >" -var p10680 = &p10680Var -var p10695Var = "thunk from >" -var p10695 = &p10695Var -var p10704Var = "function " -var p10704 = &p10704Var -var p10720Var = "thunk from >" -var p10720 = &p10720Var -var p10737Var = "thunk from >" +var p10611Var = "thunk from from >>" +var p10611 = &p10611Var +var p10617Var = "thunk from from >>>" +var p10617 = &p10617Var +var p10632Var = "thunk from from >>" +var p10632 = &p10632Var +var p10644Var = "thunk from >" +var p10644 = &p10644Var +var p10659Var = "thunk from >" +var p10659 = &p10659Var +var p10669Var = "function " +var p10669 = &p10669Var +var p10685Var = "thunk from >" +var p10685 = &p10685Var +var p10702Var = "thunk from >" +var p10702 = &p10702Var +var p10721Var = "thunk from >" +var p10721 = &p10721Var +var p10737Var = "thunk from >>" var p10737 = &p10737Var -var p10756Var = "thunk from >" -var p10756 = &p10756Var -var p10772Var = "thunk from >>" -var p10772 = &p10772Var -var p10781Var = "thunk from >>>" -var p10781 = &p10781Var -var p10800Var = "function " -var p10800 = &p10800Var -var p10811Var = "thunk from >" +var p10746Var = "thunk from >>>" +var p10746 = &p10746Var +var p10765Var = "function " +var p10765 = &p10765Var +var p10776Var = "thunk from >" +var p10776 = &p10776Var +var p10806Var = "thunk from from >>" +var p10806 = &p10806Var +var p10811Var = "thunk from from >>>" var p10811 = &p10811Var -var p10841Var = "thunk from from >>" -var p10841 = &p10841Var -var p10846Var = "thunk from from >>>" -var p10846 = &p10846Var -var p10855Var = "thunk from from >>>>" -var p10855 = &p10855Var -var p10866Var = "thunk from from >>>>" -var p10866 = &p10866Var -var p10873Var = "thunk from >" -var p10873 = &p10873Var -var p10882Var = "thunk from from >>" -var p10882 = &p10882Var -var p10898Var = "thunk from >" -var p10898 = &p10898Var -var p10907Var = "thunk from >>" -var p10907 = &p10907Var -var p10922Var = "thunk from >" -var p10922 = &p10922Var -var p10938Var = "thunk from >" -var p10938 = &p10938Var -var p10947Var = "thunk from >>" -var p10947 = &p10947Var -var p10963Var = "thunk from >>>" -var p10963 = &p10963Var -var p10972Var = "thunk from >>>>" -var p10972 = &p10972Var -var p10988Var = "thunk from >" -var p10988 = &p10988Var -var p11004Var = "thunk from >" -var p11004 = &p11004Var -var p11013Var = "thunk from >>" -var p11013 = &p11013Var -var p11027Var = "thunk from >" -var p11027 = &p11027Var -var p11044Var = "thunk from >" -var p11044 = &p11044Var -var p11056Var = "thunk from >" -var p11056 = &p11056Var -var p11095Var = "function " -var p11095 = &p11095Var -var p11121Var = "thunk from from >>" -var p11121 = &p11121Var -var p11126Var = "thunk from from >>>" -var p11126 = &p11126Var -var p11137Var = "thunk from from >>>>" -var p11137 = &p11137Var -var p11144Var = "thunk from >" -var p11144 = &p11144Var -var p11153Var = "thunk from from >>" -var p11153 = &p11153Var -var p11165Var = "thunk from >" -var p11165 = &p11165Var -var p11173Var = "thunk from >>" -var p11173 = &p11173Var -var p11180Var = "function " -var p11180 = &p11180Var -var p11193Var = "thunk from >" -var p11193 = &p11193Var -var p11216Var = "thunk from >" +var p10820Var = "thunk from from >>>>" +var p10820 = &p10820Var +var p10831Var = "thunk from from >>>>" +var p10831 = &p10831Var +var p10838Var = "thunk from >" +var p10838 = &p10838Var +var p10847Var = "thunk from from >>" +var p10847 = &p10847Var +var p10863Var = "thunk from >" +var p10863 = &p10863Var +var p10872Var = "thunk from >>" +var p10872 = &p10872Var +var p10887Var = "thunk from >" +var p10887 = &p10887Var +var p10903Var = "thunk from >" +var p10903 = &p10903Var +var p10912Var = "thunk from >>" +var p10912 = &p10912Var +var p10928Var = "thunk from >>>" +var p10928 = &p10928Var +var p10937Var = "thunk from >>>>" +var p10937 = &p10937Var +var p10953Var = "thunk from >" +var p10953 = &p10953Var +var p10969Var = "thunk from >" +var p10969 = &p10969Var +var p10978Var = "thunk from >>" +var p10978 = &p10978Var +var p10992Var = "thunk from >" +var p10992 = &p10992Var +var p11009Var = "thunk from >" +var p11009 = &p11009Var +var p11021Var = "thunk from >" +var p11021 = &p11021Var +var p11060Var = "function " +var p11060 = &p11060Var +var p11086Var = "thunk from from >>" +var p11086 = &p11086Var +var p11091Var = "thunk from from >>>" +var p11091 = &p11091Var +var p11102Var = "thunk from from >>>>" +var p11102 = &p11102Var +var p11109Var = "thunk from >" +var p11109 = &p11109Var +var p11118Var = "thunk from from >>" +var p11118 = &p11118Var +var p11130Var = "thunk from >" +var p11130 = &p11130Var +var p11138Var = "thunk from >>" +var p11138 = &p11138Var +var p11145Var = "function " +var p11145 = &p11145Var +var p11158Var = "thunk from >" +var p11158 = &p11158Var +var p11181Var = "thunk from >" +var p11181 = &p11181Var +var p11189Var = "thunk from >" +var p11189 = &p11189Var +var p11194Var = "function " +var p11194 = &p11194Var +var p11205Var = "thunk from >" +var p11205 = &p11205Var +var p11216Var = "thunk from >" var p11216 = &p11216Var -var p11224Var = "thunk from >" -var p11224 = &p11224Var -var p11229Var = "function " -var p11229 = &p11229Var -var p11240Var = "thunk from >" -var p11240 = &p11240Var -var p11251Var = "thunk from >" +var p11226Var = "thunk from >" +var p11226 = &p11226Var +var p11239Var = "thunk from from >>" +var p11239 = &p11239Var +var p11251Var = "thunk from from >>" var p11251 = &p11251Var -var p11261Var = "thunk from >" +var p11261Var = "thunk from >" var p11261 = &p11261Var -var p11274Var = "thunk from from >>" -var p11274 = &p11274Var -var p11286Var = "thunk from from >>" -var p11286 = &p11286Var -var p11296Var = "thunk from >" -var p11296 = &p11296Var -var p11311Var = "thunk from >" +var p11276Var = "thunk from >" +var p11276 = &p11276Var +var p11311Var = "thunk from >" var p11311 = &p11311Var -var p11346Var = "thunk from >" -var p11346 = &p11346Var -var p11356Var = "thunk from from >>" -var p11356 = &p11356Var -var p11380Var = "thunk from from >>>" -var p11380 = &p11380Var -var p11385Var = "thunk from from >>>>" -var p11385 = &p11385Var -var p11402Var = "thunk from from >>>" -var p11402 = &p11402Var -var p11414Var = "thunk from >" -var p11414 = &p11414Var -var p11418Var = "thunk from >>" -var p11418 = &p11418Var -var p11439Var = "thunk from >>>" -var p11439 = &p11439Var -var p11445Var = "thunk from >>>>" -var p11445 = &p11445Var -var p11460Var = "thunk from >" -var p11460 = &p11460Var -var p11468Var = "function " -var p11468 = &p11468Var -var p11473Var = "thunk from >" -var p11473 = &p11473Var -var p11484Var = "thunk from from >>" -var p11484 = &p11484Var -var p11496Var = "thunk from from >>" -var p11496 = &p11496Var -var p11500Var = "function " -var p11500 = &p11500Var -var p11509Var = "thunk from >" -var p11509 = &p11509Var -var p11522Var = "thunk from >" -var p11522 = &p11522Var -var p11527Var = "function " -var p11527 = &p11527Var -var p11542Var = "thunk from >" -var p11542 = &p11542Var -var p11566Var = "thunk from >" -var p11566 = &p11566Var -var p11573Var = "thunk from >" -var p11573 = &p11573Var -var p11622Var = "thunk from >" -var p11622 = &p11622Var -var p11654Var = "thunk from >" -var p11654 = &p11654Var -var p11661Var = "thunk from >" -var p11661 = &p11661Var -var p11750Var = "thunk from >" -var p11750 = &p11750Var -var p11769Var = "thunk from >" -var p11769 = &p11769Var -var p11891Var = "thunk from >" -var p11891 = &p11891Var -var p11909Var = "thunk from >" -var p11909 = &p11909Var -var p11918Var = "thunk from from >>" -var p11918 = &p11918Var -var p11921Var = "function " -var p11921 = &p11921Var -var p11951Var = "thunk from >" -var p11951 = &p11951Var -var p11961Var = "function " -var p11961 = &p11961Var -var p11982Var = "thunk from >" -var p11982 = &p11982Var -var p12006Var = "thunk from >" -var p12006 = &p12006Var -var p12011Var = "function " -var p12011 = &p12011Var -var p12026Var = "thunk from >" -var p12026 = &p12026Var -var p12038Var = "thunk from >" -var p12038 = &p12038Var -var p12042Var = "thunk from from >>" -var p12042 = &p12042Var -var p12080Var = "thunk from >" -var p12080 = &p12080Var -var p12100Var = "thunk from from >>" -var p12100 = &p12100Var -var p12144Var = "thunk from >" -var p12144 = &p12144Var -var p12164Var = "thunk from from >>" -var p12164 = &p12164Var -var p12205Var = "thunk from >" -var p12205 = &p12205Var +var p11321Var = "thunk from from >>" +var p11321 = &p11321Var +var p11345Var = "thunk from from >>>" +var p11345 = &p11345Var +var p11350Var = "thunk from from >>>>" +var p11350 = &p11350Var +var p11367Var = "thunk from from >>>" +var p11367 = &p11367Var +var p11379Var = "thunk from >" +var p11379 = &p11379Var +var p11383Var = "thunk from >>" +var p11383 = &p11383Var +var p11404Var = "thunk from >>>" +var p11404 = &p11404Var +var p11410Var = "thunk from >>>>" +var p11410 = &p11410Var +var p11425Var = "thunk from >" +var p11425 = &p11425Var +var p11433Var = "function " +var p11433 = &p11433Var +var p11438Var = "thunk from >" +var p11438 = &p11438Var +var p11449Var = "thunk from from >>" +var p11449 = &p11449Var +var p11461Var = "thunk from from >>" +var p11461 = &p11461Var +var p11465Var = "function " +var p11465 = &p11465Var +var p11474Var = "thunk from >" +var p11474 = &p11474Var +var p11487Var = "thunk from >" +var p11487 = &p11487Var +var p11492Var = "function " +var p11492 = &p11492Var +var p11507Var = "thunk from >" +var p11507 = &p11507Var +var p11531Var = "thunk from >" +var p11531 = &p11531Var +var p11538Var = "thunk from >" +var p11538 = &p11538Var +var p11587Var = "thunk from >" +var p11587 = &p11587Var +var p11619Var = "thunk from >" +var p11619 = &p11619Var +var p11626Var = "thunk from >" +var p11626 = &p11626Var +var p11715Var = "thunk from >" +var p11715 = &p11715Var +var p11734Var = "thunk from >" +var p11734 = &p11734Var +var p11856Var = "thunk from >" +var p11856 = &p11856Var +var p11874Var = "thunk from >" +var p11874 = &p11874Var +var p11883Var = "thunk from from >>" +var p11883 = &p11883Var +var p11886Var = "function " +var p11886 = &p11886Var +var p11916Var = "thunk from >" +var p11916 = &p11916Var +var p11926Var = "function " +var p11926 = &p11926Var +var p11947Var = "thunk from >" +var p11947 = &p11947Var +var p11971Var = "thunk from >" +var p11971 = &p11971Var +var p11976Var = "function " +var p11976 = &p11976Var +var p11991Var = "thunk from >" +var p11991 = &p11991Var +var p12003Var = "thunk from >" +var p12003 = &p12003Var +var p12007Var = "thunk from from >>" +var p12007 = &p12007Var +var p12045Var = "thunk from >" +var p12045 = &p12045Var +var p12065Var = "thunk from from >>" +var p12065 = &p12065Var +var p12109Var = "thunk from >" +var p12109 = &p12109Var +var p12129Var = "thunk from from >>" +var p12129 = &p12129Var +var p12170Var = "thunk from >" +var p12170 = &p12170Var +var p12198Var = "thunk from >" +var p12198 = &p12198Var +var p12208Var = "function " +var p12208 = &p12208Var +var p12212Var = "thunk from >" +var p12212 = &p12212Var +var p12221Var = "thunk from from >>" +var p12221 = &p12221Var var p12233Var = "thunk from >" var p12233 = &p12233Var -var p12243Var = "function " +var p12243Var = "thunk from >>" var p12243 = &p12243Var -var p12247Var = "thunk from >" +var p12247Var = "function " var p12247 = &p12247Var -var p12256Var = "thunk from from >>" +var p12256Var = "thunk from >" var p12256 = &p12256Var -var p12268Var = "thunk from >" -var p12268 = &p12268Var -var p12278Var = "thunk from >>" -var p12278 = &p12278Var -var p12282Var = "function " -var p12282 = &p12282Var +var p12266Var = "function " +var p12266 = &p12266Var +var p12270Var = "thunk from >" +var p12270 = &p12270Var +var p12279Var = "thunk from from >>" +var p12279 = &p12279Var var p12291Var = "thunk from >" var p12291 = &p12291Var -var p12301Var = "function " -var p12301 = &p12301Var -var p12305Var = "thunk from >" -var p12305 = &p12305Var -var p12314Var = "thunk from from >>" -var p12314 = &p12314Var -var p12326Var = "thunk from >" -var p12326 = &p12326Var -var p12332Var = "function " -var p12332 = &p12332Var -var p12350Var = "function " -var p12350 = &p12350Var -var p12357Var = "thunk from >" -var p12357 = &p12357Var -var p12361Var = "function " -var p12361 = &p12361Var -var p12368Var = "thunk from >" -var p12368 = &p12368Var -var p12377Var = "thunk from from >>" -var p12377 = &p12377Var -var p12393Var = "thunk from >" -var p12393 = &p12393Var -var p12405Var = "thunk from >" -var p12405 = &p12405Var -var p12411Var = "thunk from >" -var p12411 = &p12411Var -var p12417Var = "thunk from from >>" -var p12417 = &p12417Var -var p12428Var = "thunk from >" -var p12428 = &p12428Var -var p12437Var = "thunk from from >>" -var p12437 = &p12437Var -var p12446Var = "function " -var p12446 = &p12446Var -var p12474Var = "thunk from >" -var p12474 = &p12474Var -var p12483Var = "thunk from from >>" -var p12483 = &p12483Var -var p12487Var = "function " -var p12487 = &p12487Var -var p12495Var = "thunk from >" -var p12495 = &p12495Var -var p12506Var = "thunk from >" -var p12506 = &p12506Var -var p12515Var = "thunk from from >>" -var p12515 = &p12515Var -var p12519Var = "function " +var p12297Var = "function " +var p12297 = &p12297Var +var p12316Var = "function " +var p12316 = &p12316Var +var p12323Var = "thunk from >" +var p12323 = &p12323Var +var p12328Var = "function " +var p12328 = &p12328Var +var p12335Var = "thunk from >" +var p12335 = &p12335Var +var p12344Var = "thunk from from >>" +var p12344 = &p12344Var +var p12360Var = "thunk from >" +var p12360 = &p12360Var +var p12372Var = "thunk from >" +var p12372 = &p12372Var +var p12378Var = "thunk from >" +var p12378 = &p12378Var +var p12384Var = "thunk from from >>" +var p12384 = &p12384Var +var p12395Var = "thunk from >" +var p12395 = &p12395Var +var p12404Var = "thunk from from >>" +var p12404 = &p12404Var +var p12413Var = "function " +var p12413 = &p12413Var +var p12441Var = "thunk from >" +var p12441 = &p12441Var +var p12450Var = "thunk from from >>" +var p12450 = &p12450Var +var p12454Var = "function " +var p12454 = &p12454Var +var p12462Var = "thunk from >" +var p12462 = &p12462Var +var p12473Var = "thunk from >" +var p12473 = &p12473Var +var p12482Var = "thunk from from >>" +var p12482 = &p12482Var +var p12486Var = "function " +var p12486 = &p12486Var +var p12494Var = "thunk from >" +var p12494 = &p12494Var +var p12511Var = "thunk from >" +var p12511 = &p12511Var +var p12519Var = "thunk from >" var p12519 = &p12519Var -var p12527Var = "thunk from >" -var p12527 = &p12527Var -var p12544Var = "thunk from >" -var p12544 = &p12544Var -var p12552Var = "thunk from >" -var p12552 = &p12552Var -var p12564Var = "thunk from >" -var p12564 = &p12564Var -var p12573Var = "thunk from >" -var p12573 = &p12573Var -var p12578Var = "function " -var p12578 = &p12578Var -var p12582Var = "thunk from >" +var p12531Var = "thunk from >" +var p12531 = &p12531Var +var p12540Var = "thunk from >" +var p12540 = &p12540Var +var p12545Var = "function " +var p12545 = &p12545Var +var p12549Var = "thunk from >" +var p12549 = &p12549Var +var p12558Var = "thunk from from >>" +var p12558 = &p12558Var +var p12561Var = "thunk from >" +var p12561 = &p12561Var +var p12570Var = "thunk from from >>" +var p12570 = &p12570Var +var p12577Var = "thunk from >" +var p12577 = &p12577Var +var p12582Var = "function " var p12582 = &p12582Var -var p12591Var = "thunk from from >>" -var p12591 = &p12591Var -var p12594Var = "thunk from >" -var p12594 = &p12594Var -var p12603Var = "thunk from from >>" -var p12603 = &p12603Var -var p12610Var = "thunk from >" -var p12610 = &p12610Var -var p12615Var = "function " -var p12615 = &p12615Var -var p12682Var = "thunk from >" -var p12682 = &p12682Var -var p12694Var = "thunk from >" -var p12694 = &p12694Var -var p12707Var = "thunk from >" -var p12707 = &p12707Var -var p12722Var = "thunk from >>" -var p12722 = &p12722Var -var p12736Var = "thunk from >" -var p12736 = &p12736Var -var p12751Var = "thunk from >>" -var p12751 = &p12751Var -var p12764Var = "thunk from >" +var p12649Var = "thunk from >" +var p12649 = &p12649Var +var p12661Var = "thunk from >" +var p12661 = &p12661Var +var p12674Var = "thunk from >" +var p12674 = &p12674Var +var p12689Var = "thunk from >>" +var p12689 = &p12689Var +var p12703Var = "thunk from >" +var p12703 = &p12703Var +var p12718Var = "thunk from >>" +var p12718 = &p12718Var +var p12731Var = "thunk from >" +var p12731 = &p12731Var +var p12739Var = "thunk from >" +var p12739 = &p12739Var +var p12748Var = "thunk from from >>" +var p12748 = &p12748Var +var p12764Var = "thunk from >" var p12764 = &p12764Var -var p12772Var = "thunk from >" -var p12772 = &p12772Var -var p12781Var = "thunk from from >>" -var p12781 = &p12781Var -var p12797Var = "thunk from >" -var p12797 = &p12797Var -var p12807Var = "thunk from >" -var p12807 = &p12807Var -var p12818Var = "thunk from >" -var p12818 = &p12818Var -var p12827Var = "thunk from from >>" +var p12774Var = "thunk from >" +var p12774 = &p12774Var +var p12785Var = "thunk from >" +var p12785 = &p12785Var +var p12794Var = "thunk from from >>" +var p12794 = &p12794Var +var p12812Var = "thunk from >" +var p12812 = &p12812Var +var p12827Var = "thunk from >" var p12827 = &p12827Var -var p12845Var = "thunk from >" -var p12845 = &p12845Var -var p12860Var = "thunk from >" -var p12860 = &p12860Var -var p12873Var = "thunk from >" -var p12873 = &p12873Var -var p12882Var = "thunk from >>" -var p12882 = &p12882Var -var p12896Var = "thunk from >>" -var p12896 = &p12896Var -var p12906Var = "function " -var p12906 = &p12906Var -var p12913Var = "thunk from >" -var p12913 = &p12913Var -var p12918Var = "function " -var p12918 = &p12918Var -var p12931Var = "thunk from >" -var p12931 = &p12931Var -var p12939Var = "thunk from >" -var p12939 = &p12939Var -var p12952Var = "thunk from >" -var p12952 = &p12952Var -var p12967Var = "thunk from >>" -var p12967 = &p12967Var -var p12976Var = "thunk from >" -var p12976 = &p12976Var -var p12991Var = "thunk from >" -var p12991 = &p12991Var -var p13003Var = "thunk from >" +var p12840Var = "thunk from >" +var p12840 = &p12840Var +var p12849Var = "thunk from >>" +var p12849 = &p12849Var +var p12863Var = "thunk from >>" +var p12863 = &p12863Var +var p12874Var = "function " +var p12874 = &p12874Var +var p12881Var = "thunk from >" +var p12881 = &p12881Var +var p12886Var = "function " +var p12886 = &p12886Var +var p12899Var = "thunk from >" +var p12899 = &p12899Var +var p12907Var = "thunk from >" +var p12907 = &p12907Var +var p12920Var = "thunk from >" +var p12920 = &p12920Var +var p12935Var = "thunk from >>" +var p12935 = &p12935Var +var p12944Var = "thunk from >" +var p12944 = &p12944Var +var p12959Var = "thunk from >" +var p12959 = &p12959Var +var p12971Var = "thunk from >" +var p12971 = &p12971Var +var p12982Var = "function " +var p12982 = &p12982Var +var p12994Var = "thunk from >" +var p12994 = &p12994Var +var p13003Var = "thunk from >>" var p13003 = &p13003Var -var p13013Var = "function " -var p13013 = &p13013Var -var p13025Var = "thunk from >" -var p13025 = &p13025Var -var p13034Var = "thunk from >>" -var p13034 = &p13034Var -var p13045Var = "function " -var p13045 = &p13045Var -var p13060Var = "thunk from >" -var p13060 = &p13060Var -var p13069Var = "thunk from >>" +var p13015Var = "function " +var p13015 = &p13015Var +var p13030Var = "thunk from >" +var p13030 = &p13030Var +var p13039Var = "thunk from >>" +var p13039 = &p13039Var +var p13043Var = "thunk from >>>" +var p13043 = &p13043Var +var p13056Var = "function " +var p13056 = &p13056Var +var p13064Var = "thunk from >" +var p13064 = &p13064Var +var p13069Var = "function " var p13069 = &p13069Var -var p13073Var = "thunk from >>>" -var p13073 = &p13073Var -var p13085Var = "function " -var p13085 = &p13085Var -var p13093Var = "thunk from >" -var p13093 = &p13093Var -var p13098Var = "function " -var p13098 = &p13098Var -var p13113Var = "thunk from >" -var p13113 = &p13113Var -var p13150Var = "thunk from >" -var p13150 = &p13150Var -var p13177Var = "thunk from >" -var p13177 = &p13177Var -var p13183Var = "thunk from from >>" -var p13183 = &p13183Var -var p13194Var = "thunk from >" -var p13194 = &p13194Var -var p13200Var = "thunk from from >>" -var p13200 = &p13200Var -var p13222Var = "thunk from >" -var p13222 = &p13222Var -var p13244Var = "thunk from >>" -var p13244 = &p13244Var -var p13266Var = "thunk from >" -var p13266 = &p13266Var -var p13285Var = "thunk from >>" -var p13285 = &p13285Var -var p13299Var = "thunk from >" -var p13299 = &p13299Var -var p13318Var = "thunk from >>" -var p13318 = &p13318Var -var p13331Var = "thunk from >" -var p13331 = &p13331Var -var p13343Var = "function " -var p13343 = &p13343Var -var p13350Var = "thunk from >" -var p13350 = &p13350Var -var p13355Var = "function " -var p13355 = &p13355Var -var p13372Var = "thunk from >" -var p13372 = &p13372Var -var p13387Var = "thunk from >" -var p13387 = &p13387Var -var p13404Var = "thunk from >" -var p13404 = &p13404Var -var p13416Var = "thunk from >" -var p13416 = &p13416Var -var p13429Var = "thunk from >" -var p13429 = &p13429Var -var p13451Var = "thunk from >>" -var p13451 = &p13451Var -var p13468Var = "thunk from >" -var p13468 = &p13468Var -var p13480Var = "thunk from >" -var p13480 = &p13480Var -var p13493Var = "thunk from >" -var p13493 = &p13493Var -var p13514Var = "thunk from >" -var p13514 = &p13514Var -var p13534Var = "thunk from >" -var p13534 = &p13534Var -var p13546Var = "function " +var p13084Var = "thunk from >" +var p13084 = &p13084Var +var p13121Var = "thunk from >" +var p13121 = &p13121Var +var p13148Var = "thunk from >" +var p13148 = &p13148Var +var p13154Var = "thunk from from >>" +var p13154 = &p13154Var +var p13165Var = "thunk from >" +var p13165 = &p13165Var +var p13171Var = "thunk from from >>" +var p13171 = &p13171Var +var p13193Var = "thunk from >" +var p13193 = &p13193Var +var p13215Var = "thunk from >>" +var p13215 = &p13215Var +var p13237Var = "thunk from >" +var p13237 = &p13237Var +var p13256Var = "thunk from >>" +var p13256 = &p13256Var +var p13270Var = "thunk from >" +var p13270 = &p13270Var +var p13289Var = "thunk from >>" +var p13289 = &p13289Var +var p13302Var = "thunk from >" +var p13302 = &p13302Var +var p13315Var = "function " +var p13315 = &p13315Var +var p13322Var = "thunk from >" +var p13322 = &p13322Var +var p13327Var = "function " +var p13327 = &p13327Var +var p13344Var = "thunk from >" +var p13344 = &p13344Var +var p13359Var = "thunk from >" +var p13359 = &p13359Var +var p13376Var = "thunk from >" +var p13376 = &p13376Var +var p13388Var = "thunk from >" +var p13388 = &p13388Var +var p13401Var = "thunk from >" +var p13401 = &p13401Var +var p13423Var = "thunk from >>" +var p13423 = &p13423Var +var p13440Var = "thunk from >" +var p13440 = &p13440Var +var p13452Var = "thunk from >" +var p13452 = &p13452Var +var p13465Var = "thunk from >" +var p13465 = &p13465Var +var p13486Var = "thunk from >" +var p13486 = &p13486Var +var p13506Var = "thunk from >" +var p13506 = &p13506Var +var p13519Var = "function " +var p13519 = &p13519Var +var p13526Var = "thunk from >" +var p13526 = &p13526Var +var p13531Var = "function " +var p13531 = &p13531Var +var p13546Var = "thunk from >" var p13546 = &p13546Var -var p13553Var = "thunk from >" -var p13553 = &p13553Var -var p13558Var = "function " -var p13558 = &p13558Var -var p13573Var = "thunk from >" -var p13573 = &p13573Var -var p13594Var = "thunk from >" -var p13594 = &p13594Var -var p13627Var = "thunk from >" -var p13627 = &p13627Var -var p13639Var = "thunk from >" -var p13639 = &p13639Var +var p13567Var = "thunk from >" +var p13567 = &p13567Var +var p13600Var = "thunk from >" +var p13600 = &p13600Var +var p13612Var = "thunk from >" +var p13612 = &p13612Var +var p13625Var = "thunk from >" +var p13625 = &p13625Var var p13652Var = "thunk from >" var p13652 = &p13652Var -var p13679Var = "thunk from >" -var p13679 = &p13679Var -var p13691Var = "thunk from >" -var p13691 = &p13691Var -var p13704Var = "thunk from >" -var p13704 = &p13704Var -var p13723Var = "thunk from >>" -var p13723 = &p13723Var -var p13737Var = "thunk from >" -var p13737 = &p13737Var -var p13757Var = "thunk from >" -var p13757 = &p13757Var -var p13770Var = "function " -var p13770 = &p13770Var -var p13781Var = "thunk from >" -var p13781 = &p13781Var -var p13789Var = "thunk from >" -var p13789 = &p13789Var -var p13800Var = "thunk from from >>" -var p13800 = &p13800Var -var p13811Var = "thunk from >" -var p13811 = &p13811Var -var p13822Var = "thunk from from >>" -var p13822 = &p13822Var -var p13833Var = "thunk from from >>" -var p13833 = &p13833Var -var p13854Var = "thunk from >" -var p13854 = &p13854Var -var p13867Var = "thunk from from >>" -var p13867 = &p13867Var -var p13879Var = "thunk from from >>" +var p13664Var = "thunk from >" +var p13664 = &p13664Var +var p13677Var = "thunk from >" +var p13677 = &p13677Var +var p13696Var = "thunk from >>" +var p13696 = &p13696Var +var p13710Var = "thunk from >" +var p13710 = &p13710Var +var p13730Var = "thunk from >" +var p13730 = &p13730Var +var p13743Var = "function " +var p13743 = &p13743Var +var p13754Var = "thunk from >" +var p13754 = &p13754Var +var p13762Var = "thunk from >" +var p13762 = &p13762Var +var p13773Var = "thunk from from >>" +var p13773 = &p13773Var +var p13784Var = "thunk from >" +var p13784 = &p13784Var +var p13795Var = "thunk from from >>" +var p13795 = &p13795Var +var p13806Var = "thunk from from >>" +var p13806 = &p13806Var +var p13827Var = "thunk from >" +var p13827 = &p13827Var +var p13840Var = "thunk from from >>" +var p13840 = &p13840Var +var p13852Var = "thunk from from >>" +var p13852 = &p13852Var +var p13859Var = "thunk from >" +var p13859 = &p13859Var +var p13868Var = "thunk from from >>" +var p13868 = &p13868Var +var p13879Var = "thunk from from >>>" var p13879 = &p13879Var -var p13886Var = "thunk from >" -var p13886 = &p13886Var -var p13895Var = "thunk from from >>" -var p13895 = &p13895Var -var p13906Var = "thunk from from >>>" -var p13906 = &p13906Var -var p13938Var = "object " -var p13938 = &p13938Var -var p13951Var = "thunk from >" -var p13951 = &p13951Var -var p13976Var = "thunk from >" -var p13976 = &p13976Var -var p13990Var = "thunk from >" -var p13990 = &p13990Var -var p14008Var = "thunk from >" -var p14008 = &p14008Var -var p14029Var = "thunk from >" -var p14029 = &p14029Var -var p14042Var = "function " -var p14042 = &p14042Var -var p14052Var = "thunk from >" -var p14052 = &p14052Var -var p14060Var = "function " -var p14060 = &p14060Var -var p14070Var = "thunk from >" -var p14070 = &p14070Var -var p14078Var = "function " -var p14078 = &p14078Var -var p14088Var = "thunk from >" -var p14088 = &p14088Var -var p14098Var = "function " -var p14098 = &p14098Var -var p14108Var = "thunk from >" -var p14108 = &p14108Var -var p14132Var = "thunk from >" -var p14132 = &p14132Var -var p14139Var = "function " -var p14139 = &p14139Var -var p14148Var = "thunk from >" -var p14148 = &p14148Var -var p14169Var = "thunk from >" -var p14169 = &p14169Var -var p14176Var = "function " -var p14176 = &p14176Var -var p14185Var = "thunk from >" -var p14185 = &p14185Var -var p14193Var = "function " -var p14193 = &p14193Var -var p14197Var = "thunk from >" -var p14197 = &p14197Var -var p14206Var = "thunk from from >>" -var p14206 = &p14206Var -var p14213Var = "thunk from >" -var p14213 = &p14213Var -var p14222Var = "thunk from from >>" -var p14222 = &p14222Var -var p14238Var = "thunk from >" -var p14238 = &p14238Var -var p14257Var = "thunk from >" -var p14257 = &p14257Var -var p14265Var = "thunk from >" -var p14265 = &p14265Var -var p14274Var = "thunk from from >>" -var p14274 = &p14274Var -var p14290Var = "thunk from >" -var p14290 = &p14290Var -var p14301Var = "thunk from >>" -var p14301 = &p14301Var -var p14311Var = "thunk from >" -var p14311 = &p14311Var -var p14316Var = "function " -var p14316 = &p14316Var -var p14353Var = "thunk from >" -var p14353 = &p14353Var -var p14369Var = "thunk from >" -var p14369 = &p14369Var -var p14386Var = "thunk from >" -var p14386 = &p14386Var -var p14394Var = "thunk from >" -var p14394 = &p14394Var -var p14403Var = "thunk from from >>" -var p14403 = &p14403Var -var p14410Var = "thunk from >" -var p14410 = &p14410Var -var p14419Var = "thunk from from >>" -var p14419 = &p14419Var -var p14437Var = "thunk from >" -var p14437 = &p14437Var -var p14447Var = "thunk from >" -var p14447 = &p14447Var -var p14452Var = "function " -var p14452 = &p14452Var -var p14469Var = "thunk from >" -var p14469 = &p14469Var -var p14499Var = "thunk from >" -var p14499 = &p14499Var -var p14515Var = "thunk from >" -var p14515 = &p14515Var -var p14531Var = "thunk from >" -var p14531 = &p14531Var -var p14541Var = "function " -var p14541 = &p14541Var -var p14545Var = "thunk from >" -var p14545 = &p14545Var -var p14554Var = "thunk from from >>" -var p14554 = &p14554Var -var p14567Var = "thunk from >" -var p14567 = &p14567Var -var p14579Var = "thunk from >>" -var p14579 = &p14579Var -var p14590Var = "thunk from >>>" -var p14590 = &p14590Var -var p14597Var = "function " -var p14597 = &p14597Var -var p14607Var = "thunk from >>" -var p14607 = &p14607Var -var p14615Var = "function " -var p14615 = &p14615Var -var p14619Var = "thunk from >" -var p14619 = &p14619Var -var p14624Var = "function " -var p14624 = &p14624Var -var p14645Var = "thunk from >" -var p14645 = &p14645Var -var p14659Var = "thunk from >" -var p14659 = &p14659Var -var p14674Var = "thunk from >" -var p14674 = &p14674Var -var p14688Var = "thunk from >" -var p14688 = &p14688Var -var p14706Var = "thunk from >" -var p14706 = &p14706Var -var p14727Var = "thunk from >" -var p14727 = &p14727Var -var p14736Var = "thunk from >>" -var p14736 = &p14736Var -var p14742Var = "thunk from >" -var p14742 = &p14742Var -var p14751Var = "thunk from >>" -var p14751 = &p14751Var -var p14768Var = "thunk from >" -var p14768 = &p14768Var -var p14797Var = "thunk from >" -var p14797 = &p14797Var -var p14806Var = "thunk from >>" -var p14806 = &p14806Var -var p14821Var = "object " -var p14821 = &p14821Var -var p14830Var = "thunk from >" -var p14830 = &p14830Var -var p14846Var = "thunk from >" -var p14846 = &p14846Var -var p14857Var = "function " -var p14857 = &p14857Var -var p14870Var = "thunk from >" -var p14870 = &p14870Var -var p14887Var = "thunk from >" -var p14887 = &p14887Var -var p14903Var = "thunk from >" -var p14903 = &p14903Var -var p14920Var = "thunk from >" -var p14920 = &p14920Var -var p14928Var = "thunk from >" -var p14928 = &p14928Var -var p14937Var = "thunk from from >>" -var p14937 = &p14937Var -var p14944Var = "thunk from >" -var p14944 = &p14944Var -var p14953Var = "thunk from from >>" -var p14953 = &p14953Var -var p14991Var = "thunk from >" -var p14991 = &p14991Var -var p14995Var = "function " -var p14995 = &p14995Var -var p15027Var = "thunk from >>" +var p13911Var = "object " +var p13911 = &p13911Var +var p13924Var = "thunk from >" +var p13924 = &p13924Var +var p13949Var = "thunk from >" +var p13949 = &p13949Var +var p13963Var = "thunk from >" +var p13963 = &p13963Var +var p13981Var = "thunk from >" +var p13981 = &p13981Var +var p14002Var = "thunk from >" +var p14002 = &p14002Var +var p14015Var = "function " +var p14015 = &p14015Var +var p14025Var = "thunk from >" +var p14025 = &p14025Var +var p14033Var = "function " +var p14033 = &p14033Var +var p14043Var = "thunk from >" +var p14043 = &p14043Var +var p14051Var = "function " +var p14051 = &p14051Var +var p14061Var = "thunk from >" +var p14061 = &p14061Var +var p14071Var = "function " +var p14071 = &p14071Var +var p14081Var = "thunk from >" +var p14081 = &p14081Var +var p14092Var = "function " +var p14092 = &p14092Var +var p14096Var = "thunk from >" +var p14096 = &p14096Var +var p14105Var = "thunk from from >>" +var p14105 = &p14105Var +var p14112Var = "thunk from >" +var p14112 = &p14112Var +var p14121Var = "thunk from from >>" +var p14121 = &p14121Var +var p14137Var = "thunk from >" +var p14137 = &p14137Var +var p14156Var = "thunk from >" +var p14156 = &p14156Var +var p14164Var = "thunk from >" +var p14164 = &p14164Var +var p14173Var = "thunk from from >>" +var p14173 = &p14173Var +var p14189Var = "thunk from >" +var p14189 = &p14189Var +var p14200Var = "thunk from >>" +var p14200 = &p14200Var +var p14210Var = "thunk from >" +var p14210 = &p14210Var +var p14215Var = "function " +var p14215 = &p14215Var +var p14252Var = "thunk from >" +var p14252 = &p14252Var +var p14268Var = "thunk from >" +var p14268 = &p14268Var +var p14285Var = "thunk from >" +var p14285 = &p14285Var +var p14293Var = "thunk from >" +var p14293 = &p14293Var +var p14302Var = "thunk from from >>" +var p14302 = &p14302Var +var p14309Var = "thunk from >" +var p14309 = &p14309Var +var p14318Var = "thunk from from >>" +var p14318 = &p14318Var +var p14336Var = "thunk from >" +var p14336 = &p14336Var +var p14346Var = "thunk from >" +var p14346 = &p14346Var +var p14351Var = "function " +var p14351 = &p14351Var +var p14368Var = "thunk from >" +var p14368 = &p14368Var +var p14398Var = "thunk from >" +var p14398 = &p14398Var +var p14414Var = "thunk from >" +var p14414 = &p14414Var +var p14430Var = "thunk from >" +var p14430 = &p14430Var +var p14440Var = "function " +var p14440 = &p14440Var +var p14444Var = "thunk from >" +var p14444 = &p14444Var +var p14453Var = "thunk from from >>" +var p14453 = &p14453Var +var p14466Var = "thunk from >" +var p14466 = &p14466Var +var p14478Var = "thunk from >>" +var p14478 = &p14478Var +var p14489Var = "thunk from >>>" +var p14489 = &p14489Var +var p14496Var = "function " +var p14496 = &p14496Var +var p14506Var = "thunk from >>" +var p14506 = &p14506Var +var p14514Var = "function " +var p14514 = &p14514Var +var p14518Var = "thunk from >" +var p14518 = &p14518Var +var p14523Var = "function " +var p14523 = &p14523Var +var p14544Var = "thunk from >" +var p14544 = &p14544Var +var p14558Var = "thunk from >" +var p14558 = &p14558Var +var p14573Var = "thunk from >" +var p14573 = &p14573Var +var p14587Var = "thunk from >" +var p14587 = &p14587Var +var p14605Var = "thunk from >" +var p14605 = &p14605Var +var p14626Var = "thunk from >" +var p14626 = &p14626Var +var p14635Var = "thunk from >>" +var p14635 = &p14635Var +var p14641Var = "thunk from >" +var p14641 = &p14641Var +var p14650Var = "thunk from >>" +var p14650 = &p14650Var +var p14667Var = "thunk from >" +var p14667 = &p14667Var +var p14696Var = "thunk from >" +var p14696 = &p14696Var +var p14705Var = "thunk from >>" +var p14705 = &p14705Var +var p14720Var = "object " +var p14720 = &p14720Var +var p14729Var = "thunk from >" +var p14729 = &p14729Var +var p14745Var = "thunk from >" +var p14745 = &p14745Var +var p14756Var = "function " +var p14756 = &p14756Var +var p14769Var = "thunk from >" +var p14769 = &p14769Var +var p14786Var = "thunk from >" +var p14786 = &p14786Var +var p14802Var = "thunk from >" +var p14802 = &p14802Var +var p14819Var = "thunk from >" +var p14819 = &p14819Var +var p14827Var = "thunk from >" +var p14827 = &p14827Var +var p14836Var = "thunk from from >>" +var p14836 = &p14836Var +var p14843Var = "thunk from >" +var p14843 = &p14843Var +var p14852Var = "thunk from from >>" +var p14852 = &p14852Var +var p14890Var = "thunk from >" +var p14890 = &p14890Var +var p14894Var = "function " +var p14894 = &p14894Var +var p14926Var = "thunk from >>" +var p14926 = &p14926Var +var p14939Var = "function " +var p14939 = &p14939Var +var p14952Var = "thunk from >" +var p14952 = &p14952Var +var p14969Var = "thunk from >" +var p14969 = &p14969Var +var p14982Var = "thunk from >" +var p14982 = &p14982Var +var p14986Var = "function " +var p14986 = &p14986Var +var p15005Var = "thunk from >>" +var p15005 = &p15005Var +var p15017Var = "thunk from >>>" +var p15017 = &p15017Var +var p15022Var = "object " +var p15022 = &p15022Var +var p15024Var = "object " +var p15024 = &p15024Var +var p15027Var = "function " var p15027 = &p15027Var -var p15040Var = "function " -var p15040 = &p15040Var -var p15053Var = "thunk from >" +var p15030Var = "object " +var p15030 = &p15030Var +var p15036Var = "function " +var p15036 = &p15036Var +var p15039Var = "function " +var p15039 = &p15039Var +var p15042Var = "function " +var p15042 = &p15042Var +var p15045Var = "function " +var p15045 = &p15045Var +var p15047Var = "function " +var p15047 = &p15047Var +var p15050Var = "function " +var p15050 = &p15050Var +var p15053Var = "function " var p15053 = &p15053Var -var p15070Var = "thunk from >" -var p15070 = &p15070Var -var p15083Var = "thunk from >" -var p15083 = &p15083Var -var p15087Var = "function " -var p15087 = &p15087Var -var p15106Var = "thunk from >>" -var p15106 = &p15106Var -var p15118Var = "thunk from >>>" -var p15118 = &p15118Var -var p15127Var = "function " -var p15127 = &p15127Var -var p15131Var = "thunk from >" +var p15057Var = "function " +var p15057 = &p15057Var +var p15061Var = "thunk from >" +var p15061 = &p15061Var +var p15064Var = "thunk from >" +var p15064 = &p15064Var +var p15071Var = "thunk from from >>" +var p15071 = &p15071Var +var p15074Var = "function " +var p15074 = &p15074Var +var p15078Var = "thunk from >" +var p15078 = &p15078Var +var p15081Var = "thunk from >" +var p15081 = &p15081Var +var p15088Var = "thunk from from >>" +var p15088 = &p15088Var +var p15091Var = "function " +var p15091 = &p15091Var +var p15095Var = "thunk from >" +var p15095 = &p15095Var +var p15098Var = "thunk from >" +var p15098 = &p15098Var +var p15105Var = "thunk from from >>" +var p15105 = &p15105Var +var p15108Var = "function " +var p15108 = &p15108Var +var p15112Var = "thunk from >" +var p15112 = &p15112Var +var p15117Var = "function " +var p15117 = &p15117Var +var p15121Var = "thunk from >" +var p15121 = &p15121Var +var p15124Var = "thunk from >" +var p15124 = &p15124Var +var p15131Var = "thunk from from >>" var p15131 = &p15131Var -var p15140Var = "thunk from from >>" -var p15140 = &p15140Var -var p15143Var = "thunk from >" -var p15143 = &p15143Var -var p15152Var = "thunk from from >>" -var p15152 = &p15152Var -var p15196Var = "thunk from >" -var p15196 = &p15196Var -var p15262Var = "function " +var p15135Var = "function " +var p15135 = &p15135Var +var p15139Var = "thunk from >" +var p15139 = &p15139Var +var p15142Var = "thunk from >" +var p15142 = &p15142Var +var p15145Var = "thunk from >" +var p15145 = &p15145Var +var p15148Var = "thunk from >" +var p15148 = &p15148Var +var p15151Var = "thunk from >" +var p15151 = &p15151Var +var p15154Var = "thunk from >" +var p15154 = &p15154Var +var p15158Var = "thunk from >" +var p15158 = &p15158Var +var p15161Var = "thunk from >" +var p15161 = &p15161Var +var p15164Var = "thunk from >" +var p15164 = &p15164Var +var p15167Var = "thunk from >" +var p15167 = &p15167Var +var p15170Var = "thunk from >" +var p15170 = &p15170Var +var p15173Var = "thunk from >" +var p15173 = &p15173Var +var p15176Var = "thunk from >" +var p15176 = &p15176Var +var p15179Var = "thunk from >" +var p15179 = &p15179Var +var p15182Var = "thunk from >" +var p15182 = &p15182Var +var p15186Var = "thunk from >" +var p15186 = &p15186Var +var p15189Var = "thunk from >" +var p15189 = &p15189Var +var p15192Var = "thunk from >" +var p15192 = &p15192Var +var p15195Var = "thunk from >" +var p15195 = &p15195Var +var p15199Var = "thunk from >" +var p15199 = &p15199Var +var p15202Var = "thunk from >" +var p15202 = &p15202Var +var p15207Var = "function " +var p15207 = &p15207Var +var p15210Var = "function " +var p15210 = &p15210Var +var p15213Var = "function " +var p15213 = &p15213Var +var p15216Var = "function " +var p15216 = &p15216Var +var p15218Var = "function " +var p15218 = &p15218Var +var p15221Var = "function " +var p15221 = &p15221Var +var p15224Var = "function " +var p15224 = &p15224Var +var p15227Var = "function " +var p15227 = &p15227Var +var p15230Var = "function " +var p15230 = &p15230Var +var p15234Var = "function " +var p15234 = &p15234Var +var p15237Var = "function " +var p15237 = &p15237Var +var p15240Var = "function " +var p15240 = &p15240Var +var p15253Var = "function " +var p15253 = &p15253Var +var p15255Var = "function " +var p15255 = &p15255Var +var p15259Var = "thunk from >" +var p15259 = &p15259Var +var p15262Var = "thunk from >" var p15262 = &p15262Var -var p15266Var = "thunk from >" -var p15266 = &p15266Var -var p15275Var = "thunk from from >>" -var p15275 = &p15275Var -var p15278Var = "thunk from >" -var p15278 = &p15278Var -var p15287Var = "thunk from from >>" -var p15287 = &p15287Var -var p15294Var = "thunk from >" -var p15294 = &p15294Var -var p15303Var = "thunk from from >>" -var p15303 = &p15303Var -var p15312Var = "thunk from >" -var p15312 = &p15312Var -var p15317Var = "function " -var p15317 = &p15317Var -var p15330Var = "thunk from >" -var p15330 = &p15330Var -var p15339Var = "thunk from from >>" -var p15339 = &p15339Var -var p15370Var = "thunk from >" -var p15370 = &p15370Var -var p15386Var = "thunk from >" -var p15386 = &p15386Var -var p15397Var = "thunk from >" -var p15397 = &p15397Var -var p15403Var = "function " -var p15403 = &p15403Var -var p15414Var = "thunk from >" -var p15414 = &p15414Var -var p15425Var = "function " -var p15425 = &p15425Var -var p15436Var = "thunk from >" -var p15436 = &p15436Var -var p15446Var = "function " -var p15446 = &p15446Var -var p15457Var = "thunk from >" -var p15457 = &p15457Var -var p15467Var = "function " -var p15467 = &p15467Var -var p15478Var = "thunk from >" -var p15478 = &p15478Var -var p15485Var = "object " -var p15485 = &p15485Var -var p15487Var = "object " -var p15487 = &p15487Var -var p15490Var = "function " -var p15490 = &p15490Var -var p15493Var = "object " -var p15493 = &p15493Var -var p15499Var = "function " -var p15499 = &p15499Var -var p15502Var = "function " -var p15502 = &p15502Var -var p15505Var = "function " -var p15505 = &p15505Var -var p15508Var = "function " -var p15508 = &p15508Var -var p15510Var = "function " -var p15510 = &p15510Var -var p15513Var = "function " -var p15513 = &p15513Var -var p15516Var = "function " -var p15516 = &p15516Var -var p15520Var = "function " -var p15520 = &p15520Var -var p15524Var = "thunk from >" -var p15524 = &p15524Var -var p15527Var = "thunk from >" -var p15527 = &p15527Var -var p15534Var = "thunk from from >>" -var p15534 = &p15534Var -var p15537Var = "function " -var p15537 = &p15537Var -var p15541Var = "thunk from >" -var p15541 = &p15541Var -var p15544Var = "thunk from >" -var p15544 = &p15544Var -var p15551Var = "thunk from from >>" -var p15551 = &p15551Var -var p15554Var = "function " -var p15554 = &p15554Var -var p15558Var = "thunk from >" -var p15558 = &p15558Var -var p15561Var = "thunk from >" -var p15561 = &p15561Var -var p15568Var = "thunk from from >>" -var p15568 = &p15568Var -var p15571Var = "function " -var p15571 = &p15571Var -var p15575Var = "thunk from >" -var p15575 = &p15575Var -var p15580Var = "function " -var p15580 = &p15580Var -var p15584Var = "thunk from >" -var p15584 = &p15584Var -var p15587Var = "thunk from >" -var p15587 = &p15587Var -var p15594Var = "thunk from from >>" -var p15594 = &p15594Var -var p15598Var = "function " -var p15598 = &p15598Var -var p15602Var = "thunk from >" -var p15602 = &p15602Var -var p15605Var = "thunk from >" -var p15605 = &p15605Var -var p15608Var = "thunk from >" -var p15608 = &p15608Var -var p15611Var = "thunk from >" -var p15611 = &p15611Var -var p15614Var = "thunk from >" -var p15614 = &p15614Var -var p15617Var = "thunk from >" -var p15617 = &p15617Var -var p15621Var = "thunk from >" -var p15621 = &p15621Var -var p15624Var = "thunk from >" -var p15624 = &p15624Var -var p15627Var = "thunk from >" -var p15627 = &p15627Var -var p15630Var = "thunk from >" -var p15630 = &p15630Var -var p15633Var = "thunk from >" -var p15633 = &p15633Var -var p15636Var = "thunk from >" -var p15636 = &p15636Var -var p15639Var = "thunk from >" -var p15639 = &p15639Var -var p15642Var = "thunk from >" -var p15642 = &p15642Var -var p15645Var = "thunk from >" -var p15645 = &p15645Var -var p15649Var = "thunk from >" -var p15649 = &p15649Var -var p15652Var = "thunk from >" -var p15652 = &p15652Var -var p15655Var = "thunk from >" -var p15655 = &p15655Var -var p15658Var = "thunk from >" -var p15658 = &p15658Var -var p15662Var = "thunk from >" -var p15662 = &p15662Var -var p15665Var = "thunk from >" -var p15665 = &p15665Var -var p15670Var = "function " -var p15670 = &p15670Var -var p15673Var = "function " -var p15673 = &p15673Var -var p15676Var = "function " -var p15676 = &p15676Var -var p15679Var = "function " -var p15679 = &p15679Var -var p15681Var = "function " -var p15681 = &p15681Var -var p15684Var = "function " -var p15684 = &p15684Var -var p15687Var = "function " -var p15687 = &p15687Var -var p15690Var = "function " -var p15690 = &p15690Var -var p15693Var = "function " -var p15693 = &p15693Var -var p15697Var = "function " -var p15697 = &p15697Var -var p15700Var = "function " -var p15700 = &p15700Var -var p15703Var = "function " -var p15703 = &p15703Var -var p15716Var = "function " -var p15716 = &p15716Var -var p15718Var = "function " -var p15718 = &p15718Var -var p15722Var = "thunk from >" -var p15722 = &p15722Var -var p15725Var = "thunk from >" -var p15725 = &p15725Var -var p15728Var = "function " -var p15728 = &p15728Var -var p15731Var = "function " -var p15731 = &p15731Var -var p15739Var = "thunk from >" -var p15739 = &p15739Var -var p15742Var = "thunk from >" -var p15742 = &p15742Var -var p15745Var = "thunk from >" -var p15745 = &p15745Var -var p15752Var = "thunk from >>" -var p15752 = &p15752Var -var p15755Var = "thunk from >" -var p15755 = &p15755Var -var p15767Var = "function " -var p15767 = &p15767Var -var p15769Var = "function " -var p15769 = &p15769Var -var p15772Var = "object " -var p15772 = &p15772Var -var p15796Var = "object " -var p15796 = &p15796Var -var p15800Var = "object " -var p15800 = &p15800Var -var p15803Var = "object " -var p15803 = &p15803Var -var p15806Var = "object " -var p15806 = &p15806Var -var p15809Var = "object " -var p15809 = &p15809Var -var p15812Var = "object " -var p15812 = &p15812Var -var p15815Var = "object " -var p15815 = &p15815Var -var p15822Var = "thunk from >" -var p15822 = &p15822Var -var p15824Var = "thunk from >" -var p15824 = &p15824Var +var p15265Var = "function " +var p15265 = &p15265Var +var p15268Var = "function " +var p15268 = &p15268Var +var p15276Var = "thunk from >" +var p15276 = &p15276Var +var p15279Var = "thunk from >" +var p15279 = &p15279Var +var p15282Var = "thunk from >" +var p15282 = &p15282Var +var p15289Var = "thunk from >>" +var p15289 = &p15289Var +var p15292Var = "thunk from >" +var p15292 = &p15292Var +var p15304Var = "function " +var p15304 = &p15304Var +var p15306Var = "function " +var p15306 = &p15306Var +var p15309Var = "object " +var p15309 = &p15309Var +var p15333Var = "object " +var p15333 = &p15333Var +var p15337Var = "object " +var p15337 = &p15337Var +var p15340Var = "object " +var p15340 = &p15340Var +var p15343Var = "object " +var p15343 = &p15343Var +var p15346Var = "object " +var p15346 = &p15346Var +var p15349Var = "object " +var p15349 = &p15349Var +var p15352Var = "object " +var p15352 = &p15352Var +var p15359Var = "thunk from >" +var p15359 = &p15359Var +var p15361Var = "thunk from >" +var p15361 = &p15361Var var p1 = &ast.Source{ Lines: []string{ "/*\n", @@ -2420,7 +2352,7 @@ var p1 = &ast.Source{ " assert std.isString(str) : 'substr first parameter should be a string, got ' + std.type(str);\n", " assert std.isNumber(from) : 'substr second parameter should be a string, got ' + std.type(from);\n", " assert std.isNumber(len) : 'substr third parameter should be a string, got ' + std.type(len);\n", - " assert len >= 0 : 'substr third parameter should be greater than zero, got ' + len;\n", + " assert len >=0 : 'substr third parameter should be greater than zero, got ' + len;\n", " std.join('', std.makeArray(std.max(0, std.min(len, std.length(str) - from)), function(i) str[i + from])),\n", "\n", " startsWith(a, b)::\n", @@ -2493,13 +2425,13 @@ var p1 = &ast.Source{ " split(str, c)::\n", " assert std.isString(str) : 'std.split first parameter should be a string, got ' + std.type(str);\n", " assert std.isString(c) : 'std.split second parameter should be a string, got ' + std.type(c);\n", - " assert std.length(c) == 1 : 'std.split second parameter should have length 1, got ' + std.length(c);\n", + " assert std.length(c) == 1 : 'std.split second parameter should be a string, got ' + std.type(c);\n", " std.splitLimit(str, c, -1),\n", "\n", " splitLimit(str, c, maxsplits)::\n", " assert std.isString(str) : 'std.splitLimit first parameter should be a string, got ' + std.type(str);\n", " assert std.isString(c) : 'std.splitLimit second parameter should be a string, got ' + std.type(c);\n", - " assert std.length(c) == 1 : 'std.splitLimit second parameter should have length 1, got ' + std.length(c);\n", + " assert std.length(c) == 1 : 'std.splitLimit second parameter should have length 1, got ' + std.length(c);\n", " assert std.isNumber(maxsplits) : 'std.splitLimit third parameter should be a number, got ' + std.type(maxsplits);\n", " local aux(str, delim, i, arr, v) =\n", " local c = str[i];\n", @@ -2564,9 +2496,9 @@ var p1 = &ast.Source{ "\n", " repeat(what, count)::\n", " local joiner =\n", - " if std.isString(what) then ''\n", + " if std.isString(what) then \"\"\n", " else if std.isArray(what) then []\n", - " else error 'std.repeat first argument must be an array or a string';\n", + " else error \"std.repeat first argument must be an array or a string\";\n", " std.join(joiner, std.makeArray(count, function(i) what)),\n", "\n", " slice(indexable, index, end, step)::\n", @@ -2607,7 +2539,7 @@ var p1 = &ast.Source{ " std.count(arr, x) > 0\n", " else if std.isString(arr) then\n", " std.length(std.findSubstr(x, arr)) > 0\n", - " else error 'std.member first argument must be an array or a string',\n", + " else error \"std.member first argument must be an array or a string\",\n", "\n", " count(arr, x):: std.length(std.filter(function(v) v == x, arr)),\n", "\n", @@ -2721,10 +2653,10 @@ var p1 = &ast.Source{ " else if c == ' ' then\n", " consume(str, j + 1, v { blank: true })\n", " else if c == '+' then\n", - " consume(str, j + 1, v { plus: true })\n", + " consume(str, j + 1, v { sign: true })\n", " else\n", " { i: j, v: v };\n", - " consume(str, i, { alt: false, zero: false, left: false, blank: false, plus: false });\n", + " consume(str, i, { alt: false, zero: false, left: false, blank: false, sign: false });\n", "\n", " local try_parse_field_width(str, i) =\n", " if i < std.length(str) && str[i] == '*' then\n", @@ -2864,38 +2796,23 @@ var p1 = &ast.Source{ " local pad_right(str, w, s) =\n", " str + padding(w - std.length(str), s);\n", "\n", - " // Render a sign & magnitude integer (radix ranges from decimal to binary).\n", - " // neg should be a boolean, and when true indicates that we should render a negative number.\n", - " // mag must always be a whole number >= 0, it's the magnitude of the integer to render\n", - " // min_chars must be a whole number >= 0\n", - " // It is the field width, i.e. std.length() of the result should be >= min_chars\n", - " // min_digits must be a whole number >= 0. It's the number of zeroes to pad with.\n", - " // blank must be a boolean, if true adds an additional ' ' in front of a positive number, so\n", - " // that it is aligned with negative numbers with the same number of digits.\n", - " // plus must be a boolean, if true adds a '+' in front of a postive number, so that it is\n", - " // aligned with negative numbers with the same number of digits. This takes precedence over\n", - " // blank, if both are true.\n", - " // radix must be a whole number >1 and <= 10. It is the base of the system of numerals.\n", - " // zero_prefix is a string prefixed before the sign to all numbers that are not 0.\n", - " local render_int(neg, mag, min_chars, min_digits, blank, plus, radix, zero_prefix) =\n", - " // dec is the minimal string needed to represent the number as text.\n", - " local dec =\n", - " if mag == 0 then\n", - " '0'\n", + " // Render an integer (e.g., decimal or octal).\n", + " local render_int(n__, min_chars, min_digits, blank, sign, radix, zero_prefix) =\n", + " local n_ = std.abs(n__);\n", + " local aux(n) =\n", + " if n == 0 then\n", + " zero_prefix\n", " else\n", - " local aux(n) =\n", - " if n == 0 then\n", - " zero_prefix\n", - " else\n", - " aux(std.floor(n / radix)) + (n % radix);\n", - " aux(mag);\n", - " local zp = min_chars - (if neg || blank || plus then 1 else 0);\n", + " aux(std.floor(n / radix)) + (n % radix);\n", + " local dec = if std.floor(n_) == 0 then '0' else aux(std.floor(n_));\n", + " local neg = n__ < 0;\n", + " local zp = min_chars - (if neg || blank || sign then 1 else 0);\n", " local zp2 = std.max(zp, min_digits);\n", " local dec2 = pad_left(dec, zp2, '0');\n", - " (if neg then '-' else if plus then '+' else if blank then ' ' else '') + dec2;\n", + " (if neg then '-' else if sign then '+' else if blank then ' ' else '') + dec2;\n", "\n", " // Render an integer in hexadecimal.\n", - " local render_hex(n__, min_chars, min_digits, blank, plus, add_zerox, capitals) =\n", + " local render_hex(n__, min_chars, min_digits, blank, sign, add_zerox, capitals) =\n", " local numerals = [0, 1, 2, 3, 4, 5, 6, 7, 8, 9]\n", " + if capitals then ['A', 'B', 'C', 'D', 'E', 'F']\n", " else ['a', 'b', 'c', 'd', 'e', 'f'];\n", @@ -2907,12 +2824,12 @@ var p1 = &ast.Source{ " aux(std.floor(n / 16)) + numerals[n % 16];\n", " local hex = if std.floor(n_) == 0 then '0' else aux(std.floor(n_));\n", " local neg = n__ < 0;\n", - " local zp = min_chars - (if neg || blank || plus then 1 else 0)\n", + " local zp = min_chars - (if neg || blank || sign then 1 else 0)\n", " - (if add_zerox then 2 else 0);\n", " local zp2 = std.max(zp, min_digits);\n", " local hex2 = (if add_zerox then (if capitals then '0X' else '0x') else '')\n", " + pad_left(hex, zp2, '0');\n", - " (if neg then '-' else if plus then '+' else if blank then ' ' else '') + hex2;\n", + " (if neg then '-' else if sign then '+' else if blank then ' ' else '') + hex2;\n", "\n", " local strip_trailing_zero(str) =\n", " local aux(str, i) =\n", @@ -2926,27 +2843,27 @@ var p1 = &ast.Source{ " aux(str, std.length(str) - 1);\n", "\n", " // Render floating point in decimal form\n", - " local render_float_dec(n__, zero_pad, blank, plus, ensure_pt, trailing, prec) =\n", + " local render_float_dec(n__, zero_pad, blank, sign, ensure_pt, trailing, prec) =\n", " local n_ = std.abs(n__);\n", " local whole = std.floor(n_);\n", " local dot_size = if prec == 0 && !ensure_pt then 0 else 1;\n", " local zp = zero_pad - prec - dot_size;\n", - " local str = render_int(n__ < 0, whole, zp, 0, blank, plus, 10, '');\n", + " local str = render_int(std.sign(n__) * whole, zp, 0, blank, sign, 10, '');\n", " if prec == 0 then\n", " str + if ensure_pt then '.' else ''\n", " else\n", " local frac = std.floor((n_ - whole) * std.pow(10, prec) + 0.5);\n", " if trailing || frac > 0 then\n", - " local frac_str = render_int(false, frac, prec, 0, false, false, 10, '');\n", + " local frac_str = render_int(frac, prec, 0, false, false, 10, '');\n", " str + '.' + if !trailing then strip_trailing_zero(frac_str) else frac_str\n", " else\n", " str;\n", "\n", " // Render floating point in scientific form\n", - " local render_float_sci(n__, zero_pad, blank, plus, ensure_pt, trailing, caps, prec) =\n", + " local render_float_sci(n__, zero_pad, blank, sign, ensure_pt, trailing, caps, prec) =\n", " local exponent = if n__ == 0 then 0 else std.floor(std.log(std.abs(n__)) / std.log(10));\n", " local suff = (if caps then 'E' else 'e')\n", - " + render_int(exponent < 0, std.abs(exponent), 3, 0, false, true, 10, '');\n", + " + render_int(exponent, 3, 0, false, true, 10, '');\n", " local mantissa = if exponent == -324 then\n", " // Avoid a rounding error where std.pow(10, -324) is 0\n", " // -324 is the smallest exponent possible.\n", @@ -2954,7 +2871,7 @@ var p1 = &ast.Source{ " else\n", " n__ / std.pow(10, exponent);\n", " local zp2 = zero_pad - std.length(suff);\n", - " render_float_dec(mantissa, zp2, blank, plus, ensure_pt, trailing, prec) + suff;\n", + " render_float_dec(mantissa, zp2, blank, sign, ensure_pt, trailing, prec) + suff;\n", "\n", " // Render a value with an arbitrary format code.\n", " local format_code(val, code, fw, prec_or_null, i) =\n", @@ -2969,24 +2886,24 @@ var p1 = &ast.Source{ " error 'Format required number at '\n", " + i + ', got ' + std.type(val)\n", " else\n", - " render_int(val <= -1, std.floor(std.abs(val)), zp, iprec, cflags.blank, cflags.plus, 10, '')\n", + " render_int(val, zp, iprec, cflags.blank, cflags.sign, 10, '')\n", " else if code.ctype == 'o' then\n", " if std.type(val) != 'number' then\n", " error 'Format required number at '\n", " + i + ', got ' + std.type(val)\n", " else\n", " local zero_prefix = if cflags.alt then '0' else '';\n", - " render_int(val <= -1, std.floor(std.abs(val)), zp, iprec, cflags.blank, cflags.plus, 8, zero_prefix)\n", + " render_int(val, zp, iprec, cflags.blank, cflags.sign, 8, zero_prefix)\n", " else if code.ctype == 'x' then\n", " if std.type(val) != 'number' then\n", " error 'Format required number at '\n", " + i + ', got ' + std.type(val)\n", " else\n", - " render_hex(std.floor(val),\n", + " render_hex(val,\n", " zp,\n", " iprec,\n", " cflags.blank,\n", - " cflags.plus,\n", + " cflags.sign,\n", " cflags.alt,\n", " code.caps)\n", " else if code.ctype == 'f' then\n", @@ -2997,7 +2914,7 @@ var p1 = &ast.Source{ " render_float_dec(val,\n", " zp,\n", " cflags.blank,\n", - " cflags.plus,\n", + " cflags.sign,\n", " cflags.alt,\n", " true,\n", " fpprec)\n", @@ -3009,7 +2926,7 @@ var p1 = &ast.Source{ " render_float_sci(val,\n", " zp,\n", " cflags.blank,\n", - " cflags.plus,\n", + " cflags.sign,\n", " cflags.alt,\n", " true,\n", " code.caps,\n", @@ -3024,7 +2941,7 @@ var p1 = &ast.Source{ " render_float_sci(val,\n", " zp,\n", " cflags.blank,\n", - " cflags.plus,\n", + " cflags.sign,\n", " cflags.alt,\n", " cflags.alt,\n", " code.caps,\n", @@ -3034,7 +2951,7 @@ var p1 = &ast.Source{ " render_float_dec(val,\n", " zp,\n", " cflags.blank,\n", - " cflags.plus,\n", + " cflags.sign,\n", " cflags.alt,\n", " cflags.alt,\n", " fpprec - digits_before_pt)\n", @@ -3220,7 +3137,7 @@ var p1 = &ast.Source{ " if a < b then a else b,\n", "\n", " clamp(x, minVal, maxVal)::\n", - " if x < minVal then minVal\n", + " if x < minVal then minVal\n", " else if x > maxVal then maxVal\n", " else x,\n", "\n", @@ -3682,12 +3599,6 @@ var p1 = &ast.Source{ " objectHasAll(o, f)::\n", " std.objectHasEx(o, f, true),\n", "\n", - " objectValues(o)::\n", - " [o[k] for k in std.objectFields(o)],\n", - "\n", - " objectValuesAll(o)::\n", - " [o[k] for k in std.objectFieldsAll(o)],\n", - "\n", " equals(a, b)::\n", " local ta = std.type(a);\n", " local tb = std.type(b);\n", @@ -3766,44 +3677,9 @@ var p1 = &ast.Source{ " error 'find second parameter should be an array, got ' + std.type(arr)\n", " else\n", " std.filter(function(i) arr[i] == value, std.range(0, std.length(arr) - 1)),\n", - "\n", - " // Three way comparison.\n", - " // TODO(sbarzowski): consider exposing and documenting it properly\n", - " __compare(v1, v2)::\n", - " local t1 = std.type(v1), t2 = std.type(v2);\n", - " if t1 != t2 then\n", - " error \"Comparison requires matching types. Got \" + t1 + \" and \" + t2\n", - " else if t1 == \"array\" then\n", - " std.__compare_array(v1, v2)\n", - " else if t1 == \"function\" || t1 == \"object\" || t1 == \"bool\" then\n", - " error \"Values of type \" + t1 + \" are not comparable.\"\n", - " else if v1 < v2 then -1\n", - " else if v1 > v2 then 1\n", - " else 0,\n", - "\n", - " __compare_array(arr1, arr2)::\n", - " local len1 = std.length(arr1), len2 = std.length(arr2);\n", - " local minLen = std.min(len1, len2);\n", - " local aux(i) =\n", - " if i < minLen then\n", - " local cmpRes = std.__compare(arr1[i], arr2[i]);\n", - " if cmpRes != 0 then\n", - " cmpRes\n", - " else\n", - " aux(i + 1) tailstrict\n", - " else\n", - " std.__compare(len1, len2);\n", - " aux(0),\n", - "\n", - " __array_less(arr1, arr2):: std.__compare_array(arr1, arr2) == -1,\n", - " __array_greater(arr1, arr2):: std.__compare_array(arr1, arr2) == 1,\n", - " __array_less_or_equal(arr1, arr2):: std.__compare_array(arr1, arr2) <= 0,\n", - " __array_greater_or_equal(arr1, arr2):: std.__compare_array(arr1, arr2) >= 0,\n", - "\n", "}\n", "\n", }, - DiagnosticFileName: "", } // StdAst is the AST for the standard library. @@ -3811,24 +3687,18 @@ var StdAst = _StdAst var _StdAst = &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(23), Column: int(1), }, End: ast.Location{ - Line: int(1422), + Line: int(1367), Column: int(2), }, File: p1, }, Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(0), - Comment: []string{}, - }, ast.FodderElement{ Kind: ast.FodderKind(2), Blanks: int(1), @@ -3841,7 +3711,7 @@ var _StdAst = &ast.DesugaredObject{ "you may not use this file except in compliance with the License.", "You may obtain a copy of the License at", "", - " http://www.apache.org/licenses/LICENSE-2.0", + " http://www.apache.org/licenses/LICENSE-2.0", "", "Unless required by applicable law or agreed to in writing, software", "distributed under the License is distributed on an \"AS IS\" BASIS,", @@ -3857,15 +3727,15 @@ var _StdAst = &ast.DesugaredObject{ Indent: int(0), Comment: []string{ "/* This is the Jsonnet standard library, at least the parts of it that are written in Jsonnet.", - " *", - " * There are some native methods as well, which are defined in the interpreter and added to this", - " * file. It is never necessary to import std.jsonnet, it is embedded into the interpreter at", - " * compile-time and automatically imported into all other Jsonnet programs.", - " */", + "*", + "* There are some native methods as well, which are defined in the interpreter and added to this", + "* file. It is never necessary to import std.jsonnet, it is embedded into the interpreter at", + "* compile-time and automatically imported into all other Jsonnet programs.", + "*/", }, }, }, - Ctx: p7, + Ctx: p6, FreeVars: nil, }, Asserts: ast.Nodes{}, @@ -3893,7 +3763,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -3916,33 +3785,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(28), - Column: int(12), - }, - End: ast.Location{ - Line: int(28), - Column: int(13), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(28), Column: int(17), @@ -3963,7 +3821,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(28), Column: int(17), @@ -3984,7 +3842,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(28), Column: int(17), @@ -4004,7 +3862,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(28), Column: int(17), @@ -4045,9 +3903,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -4057,7 +3914,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(28), Column: int(26), @@ -4091,7 +3948,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(28), Column: int(32), @@ -4109,23 +3966,10 @@ var _StdAst = &ast.DesugaredObject{ Value: "string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(28), - Column: int(3), - }, - End: ast.Location{ - Line: int(28), - Column: int(40), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -4150,7 +3994,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -4173,33 +4016,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(29), - Column: int(12), - }, - End: ast.Location{ - Line: int(29), - Column: int(13), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(29), Column: int(17), @@ -4220,7 +4052,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(29), Column: int(17), @@ -4241,7 +4073,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(29), Column: int(17), @@ -4261,7 +4093,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(29), Column: int(17), @@ -4302,9 +4134,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -4314,7 +4145,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(29), Column: int(26), @@ -4348,7 +4179,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(29), Column: int(32), @@ -4366,23 +4197,10 @@ var _StdAst = &ast.DesugaredObject{ Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(29), - Column: int(3), - }, - End: ast.Location{ - Line: int(29), - Column: int(40), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -4407,7 +4225,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "isBoolean", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -4430,33 +4247,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(30), - Column: int(13), - }, - End: ast.Location{ - Line: int(30), - Column: int(14), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(30), Column: int(18), @@ -4477,7 +4283,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(30), Column: int(18), @@ -4498,7 +4304,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(30), Column: int(18), @@ -4518,7 +4324,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(30), Column: int(18), @@ -4559,9 +4365,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -4571,7 +4376,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(30), Column: int(27), @@ -4605,7 +4410,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(30), Column: int(33), @@ -4623,23 +4428,10 @@ var _StdAst = &ast.DesugaredObject{ Value: "boolean", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(30), - Column: int(3), - }, - End: ast.Location{ - Line: int(30), - Column: int(42), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -4664,7 +4456,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -4687,33 +4478,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(31), - Column: int(12), - }, - End: ast.Location{ - Line: int(31), - Column: int(13), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(31), Column: int(17), @@ -4734,7 +4514,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(31), Column: int(17), @@ -4755,7 +4535,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(31), Column: int(17), @@ -4775,7 +4555,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(31), Column: int(17), @@ -4816,9 +4596,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -4828,7 +4607,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(31), Column: int(26), @@ -4862,7 +4641,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(31), Column: int(32), @@ -4880,23 +4659,10 @@ var _StdAst = &ast.DesugaredObject{ Value: "object", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(31), - Column: int(3), - }, - End: ast.Location{ - Line: int(31), - Column: int(40), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -4921,7 +4687,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -4944,33 +4709,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(32), - Column: int(11), - }, - End: ast.Location{ - Line: int(32), - Column: int(12), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(32), Column: int(16), @@ -4991,7 +4745,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(32), Column: int(16), @@ -5012,7 +4766,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(32), Column: int(16), @@ -5032,7 +4786,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(32), Column: int(16), @@ -5073,9 +4827,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -5085,7 +4838,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(32), Column: int(25), @@ -5119,7 +4872,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(32), Column: int(31), @@ -5137,23 +4890,10 @@ var _StdAst = &ast.DesugaredObject{ Value: "array", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(32), - Column: int(3), - }, - End: ast.Location{ - Line: int(32), - Column: int(38), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -5178,7 +4918,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -5201,33 +4940,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(33), - Column: int(14), - }, - End: ast.Location{ - Line: int(33), - Column: int(15), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(33), Column: int(19), @@ -5248,7 +4976,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(33), Column: int(19), @@ -5269,7 +4997,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(33), Column: int(19), @@ -5289,7 +5017,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(33), Column: int(19), @@ -5330,9 +5058,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -5342,7 +5069,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(33), Column: int(28), @@ -5376,7 +5103,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(33), Column: int(34), @@ -5394,23 +5121,10 @@ var _StdAst = &ast.DesugaredObject{ Value: "function", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(33), - Column: int(3), - }, - End: ast.Location{ - Line: int(33), - Column: int(44), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -5435,7 +5149,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "toString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -5458,33 +5171,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(35), - Column: int(12), - }, - End: ast.Location{ - Line: int(35), - Column: int(13), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(5), @@ -5512,7 +5214,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(8), @@ -5533,7 +5235,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(8), @@ -5554,7 +5256,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(8), @@ -5574,7 +5276,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(8), @@ -5615,9 +5317,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -5627,7 +5328,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(17), @@ -5661,7 +5362,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(23), @@ -5679,14 +5380,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(37), @@ -5709,7 +5409,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(44), @@ -5729,7 +5429,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(44), @@ -5747,14 +5447,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(36), Column: int(49), @@ -5777,18 +5476,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(35), - Column: int(3), - }, - End: ast.Location{ - Line: int(36), - Column: int(50), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -5813,7 +5500,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -5836,64 +5522,25 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(38), - Column: int(10), - }, - End: ast.Location{ - Line: int(38), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "from", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(38), - Column: int(15), - }, - End: ast.Location{ - Line: int(38), - Column: int(19), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "len", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(38), - Column: int(21), - }, - End: ast.Location{ - Line: int(38), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "from", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "len", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -5923,7 +5570,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(12), @@ -5944,7 +5591,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(12), @@ -5964,7 +5611,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(12), @@ -6005,9 +5652,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -6017,7 +5663,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(25), @@ -6073,7 +5719,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(12), @@ -6094,7 +5740,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(12), @@ -6114,7 +5760,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(12), @@ -6155,9 +5801,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -6167,7 +5812,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(25), @@ -6223,7 +5868,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(12), @@ -6244,7 +5889,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(12), @@ -6264,7 +5909,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(12), @@ -6305,9 +5950,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -6317,7 +5961,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(25), @@ -6373,14 +6017,14 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(42), Column: int(12), }, End: ast.Location{ Line: int(42), - Column: int(20), + Column: int(19), }, File: p1, }, @@ -6393,7 +6037,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(42), Column: int(12), @@ -6417,14 +6061,14 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(42), - Column: int(19), + Column: int(18), }, End: ast.Location{ Line: int(42), - Column: int(20), + Column: int(19), }, File: p1, }, @@ -6432,6 +6076,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p160, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -6439,7 +6084,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(5), @@ -6462,7 +6107,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(5), @@ -6482,7 +6127,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(5), @@ -6530,9 +6175,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -6542,7 +6186,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(14), @@ -6560,7 +6204,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -6568,7 +6211,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(18), @@ -6591,7 +6234,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(18), @@ -6611,7 +6254,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(18), @@ -6652,9 +6295,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -6664,7 +6306,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(32), @@ -6687,7 +6329,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(32), @@ -6707,7 +6349,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(32), @@ -6748,9 +6390,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "max", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -6760,7 +6401,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(40), @@ -6775,6 +6416,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p233, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -6783,7 +6425,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(43), @@ -6806,7 +6448,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(43), @@ -6826,7 +6468,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(43), @@ -6867,9 +6509,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "min", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -6879,7 +6520,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(51), @@ -6904,7 +6545,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(56), @@ -6926,7 +6567,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(56), @@ -6947,7 +6588,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(56), @@ -6967,7 +6608,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(56), @@ -7008,9 +6649,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -7020,7 +6660,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(67), @@ -7054,7 +6694,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(74), @@ -7100,7 +6740,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(82), @@ -7119,33 +6759,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(43), - Column: int(91), - }, - End: ast.Location{ - Line: int(43), - Column: int(92), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(94), @@ -7167,7 +6796,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(94), @@ -7190,7 +6819,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(98), @@ -7211,7 +6840,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(98), @@ -7235,7 +6864,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(43), Column: int(102), @@ -7285,14 +6914,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(42), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(43), - Column: int(109), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -7303,14 +6932,14 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(42), - Column: int(23), + Column: int(22), }, End: ast.Location{ Line: int(42), - Column: int(87), + Column: int(86), }, File: p1, }, @@ -7323,14 +6952,14 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(42), - Column: int(23), + Column: int(22), }, End: ast.Location{ Line: int(42), - Column: int(81), + Column: int(80), }, File: p1, }, @@ -7341,21 +6970,20 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr third parameter should be greater than zero, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(42), - Column: int(84), + Column: int(83), }, End: ast.Location{ Line: int(42), - Column: int(87), + Column: int(86), }, File: p1, }, @@ -7376,14 +7004,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(41), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(43), - Column: int(109), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -7395,7 +7023,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(32), @@ -7416,7 +7044,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(32), @@ -7434,14 +7062,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr third parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(84), @@ -7462,7 +7089,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(84), @@ -7482,7 +7109,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(84), @@ -7523,9 +7150,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -7535,7 +7161,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(41), Column: int(93), @@ -7573,14 +7199,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(40), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(43), - Column: int(109), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -7592,7 +7218,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(33), @@ -7613,7 +7239,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(33), @@ -7631,14 +7257,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr second parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(86), @@ -7659,7 +7284,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(86), @@ -7679,7 +7304,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(86), @@ -7720,9 +7345,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -7732,7 +7356,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(40), Column: int(95), @@ -7770,14 +7394,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(39), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(43), - Column: int(109), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -7789,7 +7413,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(32), @@ -7810,7 +7434,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(32), @@ -7828,14 +7452,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr first parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(84), @@ -7856,7 +7479,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(84), @@ -7876,7 +7499,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(84), @@ -7917,9 +7540,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -7929,7 +7551,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(39), Column: int(93), @@ -7963,18 +7585,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(38), - Column: int(3), - }, - End: ast.Location{ - Line: int(43), - Column: int(109), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -7999,7 +7609,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "startsWith", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -8022,52 +7631,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(45), - Column: int(14), - }, - End: ast.Location{ - Line: int(45), - Column: int(15), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(45), - Column: int(17), - }, - End: ast.Location{ - Line: int(45), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(5), @@ -8096,7 +7680,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(8), @@ -8118,7 +7702,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(8), @@ -8139,7 +7723,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(8), @@ -8159,7 +7743,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(8), @@ -8200,9 +7784,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -8212,7 +7795,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(19), @@ -8246,7 +7829,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(24), @@ -8267,7 +7850,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(24), @@ -8287,7 +7870,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(24), @@ -8328,9 +7911,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -8340,7 +7922,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(46), Column: int(35), @@ -8374,7 +7956,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(47), Column: int(7), @@ -8409,7 +7991,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(7), @@ -8431,7 +8013,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(7), @@ -8453,7 +8035,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(7), @@ -8473,7 +8055,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(7), @@ -8521,9 +8103,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -8533,7 +8114,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(18), @@ -8558,7 +8139,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(21), @@ -8573,6 +8154,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p376, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -8581,7 +8163,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(24), @@ -8602,7 +8184,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(24), @@ -8622,7 +8204,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(24), @@ -8663,9 +8245,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -8675,7 +8256,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(35), @@ -8719,7 +8300,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(49), Column: int(42), @@ -8742,18 +8323,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(45), - Column: int(3), - }, - End: ast.Location{ - Line: int(49), - Column: int(43), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -8778,7 +8347,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "endsWith", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -8801,52 +8369,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(51), - Column: int(12), - }, - End: ast.Location{ - Line: int(51), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(51), - Column: int(15), - }, - End: ast.Location{ - Line: int(51), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(5), @@ -8875,7 +8418,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(8), @@ -8897,7 +8440,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(8), @@ -8918,7 +8461,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(8), @@ -8938,7 +8481,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(8), @@ -8979,9 +8522,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -8991,7 +8533,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(19), @@ -9025,7 +8567,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(24), @@ -9046,7 +8588,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(24), @@ -9066,7 +8608,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(24), @@ -9107,9 +8649,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -9119,7 +8660,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(52), Column: int(35), @@ -9153,7 +8694,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(53), Column: int(7), @@ -9188,7 +8729,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(7), @@ -9210,7 +8751,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(7), @@ -9232,7 +8773,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(7), @@ -9252,7 +8793,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(7), @@ -9300,9 +8841,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -9312,7 +8852,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(18), @@ -9337,7 +8877,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(21), @@ -9359,7 +8899,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(21), @@ -9380,7 +8920,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(21), @@ -9400,7 +8940,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(21), @@ -9441,9 +8981,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -9453,7 +8992,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(32), @@ -9487,7 +9026,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(37), @@ -9508,7 +9047,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(37), @@ -9528,7 +9067,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(37), @@ -9569,9 +9108,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -9581,7 +9119,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(48), @@ -9617,7 +9155,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(52), @@ -9638,7 +9176,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(52), @@ -9658,7 +9196,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(52), @@ -9699,9 +9237,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -9711,7 +9248,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(63), @@ -9755,7 +9292,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(55), Column: int(70), @@ -9778,18 +9315,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(51), - Column: int(3), - }, - End: ast.Location{ - Line: int(55), - Column: int(71), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -9814,7 +9339,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "lstripChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -9837,52 +9361,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(57), - Column: int(15), - }, - End: ast.Location{ - Line: int(57), - Column: int(18), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "chars", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(57), - Column: int(20), - }, - End: ast.Location{ - Line: int(57), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "chars", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(5), @@ -9911,7 +9410,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(8), @@ -9933,7 +9432,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(8), @@ -9954,7 +9453,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(8), @@ -9975,7 +9474,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(8), @@ -9995,7 +9494,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(8), @@ -10036,9 +9535,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -10048,7 +9546,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(19), @@ -10082,7 +9580,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(26), @@ -10097,6 +9595,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p484, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -10105,7 +9604,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(31), @@ -10127,7 +9626,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(31), @@ -10147,7 +9646,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(31), @@ -10188,9 +9687,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "member", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -10200,7 +9698,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(42), @@ -10225,7 +9723,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(49), @@ -10245,7 +9743,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(49), @@ -10268,7 +9766,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(58), Column: int(53), @@ -10283,6 +9781,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p511, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, RightBracketFodder: ast.Fodder{}, @@ -10303,7 +9802,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(59), Column: int(7), @@ -10325,7 +9824,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(59), Column: int(7), @@ -10345,7 +9844,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(59), Column: int(7), @@ -10393,9 +9892,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "lstripChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -10407,14 +9905,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(59), - Column: int(23), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(59), - Column: int(30), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -10487,7 +9985,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -10499,7 +9996,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(59), Column: int(23), @@ -10524,7 +10021,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(59), Column: int(27), @@ -10539,6 +10036,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p536, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, CommaFodder: nil, @@ -10601,7 +10099,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(59), Column: int(32), @@ -10641,7 +10139,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(61), Column: int(7), @@ -10670,18 +10168,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(57), - Column: int(3), - }, - End: ast.Location{ - Line: int(61), - Column: int(10), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -10706,7 +10192,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "rstripChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -10729,52 +10214,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(63), - Column: int(15), - }, - End: ast.Location{ - Line: int(63), - Column: int(18), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "chars", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(63), - Column: int(20), - }, - End: ast.Location{ - Line: int(63), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "chars", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(64), Column: int(5), @@ -10808,7 +10268,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(64), Column: int(17), @@ -10829,7 +10289,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(64), Column: int(17), @@ -10849,7 +10309,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(64), Column: int(17), @@ -10890,9 +10350,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -10902,7 +10361,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(64), Column: int(28), @@ -10933,24 +10392,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(64), - Column: int(11), - }, - End: ast.Location{ - Line: int(64), - Column: int(32), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(5), @@ -10980,7 +10427,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(8), @@ -11003,7 +10450,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(8), @@ -11023,7 +10470,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(8), @@ -11047,7 +10494,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(14), @@ -11062,6 +10509,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p553, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -11070,7 +10518,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(19), @@ -11093,7 +10541,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(19), @@ -11113,7 +10561,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(19), @@ -11154,9 +10602,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "member", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -11166,7 +10613,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(30), @@ -11191,7 +10638,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(37), @@ -11212,7 +10659,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(37), @@ -11235,7 +10682,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(41), @@ -11255,7 +10702,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(41), @@ -11279,7 +10726,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(65), Column: int(47), @@ -11294,6 +10741,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p587, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -11315,7 +10763,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(7), @@ -11338,7 +10786,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(7), @@ -11358,7 +10806,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(7), @@ -11406,9 +10854,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "rstripChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -11420,14 +10867,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(66), - Column: int(23), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(66), - Column: int(36), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -11501,7 +10948,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -11513,7 +10959,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(23), @@ -11560,7 +11006,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(28), @@ -11580,7 +11026,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(28), @@ -11604,7 +11050,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(34), @@ -11619,6 +11065,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p616, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -11660,7 +11107,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(66), Column: int(38), @@ -11700,7 +11147,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(68), Column: int(7), @@ -11730,18 +11177,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(63), - Column: int(3), - }, - End: ast.Location{ - Line: int(68), - Column: int(10), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -11766,7 +11201,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "stripChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -11789,52 +11223,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(70), - Column: int(14), - }, - End: ast.Location{ - Line: int(70), - Column: int(17), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "chars", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(70), - Column: int(19), - }, - End: ast.Location{ - Line: int(70), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "chars", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(5), @@ -11856,7 +11265,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(5), @@ -11876,7 +11285,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(5), @@ -11924,9 +11333,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "lstripChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -11936,7 +11344,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(21), @@ -11958,7 +11366,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(21), @@ -11978,7 +11386,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(21), @@ -12019,9 +11427,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "rstripChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -12031,7 +11438,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(37), @@ -12056,7 +11463,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(42), @@ -12091,7 +11498,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(71), Column: int(50), @@ -12122,18 +11529,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(70), - Column: int(3), - }, - End: ast.Location{ - Line: int(71), - Column: int(56), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -12158,7 +11553,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "stringChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -12181,33 +11575,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(73), - Column: int(15), - }, - End: ast.Location{ - Line: int(73), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(5), @@ -12228,7 +11611,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(5), @@ -12248,7 +11631,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(5), @@ -12296,9 +11679,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -12308,7 +11690,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(19), @@ -12329,7 +11711,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(19), @@ -12349,7 +11731,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(19), @@ -12390,9 +11772,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -12402,7 +11783,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(30), @@ -12437,7 +11818,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(36), @@ -12455,33 +11836,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(74), - Column: int(45), - }, - End: ast.Location{ - Line: int(74), - Column: int(46), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(48), @@ -12502,7 +11872,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(48), @@ -12525,7 +11895,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(74), Column: int(52), @@ -12560,18 +11930,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(73), - Column: int(3), - }, - End: ast.Location{ - Line: int(74), - Column: int(55), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -12596,7 +11954,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "parseInt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -12620,26 +11977,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(94), - Column: int(12), - }, - End: ast.Location{ - Line: int(94), - Column: int(15), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -12668,7 +12014,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(12), @@ -12689,7 +12035,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(12), @@ -12709,7 +12055,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(12), @@ -12750,9 +12096,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -12762,7 +12107,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(25), @@ -12817,7 +12162,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(12), @@ -12838,7 +12183,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(12), @@ -12859,7 +12204,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(12), @@ -12880,7 +12225,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(12), @@ -12900,7 +12245,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(12), @@ -12941,9 +12286,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -12953,7 +12297,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(23), @@ -12987,7 +12331,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(30), @@ -13002,6 +12346,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p704, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -13010,7 +12355,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(35), @@ -13030,7 +12375,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(35), @@ -13054,7 +12399,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(42), @@ -13072,7 +12417,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "-", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -13080,7 +12424,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(97), Column: int(5), @@ -13109,7 +12453,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(97), Column: int(8), @@ -13129,7 +12473,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(97), Column: int(8), @@ -13149,7 +12493,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(97), Column: int(8), @@ -13172,7 +12516,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(97), Column: int(12), @@ -13187,6 +12531,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p704, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, RightBracketFodder: ast.Fodder{}, @@ -13197,7 +12542,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(97), Column: int(18), @@ -13215,14 +12560,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "-", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(98), Column: int(7), @@ -13252,7 +12596,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(98), Column: int(8), @@ -13274,7 +12618,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(98), Column: int(8), @@ -13302,14 +12646,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(98), - Column: int(18), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(98), - Column: int(25), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -13382,7 +12726,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -13394,7 +12737,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(98), Column: int(18), @@ -13419,7 +12762,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(98), Column: int(22), @@ -13434,6 +12777,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p766, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, CommaFodder: nil, @@ -13496,7 +12840,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(98), Column: int(27), @@ -13511,6 +12855,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p766, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: nil, @@ -13535,7 +12880,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(100), Column: int(7), @@ -13556,7 +12901,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(100), Column: int(7), @@ -13589,7 +12934,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(100), Column: int(17), @@ -13614,7 +12959,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(100), Column: int(22), @@ -13629,6 +12974,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p780, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: nil, @@ -13648,14 +12994,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(96), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(100), - Column: int(25), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -13669,14 +13015,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(96), - Column: int(48), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(96), - Column: int(78), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -13749,7 +13095,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -13761,7 +13106,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(48), @@ -13779,7 +13124,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "Not an integer: \"%s\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -13787,7 +13131,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(73), @@ -13809,7 +13153,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(96), Column: int(74), @@ -13852,14 +13196,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(95), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(100), - Column: int(25), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -13871,7 +13215,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(32), @@ -13892,7 +13236,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(32), @@ -13910,14 +13254,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "Expected string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(58), @@ -13938,7 +13281,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(58), @@ -13958,7 +13301,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(58), @@ -13999,9 +13342,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -14011,7 +13353,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(95), Column: int(67), @@ -14045,18 +13387,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(94), - Column: int(3), - }, - End: ast.Location{ - Line: int(100), - Column: int(25), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -14081,7 +13411,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "parseOctal", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -14105,26 +13434,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(102), - Column: int(14), - }, - End: ast.Location{ - Line: int(102), - Column: int(17), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -14153,7 +13471,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(12), @@ -14174,7 +13492,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(12), @@ -14194,7 +13512,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(12), @@ -14235,9 +13553,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -14247,7 +13564,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(25), @@ -14302,7 +13619,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(104), Column: int(12), @@ -14323,7 +13640,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(104), Column: int(12), @@ -14344,7 +13661,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(104), Column: int(12), @@ -14364,7 +13681,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(104), Column: int(12), @@ -14405,9 +13722,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -14417,7 +13733,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(104), Column: int(23), @@ -14451,7 +13767,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(104), Column: int(30), @@ -14466,6 +13782,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p823, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -14473,7 +13790,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(105), Column: int(5), @@ -14494,7 +13811,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(105), Column: int(5), @@ -14527,7 +13844,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(105), Column: int(15), @@ -14552,7 +13869,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(105), Column: int(20), @@ -14567,6 +13884,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p857, FreeVars: nil, }, + Value: float64(8), OriginalString: "8", }, CommaFodder: nil, @@ -14585,14 +13903,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(104), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(105), - Column: int(22), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -14601,7 +13919,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(104), Column: int(34), @@ -14619,7 +13937,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "Not an octal number: \"\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -14629,14 +13946,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(103), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(105), - Column: int(22), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -14648,7 +13965,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(32), @@ -14669,7 +13986,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(32), @@ -14687,14 +14004,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "Expected string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(58), @@ -14715,7 +14031,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(58), @@ -14735,7 +14051,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(58), @@ -14776,9 +14092,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -14788,7 +14103,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(103), Column: int(67), @@ -14822,18 +14137,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(102), - Column: int(3), - }, - End: ast.Location{ - Line: int(105), - Column: int(22), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -14858,7 +14161,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "parseHex", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -14882,26 +14184,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(107), - Column: int(12), - }, - End: ast.Location{ - Line: int(107), - Column: int(15), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -14930,7 +14221,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(12), @@ -14951,7 +14242,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(12), @@ -14971,7 +14262,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(12), @@ -15012,9 +14303,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -15024,7 +14314,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(25), @@ -15079,7 +14369,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(109), Column: int(12), @@ -15100,7 +14390,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(109), Column: int(12), @@ -15121,7 +14411,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(109), Column: int(12), @@ -15141,7 +14431,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(109), Column: int(12), @@ -15182,9 +14472,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -15194,7 +14483,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(109), Column: int(23), @@ -15228,7 +14517,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(109), Column: int(30), @@ -15243,6 +14532,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p885, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -15250,7 +14540,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(110), Column: int(5), @@ -15271,7 +14561,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(110), Column: int(5), @@ -15304,7 +14594,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(110), Column: int(15), @@ -15329,7 +14619,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(110), Column: int(20), @@ -15344,6 +14634,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p919, FreeVars: nil, }, + Value: float64(16), OriginalString: "16", }, CommaFodder: nil, @@ -15362,14 +14653,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(109), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(110), - Column: int(23), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -15378,7 +14669,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(109), Column: int(34), @@ -15396,7 +14687,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "Not hexadecimal: \"\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -15406,14 +14696,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(108), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(110), - Column: int(23), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -15425,7 +14715,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(32), @@ -15446,7 +14736,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(32), @@ -15464,14 +14754,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "Expected string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(58), @@ -15492,7 +14781,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(58), @@ -15512,7 +14801,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(58), @@ -15553,9 +14842,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -15565,7 +14853,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(108), Column: int(67), @@ -15599,18 +14887,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(107), - Column: int(3), - }, - End: ast.Location{ - Line: int(110), - Column: int(23), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -15635,7 +14911,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "split", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -15658,45 +14933,20 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(112), - Column: int(9), - }, - End: ast.Location{ - Line: int(112), - Column: int(12), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "c", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(112), - Column: int(14), - }, - End: ast.Location{ - Line: int(112), - Column: int(15), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "c", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -15725,7 +14975,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(12), @@ -15746,7 +14996,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(12), @@ -15766,7 +15016,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(12), @@ -15807,9 +15057,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -15819,7 +15068,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(25), @@ -15874,7 +15123,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(12), @@ -15895,7 +15144,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(12), @@ -15915,7 +15164,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(12), @@ -15956,9 +15205,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -15968,7 +15216,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(25), @@ -16023,7 +15271,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), Column: int(12), @@ -16044,7 +15292,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), Column: int(12), @@ -16065,7 +15313,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), Column: int(12), @@ -16085,7 +15333,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), Column: int(12), @@ -16126,9 +15374,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -16138,7 +15385,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), Column: int(23), @@ -16172,7 +15419,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), Column: int(29), @@ -16187,6 +15434,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p947, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -16194,7 +15442,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(116), Column: int(5), @@ -16216,7 +15464,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(116), Column: int(5), @@ -16236,7 +15484,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(116), Column: int(5), @@ -16284,9 +15532,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "splitLimit", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -16296,7 +15543,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(116), Column: int(20), @@ -16321,7 +15568,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(116), Column: int(25), @@ -16346,7 +15593,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(116), Column: int(28), @@ -16365,7 +15612,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(116), Column: int(29), @@ -16380,6 +15627,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p997, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -16399,14 +15647,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(115), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(116), - Column: int(31), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -16418,14 +15666,14 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), - Column: int(33), + Column: int(34), }, End: ast.Location{ Line: int(115), - Column: int(104), + Column: int(101), }, File: p1, }, @@ -16439,14 +15687,14 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), - Column: int(33), + Column: int(34), }, End: ast.Location{ Line: int(115), - Column: int(88), + Column: int(87), }, File: p1, }, @@ -16454,24 +15702,23 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p947, FreeVars: nil, }, - Value: "std.split second parameter should have length 1, got ", + Value: "std.split second parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), - Column: int(91), + Column: int(90), }, End: ast.Location{ Line: int(115), - Column: int(104), + Column: int(101), }, File: p1, }, @@ -16485,14 +15732,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), - Column: int(91), + Column: int(90), }, End: ast.Location{ Line: int(115), - Column: int(101), + Column: int(98), }, File: p1, }, @@ -16505,14 +15752,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), - Column: int(91), + Column: int(90), }, End: ast.Location{ Line: int(115), - Column: int(94), + Column: int(93), }, File: p1, }, @@ -16543,12 +15790,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "length", + Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -16558,14 +15804,14 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(115), - Column: int(102), + Column: int(99), }, End: ast.Location{ Line: int(115), - Column: int(103), + Column: int(100), }, File: p1, }, @@ -16596,14 +15842,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(114), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(116), - Column: int(31), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -16615,7 +15861,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(30), @@ -16636,7 +15882,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(30), @@ -16654,14 +15900,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.split second parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(86), @@ -16682,7 +15927,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(86), @@ -16702,7 +15947,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(86), @@ -16743,9 +15988,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -16755,7 +15999,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(114), Column: int(95), @@ -16793,14 +16037,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(113), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(116), - Column: int(31), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -16812,7 +16056,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(32), @@ -16833,7 +16077,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(32), @@ -16851,14 +16095,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.split first parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(87), @@ -16879,7 +16122,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(87), @@ -16899,7 +16142,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(87), @@ -16940,9 +16183,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -16952,7 +16194,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(113), Column: int(96), @@ -16986,18 +16228,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(112), - Column: int(3), - }, - End: ast.Location{ - Line: int(116), - Column: int(31), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -17022,7 +16252,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "splitLimit", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -17045,64 +16274,25 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(118), - Column: int(14), - }, - End: ast.Location{ - Line: int(118), - Column: int(17), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "c", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(118), - Column: int(19), - }, - End: ast.Location{ - Line: int(118), - Column: int(20), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "maxsplits", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(118), - Column: int(22), - }, - End: ast.Location{ - Line: int(118), - Column: int(31), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "c", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "maxsplits", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -17132,7 +16322,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(12), @@ -17153,7 +16343,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(12), @@ -17173,7 +16363,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(12), @@ -17214,9 +16404,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -17226,7 +16415,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(25), @@ -17282,7 +16471,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(12), @@ -17303,7 +16492,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(12), @@ -17323,7 +16512,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(12), @@ -17364,9 +16553,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -17376,7 +16564,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(25), @@ -17432,7 +16620,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), Column: int(12), @@ -17453,7 +16641,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), Column: int(12), @@ -17474,7 +16662,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), Column: int(12), @@ -17494,7 +16682,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), Column: int(12), @@ -17535,9 +16723,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -17547,7 +16734,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), Column: int(23), @@ -17581,7 +16768,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), Column: int(29), @@ -17596,6 +16783,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1058, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -17626,7 +16814,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(12), @@ -17647,7 +16835,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(12), @@ -17667,7 +16855,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(12), @@ -17708,9 +16896,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -17720,7 +16907,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(25), @@ -17753,7 +16940,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(123), Column: int(5), @@ -17788,7 +16975,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(123), Column: int(11), @@ -17808,109 +16995,42 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(123), - Column: int(15), - }, - End: ast.Location{ - Line: int(123), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "delim", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(123), - Column: int(20), - }, - End: ast.Location{ - Line: int(123), - Column: int(25), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "delim", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(123), - Column: int(27), - }, - End: ast.Location{ - Line: int(123), - Column: int(28), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(123), - Column: int(30), - }, - End: ast.Location{ - Line: int(123), - Column: int(33), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(123), - Column: int(35), - }, - End: ast.Location{ - Line: int(123), - Column: int(36), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(124), Column: int(7), @@ -17949,7 +17069,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(124), Column: int(17), @@ -17970,7 +17090,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(124), Column: int(17), @@ -17993,7 +17113,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(124), Column: int(21), @@ -18017,24 +17137,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(124), - Column: int(13), - }, - End: ast.Location{ - Line: int(124), - Column: int(23), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(125), Column: int(7), @@ -18074,7 +17182,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(125), Column: int(18), @@ -18094,7 +17202,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(125), Column: int(18), @@ -18118,7 +17226,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(125), Column: int(22), @@ -18133,29 +17241,18 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1136, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(125), - Column: int(13), - }, - End: ast.Location{ - Line: int(125), - Column: int(23), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(126), Column: int(7), @@ -18191,7 +17288,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(126), Column: int(10), @@ -18213,7 +17310,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(126), Column: int(10), @@ -18237,7 +17334,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(126), Column: int(15), @@ -18258,7 +17355,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(126), Column: int(15), @@ -18278,7 +17375,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(126), Column: int(15), @@ -18319,9 +17416,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -18331,7 +17427,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(126), Column: int(26), @@ -18365,7 +17461,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(127), Column: int(9), @@ -18386,7 +17482,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(127), Column: int(9), @@ -18417,7 +17513,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(127), Column: int(15), @@ -18439,7 +17535,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(127), Column: int(16), @@ -18476,7 +17572,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(12), @@ -18504,7 +17600,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(15), @@ -18528,7 +17624,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(15), @@ -18549,7 +17645,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(15), @@ -18573,7 +17669,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(20), @@ -18598,7 +17694,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(30), @@ -18620,7 +17716,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(30), @@ -18640,7 +17736,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(30), @@ -18664,7 +17760,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(43), @@ -18683,7 +17779,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(44), @@ -18698,6 +17794,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1121, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -18707,7 +17804,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(49), @@ -18729,7 +17826,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(49), @@ -18750,7 +17847,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(49), @@ -18770,7 +17867,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(49), @@ -18811,9 +17908,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -18823,7 +17919,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(60), @@ -18857,7 +17953,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(128), Column: int(67), @@ -18883,7 +17979,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(9), @@ -18908,7 +18004,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(9), @@ -18941,7 +18037,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(13), @@ -18966,7 +18062,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(18), @@ -18991,7 +18087,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(25), @@ -19016,7 +18112,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(29), @@ -19037,7 +18133,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(29), @@ -19061,7 +18157,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(35), @@ -19083,7 +18179,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(36), @@ -19115,7 +18211,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(129), Column: int(40), @@ -19133,7 +18229,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -19156,7 +18251,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(9), @@ -19182,7 +18277,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(9), @@ -19215,7 +18310,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(13), @@ -19240,7 +18335,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(18), @@ -19265,7 +18360,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(25), @@ -19290,7 +18385,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(29), @@ -19315,7 +18410,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(34), @@ -19336,7 +18431,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(34), @@ -19360,7 +18455,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(131), Column: int(38), @@ -19397,24 +18492,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(132), Column: int(5), @@ -19436,7 +18519,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(132), Column: int(5), @@ -19469,7 +18552,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(132), Column: int(9), @@ -19494,7 +18577,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(132), Column: int(14), @@ -19519,7 +18602,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(132), Column: int(17), @@ -19534,6 +18617,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1257, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -19542,7 +18626,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(132), Column: int(20), @@ -19567,7 +18651,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(132), Column: int(24), @@ -19585,7 +18669,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -19604,14 +18687,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(122), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(132), - Column: int(27), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -19623,7 +18706,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(38), @@ -19644,7 +18727,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(38), @@ -19662,14 +18745,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.splitLimit third parameter should be a number, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(98), @@ -19690,7 +18772,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(98), @@ -19710,7 +18792,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(98), @@ -19751,9 +18833,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -19763,7 +18844,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(122), Column: int(107), @@ -19801,14 +18882,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(121), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(132), - Column: int(27), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -19820,14 +18901,14 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), - Column: int(33), + Column: int(34), }, End: ast.Location{ Line: int(121), - Column: int(109), + Column: int(110), }, File: p1, }, @@ -19841,14 +18922,14 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), - Column: int(33), + Column: int(34), }, End: ast.Location{ Line: int(121), - Column: int(93), + Column: int(94), }, File: p1, }, @@ -19859,21 +18940,20 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.splitLimit second parameter should have length 1, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), - Column: int(96), + Column: int(97), }, End: ast.Location{ Line: int(121), - Column: int(109), + Column: int(110), }, File: p1, }, @@ -19887,14 +18967,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), - Column: int(96), + Column: int(97), }, End: ast.Location{ Line: int(121), - Column: int(106), + Column: int(107), }, File: p1, }, @@ -19907,14 +18987,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), - Column: int(96), + Column: int(97), }, End: ast.Location{ Line: int(121), - Column: int(99), + Column: int(100), }, File: p1, }, @@ -19948,9 +19028,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -19960,14 +19039,14 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(121), - Column: int(107), + Column: int(108), }, End: ast.Location{ Line: int(121), - Column: int(108), + Column: int(109), }, File: p1, }, @@ -19998,14 +19077,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(120), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(132), - Column: int(27), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -20017,7 +19096,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(30), @@ -20038,7 +19117,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(30), @@ -20056,14 +19135,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.splitLimit second parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(91), @@ -20084,7 +19162,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(91), @@ -20104,7 +19182,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(91), @@ -20145,9 +19223,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -20157,7 +19234,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(120), Column: int(100), @@ -20195,14 +19272,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(119), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(132), - Column: int(27), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -20214,7 +19291,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(32), @@ -20235,7 +19312,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(32), @@ -20253,14 +19330,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.splitLimit first parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(92), @@ -20281,7 +19357,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(92), @@ -20301,7 +19377,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(92), @@ -20342,9 +19418,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -20354,7 +19429,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(119), Column: int(101), @@ -20388,18 +19463,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(118), - Column: int(3), - }, - End: ast.Location{ - Line: int(132), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -20424,7 +19487,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "strReplace", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -20447,64 +19509,25 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(134), - Column: int(14), - }, - End: ast.Location{ - Line: int(134), - Column: int(17), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "from", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(134), - Column: int(19), - }, - End: ast.Location{ - Line: int(134), - Column: int(23), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "to", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(134), - Column: int(25), - }, - End: ast.Location{ - Line: int(134), - Column: int(27), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "from", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "to", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -20534,7 +19557,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(135), Column: int(12), @@ -20555,7 +19578,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(135), Column: int(12), @@ -20575,7 +19598,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(135), Column: int(12), @@ -20616,9 +19639,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -20628,7 +19650,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(135), Column: int(25), @@ -20684,7 +19706,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(136), Column: int(12), @@ -20705,7 +19727,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(136), Column: int(12), @@ -20725,7 +19747,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(136), Column: int(12), @@ -20766,9 +19788,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -20778,7 +19799,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(136), Column: int(25), @@ -20834,7 +19855,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(137), Column: int(12), @@ -20855,7 +19876,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(137), Column: int(12), @@ -20875,7 +19896,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(137), Column: int(12), @@ -20916,9 +19937,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -20928,7 +19948,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(137), Column: int(25), @@ -20984,7 +20004,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(138), Column: int(12), @@ -21004,7 +20024,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(138), Column: int(12), @@ -21028,7 +20048,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(138), Column: int(20), @@ -21046,14 +20066,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: nil, BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(141), Column: int(5), @@ -21072,7 +20091,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -21096,7 +20115,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(141), Column: int(21), @@ -21117,7 +20136,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(141), Column: int(21), @@ -21137,7 +20156,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(141), Column: int(21), @@ -21178,9 +20197,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -21190,7 +20208,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(141), Column: int(32), @@ -21221,24 +20239,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(141), - Column: int(11), - }, - End: ast.Location{ - Line: int(141), - Column: int(36), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(142), Column: int(5), @@ -21274,7 +20280,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(142), Column: int(22), @@ -21295,7 +20301,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(142), Column: int(22), @@ -21315,7 +20321,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(142), Column: int(22), @@ -21356,9 +20362,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -21368,7 +20373,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(142), Column: int(33), @@ -21399,24 +20404,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(142), - Column: int(11), - }, - End: ast.Location{ - Line: int(142), - Column: int(38), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(5), @@ -21435,7 +20428,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -21461,7 +20454,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(11), @@ -21482,33 +20475,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(145), - Column: int(20), - }, - End: ast.Location{ - Line: int(145), - Column: int(21), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(25), @@ -21534,14 +20516,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(145), - Column: int(25), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(145), - Column: int(44), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -21616,7 +20598,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -21628,7 +20609,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(25), @@ -21653,7 +20634,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(29), @@ -21678,7 +20659,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(31), @@ -21699,7 +20680,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(31), @@ -21723,7 +20704,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(35), @@ -21780,7 +20761,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(145), Column: int(48), @@ -21803,24 +20784,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(149), Column: int(5), @@ -21839,7 +20808,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -21847,7 +20816,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -21874,7 +20843,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(149), Column: int(11), @@ -21898,71 +20867,32 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "start_index", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(149), - Column: int(25), - }, - End: ast.Location{ - Line: int(149), - Column: int(36), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "start_index", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "curr_index", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(149), - Column: int(38), - }, - End: ast.Location{ - Line: int(149), - Column: int(48), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "curr_index", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "acc", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(149), - Column: int(50), - }, - End: ast.Location{ - Line: int(149), - Column: int(53), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "acc", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(150), Column: int(7), @@ -21998,7 +20928,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(150), Column: int(10), @@ -22019,7 +20949,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(150), Column: int(10), @@ -22043,7 +20973,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(150), Column: int(23), @@ -22067,7 +20997,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(151), Column: int(9), @@ -22091,7 +21021,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(151), Column: int(9), @@ -22124,14 +21054,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(151), - Column: int(15), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(151), - Column: int(42), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -22206,7 +21136,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -22218,7 +21147,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(151), Column: int(15), @@ -22243,7 +21172,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(151), Column: int(19), @@ -22268,7 +21197,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(151), Column: int(31), @@ -22331,7 +21260,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(152), Column: int(12), @@ -22359,7 +21288,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(152), Column: int(15), @@ -22380,7 +21309,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(152), Column: int(15), @@ -22406,7 +21335,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(152), Column: int(24), @@ -22439,7 +21368,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(153), Column: int(9), @@ -22478,7 +21407,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(153), Column: int(27), @@ -22500,7 +21429,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(153), Column: int(27), @@ -22524,7 +21453,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(153), Column: int(40), @@ -22545,7 +21474,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(153), Column: int(40), @@ -22565,7 +21494,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(153), Column: int(40), @@ -22606,9 +21535,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -22618,7 +21546,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(153), Column: int(51), @@ -22650,24 +21578,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(153), - Column: int(15), - }, - End: ast.Location{ - Line: int(153), - Column: int(56), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(9), @@ -22694,7 +21610,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(9), @@ -22727,7 +21643,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(23), @@ -22752,7 +21668,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(34), @@ -22777,7 +21693,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(45), @@ -22802,7 +21718,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(45), @@ -22826,7 +21742,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(45), @@ -22852,14 +21768,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(154), - Column: int(51), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(154), - Column: int(78), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -22934,7 +21850,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -22946,7 +21861,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(51), @@ -22971,7 +21886,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(55), @@ -22996,7 +21911,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(67), @@ -23053,7 +21968,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(154), Column: int(81), @@ -23095,7 +22010,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(156), Column: int(9), @@ -23118,7 +22033,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(156), Column: int(9), @@ -23151,7 +22066,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(156), Column: int(23), @@ -23176,7 +22091,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(156), Column: int(36), @@ -23196,7 +22111,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(156), Column: int(36), @@ -23220,7 +22135,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(156), Column: int(49), @@ -23235,6 +22150,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1558, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -23244,7 +22160,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(156), Column: int(52), @@ -23278,24 +22194,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(160), Column: int(5), @@ -23314,7 +22218,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -23322,7 +22226,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -23343,7 +22247,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(160), Column: int(8), @@ -23363,7 +22267,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(160), Column: int(8), @@ -23387,7 +22291,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(160), Column: int(20), @@ -23402,6 +22306,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1335, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -23409,7 +22314,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(7), @@ -23432,7 +22337,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(7), @@ -23452,7 +22357,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(7), @@ -23500,9 +22405,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -23512,7 +22416,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(16), @@ -23537,7 +22441,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(20), @@ -23559,7 +22463,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(20), @@ -23579,7 +22483,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(20), @@ -23620,9 +22524,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "split", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -23632,7 +22535,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(30), @@ -23657,7 +22560,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(161), Column: int(35), @@ -23707,7 +22610,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(163), Column: int(7), @@ -23727,7 +22630,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(163), Column: int(7), @@ -23760,7 +22663,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(163), Column: int(21), @@ -23775,6 +22678,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1610, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -23783,7 +22687,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(163), Column: int(24), @@ -23798,6 +22702,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1610, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -23806,7 +22711,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(163), Column: int(27), @@ -23824,7 +22729,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -23847,14 +22751,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(138), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(163), - Column: int(30), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -23863,7 +22767,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(138), Column: int(25), @@ -23881,7 +22785,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "'from' string must not be zero length.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -23891,14 +22794,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(137), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(163), - Column: int(30), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -23925,7 +22828,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "Assertion failed", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -23935,14 +22837,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(136), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(163), - Column: int(30), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -23969,7 +22871,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "Assertion failed", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -23979,14 +22880,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(135), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(163), - Column: int(30), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -24013,24 +22914,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "Assertion failed", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(134), - Column: int(3), - }, - End: ast.Location{ - Line: int(163), - Column: int(30), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -24055,7 +22943,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "asciiUpper", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -24078,33 +22965,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(165), - Column: int(14), - }, - End: ast.Location{ - Line: int(165), - Column: int(17), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(166), Column: int(5), @@ -24137,7 +23013,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(166), Column: int(16), @@ -24157,7 +23033,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(166), Column: int(16), @@ -24198,31 +23074,18 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(166), - Column: int(11), - }, - End: ast.Location{ - Line: int(166), - Column: int(29), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(5), @@ -24256,7 +23119,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(11), @@ -24275,33 +23138,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "c", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(167), - Column: int(21), - }, - End: ast.Location{ - Line: int(167), - Column: int(22), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "c", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(26), @@ -24323,7 +23175,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(29), @@ -24344,7 +23196,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(29), @@ -24365,7 +23217,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(29), @@ -24386,7 +23238,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(29), @@ -24412,7 +23264,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(32), @@ -24446,7 +23298,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(38), @@ -24461,6 +23313,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1645, FreeVars: nil, }, + Value: float64(97), OriginalString: "97", }, }, @@ -24469,7 +23322,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(44), @@ -24490,7 +23343,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(44), @@ -24511,7 +23364,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(44), @@ -24537,7 +23390,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(47), @@ -24571,7 +23424,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(167), Column: int(52), @@ -24586,6 +23439,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1645, FreeVars: nil, }, + Value: float64(123), OriginalString: "123", }, }, @@ -24594,7 +23448,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(7), @@ -24616,7 +23470,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(7), @@ -24636,7 +23490,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(7), @@ -24684,9 +23538,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "char", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -24696,7 +23549,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(16), @@ -24717,7 +23570,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(16), @@ -24738,7 +23591,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(16), @@ -24764,7 +23617,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(19), @@ -24798,7 +23651,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(168), Column: int(24), @@ -24813,6 +23666,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1681, FreeVars: nil, }, + Value: float64(32), OriginalString: "32", }, }, @@ -24837,7 +23691,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(170), Column: int(7), @@ -24867,24 +23721,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(5), @@ -24906,7 +23748,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(5), @@ -24926,7 +23768,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(5), @@ -24974,9 +23816,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -24986,7 +23827,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(14), @@ -25004,7 +23845,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -25012,7 +23852,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(18), @@ -25034,7 +23874,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(18), @@ -25054,7 +23894,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(18), @@ -25095,9 +23935,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "map", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -25107,7 +23946,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(26), @@ -25132,7 +23971,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(37), @@ -25153,7 +23992,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(37), @@ -25173,7 +24012,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(37), @@ -25214,9 +24053,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "stringChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -25226,7 +24064,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(171), Column: int(53), @@ -25279,18 +24117,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(165), - Column: int(3), - }, - End: ast.Location{ - Line: int(171), - Column: int(59), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -25315,7 +24141,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "asciiLower", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -25338,33 +24163,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(173), - Column: int(14), - }, - End: ast.Location{ - Line: int(173), - Column: int(17), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(174), Column: int(5), @@ -25397,7 +24211,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(174), Column: int(16), @@ -25417,7 +24231,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(174), Column: int(16), @@ -25458,31 +24272,18 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(174), - Column: int(11), - }, - End: ast.Location{ - Line: int(174), - Column: int(29), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(5), @@ -25516,7 +24317,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(11), @@ -25535,33 +24336,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "c", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(175), - Column: int(23), - }, - End: ast.Location{ - Line: int(175), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "c", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(28), @@ -25583,7 +24373,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(31), @@ -25604,7 +24394,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(31), @@ -25625,7 +24415,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(31), @@ -25646,7 +24436,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(31), @@ -25672,7 +24462,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(34), @@ -25706,7 +24496,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(40), @@ -25721,6 +24511,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1753, FreeVars: nil, }, + Value: float64(65), OriginalString: "65", }, }, @@ -25729,7 +24520,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(46), @@ -25750,7 +24541,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(46), @@ -25771,7 +24562,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(46), @@ -25797,7 +24588,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(49), @@ -25831,7 +24622,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(175), Column: int(54), @@ -25846,6 +24637,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1753, FreeVars: nil, }, + Value: float64(91), OriginalString: "91", }, }, @@ -25854,7 +24646,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(7), @@ -25876,7 +24668,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(7), @@ -25896,7 +24688,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(7), @@ -25944,9 +24736,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "char", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -25956,7 +24747,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(16), @@ -25977,7 +24768,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(16), @@ -25998,7 +24789,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(16), @@ -26024,7 +24815,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(19), @@ -26058,7 +24849,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(176), Column: int(24), @@ -26073,6 +24864,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1789, FreeVars: nil, }, + Value: float64(32), OriginalString: "32", }, }, @@ -26097,7 +24889,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(178), Column: int(7), @@ -26127,24 +24919,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(5), @@ -26166,7 +24946,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(5), @@ -26186,7 +24966,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(5), @@ -26234,9 +25014,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -26246,7 +25025,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(14), @@ -26264,7 +25043,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -26272,7 +25050,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(18), @@ -26294,7 +25072,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(18), @@ -26314,7 +25092,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(18), @@ -26355,9 +25133,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "map", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -26367,7 +25144,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(26), @@ -26392,7 +25169,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(39), @@ -26413,7 +25190,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(39), @@ -26433,7 +25210,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(39), @@ -26474,9 +25251,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "stringChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -26486,7 +25262,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(179), Column: int(55), @@ -26539,18 +25315,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(173), - Column: int(3), - }, - End: ast.Location{ - Line: int(179), - Column: int(61), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -26575,7 +25339,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "range", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -26598,52 +25361,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "from", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(181), - Column: int(9), - }, - End: ast.Location{ - Line: int(181), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "to", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(181), - Column: int(15), - }, - End: ast.Location{ - Line: int(181), - Column: int(17), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "from", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "to", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(5), @@ -26665,7 +25403,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(5), @@ -26685,7 +25423,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(5), @@ -26733,9 +25471,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -26745,7 +25482,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(19), @@ -26766,7 +25503,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(19), @@ -26787,7 +25524,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(19), @@ -26811,7 +25548,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(24), @@ -26836,7 +25573,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(31), @@ -26851,6 +25588,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1852, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -26860,7 +25598,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(34), @@ -26878,33 +25616,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(182), - Column: int(43), - }, - End: ast.Location{ - Line: int(182), - Column: int(44), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(46), @@ -26925,7 +25652,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(46), @@ -26949,7 +25676,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(182), Column: int(50), @@ -26982,18 +25709,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(181), - Column: int(3), - }, - End: ast.Location{ - Line: int(182), - Column: int(55), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -27018,7 +25733,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "repeat", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -27041,52 +25755,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "what", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(184), - Column: int(10), - }, - End: ast.Location{ - Line: int(184), - Column: int(14), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "count", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(184), - Column: int(16), - }, - End: ast.Location{ - Line: int(184), - Column: int(21), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "what", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "count", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(185), Column: int(5), @@ -27120,7 +25809,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(186), Column: int(7), @@ -27148,7 +25837,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(186), Column: int(10), @@ -27169,7 +25858,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(186), Column: int(10), @@ -27189,7 +25878,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(186), Column: int(10), @@ -27230,9 +25919,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -27242,7 +25930,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(186), Column: int(23), @@ -27275,7 +25963,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(186), Column: int(34), @@ -27293,7 +25981,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -27306,7 +25993,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(187), Column: int(12), @@ -27327,7 +26014,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(187), Column: int(15), @@ -27348,7 +26035,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(187), Column: int(15), @@ -27368,7 +26055,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(187), Column: int(15), @@ -27409,9 +26096,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -27421,7 +26107,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(187), Column: int(27), @@ -27454,7 +26140,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(187), Column: int(38), @@ -27484,7 +26170,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(188), Column: int(12), @@ -27502,7 +26188,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(188), Column: int(18), @@ -27520,31 +26206,18 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.repeat first argument must be an array or a string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(185), - Column: int(11), - }, - End: ast.Location{ - Line: int(188), - Column: int(74), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(5), @@ -27567,7 +26240,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(5), @@ -27587,7 +26260,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(5), @@ -27635,9 +26308,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -27647,7 +26319,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(14), @@ -27672,7 +26344,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(22), @@ -27694,7 +26366,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(22), @@ -27714,7 +26386,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(22), @@ -27755,9 +26427,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -27767,7 +26438,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(36), @@ -27792,7 +26463,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(43), @@ -27810,33 +26481,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(189), - Column: int(52), - }, - End: ast.Location{ - Line: int(189), - Column: int(53), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(189), Column: int(55), @@ -27879,18 +26539,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(184), - Column: int(3), - }, - End: ast.Location{ - Line: int(189), - Column: int(61), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -27915,7 +26563,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -27938,90 +26585,37 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "indexable", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(191), - Column: int(9), - }, - End: ast.Location{ - Line: int(191), - Column: int(18), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "index", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(191), - Column: int(20), - }, - End: ast.Location{ - Line: int(191), - Column: int(25), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "end", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(191), - Column: int(27), - }, - End: ast.Location{ - Line: int(191), - Column: int(30), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "step", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(191), - Column: int(32), - }, - End: ast.Location{ - Line: int(191), - Column: int(36), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "indexable", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "index", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "end", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "step", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(192), Column: int(5), @@ -28057,7 +26651,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(194), Column: int(7), @@ -28076,7 +26670,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(6), Comment: []string{ @@ -28118,12 +26712,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "indexable", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(195), Column: int(20), @@ -28143,18 +26736,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "indexable", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(195), - Column: int(9), - }, - End: ast.Location{ - Line: int(195), - Column: int(29), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -28179,12 +26760,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "index", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(197), Column: int(11), @@ -28211,7 +26791,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(197), Column: int(14), @@ -28231,7 +26811,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(197), Column: int(14), @@ -28255,7 +26835,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(197), Column: int(23), @@ -28276,7 +26856,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(197), Column: int(33), @@ -28291,6 +26871,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1960, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, ElseFodder: ast.Fodder{ @@ -28304,7 +26885,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(198), Column: int(16), @@ -28325,18 +26906,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(196), - Column: int(9), - }, - End: ast.Location{ - Line: int(198), - Column: int(21), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -28361,12 +26930,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "end", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(11), @@ -28395,7 +26963,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(14), @@ -28415,7 +26983,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(14), @@ -28439,7 +27007,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(21), @@ -28460,7 +27028,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(31), @@ -28481,7 +27049,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(31), @@ -28501,7 +27069,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(31), @@ -28542,9 +27110,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -28554,7 +27121,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(200), Column: int(42), @@ -28594,7 +27161,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(201), Column: int(16), @@ -28615,18 +27182,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(199), - Column: int(9), - }, - End: ast.Location{ - Line: int(201), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -28651,12 +27206,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "step", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(203), Column: int(11), @@ -28683,7 +27237,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(203), Column: int(14), @@ -28703,7 +27257,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(203), Column: int(14), @@ -28727,7 +27281,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(203), Column: int(22), @@ -28748,7 +27302,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(203), Column: int(32), @@ -28763,6 +27317,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1960, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, ElseFodder: ast.Fodder{ @@ -28776,7 +27331,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(204), Column: int(16), @@ -28797,18 +27352,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(202), - Column: int(9), - }, - End: ast.Location{ - Line: int(204), - Column: int(20), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -28833,12 +27376,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(205), Column: int(17), @@ -28859,7 +27401,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(205), Column: int(17), @@ -28879,7 +27421,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(205), Column: int(17), @@ -28920,9 +27462,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -28932,7 +27473,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(205), Column: int(28), @@ -28962,18 +27503,6 @@ var _StdAst = &ast.DesugaredObject{ TailStrictFodder: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(205), - Column: int(9), - }, - End: ast.Location{ - Line: int(205), - Column: int(38), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -28998,12 +27527,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(206), Column: int(15), @@ -29024,7 +27552,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(206), Column: int(15), @@ -29044,7 +27572,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(206), Column: int(15), @@ -29085,9 +27613,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -29097,7 +27624,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(206), Column: int(24), @@ -29127,36 +27654,12 @@ var _StdAst = &ast.DesugaredObject{ TailStrictFodder: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(206), - Column: int(9), - }, - End: ast.Location{ - Line: int(206), - Column: int(34), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(192), - Column: int(11), - }, - End: ast.Location{ - Line: int(207), - Column: int(8), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ @@ -29185,7 +27688,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(12), @@ -29205,7 +27708,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(12), @@ -29225,7 +27728,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(12), @@ -29245,7 +27748,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(12), @@ -29265,7 +27768,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(12), @@ -29306,9 +27809,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "index", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -29316,7 +27818,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(27), @@ -29331,6 +27833,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1949, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -29339,7 +27842,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(32), @@ -29359,7 +27862,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(32), @@ -29379,7 +27882,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(32), @@ -29420,9 +27923,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "end", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -29430,7 +27932,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(45), @@ -29445,6 +27947,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1949, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -29454,7 +27957,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(50), @@ -29474,7 +27977,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(50), @@ -29494,7 +27997,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(50), @@ -29535,9 +28038,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "step", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -29545,7 +28047,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(64), @@ -29560,6 +28062,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1949, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -29591,7 +28094,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(209), Column: int(12), @@ -29611,7 +28114,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(209), Column: int(12), @@ -29635,7 +28138,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(209), Column: int(20), @@ -29650,6 +28153,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p1949, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -29679,7 +28183,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(12), @@ -29700,7 +28204,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(12), @@ -29721,7 +28225,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(12), @@ -29741,7 +28245,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(12), @@ -29782,9 +28286,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -29794,7 +28297,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(25), @@ -29828,7 +28331,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(39), @@ -29849,7 +28352,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(39), @@ -29869,7 +28372,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(39), @@ -29910,9 +28413,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -29922,7 +28424,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(51), @@ -29956,7 +28458,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(211), Column: int(5), @@ -29988,7 +28490,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(211), Column: int(11), @@ -30007,52 +28509,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "slice", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(211), - Column: int(17), - }, - End: ast.Location{ - Line: int(211), - Column: int(22), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "slice", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "cur", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(211), - Column: int(24), - }, - End: ast.Location{ - Line: int(211), - Column: int(27), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "cur", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(7), @@ -30082,7 +28559,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(10), @@ -30103,7 +28580,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(10), @@ -30124,7 +28601,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(10), @@ -30148,7 +28625,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(17), @@ -30168,7 +28645,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(17), @@ -30209,9 +28686,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "end", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -30220,7 +28696,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(30), @@ -30241,7 +28717,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(30), @@ -30265,7 +28741,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(37), @@ -30285,7 +28761,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(212), Column: int(37), @@ -30326,9 +28802,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -30337,7 +28812,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(213), Column: int(9), @@ -30374,7 +28849,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(215), Column: int(9), @@ -30397,7 +28872,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(215), Column: int(9), @@ -30430,7 +28905,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(216), Column: int(11), @@ -30459,7 +28934,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(216), Column: int(14), @@ -30479,7 +28954,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(216), Column: int(14), @@ -30499,7 +28974,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(216), Column: int(14), @@ -30540,9 +29015,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -30550,7 +29024,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(216), Column: int(28), @@ -30568,14 +29042,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(217), Column: int(13), @@ -30597,7 +29070,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(217), Column: int(13), @@ -30628,7 +29101,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(217), Column: int(21), @@ -30649,7 +29122,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(217), Column: int(21), @@ -30669,7 +29142,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(217), Column: int(21), @@ -30710,16 +29183,15 @@ var _StdAst = &ast.DesugaredObject{ Value: "indexable", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, LeftBracketFodder: ast.Fodder{}, Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(217), Column: int(37), @@ -30753,7 +29225,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(219), Column: int(13), @@ -30775,7 +29247,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(219), Column: int(13), @@ -30806,7 +29278,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(219), Column: int(21), @@ -30829,7 +29301,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(219), Column: int(22), @@ -30850,7 +29322,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(219), Column: int(22), @@ -30870,7 +29342,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(219), Column: int(22), @@ -30911,16 +29383,15 @@ var _StdAst = &ast.DesugaredObject{ Value: "indexable", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, LeftBracketFodder: ast.Fodder{}, Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(219), Column: int(38), @@ -30956,7 +29427,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(220), Column: int(11), @@ -30977,7 +29448,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(220), Column: int(11), @@ -31008,7 +29479,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(220), Column: int(17), @@ -31028,7 +29499,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(220), Column: int(17), @@ -31069,9 +29540,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "step", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -31096,24 +29566,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(5), @@ -31134,7 +29592,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(5), @@ -31167,7 +29625,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(11), @@ -31187,7 +29645,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(14), @@ -31207,7 +29665,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(14), @@ -31227,7 +29685,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(14), @@ -31268,9 +29726,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -31278,7 +29735,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(28), @@ -31296,14 +29753,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(42), @@ -31321,13 +29777,12 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(50), @@ -31353,7 +29808,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(54), @@ -31373,7 +29828,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(222), Column: int(54), @@ -31414,9 +29869,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "index", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -31436,14 +29890,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(210), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(222), - Column: int(66), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -31457,14 +29911,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(210), - Column: int(64), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(210), - Column: int(139), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -31537,7 +29991,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -31549,7 +30002,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(64), @@ -31567,7 +30020,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.slice accepts a string or an array, but got: %s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -31575,7 +30027,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(120), @@ -31596,7 +30048,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(120), @@ -31616,7 +30068,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(120), @@ -31657,9 +30109,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -31669,7 +30120,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(210), Column: int(129), @@ -31716,14 +30167,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(209), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(222), - Column: int(66), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -31737,14 +30188,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(209), - Column: int(24), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(209), - Column: int(71), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -31817,7 +30268,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -31829,7 +30279,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(209), Column: int(24), @@ -31847,7 +30297,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "got %s but step must be greater than 0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -31855,7 +30304,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(209), Column: int(67), @@ -31892,14 +30341,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(208), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(222), - Column: int(66), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -31913,14 +30362,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(208), - Column: int(68), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(208), - Column: int(176), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -31993,7 +30442,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -32005,7 +30453,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(68), @@ -32023,7 +30471,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "got [%s:%s:%s] but negative index, end, and steps are not supported", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -32031,7 +30478,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(140), @@ -32053,7 +30500,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(141), @@ -32073,7 +30520,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(141), @@ -32114,9 +30561,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "index", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -32125,7 +30571,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(154), @@ -32145,7 +30591,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(154), @@ -32186,9 +30632,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "end", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -32197,7 +30642,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(165), @@ -32217,7 +30662,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(208), Column: int(165), @@ -32258,9 +30703,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "step", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -32284,18 +30728,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(191), - Column: int(3), - }, - End: ast.Location{ - Line: int(222), - Column: int(66), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -32320,7 +30752,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "member", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -32343,52 +30774,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(224), - Column: int(10), - }, - End: ast.Location{ - Line: int(224), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(224), - Column: int(15), - }, - End: ast.Location{ - Line: int(224), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(225), Column: int(5), @@ -32417,7 +30823,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(225), Column: int(8), @@ -32438,7 +30844,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(225), Column: int(8), @@ -32458,7 +30864,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(225), Column: int(8), @@ -32499,9 +30905,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -32511,7 +30916,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(225), Column: int(20), @@ -32544,7 +30949,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(226), Column: int(7), @@ -32566,7 +30971,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(226), Column: int(7), @@ -32588,7 +30993,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(226), Column: int(7), @@ -32608,7 +31013,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(226), Column: int(7), @@ -32656,9 +31061,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "count", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -32668,7 +31072,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(226), Column: int(17), @@ -32693,7 +31097,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(226), Column: int(22), @@ -32727,7 +31131,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(226), Column: int(27), @@ -32742,6 +31146,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p2291, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -32756,7 +31161,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(227), Column: int(10), @@ -32778,7 +31183,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(227), Column: int(13), @@ -32799,7 +31204,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(227), Column: int(13), @@ -32819,7 +31224,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(227), Column: int(13), @@ -32860,9 +31265,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -32872,7 +31276,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(227), Column: int(26), @@ -32905,7 +31309,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(7), @@ -32927,7 +31331,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(7), @@ -32949,7 +31353,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(7), @@ -32969,7 +31373,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(7), @@ -33017,9 +31421,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -33029,7 +31432,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(18), @@ -33051,7 +31454,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(18), @@ -33071,7 +31474,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(18), @@ -33112,9 +31515,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "findSubstr", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -33124,7 +31526,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(33), @@ -33149,7 +31551,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(36), @@ -33193,7 +31595,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(228), Column: int(44), @@ -33208,6 +31610,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p2291, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -33222,7 +31625,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(229), Column: int(10), @@ -33240,7 +31643,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(229), Column: int(16), @@ -33258,25 +31661,12 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.member first argument must be an array or a string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(224), - Column: int(3), - }, - End: ast.Location{ - Line: int(229), - Column: int(72), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -33301,7 +31691,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "count", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -33324,52 +31713,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(231), - Column: int(9), - }, - End: ast.Location{ - Line: int(231), - Column: int(12), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(231), - Column: int(14), - }, - End: ast.Location{ - Line: int(231), - Column: int(15), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(19), @@ -33391,7 +31755,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(19), @@ -33411,7 +31775,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(19), @@ -33452,9 +31816,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -33464,7 +31827,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(30), @@ -33486,7 +31849,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(30), @@ -33506,7 +31869,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(30), @@ -33547,9 +31910,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "filter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -33559,7 +31921,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(41), @@ -33577,33 +31939,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(231), - Column: int(50), - }, - End: ast.Location{ - Line: int(231), - Column: int(51), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(53), @@ -33624,7 +31975,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(53), @@ -33648,7 +31999,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(58), @@ -33675,7 +32026,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(231), Column: int(61), @@ -33716,18 +32067,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(231), - Column: int(3), - }, - End: ast.Location{ - Line: int(231), - Column: int(66), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -33752,7 +32091,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -33775,52 +32113,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(233), - Column: int(7), - }, - End: ast.Location{ - Line: int(233), - Column: int(8), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(233), - Column: int(10), - }, - End: ast.Location{ - Line: int(233), - Column: int(11), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(5), @@ -33849,7 +32162,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(8), @@ -33871,7 +32184,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(8), @@ -33892,7 +32205,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(8), @@ -33912,7 +32225,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(8), @@ -33953,9 +32266,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -33965,7 +32277,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(21), @@ -33999,7 +32311,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(27), @@ -34020,7 +32332,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(27), @@ -34040,7 +32352,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(27), @@ -34081,9 +32393,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -34093,7 +32404,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(234), Column: int(40), @@ -34127,7 +32438,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(235), Column: int(7), @@ -34149,7 +32460,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(235), Column: int(7), @@ -34169,7 +32480,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(235), Column: int(7), @@ -34217,9 +32528,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "modulo", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -34229,7 +32539,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(235), Column: int(18), @@ -34254,7 +32564,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(235), Column: int(21), @@ -34294,7 +32604,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(236), Column: int(10), @@ -34316,7 +32626,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(236), Column: int(13), @@ -34337,7 +32647,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(236), Column: int(13), @@ -34357,7 +32667,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(236), Column: int(13), @@ -34398,9 +32708,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -34410,7 +32719,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(236), Column: int(26), @@ -34443,7 +32752,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(237), Column: int(7), @@ -34465,7 +32774,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(237), Column: int(7), @@ -34485,7 +32794,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(237), Column: int(7), @@ -34533,9 +32842,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "format", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -34545,7 +32853,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(237), Column: int(18), @@ -34570,7 +32878,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(237), Column: int(21), @@ -34610,7 +32918,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(7), @@ -34639,7 +32947,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(13), @@ -34661,7 +32969,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(13), @@ -34683,7 +32991,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(13), @@ -34704,7 +33012,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(13), @@ -34725,7 +33033,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(13), @@ -34743,14 +33051,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "Operator % cannot be used on types ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(53), @@ -34771,7 +33078,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(53), @@ -34791,7 +33098,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(53), @@ -34832,9 +33139,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -34844,7 +33150,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(62), @@ -34879,7 +33185,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(67), @@ -34897,7 +33203,6 @@ var _StdAst = &ast.DesugaredObject{ Value: " and ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -34905,7 +33210,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(77), @@ -34926,7 +33231,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(77), @@ -34946,7 +33251,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(77), @@ -34987,9 +33292,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -34999,7 +33303,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(86), @@ -35034,7 +33338,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(239), Column: int(91), @@ -35052,7 +33356,6 @@ var _StdAst = &ast.DesugaredObject{ Value: ".", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -35060,18 +33363,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(233), - Column: int(3), - }, - End: ast.Location{ - Line: int(239), - Column: int(94), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -35096,7 +33387,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "map", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -35119,52 +33409,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(241), - Column: int(7), - }, - End: ast.Location{ - Line: int(241), - Column: int(11), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(241), - Column: int(13), - }, - End: ast.Location{ - Line: int(241), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(242), Column: int(5), @@ -35193,7 +33458,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(242), Column: int(8), @@ -35215,7 +33480,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(242), Column: int(9), @@ -35236,7 +33501,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(242), Column: int(9), @@ -35256,7 +33521,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(242), Column: int(9), @@ -35297,9 +33562,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -35309,7 +33573,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(242), Column: int(24), @@ -35343,7 +33607,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(243), Column: int(7), @@ -35371,7 +33635,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(243), Column: int(14), @@ -35392,7 +33656,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(243), Column: int(14), @@ -35410,14 +33674,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.map first param must be function, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(243), Column: int(61), @@ -35438,7 +33701,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(243), Column: int(61), @@ -35458,7 +33721,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(243), Column: int(61), @@ -35499,9 +33762,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -35511,7 +33773,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(243), Column: int(70), @@ -35553,7 +33815,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(10), @@ -35575,7 +33837,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(13), @@ -35596,7 +33858,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(13), @@ -35618,7 +33880,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(14), @@ -35639,7 +33901,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(14), @@ -35659,7 +33921,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(14), @@ -35700,9 +33962,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -35712,7 +33973,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(26), @@ -35747,7 +34008,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(34), @@ -35769,7 +34030,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(35), @@ -35790,7 +34051,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(35), @@ -35810,7 +34071,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(35), @@ -35851,9 +34112,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -35863,7 +34123,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(244), Column: int(48), @@ -35898,7 +34158,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(245), Column: int(7), @@ -35926,7 +34186,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(245), Column: int(14), @@ -35947,7 +34207,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(245), Column: int(14), @@ -35965,14 +34225,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.map second param must be array / string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(245), Column: int(68), @@ -35993,7 +34252,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(245), Column: int(68), @@ -36013,7 +34272,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(245), Column: int(68), @@ -36054,9 +34313,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -36066,7 +34324,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(245), Column: int(77), @@ -36108,7 +34366,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(7), @@ -36130,7 +34388,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(7), @@ -36150,7 +34408,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(7), @@ -36198,9 +34456,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -36210,7 +34467,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(21), @@ -36231,7 +34488,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(21), @@ -36251,7 +34508,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(21), @@ -36292,9 +34549,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -36304,7 +34560,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(32), @@ -36339,7 +34595,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(38), @@ -36358,33 +34614,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(247), - Column: int(47), - }, - End: ast.Location{ - Line: int(247), - Column: int(48), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(50), @@ -36406,7 +34651,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(50), @@ -36432,7 +34677,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(55), @@ -36453,7 +34698,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(55), @@ -36476,7 +34721,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(247), Column: int(59), @@ -36523,18 +34768,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(241), - Column: int(3), - }, - End: ast.Location{ - Line: int(247), - Column: int(63), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -36559,7 +34792,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mapWithIndex", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -36582,52 +34814,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(249), - Column: int(16), - }, - End: ast.Location{ - Line: int(249), - Column: int(20), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(249), - Column: int(22), - }, - End: ast.Location{ - Line: int(249), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(250), Column: int(5), @@ -36656,7 +34863,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(250), Column: int(8), @@ -36678,7 +34885,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(250), Column: int(9), @@ -36699,7 +34906,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(250), Column: int(9), @@ -36719,7 +34926,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(250), Column: int(9), @@ -36760,9 +34967,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -36772,7 +34978,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(250), Column: int(24), @@ -36806,7 +35012,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(251), Column: int(7), @@ -36834,7 +35040,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(251), Column: int(14), @@ -36855,7 +35061,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(251), Column: int(14), @@ -36873,14 +35079,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.mapWithIndex first param must be function, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(251), Column: int(70), @@ -36901,7 +35106,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(251), Column: int(70), @@ -36921,7 +35126,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(251), Column: int(70), @@ -36962,9 +35167,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -36974,7 +35178,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(251), Column: int(79), @@ -37016,7 +35220,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(10), @@ -37038,7 +35242,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(13), @@ -37059,7 +35263,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(13), @@ -37081,7 +35285,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(14), @@ -37102,7 +35306,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(14), @@ -37122,7 +35326,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(14), @@ -37163,9 +35367,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -37175,7 +35378,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(26), @@ -37210,7 +35413,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(34), @@ -37232,7 +35435,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(35), @@ -37253,7 +35456,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(35), @@ -37273,7 +35476,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(35), @@ -37314,9 +35517,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -37326,7 +35528,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(252), Column: int(48), @@ -37361,7 +35563,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(253), Column: int(7), @@ -37389,7 +35591,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(253), Column: int(14), @@ -37410,7 +35612,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(253), Column: int(14), @@ -37428,14 +35630,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.mapWithIndex second param must be array, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(253), Column: int(68), @@ -37456,7 +35657,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(253), Column: int(68), @@ -37476,7 +35677,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(253), Column: int(68), @@ -37517,9 +35718,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -37529,7 +35729,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(253), Column: int(77), @@ -37571,7 +35771,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(7), @@ -37593,7 +35793,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(7), @@ -37613,7 +35813,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(7), @@ -37661,9 +35861,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -37673,7 +35872,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(21), @@ -37694,7 +35893,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(21), @@ -37714,7 +35913,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(21), @@ -37755,9 +35954,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -37767,7 +35965,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(32), @@ -37802,7 +36000,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(38), @@ -37821,33 +36019,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(255), - Column: int(47), - }, - End: ast.Location{ - Line: int(255), - Column: int(48), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(50), @@ -37869,7 +36056,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(50), @@ -37895,7 +36082,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(55), @@ -37920,7 +36107,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(58), @@ -37941,7 +36128,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(58), @@ -37964,7 +36151,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(255), Column: int(62), @@ -38011,18 +36198,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(249), - Column: int(3), - }, - End: ast.Location{ - Line: int(255), - Column: int(66), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -38047,7 +36222,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mapWithKey", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -38070,52 +36244,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(257), - Column: int(14), - }, - End: ast.Location{ - Line: int(257), - Column: int(18), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "obj", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(257), - Column: int(20), - }, - End: ast.Location{ - Line: int(257), - Column: int(23), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "obj", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(258), Column: int(5), @@ -38144,7 +36293,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(258), Column: int(8), @@ -38166,7 +36315,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(258), Column: int(9), @@ -38187,7 +36336,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(258), Column: int(9), @@ -38207,7 +36356,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(258), Column: int(9), @@ -38248,9 +36397,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -38260,7 +36408,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(258), Column: int(24), @@ -38294,7 +36442,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(259), Column: int(7), @@ -38322,7 +36470,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(259), Column: int(14), @@ -38343,7 +36491,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(259), Column: int(14), @@ -38361,14 +36509,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.mapWithKey first param must be function, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(259), Column: int(68), @@ -38389,7 +36536,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(259), Column: int(68), @@ -38409,7 +36556,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(259), Column: int(68), @@ -38450,9 +36597,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -38462,7 +36608,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(259), Column: int(77), @@ -38504,7 +36650,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(260), Column: int(10), @@ -38526,7 +36672,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(260), Column: int(13), @@ -38548,7 +36694,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(260), Column: int(14), @@ -38569,7 +36715,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(260), Column: int(14), @@ -38589,7 +36735,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(260), Column: int(14), @@ -38630,9 +36776,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -38642,7 +36787,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(260), Column: int(27), @@ -38676,7 +36821,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(261), Column: int(7), @@ -38704,7 +36849,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(261), Column: int(14), @@ -38725,7 +36870,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(261), Column: int(14), @@ -38743,14 +36888,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.mapWithKey second param must be object, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(261), Column: int(67), @@ -38771,7 +36915,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(261), Column: int(67), @@ -38791,7 +36935,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(261), Column: int(67), @@ -38832,9 +36976,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -38844,7 +36987,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(261), Column: int(76), @@ -38888,14 +37031,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(263), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(263), - Column: int(62), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -38969,7 +37112,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "$objectFlatMerge", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -38983,14 +37125,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(263), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(263), - Column: int(62), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -39064,7 +37206,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -39095,26 +37236,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -39145,7 +37275,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(7), @@ -39178,7 +37308,7 @@ var _StdAst = &ast.DesugaredObject{ Name: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(10), @@ -39200,7 +37330,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(14), @@ -39222,7 +37352,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(14), @@ -39248,7 +37378,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(19), @@ -39273,7 +37403,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(22), @@ -39294,7 +37424,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(22), @@ -39317,7 +37447,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(26), @@ -39350,18 +37480,6 @@ var _StdAst = &ast.DesugaredObject{ TailStrictFodder: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(263), - Column: int(9), - }, - End: ast.Location{ - Line: int(263), - Column: int(29), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -39379,7 +37497,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(39), @@ -39400,7 +37518,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(39), @@ -39420,7 +37538,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(39), @@ -39461,9 +37579,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -39473,7 +37590,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(263), Column: int(56), @@ -39526,18 +37643,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(257), - Column: int(3), - }, - End: ast.Location{ - Line: int(263), - Column: int(62), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -39562,7 +37667,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -39585,52 +37689,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(265), - Column: int(11), - }, - End: ast.Location{ - Line: int(265), - Column: int(15), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(265), - Column: int(17), - }, - End: ast.Location{ - Line: int(265), - Column: int(20), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(266), Column: int(5), @@ -39659,7 +37738,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(266), Column: int(8), @@ -39681,7 +37760,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(266), Column: int(9), @@ -39702,7 +37781,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(266), Column: int(9), @@ -39722,7 +37801,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(266), Column: int(9), @@ -39763,9 +37842,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -39775,7 +37853,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(266), Column: int(24), @@ -39809,7 +37887,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(267), Column: int(7), @@ -39837,7 +37915,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(267), Column: int(14), @@ -39858,7 +37936,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(267), Column: int(14), @@ -39876,14 +37954,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.flatMap first param must be function, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(267), Column: int(65), @@ -39904,7 +37981,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(267), Column: int(65), @@ -39924,7 +38001,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(267), Column: int(65), @@ -39965,9 +38042,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -39977,7 +38053,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(267), Column: int(74), @@ -40019,7 +38095,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(268), Column: int(10), @@ -40041,7 +38117,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(268), Column: int(13), @@ -40062,7 +38138,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(268), Column: int(13), @@ -40082,7 +38158,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(268), Column: int(13), @@ -40123,9 +38199,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -40135,7 +38210,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(268), Column: int(25), @@ -40168,7 +38243,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(7), @@ -40190,7 +38265,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(7), @@ -40210,7 +38285,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(7), @@ -40258,9 +38333,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "flattenArrays", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -40270,7 +38344,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(25), @@ -40292,7 +38366,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(25), @@ -40312,7 +38386,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(25), @@ -40353,9 +38427,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -40365,7 +38438,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(39), @@ -40386,7 +38459,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(39), @@ -40406,7 +38479,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(39), @@ -40447,9 +38520,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -40459,7 +38531,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(50), @@ -40494,7 +38566,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(56), @@ -40513,33 +38585,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(269), - Column: int(65), - }, - End: ast.Location{ - Line: int(269), - Column: int(66), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(68), @@ -40561,7 +38622,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(68), @@ -40587,7 +38648,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(73), @@ -40608,7 +38669,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(73), @@ -40631,7 +38692,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(269), Column: int(77), @@ -40695,7 +38756,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(270), Column: int(10), @@ -40717,7 +38778,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(270), Column: int(13), @@ -40738,7 +38799,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(270), Column: int(13), @@ -40758,7 +38819,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(270), Column: int(13), @@ -40799,9 +38860,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -40811,7 +38871,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(270), Column: int(26), @@ -40844,7 +38904,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(7), @@ -40866,7 +38926,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(7), @@ -40886,7 +38946,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(7), @@ -40934,9 +38994,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -40946,7 +39005,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(16), @@ -40964,7 +39023,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -40972,7 +39030,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(20), @@ -40994,7 +39052,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(20), @@ -41014,7 +39072,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(20), @@ -41055,9 +39113,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -41067,7 +39124,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(34), @@ -41088,7 +39145,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(34), @@ -41108,7 +39165,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(34), @@ -41149,9 +39206,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -41161,7 +39217,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(45), @@ -41196,7 +39252,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(51), @@ -41215,33 +39271,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(271), - Column: int(60), - }, - End: ast.Location{ - Line: int(271), - Column: int(61), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(63), @@ -41263,7 +39308,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(63), @@ -41289,7 +39334,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(68), @@ -41310,7 +39355,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(68), @@ -41333,7 +39378,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(271), Column: int(72), @@ -41397,7 +39442,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(272), Column: int(10), @@ -41418,7 +39463,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(272), Column: int(17), @@ -41439,7 +39484,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(272), Column: int(17), @@ -41457,14 +39502,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "std.flatMap second param must be array / string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(272), Column: int(75), @@ -41485,7 +39529,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(272), Column: int(75), @@ -41505,7 +39549,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(272), Column: int(75), @@ -41546,9 +39590,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -41558,7 +39601,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(272), Column: int(84), @@ -41594,18 +39637,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(265), - Column: int(3), - }, - End: ast.Location{ - Line: int(272), - Column: int(89), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -41630,7 +39661,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -41653,52 +39683,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "sep", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(274), - Column: int(8), - }, - End: ast.Location{ - Line: int(274), - Column: int(11), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(274), - Column: int(13), - }, - End: ast.Location{ - Line: int(274), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "sep", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(275), Column: int(5), @@ -41732,7 +39737,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(275), Column: int(11), @@ -41752,90 +39757,37 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(275), - Column: int(15), - }, - End: ast.Location{ - Line: int(275), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(275), - Column: int(20), - }, - End: ast.Location{ - Line: int(275), - Column: int(21), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "first", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(275), - Column: int(23), - }, - End: ast.Location{ - Line: int(275), - Column: int(28), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "first", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "running", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(275), - Column: int(30), - }, - End: ast.Location{ - Line: int(275), - Column: int(37), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "running", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(276), Column: int(7), @@ -41868,7 +39820,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(276), Column: int(10), @@ -41890,7 +39842,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(276), Column: int(10), @@ -41914,7 +39866,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(276), Column: int(15), @@ -41935,7 +39887,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(276), Column: int(15), @@ -41955,7 +39907,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(276), Column: int(15), @@ -41996,9 +39948,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -42008,7 +39959,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(276), Column: int(26), @@ -42042,7 +39993,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(277), Column: int(9), @@ -42079,7 +40030,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(278), Column: int(12), @@ -42105,7 +40056,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(278), Column: int(15), @@ -42126,7 +40077,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(278), Column: int(15), @@ -42147,7 +40098,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(278), Column: int(15), @@ -42170,7 +40121,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(278), Column: int(19), @@ -42197,7 +40148,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(278), Column: int(25), @@ -42218,7 +40169,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(9), @@ -42242,7 +40193,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(9), @@ -42275,7 +40226,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(13), @@ -42300,7 +40251,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(18), @@ -42320,7 +40271,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(18), @@ -42344,7 +40295,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(22), @@ -42359,6 +40310,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3116, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -42368,7 +40320,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(25), @@ -42393,7 +40345,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(279), Column: int(32), @@ -42433,7 +40385,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(12), @@ -42459,7 +40411,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(15), @@ -42482,7 +40434,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(15), @@ -42504,7 +40456,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(15), @@ -42524,7 +40476,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(15), @@ -42565,9 +40517,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -42577,7 +40528,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(24), @@ -42598,7 +40549,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(24), @@ -42621,7 +40572,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(28), @@ -42658,7 +40609,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(35), @@ -42679,7 +40630,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(35), @@ -42699,7 +40650,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(35), @@ -42740,9 +40691,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -42752,7 +40702,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(280), Column: int(44), @@ -42786,7 +40736,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(9), @@ -42818,14 +40768,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(281), - Column: int(15), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(281), - Column: int(87), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -42900,7 +40850,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -42912,7 +40861,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(15), @@ -42930,7 +40879,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "expected %s but arr[%d] was %s ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -42938,7 +40886,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(51), @@ -42963,7 +40911,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(52), @@ -42984,7 +40932,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(52), @@ -43004,7 +40952,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(52), @@ -43045,9 +40993,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -43057,7 +41004,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(61), @@ -43092,7 +41039,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(67), @@ -43117,7 +41064,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(70), @@ -43139,7 +41086,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(70), @@ -43159,7 +41106,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(70), @@ -43200,9 +41147,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -43212,7 +41158,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(79), @@ -43233,7 +41179,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(79), @@ -43256,7 +41202,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(281), Column: int(83), @@ -43316,7 +41262,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(282), Column: int(12), @@ -43341,7 +41287,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(282), Column: int(15), @@ -43364,7 +41310,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(9), @@ -43387,7 +41333,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(9), @@ -43420,7 +41366,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(13), @@ -43445,7 +41391,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(18), @@ -43465,7 +41411,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(18), @@ -43489,7 +41435,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(22), @@ -43504,6 +41450,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3214, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -43513,7 +41460,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(25), @@ -43536,7 +41483,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(32), @@ -43558,7 +41505,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(32), @@ -43582,7 +41529,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(42), @@ -43603,7 +41550,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(42), @@ -43626,7 +41573,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(283), Column: int(46), @@ -43670,7 +41617,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(9), @@ -43694,7 +41641,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(9), @@ -43727,7 +41674,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(13), @@ -43752,7 +41699,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(18), @@ -43772,7 +41719,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(18), @@ -43796,7 +41743,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(22), @@ -43811,6 +41758,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3240, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -43820,7 +41768,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(25), @@ -43843,7 +41791,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(32), @@ -43866,7 +41814,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(32), @@ -43887,7 +41835,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(32), @@ -43911,7 +41859,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(42), @@ -43936,7 +41884,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(48), @@ -43957,7 +41905,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(48), @@ -43980,7 +41928,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(285), Column: int(52), @@ -44020,24 +41968,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(286), Column: int(5), @@ -44067,7 +42003,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(286), Column: int(8), @@ -44089,7 +42025,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(286), Column: int(9), @@ -44110,7 +42046,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(286), Column: int(9), @@ -44130,7 +42066,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(286), Column: int(9), @@ -44171,9 +42107,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -44183,7 +42118,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(286), Column: int(21), @@ -44217,7 +42152,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(287), Column: int(7), @@ -44245,7 +42180,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(287), Column: int(13), @@ -44266,7 +42201,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(287), Column: int(13), @@ -44284,14 +42219,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "join second parameter should be array, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(287), Column: int(61), @@ -44312,7 +42246,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(287), Column: int(61), @@ -44332,7 +42266,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(287), Column: int(61), @@ -44373,9 +42307,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -44385,7 +42318,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(287), Column: int(70), @@ -44427,7 +42360,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(288), Column: int(10), @@ -44450,7 +42383,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(288), Column: int(13), @@ -44471,7 +42404,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(288), Column: int(13), @@ -44491,7 +42424,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(288), Column: int(13), @@ -44532,9 +42465,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -44544,7 +42476,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(288), Column: int(26), @@ -44577,7 +42509,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(289), Column: int(7), @@ -44598,7 +42530,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(289), Column: int(7), @@ -44631,7 +42563,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(289), Column: int(11), @@ -44656,7 +42588,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(289), Column: int(16), @@ -44671,6 +42603,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3316, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -44679,7 +42612,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(289), Column: int(19), @@ -44702,7 +42635,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(289), Column: int(25), @@ -44720,7 +42653,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -44743,7 +42675,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(290), Column: int(10), @@ -44766,7 +42698,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(290), Column: int(13), @@ -44787,7 +42719,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(290), Column: int(13), @@ -44807,7 +42739,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(290), Column: int(13), @@ -44848,9 +42780,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -44860,7 +42791,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(290), Column: int(25), @@ -44893,7 +42824,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(291), Column: int(7), @@ -44914,7 +42845,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(291), Column: int(7), @@ -44947,7 +42878,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(291), Column: int(11), @@ -44972,7 +42903,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(291), Column: int(16), @@ -44987,6 +42918,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3342, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -44995,7 +42927,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(291), Column: int(19), @@ -45018,7 +42950,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(291), Column: int(25), @@ -45058,7 +42990,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(293), Column: int(7), @@ -45086,7 +43018,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(293), Column: int(13), @@ -45107,7 +43039,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(293), Column: int(13), @@ -45125,14 +43057,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "join first parameter should be string or array, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(293), Column: int(70), @@ -45153,7 +43084,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(293), Column: int(70), @@ -45173,7 +43104,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(293), Column: int(70), @@ -45214,9 +43145,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -45226,7 +43156,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(293), Column: int(79), @@ -45263,18 +43193,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(274), - Column: int(3), - }, - End: ast.Location{ - Line: int(293), - Column: int(83), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -45299,7 +43217,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "lines", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -45322,33 +43239,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(295), - Column: int(9), - }, - End: ast.Location{ - Line: int(295), - Column: int(12), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(5), @@ -45369,7 +43275,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(5), @@ -45389,7 +43295,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(5), @@ -45437,9 +43343,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -45449,7 +43354,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(14), @@ -45467,7 +43372,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -45475,7 +43379,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(20), @@ -45495,7 +43399,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(20), @@ -45519,7 +43423,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(26), @@ -45539,7 +43443,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(296), Column: int(27), @@ -45557,7 +43461,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -45578,18 +43481,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(295), - Column: int(3), - }, - End: ast.Location{ - Line: int(296), - Column: int(31), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -45614,7 +43505,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "deepJoin", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -45637,33 +43527,22 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(298), - Column: int(12), - }, - End: ast.Location{ - Line: int(298), - Column: int(15), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(299), Column: int(5), @@ -45691,7 +43570,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(299), Column: int(8), @@ -45712,7 +43591,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(299), Column: int(8), @@ -45732,7 +43611,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(299), Column: int(8), @@ -45773,9 +43652,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -45785,7 +43663,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(299), Column: int(21), @@ -45818,7 +43696,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(300), Column: int(7), @@ -45855,7 +43733,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(301), Column: int(10), @@ -45876,7 +43754,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(301), Column: int(13), @@ -45897,7 +43775,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(301), Column: int(13), @@ -45917,7 +43795,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(301), Column: int(13), @@ -45958,9 +43836,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -45970,7 +43847,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(301), Column: int(25), @@ -46003,7 +43880,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(7), @@ -46024,7 +43901,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(7), @@ -46044,7 +43921,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(7), @@ -46092,9 +43969,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -46104,7 +43980,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(16), @@ -46122,7 +43998,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -46132,14 +44007,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(302), - Column: int(20), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(302), - Column: int(50), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -46212,7 +44087,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -46242,26 +44116,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -46291,7 +44154,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(21), @@ -46312,7 +44175,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(21), @@ -46332,7 +44195,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(21), @@ -46373,9 +44236,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "deepJoin", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -46385,7 +44247,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(34), @@ -46427,7 +44289,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(302), Column: int(46), @@ -46477,7 +44339,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(304), Column: int(7), @@ -46507,14 +44369,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(304), - Column: int(13), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(304), - Column: int(63), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -46587,7 +44449,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -46599,7 +44460,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(304), Column: int(13), @@ -46617,7 +44478,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "Expected string or array, got %s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -46625,7 +44485,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(304), Column: int(50), @@ -46646,7 +44506,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(304), Column: int(50), @@ -46666,7 +44526,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(304), Column: int(50), @@ -46707,9 +44567,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -46719,7 +44578,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(304), Column: int(59), @@ -46763,18 +44622,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(298), - Column: int(3), - }, - End: ast.Location{ - Line: int(304), - Column: int(63), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -46799,7 +44646,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "format", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -46822,58 +44668,33 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(307), - Column: int(10), - }, - End: ast.Location{ - Line: int(307), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "vals", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(307), - Column: int(15), - }, - End: ast.Location{ - Line: int(307), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "vals", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(313), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -46886,7 +44707,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -46894,7 +44715,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -46902,7 +44723,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(1), Indent: int(4), Comment: []string{ @@ -46925,7 +44746,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(313), Column: int(11), @@ -46943,45 +44764,20 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(313), - Column: int(33), - }, - End: ast.Location{ - Line: int(313), - Column: int(36), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(313), - Column: int(38), - }, - End: ast.Location{ - Line: int(313), - Column: int(39), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -47010,7 +44806,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(314), Column: int(14), @@ -47032,7 +44828,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(314), Column: int(14), @@ -47056,7 +44852,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(314), Column: int(18), @@ -47077,7 +44873,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(314), Column: int(18), @@ -47097,7 +44893,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(314), Column: int(18), @@ -47138,9 +44934,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -47150,7 +44945,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(314), Column: int(29), @@ -47184,7 +44979,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(315), Column: int(7), @@ -47218,7 +45013,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(315), Column: int(17), @@ -47239,7 +45034,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(315), Column: int(17), @@ -47262,7 +45057,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(315), Column: int(21), @@ -47286,24 +45081,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(315), - Column: int(13), - }, - End: ast.Location{ - Line: int(315), - Column: int(23), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(316), Column: int(7), @@ -47333,7 +45116,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(316), Column: int(10), @@ -47353,7 +45136,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(316), Column: int(10), @@ -47377,7 +45160,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(316), Column: int(15), @@ -47395,14 +45178,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "(", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(317), Column: int(9), @@ -47436,7 +45218,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(317), Column: int(15), @@ -47455,71 +45237,32 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(317), - Column: int(23), - }, - End: ast.Location{ - Line: int(317), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(317), - Column: int(28), - }, - End: ast.Location{ - Line: int(317), - Column: int(29), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(317), - Column: int(31), - }, - End: ast.Location{ - Line: int(317), - Column: int(32), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(318), Column: int(11), @@ -47550,7 +45293,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(318), Column: int(14), @@ -47572,7 +45315,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(318), Column: int(14), @@ -47596,7 +45339,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(318), Column: int(19), @@ -47617,7 +45360,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(318), Column: int(19), @@ -47637,7 +45380,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(318), Column: int(19), @@ -47678,9 +45421,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -47690,7 +45432,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(318), Column: int(30), @@ -47724,7 +45466,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(319), Column: int(13), @@ -47749,7 +45491,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(319), Column: int(19), @@ -47767,7 +45509,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{ @@ -47781,7 +45522,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(321), Column: int(13), @@ -47816,7 +45557,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(321), Column: int(23), @@ -47837,7 +45578,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(321), Column: int(23), @@ -47860,7 +45601,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(321), Column: int(27), @@ -47884,24 +45625,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(321), - Column: int(19), - }, - End: ast.Location{ - Line: int(321), - Column: int(29), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(322), Column: int(13), @@ -47932,7 +45661,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(322), Column: int(16), @@ -47952,7 +45681,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(322), Column: int(16), @@ -47976,7 +45705,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(322), Column: int(21), @@ -47994,14 +45723,13 @@ var _StdAst = &ast.DesugaredObject{ Value: ")", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(15), @@ -48025,7 +45753,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(15), @@ -48058,7 +45786,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(23), @@ -48083,7 +45811,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(28), @@ -48103,7 +45831,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(28), @@ -48127,7 +45855,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(32), @@ -48142,6 +45870,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3598, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -48151,7 +45880,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(35), @@ -48172,7 +45901,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(35), @@ -48196,7 +45925,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(323), Column: int(39), @@ -48237,7 +45966,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(325), Column: int(15), @@ -48287,12 +46016,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(325), Column: int(20), @@ -48312,7 +46040,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(325), Column: int(20), @@ -48336,7 +46064,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(325), Column: int(24), @@ -48351,22 +46079,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3618, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(325), - Column: int(17), - }, - End: ast.Location{ - Line: int(325), - Column: int(25), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -48391,12 +46108,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(325), Column: int(30), @@ -48416,18 +46132,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "v", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(325), - Column: int(27), - }, - End: ast.Location{ - Line: int(325), - Column: int(31), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -48438,24 +46142,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(326), Column: int(9), @@ -48477,7 +46169,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(326), Column: int(9), @@ -48510,7 +46202,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(326), Column: int(17), @@ -48535,7 +46227,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(326), Column: int(22), @@ -48555,7 +46247,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(326), Column: int(22), @@ -48579,7 +46271,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(326), Column: int(26), @@ -48594,6 +46286,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3633, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -48603,7 +46296,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(326), Column: int(29), @@ -48621,7 +46314,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -48645,7 +46337,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(328), Column: int(9), @@ -48694,12 +46386,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(328), Column: int(14), @@ -48719,18 +46410,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "i", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(328), - Column: int(11), - }, - End: ast.Location{ - Line: int(328), - Column: int(15), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -48755,12 +46434,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(328), Column: int(20), @@ -48777,18 +46455,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(328), - Column: int(17), - }, - End: ast.Location{ - Line: int(328), - Column: int(24), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -48801,14 +46467,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(314), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(328), - Column: int(26), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -48817,7 +46483,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(314), Column: int(36), @@ -48835,37 +46501,24 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(330), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -48894,7 +46547,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(330), Column: int(11), @@ -48912,52 +46565,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(330), - Column: int(28), - }, - End: ast.Location{ - Line: int(330), - Column: int(31), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(330), - Column: int(33), - }, - End: ast.Location{ - Line: int(330), - Column: int(34), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(331), Column: int(7), @@ -48991,7 +46619,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(331), Column: int(13), @@ -49010,64 +46638,25 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(331), - Column: int(21), - }, - End: ast.Location{ - Line: int(331), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(331), - Column: int(26), - }, - End: ast.Location{ - Line: int(331), - Column: int(27), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(331), - Column: int(29), - }, - End: ast.Location{ - Line: int(331), - Column: int(30), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -49098,7 +46687,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(332), Column: int(16), @@ -49120,7 +46709,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(332), Column: int(16), @@ -49144,7 +46733,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(332), Column: int(20), @@ -49165,7 +46754,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(332), Column: int(20), @@ -49185,7 +46774,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(332), Column: int(20), @@ -49226,9 +46815,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -49238,7 +46826,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(332), Column: int(31), @@ -49272,7 +46860,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(333), Column: int(9), @@ -49307,7 +46895,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(333), Column: int(19), @@ -49328,7 +46916,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(333), Column: int(19), @@ -49351,7 +46939,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(333), Column: int(23), @@ -49375,24 +46963,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(333), - Column: int(15), - }, - End: ast.Location{ - Line: int(333), - Column: int(25), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(334), Column: int(9), @@ -49423,7 +46999,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(334), Column: int(12), @@ -49443,7 +47019,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(334), Column: int(12), @@ -49467,7 +47043,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(334), Column: int(17), @@ -49485,14 +47061,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "#", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(11), @@ -49515,7 +47090,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(11), @@ -49548,7 +47123,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(19), @@ -49573,7 +47148,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(24), @@ -49593,7 +47168,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(24), @@ -49617,7 +47192,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(28), @@ -49632,6 +47207,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3715, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -49641,7 +47217,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(31), @@ -49661,7 +47237,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(31), @@ -49685,7 +47261,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(33), @@ -49725,12 +47301,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(335), Column: int(40), @@ -49748,18 +47323,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(335), - Column: int(35), - }, - End: ast.Location{ - Line: int(335), - Column: int(44), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -49786,7 +47349,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(336), Column: int(14), @@ -49810,7 +47373,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(336), Column: int(17), @@ -49830,7 +47393,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(336), Column: int(17), @@ -49854,7 +47417,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(336), Column: int(22), @@ -49872,14 +47435,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(11), @@ -49902,7 +47464,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(11), @@ -49935,7 +47497,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(19), @@ -49960,7 +47522,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(24), @@ -49980,7 +47542,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(24), @@ -50004,7 +47566,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(28), @@ -50019,6 +47581,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3746, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -50028,7 +47591,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(31), @@ -50048,7 +47611,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(31), @@ -50072,7 +47635,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(33), @@ -50112,12 +47675,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "zero", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(337), Column: int(41), @@ -50135,18 +47697,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(337), - Column: int(35), - }, - End: ast.Location{ - Line: int(337), - Column: int(45), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -50173,7 +47723,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(338), Column: int(14), @@ -50197,7 +47747,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(338), Column: int(17), @@ -50217,7 +47767,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(338), Column: int(17), @@ -50241,7 +47791,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(338), Column: int(22), @@ -50259,14 +47809,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "-", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(11), @@ -50289,7 +47838,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(11), @@ -50322,7 +47871,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(19), @@ -50347,7 +47896,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(24), @@ -50367,7 +47916,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(24), @@ -50391,7 +47940,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(28), @@ -50406,6 +47955,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3777, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -50415,7 +47965,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(31), @@ -50435,7 +47985,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(31), @@ -50459,7 +48009,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(33), @@ -50499,12 +48049,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "left", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(339), Column: int(41), @@ -50522,18 +48071,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(339), - Column: int(35), - }, - End: ast.Location{ - Line: int(339), - Column: int(45), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -50560,7 +48097,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(340), Column: int(14), @@ -50584,7 +48121,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(340), Column: int(17), @@ -50604,7 +48141,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(340), Column: int(17), @@ -50628,7 +48165,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(340), Column: int(22), @@ -50646,14 +48183,13 @@ var _StdAst = &ast.DesugaredObject{ Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(11), @@ -50676,7 +48212,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(11), @@ -50709,7 +48245,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(19), @@ -50734,7 +48270,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(24), @@ -50754,7 +48290,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(24), @@ -50778,7 +48314,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(28), @@ -50793,6 +48329,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3808, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -50802,7 +48339,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(31), @@ -50822,7 +48359,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(31), @@ -50846,7 +48383,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(33), @@ -50886,12 +48423,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(341), Column: int(42), @@ -50909,18 +48445,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(341), - Column: int(35), - }, - End: ast.Location{ - Line: int(341), - Column: int(46), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -50947,7 +48471,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(342), Column: int(14), @@ -50971,7 +48495,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(342), Column: int(17), @@ -50991,7 +48515,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(342), Column: int(17), @@ -51015,7 +48539,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(342), Column: int(22), @@ -51033,14 +48557,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "+", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(11), @@ -51063,7 +48586,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(11), @@ -51096,7 +48619,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(19), @@ -51121,7 +48644,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(24), @@ -51141,7 +48664,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(24), @@ -51165,7 +48688,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(28), @@ -51180,6 +48703,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3839, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -51189,7 +48713,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(31), @@ -51209,7 +48733,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(31), @@ -51233,7 +48757,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(33), @@ -51270,15 +48794,14 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(343), Column: int(41), @@ -51296,18 +48819,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(343), - Column: int(35), - }, - End: ast.Location{ - Line: int(343), - Column: int(45), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -51334,7 +48845,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(345), Column: int(11), @@ -51384,12 +48895,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(345), Column: int(16), @@ -51409,18 +48919,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "j", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(345), - Column: int(13), - }, - End: ast.Location{ - Line: int(345), - Column: int(17), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -51445,12 +48943,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(345), Column: int(22), @@ -51470,18 +48967,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "v", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(345), - Column: int(19), - }, - End: ast.Location{ - Line: int(345), - Column: int(23), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -51498,14 +48983,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(332), - Column: int(9), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(345), - Column: int(25), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -51514,7 +48999,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(332), Column: int(38), @@ -51532,31 +49017,18 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(7), @@ -51578,7 +49050,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(7), @@ -51611,7 +49083,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(15), @@ -51636,7 +49108,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(20), @@ -51661,7 +49133,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(23), @@ -51701,12 +49173,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(30), @@ -51724,18 +49195,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(346), - Column: int(25), - }, - End: ast.Location{ - Line: int(346), - Column: int(35), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -51760,12 +49219,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "zero", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(43), @@ -51783,18 +49241,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(346), - Column: int(37), - }, - End: ast.Location{ - Line: int(346), - Column: int(48), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -51819,12 +49265,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "left", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(56), @@ -51842,18 +49287,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(346), - Column: int(50), - }, - End: ast.Location{ - Line: int(346), - Column: int(61), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -51878,12 +49311,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(70), @@ -51901,18 +49333,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(346), - Column: int(63), - }, - End: ast.Location{ - Line: int(346), - Column: int(75), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -51934,15 +49354,14 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(346), Column: int(83), @@ -51960,18 +49379,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(346), - Column: int(77), - }, - End: ast.Location{ - Line: int(346), - Column: int(88), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -51990,30 +49397,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(348), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -52043,7 +49438,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(348), Column: int(11), @@ -52061,52 +49456,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(348), - Column: int(33), - }, - End: ast.Location{ - Line: int(348), - Column: int(36), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(348), - Column: int(38), - }, - End: ast.Location{ - Line: int(348), - Column: int(39), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(7), @@ -52135,7 +49505,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(10), @@ -52157,7 +49527,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(10), @@ -52179,7 +49549,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(10), @@ -52203,7 +49573,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(14), @@ -52224,7 +49594,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(14), @@ -52244,7 +49614,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(14), @@ -52285,9 +49655,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -52297,7 +49666,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(25), @@ -52332,7 +49701,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(33), @@ -52353,7 +49722,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(33), @@ -52374,7 +49743,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(33), @@ -52397,7 +49766,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(37), @@ -52424,7 +49793,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(349), Column: int(43), @@ -52442,7 +49811,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "*", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -52450,7 +49818,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(350), Column: int(9), @@ -52499,12 +49867,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(350), Column: int(14), @@ -52524,7 +49891,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(350), Column: int(14), @@ -52548,7 +49915,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(350), Column: int(18), @@ -52563,22 +49930,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3937, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(350), - Column: int(11), - }, - End: ast.Location{ - Line: int(350), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -52603,12 +49959,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(350), Column: int(24), @@ -52626,21 +49981,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "*", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(350), - Column: int(21), - }, - End: ast.Location{ - Line: int(350), - Column: int(27), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -52656,7 +49998,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(352), Column: int(9), @@ -52690,7 +50032,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(352), Column: int(15), @@ -52709,64 +50051,25 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(352), - Column: int(23), - }, - End: ast.Location{ - Line: int(352), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(352), - Column: int(28), - }, - End: ast.Location{ - Line: int(352), - Column: int(29), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(352), - Column: int(31), - }, - End: ast.Location{ - Line: int(352), - Column: int(32), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -52797,7 +50100,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(353), Column: int(18), @@ -52819,7 +50122,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(353), Column: int(18), @@ -52843,7 +50146,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(353), Column: int(22), @@ -52864,7 +50167,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(353), Column: int(22), @@ -52884,7 +50187,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(353), Column: int(22), @@ -52925,9 +50228,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -52937,7 +50239,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(353), Column: int(33), @@ -52971,7 +50273,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(354), Column: int(11), @@ -53006,7 +50308,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(354), Column: int(21), @@ -53027,7 +50329,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(354), Column: int(21), @@ -53050,7 +50352,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(354), Column: int(25), @@ -53074,24 +50376,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(354), - Column: int(17), - }, - End: ast.Location{ - Line: int(354), - Column: int(27), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(355), Column: int(11), @@ -53122,7 +50412,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(355), Column: int(14), @@ -53142,7 +50432,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(355), Column: int(14), @@ -53166,7 +50456,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(355), Column: int(19), @@ -53184,14 +50474,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(13), @@ -53214,7 +50503,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(13), @@ -53247,7 +50536,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(21), @@ -53272,7 +50561,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(26), @@ -53292,7 +50581,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(26), @@ -53316,7 +50605,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(30), @@ -53331,6 +50620,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3997, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -53340,7 +50630,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(33), @@ -53360,7 +50650,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(33), @@ -53380,7 +50670,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(33), @@ -53404,7 +50694,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(37), @@ -53419,6 +50709,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3997, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -53427,7 +50718,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(356), Column: int(42), @@ -53442,6 +50733,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p3997, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -53466,7 +50758,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(357), Column: int(16), @@ -53490,7 +50782,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(357), Column: int(19), @@ -53510,7 +50802,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(357), Column: int(19), @@ -53534,7 +50826,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(357), Column: int(24), @@ -53552,14 +50844,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "1", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(13), @@ -53582,7 +50873,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(13), @@ -53615,7 +50906,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(21), @@ -53640,7 +50931,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(26), @@ -53660,7 +50951,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(26), @@ -53684,7 +50975,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(30), @@ -53699,6 +50990,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4027, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -53708,7 +51000,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(33), @@ -53728,7 +51020,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(33), @@ -53748,7 +51040,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(33), @@ -53772,7 +51064,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(37), @@ -53787,6 +51079,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4027, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -53795,7 +51088,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(358), Column: int(42), @@ -53810,6 +51103,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4027, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -53834,7 +51128,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(359), Column: int(16), @@ -53858,7 +51152,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(359), Column: int(19), @@ -53878,7 +51172,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(359), Column: int(19), @@ -53902,7 +51196,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(359), Column: int(24), @@ -53920,14 +51214,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "2", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(13), @@ -53950,7 +51243,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(13), @@ -53983,7 +51276,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(21), @@ -54008,7 +51301,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(26), @@ -54028,7 +51321,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(26), @@ -54052,7 +51345,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(30), @@ -54067,6 +51360,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4057, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -54076,7 +51370,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(33), @@ -54096,7 +51390,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(33), @@ -54116,7 +51410,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(33), @@ -54140,7 +51434,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(37), @@ -54155,6 +51449,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4057, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -54163,7 +51458,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(360), Column: int(42), @@ -54178,6 +51473,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4057, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -54202,7 +51498,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(361), Column: int(16), @@ -54226,7 +51522,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(361), Column: int(19), @@ -54246,7 +51542,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(361), Column: int(19), @@ -54270,7 +51566,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(361), Column: int(24), @@ -54288,14 +51584,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "3", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(13), @@ -54318,7 +51613,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(13), @@ -54351,7 +51646,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(21), @@ -54376,7 +51671,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(26), @@ -54396,7 +51691,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(26), @@ -54420,7 +51715,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(30), @@ -54435,6 +51730,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4087, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -54444,7 +51740,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(33), @@ -54464,7 +51760,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(33), @@ -54484,7 +51780,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(33), @@ -54508,7 +51804,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(37), @@ -54523,6 +51819,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4087, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -54531,7 +51828,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(362), Column: int(42), @@ -54546,6 +51843,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4087, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -54570,7 +51868,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(363), Column: int(16), @@ -54594,7 +51892,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(363), Column: int(19), @@ -54614,7 +51912,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(363), Column: int(19), @@ -54638,7 +51936,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(363), Column: int(24), @@ -54656,14 +51954,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "4", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(13), @@ -54686,7 +51983,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(13), @@ -54719,7 +52016,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(21), @@ -54744,7 +52041,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(26), @@ -54764,7 +52061,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(26), @@ -54788,7 +52085,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(30), @@ -54803,6 +52100,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4117, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -54812,7 +52110,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(33), @@ -54832,7 +52130,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(33), @@ -54852,7 +52150,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(33), @@ -54876,7 +52174,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(37), @@ -54891,6 +52189,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4117, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -54899,7 +52198,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(364), Column: int(42), @@ -54914,6 +52213,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4117, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -54938,7 +52238,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(365), Column: int(16), @@ -54962,7 +52262,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(365), Column: int(19), @@ -54982,7 +52282,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(365), Column: int(19), @@ -55006,7 +52306,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(365), Column: int(24), @@ -55024,14 +52324,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "5", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(13), @@ -55054,7 +52353,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(13), @@ -55087,7 +52386,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(21), @@ -55112,7 +52411,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(26), @@ -55132,7 +52431,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(26), @@ -55156,7 +52455,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(30), @@ -55171,6 +52470,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4147, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -55180,7 +52480,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(33), @@ -55200,7 +52500,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(33), @@ -55220,7 +52520,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(33), @@ -55244,7 +52544,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(37), @@ -55259,6 +52559,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4147, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -55267,7 +52568,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(366), Column: int(42), @@ -55282,6 +52583,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4147, FreeVars: nil, }, + Value: float64(5), OriginalString: "5", }, }, @@ -55306,7 +52608,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(367), Column: int(16), @@ -55330,7 +52632,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(367), Column: int(19), @@ -55350,7 +52652,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(367), Column: int(19), @@ -55374,7 +52676,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(367), Column: int(24), @@ -55392,14 +52694,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "6", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(13), @@ -55422,7 +52723,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(13), @@ -55455,7 +52756,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(21), @@ -55480,7 +52781,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(26), @@ -55500,7 +52801,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(26), @@ -55524,7 +52825,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(30), @@ -55539,6 +52840,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4177, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -55548,7 +52850,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(33), @@ -55568,7 +52870,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(33), @@ -55588,7 +52890,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(33), @@ -55612,7 +52914,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(37), @@ -55627,6 +52929,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4177, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -55635,7 +52938,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(368), Column: int(42), @@ -55650,6 +52953,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4177, FreeVars: nil, }, + Value: float64(6), OriginalString: "6", }, }, @@ -55674,7 +52978,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(369), Column: int(16), @@ -55698,7 +53002,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(369), Column: int(19), @@ -55718,7 +53022,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(369), Column: int(19), @@ -55742,7 +53046,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(369), Column: int(24), @@ -55760,14 +53064,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "7", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(13), @@ -55790,7 +53093,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(13), @@ -55823,7 +53126,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(21), @@ -55848,7 +53151,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(26), @@ -55868,7 +53171,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(26), @@ -55892,7 +53195,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(30), @@ -55907,6 +53210,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4207, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -55916,7 +53220,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(33), @@ -55936,7 +53240,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(33), @@ -55956,7 +53260,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(33), @@ -55980,7 +53284,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(37), @@ -55995,6 +53299,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4207, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -56003,7 +53308,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(370), Column: int(42), @@ -56018,6 +53323,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4207, FreeVars: nil, }, + Value: float64(7), OriginalString: "7", }, }, @@ -56042,7 +53348,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(371), Column: int(16), @@ -56066,7 +53372,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(371), Column: int(19), @@ -56086,7 +53392,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(371), Column: int(19), @@ -56110,7 +53416,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(371), Column: int(24), @@ -56128,14 +53434,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "8", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(13), @@ -56158,7 +53463,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(13), @@ -56191,7 +53496,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(21), @@ -56216,7 +53521,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(26), @@ -56236,7 +53541,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(26), @@ -56260,7 +53565,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(30), @@ -56275,6 +53580,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4237, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -56284,7 +53590,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(33), @@ -56304,7 +53610,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(33), @@ -56324,7 +53630,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(33), @@ -56348,7 +53654,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(37), @@ -56363,6 +53669,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4237, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -56371,7 +53678,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(372), Column: int(42), @@ -56386,6 +53693,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4237, FreeVars: nil, }, + Value: float64(8), OriginalString: "8", }, }, @@ -56410,7 +53718,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(373), Column: int(16), @@ -56434,7 +53742,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(373), Column: int(19), @@ -56454,7 +53762,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(373), Column: int(19), @@ -56478,7 +53786,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(373), Column: int(24), @@ -56496,14 +53804,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "9", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(13), @@ -56526,7 +53833,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(13), @@ -56559,7 +53866,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(21), @@ -56584,7 +53891,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(26), @@ -56604,7 +53911,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(26), @@ -56628,7 +53935,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(30), @@ -56643,6 +53950,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4267, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -56652,7 +53960,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(33), @@ -56672,7 +53980,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(33), @@ -56692,7 +54000,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(33), @@ -56716,7 +54024,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(37), @@ -56731,6 +54039,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4267, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -56739,7 +54048,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(374), Column: int(42), @@ -56754,6 +54063,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4267, FreeVars: nil, }, + Value: float64(9), OriginalString: "9", }, }, @@ -56778,7 +54088,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(376), Column: int(13), @@ -56828,12 +54138,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(376), Column: int(18), @@ -56853,18 +54162,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "j", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(376), - Column: int(15), - }, - End: ast.Location{ - Line: int(376), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -56889,12 +54186,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(376), Column: int(24), @@ -56914,18 +54210,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "v", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(376), - Column: int(21), - }, - End: ast.Location{ - Line: int(376), - Column: int(25), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -56947,14 +54231,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(353), - Column: int(11), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(376), - Column: int(27), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -56963,7 +54247,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(353), Column: int(40), @@ -56981,31 +54265,18 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(377), Column: int(9), @@ -57027,7 +54298,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(377), Column: int(9), @@ -57060,7 +54331,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(377), Column: int(17), @@ -57085,7 +54356,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(377), Column: int(22), @@ -57110,7 +54381,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(377), Column: int(25), @@ -57125,6 +54396,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4303, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: nil, @@ -57142,30 +54414,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(379), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -57196,7 +54456,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(379), Column: int(11), @@ -57215,45 +54475,20 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(379), - Column: int(31), - }, - End: ast.Location{ - Line: int(379), - Column: int(34), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(379), - Column: int(36), - }, - End: ast.Location{ - Line: int(379), - Column: int(37), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -57283,7 +54518,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(380), Column: int(14), @@ -57305,7 +54540,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(380), Column: int(14), @@ -57329,7 +54564,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(380), Column: int(18), @@ -57350,7 +54585,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(380), Column: int(18), @@ -57370,7 +54605,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(380), Column: int(18), @@ -57411,9 +54646,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -57423,7 +54657,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(380), Column: int(29), @@ -57457,7 +54691,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(381), Column: int(7), @@ -57491,7 +54725,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(381), Column: int(17), @@ -57512,7 +54746,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(381), Column: int(17), @@ -57535,7 +54769,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(381), Column: int(21), @@ -57559,24 +54793,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(381), - Column: int(13), - }, - End: ast.Location{ - Line: int(381), - Column: int(23), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(382), Column: int(7), @@ -57606,7 +54828,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(382), Column: int(10), @@ -57626,7 +54848,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(382), Column: int(10), @@ -57650,7 +54872,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(382), Column: int(15), @@ -57668,14 +54890,13 @@ var _StdAst = &ast.DesugaredObject{ Value: ".", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(383), Column: int(9), @@ -57697,7 +54918,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(383), Column: int(9), @@ -57730,7 +54951,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(383), Column: int(31), @@ -57755,7 +54976,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(383), Column: int(36), @@ -57775,7 +54996,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(383), Column: int(36), @@ -57799,7 +55020,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(383), Column: int(40), @@ -57814,6 +55035,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4360, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -57838,7 +55060,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(385), Column: int(9), @@ -57887,12 +55109,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(385), Column: int(14), @@ -57912,18 +55133,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "i", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(385), - Column: int(11), - }, - End: ast.Location{ - Line: int(385), - Column: int(15), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -57948,12 +55157,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(385), Column: int(20), @@ -57970,18 +55178,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(385), - Column: int(17), - }, - End: ast.Location{ - Line: int(385), - Column: int(24), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -57994,14 +55190,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(380), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(385), - Column: int(26), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -58010,7 +55206,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(380), Column: int(36), @@ -58028,37 +55224,24 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(388), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -58071,7 +55254,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -58098,7 +55281,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(388), Column: int(11), @@ -58116,45 +55299,20 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(388), - Column: int(37), - }, - End: ast.Location{ - Line: int(388), - Column: int(40), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(388), - Column: int(42), - }, - End: ast.Location{ - Line: int(388), - Column: int(43), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -58183,7 +55341,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(389), Column: int(14), @@ -58205,7 +55363,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(389), Column: int(14), @@ -58229,7 +55387,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(389), Column: int(18), @@ -58250,7 +55408,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(389), Column: int(18), @@ -58270,7 +55428,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(389), Column: int(18), @@ -58311,9 +55469,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -58323,7 +55480,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(389), Column: int(29), @@ -58357,7 +55514,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(390), Column: int(7), @@ -58390,7 +55547,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(390), Column: int(17), @@ -58411,7 +55568,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(390), Column: int(17), @@ -58434,7 +55591,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(390), Column: int(21), @@ -58458,24 +55615,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(390), - Column: int(13), - }, - End: ast.Location{ - Line: int(390), - Column: int(23), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(7), @@ -58503,7 +55648,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(10), @@ -58523,7 +55668,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(10), @@ -58543,7 +55688,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(10), @@ -58563,7 +55708,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(10), @@ -58587,7 +55732,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(15), @@ -58605,7 +55750,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "h", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -58613,7 +55757,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(22), @@ -58633,7 +55777,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(22), @@ -58657,7 +55801,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(27), @@ -58675,7 +55819,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "l", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -58684,7 +55827,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(34), @@ -58704,7 +55847,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(34), @@ -58728,7 +55871,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(391), Column: int(39), @@ -58746,7 +55889,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "L", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -58754,7 +55896,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(392), Column: int(9), @@ -58774,7 +55916,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(392), Column: int(9), @@ -58805,7 +55947,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(392), Column: int(13), @@ -58820,6 +55962,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4392, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -58834,7 +55977,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(394), Column: int(9), @@ -58868,14 +56011,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(389), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(394), - Column: int(10), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -58884,7 +56027,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(389), Column: int(36), @@ -58902,37 +56045,24 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(396), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -58965,7 +56095,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(396), Column: int(11), @@ -58983,45 +56113,20 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(396), - Column: int(27), - }, - End: ast.Location{ - Line: int(396), - Column: int(30), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(396), - Column: int(32), - }, - End: ast.Location{ - Line: int(396), - Column: int(33), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -59050,7 +56155,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(397), Column: int(14), @@ -59072,7 +56177,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(397), Column: int(14), @@ -59096,7 +56201,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(397), Column: int(18), @@ -59117,7 +56222,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(397), Column: int(18), @@ -59137,7 +56242,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(397), Column: int(18), @@ -59178,9 +56283,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -59190,7 +56294,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(397), Column: int(29), @@ -59224,7 +56328,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(398), Column: int(7), @@ -59257,7 +56361,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(398), Column: int(17), @@ -59278,7 +56382,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(398), Column: int(17), @@ -59301,7 +56405,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(398), Column: int(21), @@ -59325,24 +56429,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(398), - Column: int(13), - }, - End: ast.Location{ - Line: int(398), - Column: int(23), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(7), @@ -59370,7 +56462,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(10), @@ -59390,7 +56482,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(10), @@ -59410,7 +56502,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(10), @@ -59430,7 +56522,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(10), @@ -59454,7 +56546,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(15), @@ -59472,7 +56564,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "d", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -59480,7 +56571,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(22), @@ -59500,7 +56591,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(22), @@ -59524,7 +56615,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(27), @@ -59542,7 +56633,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -59551,7 +56641,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(34), @@ -59571,7 +56661,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(34), @@ -59595,7 +56685,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(399), Column: int(39), @@ -59613,7 +56703,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "u", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -59621,7 +56710,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(400), Column: int(9), @@ -59670,12 +56759,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(400), Column: int(14), @@ -59695,7 +56783,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(400), Column: int(14), @@ -59719,7 +56807,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(400), Column: int(18), @@ -59734,22 +56822,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4517, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(400), - Column: int(11), - }, - End: ast.Location{ - Line: int(400), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -59774,12 +56851,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(400), Column: int(24), @@ -59797,21 +56873,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "d", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(400), - Column: int(21), - }, - End: ast.Location{ - Line: int(400), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -59836,12 +56899,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(400), Column: int(35), @@ -59859,18 +56921,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(400), - Column: int(29), - }, - End: ast.Location{ - Line: int(400), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -59886,7 +56936,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(401), Column: int(12), @@ -59907,7 +56957,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(401), Column: int(15), @@ -59927,7 +56977,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(401), Column: int(15), @@ -59951,7 +57001,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(401), Column: int(20), @@ -59969,14 +57019,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "o", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(402), Column: int(9), @@ -60025,12 +57074,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(402), Column: int(14), @@ -60050,7 +57098,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(402), Column: int(14), @@ -60074,7 +57122,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(402), Column: int(18), @@ -60089,22 +57137,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4540, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(402), - Column: int(11), - }, - End: ast.Location{ - Line: int(402), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -60129,12 +57166,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(402), Column: int(24), @@ -60152,21 +57188,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "o", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(402), - Column: int(21), - }, - End: ast.Location{ - Line: int(402), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -60191,12 +57214,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(402), Column: int(35), @@ -60214,18 +57236,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(402), - Column: int(29), - }, - End: ast.Location{ - Line: int(402), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -60241,7 +57251,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(403), Column: int(12), @@ -60262,7 +57272,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(403), Column: int(15), @@ -60282,7 +57292,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(403), Column: int(15), @@ -60306,7 +57316,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(403), Column: int(20), @@ -60324,14 +57334,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "x", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(404), Column: int(9), @@ -60380,12 +57389,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(404), Column: int(14), @@ -60405,7 +57413,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(404), Column: int(14), @@ -60429,7 +57437,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(404), Column: int(18), @@ -60444,22 +57452,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4563, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(404), - Column: int(11), - }, - End: ast.Location{ - Line: int(404), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -60484,12 +57481,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(404), Column: int(24), @@ -60507,21 +57503,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "x", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(404), - Column: int(21), - }, - End: ast.Location{ - Line: int(404), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -60546,12 +57529,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(404), Column: int(35), @@ -60569,18 +57551,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(404), - Column: int(29), - }, - End: ast.Location{ - Line: int(404), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -60596,7 +57566,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(405), Column: int(12), @@ -60617,7 +57587,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(405), Column: int(15), @@ -60637,7 +57607,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(405), Column: int(15), @@ -60661,7 +57631,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(405), Column: int(20), @@ -60679,14 +57649,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "X", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(406), Column: int(9), @@ -60735,12 +57704,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(406), Column: int(14), @@ -60760,7 +57728,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(406), Column: int(14), @@ -60784,7 +57752,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(406), Column: int(18), @@ -60799,22 +57767,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4586, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(406), - Column: int(11), - }, - End: ast.Location{ - Line: int(406), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -60839,12 +57796,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(406), Column: int(24), @@ -60862,21 +57818,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "x", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(406), - Column: int(21), - }, - End: ast.Location{ - Line: int(406), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -60901,12 +57844,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(406), Column: int(35), @@ -60924,18 +57866,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(406), - Column: int(29), - }, - End: ast.Location{ - Line: int(406), - Column: int(39), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -60951,7 +57881,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(407), Column: int(12), @@ -60972,7 +57902,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(407), Column: int(15), @@ -60992,7 +57922,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(407), Column: int(15), @@ -61016,7 +57946,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(407), Column: int(20), @@ -61034,14 +57964,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "e", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(408), Column: int(9), @@ -61090,12 +58019,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(408), Column: int(14), @@ -61115,7 +58043,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(408), Column: int(14), @@ -61139,7 +58067,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(408), Column: int(18), @@ -61154,22 +58082,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4609, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(408), - Column: int(11), - }, - End: ast.Location{ - Line: int(408), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -61194,12 +58111,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(408), Column: int(24), @@ -61217,21 +58133,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "e", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(408), - Column: int(21), - }, - End: ast.Location{ - Line: int(408), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -61256,12 +58159,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(408), Column: int(35), @@ -61279,18 +58181,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(408), - Column: int(29), - }, - End: ast.Location{ - Line: int(408), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -61306,7 +58196,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(409), Column: int(12), @@ -61327,7 +58217,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(409), Column: int(15), @@ -61347,7 +58237,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(409), Column: int(15), @@ -61371,7 +58261,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(409), Column: int(20), @@ -61389,14 +58279,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "E", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(410), Column: int(9), @@ -61445,12 +58334,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(410), Column: int(14), @@ -61470,7 +58358,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(410), Column: int(14), @@ -61494,7 +58382,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(410), Column: int(18), @@ -61509,22 +58397,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4632, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(410), - Column: int(11), - }, - End: ast.Location{ - Line: int(410), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -61549,12 +58426,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(410), Column: int(24), @@ -61572,21 +58448,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "e", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(410), - Column: int(21), - }, - End: ast.Location{ - Line: int(410), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -61611,12 +58474,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(410), Column: int(35), @@ -61634,18 +58496,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(410), - Column: int(29), - }, - End: ast.Location{ - Line: int(410), - Column: int(39), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -61661,7 +58511,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(411), Column: int(12), @@ -61682,7 +58532,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(411), Column: int(15), @@ -61702,7 +58552,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(411), Column: int(15), @@ -61726,7 +58576,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(411), Column: int(20), @@ -61744,14 +58594,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "f", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(412), Column: int(9), @@ -61800,12 +58649,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(412), Column: int(14), @@ -61825,7 +58673,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(412), Column: int(14), @@ -61849,7 +58697,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(412), Column: int(18), @@ -61864,22 +58712,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4655, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(412), - Column: int(11), - }, - End: ast.Location{ - Line: int(412), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -61904,12 +58741,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(412), Column: int(24), @@ -61927,21 +58763,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "f", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(412), - Column: int(21), - }, - End: ast.Location{ - Line: int(412), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -61966,12 +58789,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(412), Column: int(35), @@ -61989,18 +58811,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(412), - Column: int(29), - }, - End: ast.Location{ - Line: int(412), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -62016,7 +58826,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(413), Column: int(12), @@ -62037,7 +58847,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(413), Column: int(15), @@ -62057,7 +58867,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(413), Column: int(15), @@ -62081,7 +58891,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(413), Column: int(20), @@ -62099,14 +58909,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "F", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(414), Column: int(9), @@ -62155,12 +58964,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(414), Column: int(14), @@ -62180,7 +58988,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(414), Column: int(14), @@ -62204,7 +59012,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(414), Column: int(18), @@ -62219,22 +59027,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4678, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(414), - Column: int(11), - }, - End: ast.Location{ - Line: int(414), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -62259,12 +59056,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(414), Column: int(24), @@ -62282,21 +59078,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "f", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(414), - Column: int(21), - }, - End: ast.Location{ - Line: int(414), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -62321,12 +59104,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(414), Column: int(35), @@ -62344,18 +59126,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(414), - Column: int(29), - }, - End: ast.Location{ - Line: int(414), - Column: int(39), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -62371,7 +59141,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(415), Column: int(12), @@ -62392,7 +59162,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(415), Column: int(15), @@ -62412,7 +59182,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(415), Column: int(15), @@ -62436,7 +59206,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(415), Column: int(20), @@ -62454,14 +59224,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "g", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(416), Column: int(9), @@ -62510,12 +59279,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(416), Column: int(14), @@ -62535,7 +59303,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(416), Column: int(14), @@ -62559,7 +59327,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(416), Column: int(18), @@ -62574,22 +59342,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4701, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(416), - Column: int(11), - }, - End: ast.Location{ - Line: int(416), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -62614,12 +59371,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(416), Column: int(24), @@ -62637,21 +59393,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "g", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(416), - Column: int(21), - }, - End: ast.Location{ - Line: int(416), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -62676,12 +59419,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(416), Column: int(35), @@ -62699,18 +59441,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(416), - Column: int(29), - }, - End: ast.Location{ - Line: int(416), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -62726,7 +59456,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(417), Column: int(12), @@ -62747,7 +59477,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(417), Column: int(15), @@ -62767,7 +59497,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(417), Column: int(15), @@ -62791,7 +59521,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(417), Column: int(20), @@ -62809,14 +59539,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "G", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(418), Column: int(9), @@ -62865,12 +59594,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(418), Column: int(14), @@ -62890,7 +59618,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(418), Column: int(14), @@ -62914,7 +59642,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(418), Column: int(18), @@ -62929,22 +59657,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4724, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(418), - Column: int(11), - }, - End: ast.Location{ - Line: int(418), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -62969,12 +59686,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(418), Column: int(24), @@ -62992,21 +59708,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "g", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(418), - Column: int(21), - }, - End: ast.Location{ - Line: int(418), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -63031,12 +59734,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(418), Column: int(35), @@ -63054,18 +59756,6 @@ var _StdAst = &ast.DesugaredObject{ Value: true, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(418), - Column: int(29), - }, - End: ast.Location{ - Line: int(418), - Column: int(39), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -63081,7 +59771,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(419), Column: int(12), @@ -63102,7 +59792,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(419), Column: int(15), @@ -63122,7 +59812,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(419), Column: int(15), @@ -63146,7 +59836,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(419), Column: int(20), @@ -63164,14 +59854,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "c", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(420), Column: int(9), @@ -63220,12 +59909,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(420), Column: int(14), @@ -63245,7 +59933,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(420), Column: int(14), @@ -63269,7 +59957,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(420), Column: int(18), @@ -63284,22 +59972,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4747, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(420), - Column: int(11), - }, - End: ast.Location{ - Line: int(420), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -63324,12 +60001,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(420), Column: int(24), @@ -63347,21 +60023,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "c", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(420), - Column: int(21), - }, - End: ast.Location{ - Line: int(420), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -63386,12 +60049,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(420), Column: int(35), @@ -63409,18 +60071,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(420), - Column: int(29), - }, - End: ast.Location{ - Line: int(420), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -63436,7 +60086,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(421), Column: int(12), @@ -63457,7 +60107,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(421), Column: int(15), @@ -63477,7 +60127,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(421), Column: int(15), @@ -63501,7 +60151,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(421), Column: int(20), @@ -63519,14 +60169,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(422), Column: int(9), @@ -63575,12 +60224,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(422), Column: int(14), @@ -63600,7 +60248,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(422), Column: int(14), @@ -63624,7 +60272,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(422), Column: int(18), @@ -63639,22 +60287,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4770, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(422), - Column: int(11), - }, - End: ast.Location{ - Line: int(422), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -63679,12 +60316,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(422), Column: int(24), @@ -63702,21 +60338,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(422), - Column: int(21), - }, - End: ast.Location{ - Line: int(422), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -63741,12 +60364,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(422), Column: int(35), @@ -63764,18 +60386,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(422), - Column: int(29), - }, - End: ast.Location{ - Line: int(422), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -63791,7 +60401,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(423), Column: int(12), @@ -63812,7 +60422,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(423), Column: int(15), @@ -63832,7 +60442,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(423), Column: int(15), @@ -63856,7 +60466,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(423), Column: int(20), @@ -63874,14 +60484,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(424), Column: int(9), @@ -63930,12 +60539,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(424), Column: int(14), @@ -63955,7 +60563,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(424), Column: int(14), @@ -63979,7 +60587,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(424), Column: int(18), @@ -63994,22 +60602,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p4793, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(424), - Column: int(11), - }, - End: ast.Location{ - Line: int(424), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -64034,12 +60631,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(424), Column: int(24), @@ -64057,21 +60653,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(424), - Column: int(21), - }, - End: ast.Location{ - Line: int(424), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -64096,12 +60679,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(424), Column: int(35), @@ -64119,18 +60701,6 @@ var _StdAst = &ast.DesugaredObject{ Value: false, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(424), - Column: int(29), - }, - End: ast.Location{ - Line: int(424), - Column: int(40), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -64146,7 +60716,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(426), Column: int(9), @@ -64173,7 +60743,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(426), Column: int(15), @@ -64193,7 +60763,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(426), Column: int(15), @@ -64211,14 +60781,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "Unrecognised conversion type: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(426), Column: int(50), @@ -64259,14 +60828,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(397), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(426), - Column: int(51), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -64275,7 +60844,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(397), Column: int(36), @@ -64293,37 +60862,24 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(430), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -64336,7 +60892,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -64365,7 +60921,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(430), Column: int(11), @@ -64389,45 +60945,20 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(430), - Column: int(22), - }, - End: ast.Location{ - Line: int(430), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(430), - Column: int(27), - }, - End: ast.Location{ - Line: int(430), - Column: int(28), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -64462,7 +60993,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(431), Column: int(14), @@ -64484,7 +61015,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(431), Column: int(14), @@ -64508,7 +61039,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(431), Column: int(18), @@ -64529,7 +61060,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(431), Column: int(18), @@ -64549,7 +61080,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(431), Column: int(18), @@ -64590,9 +61121,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -64602,7 +61132,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(431), Column: int(29), @@ -64636,7 +61166,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(432), Column: int(7), @@ -64675,7 +61205,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(432), Column: int(20), @@ -64697,7 +61227,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(432), Column: int(20), @@ -64723,7 +61253,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(432), Column: int(42), @@ -64748,7 +61278,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(432), Column: int(47), @@ -64779,24 +61309,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(432), - Column: int(13), - }, - End: ast.Location{ - Line: int(432), - Column: int(49), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(433), Column: int(7), @@ -64834,7 +61352,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(433), Column: int(22), @@ -64856,7 +61374,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(433), Column: int(22), @@ -64882,7 +61400,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(433), Column: int(39), @@ -64907,7 +61425,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(433), Column: int(44), @@ -64927,7 +61445,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(433), Column: int(44), @@ -64968,9 +61486,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -64985,24 +61502,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(433), - Column: int(13), - }, - End: ast.Location{ - Line: int(433), - Column: int(51), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(434), Column: int(7), @@ -65040,7 +61545,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(434), Column: int(18), @@ -65062,7 +61567,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(434), Column: int(18), @@ -65088,7 +61593,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(434), Column: int(40), @@ -65113,7 +61618,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(434), Column: int(45), @@ -65133,7 +61638,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(434), Column: int(45), @@ -65174,9 +61679,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -65191,24 +61695,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(434), - Column: int(13), - }, - End: ast.Location{ - Line: int(434), - Column: int(54), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(435), Column: int(7), @@ -65246,7 +61738,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(435), Column: int(20), @@ -65268,7 +61760,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(435), Column: int(20), @@ -65294,7 +61786,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(435), Column: int(40), @@ -65319,7 +61811,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(435), Column: int(45), @@ -65339,7 +61831,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(435), Column: int(45), @@ -65380,9 +61872,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -65397,24 +61888,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(435), - Column: int(13), - }, - End: ast.Location{ - Line: int(435), - Column: int(50), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(436), Column: int(7), @@ -65452,7 +61931,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(436), Column: int(23), @@ -65474,7 +61953,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(436), Column: int(23), @@ -65500,7 +61979,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(436), Column: int(49), @@ -65525,7 +62004,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(436), Column: int(54), @@ -65545,7 +62024,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(436), Column: int(54), @@ -65586,9 +62065,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -65603,24 +62081,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(436), - Column: int(13), - }, - End: ast.Location{ - Line: int(436), - Column: int(61), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(437), Column: int(7), @@ -65658,7 +62124,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(437), Column: int(21), @@ -65680,7 +62146,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(437), Column: int(21), @@ -65706,7 +62172,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(437), Column: int(37), @@ -65731,7 +62197,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(437), Column: int(42), @@ -65762,24 +62228,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(437), - Column: int(13), - }, - End: ast.Location{ - Line: int(437), - Column: int(50), - }, - File: p1, - }, }, }, Body: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(438), Column: int(7), @@ -65832,12 +62286,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(439), Column: int(12), @@ -65857,7 +62310,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(439), Column: int(12), @@ -65898,24 +62351,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(439), - Column: int(9), - }, - End: ast.Location{ - Line: int(439), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -65940,12 +62380,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "code", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(440), Column: int(15), @@ -65991,12 +62430,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "mkey", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(441), Column: int(17), @@ -66016,7 +62454,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(441), Column: int(17), @@ -66057,24 +62495,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(441), - Column: int(11), - }, - End: ast.Location{ - Line: int(441), - Column: int(23), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -66099,12 +62524,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "cflags", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(442), Column: int(19), @@ -66124,7 +62548,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(442), Column: int(19), @@ -66165,24 +62589,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(442), - Column: int(11), - }, - End: ast.Location{ - Line: int(442), - Column: int(27), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -66207,12 +62618,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(443), Column: int(15), @@ -66232,7 +62642,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(443), Column: int(15), @@ -66273,24 +62683,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(443), - Column: int(11), - }, - End: ast.Location{ - Line: int(443), - Column: int(19), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -66315,12 +62712,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(444), Column: int(17), @@ -66340,7 +62736,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(444), Column: int(17), @@ -66381,24 +62777,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(444), - Column: int(11), - }, - End: ast.Location{ - Line: int(444), - Column: int(23), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -66423,12 +62806,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(445), Column: int(18), @@ -66448,7 +62830,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(445), Column: int(18), @@ -66489,24 +62871,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "v", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(445), - Column: int(11), - }, - End: ast.Location{ - Line: int(445), - Column: int(25), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -66531,12 +62900,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(446), Column: int(17), @@ -66556,7 +62924,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(446), Column: int(17), @@ -66597,41 +62965,16 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(446), - Column: int(11), - }, - End: ast.Location{ - Line: int(446), - Column: int(27), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(440), - Column: int(9), - }, - End: ast.Location{ - Line: int(447), - Column: int(10), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -66648,14 +62991,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(431), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(448), - Column: int(8), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -66664,7 +63007,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(431), Column: int(36), @@ -66682,37 +63025,24 @@ var _StdAst = &ast.DesugaredObject{ Value: "Truncated format code.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(451), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -66725,7 +63055,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -66749,7 +63079,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(451), Column: int(11), @@ -66769,90 +63099,37 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(451), - Column: int(23), - }, - End: ast.Location{ - Line: int(451), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(451), - Column: int(28), - }, - End: ast.Location{ - Line: int(451), - Column: int(29), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "out", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(451), - Column: int(31), - }, - End: ast.Location{ - Line: int(451), - Column: int(34), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "out", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "cur", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(451), - Column: int(36), - }, - End: ast.Location{ - Line: int(451), - Column: int(39), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "cur", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(452), Column: int(7), @@ -66885,7 +63162,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(452), Column: int(10), @@ -66907,7 +63184,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(452), Column: int(10), @@ -66931,7 +63208,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(452), Column: int(15), @@ -66952,7 +63229,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(452), Column: int(15), @@ -66972,7 +63249,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(452), Column: int(15), @@ -67013,9 +63290,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -67025,7 +63301,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(452), Column: int(26), @@ -67059,7 +63335,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(453), Column: int(9), @@ -67080,7 +63356,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(453), Column: int(9), @@ -67111,7 +63387,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(453), Column: int(15), @@ -67133,7 +63409,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(453), Column: int(16), @@ -67170,7 +63446,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(455), Column: int(9), @@ -67207,7 +63483,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(455), Column: int(19), @@ -67228,7 +63504,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(455), Column: int(19), @@ -67251,7 +63527,7 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(455), Column: int(23), @@ -67275,24 +63551,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(455), - Column: int(15), - }, - End: ast.Location{ - Line: int(455), - Column: int(25), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(456), Column: int(9), @@ -67325,7 +63589,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(456), Column: int(12), @@ -67345,7 +63609,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(456), Column: int(12), @@ -67369,7 +63633,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(456), Column: int(17), @@ -67387,14 +63651,13 @@ var _StdAst = &ast.DesugaredObject{ Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(457), Column: int(11), @@ -67431,7 +63694,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(457), Column: int(21), @@ -67453,7 +63716,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(457), Column: int(21), @@ -67479,7 +63742,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(457), Column: int(32), @@ -67504,7 +63767,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(457), Column: int(37), @@ -67524,7 +63787,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(457), Column: int(37), @@ -67548,7 +63811,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(457), Column: int(41), @@ -67563,6 +63826,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p5066, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -67578,24 +63842,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(457), - Column: int(17), - }, - End: ast.Location{ - Line: int(457), - Column: int(43), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(11), @@ -67619,7 +63871,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(11), @@ -67652,7 +63904,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(23), @@ -67677,7 +63929,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(28), @@ -67697,7 +63949,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(28), @@ -67738,9 +63990,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "i", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -67749,7 +64000,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(33), @@ -67771,7 +64022,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(33), @@ -67795,7 +64046,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(39), @@ -67818,7 +64069,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(40), @@ -67843,7 +64094,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(45), @@ -67863,7 +64114,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(45), @@ -67904,9 +64155,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "code", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -67922,7 +64172,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(458), Column: int(54), @@ -67940,7 +64190,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -67964,7 +64213,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(11), @@ -67989,7 +64238,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(11), @@ -68022,7 +64271,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(23), @@ -68047,7 +64296,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(28), @@ -68067,7 +64316,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(28), @@ -68091,7 +64340,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(32), @@ -68106,6 +64355,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p5111, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -68115,7 +64365,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(35), @@ -68140,7 +64390,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(40), @@ -68161,7 +64411,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(40), @@ -68185,7 +64435,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(460), Column: int(46), @@ -68221,30 +64471,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(462), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -68273,7 +64511,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(462), Column: int(19), @@ -68294,7 +64532,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(462), Column: int(19), @@ -68320,7 +64558,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(462), Column: int(31), @@ -68345,7 +64583,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(462), Column: int(36), @@ -68360,6 +64598,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p5137, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -68368,7 +64607,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(462), Column: int(39), @@ -68393,7 +64632,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(462), Column: int(43), @@ -68411,7 +64650,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -68425,30 +64663,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(462), - Column: int(11), - }, - End: ast.Location{ - Line: int(462), - Column: int(46), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(470), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -68461,7 +64687,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -68469,7 +64695,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -68477,7 +64703,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(1), Indent: int(4), Comment: []string{ @@ -68485,7 +64711,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -68508,7 +64734,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(470), Column: int(11), @@ -68524,52 +64750,27 @@ var _StdAst = &ast.DesugaredObject{ FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "w", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(470), - Column: int(19), - }, - End: ast.Location{ - Line: int(470), - Column: int(20), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "w", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "s", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(470), - Column: int(22), - }, - End: ast.Location{ - Line: int(470), - Column: int(23), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "s", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(471), Column: int(7), @@ -68602,7 +64803,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(471), Column: int(13), @@ -68621,52 +64822,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "w", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(471), - Column: int(17), - }, - End: ast.Location{ - Line: int(471), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "w", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(471), - Column: int(20), - }, - End: ast.Location{ - Line: int(471), - Column: int(21), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(472), Column: int(9), @@ -68696,7 +64872,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(472), Column: int(12), @@ -68716,7 +64892,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(472), Column: int(12), @@ -68740,7 +64916,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(472), Column: int(17), @@ -68755,6 +64931,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p5164, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -68762,7 +64939,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(473), Column: int(11), @@ -68799,7 +64976,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(11), @@ -68822,7 +64999,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(11), @@ -68855,7 +65032,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(15), @@ -68875,7 +65052,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(15), @@ -68899,7 +65076,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(19), @@ -68914,6 +65091,7 @@ var _StdAst = &ast.DesugaredObject{ Ctx: p5182, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -68923,7 +65101,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(22), @@ -68944,7 +65122,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(22), @@ -68968,7 +65146,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(475), Column: int(26), @@ -69002,24 +65180,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(476), Column: int(7), @@ -69040,7 +65206,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(476), Column: int(7), @@ -69073,7 +65239,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(476), Column: int(11), @@ -69098,7 +65264,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(476), Column: int(14), @@ -69116,7 +65282,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -69132,30 +65297,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(479), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -69168,7 +65321,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -69192,7 +65345,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(479), Column: int(11), @@ -69211,71 +65364,32 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(479), - Column: int(20), - }, - End: ast.Location{ - Line: int(479), - Column: int(23), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "w", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(479), - Column: int(25), - }, - End: ast.Location{ - Line: int(479), - Column: int(26), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "w", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "s", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(479), - Column: int(28), - }, - End: ast.Location{ - Line: int(479), - Column: int(29), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "s", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(7), @@ -69299,7 +65413,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(7), @@ -69323,7 +65437,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(7), @@ -69356,7 +65470,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(15), @@ -69378,7 +65492,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(15), @@ -69402,7 +65516,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(19), @@ -69423,7 +65537,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(19), @@ -69443,7 +65557,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(19), @@ -69484,9 +65598,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -69496,7 +65609,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(30), @@ -69532,7 +65645,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(36), @@ -69566,7 +65679,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(480), Column: int(41), @@ -69589,30 +65702,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(483), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -69625,7 +65726,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -69650,7 +65751,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(483), Column: int(11), @@ -69669,71 +65770,32 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(483), - Column: int(21), - }, - End: ast.Location{ - Line: int(483), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "w", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(483), - Column: int(26), - }, - End: ast.Location{ - Line: int(483), - Column: int(27), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "w", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "s", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(483), - Column: int(29), - }, - End: ast.Location{ - Line: int(483), - Column: int(30), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "s", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(7), @@ -69757,7 +65819,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(7), @@ -69788,7 +65850,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(13), @@ -69812,7 +65874,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(13), @@ -69838,7 +65900,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(21), @@ -69860,7 +65922,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(21), @@ -69884,7 +65946,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(25), @@ -69905,7 +65967,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(25), @@ -69925,7 +65987,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(25), @@ -69966,9 +66028,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -69978,7 +66039,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(36), @@ -70014,7 +66075,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(484), Column: int(42), @@ -70047,30 +66108,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(499), + Line: int(487), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -70083,107 +66132,11 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// Render a sign & magnitude integer (radix ranges from decimal to binary).", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// neg should be a boolean, and when true indicates that we should render a negative number.", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// mag must always be a whole number >= 0, it's the magnitude of the integer to render", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// min_chars must be a whole number >= 0", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// It is the field width, i.e. std.length() of the result should be >= min_chars", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// min_digits must be a whole number >= 0. It's the number of zeroes to pad with.", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// blank must be a boolean, if true adds an additional ' ' in front of a positive number, so", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// that it is aligned with negative numbers with the same number of digits.", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// plus must be a boolean, if true adds a '+' in front of a postive number, so that it is", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// aligned with negative numbers with the same number of digits. This takes precedence over", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// blank, if both are true.", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(4), - Comment: []string{ - "// radix must be a whole number >1 and <= 10. It is the base of the system of numerals.", - }, - }, - ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ - "// zero_prefix is a string prefixed before the sign to all numbers that are not 0.", + "// Render an integer (e.g., decimal or octal).", }, }, }, @@ -70204,191 +66157,77 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(499), + Line: int(487), Column: int(11), }, End: ast.Location{ - Line: int(514), + Line: int(499), Column: int(84), }, File: p1, }, Fodder: nil, - Ctx: p5297, + Ctx: p5285, FreeVars: ast.Identifiers{ "pad_left", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "neg", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(22), - }, - End: ast.Location{ - Line: int(499), - Column: int(25), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "mag", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(27), - }, - End: ast.Location{ - Line: int(499), - Column: int(30), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n__", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "min_chars", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(32), - }, - End: ast.Location{ - Line: int(499), - Column: int(41), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "min_chars", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "min_digits", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(43), - }, - End: ast.Location{ - Line: int(499), - Column: int(53), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "min_digits", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "blank", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(55), - }, - End: ast.Location{ - Line: int(499), - Column: int(60), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "blank", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "plus", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(62), - }, - End: ast.Location{ - Line: int(499), - Column: int(66), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "sign", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "radix", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(68), - }, - End: ast.Location{ - Line: int(499), - Column: int(73), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "radix", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "zero_prefix", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(499), - Column: int(75), - }, - End: ast.Location{ - Line: int(499), - Column: int(86), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "zero_prefix", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(501), + Line: int(488), Column: int(7), }, End: ast.Location{ - Line: int(514), + Line: int(499), Column: int(84), }, File: p1, @@ -70400,25 +66239,16 @@ var _StdAst = &ast.DesugaredObject{ Indent: int(6), Comment: []string{}, }, - ast.FodderElement{ - Kind: ast.FodderKind(2), - Blanks: int(0), - Indent: int(6), - Comment: []string{ - "// dec is the minimal string needed to represent the number as text.", - }, - }, }, - Ctx: p5303, + Ctx: p5290, FreeVars: ast.Identifiers{ "blank", - "mag", "min_chars", "min_digits", - "neg", + "n__", "pad_left", - "plus", "radix", + "sign", "std", "zero_prefix", }, @@ -70426,235 +66256,406 @@ var _StdAst = &ast.DesugaredObject{ Binds: ast.LocalBinds{ ast.LocalBind{ VarFodder: ast.Fodder{}, - Variable: "dec", + Variable: "n_", EqFodder: ast.Fodder{}, - Body: &ast.Conditional{ + Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(502), - Column: int(9), + Line: int(488), + Column: int(18), }, End: ast.Location{ - Line: int(510), - Column: int(19), + Line: int(488), + Column: int(30), }, File: p1, }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - Ctx: p5308, + Fodder: ast.Fodder{}, + Ctx: p5294, FreeVars: ast.Identifiers{ - "mag", - "radix", + "n__", "std", - "zero_prefix", }, }, - Cond: &ast.Binary{ + Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(502), - Column: int(12), + Line: int(488), + Column: int(18), }, End: ast.Location{ - Line: int(502), - Column: int(20), + Line: int(488), + Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5308, + Ctx: p5294, FreeVars: ast.Identifiers{ - "mag", + "std", }, }, - Left: &ast.Var{ + Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(502), - Column: int(12), + Line: int(488), + Column: int(18), }, End: ast.Location{ - Line: int(502), - Column: int(15), + Line: int(488), + Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5308, + Ctx: nil, FreeVars: ast.Identifiers{ - "mag", + "std", }, }, - Id: "mag", + Id: "std", }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.LiteralNumber{ + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(502), - Column: int(19), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(502), - Column: int(20), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, - Fodder: ast.Fodder{}, - Ctx: p5308, + Fodder: nil, + Ctx: nil, FreeVars: nil, }, - OriginalString: "0", + Value: "abs", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", }, + RightBracketFodder: nil, + Id: nil, }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(503), - Column: int(11), - }, - End: ast.Location{ - Line: int(503), - Column: int(14), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(10), - Comment: []string{}, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(488), + Column: int(26), + }, + End: ast.Location{ + Line: int(488), + Column: int(29), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5303, + FreeVars: ast.Identifiers{ + "n__", + }, + }, + Id: "n__", }, + CommaFodder: nil, }, - Ctx: p5308, - FreeVars: nil, }, - Value: "0", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", + Named: nil, }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, + }, + Fun: nil, + CloseFodder: ast.Fodder{}, + }, + }, + Body: &ast.Local{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(489), + Column: int(7), + }, + End: ast.Location{ + Line: int(499), + Column: int(84), + }, + File: p1, + }, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(6), + Comment: []string{}, }, - BranchFalse: &ast.Local{ + }, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "blank", + "min_chars", + "min_digits", + "n_", + "n__", + "pad_left", + "radix", + "sign", + "std", + "zero_prefix", + }, + }, + Binds: ast.LocalBinds{ + ast.LocalBind{ + VarFodder: nil, + Variable: "aux", + EqFodder: nil, + Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(505), - Column: int(11), + Line: int(489), + Column: int(13), }, End: ast.Location{ - Line: int(510), - Column: int(19), + Line: int(493), + Column: int(50), }, File: p1, }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(10), - Comment: []string{}, - }, - }, - Ctx: p5308, + Fodder: nil, + Ctx: p5310, FreeVars: ast.Identifiers{ - "mag", + "aux", "radix", "std", "zero_prefix", }, }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: nil, - Variable: "aux", - EqFodder: nil, - Body: &ast.Function{ + ParenLeftFodder: ast.Fodder{}, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n", + CommaFodder: nil, + }, + }, + Optional: nil, + }, + TrailingComma: false, + ParenRightFodder: ast.Fodder{}, + Body: &ast.Conditional{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(490), + Column: int(9), + }, + End: ast.Location{ + Line: int(493), + Column: int(50), + }, + File: p1, + }, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(8), + Comment: []string{}, + }, + }, + Ctx: p5315, + FreeVars: ast.Identifiers{ + "aux", + "n", + "radix", + "std", + "zero_prefix", + }, + }, + Cond: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(490), + Column: int(12), + }, + End: ast.Location{ + Line: int(490), + Column: int(18), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5315, + FreeVars: ast.Identifiers{ + "n", + }, + }, + Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", + Begin: ast.Location{ + Line: int(490), + Column: int(12), + }, + End: ast.Location{ + Line: int(490), + Column: int(13), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5315, + FreeVars: ast.Identifiers{ + "n", + }, + }, + Id: "n", + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(12), + Right: &ast.LiteralNumber{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", Begin: ast.Location{ - Line: int(505), + Line: int(490), Column: int(17), }, End: ast.Location{ - Line: int(509), - Column: int(54), + Line: int(490), + Column: int(18), }, File: p1, }, - Fodder: nil, - Ctx: p5323, + Fodder: ast.Fodder{}, + Ctx: p5315, + FreeVars: nil, + }, + Value: float64(0), + OriginalString: "0", + }, + }, + ThenFodder: ast.Fodder{}, + BranchTrue: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(491), + Column: int(11), + }, + End: ast.Location{ + Line: int(491), + Column: int(22), + }, + File: p1, + }, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(10), + Comment: []string{}, + }, + }, + Ctx: p5315, + FreeVars: ast.Identifiers{ + "zero_prefix", + }, + }, + Id: "zero_prefix", + }, + ElseFodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(8), + Comment: []string{}, + }, + }, + BranchFalse: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(493), + Column: int(11), + }, + End: ast.Location{ + Line: int(493), + Column: int(50), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5315, + FreeVars: ast.Identifiers{ + "aux", + "n", + "radix", + "std", + }, + }, + Left: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(493), + Column: int(11), + }, + End: ast.Location{ + Line: int(493), + Column: int(36), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5315, FreeVars: ast.Identifiers{ "aux", + "n", "radix", "std", - "zero_prefix", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "n", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(505), - Column: int(21), - }, - End: ast.Location{ - Line: int(505), - Column: int(22), - }, - File: p1, - }, }, }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Conditional{ + Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(506), - Column: int(13), + Line: int(493), + Column: int(11), }, End: ast.Location{ - Line: int(509), - Column: int(54), + Line: int(493), + Column: int(14), }, File: p1, }, @@ -70662,1161 +66663,817 @@ var _StdAst = &ast.DesugaredObject{ ast.FodderElement{ Kind: ast.FodderKind(0), Blanks: int(0), - Indent: int(12), + Indent: int(10), Comment: []string{}, }, }, - Ctx: p5328, + Ctx: p5315, FreeVars: ast.Identifiers{ "aux", - "n", - "radix", - "std", - "zero_prefix", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(506), - Column: int(16), - }, - End: ast.Location{ - Line: int(506), - Column: int(22), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5328, - FreeVars: ast.Identifiers{ - "n", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(506), - Column: int(16), - }, - End: ast.Location{ - Line: int(506), - Column: int(17), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5328, - FreeVars: ast.Identifiers{ - "n", - }, - }, - Id: "n", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(506), - Column: int(21), - }, - End: ast.Location{ - Line: int(506), - Column: int(22), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5328, - FreeVars: nil, - }, - OriginalString: "0", - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(507), - Column: int(15), - }, - End: ast.Location{ - Line: int(507), - Column: int(26), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(14), - Comment: []string{}, - }, - }, - Ctx: p5328, - FreeVars: ast.Identifiers{ - "zero_prefix", - }, - }, - Id: "zero_prefix", - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(12), - Comment: []string{}, }, }, - BranchFalse: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(15), - }, - End: ast.Location{ - Line: int(509), - Column: int(54), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5328, - FreeVars: ast.Identifiers{ - "aux", - "n", - "radix", - "std", - }, - }, - Left: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(15), - }, - End: ast.Location{ - Line: int(509), - Column: int(40), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5328, - FreeVars: ast.Identifiers{ - "aux", - "n", - "radix", - "std", - }, - }, - Target: &ast.Var{ + Id: "aux", + }, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(509), + Line: int(493), Column: int(15), }, End: ast.Location{ - Line: int(509), - Column: int(18), + Line: int(493), + Column: int(35), }, File: p1, }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(14), - Comment: []string{}, - }, - }, - Ctx: p5328, + Fodder: ast.Fodder{}, + Ctx: p5335, FreeVars: ast.Identifiers{ - "aux", + "n", + "radix", + "std", }, }, - Id: "aux", - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(19), - }, - End: ast.Location{ - Line: int(509), - Column: int(39), - }, - File: p1, + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(493), + Column: int(15), + }, + End: ast.Location{ + Line: int(493), + Column: int(24), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5335, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(493), + Column: int(15), }, - Fodder: ast.Fodder{}, - Ctx: p5348, - FreeVars: ast.Identifiers{ - "n", - "radix", - "std", + End: ast.Location{ + Line: int(493), + Column: int(18), }, + File: p1, }, - Target: &ast.Index{ + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "floor", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, + }, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(509), - Column: int(19), + Line: int(493), + Column: int(25), }, End: ast.Location{ - Line: int(509), - Column: int(28), + Line: int(493), + Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5348, + Ctx: p5344, FreeVars: ast.Identifiers{ - "std", + "n", + "radix", }, }, - Target: &ast.Var{ + Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(509), - Column: int(19), + Line: int(493), + Column: int(25), }, End: ast.Location{ - Line: int(509), - Column: int(22), + Line: int(493), + Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: nil, + Ctx: p5344, FreeVars: ast.Identifiers{ - "std", + "n", }, }, - Id: "std", + Id: "n", }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(1), + Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(493), + Column: int(29), }, End: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(493), + Column: int(34), }, - File: nil, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5344, + FreeVars: ast.Identifiers{ + "radix", }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, }, - Value: "floor", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", + Id: "radix", }, - RightBracketFodder: ast.Fodder{}, - Id: nil, }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(29), - }, - End: ast.Location{ - Line: int(509), - Column: int(38), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5357, - FreeVars: ast.Identifiers{ - "n", - "radix", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(29), - }, - End: ast.Location{ - Line: int(509), - Column: int(30), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5357, - FreeVars: ast.Identifiers{ - "n", - }, - }, - Id: "n", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(1), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(33), - }, - End: ast.Location{ - Line: int(509), - Column: int(38), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5357, - FreeVars: ast.Identifiers{ - "radix", - }, - }, - Id: "radix", - }, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, + CommaFodder: nil, }, - CommaFodder: nil, }, + Named: nil, }, - Named: nil, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, + CommaFodder: nil, }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(44), - }, - End: ast.Location{ - Line: int(509), - Column: int(53), - }, - File: p1, + }, + Named: nil, + }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(3), + Right: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: ast.Identifiers{ + "n", + "radix", + "std", + }, + }, + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "n", - "radix", - "std", + End: ast.Location{ + Line: int(0), + Column: int(0), }, + File: nil, }, - Target: &ast.Index{ + Fodder: nil, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", + }, + LeftBracketFodder: nil, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "mod", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, + }, + FodderLeft: nil, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(493), + Column: int(40), }, End: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(493), + Column: int(41), }, - File: nil, + File: p1, }, - Fodder: nil, - Ctx: nil, + Fodder: ast.Fodder{}, + Ctx: p5315, FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: nil, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, + "n", }, - Value: "mod", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: nil, - Id: nil, + Id: "n", }, - FodderLeft: nil, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(44), - }, - End: ast.Location{ - Line: int(509), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5328, - FreeVars: ast.Identifiers{ - "n", - }, - }, - Id: "n", + CommaFodder: nil, + }, + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(493), + Column: int(44), }, - CommaFodder: nil, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(509), - Column: int(48), - }, - End: ast.Location{ - Line: int(509), - Column: int(53), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5328, - FreeVars: ast.Identifiers{ - "radix", - }, - }, - Id: "radix", + End: ast.Location{ + Line: int(493), + Column: int(49), }, - CommaFodder: nil, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5315, + FreeVars: ast.Identifiers{ + "radix", }, }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: nil, - TailStrictFodder: nil, - }, - }, - }, - }, - Fun: nil, - CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - }, - }, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(510), - Column: int(11), - }, - End: ast.Location{ - Line: int(510), - Column: int(19), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5308, - FreeVars: ast.Identifiers{ - "aux", - "mag", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(510), - Column: int(11), - }, - End: ast.Location{ - Line: int(510), - Column: int(14), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(10), - Comment: []string{}, - }, - }, - Ctx: p5308, - FreeVars: ast.Identifiers{ - "aux", - }, - }, - Id: "aux", - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(510), - Column: int(15), - }, - End: ast.Location{ - Line: int(510), - Column: int(18), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5382, - FreeVars: ast.Identifiers{ - "mag", + Id: "radix", }, + CommaFodder: nil, }, - Id: "mag", }, - CommaFodder: nil, + Named: nil, }, + TrailingComma: false, + TailStrict: false, + FodderRight: nil, + TailStrictFodder: nil, }, - Named: nil, }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, }, }, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(501), - Column: int(13), - }, - End: ast.Location{ - Line: int(510), - Column: int(19), - }, - File: p1, + Fun: nil, + CloseFodder: nil, }, }, - }, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(511), - Column: int(7), + Body: &ast.Local{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(494), + Column: int(7), + }, + End: ast.Location{ + Line: int(499), + Column: int(84), + }, + File: p1, }, - End: ast.Location{ - Line: int(514), - Column: int(84), + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(6), + Comment: []string{}, + }, }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "aux", + "blank", + "min_chars", + "min_digits", + "n_", + "n__", + "pad_left", + "sign", + "std", }, }, - Ctx: p5303, - FreeVars: ast.Identifiers{ - "blank", - "dec", - "min_chars", - "min_digits", - "neg", - "pad_left", - "plus", - "std", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "zp", - EqFodder: ast.Fodder{}, - Body: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(511), - Column: int(18), - }, - End: ast.Location{ - Line: int(511), - Column: int(69), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5389, - FreeVars: ast.Identifiers{ - "blank", - "min_chars", - "neg", - "plus", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(511), - Column: int(18), - }, - End: ast.Location{ - Line: int(511), - Column: int(27), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5389, - FreeVars: ast.Identifiers{ - "min_chars", - }, - }, - Id: "min_chars", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(4), - Right: &ast.Conditional{ + Binds: ast.LocalBinds{ + ast.LocalBind{ + VarFodder: ast.Fodder{}, + Variable: "dec", + EqFodder: ast.Fodder{}, + Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(511), - Column: int(31), + Line: int(494), + Column: int(19), }, End: ast.Location{ - Line: int(511), - Column: int(68), + Line: int(494), + Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5389, + Ctx: p5367, FreeVars: ast.Identifiers{ - "blank", - "neg", - "plus", + "aux", + "n_", + "std", }, }, Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(511), - Column: int(34), + Line: int(494), + Column: int(22), }, End: ast.Location{ - Line: int(511), - Column: int(54), + Line: int(494), + Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5389, + Ctx: p5367, FreeVars: ast.Identifiers{ - "blank", - "neg", - "plus", + "n_", + "std", }, }, - Left: &ast.Binary{ + Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(511), - Column: int(34), + Line: int(494), + Column: int(22), }, End: ast.Location{ - Line: int(511), - Column: int(46), + Line: int(494), + Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5389, + Ctx: p5367, FreeVars: ast.Identifiers{ - "blank", - "neg", + "n_", + "std", }, }, - Left: &ast.Var{ + Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(511), - Column: int(34), + Line: int(494), + Column: int(22), }, End: ast.Location{ - Line: int(511), - Column: int(37), + Line: int(494), + Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5389, + Ctx: p5367, FreeVars: ast.Identifiers{ - "neg", + "std", }, }, - Id: "neg", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(18), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(511), - Column: int(41), + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(494), + Column: int(22), + }, + End: ast.Location{ + Line: int(494), + Column: int(25), + }, + File: p1, }, - End: ast.Location{ - Line: int(511), - Column: int(46), + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p5389, - FreeVars: ast.Identifiers{ - "blank", + Id: "std", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "floor", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, + }, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(494), + Column: int(32), + }, + End: ast.Location{ + Line: int(494), + Column: int(34), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5380, + FreeVars: ast.Identifiers{ + "n_", + }, + }, + Id: "n_", + }, + CommaFodder: nil, }, }, - Id: "blank", + Named: nil, }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(18), - Right: &ast.Var{ + Op: ast.BinaryOp(12), + Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(511), - Column: int(50), + Line: int(494), + Column: int(39), }, End: ast.Location{ - Line: int(511), - Column: int(54), + Line: int(494), + Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5389, - FreeVars: ast.Identifiers{ - "plus", - }, + Ctx: p5367, + FreeVars: nil, }, - Id: "plus", + Value: float64(0), + OriginalString: "0", }, }, ThenFodder: ast.Fodder{}, - BranchTrue: &ast.LiteralNumber{ + BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(511), - Column: int(60), + Line: int(494), + Column: int(46), }, End: ast.Location{ - Line: int(511), - Column: int(61), + Line: int(494), + Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5389, + Ctx: p5367, FreeVars: nil, }, - OriginalString: "1", + Value: "0", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", }, ElseFodder: ast.Fodder{}, - BranchFalse: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(511), - Column: int(67), - }, - End: ast.Location{ - Line: int(511), - Column: int(68), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5389, - FreeVars: nil, - }, - OriginalString: "0", - }, - }, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(511), - Column: int(13), - }, - End: ast.Location{ - Line: int(511), - Column: int(69), - }, - File: p1, - }, - }, - }, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(512), - Column: int(7), - }, - End: ast.Location{ - Line: int(514), - Column: int(84), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - Ctx: p5303, - FreeVars: ast.Identifiers{ - "blank", - "dec", - "min_digits", - "neg", - "pad_left", - "plus", - "std", - "zp", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "zp2", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(512), - Column: int(19), - }, - End: ast.Location{ - Line: int(512), - Column: int(42), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5412, - FreeVars: ast.Identifiers{ - "min_digits", - "std", - "zp", - }, - }, - Target: &ast.Index{ + BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(512), - Column: int(19), + Line: int(494), + Column: int(55), }, End: ast.Location{ - Line: int(512), - Column: int(26), + Line: int(494), + Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5412, + Ctx: p5367, FreeVars: ast.Identifiers{ + "aux", + "n_", "std", }, }, Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(512), - Column: int(19), + Line: int(494), + Column: int(55), }, End: ast.Location{ - Line: int(512), - Column: int(22), + Line: int(494), + Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: nil, + Ctx: p5367, FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + "aux", }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, }, - Value: "max", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", + Id: "aux", }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(512), - Column: int(27), + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(494), + Column: int(59), + }, + End: ast.Location{ + Line: int(494), + Column: int(72), + }, + File: p1, }, - End: ast.Location{ - Line: int(512), - Column: int(29), + Fodder: ast.Fodder{}, + Ctx: p5390, + FreeVars: ast.Identifiers{ + "n_", + "std", }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5421, - FreeVars: ast.Identifiers{ - "zp", }, - }, - Id: "zp", - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(512), - Column: int(31), + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(494), + Column: int(59), + }, + End: ast.Location{ + Line: int(494), + Column: int(68), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5390, + FreeVars: ast.Identifiers{ + "std", + }, }, - End: ast.Location{ - Line: int(512), - Column: int(41), + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(494), + Column: int(59), + }, + End: ast.Location{ + Line: int(494), + Column: int(62), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", }, - File: p1, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "floor", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, }, - Fodder: ast.Fodder{}, - Ctx: p5421, - FreeVars: ast.Identifiers{ - "min_digits", + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(494), + Column: int(69), + }, + End: ast.Location{ + Line: int(494), + Column: int(71), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5399, + FreeVars: ast.Identifiers{ + "n_", + }, + }, + Id: "n_", + }, + CommaFodder: nil, + }, + }, + Named: nil, }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, - Id: "min_digits", + CommaFodder: nil, }, - CommaFodder: nil, }, + Named: nil, }, - Named: nil, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(512), - Column: int(13), - }, - End: ast.Location{ - Line: int(512), - Column: int(42), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(513), + Line: int(495), Column: int(7), }, End: ast.Location{ - Line: int(514), + Line: int(499), Column: int(84), }, File: p1, @@ -71829,444 +67486,1006 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5303, + Ctx: p5290, FreeVars: ast.Identifiers{ "blank", "dec", - "neg", + "min_chars", + "min_digits", + "n__", "pad_left", - "plus", - "zp2", + "sign", + "std", }, }, Binds: ast.LocalBinds{ ast.LocalBind{ VarFodder: ast.Fodder{}, - Variable: "dec2", + Variable: "neg", EqFodder: ast.Fodder{}, - Body: &ast.Apply{ + Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(513), - Column: int(20), + Line: int(495), + Column: int(19), }, End: ast.Location{ - Line: int(513), - Column: int(43), + Line: int(495), + Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5430, + Ctx: p5406, FreeVars: ast.Identifiers{ - "dec", - "pad_left", - "zp2", + "n__", }, }, - Target: &ast.Var{ + Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(513), - Column: int(20), + Line: int(495), + Column: int(19), }, End: ast.Location{ - Line: int(513), - Column: int(28), + Line: int(495), + Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5430, + Ctx: p5406, FreeVars: ast.Identifiers{ - "pad_left", + "n__", }, }, - Id: "pad_left", + Id: "n__", }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(513), - Column: int(29), - }, - End: ast.Location{ - Line: int(513), - Column: int(32), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5436, - FreeVars: ast.Identifiers{ - "dec", - }, - }, - Id: "dec", + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(9), + Right: &ast.LiteralNumber{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(495), + Column: int(25), }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(513), - Column: int(34), - }, - End: ast.Location{ - Line: int(513), - Column: int(37), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5436, - FreeVars: ast.Identifiers{ - "zp2", - }, - }, - Id: "zp2", + End: ast.Location{ + Line: int(495), + Column: int(26), }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(513), - Column: int(39), - }, - End: ast.Location{ - Line: int(513), - Column: int(42), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5436, - FreeVars: nil, - }, - Value: "0", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - CommaFodder: nil, + File: p1, }, + Fodder: ast.Fodder{}, + Ctx: p5406, + FreeVars: nil, }, - Named: nil, + Value: float64(0), + OriginalString: "0", }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(513), - Column: int(13), - }, - End: ast.Location{ - Line: int(513), - Column: int(43), - }, - File: p1, - }, }, }, - Body: &ast.Binary{ + Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(514), + Line: int(496), Column: int(7), }, End: ast.Location{ - Line: int(514), + Line: int(499), Column: int(84), }, File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p5303, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(6), + Comment: []string{}, + }, + }, + Ctx: p5290, FreeVars: ast.Identifiers{ "blank", - "dec2", + "dec", + "min_chars", + "min_digits", "neg", - "plus", + "pad_left", + "sign", + "std", + }, + }, + Binds: ast.LocalBinds{ + ast.LocalBind{ + VarFodder: ast.Fodder{}, + Variable: "zp", + EqFodder: ast.Fodder{}, + Body: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(18), + }, + End: ast.Location{ + Line: int(496), + Column: int(69), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "blank", + "min_chars", + "neg", + "sign", + }, + }, + Left: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(18), + }, + End: ast.Location{ + Line: int(496), + Column: int(27), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "min_chars", + }, + }, + Id: "min_chars", + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(4), + Right: &ast.Conditional{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(31), + }, + End: ast.Location{ + Line: int(496), + Column: int(68), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "blank", + "neg", + "sign", + }, + }, + Cond: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(34), + }, + End: ast.Location{ + Line: int(496), + Column: int(54), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "blank", + "neg", + "sign", + }, + }, + Left: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(34), + }, + End: ast.Location{ + Line: int(496), + Column: int(46), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "blank", + "neg", + }, + }, + Left: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(34), + }, + End: ast.Location{ + Line: int(496), + Column: int(37), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "neg", + }, + }, + Id: "neg", + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(18), + Right: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(41), + }, + End: ast.Location{ + Line: int(496), + Column: int(46), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "blank", + }, + }, + Id: "blank", + }, + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(18), + Right: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(50), + }, + End: ast.Location{ + Line: int(496), + Column: int(54), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: ast.Identifiers{ + "sign", + }, + }, + Id: "sign", + }, + }, + ThenFodder: ast.Fodder{}, + BranchTrue: &ast.LiteralNumber{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(60), + }, + End: ast.Location{ + Line: int(496), + Column: int(61), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: nil, + }, + Value: float64(1), + OriginalString: "1", + }, + ElseFodder: ast.Fodder{}, + BranchFalse: &ast.LiteralNumber{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(496), + Column: int(67), + }, + End: ast.Location{ + Line: int(496), + Column: int(68), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5416, + FreeVars: nil, + }, + Value: float64(0), + OriginalString: "0", + }, + }, + }, + Fun: nil, + CloseFodder: ast.Fodder{}, }, }, - Left: &ast.Conditional{ + Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(514), - Column: int(8), + Line: int(497), + Column: int(7), }, End: ast.Location{ - Line: int(514), - Column: int(76), + Line: int(499), + Column: int(84), }, File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p5303, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(6), + Comment: []string{}, + }, + }, + Ctx: p5290, FreeVars: ast.Identifiers{ "blank", + "dec", + "min_digits", "neg", - "plus", + "pad_left", + "sign", + "std", + "zp", }, }, - Cond: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(514), - Column: int(11), + Binds: ast.LocalBinds{ + ast.LocalBind{ + VarFodder: ast.Fodder{}, + Variable: "zp2", + EqFodder: ast.Fodder{}, + Body: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(497), + Column: int(19), + }, + End: ast.Location{ + Line: int(497), + Column: int(42), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5439, + FreeVars: ast.Identifiers{ + "min_digits", + "std", + "zp", + }, }, - End: ast.Location{ - Line: int(514), - Column: int(14), + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(497), + Column: int(19), + }, + End: ast.Location{ + Line: int(497), + Column: int(26), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5439, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(497), + Column: int(19), + }, + End: ast.Location{ + Line: int(497), + Column: int(22), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "max", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5303, - FreeVars: ast.Identifiers{ - "neg", + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(497), + Column: int(27), + }, + End: ast.Location{ + Line: int(497), + Column: int(29), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5448, + FreeVars: ast.Identifiers{ + "zp", + }, + }, + Id: "zp", + }, + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(497), + Column: int(31), + }, + End: ast.Location{ + Line: int(497), + Column: int(41), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5448, + FreeVars: ast.Identifiers{ + "min_digits", + }, + }, + Id: "min_digits", + }, + CommaFodder: nil, + }, + }, + Named: nil, + }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, + Fun: nil, + CloseFodder: ast.Fodder{}, }, - Id: "neg", }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.LiteralString{ + Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(514), - Column: int(20), + Line: int(498), + Column: int(7), }, End: ast.Location{ - Line: int(514), - Column: int(23), + Line: int(499), + Column: int(84), }, File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p5303, - FreeVars: nil, - }, - Value: "-", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - ElseFodder: ast.Fodder{}, - BranchFalse: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(514), - Column: int(29), - }, - End: ast.Location{ - Line: int(514), - Column: int(76), + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(6), + Comment: []string{}, }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p5303, + Ctx: p5290, FreeVars: ast.Identifiers{ "blank", - "plus", + "dec", + "neg", + "pad_left", + "sign", + "zp2", }, }, - Cond: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(514), - Column: int(32), - }, - End: ast.Location{ - Line: int(514), - Column: int(36), + Binds: ast.LocalBinds{ + ast.LocalBind{ + VarFodder: ast.Fodder{}, + Variable: "dec2", + EqFodder: ast.Fodder{}, + Body: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(498), + Column: int(20), + }, + End: ast.Location{ + Line: int(498), + Column: int(43), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5457, + FreeVars: ast.Identifiers{ + "dec", + "pad_left", + "zp2", + }, }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5303, - FreeVars: ast.Identifiers{ - "plus", - }, - }, - Id: "plus", - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(514), - Column: int(42), + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(498), + Column: int(20), + }, + End: ast.Location{ + Line: int(498), + Column: int(28), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5457, + FreeVars: ast.Identifiers{ + "pad_left", + }, + }, + Id: "pad_left", }, - End: ast.Location{ - Line: int(514), - Column: int(45), + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(498), + Column: int(29), + }, + End: ast.Location{ + Line: int(498), + Column: int(32), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5463, + FreeVars: ast.Identifiers{ + "dec", + }, + }, + Id: "dec", + }, + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(498), + Column: int(34), + }, + End: ast.Location{ + Line: int(498), + Column: int(37), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5463, + FreeVars: ast.Identifiers{ + "zp2", + }, + }, + Id: "zp2", + }, + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedExpr{ + Expr: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(498), + Column: int(39), + }, + End: ast.Location{ + Line: int(498), + Column: int(42), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5463, + FreeVars: nil, + }, + Value: "0", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + CommaFodder: nil, + }, + }, + Named: nil, }, - File: p1, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, - Fodder: ast.Fodder{}, - Ctx: p5303, - FreeVars: nil, + Fun: nil, + CloseFodder: ast.Fodder{}, }, - Value: "+", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", }, - ElseFodder: ast.Fodder{}, - BranchFalse: &ast.Conditional{ + Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(514), - Column: int(51), + Line: int(499), + Column: int(7), }, End: ast.Location{ - Line: int(514), - Column: int(76), + Line: int(499), + Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5303, + Ctx: p5290, FreeVars: ast.Identifiers{ "blank", + "dec2", + "neg", + "sign", }, }, - Cond: &ast.Var{ + Left: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(514), - Column: int(54), + Line: int(499), + Column: int(8), }, End: ast.Location{ - Line: int(514), - Column: int(59), + Line: int(499), + Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5303, + Ctx: p5290, FreeVars: ast.Identifiers{ "blank", + "neg", + "sign", }, }, - Id: "blank", - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(514), - Column: int(65), + Cond: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(11), + }, + End: ast.Location{ + Line: int(499), + Column: int(14), + }, + File: p1, }, - End: ast.Location{ - Line: int(514), - Column: int(68), + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "neg", + }, + }, + Id: "neg", + }, + ThenFodder: ast.Fodder{}, + BranchTrue: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(20), + }, + End: ast.Location{ + Line: int(499), + Column: int(23), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: nil, + }, + Value: "-", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + ElseFodder: ast.Fodder{}, + BranchFalse: &ast.Conditional{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(29), + }, + End: ast.Location{ + Line: int(499), + Column: int(76), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "blank", + "sign", + }, + }, + Cond: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(32), + }, + End: ast.Location{ + Line: int(499), + Column: int(36), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "sign", + }, + }, + Id: "sign", + }, + ThenFodder: ast.Fodder{}, + BranchTrue: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(42), + }, + End: ast.Location{ + Line: int(499), + Column: int(45), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: nil, + }, + Value: "+", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + ElseFodder: ast.Fodder{}, + BranchFalse: &ast.Conditional{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(51), + }, + End: ast.Location{ + Line: int(499), + Column: int(76), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "blank", + }, + }, + Cond: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(54), + }, + End: ast.Location{ + Line: int(499), + Column: int(59), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "blank", + }, + }, + Id: "blank", + }, + ThenFodder: ast.Fodder{}, + BranchTrue: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(65), + }, + End: ast.Location{ + Line: int(499), + Column: int(68), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: nil, + }, + Value: " ", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + ElseFodder: ast.Fodder{}, + BranchFalse: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(499), + Column: int(74), + }, + End: ast.Location{ + Line: int(499), + Column: int(76), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5290, + FreeVars: nil, + }, + Value: "", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p5303, - FreeVars: nil, }, - Value: " ", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", }, - ElseFodder: ast.Fodder{}, - BranchFalse: &ast.LiteralString{ + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(3), + Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(514), - Column: int(74), + Line: int(499), + Column: int(80), }, End: ast.Location{ - Line: int(514), - Column: int(76), + Line: int(499), + Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5303, - FreeVars: nil, + Ctx: p5290, + FreeVars: ast.Identifiers{ + "dec2", + }, }, - Value: "", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - }, - }, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(514), - Column: int(80), - }, - End: ast.Location{ - Line: int(514), - Column: int(84), + Id: "dec2", }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5303, - FreeVars: ast.Identifiers{ - "dec2", }, }, - Id: "dec2", }, }, }, @@ -72276,30 +68495,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(517), + Line: int(502), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -72312,7 +68519,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -72338,172 +68545,77 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(517), + Line: int(502), Column: int(11), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, }, Fodder: nil, - Ctx: p5467, + Ctx: p5494, FreeVars: ast.Identifiers{ "pad_left", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "n__", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(517), - Column: int(22), - }, - End: ast.Location{ - Line: int(517), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n__", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "min_chars", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(517), - Column: int(27), - }, - End: ast.Location{ - Line: int(517), - Column: int(36), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "min_chars", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "min_digits", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(517), - Column: int(38), - }, - End: ast.Location{ - Line: int(517), - Column: int(48), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "min_digits", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "blank", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(517), - Column: int(50), - }, - End: ast.Location{ - Line: int(517), - Column: int(55), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "blank", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "plus", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(517), - Column: int(57), - }, - End: ast.Location{ - Line: int(517), - Column: int(61), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "sign", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "add_zerox", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(517), - Column: int(63), - }, - End: ast.Location{ - Line: int(517), - Column: int(72), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "add_zerox", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "capitals", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(517), - Column: int(74), - }, - End: ast.Location{ - Line: int(517), - Column: int(82), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "capitals", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -72516,7 +68628,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "blank", @@ -72525,7 +68637,7 @@ var _StdAst = &ast.DesugaredObject{ "min_digits", "n__", "pad_left", - "plus", + "sign", "std", }, }, @@ -72537,19 +68649,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(24), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5476, + Ctx: p5503, FreeVars: ast.Identifiers{ "capitals", }, @@ -72557,19 +68669,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(24), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5476, + Ctx: p5503, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -72577,21 +68689,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(25), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -72600,21 +68713,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(28), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, CommaFodder: ast.Fodder{}, @@ -72623,21 +68737,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(31), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, CommaFodder: ast.Fodder{}, @@ -72646,21 +68761,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(34), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, CommaFodder: ast.Fodder{}, @@ -72669,21 +68785,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(37), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, CommaFodder: ast.Fodder{}, @@ -72692,21 +68809,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(40), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(5), OriginalString: "5", }, CommaFodder: ast.Fodder{}, @@ -72715,21 +68833,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(43), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(6), OriginalString: "6", }, CommaFodder: ast.Fodder{}, @@ -72738,21 +68857,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(46), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(7), OriginalString: "7", }, CommaFodder: ast.Fodder{}, @@ -72761,21 +68881,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(49), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(8), OriginalString: "8", }, CommaFodder: ast.Fodder{}, @@ -72784,21 +68905,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(518), + Line: int(503), Column: int(52), }, End: ast.Location{ - Line: int(518), + Line: int(503), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5481, + Ctx: p5508, FreeVars: nil, }, + Value: float64(9), OriginalString: "9", }, CommaFodder: nil, @@ -72819,19 +68941,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(26), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5476, + Ctx: p5503, FreeVars: ast.Identifiers{ "capitals", }, @@ -72839,19 +68961,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(29), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5476, + Ctx: p5503, FreeVars: ast.Identifiers{ "capitals", }, @@ -72862,19 +68984,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(43), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5476, + Ctx: p5503, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -72882,25 +69004,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(44), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5499, + Ctx: p5526, FreeVars: nil, }, Value: "A", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -72908,25 +69029,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(49), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5499, + Ctx: p5526, FreeVars: nil, }, Value: "B", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -72934,25 +69054,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(54), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5499, + Ctx: p5526, FreeVars: nil, }, Value: "C", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -72960,25 +69079,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(59), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5499, + Ctx: p5526, FreeVars: nil, }, Value: "D", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -72986,25 +69104,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(64), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5499, + Ctx: p5526, FreeVars: nil, }, Value: "E", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -73012,25 +69129,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(519), + Line: int(504), Column: int(69), }, End: ast.Location{ - Line: int(519), + Line: int(504), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5499, + Ctx: p5526, FreeVars: nil, }, Value: "F", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -73049,19 +69165,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(520), + Line: int(505), Column: int(29), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5476, + Ctx: p5503, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -73069,25 +69185,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(520), + Line: int(505), Column: int(30), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5509, + Ctx: p5536, FreeVars: nil, }, Value: "a", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -73095,25 +69210,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(520), + Line: int(505), Column: int(35), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5509, + Ctx: p5536, FreeVars: nil, }, Value: "b", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -73121,25 +69235,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(520), + Line: int(505), Column: int(40), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5509, + Ctx: p5536, FreeVars: nil, }, Value: "c", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -73147,25 +69260,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(520), + Line: int(505), Column: int(45), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5509, + Ctx: p5536, FreeVars: nil, }, Value: "d", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -73173,25 +69285,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(520), + Line: int(505), Column: int(50), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5509, + Ctx: p5536, FreeVars: nil, }, Value: "e", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -73199,25 +69310,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(520), + Line: int(505), Column: int(55), }, End: ast.Location{ - Line: int(520), + Line: int(505), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5509, + Ctx: p5536, FreeVars: nil, }, Value: "f", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -73229,30 +69339,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(518), - Column: int(13), - }, - End: ast.Location{ - Line: int(520), - Column: int(59), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(521), + Line: int(506), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -73265,7 +69363,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "blank", @@ -73275,7 +69373,7 @@ var _StdAst = &ast.DesugaredObject{ "n__", "numerals", "pad_left", - "plus", + "sign", "std", }, }, @@ -73287,19 +69385,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(521), + Line: int(506), Column: int(18), }, End: ast.Location{ - Line: int(521), + Line: int(506), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5520, + Ctx: p5547, FreeVars: ast.Identifiers{ "n__", "std", @@ -73308,19 +69406,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(521), + Line: int(506), Column: int(18), }, End: ast.Location{ - Line: int(521), + Line: int(506), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5520, + Ctx: p5547, FreeVars: ast.Identifiers{ "std", }, @@ -73328,13 +69426,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(521), + Line: int(506), Column: int(18), }, End: ast.Location{ - Line: int(521), + Line: int(506), Column: int(21), }, File: p1, @@ -73369,9 +69467,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "abs", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -73381,19 +69478,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(521), + Line: int(506), Column: int(26), }, End: ast.Location{ - Line: int(521), + Line: int(506), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5529, + Ctx: p5556, FreeVars: ast.Identifiers{ "n__", }, @@ -73412,30 +69509,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(521), - Column: int(13), - }, - End: ast.Location{ - Line: int(521), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(522), + Line: int(507), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -73448,7 +69533,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "blank", @@ -73459,7 +69544,7 @@ var _StdAst = &ast.DesugaredObject{ "n__", "numerals", "pad_left", - "plus", + "sign", "std", }, }, @@ -73471,19 +69556,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(522), + Line: int(507), Column: int(13), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(52), }, File: p1, }, Fodder: nil, - Ctx: p5536, + Ctx: p5563, FreeVars: ast.Identifiers{ "aux", "numerals", @@ -73491,39 +69576,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "n", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(522), - Column: int(17), - }, - End: ast.Location{ - Line: int(522), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(523), + Line: int(508), Column: int(9), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(52), }, File: p1, @@ -73536,7 +69610,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "aux", "n", @@ -73547,19 +69621,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(523), + Line: int(508), Column: int(12), }, End: ast.Location{ - Line: int(523), + Line: int(508), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "n", }, @@ -73567,19 +69641,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(523), + Line: int(508), Column: int(12), }, End: ast.Location{ - Line: int(523), + Line: int(508), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "n", }, @@ -73591,21 +69665,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(523), + Line: int(508), Column: int(17), }, End: ast.Location{ - Line: int(523), + Line: int(508), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -73613,13 +69688,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(524), + Line: int(509), Column: int(11), }, End: ast.Location{ - Line: int(524), + Line: int(509), Column: int(13), }, File: p1, @@ -73632,13 +69707,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5541, + Ctx: p5568, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -73651,19 +69725,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(11), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "aux", "n", @@ -73674,19 +69748,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(11), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "aux", "n", @@ -73696,13 +69770,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(11), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(14), }, File: p1, @@ -73715,7 +69789,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "aux", }, @@ -73729,19 +69803,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(15), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5560, + Ctx: p5587, FreeVars: ast.Identifiers{ "n", "std", @@ -73750,19 +69824,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(15), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5560, + Ctx: p5587, FreeVars: ast.Identifiers{ "std", }, @@ -73770,13 +69844,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(15), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(18), }, File: p1, @@ -73811,9 +69885,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -73823,19 +69896,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(25), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5569, + Ctx: p5596, FreeVars: ast.Identifiers{ "n", }, @@ -73843,19 +69916,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(25), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5569, + Ctx: p5596, FreeVars: ast.Identifiers{ "n", }, @@ -73867,21 +69940,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(29), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5569, + Ctx: p5596, FreeVars: nil, }, + Value: float64(16), OriginalString: "16", }, }, @@ -73910,19 +69984,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(36), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "n", "numerals", @@ -73932,19 +70006,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(36), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "numerals", }, @@ -73957,14 +70031,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(526), - Column: int(45), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(526), - Column: int(51), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -74037,7 +70111,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -74049,19 +70122,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(45), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: ast.Identifiers{ "n", }, @@ -74074,21 +70147,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(526), + Line: int(511), Column: int(49), }, End: ast.Location{ - Line: int(526), + Line: int(511), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5541, + Ctx: p5568, FreeVars: nil, }, + Value: float64(16), OriginalString: "16", }, CommaFodder: nil, @@ -74109,30 +70183,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -74145,7 +70207,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "aux", @@ -74156,7 +70218,7 @@ var _StdAst = &ast.DesugaredObject{ "n_", "n__", "pad_left", - "plus", + "sign", "std", }, }, @@ -74168,19 +70230,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(19), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: ast.Identifiers{ "aux", "n_", @@ -74190,19 +70252,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(22), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: ast.Identifiers{ "n_", "std", @@ -74211,19 +70273,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(22), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: ast.Identifiers{ "n_", "std", @@ -74232,19 +70294,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(22), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: ast.Identifiers{ "std", }, @@ -74252,13 +70314,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(22), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(25), }, File: p1, @@ -74293,9 +70355,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -74305,19 +70366,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(32), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5607, + Ctx: p5634, FreeVars: ast.Identifiers{ "n_", }, @@ -74339,21 +70400,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(39), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -74361,43 +70423,42 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(46), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: nil, }, Value: "0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(55), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: ast.Identifiers{ "aux", "n_", @@ -74407,19 +70468,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(55), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5594, + Ctx: p5621, FreeVars: ast.Identifiers{ "aux", }, @@ -74433,19 +70494,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(59), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5617, + Ctx: p5644, FreeVars: ast.Identifiers{ "n_", "std", @@ -74454,19 +70515,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(59), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5617, + Ctx: p5644, FreeVars: ast.Identifiers{ "std", }, @@ -74474,13 +70535,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(59), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(62), }, File: p1, @@ -74515,9 +70576,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -74527,19 +70587,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(527), + Line: int(512), Column: int(69), }, End: ast.Location{ - Line: int(527), + Line: int(512), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5626, + Ctx: p5653, FreeVars: ast.Identifiers{ "n_", }, @@ -74569,30 +70629,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(527), - Column: int(13), - }, - End: ast.Location{ - Line: int(527), - Column: int(73), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(528), + Line: int(513), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -74605,7 +70653,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "blank", @@ -74615,7 +70663,7 @@ var _StdAst = &ast.DesugaredObject{ "min_digits", "n__", "pad_left", - "plus", + "sign", "std", }, }, @@ -74627,19 +70675,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(528), + Line: int(513), Column: int(19), }, End: ast.Location{ - Line: int(528), + Line: int(513), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5633, + Ctx: p5660, FreeVars: ast.Identifiers{ "n__", }, @@ -74647,19 +70695,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(528), + Line: int(513), Column: int(19), }, End: ast.Location{ - Line: int(528), + Line: int(513), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5633, + Ctx: p5660, FreeVars: ast.Identifiers{ "n__", }, @@ -74671,50 +70719,39 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(528), + Line: int(513), Column: int(25), }, End: ast.Location{ - Line: int(528), + Line: int(513), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5633, + Ctx: p5660, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(528), - Column: int(13), - }, - End: ast.Location{ - Line: int(528), - Column: int(26), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -74727,7 +70764,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "blank", @@ -74737,7 +70774,7 @@ var _StdAst = &ast.DesugaredObject{ "min_digits", "neg", "pad_left", - "plus", + "sign", "std", }, }, @@ -74749,66 +70786,66 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(18), }, End: ast.Location{ - Line: int(530), + Line: int(515), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "add_zerox", "blank", "min_chars", "neg", - "plus", + "sign", }, }, Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(18), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "blank", "min_chars", "neg", - "plus", + "sign", }, }, Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(18), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "min_chars", }, @@ -74820,63 +70857,63 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(31), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "blank", "neg", - "plus", + "sign", }, }, Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(34), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "blank", "neg", - "plus", + "sign", }, }, Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(34), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "blank", "neg", @@ -74885,19 +70922,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(34), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "neg", }, @@ -74909,19 +70946,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(41), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "blank", }, @@ -74934,66 +70971,68 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(50), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ - "plus", + "sign", }, }, - Id: "plus", + Id: "sign", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(60), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(529), + Line: int(514), Column: int(67), }, End: ast.Location{ - Line: int(529), + Line: int(514), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -75010,19 +71049,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(530), + Line: int(515), Column: int(21), }, End: ast.Location{ - Line: int(530), + Line: int(515), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "add_zerox", }, @@ -75030,19 +71069,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(530), + Line: int(515), Column: int(24), }, End: ast.Location{ - Line: int(530), + Line: int(515), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: ast.Identifiers{ "add_zerox", }, @@ -75053,72 +71092,62 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(530), + Line: int(515), Column: int(39), }, End: ast.Location{ - Line: int(530), + Line: int(515), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(530), + Line: int(515), Column: int(46), }, End: ast.Location{ - Line: int(530), + Line: int(515), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5643, + Ctx: p5670, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(529), - Column: int(13), - }, - End: ast.Location{ - Line: int(530), - Column: int(48), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(531), + Line: int(516), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -75131,7 +71160,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "blank", @@ -75140,7 +71169,7 @@ var _StdAst = &ast.DesugaredObject{ "min_digits", "neg", "pad_left", - "plus", + "sign", "std", "zp", }, @@ -75153,19 +71182,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(531), + Line: int(516), Column: int(19), }, End: ast.Location{ - Line: int(531), + Line: int(516), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5675, + Ctx: p5702, FreeVars: ast.Identifiers{ "min_digits", "std", @@ -75175,19 +71204,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(531), + Line: int(516), Column: int(19), }, End: ast.Location{ - Line: int(531), + Line: int(516), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5675, + Ctx: p5702, FreeVars: ast.Identifiers{ "std", }, @@ -75195,13 +71224,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(531), + Line: int(516), Column: int(19), }, End: ast.Location{ - Line: int(531), + Line: int(516), Column: int(22), }, File: p1, @@ -75236,9 +71265,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "max", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -75248,19 +71276,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(531), + Line: int(516), Column: int(27), }, End: ast.Location{ - Line: int(531), + Line: int(516), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5684, + Ctx: p5711, FreeVars: ast.Identifiers{ "zp", }, @@ -75273,19 +71301,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(531), + Line: int(516), Column: int(31), }, End: ast.Location{ - Line: int(531), + Line: int(516), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5684, + Ctx: p5711, FreeVars: ast.Identifiers{ "min_digits", }, @@ -75304,30 +71332,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(531), - Column: int(13), - }, - End: ast.Location{ - Line: int(531), - Column: int(42), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, @@ -75340,7 +71356,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "add_zerox", "blank", @@ -75348,7 +71364,7 @@ var _StdAst = &ast.DesugaredObject{ "hex", "neg", "pad_left", - "plus", + "sign", "zp2", }, }, @@ -75360,19 +71376,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(20), }, End: ast.Location{ - Line: int(533), + Line: int(518), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: ast.Identifiers{ "add_zerox", "capitals", @@ -75384,19 +71400,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(21), }, End: ast.Location{ - Line: int(532), + Line: int(517), Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: ast.Identifiers{ "add_zerox", "capitals", @@ -75405,19 +71421,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(24), }, End: ast.Location{ - Line: int(532), + Line: int(517), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: ast.Identifiers{ "add_zerox", }, @@ -75428,19 +71444,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(40), }, End: ast.Location{ - Line: int(532), + Line: int(517), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: ast.Identifiers{ "capitals", }, @@ -75448,19 +71464,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(43), }, End: ast.Location{ - Line: int(532), + Line: int(517), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: ast.Identifiers{ "capitals", }, @@ -75471,74 +71487,71 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(57), }, End: ast.Location{ - Line: int(532), + Line: int(517), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: nil, }, Value: "0X", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(67), }, End: ast.Location{ - Line: int(532), + Line: int(517), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: nil, }, Value: "0x", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(532), + Line: int(517), Column: int(78), }, End: ast.Location{ - Line: int(532), + Line: int(517), Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{ @@ -75553,19 +71566,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(533), + Line: int(518), Column: int(22), }, End: ast.Location{ - Line: int(533), + Line: int(518), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: ast.Identifiers{ "hex", "pad_left", @@ -75575,19 +71588,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(533), + Line: int(518), Column: int(22), }, End: ast.Location{ - Line: int(533), + Line: int(518), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5693, + Ctx: p5720, FreeVars: ast.Identifiers{ "pad_left", }, @@ -75601,19 +71614,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(533), + Line: int(518), Column: int(31), }, End: ast.Location{ - Line: int(533), + Line: int(518), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5713, + Ctx: p5740, FreeVars: ast.Identifiers{ "hex", }, @@ -75626,19 +71639,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(533), + Line: int(518), Column: int(36), }, End: ast.Location{ - Line: int(533), + Line: int(518), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5713, + Ctx: p5740, FreeVars: ast.Identifiers{ "zp2", }, @@ -75651,25 +71664,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(533), + Line: int(518), Column: int(41), }, End: ast.Location{ - Line: int(533), + Line: int(518), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5713, + Ctx: p5740, FreeVars: nil, }, Value: "0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -75684,81 +71696,69 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(532), - Column: int(13), - }, - End: ast.Location{ - Line: int(533), - Column: int(45), - }, - File: p1, - }, }, }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(7), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "blank", "hex2", "neg", - "plus", + "sign", }, }, Left: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(8), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "blank", "neg", - "plus", + "sign", }, }, Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(11), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "neg", }, @@ -75769,111 +71769,109 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(20), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: nil, }, Value: "-", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(29), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "blank", - "plus", + "sign", }, }, Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(32), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ - "plus", + "sign", }, }, - Id: "plus", + Id: "sign", }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(42), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: nil, }, Value: "+", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(51), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "blank", }, @@ -75881,19 +71879,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(54), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "blank", }, @@ -75904,49 +71902,47 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(65), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(74), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -75956,19 +71952,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(534), + Line: int(519), Column: int(80), }, End: ast.Location{ - Line: int(534), + Line: int(519), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5472, + Ctx: p5499, FreeVars: ast.Identifiers{ "hex2", }, @@ -75987,30 +71983,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(536), + Line: int(521), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -76042,57 +72026,46 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(536), + Line: int(521), Column: int(11), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(36), }, File: p1, }, Fodder: nil, - Ctx: p5743, + Ctx: p5770, FreeVars: ast.Identifiers{ "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(536), - Column: int(31), - }, - End: ast.Location{ - Line: int(536), - Column: int(34), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(537), + Line: int(522), Column: int(7), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(36), }, File: p1, @@ -76105,7 +72078,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5748, + Ctx: p5775, FreeVars: ast.Identifiers{ "std", "str", @@ -76119,77 +72092,52 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(537), + Line: int(522), Column: int(13), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(38), }, File: p1, }, Fodder: nil, - Ctx: p5752, + Ctx: p5779, FreeVars: ast.Identifiers{ "aux", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(537), - Column: int(17), - }, - End: ast.Location{ - Line: int(537), - Column: int(20), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(537), - Column: int(22), - }, - End: ast.Location{ - Line: int(537), - Column: int(23), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(538), + Line: int(523), Column: int(9), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(38), }, File: p1, @@ -76202,7 +72150,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "aux", "i", @@ -76213,19 +72161,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(538), + Line: int(523), Column: int(12), }, End: ast.Location{ - Line: int(538), + Line: int(523), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "i", }, @@ -76233,19 +72181,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(538), + Line: int(523), Column: int(12), }, End: ast.Location{ - Line: int(538), + Line: int(523), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "i", }, @@ -76257,21 +72205,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(538), + Line: int(523), Column: int(16), }, End: ast.Location{ - Line: int(538), + Line: int(523), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -76279,13 +72228,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(539), + Line: int(524), Column: int(11), }, End: ast.Location{ - Line: int(539), + Line: int(524), Column: int(13), }, File: p1, @@ -76298,13 +72247,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5757, + Ctx: p5784, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -76317,13 +72265,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(541), + Line: int(526), Column: int(11), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(38), }, File: p1, @@ -76336,7 +72284,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "aux", "i", @@ -76347,19 +72295,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(541), + Line: int(526), Column: int(14), }, End: ast.Location{ - Line: int(541), + Line: int(526), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "i", "str", @@ -76368,19 +72316,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(541), + Line: int(526), Column: int(14), }, End: ast.Location{ - Line: int(541), + Line: int(526), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "i", "str", @@ -76389,19 +72337,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(541), + Line: int(526), Column: int(14), }, End: ast.Location{ - Line: int(541), + Line: int(526), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "str", }, @@ -76412,19 +72360,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(541), + Line: int(526), Column: int(18), }, End: ast.Location{ - Line: int(541), + Line: int(526), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "i", }, @@ -76439,44 +72387,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(541), + Line: int(526), Column: int(24), }, End: ast.Location{ - Line: int(541), + Line: int(526), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: nil, }, Value: "0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(542), + Line: int(527), Column: int(13), }, End: ast.Location{ - Line: int(542), + Line: int(527), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "aux", "i", @@ -76486,13 +72433,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(542), + Line: int(527), Column: int(13), }, End: ast.Location{ - Line: int(542), + Line: int(527), Column: int(16), }, File: p1, @@ -76505,7 +72452,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "aux", }, @@ -76519,19 +72466,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(542), + Line: int(527), Column: int(17), }, End: ast.Location{ - Line: int(542), + Line: int(527), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5786, + Ctx: p5813, FreeVars: ast.Identifiers{ "str", }, @@ -76544,19 +72491,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(542), + Line: int(527), Column: int(22), }, End: ast.Location{ - Line: int(542), + Line: int(527), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5786, + Ctx: p5813, FreeVars: ast.Identifiers{ "i", }, @@ -76564,19 +72511,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(542), + Line: int(527), Column: int(22), }, End: ast.Location{ - Line: int(542), + Line: int(527), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5786, + Ctx: p5813, FreeVars: ast.Identifiers{ "i", }, @@ -76588,21 +72535,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(542), + Line: int(527), Column: int(26), }, End: ast.Location{ - Line: int(542), + Line: int(527), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5786, + Ctx: p5813, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -76627,19 +72575,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(13), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "i", "std", @@ -76649,19 +72597,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(13), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5757, + Ctx: p5784, FreeVars: ast.Identifiers{ "std", }, @@ -76669,13 +72617,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(13), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(16), }, File: p1, @@ -76717,9 +72665,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "substr", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -76729,19 +72676,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(24), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5804, + Ctx: p5831, FreeVars: ast.Identifiers{ "str", }, @@ -76754,21 +72701,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(29), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5804, + Ctx: p5831, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -76777,19 +72725,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(32), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5804, + Ctx: p5831, FreeVars: ast.Identifiers{ "i", }, @@ -76797,19 +72745,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(32), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5804, + Ctx: p5831, FreeVars: ast.Identifiers{ "i", }, @@ -76821,21 +72769,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(544), + Line: int(529), Column: int(36), }, End: ast.Location{ - Line: int(544), + Line: int(529), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5804, + Ctx: p5831, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -76854,36 +72803,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(7), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5748, + Ctx: p5775, FreeVars: ast.Identifiers{ "aux", "std", @@ -76893,13 +72830,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(7), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(10), }, File: p1, @@ -76912,7 +72849,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5748, + Ctx: p5775, FreeVars: ast.Identifiers{ "aux", }, @@ -76926,19 +72863,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(11), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5819, + Ctx: p5846, FreeVars: ast.Identifiers{ "str", }, @@ -76951,19 +72888,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(16), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5819, + Ctx: p5846, FreeVars: ast.Identifiers{ "std", "str", @@ -76972,19 +72909,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(16), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5819, + Ctx: p5846, FreeVars: ast.Identifiers{ "std", "str", @@ -76993,19 +72930,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(16), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5819, + Ctx: p5846, FreeVars: ast.Identifiers{ "std", }, @@ -77013,13 +72950,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(16), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(19), }, File: p1, @@ -77054,9 +72991,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -77066,19 +73002,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(27), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5832, + Ctx: p5859, FreeVars: ast.Identifiers{ "str", }, @@ -77100,21 +73036,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(545), + Line: int(530), Column: int(34), }, End: ast.Location{ - Line: int(545), + Line: int(530), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5819, + Ctx: p5846, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -77132,30 +73069,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(548), + Line: int(533), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -77168,7 +73093,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -77196,19 +73121,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(548), + Line: int(533), Column: int(11), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, }, Fodder: nil, - Ctx: p5841, + Ctx: p5868, FreeVars: ast.Identifiers{ "render_int", "std", @@ -77216,153 +73141,58 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "n__", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(548), - Column: int(28), - }, - End: ast.Location{ - Line: int(548), - Column: int(31), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n__", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "zero_pad", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(548), - Column: int(33), - }, - End: ast.Location{ - Line: int(548), - Column: int(41), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "zero_pad", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "blank", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(548), - Column: int(43), - }, - End: ast.Location{ - Line: int(548), - Column: int(48), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "blank", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "plus", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(548), - Column: int(50), - }, - End: ast.Location{ - Line: int(548), - Column: int(54), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "sign", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "ensure_pt", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(548), - Column: int(56), - }, - End: ast.Location{ - Line: int(548), - Column: int(65), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "ensure_pt", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "trailing", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(548), - Column: int(67), - }, - End: ast.Location{ - Line: int(548), - Column: int(75), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "trailing", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "prec", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(548), - Column: int(77), - }, - End: ast.Location{ - Line: int(548), - Column: int(81), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "prec", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(549), + Line: int(534), Column: int(7), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -77375,14 +73205,14 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "n__", - "plus", "prec", "render_int", + "sign", "std", "strip_trailing_zero", "trailing", @@ -77397,19 +73227,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(549), + Line: int(534), Column: int(18), }, End: ast.Location{ - Line: int(549), + Line: int(534), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5850, + Ctx: p5877, FreeVars: ast.Identifiers{ "n__", "std", @@ -77418,19 +73248,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(549), + Line: int(534), Column: int(18), }, End: ast.Location{ - Line: int(549), + Line: int(534), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5850, + Ctx: p5877, FreeVars: ast.Identifiers{ "std", }, @@ -77438,13 +73268,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(549), + Line: int(534), Column: int(18), }, End: ast.Location{ - Line: int(549), + Line: int(534), Column: int(21), }, File: p1, @@ -77479,9 +73309,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "abs", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -77491,19 +73320,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(549), + Line: int(534), Column: int(26), }, End: ast.Location{ - Line: int(549), + Line: int(534), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5859, + Ctx: p5886, FreeVars: ast.Identifiers{ "n__", }, @@ -77522,30 +73351,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(549), - Column: int(13), - }, - End: ast.Location{ - Line: int(549), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(550), + Line: int(535), Column: int(7), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -77558,15 +73375,15 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "n_", "n__", - "plus", "prec", "render_int", + "sign", "std", "strip_trailing_zero", "trailing", @@ -77581,19 +73398,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(550), + Line: int(535), Column: int(21), }, End: ast.Location{ - Line: int(550), + Line: int(535), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5866, + Ctx: p5893, FreeVars: ast.Identifiers{ "n_", "std", @@ -77602,19 +73419,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(550), + Line: int(535), Column: int(21), }, End: ast.Location{ - Line: int(550), + Line: int(535), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5866, + Ctx: p5893, FreeVars: ast.Identifiers{ "std", }, @@ -77622,13 +73439,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(550), + Line: int(535), Column: int(21), }, End: ast.Location{ - Line: int(550), + Line: int(535), Column: int(24), }, File: p1, @@ -77663,9 +73480,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -77675,19 +73491,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(550), + Line: int(535), Column: int(31), }, End: ast.Location{ - Line: int(550), + Line: int(535), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5875, + Ctx: p5902, FreeVars: ast.Identifiers{ "n_", }, @@ -77706,30 +73522,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(550), - Column: int(13), - }, - End: ast.Location{ - Line: int(550), - Column: int(34), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(7), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -77742,15 +73546,15 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "n_", "n__", - "plus", "prec", "render_int", + "sign", "std", "strip_trailing_zero", "trailing", @@ -77766,19 +73570,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(24), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: ast.Identifiers{ "ensure_pt", "prec", @@ -77787,19 +73591,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(27), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: ast.Identifiers{ "ensure_pt", "prec", @@ -77808,19 +73612,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(27), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: ast.Identifiers{ "prec", }, @@ -77828,19 +73632,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(27), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: ast.Identifiers{ "prec", }, @@ -77852,21 +73656,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(35), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -77875,19 +73680,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(40), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: ast.Identifiers{ "ensure_pt", }, @@ -77896,19 +73701,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(41), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: ast.Identifiers{ "ensure_pt", }, @@ -77921,71 +73726,61 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(56), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(551), + Line: int(536), Column: int(63), }, End: ast.Location{ - Line: int(551), + Line: int(536), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5882, + Ctx: p5909, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(551), - Column: int(13), - }, - End: ast.Location{ - Line: int(551), - Column: int(64), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(552), + Line: int(537), Column: int(7), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -77998,16 +73793,16 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "blank", "dot_size", "ensure_pt", "n_", "n__", - "plus", "prec", "render_int", + "sign", "std", "strip_trailing_zero", "trailing", @@ -78023,19 +73818,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(552), + Line: int(537), Column: int(18), }, End: ast.Location{ - Line: int(552), + Line: int(537), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5902, + Ctx: p5929, FreeVars: ast.Identifiers{ "dot_size", "prec", @@ -78045,19 +73840,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(552), + Line: int(537), Column: int(18), }, End: ast.Location{ - Line: int(552), + Line: int(537), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5902, + Ctx: p5929, FreeVars: ast.Identifiers{ "prec", "zero_pad", @@ -78066,19 +73861,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(552), + Line: int(537), Column: int(18), }, End: ast.Location{ - Line: int(552), + Line: int(537), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5902, + Ctx: p5929, FreeVars: ast.Identifiers{ "zero_pad", }, @@ -78090,19 +73885,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(552), + Line: int(537), Column: int(29), }, End: ast.Location{ - Line: int(552), + Line: int(537), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5902, + Ctx: p5929, FreeVars: ast.Identifiers{ "prec", }, @@ -78115,19 +73910,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(552), + Line: int(537), Column: int(36), }, End: ast.Location{ - Line: int(552), + Line: int(537), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5902, + Ctx: p5929, FreeVars: ast.Identifiers{ "dot_size", }, @@ -78137,30 +73932,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(552), - Column: int(13), - }, - End: ast.Location{ - Line: int(552), - Column: int(44), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), + Line: int(538), Column: int(7), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -78173,15 +73956,15 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "n_", "n__", - "plus", "prec", "render_int", + "sign", "std", "strip_trailing_zero", "trailing", @@ -78197,24 +73980,25 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), + Line: int(538), Column: int(19), }, End: ast.Location{ - Line: int(553), - Column: int(73), + Line: int(538), + Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5917, + Ctx: p5944, FreeVars: ast.Identifiers{ "blank", "n__", - "plus", "render_int", + "sign", + "std", "whole", "zp", }, @@ -78222,19 +74006,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), + Line: int(538), Column: int(19), }, End: ast.Location{ - Line: int(553), + Line: int(538), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5917, + Ctx: p5944, FreeVars: ast.Identifiers{ "render_int", }, @@ -78248,92 +74032,174 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), + Line: int(538), Column: int(30), }, End: ast.Location{ - Line: int(553), - Column: int(37), + Line: int(538), + Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: ast.Identifiers{ "n__", + "std", + "whole", }, }, - Left: &ast.Var{ + Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), + Line: int(538), Column: int(30), }, End: ast.Location{ - Line: int(553), - Column: int(33), + Line: int(538), + Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: ast.Identifiers{ "n__", + "std", }, }, - Id: "n__", + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(538), + Column: int(30), + }, + End: ast.Location{ + Line: int(538), + Column: int(38), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5950, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(538), + Column: int(30), + }, + End: ast.Location{ + Line: int(538), + Column: int(33), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "sign", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, + }, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(538), + Column: int(39), + }, + End: ast.Location{ + Line: int(538), + Column: int(42), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p5961, + FreeVars: ast.Identifiers{ + "n__", + }, + }, + Id: "n__", + }, + CommaFodder: nil, + }, + }, + Named: nil, + }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(9), - Right: &ast.LiteralNumber{ + Op: ast.BinaryOp(0), + Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), - Column: int(36), + Line: int(538), + Column: int(46), }, End: ast.Location{ - Line: int(553), - Column: int(37), + Line: int(538), + Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, - FreeVars: nil, - }, - OriginalString: "0", - }, - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(553), - Column: int(39), - }, - End: ast.Location{ - Line: int(553), - Column: int(44), + Ctx: p5950, + FreeVars: ast.Identifiers{ + "whole", }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p5923, - FreeVars: ast.Identifiers{ - "whole", }, + Id: "whole", }, - Id: "whole", }, CommaFodder: ast.Fodder{}, }, @@ -78341,19 +74207,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), - Column: int(46), + Line: int(538), + Column: int(53), }, End: ast.Location{ - Line: int(553), - Column: int(48), + Line: int(538), + Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: ast.Identifiers{ "zp", }, @@ -78366,21 +74232,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), - Column: int(50), + Line: int(538), + Column: int(57), }, End: ast.Location{ - Line: int(553), - Column: int(51), + Line: int(538), + Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -78389,19 +74256,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), - Column: int(53), + Line: int(538), + Column: int(60), }, End: ast.Location{ - Line: int(553), - Column: int(58), + Line: int(538), + Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: ast.Identifiers{ "blank", }, @@ -78414,24 +74281,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), - Column: int(60), + Line: int(538), + Column: int(67), }, End: ast.Location{ - Line: int(553), - Column: int(64), + Line: int(538), + Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: ast.Identifiers{ - "plus", + "sign", }, }, - Id: "plus", + Id: "sign", }, CommaFodder: ast.Fodder{}, }, @@ -78439,21 +74306,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), - Column: int(66), + Line: int(538), + Column: int(73), }, End: ast.Location{ - Line: int(553), - Column: int(68), + Line: int(538), + Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: ast.Fodder{}, @@ -78462,25 +74330,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(553), - Column: int(70), + Line: int(538), + Column: int(77), }, End: ast.Location{ - Line: int(553), - Column: int(72), + Line: int(538), + Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5923, + Ctx: p5950, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -78494,30 +74361,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(553), - Column: int(13), - }, - End: ast.Location{ - Line: int(553), - Column: int(73), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(554), + Line: int(539), Column: int(7), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -78530,7 +74385,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "ensure_pt", "n_", @@ -78546,19 +74401,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(554), + Line: int(539), Column: int(10), }, End: ast.Location{ - Line: int(554), + Line: int(539), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "prec", }, @@ -78566,19 +74421,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(554), + Line: int(539), Column: int(10), }, End: ast.Location{ - Line: int(554), + Line: int(539), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "prec", }, @@ -78590,21 +74445,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(554), + Line: int(539), Column: int(18), }, End: ast.Location{ - Line: int(554), + Line: int(539), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -78612,19 +74468,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(555), + Line: int(540), Column: int(9), }, End: ast.Location{ - Line: int(555), + Line: int(540), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "ensure_pt", "str", @@ -78633,13 +74489,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(555), + Line: int(540), Column: int(9), }, End: ast.Location{ - Line: int(555), + Line: int(540), Column: int(12), }, File: p1, @@ -78652,7 +74508,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "str", }, @@ -78664,19 +74520,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(555), + Line: int(540), Column: int(15), }, End: ast.Location{ - Line: int(555), + Line: int(540), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "ensure_pt", }, @@ -78684,19 +74540,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(555), + Line: int(540), Column: int(18), }, End: ast.Location{ - Line: int(555), + Line: int(540), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "ensure_pt", }, @@ -78707,49 +74563,47 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(555), + Line: int(540), Column: int(33), }, End: ast.Location{ - Line: int(555), + Line: int(540), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: nil, }, Value: ".", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(555), + Line: int(540), Column: int(42), }, End: ast.Location{ - Line: int(555), + Line: int(540), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, @@ -78764,13 +74618,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(9), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -78783,7 +74637,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "n_", "prec", @@ -78803,19 +74657,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(22), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5964, + Ctx: p5999, FreeVars: ast.Identifiers{ "n_", "prec", @@ -78826,19 +74680,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(22), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5964, + Ctx: p5999, FreeVars: ast.Identifiers{ "std", }, @@ -78846,13 +74700,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(22), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(25), }, File: p1, @@ -78887,9 +74741,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -78899,19 +74752,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(32), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: ast.Identifiers{ "n_", "prec", @@ -78922,19 +74775,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(32), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: ast.Identifiers{ "n_", "prec", @@ -78945,19 +74798,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(33), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: ast.Identifiers{ "n_", "whole", @@ -78966,19 +74819,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(33), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: ast.Identifiers{ "n_", }, @@ -78990,19 +74843,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(38), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: ast.Identifiers{ "whole", }, @@ -79015,19 +74868,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(47), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: ast.Identifiers{ "prec", "std", @@ -79036,19 +74889,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(47), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: ast.Identifiers{ "std", }, @@ -79056,13 +74909,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(47), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(50), }, File: p1, @@ -79097,9 +74950,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "pow", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -79109,21 +74961,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(55), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5992, + Ctx: p6027, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: ast.Fodder{}, @@ -79132,19 +74985,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(59), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5992, + Ctx: p6027, FreeVars: ast.Identifiers{ "prec", }, @@ -79167,21 +75020,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(557), + Line: int(542), Column: int(67), }, End: ast.Location{ - Line: int(557), + Line: int(542), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5973, + Ctx: p6008, FreeVars: nil, }, + Value: float64(0.5), OriginalString: "0.5", }, }, @@ -79197,30 +75051,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(557), - Column: int(15), - }, - End: ast.Location{ - Line: int(557), - Column: int(71), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(558), + Line: int(543), Column: int(9), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -79233,7 +75075,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac", "prec", @@ -79246,19 +75088,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(558), + Line: int(543), Column: int(12), }, End: ast.Location{ - Line: int(558), + Line: int(543), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac", "trailing", @@ -79267,19 +75109,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(558), + Line: int(543), Column: int(12), }, End: ast.Location{ - Line: int(558), + Line: int(543), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "trailing", }, @@ -79291,19 +75133,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(558), + Line: int(543), Column: int(24), }, End: ast.Location{ - Line: int(558), + Line: int(543), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac", }, @@ -79311,19 +75153,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(558), + Line: int(543), Column: int(24), }, End: ast.Location{ - Line: int(558), + Line: int(543), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac", }, @@ -79335,21 +75177,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(558), + Line: int(543), Column: int(31), }, End: ast.Location{ - Line: int(558), + Line: int(543), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -79358,13 +75201,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), + Line: int(544), Column: int(11), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(84), }, File: p1, @@ -79377,7 +75220,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac", "prec", @@ -79395,19 +75238,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), + Line: int(544), Column: int(28), }, End: ast.Location{ - Line: int(559), - Column: int(82), + Line: int(544), + Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6013, + Ctx: p6048, FreeVars: ast.Identifiers{ "frac", "prec", @@ -79417,19 +75260,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), + Line: int(544), Column: int(28), }, End: ast.Location{ - Line: int(559), + Line: int(544), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6013, + Ctx: p6048, FreeVars: ast.Identifiers{ "render_int", }, @@ -79440,45 +75283,22 @@ var _StdAst = &ast.DesugaredObject{ Arguments: ast.Arguments{ Positional: []ast.CommaSeparatedExpr{ ast.CommaSeparatedExpr{ - Expr: &ast.LiteralBoolean{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(559), - Column: int(39), - }, - End: ast.Location{ - Line: int(559), - Column: int(44), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6019, - FreeVars: nil, - }, - Value: false, - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), - Column: int(46), + Line: int(544), + Column: int(39), }, End: ast.Location{ - Line: int(559), - Column: int(50), + Line: int(544), + Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6019, + Ctx: p6054, FreeVars: ast.Identifiers{ "frac", }, @@ -79491,19 +75311,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), - Column: int(52), + Line: int(544), + Column: int(45), }, End: ast.Location{ - Line: int(559), - Column: int(56), + Line: int(544), + Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6019, + Ctx: p6054, FreeVars: ast.Identifiers{ "prec", }, @@ -79516,21 +75336,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), - Column: int(58), + Line: int(544), + Column: int(51), }, End: ast.Location{ - Line: int(559), - Column: int(59), + Line: int(544), + Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6019, + Ctx: p6054, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -79539,19 +75360,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), - Column: int(61), + Line: int(544), + Column: int(54), }, End: ast.Location{ - Line: int(559), - Column: int(66), + Line: int(544), + Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6019, + Ctx: p6054, FreeVars: nil, }, Value: false, @@ -79562,19 +75383,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), - Column: int(68), + Line: int(544), + Column: int(61), }, End: ast.Location{ - Line: int(559), - Column: int(73), + Line: int(544), + Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6019, + Ctx: p6054, FreeVars: nil, }, Value: false, @@ -79585,21 +75406,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), - Column: int(75), + Line: int(544), + Column: int(68), }, End: ast.Location{ - Line: int(559), - Column: int(77), + Line: int(544), + Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6019, + Ctx: p6054, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: ast.Fodder{}, @@ -79608,25 +75430,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(559), - Column: int(79), + Line: int(544), + Column: int(72), }, End: ast.Location{ - Line: int(559), - Column: int(81), + Line: int(544), + Column: int(74), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6019, + Ctx: p6054, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -79640,36 +75461,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(559), - Column: int(17), - }, - End: ast.Location{ - Line: int(559), - Column: int(82), - }, - File: p1, - }, }, }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(11), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac_str", "str", @@ -79680,19 +75489,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(11), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "str", }, @@ -79700,13 +75509,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(11), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(14), }, File: p1, @@ -79719,7 +75528,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "str", }, @@ -79731,25 +75540,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(17), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: nil, }, Value: ".", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -79757,19 +75565,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(23), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac_str", "strip_trailing_zero", @@ -79779,19 +75587,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(26), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "trailing", }, @@ -79800,19 +75608,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(27), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "trailing", }, @@ -79824,19 +75632,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(41), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac_str", "strip_trailing_zero", @@ -79845,19 +75653,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(41), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "strip_trailing_zero", }, @@ -79871,19 +75679,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(61), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6049, + Ctx: p6083, FreeVars: ast.Identifiers{ "frac_str", }, @@ -79904,19 +75712,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(560), + Line: int(545), Column: int(76), }, End: ast.Location{ - Line: int(560), + Line: int(545), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "frac_str", }, @@ -79937,13 +75745,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(562), + Line: int(547), Column: int(11), }, End: ast.Location{ - Line: int(562), + Line: int(547), Column: int(14), }, File: p1, @@ -79956,7 +75764,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p5846, + Ctx: p5873, FreeVars: ast.Identifiers{ "str", }, @@ -79974,30 +75782,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(565), + Line: int(550), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -80010,7 +75806,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -80038,19 +75834,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(565), + Line: int(550), Column: int(11), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(85), }, File: p1, }, Fodder: nil, - Ctx: p6063, + Ctx: p6097, FreeVars: ast.Identifiers{ "render_float_dec", "render_int", @@ -80058,172 +75854,63 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "n__", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(28), - }, - End: ast.Location{ - Line: int(565), - Column: int(31), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n__", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "zero_pad", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(33), - }, - End: ast.Location{ - Line: int(565), - Column: int(41), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "zero_pad", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "blank", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(43), - }, - End: ast.Location{ - Line: int(565), - Column: int(48), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "blank", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "plus", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(50), - }, - End: ast.Location{ - Line: int(565), - Column: int(54), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "sign", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "ensure_pt", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(56), - }, - End: ast.Location{ - Line: int(565), - Column: int(65), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "ensure_pt", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "trailing", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(67), - }, - End: ast.Location{ - Line: int(565), - Column: int(75), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "trailing", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "caps", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(77), - }, - End: ast.Location{ - Line: int(565), - Column: int(81), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "caps", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "prec", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(565), - Column: int(83), - }, - End: ast.Location{ - Line: int(565), - Column: int(87), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "prec", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(7), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(85), }, File: p1, @@ -80236,16 +75923,16 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "blank", "caps", "ensure_pt", "n__", - "plus", "prec", "render_float_dec", "render_int", + "sign", "std", "trailing", "zero_pad", @@ -80259,19 +75946,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(24), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(94), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6072, + Ctx: p6106, FreeVars: ast.Identifiers{ "n__", "std", @@ -80280,19 +75967,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(27), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6072, + Ctx: p6106, FreeVars: ast.Identifiers{ "n__", }, @@ -80300,19 +75987,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(27), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6072, + Ctx: p6106, FreeVars: ast.Identifiers{ "n__", }, @@ -80324,21 +76011,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(34), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6072, + Ctx: p6106, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -80346,40 +76034,41 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(41), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6072, + Ctx: p6106, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(48), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(94), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6072, + Ctx: p6106, FreeVars: ast.Identifiers{ "n__", "std", @@ -80388,19 +76077,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(48), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6072, + Ctx: p6106, FreeVars: ast.Identifiers{ "std", }, @@ -80408,13 +76097,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(48), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(51), }, File: p1, @@ -80449,9 +76138,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -80461,19 +76149,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(58), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(93), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6089, + Ctx: p6123, FreeVars: ast.Identifiers{ "n__", "std", @@ -80482,19 +76170,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(58), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6089, + Ctx: p6123, FreeVars: ast.Identifiers{ "n__", "std", @@ -80503,19 +76191,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(58), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6089, + Ctx: p6123, FreeVars: ast.Identifiers{ "std", }, @@ -80523,13 +76211,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(58), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(61), }, File: p1, @@ -80564,9 +76252,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "log", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -80576,19 +76263,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(66), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6100, + Ctx: p6134, FreeVars: ast.Identifiers{ "n__", "std", @@ -80597,19 +76284,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(66), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6100, + Ctx: p6134, FreeVars: ast.Identifiers{ "std", }, @@ -80617,13 +76304,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(66), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(69), }, File: p1, @@ -80658,9 +76345,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "abs", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -80670,19 +76356,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(74), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6109, + Ctx: p6143, FreeVars: ast.Identifiers{ "n__", }, @@ -80714,19 +76400,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(82), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(93), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6089, + Ctx: p6123, FreeVars: ast.Identifiers{ "std", }, @@ -80734,19 +76420,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(82), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(89), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6089, + Ctx: p6123, FreeVars: ast.Identifiers{ "std", }, @@ -80754,13 +76440,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(82), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(85), }, File: p1, @@ -80795,9 +76481,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "log", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -80807,21 +76492,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(566), + Line: int(551), Column: int(90), }, End: ast.Location{ - Line: int(566), + Line: int(551), Column: int(92), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6120, + Ctx: p6154, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: nil, @@ -80848,30 +76534,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(566), - Column: int(13), - }, - End: ast.Location{ - Line: int(566), - Column: int(94), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(567), + Line: int(552), Column: int(7), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(85), }, File: p1, @@ -80884,17 +76558,17 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "blank", "caps", "ensure_pt", "exponent", "n__", - "plus", "prec", "render_float_dec", "render_int", + "sign", "std", "trailing", "zero_pad", @@ -80908,42 +76582,41 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(567), + Line: int(552), Column: int(20), }, End: ast.Location{ - Line: int(568), - Column: int(92), + Line: int(553), + Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6126, + Ctx: p6160, FreeVars: ast.Identifiers{ "caps", "exponent", "render_int", - "std", }, }, Left: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(567), + Line: int(552), Column: int(21), }, End: ast.Location{ - Line: int(567), + Line: int(552), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6126, + Ctx: p6160, FreeVars: ast.Identifiers{ "caps", }, @@ -80951,19 +76624,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(567), + Line: int(552), Column: int(24), }, End: ast.Location{ - Line: int(567), + Line: int(552), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6126, + Ctx: p6160, FreeVars: ast.Identifiers{ "caps", }, @@ -80974,49 +76647,47 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(567), + Line: int(552), Column: int(34), }, End: ast.Location{ - Line: int(567), + Line: int(552), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6126, + Ctx: p6160, FreeVars: nil, }, Value: "E", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(567), + Line: int(552), Column: int(43), }, End: ast.Location{ - Line: int(567), + Line: int(552), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6126, + Ctx: p6160, FreeVars: nil, }, Value: "e", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{ @@ -81031,41 +76702,40 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), + Line: int(553), Column: int(22), }, End: ast.Location{ - Line: int(568), - Column: int(92), + Line: int(553), + Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6126, + Ctx: p6160, FreeVars: ast.Identifiers{ "exponent", "render_int", - "std", }, }, Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), + Line: int(553), Column: int(22), }, End: ast.Location{ - Line: int(568), + Line: int(553), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6126, + Ctx: p6160, FreeVars: ast.Identifiers{ "render_int", }, @@ -81076,199 +76746,27 @@ var _StdAst = &ast.DesugaredObject{ Arguments: ast.Arguments{ Positional: []ast.CommaSeparatedExpr{ ast.CommaSeparatedExpr{ - Expr: &ast.Binary{ + Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), + Line: int(553), Column: int(33), }, End: ast.Location{ - Line: int(568), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6141, - FreeVars: ast.Identifiers{ - "exponent", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(568), - Column: int(33), - }, - End: ast.Location{ - Line: int(568), - Column: int(41), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6141, - FreeVars: ast.Identifiers{ - "exponent", - }, - }, - Id: "exponent", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(9), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(568), - Column: int(44), - }, - End: ast.Location{ - Line: int(568), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6141, - FreeVars: nil, - }, - OriginalString: "0", - }, - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(568), - Column: int(47), - }, - End: ast.Location{ - Line: int(568), - Column: int(64), + Line: int(553), + Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6141, + Ctx: p6175, FreeVars: ast.Identifiers{ "exponent", - "std", }, }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(568), - Column: int(47), - }, - End: ast.Location{ - Line: int(568), - Column: int(54), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6141, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(568), - Column: int(47), - }, - End: ast.Location{ - Line: int(568), - Column: int(50), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "abs", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(568), - Column: int(55), - }, - End: ast.Location{ - Line: int(568), - Column: int(63), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6155, - FreeVars: ast.Identifiers{ - "exponent", - }, - }, - Id: "exponent", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, + Id: "exponent", }, CommaFodder: ast.Fodder{}, }, @@ -81276,21 +76774,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), - Column: int(66), + Line: int(553), + Column: int(43), }, End: ast.Location{ - Line: int(568), - Column: int(67), + Line: int(553), + Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6141, + Ctx: p6175, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, CommaFodder: ast.Fodder{}, @@ -81299,21 +76798,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), - Column: int(69), + Line: int(553), + Column: int(46), }, End: ast.Location{ - Line: int(568), - Column: int(70), + Line: int(553), + Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6141, + Ctx: p6175, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -81322,19 +76822,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), - Column: int(72), + Line: int(553), + Column: int(49), }, End: ast.Location{ - Line: int(568), - Column: int(77), + Line: int(553), + Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6141, + Ctx: p6175, FreeVars: nil, }, Value: false, @@ -81345,19 +76845,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), - Column: int(79), + Line: int(553), + Column: int(56), }, End: ast.Location{ - Line: int(568), - Column: int(83), + Line: int(553), + Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6141, + Ctx: p6175, FreeVars: nil, }, Value: true, @@ -81368,21 +76868,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), - Column: int(85), + Line: int(553), + Column: int(62), }, End: ast.Location{ - Line: int(568), - Column: int(87), + Line: int(553), + Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6141, + Ctx: p6175, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: ast.Fodder{}, @@ -81391,25 +76892,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(568), - Column: int(89), + Line: int(553), + Column: int(66), }, End: ast.Location{ - Line: int(568), - Column: int(91), + Line: int(553), + Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6141, + Ctx: p6175, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -81424,30 +76924,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(567), - Column: int(13), - }, - End: ast.Location{ - Line: int(568), - Column: int(92), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(569), + Line: int(554), Column: int(7), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(85), }, File: p1, @@ -81460,15 +76948,15 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "exponent", "n__", - "plus", "prec", "render_float_dec", + "sign", "std", "suff", "trailing", @@ -81483,19 +76971,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(569), + Line: int(554), Column: int(24), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "exponent", "n__", @@ -81505,19 +76993,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(569), + Line: int(554), Column: int(27), }, End: ast.Location{ - Line: int(569), + Line: int(554), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "exponent", }, @@ -81525,19 +77013,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(569), + Line: int(554), Column: int(27), }, End: ast.Location{ - Line: int(569), + Line: int(554), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "exponent", }, @@ -81549,40 +77037,41 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(569), + Line: int(554), Column: int(39), }, End: ast.Location{ - Line: int(569), + Line: int(554), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: nil, }, Op: ast.UnaryOp(3), Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(569), + Line: int(554), Column: int(40), }, End: ast.Location{ - Line: int(569), + Line: int(554), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: nil, }, + Value: float64(324), OriginalString: "324", }, }, @@ -81591,19 +77080,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(9), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "exponent", "n__", @@ -81613,19 +77102,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(9), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "n__", }, @@ -81633,13 +77122,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(9), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(12), }, File: p1, @@ -81652,7 +77141,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(8), Comment: []string{ @@ -81660,7 +77149,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(8), Comment: []string{ @@ -81668,7 +77157,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "n__", }, @@ -81680,21 +77169,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(15), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -81703,19 +77193,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(20), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "exponent", "std", @@ -81724,19 +77214,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(20), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "std", }, @@ -81744,13 +77234,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(20), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(23), }, File: p1, @@ -81785,9 +77275,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "pow", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -81797,21 +77286,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(28), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6195, + Ctx: p6215, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: ast.Fodder{}, @@ -81820,19 +77310,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(32), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6195, + Ctx: p6215, FreeVars: ast.Identifiers{ "exponent", }, @@ -81840,19 +77330,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(32), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6195, + Ctx: p6215, FreeVars: ast.Identifiers{ "exponent", }, @@ -81864,21 +77354,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(572), + Line: int(557), Column: int(43), }, End: ast.Location{ - Line: int(572), + Line: int(557), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6195, + Ctx: p6215, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -81904,19 +77395,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(574), + Line: int(559), Column: int(9), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "exponent", "n__", @@ -81926,13 +77417,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(574), + Line: int(559), Column: int(9), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(12), }, File: p1, @@ -81945,7 +77436,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "n__", }, @@ -81957,19 +77448,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(574), + Line: int(559), Column: int(15), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "exponent", "std", @@ -81978,19 +77469,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(574), + Line: int(559), Column: int(15), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6168, + Ctx: p6188, FreeVars: ast.Identifiers{ "std", }, @@ -81998,13 +77489,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(574), + Line: int(559), Column: int(15), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(18), }, File: p1, @@ -82039,9 +77530,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "pow", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -82051,21 +77541,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(574), + Line: int(559), Column: int(23), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6216, + Ctx: p6236, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: ast.Fodder{}, @@ -82074,19 +77565,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(574), + Line: int(559), Column: int(27), }, End: ast.Location{ - Line: int(574), + Line: int(559), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6216, + Ctx: p6236, FreeVars: ast.Identifiers{ "exponent", }, @@ -82107,30 +77598,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(569), - Column: int(13), - }, - End: ast.Location{ - Line: int(574), - Column: int(36), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(575), + Line: int(560), Column: int(7), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(85), }, File: p1, @@ -82143,14 +77622,14 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "mantissa", - "plus", "prec", "render_float_dec", + "sign", "std", "suff", "trailing", @@ -82165,19 +77644,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(575), + Line: int(560), Column: int(19), }, End: ast.Location{ - Line: int(575), + Line: int(560), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6224, + Ctx: p6244, FreeVars: ast.Identifiers{ "std", "suff", @@ -82187,19 +77666,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(575), + Line: int(560), Column: int(19), }, End: ast.Location{ - Line: int(575), + Line: int(560), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6224, + Ctx: p6244, FreeVars: ast.Identifiers{ "zero_pad", }, @@ -82211,19 +77690,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(575), + Line: int(560), Column: int(30), }, End: ast.Location{ - Line: int(575), + Line: int(560), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6224, + Ctx: p6244, FreeVars: ast.Identifiers{ "std", "suff", @@ -82232,19 +77711,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(575), + Line: int(560), Column: int(30), }, End: ast.Location{ - Line: int(575), + Line: int(560), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6224, + Ctx: p6244, FreeVars: ast.Identifiers{ "std", }, @@ -82252,13 +77731,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(575), + Line: int(560), Column: int(30), }, End: ast.Location{ - Line: int(575), + Line: int(560), Column: int(33), }, File: p1, @@ -82293,9 +77772,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -82305,19 +77783,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(575), + Line: int(560), Column: int(41), }, End: ast.Location{ - Line: int(575), + Line: int(560), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6237, + Ctx: p6257, FreeVars: ast.Identifiers{ "suff", }, @@ -82337,43 +77815,31 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(575), - Column: int(13), - }, - End: ast.Location{ - Line: int(575), - Column: int(46), - }, - File: p1, - }, }, }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(7), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "mantissa", - "plus", "prec", "render_float_dec", + "sign", "suff", "trailing", "zp2", @@ -82382,26 +77848,26 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(7), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "blank", "ensure_pt", "mantissa", - "plus", "prec", "render_float_dec", + "sign", "trailing", "zp2", }, @@ -82409,13 +77875,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(7), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(23), }, File: p1, @@ -82428,7 +77894,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "render_float_dec", }, @@ -82442,19 +77908,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(24), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6248, + Ctx: p6268, FreeVars: ast.Identifiers{ "mantissa", }, @@ -82467,19 +77933,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(34), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6248, + Ctx: p6268, FreeVars: ast.Identifiers{ "zp2", }, @@ -82492,19 +77958,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(39), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6248, + Ctx: p6268, FreeVars: ast.Identifiers{ "blank", }, @@ -82517,24 +77983,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(46), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6248, + Ctx: p6268, FreeVars: ast.Identifiers{ - "plus", + "sign", }, }, - Id: "plus", + Id: "sign", }, CommaFodder: ast.Fodder{}, }, @@ -82542,19 +78008,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(52), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6248, + Ctx: p6268, FreeVars: ast.Identifiers{ "ensure_pt", }, @@ -82567,19 +78033,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(63), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6248, + Ctx: p6268, FreeVars: ast.Identifiers{ "trailing", }, @@ -82592,19 +78058,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(73), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6248, + Ctx: p6268, FreeVars: ast.Identifiers{ "prec", }, @@ -82626,19 +78092,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(576), + Line: int(561), Column: int(81), }, End: ast.Location{ - Line: int(576), + Line: int(561), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6068, + Ctx: p6102, FreeVars: ast.Identifiers{ "suff", }, @@ -82653,30 +78119,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(579), + Line: int(564), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -82689,7 +78143,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -82718,19 +78172,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(579), + Line: int(564), Column: int(11), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: nil, - Ctx: p6270, + Ctx: p6290, FreeVars: ast.Identifiers{ "render_float_dec", "render_float_sci", @@ -82740,115 +78194,48 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "val", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(579), - Column: int(23), - }, - End: ast.Location{ - Line: int(579), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "val", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "code", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(579), - Column: int(28), - }, - End: ast.Location{ - Line: int(579), - Column: int(32), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "code", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "fw", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(579), - Column: int(34), - }, - End: ast.Location{ - Line: int(579), - Column: int(36), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "fw", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "prec_or_null", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(579), - Column: int(38), - }, - End: ast.Location{ - Line: int(579), - Column: int(50), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "prec_or_null", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(579), - Column: int(52), - }, - End: ast.Location{ - Line: int(579), - Column: int(53), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(580), + Line: int(565), Column: int(7), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, @@ -82861,7 +78248,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", "fw", @@ -82883,19 +78270,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(580), + Line: int(565), Column: int(22), }, End: ast.Location{ - Line: int(580), + Line: int(565), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6279, + Ctx: p6299, FreeVars: ast.Identifiers{ "code", }, @@ -82903,13 +78290,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(580), + Line: int(565), Column: int(22), }, End: ast.Location{ - Line: int(580), + Line: int(565), Column: int(26), }, File: p1, @@ -82944,37 +78331,24 @@ var _StdAst = &ast.DesugaredObject{ Value: "cflags", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(580), - Column: int(13), - }, - End: ast.Location{ - Line: int(580), - Column: int(33), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(581), + Line: int(566), Column: int(7), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, @@ -82987,7 +78361,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -83010,19 +78384,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(581), + Line: int(566), Column: int(22), }, End: ast.Location{ - Line: int(581), + Line: int(566), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6289, + Ctx: p6309, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83030,19 +78404,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(581), + Line: int(566), Column: int(25), }, End: ast.Location{ - Line: int(581), + Line: int(566), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6289, + Ctx: p6309, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83050,19 +78424,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(581), + Line: int(566), Column: int(25), }, End: ast.Location{ - Line: int(581), + Line: int(566), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6289, + Ctx: p6309, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83074,19 +78448,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(581), + Line: int(566), Column: int(41), }, End: ast.Location{ - Line: int(581), + Line: int(566), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6289, + Ctx: p6309, FreeVars: nil, }, }, @@ -83095,19 +78469,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(581), + Line: int(566), Column: int(51), }, End: ast.Location{ - Line: int(581), + Line: int(566), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6289, + Ctx: p6309, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83118,50 +78492,39 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(581), + Line: int(566), Column: int(69), }, End: ast.Location{ - Line: int(581), + Line: int(566), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6289, + Ctx: p6309, FreeVars: nil, }, + Value: float64(6), OriginalString: "6", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(581), - Column: int(13), - }, - End: ast.Location{ - Line: int(581), - Column: int(70), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(582), + Line: int(567), Column: int(7), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, @@ -83174,7 +78537,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -83198,19 +78561,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(582), + Line: int(567), Column: int(21), }, End: ast.Location{ - Line: int(582), + Line: int(567), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6304, + Ctx: p6324, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83218,19 +78581,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(582), + Line: int(567), Column: int(24), }, End: ast.Location{ - Line: int(582), + Line: int(567), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6304, + Ctx: p6324, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83238,19 +78601,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(582), + Line: int(567), Column: int(24), }, End: ast.Location{ - Line: int(582), + Line: int(567), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6304, + Ctx: p6324, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83262,19 +78625,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(582), + Line: int(567), Column: int(40), }, End: ast.Location{ - Line: int(582), + Line: int(567), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6304, + Ctx: p6324, FreeVars: nil, }, }, @@ -83283,19 +78646,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(582), + Line: int(567), Column: int(50), }, End: ast.Location{ - Line: int(582), + Line: int(567), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6304, + Ctx: p6324, FreeVars: ast.Identifiers{ "prec_or_null", }, @@ -83306,50 +78669,39 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(582), + Line: int(567), Column: int(68), }, End: ast.Location{ - Line: int(582), + Line: int(567), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6304, + Ctx: p6324, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(582), - Column: int(13), - }, - End: ast.Location{ - Line: int(582), - Column: int(69), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(7), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, @@ -83362,7 +78714,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -83386,19 +78738,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(18), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6319, + Ctx: p6339, FreeVars: ast.Identifiers{ "cflags", "fw", @@ -83407,19 +78759,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(21), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6319, + Ctx: p6339, FreeVars: ast.Identifiers{ "cflags", }, @@ -83427,19 +78779,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(21), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6319, + Ctx: p6339, FreeVars: ast.Identifiers{ "cflags", }, @@ -83447,13 +78799,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(21), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(27), }, File: p1, @@ -83488,9 +78840,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "zero", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -83498,19 +78849,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(36), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6319, + Ctx: p6339, FreeVars: ast.Identifiers{ "cflags", }, @@ -83519,19 +78870,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(37), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6319, + Ctx: p6339, FreeVars: ast.Identifiers{ "cflags", }, @@ -83539,13 +78890,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(37), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(43), }, File: p1, @@ -83580,9 +78931,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "left", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -83591,19 +78941,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(54), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6319, + Ctx: p6339, FreeVars: ast.Identifiers{ "fw", }, @@ -83614,50 +78964,39 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(583), + Line: int(568), Column: int(62), }, End: ast.Location{ - Line: int(583), + Line: int(568), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6319, + Ctx: p6339, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(583), - Column: int(13), - }, - End: ast.Location{ - Line: int(583), - Column: int(63), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(584), + Line: int(569), Column: int(7), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, @@ -83670,7 +79009,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -83689,19 +79028,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(584), + Line: int(569), Column: int(10), }, End: ast.Location{ - Line: int(584), + Line: int(569), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -83709,19 +79048,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(584), + Line: int(569), Column: int(10), }, End: ast.Location{ - Line: int(584), + Line: int(569), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -83729,13 +79068,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(584), + Line: int(569), Column: int(10), }, End: ast.Location{ - Line: int(584), + Line: int(569), Column: int(14), }, File: p1, @@ -83770,9 +79109,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -83780,44 +79118,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(584), + Line: int(569), Column: int(24), }, End: ast.Location{ - Line: int(584), + Line: int(569), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(585), + Line: int(570), Column: int(9), }, End: ast.Location{ - Line: int(585), + Line: int(570), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -83826,19 +79163,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(585), + Line: int(570), Column: int(9), }, End: ast.Location{ - Line: int(585), + Line: int(570), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -83846,13 +79183,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(585), + Line: int(570), Column: int(9), }, End: ast.Location{ - Line: int(585), + Line: int(570), Column: int(12), }, File: p1, @@ -83894,9 +79231,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "toString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -83906,19 +79242,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(585), + Line: int(570), Column: int(22), }, End: ast.Location{ - Line: int(585), + Line: int(570), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6359, + Ctx: p6379, FreeVars: ast.Identifiers{ "val", }, @@ -83946,19 +79282,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(586), + Line: int(571), Column: int(12), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -83977,19 +79313,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(586), + Line: int(571), Column: int(15), }, End: ast.Location{ - Line: int(586), + Line: int(571), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -83997,19 +79333,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(586), + Line: int(571), Column: int(15), }, End: ast.Location{ - Line: int(586), + Line: int(571), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -84017,13 +79353,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(586), + Line: int(571), Column: int(15), }, End: ast.Location{ - Line: int(586), + Line: int(571), Column: int(19), }, File: p1, @@ -84058,9 +79394,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -84068,39 +79403,38 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(586), + Line: int(571), Column: int(29), }, End: ast.Location{ - Line: int(586), + Line: int(571), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "d", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(587), + Line: int(572), Column: int(9), }, End: ast.Location{ - Line: int(591), - Column: int(103), + Line: int(576), + Column: int(72), }, File: p1, }, @@ -84112,7 +79446,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "i", @@ -84126,19 +79460,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(587), + Line: int(572), Column: int(12), }, End: ast.Location{ - Line: int(587), + Line: int(572), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -84147,19 +79481,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(587), + Line: int(572), Column: int(12), }, End: ast.Location{ - Line: int(587), + Line: int(572), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -84168,19 +79502,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(587), + Line: int(572), Column: int(12), }, End: ast.Location{ - Line: int(587), + Line: int(572), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -84188,13 +79522,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(587), + Line: int(572), Column: int(12), }, End: ast.Location{ - Line: int(587), + Line: int(572), Column: int(15), }, File: p1, @@ -84229,9 +79563,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -84241,19 +79574,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(587), + Line: int(572), Column: int(21), }, End: ast.Location{ - Line: int(587), + Line: int(572), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6386, + Ctx: p6406, FreeVars: ast.Identifiers{ "val", }, @@ -84275,38 +79608,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(587), + Line: int(572), Column: int(29), }, End: ast.Location{ - Line: int(587), + Line: int(572), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(588), + Line: int(573), Column: int(11), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(47), }, File: p1, @@ -84319,7 +79651,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -84329,19 +79661,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(588), + Line: int(573), Column: int(17), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -84351,19 +79683,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(588), + Line: int(573), Column: int(17), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -84371,19 +79703,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(588), + Line: int(573), Column: int(17), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -84391,25 +79723,24 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(588), + Line: int(573), Column: int(17), }, End: ast.Location{ - Line: int(588), + Line: int(573), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "Format required number at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{ ast.FodderElement{ @@ -84423,19 +79754,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(589), + Line: int(574), Column: int(19), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -84448,25 +79779,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(589), + Line: int(574), Column: int(23), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: ", got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -84474,19 +79804,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(589), + Line: int(574), Column: int(34), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -84495,19 +79825,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(589), + Line: int(574), Column: int(34), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -84515,13 +79845,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(589), + Line: int(574), Column: int(34), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(37), }, File: p1, @@ -84556,9 +79886,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -84568,19 +79897,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(589), + Line: int(574), Column: int(43), }, End: ast.Location{ - Line: int(589), + Line: int(574), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6412, + Ctx: p6432, FreeVars: ast.Identifiers{ "val", }, @@ -84610,24 +79939,23 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), + Line: int(576), Column: int(11), }, End: ast.Location{ - Line: int(591), - Column: int(103), + Line: int(576), + Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "iprec", "render_int", - "std", "val", "zp", }, @@ -84635,13 +79963,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), + Line: int(576), Column: int(11), }, End: ast.Location{ - Line: int(591), + Line: int(576), Column: int(21), }, File: p1, @@ -84654,7 +79982,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "render_int", }, @@ -84665,323 +79993,27 @@ var _StdAst = &ast.DesugaredObject{ Arguments: ast.Arguments{ Positional: []ast.CommaSeparatedExpr{ ast.CommaSeparatedExpr{ - Expr: &ast.Binary{ + Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), + Line: int(576), Column: int(22), }, End: ast.Location{ - Line: int(591), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6422, - FreeVars: ast.Identifiers{ - "val", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(22), - }, - End: ast.Location{ - Line: int(591), - Column: int(25), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6422, - FreeVars: ast.Identifiers{ - "val", - }, - }, - Id: "val", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(10), - Right: &ast.Unary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(29), - }, - End: ast.Location{ - Line: int(591), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6422, - FreeVars: nil, - }, - Op: ast.UnaryOp(3), - Expr: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(30), - }, - End: ast.Location{ - Line: int(591), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6422, - FreeVars: nil, - }, - OriginalString: "1", - }, - }, - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(33), - }, - End: ast.Location{ - Line: int(591), - Column: int(56), + Line: int(576), + Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6422, + Ctx: p6442, FreeVars: ast.Identifiers{ - "std", "val", }, }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(33), - }, - End: ast.Location{ - Line: int(591), - Column: int(42), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6422, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(33), - }, - End: ast.Location{ - Line: int(591), - Column: int(36), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "floor", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(43), - }, - End: ast.Location{ - Line: int(591), - Column: int(55), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6437, - FreeVars: ast.Identifiers{ - "std", - "val", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(43), - }, - End: ast.Location{ - Line: int(591), - Column: int(50), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6437, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(43), - }, - End: ast.Location{ - Line: int(591), - Column: int(46), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "abs", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(591), - Column: int(51), - }, - End: ast.Location{ - Line: int(591), - Column: int(54), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6446, - FreeVars: ast.Identifiers{ - "val", - }, - }, - Id: "val", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, + Id: "val", }, CommaFodder: ast.Fodder{}, }, @@ -84989,19 +80021,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(58), + Line: int(576), + Column: int(27), }, End: ast.Location{ - Line: int(591), - Column: int(60), + Line: int(576), + Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6422, + Ctx: p6442, FreeVars: ast.Identifiers{ "zp", }, @@ -85014,19 +80046,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(62), + Line: int(576), + Column: int(31), }, End: ast.Location{ - Line: int(591), - Column: int(67), + Line: int(576), + Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6422, + Ctx: p6442, FreeVars: ast.Identifiers{ "iprec", }, @@ -85039,19 +80071,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(69), + Line: int(576), + Column: int(38), }, End: ast.Location{ - Line: int(591), - Column: int(81), + Line: int(576), + Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6422, + Ctx: p6442, FreeVars: ast.Identifiers{ "cflags", }, @@ -85059,14 +80091,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(69), + Line: int(576), + Column: int(38), }, End: ast.Location{ - Line: int(591), - Column: int(75), + Line: int(576), + Column: int(44), }, File: p1, }, @@ -85100,9 +80132,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -85111,19 +80142,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(83), + Line: int(576), + Column: int(52), }, End: ast.Location{ - Line: int(591), - Column: int(94), + Line: int(576), + Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6422, + Ctx: p6442, FreeVars: ast.Identifiers{ "cflags", }, @@ -85131,14 +80162,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(83), + Line: int(576), + Column: int(52), }, End: ast.Location{ - Line: int(591), - Column: int(89), + Line: int(576), + Column: int(58), }, File: p1, }, @@ -85169,12 +80200,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -85183,21 +80213,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(96), + Line: int(576), + Column: int(65), }, End: ast.Location{ - Line: int(591), - Column: int(98), + Line: int(576), + Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6422, + Ctx: p6442, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: ast.Fodder{}, @@ -85206,25 +80237,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(591), - Column: int(100), + Line: int(576), + Column: int(69), }, End: ast.Location{ - Line: int(591), - Column: int(102), + Line: int(576), + Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6422, + Ctx: p6442, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -85248,19 +80278,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(592), + Line: int(577), Column: int(12), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -85279,19 +80309,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(592), + Line: int(577), Column: int(15), }, End: ast.Location{ - Line: int(592), + Line: int(577), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -85299,19 +80329,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(592), + Line: int(577), Column: int(15), }, End: ast.Location{ - Line: int(592), + Line: int(577), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -85319,13 +80349,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(592), + Line: int(577), Column: int(15), }, End: ast.Location{ - Line: int(592), + Line: int(577), Column: int(19), }, File: p1, @@ -85360,9 +80390,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -85370,39 +80399,38 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(592), + Line: int(577), Column: int(29), }, End: ast.Location{ - Line: int(592), + Line: int(577), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "o", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(593), + Line: int(578), Column: int(9), }, End: ast.Location{ - Line: int(598), - Column: int(111), + Line: int(583), + Column: int(80), }, File: p1, }, @@ -85414,7 +80442,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "i", @@ -85428,19 +80456,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(593), + Line: int(578), Column: int(12), }, End: ast.Location{ - Line: int(593), + Line: int(578), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -85449,19 +80477,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(593), + Line: int(578), Column: int(12), }, End: ast.Location{ - Line: int(593), + Line: int(578), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -85470,19 +80498,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(593), + Line: int(578), Column: int(12), }, End: ast.Location{ - Line: int(593), + Line: int(578), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -85490,13 +80518,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(593), + Line: int(578), Column: int(12), }, End: ast.Location{ - Line: int(593), + Line: int(578), Column: int(15), }, File: p1, @@ -85531,9 +80559,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -85543,19 +80570,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(593), + Line: int(578), Column: int(21), }, End: ast.Location{ - Line: int(593), + Line: int(578), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6489, + Ctx: p6485, FreeVars: ast.Identifiers{ "val", }, @@ -85577,38 +80604,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(593), + Line: int(578), Column: int(29), }, End: ast.Location{ - Line: int(593), + Line: int(578), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(594), + Line: int(579), Column: int(11), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(47), }, File: p1, @@ -85621,7 +80647,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -85631,19 +80657,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(594), + Line: int(579), Column: int(17), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -85653,19 +80679,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(594), + Line: int(579), Column: int(17), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -85673,19 +80699,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(594), + Line: int(579), Column: int(17), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -85693,25 +80719,24 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(594), + Line: int(579), Column: int(17), }, End: ast.Location{ - Line: int(594), + Line: int(579), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "Format required number at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{ ast.FodderElement{ @@ -85725,19 +80750,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(595), + Line: int(580), Column: int(19), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -85750,25 +80775,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(595), + Line: int(580), Column: int(23), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: ", got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -85776,19 +80800,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(595), + Line: int(580), Column: int(34), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -85797,19 +80821,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(595), + Line: int(580), Column: int(34), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -85817,13 +80841,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(595), + Line: int(580), Column: int(34), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(37), }, File: p1, @@ -85858,9 +80882,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -85870,19 +80893,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(595), + Line: int(580), Column: int(43), }, End: ast.Location{ - Line: int(595), + Line: int(580), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6515, + Ctx: p6511, FreeVars: ast.Identifiers{ "val", }, @@ -85912,14 +80935,14 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(597), + Line: int(582), Column: int(11), }, End: ast.Location{ - Line: int(598), - Column: int(111), + Line: int(583), + Column: int(80), }, File: p1, }, @@ -85931,12 +80954,11 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "iprec", "render_int", - "std", "val", "zp", }, @@ -85949,19 +80971,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(597), + Line: int(582), Column: int(31), }, End: ast.Location{ - Line: int(597), + Line: int(582), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6523, + Ctx: p6519, FreeVars: ast.Identifiers{ "cflags", }, @@ -85969,19 +80991,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(597), + Line: int(582), Column: int(34), }, End: ast.Location{ - Line: int(597), + Line: int(582), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6523, + Ctx: p6519, FreeVars: ast.Identifiers{ "cflags", }, @@ -85989,13 +81011,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(597), + Line: int(582), Column: int(34), }, End: ast.Location{ - Line: int(597), + Line: int(582), Column: int(40), }, File: p1, @@ -86030,97 +81052,81 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(597), + Line: int(582), Column: int(50), }, End: ast.Location{ - Line: int(597), + Line: int(582), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6523, + Ctx: p6519, FreeVars: nil, }, Value: "0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(597), + Line: int(582), Column: int(59), }, End: ast.Location{ - Line: int(597), + Line: int(582), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6523, + Ctx: p6519, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(597), - Column: int(17), - }, - End: ast.Location{ - Line: int(597), - Column: int(61), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), + Line: int(583), Column: int(11), }, End: ast.Location{ - Line: int(598), - Column: int(111), + Line: int(583), + Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "iprec", "render_int", - "std", "val", "zero_prefix", "zp", @@ -86129,13 +81135,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), + Line: int(583), Column: int(11), }, End: ast.Location{ - Line: int(598), + Line: int(583), Column: int(21), }, File: p1, @@ -86148,7 +81154,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "render_int", }, @@ -86159,323 +81165,27 @@ var _StdAst = &ast.DesugaredObject{ Arguments: ast.Arguments{ Positional: []ast.CommaSeparatedExpr{ ast.CommaSeparatedExpr{ - Expr: &ast.Binary{ + Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), + Line: int(583), Column: int(22), }, End: ast.Location{ - Line: int(598), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6539, - FreeVars: ast.Identifiers{ - "val", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(22), - }, - End: ast.Location{ - Line: int(598), - Column: int(25), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6539, - FreeVars: ast.Identifiers{ - "val", - }, - }, - Id: "val", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(10), - Right: &ast.Unary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(29), - }, - End: ast.Location{ - Line: int(598), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6539, - FreeVars: nil, - }, - Op: ast.UnaryOp(3), - Expr: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(30), - }, - End: ast.Location{ - Line: int(598), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6539, - FreeVars: nil, - }, - OriginalString: "1", - }, - }, - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(33), - }, - End: ast.Location{ - Line: int(598), - Column: int(56), + Line: int(583), + Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6539, + Ctx: p6535, FreeVars: ast.Identifiers{ - "std", "val", }, }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(33), - }, - End: ast.Location{ - Line: int(598), - Column: int(42), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6539, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(33), - }, - End: ast.Location{ - Line: int(598), - Column: int(36), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "floor", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(43), - }, - End: ast.Location{ - Line: int(598), - Column: int(55), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6554, - FreeVars: ast.Identifiers{ - "std", - "val", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(43), - }, - End: ast.Location{ - Line: int(598), - Column: int(50), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6554, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(43), - }, - End: ast.Location{ - Line: int(598), - Column: int(46), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "abs", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(598), - Column: int(51), - }, - End: ast.Location{ - Line: int(598), - Column: int(54), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6563, - FreeVars: ast.Identifiers{ - "val", - }, - }, - Id: "val", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, + Id: "val", }, CommaFodder: ast.Fodder{}, }, @@ -86483,19 +81193,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(58), + Line: int(583), + Column: int(27), }, End: ast.Location{ - Line: int(598), - Column: int(60), + Line: int(583), + Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6539, + Ctx: p6535, FreeVars: ast.Identifiers{ "zp", }, @@ -86508,19 +81218,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(62), + Line: int(583), + Column: int(31), }, End: ast.Location{ - Line: int(598), - Column: int(67), + Line: int(583), + Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6539, + Ctx: p6535, FreeVars: ast.Identifiers{ "iprec", }, @@ -86533,19 +81243,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(69), + Line: int(583), + Column: int(38), }, End: ast.Location{ - Line: int(598), - Column: int(81), + Line: int(583), + Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6539, + Ctx: p6535, FreeVars: ast.Identifiers{ "cflags", }, @@ -86553,14 +81263,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(69), + Line: int(583), + Column: int(38), }, End: ast.Location{ - Line: int(598), - Column: int(75), + Line: int(583), + Column: int(44), }, File: p1, }, @@ -86594,9 +81304,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -86605,19 +81314,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(83), + Line: int(583), + Column: int(52), }, End: ast.Location{ - Line: int(598), - Column: int(94), + Line: int(583), + Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6539, + Ctx: p6535, FreeVars: ast.Identifiers{ "cflags", }, @@ -86625,14 +81334,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(83), + Line: int(583), + Column: int(52), }, End: ast.Location{ - Line: int(598), - Column: int(89), + Line: int(583), + Column: int(58), }, File: p1, }, @@ -86663,12 +81372,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -86677,21 +81385,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(96), + Line: int(583), + Column: int(65), }, End: ast.Location{ - Line: int(598), - Column: int(97), + Line: int(583), + Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6539, + Ctx: p6535, FreeVars: nil, }, + Value: float64(8), OriginalString: "8", }, CommaFodder: ast.Fodder{}, @@ -86700,19 +81409,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(598), - Column: int(99), + Line: int(583), + Column: int(68), }, End: ast.Location{ - Line: int(598), - Column: int(110), + Line: int(583), + Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6539, + Ctx: p6535, FreeVars: ast.Identifiers{ "zero_prefix", }, @@ -86742,19 +81451,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(599), + Line: int(584), Column: int(12), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -86772,19 +81481,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(599), + Line: int(584), Column: int(15), }, End: ast.Location{ - Line: int(599), + Line: int(584), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -86792,19 +81501,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(599), + Line: int(584), Column: int(15), }, End: ast.Location{ - Line: int(599), + Line: int(584), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -86812,13 +81521,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(599), + Line: int(584), Column: int(15), }, End: ast.Location{ - Line: int(599), + Line: int(584), Column: int(19), }, File: p1, @@ -86853,9 +81562,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -86863,38 +81571,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(599), + Line: int(584), Column: int(29), }, End: ast.Location{ - Line: int(599), + Line: int(584), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "x", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(600), + Line: int(585), Column: int(9), }, End: ast.Location{ - Line: int(610), + Line: int(595), Column: int(32), }, File: p1, @@ -86907,7 +81614,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -86922,19 +81629,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(600), + Line: int(585), Column: int(12), }, End: ast.Location{ - Line: int(600), + Line: int(585), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -86943,19 +81650,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(600), + Line: int(585), Column: int(12), }, End: ast.Location{ - Line: int(600), + Line: int(585), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -86964,19 +81671,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(600), + Line: int(585), Column: int(12), }, End: ast.Location{ - Line: int(600), + Line: int(585), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -86984,13 +81691,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(600), + Line: int(585), Column: int(12), }, End: ast.Location{ - Line: int(600), + Line: int(585), Column: int(15), }, File: p1, @@ -87025,9 +81732,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -87037,19 +81743,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(600), + Line: int(585), Column: int(21), }, End: ast.Location{ - Line: int(600), + Line: int(585), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6607, + Ctx: p6579, FreeVars: ast.Identifiers{ "val", }, @@ -87071,38 +81777,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(600), + Line: int(585), Column: int(29), }, End: ast.Location{ - Line: int(600), + Line: int(585), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(601), + Line: int(586), Column: int(11), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(47), }, File: p1, @@ -87115,7 +81820,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -87125,19 +81830,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(601), + Line: int(586), Column: int(17), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -87147,19 +81852,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(601), + Line: int(586), Column: int(17), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -87167,19 +81872,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(601), + Line: int(586), Column: int(17), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -87187,25 +81892,24 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(601), + Line: int(586), Column: int(17), }, End: ast.Location{ - Line: int(601), + Line: int(586), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "Format required number at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{ ast.FodderElement{ @@ -87219,19 +81923,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(602), + Line: int(587), Column: int(19), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -87244,25 +81948,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(602), + Line: int(587), Column: int(23), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: ", got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -87270,19 +81973,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(602), + Line: int(587), Column: int(34), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -87291,19 +81994,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(602), + Line: int(587), Column: int(34), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -87311,13 +82014,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(602), + Line: int(587), Column: int(34), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(37), }, File: p1, @@ -87352,9 +82055,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -87364,19 +82066,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(602), + Line: int(587), Column: int(43), }, End: ast.Location{ - Line: int(602), + Line: int(587), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6633, + Ctx: p6605, FreeVars: ast.Identifiers{ "val", }, @@ -87406,25 +82108,24 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(604), + Line: int(589), Column: int(11), }, End: ast.Location{ - Line: int(610), + Line: int(595), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", "iprec", "render_hex", - "std", "val", "zp", }, @@ -87432,13 +82133,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(604), + Line: int(589), Column: int(11), }, End: ast.Location{ - Line: int(604), + Line: int(589), Column: int(21), }, File: p1, @@ -87451,7 +82152,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "render_hex", }, @@ -87462,131 +82163,27 @@ var _StdAst = &ast.DesugaredObject{ Arguments: ast.Arguments{ Positional: []ast.CommaSeparatedExpr{ ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ + Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(604), + Line: int(589), Column: int(22), }, End: ast.Location{ - Line: int(604), - Column: int(36), + Line: int(589), + Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6643, + Ctx: p6615, FreeVars: ast.Identifiers{ - "std", "val", }, }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(604), - Column: int(22), - }, - End: ast.Location{ - Line: int(604), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6643, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(604), - Column: int(22), - }, - End: ast.Location{ - Line: int(604), - Column: int(25), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "floor", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(604), - Column: int(32), - }, - End: ast.Location{ - Line: int(604), - Column: int(35), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p6652, - FreeVars: ast.Identifiers{ - "val", - }, - }, - Id: "val", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, + Id: "val", }, CommaFodder: ast.Fodder{}, }, @@ -87594,13 +82191,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(605), + Line: int(590), Column: int(22), }, End: ast.Location{ - Line: int(605), + Line: int(590), Column: int(24), }, File: p1, @@ -87613,7 +82210,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6643, + Ctx: p6615, FreeVars: ast.Identifiers{ "zp", }, @@ -87626,13 +82223,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(606), + Line: int(591), Column: int(22), }, End: ast.Location{ - Line: int(606), + Line: int(591), Column: int(27), }, File: p1, @@ -87645,7 +82242,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6643, + Ctx: p6615, FreeVars: ast.Identifiers{ "iprec", }, @@ -87658,19 +82255,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(607), + Line: int(592), Column: int(22), }, End: ast.Location{ - Line: int(607), + Line: int(592), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6643, + Ctx: p6615, FreeVars: ast.Identifiers{ "cflags", }, @@ -87678,13 +82275,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(607), + Line: int(592), Column: int(22), }, End: ast.Location{ - Line: int(607), + Line: int(592), Column: int(28), }, File: p1, @@ -87726,9 +82323,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -87737,19 +82333,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(608), + Line: int(593), Column: int(22), }, End: ast.Location{ - Line: int(608), + Line: int(593), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6643, + Ctx: p6615, FreeVars: ast.Identifiers{ "cflags", }, @@ -87757,13 +82353,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(608), + Line: int(593), Column: int(22), }, End: ast.Location{ - Line: int(608), + Line: int(593), Column: int(28), }, File: p1, @@ -87802,12 +82398,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -87816,19 +82411,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(609), + Line: int(594), Column: int(22), }, End: ast.Location{ - Line: int(609), + Line: int(594), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6643, + Ctx: p6615, FreeVars: ast.Identifiers{ "cflags", }, @@ -87836,13 +82431,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(609), + Line: int(594), Column: int(22), }, End: ast.Location{ - Line: int(609), + Line: int(594), Column: int(28), }, File: p1, @@ -87884,9 +82479,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -87895,19 +82489,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(610), + Line: int(595), Column: int(22), }, End: ast.Location{ - Line: int(610), + Line: int(595), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6643, + Ctx: p6615, FreeVars: ast.Identifiers{ "code", }, @@ -87915,13 +82509,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(610), + Line: int(595), Column: int(22), }, End: ast.Location{ - Line: int(610), + Line: int(595), Column: int(26), }, File: p1, @@ -87963,9 +82557,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -87990,19 +82583,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(611), + Line: int(596), Column: int(12), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -88018,19 +82611,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(611), + Line: int(596), Column: int(15), }, End: ast.Location{ - Line: int(611), + Line: int(596), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -88038,19 +82631,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(611), + Line: int(596), Column: int(15), }, End: ast.Location{ - Line: int(611), + Line: int(596), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -88058,13 +82651,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(611), + Line: int(596), Column: int(15), }, End: ast.Location{ - Line: int(611), + Line: int(596), Column: int(19), }, File: p1, @@ -88099,9 +82692,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -88109,38 +82701,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(611), + Line: int(596), Column: int(29), }, End: ast.Location{ - Line: int(611), + Line: int(596), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "f", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(612), + Line: int(597), Column: int(9), }, End: ast.Location{ - Line: int(622), + Line: int(607), Column: int(35), }, File: p1, @@ -88153,7 +82744,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "fpprec", @@ -88167,19 +82758,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(612), + Line: int(597), Column: int(12), }, End: ast.Location{ - Line: int(612), + Line: int(597), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -88188,19 +82779,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(612), + Line: int(597), Column: int(12), }, End: ast.Location{ - Line: int(612), + Line: int(597), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -88209,19 +82800,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(612), + Line: int(597), Column: int(12), }, End: ast.Location{ - Line: int(612), + Line: int(597), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -88229,13 +82820,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(612), + Line: int(597), Column: int(12), }, End: ast.Location{ - Line: int(612), + Line: int(597), Column: int(15), }, File: p1, @@ -88270,9 +82861,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -88282,19 +82872,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(612), + Line: int(597), Column: int(21), }, End: ast.Location{ - Line: int(612), + Line: int(597), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6709, + Ctx: p6672, FreeVars: ast.Identifiers{ "val", }, @@ -88316,38 +82906,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(612), + Line: int(597), Column: int(29), }, End: ast.Location{ - Line: int(612), + Line: int(597), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(613), + Line: int(598), Column: int(11), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(47), }, File: p1, @@ -88360,7 +82949,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -88370,19 +82959,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(613), + Line: int(598), Column: int(17), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -88392,19 +82981,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(613), + Line: int(598), Column: int(17), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -88412,19 +83001,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(613), + Line: int(598), Column: int(17), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -88432,25 +83021,24 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(613), + Line: int(598), Column: int(17), }, End: ast.Location{ - Line: int(613), + Line: int(598), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "Format required number at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{ ast.FodderElement{ @@ -88464,19 +83052,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(614), + Line: int(599), Column: int(19), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -88489,25 +83077,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(614), + Line: int(599), Column: int(23), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: ", got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -88515,19 +83102,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(614), + Line: int(599), Column: int(34), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -88536,19 +83123,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(614), + Line: int(599), Column: int(34), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -88556,13 +83143,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(614), + Line: int(599), Column: int(34), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(37), }, File: p1, @@ -88597,9 +83184,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -88609,19 +83195,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(614), + Line: int(599), Column: int(43), }, End: ast.Location{ - Line: int(614), + Line: int(599), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6735, + Ctx: p6698, FreeVars: ast.Identifiers{ "val", }, @@ -88651,19 +83237,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(616), + Line: int(601), Column: int(11), }, End: ast.Location{ - Line: int(622), + Line: int(607), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "fpprec", @@ -88675,13 +83261,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(616), + Line: int(601), Column: int(11), }, End: ast.Location{ - Line: int(616), + Line: int(601), Column: int(27), }, File: p1, @@ -88694,7 +83280,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "render_float_dec", }, @@ -88708,19 +83294,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(616), + Line: int(601), Column: int(28), }, End: ast.Location{ - Line: int(616), + Line: int(601), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6745, + Ctx: p6708, FreeVars: ast.Identifiers{ "val", }, @@ -88733,13 +83319,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(617), + Line: int(602), Column: int(28), }, End: ast.Location{ - Line: int(617), + Line: int(602), Column: int(30), }, File: p1, @@ -88752,7 +83338,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6745, + Ctx: p6708, FreeVars: ast.Identifiers{ "zp", }, @@ -88765,19 +83351,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(618), + Line: int(603), Column: int(28), }, End: ast.Location{ - Line: int(618), + Line: int(603), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6745, + Ctx: p6708, FreeVars: ast.Identifiers{ "cflags", }, @@ -88785,13 +83371,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(618), + Line: int(603), Column: int(28), }, End: ast.Location{ - Line: int(618), + Line: int(603), Column: int(34), }, File: p1, @@ -88833,9 +83419,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -88844,19 +83429,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(619), + Line: int(604), Column: int(28), }, End: ast.Location{ - Line: int(619), + Line: int(604), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6745, + Ctx: p6708, FreeVars: ast.Identifiers{ "cflags", }, @@ -88864,13 +83449,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(619), + Line: int(604), Column: int(28), }, End: ast.Location{ - Line: int(619), + Line: int(604), Column: int(34), }, File: p1, @@ -88909,12 +83494,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -88923,19 +83507,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(620), + Line: int(605), Column: int(28), }, End: ast.Location{ - Line: int(620), + Line: int(605), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6745, + Ctx: p6708, FreeVars: ast.Identifiers{ "cflags", }, @@ -88943,13 +83527,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(620), + Line: int(605), Column: int(28), }, End: ast.Location{ - Line: int(620), + Line: int(605), Column: int(34), }, File: p1, @@ -88991,9 +83575,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -89002,13 +83585,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(621), + Line: int(606), Column: int(28), }, End: ast.Location{ - Line: int(621), + Line: int(606), Column: int(32), }, File: p1, @@ -89021,7 +83604,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6745, + Ctx: p6708, FreeVars: nil, }, Value: true, @@ -89032,13 +83615,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(622), + Line: int(607), Column: int(28), }, End: ast.Location{ - Line: int(622), + Line: int(607), Column: int(34), }, File: p1, @@ -89051,7 +83634,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6745, + Ctx: p6708, FreeVars: ast.Identifiers{ "fpprec", }, @@ -89080,19 +83663,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(623), + Line: int(608), Column: int(12), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -89108,19 +83691,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(623), + Line: int(608), Column: int(15), }, End: ast.Location{ - Line: int(623), + Line: int(608), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -89128,19 +83711,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(623), + Line: int(608), Column: int(15), }, End: ast.Location{ - Line: int(623), + Line: int(608), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -89148,13 +83731,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(623), + Line: int(608), Column: int(15), }, End: ast.Location{ - Line: int(623), + Line: int(608), Column: int(19), }, File: p1, @@ -89189,9 +83772,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -89199,38 +83781,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(623), + Line: int(608), Column: int(29), }, End: ast.Location{ - Line: int(623), + Line: int(608), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "e", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(624), + Line: int(609), Column: int(9), }, End: ast.Location{ - Line: int(635), + Line: int(620), Column: int(35), }, File: p1, @@ -89243,7 +83824,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -89258,19 +83839,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(624), + Line: int(609), Column: int(12), }, End: ast.Location{ - Line: int(624), + Line: int(609), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -89279,19 +83860,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(624), + Line: int(609), Column: int(12), }, End: ast.Location{ - Line: int(624), + Line: int(609), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -89300,19 +83881,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(624), + Line: int(609), Column: int(12), }, End: ast.Location{ - Line: int(624), + Line: int(609), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -89320,13 +83901,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(624), + Line: int(609), Column: int(12), }, End: ast.Location{ - Line: int(624), + Line: int(609), Column: int(15), }, File: p1, @@ -89361,9 +83942,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -89373,19 +83953,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(624), + Line: int(609), Column: int(21), }, End: ast.Location{ - Line: int(624), + Line: int(609), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6798, + Ctx: p6761, FreeVars: ast.Identifiers{ "val", }, @@ -89407,38 +83987,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(624), + Line: int(609), Column: int(29), }, End: ast.Location{ - Line: int(624), + Line: int(609), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(625), + Line: int(610), Column: int(11), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(47), }, File: p1, @@ -89451,7 +84030,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -89461,19 +84040,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(625), + Line: int(610), Column: int(17), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -89483,19 +84062,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(625), + Line: int(610), Column: int(17), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -89503,19 +84082,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(625), + Line: int(610), Column: int(17), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -89523,25 +84102,24 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(625), + Line: int(610), Column: int(17), }, End: ast.Location{ - Line: int(625), + Line: int(610), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "Format required number at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{ ast.FodderElement{ @@ -89555,19 +84133,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(626), + Line: int(611), Column: int(19), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -89580,25 +84158,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(626), + Line: int(611), Column: int(23), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: ", got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -89606,19 +84183,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(626), + Line: int(611), Column: int(34), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -89627,19 +84204,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(626), + Line: int(611), Column: int(34), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -89647,13 +84224,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(626), + Line: int(611), Column: int(34), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(37), }, File: p1, @@ -89688,9 +84265,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -89700,19 +84276,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(626), + Line: int(611), Column: int(43), }, End: ast.Location{ - Line: int(626), + Line: int(611), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6824, + Ctx: p6787, FreeVars: ast.Identifiers{ "val", }, @@ -89742,19 +84318,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(628), + Line: int(613), Column: int(11), }, End: ast.Location{ - Line: int(635), + Line: int(620), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -89767,13 +84343,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(628), + Line: int(613), Column: int(11), }, End: ast.Location{ - Line: int(628), + Line: int(613), Column: int(27), }, File: p1, @@ -89786,7 +84362,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "render_float_sci", }, @@ -89800,19 +84376,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(628), + Line: int(613), Column: int(28), }, End: ast.Location{ - Line: int(628), + Line: int(613), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6834, + Ctx: p6797, FreeVars: ast.Identifiers{ "val", }, @@ -89825,13 +84401,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(629), + Line: int(614), Column: int(28), }, End: ast.Location{ - Line: int(629), + Line: int(614), Column: int(30), }, File: p1, @@ -89844,7 +84420,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6834, + Ctx: p6797, FreeVars: ast.Identifiers{ "zp", }, @@ -89857,19 +84433,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(630), + Line: int(615), Column: int(28), }, End: ast.Location{ - Line: int(630), + Line: int(615), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6834, + Ctx: p6797, FreeVars: ast.Identifiers{ "cflags", }, @@ -89877,13 +84453,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(630), + Line: int(615), Column: int(28), }, End: ast.Location{ - Line: int(630), + Line: int(615), Column: int(34), }, File: p1, @@ -89925,9 +84501,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -89936,19 +84511,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(631), + Line: int(616), Column: int(28), }, End: ast.Location{ - Line: int(631), + Line: int(616), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6834, + Ctx: p6797, FreeVars: ast.Identifiers{ "cflags", }, @@ -89956,13 +84531,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(631), + Line: int(616), Column: int(28), }, End: ast.Location{ - Line: int(631), + Line: int(616), Column: int(34), }, File: p1, @@ -90001,12 +84576,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -90015,19 +84589,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(632), + Line: int(617), Column: int(28), }, End: ast.Location{ - Line: int(632), + Line: int(617), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6834, + Ctx: p6797, FreeVars: ast.Identifiers{ "cflags", }, @@ -90035,13 +84609,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(632), + Line: int(617), Column: int(28), }, End: ast.Location{ - Line: int(632), + Line: int(617), Column: int(34), }, File: p1, @@ -90083,9 +84657,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -90094,13 +84667,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(633), + Line: int(618), Column: int(28), }, End: ast.Location{ - Line: int(633), + Line: int(618), Column: int(32), }, File: p1, @@ -90113,7 +84686,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6834, + Ctx: p6797, FreeVars: nil, }, Value: true, @@ -90124,19 +84697,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(634), + Line: int(619), Column: int(28), }, End: ast.Location{ - Line: int(634), + Line: int(619), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6834, + Ctx: p6797, FreeVars: ast.Identifiers{ "code", }, @@ -90144,13 +84717,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(634), + Line: int(619), Column: int(28), }, End: ast.Location{ - Line: int(634), + Line: int(619), Column: int(32), }, File: p1, @@ -90192,9 +84765,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -90203,13 +84775,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(635), + Line: int(620), Column: int(28), }, End: ast.Location{ - Line: int(635), + Line: int(620), Column: int(34), }, File: p1, @@ -90222,7 +84794,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6834, + Ctx: p6797, FreeVars: ast.Identifiers{ "fpprec", }, @@ -90251,19 +84823,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(636), + Line: int(621), Column: int(12), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -90279,19 +84851,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(636), + Line: int(621), Column: int(15), }, End: ast.Location{ - Line: int(636), + Line: int(621), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -90299,19 +84871,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(636), + Line: int(621), Column: int(15), }, End: ast.Location{ - Line: int(636), + Line: int(621), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -90319,13 +84891,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(636), + Line: int(621), Column: int(15), }, End: ast.Location{ - Line: int(636), + Line: int(621), Column: int(19), }, File: p1, @@ -90360,9 +84932,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -90370,38 +84941,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(636), + Line: int(621), Column: int(29), }, End: ast.Location{ - Line: int(636), + Line: int(621), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "g", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(637), + Line: int(622), Column: int(9), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(56), }, File: p1, @@ -90414,7 +84984,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -90430,19 +85000,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(637), + Line: int(622), Column: int(12), }, End: ast.Location{ - Line: int(637), + Line: int(622), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -90451,19 +85021,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(637), + Line: int(622), Column: int(12), }, End: ast.Location{ - Line: int(637), + Line: int(622), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -90472,19 +85042,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(637), + Line: int(622), Column: int(12), }, End: ast.Location{ - Line: int(637), + Line: int(622), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -90492,13 +85062,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(637), + Line: int(622), Column: int(12), }, End: ast.Location{ - Line: int(637), + Line: int(622), Column: int(15), }, File: p1, @@ -90533,9 +85103,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -90545,19 +85114,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(637), + Line: int(622), Column: int(21), }, End: ast.Location{ - Line: int(637), + Line: int(622), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6893, + Ctx: p6856, FreeVars: ast.Identifiers{ "val", }, @@ -90579,38 +85148,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(637), + Line: int(622), Column: int(29), }, End: ast.Location{ - Line: int(637), + Line: int(622), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(638), + Line: int(623), Column: int(11), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(47), }, File: p1, @@ -90623,7 +85191,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -90633,19 +85201,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(638), + Line: int(623), Column: int(17), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", "std", @@ -90655,19 +85223,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(638), + Line: int(623), Column: int(17), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -90675,19 +85243,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(638), + Line: int(623), Column: int(17), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -90695,25 +85263,24 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(638), + Line: int(623), Column: int(17), }, End: ast.Location{ - Line: int(638), + Line: int(623), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "Format required number at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{ ast.FodderElement{ @@ -90727,19 +85294,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(639), + Line: int(624), Column: int(19), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "i", }, @@ -90752,25 +85319,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(639), + Line: int(624), Column: int(23), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: ", got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -90778,19 +85344,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(639), + Line: int(624), Column: int(34), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -90799,19 +85365,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(639), + Line: int(624), Column: int(34), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -90819,13 +85385,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(639), + Line: int(624), Column: int(34), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(37), }, File: p1, @@ -90860,9 +85426,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -90872,19 +85437,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(639), + Line: int(624), Column: int(43), }, End: ast.Location{ - Line: int(639), + Line: int(624), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6919, + Ctx: p6882, FreeVars: ast.Identifiers{ "val", }, @@ -90914,13 +85479,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(11), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(56), }, File: p1, @@ -90933,7 +85498,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -90953,19 +85518,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(28), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(74), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6927, + Ctx: p6890, FreeVars: ast.Identifiers{ "std", "val", @@ -90974,19 +85539,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(28), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6927, + Ctx: p6890, FreeVars: ast.Identifiers{ "std", }, @@ -90994,13 +85559,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(28), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(31), }, File: p1, @@ -91035,9 +85600,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -91047,19 +85611,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(38), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6936, + Ctx: p6899, FreeVars: ast.Identifiers{ "std", "val", @@ -91068,19 +85632,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(38), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6936, + Ctx: p6899, FreeVars: ast.Identifiers{ "std", "val", @@ -91089,19 +85653,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(38), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6936, + Ctx: p6899, FreeVars: ast.Identifiers{ "std", }, @@ -91109,13 +85673,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(38), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(41), }, File: p1, @@ -91150,9 +85714,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "log", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -91162,19 +85725,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(46), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6947, + Ctx: p6910, FreeVars: ast.Identifiers{ "std", "val", @@ -91183,19 +85746,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(46), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6947, + Ctx: p6910, FreeVars: ast.Identifiers{ "std", }, @@ -91203,13 +85766,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(46), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(49), }, File: p1, @@ -91244,9 +85807,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "abs", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -91256,19 +85818,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(54), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6956, + Ctx: p6919, FreeVars: ast.Identifiers{ "val", }, @@ -91300,19 +85862,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(62), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6936, + Ctx: p6899, FreeVars: ast.Identifiers{ "std", }, @@ -91320,19 +85882,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(62), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6936, + Ctx: p6899, FreeVars: ast.Identifiers{ "std", }, @@ -91340,13 +85902,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(62), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(65), }, File: p1, @@ -91381,9 +85943,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "log", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -91393,21 +85954,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(641), + Line: int(626), Column: int(70), }, End: ast.Location{ - Line: int(641), + Line: int(626), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6967, + Ctx: p6930, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, CommaFodder: nil, @@ -91433,30 +85995,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(641), - Column: int(17), - }, - End: ast.Location{ - Line: int(641), - Column: int(74), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(11), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(56), }, File: p1, @@ -91469,7 +86019,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -91485,19 +86035,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(14), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "exponent", "fpprec", @@ -91506,19 +86056,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(14), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "exponent", }, @@ -91526,19 +86076,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(14), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "exponent", }, @@ -91550,40 +86100,41 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(25), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Op: ast.UnaryOp(3), Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(26), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -91593,19 +86144,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(31), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "exponent", "fpprec", @@ -91614,19 +86165,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(31), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "exponent", }, @@ -91638,19 +86189,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(642), + Line: int(627), Column: int(43), }, End: ast.Location{ - Line: int(642), + Line: int(627), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "fpprec", }, @@ -91663,19 +86214,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(643), + Line: int(628), Column: int(13), }, End: ast.Location{ - Line: int(650), + Line: int(635), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "code", @@ -91688,13 +86239,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(643), + Line: int(628), Column: int(13), }, End: ast.Location{ - Line: int(643), + Line: int(628), Column: int(29), }, File: p1, @@ -91707,7 +86258,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "render_float_sci", }, @@ -91721,19 +86272,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(643), + Line: int(628), Column: int(30), }, End: ast.Location{ - Line: int(643), + Line: int(628), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "val", }, @@ -91746,13 +86297,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(644), + Line: int(629), Column: int(30), }, End: ast.Location{ - Line: int(644), + Line: int(629), Column: int(32), }, File: p1, @@ -91765,7 +86316,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "zp", }, @@ -91778,19 +86329,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(645), + Line: int(630), Column: int(30), }, End: ast.Location{ - Line: int(645), + Line: int(630), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "cflags", }, @@ -91798,13 +86349,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(645), + Line: int(630), Column: int(30), }, End: ast.Location{ - Line: int(645), + Line: int(630), Column: int(36), }, File: p1, @@ -91846,9 +86397,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -91857,19 +86407,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(646), + Line: int(631), Column: int(30), }, End: ast.Location{ - Line: int(646), + Line: int(631), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "cflags", }, @@ -91877,13 +86427,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(646), + Line: int(631), Column: int(30), }, End: ast.Location{ - Line: int(646), + Line: int(631), Column: int(36), }, File: p1, @@ -91922,12 +86472,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -91936,19 +86485,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(647), + Line: int(632), Column: int(30), }, End: ast.Location{ - Line: int(647), + Line: int(632), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "cflags", }, @@ -91956,13 +86505,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(647), + Line: int(632), Column: int(30), }, End: ast.Location{ - Line: int(647), + Line: int(632), Column: int(36), }, File: p1, @@ -92004,9 +86553,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -92015,19 +86563,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(648), + Line: int(633), Column: int(30), }, End: ast.Location{ - Line: int(648), + Line: int(633), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "cflags", }, @@ -92035,13 +86583,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(648), + Line: int(633), Column: int(30), }, End: ast.Location{ - Line: int(648), + Line: int(633), Column: int(36), }, File: p1, @@ -92083,9 +86631,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -92094,19 +86641,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(649), + Line: int(634), Column: int(30), }, End: ast.Location{ - Line: int(649), + Line: int(634), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "code", }, @@ -92114,13 +86661,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(649), + Line: int(634), Column: int(30), }, End: ast.Location{ - Line: int(649), + Line: int(634), Column: int(34), }, File: p1, @@ -92162,9 +86709,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "caps", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -92173,19 +86719,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(650), + Line: int(635), Column: int(30), }, End: ast.Location{ - Line: int(650), + Line: int(635), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "fpprec", }, @@ -92193,13 +86739,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(650), + Line: int(635), Column: int(30), }, End: ast.Location{ - Line: int(650), + Line: int(635), Column: int(36), }, File: p1, @@ -92212,7 +86758,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6992, + Ctx: p6955, FreeVars: ast.Identifiers{ "fpprec", }, @@ -92224,21 +86770,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(650), + Line: int(635), Column: int(39), }, End: ast.Location{ - Line: int(650), + Line: int(635), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6992, + Ctx: p6955, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -92263,13 +86810,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(13), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(56), }, File: p1, @@ -92282,7 +86829,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "exponent", @@ -92301,19 +86848,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(38), }, End: ast.Location{ - Line: int(652), + Line: int(637), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7039, + Ctx: p7002, FreeVars: ast.Identifiers{ "exponent", "std", @@ -92322,19 +86869,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(38), }, End: ast.Location{ - Line: int(652), + Line: int(637), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7039, + Ctx: p7002, FreeVars: ast.Identifiers{ "std", }, @@ -92342,13 +86889,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(38), }, End: ast.Location{ - Line: int(652), + Line: int(637), Column: int(41), }, File: p1, @@ -92383,9 +86930,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "max", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -92395,21 +86941,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(46), }, End: ast.Location{ - Line: int(652), + Line: int(637), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7048, + Ctx: p7011, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, CommaFodder: ast.Fodder{}, @@ -92418,19 +86965,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(49), }, End: ast.Location{ - Line: int(652), + Line: int(637), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7048, + Ctx: p7011, FreeVars: ast.Identifiers{ "exponent", }, @@ -92438,19 +86985,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(49), }, End: ast.Location{ - Line: int(652), + Line: int(637), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7048, + Ctx: p7011, FreeVars: ast.Identifiers{ "exponent", }, @@ -92462,21 +87009,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(652), + Line: int(637), Column: int(60), }, End: ast.Location{ - Line: int(652), + Line: int(637), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7048, + Ctx: p7011, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -92492,36 +87040,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(652), - Column: int(19), - }, - End: ast.Location{ - Line: int(652), - Column: int(62), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(653), + Line: int(638), Column: int(13), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "cflags", "digits_before_pt", @@ -92534,13 +87070,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(653), + Line: int(638), Column: int(13), }, End: ast.Location{ - Line: int(653), + Line: int(638), Column: int(29), }, File: p1, @@ -92553,7 +87089,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "render_float_dec", }, @@ -92567,19 +87103,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(653), + Line: int(638), Column: int(30), }, End: ast.Location{ - Line: int(653), + Line: int(638), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "val", }, @@ -92592,13 +87128,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(654), + Line: int(639), Column: int(30), }, End: ast.Location{ - Line: int(654), + Line: int(639), Column: int(32), }, File: p1, @@ -92611,7 +87147,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "zp", }, @@ -92624,19 +87160,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(655), + Line: int(640), Column: int(30), }, End: ast.Location{ - Line: int(655), + Line: int(640), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "cflags", }, @@ -92644,13 +87180,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(655), + Line: int(640), Column: int(30), }, End: ast.Location{ - Line: int(655), + Line: int(640), Column: int(36), }, File: p1, @@ -92692,9 +87228,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "blank", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -92703,19 +87238,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(656), + Line: int(641), Column: int(30), }, End: ast.Location{ - Line: int(656), + Line: int(641), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "cflags", }, @@ -92723,13 +87258,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(656), + Line: int(641), Column: int(30), }, End: ast.Location{ - Line: int(656), + Line: int(641), Column: int(36), }, File: p1, @@ -92768,12 +87303,11 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "plus", + Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -92782,19 +87316,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(657), + Line: int(642), Column: int(30), }, End: ast.Location{ - Line: int(657), + Line: int(642), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "cflags", }, @@ -92802,13 +87336,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(657), + Line: int(642), Column: int(30), }, End: ast.Location{ - Line: int(657), + Line: int(642), Column: int(36), }, File: p1, @@ -92850,9 +87384,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -92861,19 +87394,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(658), + Line: int(643), Column: int(30), }, End: ast.Location{ - Line: int(658), + Line: int(643), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "cflags", }, @@ -92881,13 +87414,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(658), + Line: int(643), Column: int(30), }, End: ast.Location{ - Line: int(658), + Line: int(643), Column: int(36), }, File: p1, @@ -92929,9 +87462,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "alt", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -92940,19 +87472,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(659), + Line: int(644), Column: int(30), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "digits_before_pt", "fpprec", @@ -92961,13 +87493,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(659), + Line: int(644), Column: int(30), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(36), }, File: p1, @@ -92980,7 +87512,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "fpprec", }, @@ -92992,19 +87524,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(659), + Line: int(644), Column: int(39), }, End: ast.Location{ - Line: int(659), + Line: int(644), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7061, + Ctx: p7024, FreeVars: ast.Identifiers{ "digits_before_pt", }, @@ -93037,19 +87569,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(660), + Line: int(645), Column: int(12), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", "std", @@ -93059,19 +87591,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(660), + Line: int(645), Column: int(15), }, End: ast.Location{ - Line: int(660), + Line: int(645), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -93079,19 +87611,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(660), + Line: int(645), Column: int(15), }, End: ast.Location{ - Line: int(660), + Line: int(645), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -93099,13 +87631,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(660), + Line: int(645), Column: int(15), }, End: ast.Location{ - Line: int(660), + Line: int(645), Column: int(19), }, File: p1, @@ -93140,9 +87672,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -93150,38 +87681,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(660), + Line: int(645), Column: int(29), }, End: ast.Location{ - Line: int(660), + Line: int(645), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "c", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(661), + Line: int(646), Column: int(9), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(69), }, File: p1, @@ -93194,7 +87724,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93203,19 +87733,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(661), + Line: int(646), Column: int(12), }, End: ast.Location{ - Line: int(661), + Line: int(646), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93224,19 +87754,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(661), + Line: int(646), Column: int(12), }, End: ast.Location{ - Line: int(661), + Line: int(646), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93245,19 +87775,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(661), + Line: int(646), Column: int(12), }, End: ast.Location{ - Line: int(661), + Line: int(646), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -93265,13 +87795,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(661), + Line: int(646), Column: int(12), }, End: ast.Location{ - Line: int(661), + Line: int(646), Column: int(15), }, File: p1, @@ -93306,9 +87836,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -93318,19 +87847,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(661), + Line: int(646), Column: int(21), }, End: ast.Location{ - Line: int(661), + Line: int(646), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7122, + Ctx: p7085, FreeVars: ast.Identifiers{ "val", }, @@ -93352,44 +87881,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(661), + Line: int(646), Column: int(29), }, End: ast.Location{ - Line: int(661), + Line: int(646), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "number", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(662), + Line: int(647), Column: int(11), }, End: ast.Location{ - Line: int(662), + Line: int(647), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93398,19 +87926,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(662), + Line: int(647), Column: int(11), }, End: ast.Location{ - Line: int(662), + Line: int(647), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -93418,13 +87946,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(662), + Line: int(647), Column: int(11), }, End: ast.Location{ - Line: int(662), + Line: int(647), Column: int(14), }, File: p1, @@ -93466,9 +87994,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "char", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -93478,19 +88005,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(662), + Line: int(647), Column: int(20), }, End: ast.Location{ - Line: int(662), + Line: int(647), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7135, + Ctx: p7098, FreeVars: ast.Identifiers{ "val", }, @@ -93518,19 +88045,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(663), + Line: int(648), Column: int(14), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93539,19 +88066,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(663), + Line: int(648), Column: int(17), }, End: ast.Location{ - Line: int(663), + Line: int(648), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93560,19 +88087,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(663), + Line: int(648), Column: int(17), }, End: ast.Location{ - Line: int(663), + Line: int(648), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93581,19 +88108,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(663), + Line: int(648), Column: int(17), }, End: ast.Location{ - Line: int(663), + Line: int(648), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -93601,13 +88128,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(663), + Line: int(648), Column: int(17), }, End: ast.Location{ - Line: int(663), + Line: int(648), Column: int(20), }, File: p1, @@ -93642,9 +88169,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -93654,19 +88180,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(663), + Line: int(648), Column: int(26), }, End: ast.Location{ - Line: int(663), + Line: int(648), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7151, + Ctx: p7114, FreeVars: ast.Identifiers{ "val", }, @@ -93688,38 +88214,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(663), + Line: int(648), Column: int(34), }, End: ast.Location{ - Line: int(663), + Line: int(648), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(664), + Line: int(649), Column: int(11), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(71), }, File: p1, @@ -93732,7 +88257,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93741,19 +88266,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(664), + Line: int(649), Column: int(14), }, End: ast.Location{ - Line: int(664), + Line: int(649), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93762,19 +88287,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(664), + Line: int(649), Column: int(14), }, End: ast.Location{ - Line: int(664), + Line: int(649), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93783,19 +88308,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(664), + Line: int(649), Column: int(14), }, End: ast.Location{ - Line: int(664), + Line: int(649), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -93803,13 +88328,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(664), + Line: int(649), Column: int(14), }, End: ast.Location{ - Line: int(664), + Line: int(649), Column: int(17), }, File: p1, @@ -93844,9 +88369,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -93856,19 +88380,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(664), + Line: int(649), Column: int(25), }, End: ast.Location{ - Line: int(664), + Line: int(649), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7168, + Ctx: p7131, FreeVars: ast.Identifiers{ "val", }, @@ -93890,21 +88414,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(664), + Line: int(649), Column: int(33), }, End: ast.Location{ - Line: int(664), + Line: int(649), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -93912,13 +88437,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(665), + Line: int(650), Column: int(13), }, End: ast.Location{ - Line: int(665), + Line: int(650), Column: int(16), }, File: p1, @@ -93931,7 +88456,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "val", }, @@ -93949,13 +88474,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(667), + Line: int(652), Column: int(13), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(71), }, File: p1, @@ -93968,7 +88493,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93977,19 +88502,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(667), + Line: int(652), Column: int(19), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -93998,44 +88523,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(667), + Line: int(652), Column: int(19), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "%c expected 1-sized string got: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(667), + Line: int(652), Column: int(56), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -94044,19 +88568,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(667), + Line: int(652), Column: int(56), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -94064,13 +88588,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(667), + Line: int(652), Column: int(56), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(59), }, File: p1, @@ -94105,9 +88629,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -94117,19 +88640,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(667), + Line: int(652), Column: int(67), }, End: ast.Location{ - Line: int(667), + Line: int(652), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7190, + Ctx: p7153, FreeVars: ast.Identifiers{ "val", }, @@ -94160,13 +88683,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(669), + Line: int(654), Column: int(11), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(69), }, File: p1, @@ -94179,7 +88702,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -94188,19 +88711,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(669), + Line: int(654), Column: int(17), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -94209,44 +88732,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(669), + Line: int(654), Column: int(17), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "%c expected number / string, got: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(669), + Line: int(654), Column: int(56), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", "val", @@ -94255,19 +88777,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(669), + Line: int(654), Column: int(56), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "std", }, @@ -94275,13 +88797,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(669), + Line: int(654), Column: int(56), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(59), }, File: p1, @@ -94316,9 +88838,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -94328,19 +88849,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(669), + Line: int(654), Column: int(65), }, End: ast.Location{ - Line: int(669), + Line: int(654), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7208, + Ctx: p7171, FreeVars: ast.Identifiers{ "val", }, @@ -94372,13 +88893,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(671), + Line: int(656), Column: int(9), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, @@ -94391,7 +88912,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -94399,19 +88920,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(671), + Line: int(656), Column: int(15), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -94419,44 +88940,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(671), + Line: int(656), Column: int(15), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: nil, }, Value: "Unknown code: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(671), + Line: int(656), Column: int(34), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p6275, + Ctx: p6295, FreeVars: ast.Identifiers{ "code", }, @@ -94464,13 +88984,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(671), + Line: int(656), Column: int(34), }, End: ast.Location{ - Line: int(671), + Line: int(656), Column: int(38), }, File: p1, @@ -94505,9 +89025,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -94527,30 +89046,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(674), + Line: int(659), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -94563,7 +89070,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -94589,19 +89096,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(674), + Line: int(659), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, }, Fodder: nil, - Ctx: p7228, + Ctx: p7191, FreeVars: ast.Identifiers{ "format_code", "format_codes_arr", @@ -94611,115 +89118,48 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "codes", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(674), - Column: int(28), - }, - End: ast.Location{ - Line: int(674), - Column: int(33), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "codes", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(674), - Column: int(35), - }, - End: ast.Location{ - Line: int(674), - Column: int(38), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(674), - Column: int(40), - }, - End: ast.Location{ - Line: int(674), - Column: int(41), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(674), - Column: int(43), - }, - End: ast.Location{ - Line: int(674), - Column: int(44), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(674), - Column: int(46), - }, - End: ast.Location{ - Line: int(674), - Column: int(47), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(675), + Line: int(660), Column: int(7), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -94732,7 +89172,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "codes", @@ -94749,19 +89189,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(675), + Line: int(660), Column: int(10), }, End: ast.Location{ - Line: int(675), + Line: int(660), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "codes", "i", @@ -94771,19 +89211,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(675), + Line: int(660), Column: int(10), }, End: ast.Location{ - Line: int(675), + Line: int(660), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "i", }, @@ -94795,19 +89235,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(675), + Line: int(660), Column: int(15), }, End: ast.Location{ - Line: int(675), + Line: int(660), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "codes", "std", @@ -94816,19 +89256,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(675), + Line: int(660), Column: int(15), }, End: ast.Location{ - Line: int(675), + Line: int(660), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "std", }, @@ -94836,13 +89276,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(675), + Line: int(660), Column: int(15), }, End: ast.Location{ - Line: int(675), + Line: int(660), Column: int(18), }, File: p1, @@ -94877,9 +89317,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -94889,19 +89328,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(675), + Line: int(660), Column: int(26), }, End: ast.Location{ - Line: int(675), + Line: int(660), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7248, + Ctx: p7211, FreeVars: ast.Identifiers{ "codes", }, @@ -94923,13 +89362,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(676), + Line: int(661), Column: int(9), }, End: ast.Location{ - Line: int(679), + Line: int(664), Column: int(12), }, File: p1, @@ -94942,7 +89381,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "j", @@ -94953,19 +89392,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(676), + Line: int(661), Column: int(12), }, End: ast.Location{ - Line: int(676), + Line: int(661), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "j", @@ -94975,19 +89414,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(676), + Line: int(661), Column: int(12), }, End: ast.Location{ - Line: int(676), + Line: int(661), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "j", }, @@ -94999,19 +89438,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(676), + Line: int(661), Column: int(16), }, End: ast.Location{ - Line: int(676), + Line: int(661), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "std", @@ -95020,19 +89459,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(676), + Line: int(661), Column: int(16), }, End: ast.Location{ - Line: int(676), + Line: int(661), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "std", }, @@ -95040,13 +89479,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(676), + Line: int(661), Column: int(16), }, End: ast.Location{ - Line: int(676), + Line: int(661), Column: int(19), }, File: p1, @@ -95081,9 +89520,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -95093,19 +89531,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(676), + Line: int(661), Column: int(27), }, End: ast.Location{ - Line: int(676), + Line: int(661), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7266, + Ctx: p7229, FreeVars: ast.Identifiers{ "arr", }, @@ -95127,13 +89565,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(11), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(86), }, File: p1, @@ -95146,7 +89584,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "j", @@ -95156,19 +89594,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(18), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "j", @@ -95178,19 +89616,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(18), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(81), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "std", @@ -95199,19 +89637,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(18), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "std", @@ -95220,44 +89658,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(18), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: nil, }, Value: "Too many values to format: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(50), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "std", @@ -95266,19 +89703,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(50), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "std", }, @@ -95286,13 +89723,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(50), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(53), }, File: p1, @@ -95327,9 +89764,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -95339,19 +89775,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(61), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7287, + Ctx: p7250, FreeVars: ast.Identifiers{ "arr", }, @@ -95374,25 +89810,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(68), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(81), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: nil, }, Value: ", expected ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -95400,19 +89835,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(677), + Line: int(662), Column: int(84), }, End: ast.Location{ - Line: int(677), + Line: int(662), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "j", }, @@ -95432,13 +89867,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(679), + Line: int(664), Column: int(11), }, End: ast.Location{ - Line: int(679), + Line: int(664), Column: int(12), }, File: p1, @@ -95451,7 +89886,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "v", }, @@ -95470,13 +89905,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(681), + Line: int(666), Column: int(9), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -95489,7 +89924,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "codes", @@ -95511,19 +89946,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(681), + Line: int(666), Column: int(22), }, End: ast.Location{ - Line: int(681), + Line: int(666), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7302, + Ctx: p7265, FreeVars: ast.Identifiers{ "codes", "i", @@ -95532,19 +89967,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(681), + Line: int(666), Column: int(22), }, End: ast.Location{ - Line: int(681), + Line: int(666), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7302, + Ctx: p7265, FreeVars: ast.Identifiers{ "codes", }, @@ -95555,19 +89990,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(681), + Line: int(666), Column: int(28), }, End: ast.Location{ - Line: int(681), + Line: int(666), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7302, + Ctx: p7265, FreeVars: ast.Identifiers{ "i", }, @@ -95579,30 +90014,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(681), - Column: int(15), - }, - End: ast.Location{ - Line: int(681), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(682), + Line: int(667), Column: int(9), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -95615,7 +90038,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -95633,19 +90056,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(682), + Line: int(667), Column: int(12), }, End: ast.Location{ - Line: int(682), + Line: int(667), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "code", "std", @@ -95654,19 +90077,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(682), + Line: int(667), Column: int(12), }, End: ast.Location{ - Line: int(682), + Line: int(667), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "code", "std", @@ -95675,19 +90098,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(682), + Line: int(667), Column: int(12), }, End: ast.Location{ - Line: int(682), + Line: int(667), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "std", }, @@ -95695,13 +90118,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(682), + Line: int(667), Column: int(12), }, End: ast.Location{ - Line: int(682), + Line: int(667), Column: int(15), }, File: p1, @@ -95736,9 +90159,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -95748,19 +90170,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(682), + Line: int(667), Column: int(21), }, End: ast.Location{ - Line: int(682), + Line: int(667), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7322, + Ctx: p7285, FreeVars: ast.Identifiers{ "code", }, @@ -95782,44 +90204,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(682), + Line: int(667), Column: int(30), }, End: ast.Location{ - Line: int(682), + Line: int(667), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: nil, }, Value: "string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(11), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -95833,13 +90254,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(11), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(27), }, File: p1, @@ -95852,7 +90273,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "format_codes_arr", }, @@ -95866,19 +90287,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(28), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "codes", }, @@ -95891,19 +90312,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(35), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "arr", }, @@ -95916,19 +90337,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(40), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "i", }, @@ -95936,19 +90357,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(40), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "i", }, @@ -95960,21 +90381,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(44), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -95984,19 +90406,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(47), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "j", }, @@ -96009,19 +90431,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(50), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "code", "v", @@ -96030,19 +90452,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(50), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "v", }, @@ -96054,19 +90476,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(683), + Line: int(668), Column: int(54), }, End: ast.Location{ - Line: int(683), + Line: int(668), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7332, + Ctx: p7295, FreeVars: ast.Identifiers{ "code", }, @@ -96095,13 +90517,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(685), + Line: int(670), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -96114,7 +90536,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -96137,19 +90559,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(685), + Line: int(670), Column: int(23), }, End: ast.Location{ - Line: int(694), + Line: int(679), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7355, + Ctx: p7318, FreeVars: ast.Identifiers{ "arr", "code", @@ -96160,19 +90582,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(685), + Line: int(670), Column: int(26), }, End: ast.Location{ - Line: int(685), + Line: int(670), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7355, + Ctx: p7318, FreeVars: ast.Identifiers{ "code", }, @@ -96180,19 +90602,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(685), + Line: int(670), Column: int(26), }, End: ast.Location{ - Line: int(685), + Line: int(670), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7355, + Ctx: p7318, FreeVars: ast.Identifiers{ "code", }, @@ -96200,13 +90622,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(685), + Line: int(670), Column: int(26), }, End: ast.Location{ - Line: int(685), + Line: int(670), Column: int(30), }, File: p1, @@ -96241,9 +90663,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -96251,44 +90672,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(685), + Line: int(670), Column: int(37), }, End: ast.Location{ - Line: int(685), + Line: int(670), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7355, + Ctx: p7318, FreeVars: nil, }, Value: "*", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(685), + Line: int(670), Column: int(46), }, End: ast.Location{ - Line: int(691), + Line: int(676), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7355, + Ctx: p7318, FreeVars: ast.Identifiers{ "arr", "j", @@ -96320,24 +90740,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(686), + Line: int(671), Column: int(16), }, End: ast.Location{ - Line: int(686), + Line: int(671), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "j", }, @@ -96345,19 +90764,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(686), + Line: int(671), Column: int(16), }, End: ast.Location{ - Line: int(686), + Line: int(671), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "j", }, @@ -96369,37 +90788,26 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(686), + Line: int(671), Column: int(20), }, End: ast.Location{ - Line: int(686), + Line: int(671), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(686), - Column: int(13), - }, - End: ast.Location{ - Line: int(686), - Column: int(21), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -96424,24 +90832,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(687), + Line: int(672), Column: int(17), }, End: ast.Location{ - Line: int(690), + Line: int(675), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "j", @@ -96451,19 +90858,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(687), + Line: int(672), Column: int(20), }, End: ast.Location{ - Line: int(687), + Line: int(672), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "j", @@ -96473,19 +90880,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(687), + Line: int(672), Column: int(20), }, End: ast.Location{ - Line: int(687), + Line: int(672), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "j", }, @@ -96497,19 +90904,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(687), + Line: int(672), Column: int(25), }, End: ast.Location{ - Line: int(687), + Line: int(672), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "std", @@ -96518,19 +90925,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(687), + Line: int(672), Column: int(25), }, End: ast.Location{ - Line: int(687), + Line: int(672), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "std", }, @@ -96538,13 +90945,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(687), + Line: int(672), Column: int(25), }, End: ast.Location{ - Line: int(687), + Line: int(672), Column: int(28), }, File: p1, @@ -96579,9 +90986,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -96591,19 +90997,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(687), + Line: int(672), Column: int(36), }, End: ast.Location{ - Line: int(687), + Line: int(672), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7391, + Ctx: p7354, FreeVars: ast.Identifiers{ "arr", }, @@ -96625,13 +91031,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(15), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(101), }, File: p1, @@ -96644,7 +91050,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "j", @@ -96654,19 +91060,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(22), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(100), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "j", @@ -96676,19 +91082,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(22), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(96), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "std", @@ -96697,19 +91103,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(22), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "std", @@ -96718,44 +91124,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(22), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: nil, }, Value: "Not enough values to format: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(56), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "std", @@ -96764,19 +91169,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(56), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "std", }, @@ -96784,13 +91189,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(56), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(59), }, File: p1, @@ -96825,9 +91230,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -96837,19 +91241,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(67), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7412, + Ctx: p7375, FreeVars: ast.Identifiers{ "arr", }, @@ -96872,25 +91276,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(74), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(96), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: nil, }, Value: ", expected at least ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -96898,19 +91301,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(688), + Line: int(673), Column: int(99), }, End: ast.Location{ - Line: int(688), + Line: int(673), Column: int(100), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "j", }, @@ -96930,19 +91333,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(690), + Line: int(675), Column: int(15), }, End: ast.Location{ - Line: int(690), + Line: int(675), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", "j", @@ -96951,13 +91354,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(690), + Line: int(675), Column: int(15), }, End: ast.Location{ - Line: int(690), + Line: int(675), Column: int(18), }, File: p1, @@ -96970,7 +91373,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "arr", }, @@ -96981,19 +91384,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(690), + Line: int(675), Column: int(19), }, End: ast.Location{ - Line: int(690), + Line: int(675), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7370, + Ctx: p7333, FreeVars: ast.Identifiers{ "j", }, @@ -97005,18 +91408,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(687), - Column: int(13), - }, - End: ast.Location{ - Line: int(690), - Column: int(21), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -97025,19 +91416,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(691), + Line: int(676), Column: int(18), }, End: ast.Location{ - Line: int(694), + Line: int(679), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7355, + Ctx: p7318, FreeVars: ast.Identifiers{ "code", "j", @@ -97068,24 +91459,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(692), + Line: int(677), Column: int(16), }, End: ast.Location{ - Line: int(692), + Line: int(677), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7430, + Ctx: p7393, FreeVars: ast.Identifiers{ "j", }, @@ -97093,18 +91483,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "j", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(692), - Column: int(13), - }, - End: ast.Location{ - Line: int(692), - Column: int(17), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -97129,24 +91507,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(693), + Line: int(678), Column: int(17), }, End: ast.Location{ - Line: int(693), + Line: int(678), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7430, + Ctx: p7393, FreeVars: ast.Identifiers{ "code", }, @@ -97154,13 +91531,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(693), + Line: int(678), Column: int(17), }, End: ast.Location{ - Line: int(693), + Line: int(678), Column: int(21), }, File: p1, @@ -97195,24 +91572,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(693), - Column: int(13), - }, - End: ast.Location{ - Line: int(693), - Column: int(24), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -97220,30 +91584,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(685), - Column: int(17), - }, - End: ast.Location{ - Line: int(694), - Column: int(12), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(695), + Line: int(680), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -97256,7 +91608,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -97279,19 +91631,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(695), + Line: int(680), Column: int(24), }, End: ast.Location{ - Line: int(704), + Line: int(689), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7443, + Ctx: p7406, FreeVars: ast.Identifiers{ "arr", "code", @@ -97302,19 +91654,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(695), + Line: int(680), Column: int(27), }, End: ast.Location{ - Line: int(695), + Line: int(680), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7443, + Ctx: p7406, FreeVars: ast.Identifiers{ "code", }, @@ -97322,19 +91674,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(695), + Line: int(680), Column: int(27), }, End: ast.Location{ - Line: int(695), + Line: int(680), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7443, + Ctx: p7406, FreeVars: ast.Identifiers{ "code", }, @@ -97342,13 +91694,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(695), + Line: int(680), Column: int(27), }, End: ast.Location{ - Line: int(695), + Line: int(680), Column: int(31), }, File: p1, @@ -97383,9 +91735,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -97393,44 +91744,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(695), + Line: int(680), Column: int(40), }, End: ast.Location{ - Line: int(695), + Line: int(680), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7443, + Ctx: p7406, FreeVars: nil, }, Value: "*", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(695), + Line: int(680), Column: int(49), }, End: ast.Location{ - Line: int(701), + Line: int(686), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7443, + Ctx: p7406, FreeVars: ast.Identifiers{ "arr", "std", @@ -97462,24 +91812,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(696), + Line: int(681), Column: int(16), }, End: ast.Location{ - Line: int(696), + Line: int(681), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "tmp", }, @@ -97487,19 +91836,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(696), + Line: int(681), Column: int(16), }, End: ast.Location{ - Line: int(696), + Line: int(681), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "tmp", }, @@ -97507,13 +91856,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(696), + Line: int(681), Column: int(16), }, End: ast.Location{ - Line: int(696), + Line: int(681), Column: int(19), }, File: p1, @@ -97548,9 +91897,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -97558,37 +91906,26 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(696), + Line: int(681), Column: int(24), }, End: ast.Location{ - Line: int(696), + Line: int(681), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(696), - Column: int(13), - }, - End: ast.Location{ - Line: int(696), - Column: int(25), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -97613,24 +91950,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(19), }, End: ast.Location{ - Line: int(700), + Line: int(685), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -97640,19 +91976,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(22), }, End: ast.Location{ - Line: int(697), + Line: int(682), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -97662,19 +91998,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(22), }, End: ast.Location{ - Line: int(697), + Line: int(682), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "tmp", }, @@ -97682,13 +92018,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(22), }, End: ast.Location{ - Line: int(697), + Line: int(682), Column: int(25), }, File: p1, @@ -97723,9 +92059,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -97733,19 +92068,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(31), }, End: ast.Location{ - Line: int(697), + Line: int(682), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -97754,19 +92089,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(31), }, End: ast.Location{ - Line: int(697), + Line: int(682), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "std", }, @@ -97774,13 +92109,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(31), }, End: ast.Location{ - Line: int(697), + Line: int(682), Column: int(34), }, File: p1, @@ -97815,9 +92150,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -97827,19 +92161,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(697), + Line: int(682), Column: int(42), }, End: ast.Location{ - Line: int(697), + Line: int(682), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7485, + Ctx: p7448, FreeVars: ast.Identifiers{ "arr", }, @@ -97861,13 +92195,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(15), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(105), }, File: p1, @@ -97880,7 +92214,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -97890,19 +92224,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(22), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(104), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -97912,19 +92246,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(22), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(96), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -97933,19 +92267,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(22), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -97954,44 +92288,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(22), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: nil, }, Value: "Not enough values to format: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(56), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "std", @@ -98000,19 +92333,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(56), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "std", }, @@ -98020,13 +92353,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(56), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(59), }, File: p1, @@ -98061,9 +92394,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -98073,19 +92405,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(67), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7506, + Ctx: p7469, FreeVars: ast.Identifiers{ "arr", }, @@ -98108,25 +92440,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(74), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(96), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: nil, }, Value: ", expected at least ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -98134,19 +92465,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(99), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(104), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "tmp", }, @@ -98154,13 +92485,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(698), + Line: int(683), Column: int(99), }, End: ast.Location{ - Line: int(698), + Line: int(683), Column: int(102), }, File: p1, @@ -98195,9 +92526,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -98213,19 +92543,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(700), + Line: int(685), Column: int(15), }, End: ast.Location{ - Line: int(700), + Line: int(685), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", "tmp", @@ -98234,13 +92564,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(700), + Line: int(685), Column: int(15), }, End: ast.Location{ - Line: int(700), + Line: int(685), Column: int(18), }, File: p1, @@ -98253,7 +92583,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "arr", }, @@ -98264,19 +92594,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(700), + Line: int(685), Column: int(19), }, End: ast.Location{ - Line: int(700), + Line: int(685), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7458, + Ctx: p7421, FreeVars: ast.Identifiers{ "tmp", }, @@ -98284,13 +92614,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(700), + Line: int(685), Column: int(19), }, End: ast.Location{ - Line: int(700), + Line: int(685), Column: int(22), }, File: p1, @@ -98325,9 +92655,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, RightBracketFodder: ast.Fodder{}, @@ -98335,18 +92664,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(697), - Column: int(13), - }, - End: ast.Location{ - Line: int(700), - Column: int(25), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -98355,19 +92672,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(701), + Line: int(686), Column: int(18), }, End: ast.Location{ - Line: int(704), + Line: int(689), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7443, + Ctx: p7406, FreeVars: ast.Identifiers{ "code", "tmp", @@ -98398,24 +92715,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(702), + Line: int(687), Column: int(16), }, End: ast.Location{ - Line: int(702), + Line: int(687), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7530, + Ctx: p7493, FreeVars: ast.Identifiers{ "tmp", }, @@ -98423,13 +92739,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(702), + Line: int(687), Column: int(16), }, End: ast.Location{ - Line: int(702), + Line: int(687), Column: int(19), }, File: p1, @@ -98464,24 +92780,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(702), - Column: int(13), - }, - End: ast.Location{ - Line: int(702), - Column: int(21), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -98506,24 +92809,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(703), + Line: int(688), Column: int(19), }, End: ast.Location{ - Line: int(703), + Line: int(688), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7530, + Ctx: p7493, FreeVars: ast.Identifiers{ "code", }, @@ -98531,13 +92833,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(703), + Line: int(688), Column: int(19), }, End: ast.Location{ - Line: int(703), + Line: int(688), Column: int(23), }, File: p1, @@ -98572,24 +92874,11 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(703), - Column: int(13), - }, - End: ast.Location{ - Line: int(703), - Column: int(28), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -98597,30 +92886,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(695), - Column: int(17), - }, - End: ast.Location{ - Line: int(704), - Column: int(12), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(705), + Line: int(690), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -98633,7 +92910,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -98657,19 +92934,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(705), + Line: int(690), Column: int(22), }, End: ast.Location{ - Line: int(705), + Line: int(690), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7546, + Ctx: p7509, FreeVars: ast.Identifiers{ "tmp2", }, @@ -98677,13 +92954,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(705), + Line: int(690), Column: int(22), }, End: ast.Location{ - Line: int(705), + Line: int(690), Column: int(26), }, File: p1, @@ -98718,37 +92995,24 @@ var _StdAst = &ast.DesugaredObject{ Value: "j", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(705), - Column: int(17), - }, - End: ast.Location{ - Line: int(705), - Column: int(28), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(706), + Line: int(691), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -98761,7 +93025,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -98786,13 +93050,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(707), + Line: int(692), Column: int(13), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(103), }, File: p1, @@ -98805,7 +93069,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "j2", @@ -98815,19 +93079,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(707), + Line: int(692), Column: int(16), }, End: ast.Location{ - Line: int(707), + Line: int(692), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "j2", @@ -98837,19 +93101,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(707), + Line: int(692), Column: int(16), }, End: ast.Location{ - Line: int(707), + Line: int(692), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "j2", }, @@ -98861,19 +93125,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(707), + Line: int(692), Column: int(21), }, End: ast.Location{ - Line: int(707), + Line: int(692), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "std", @@ -98882,19 +93146,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(707), + Line: int(692), Column: int(21), }, End: ast.Location{ - Line: int(707), + Line: int(692), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "std", }, @@ -98902,13 +93166,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(707), + Line: int(692), Column: int(21), }, End: ast.Location{ - Line: int(707), + Line: int(692), Column: int(24), }, File: p1, @@ -98943,9 +93207,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -98955,19 +93218,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(707), + Line: int(692), Column: int(32), }, End: ast.Location{ - Line: int(707), + Line: int(692), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7572, + Ctx: p7535, FreeVars: ast.Identifiers{ "arr", }, @@ -98989,19 +93252,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(708), + Line: int(693), Column: int(15), }, End: ast.Location{ - Line: int(708), + Line: int(693), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "j2", @@ -99010,13 +93273,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(708), + Line: int(693), Column: int(15), }, End: ast.Location{ - Line: int(708), + Line: int(693), Column: int(18), }, File: p1, @@ -99029,7 +93292,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", }, @@ -99040,19 +93303,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(708), + Line: int(693), Column: int(19), }, End: ast.Location{ - Line: int(708), + Line: int(693), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "j2", }, @@ -99073,13 +93336,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(15), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(103), }, File: p1, @@ -99092,7 +93355,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "j2", @@ -99102,19 +93365,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(22), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(102), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "j2", @@ -99124,19 +93387,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(22), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(97), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "std", @@ -99145,19 +93408,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(22), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "std", @@ -99166,44 +93429,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(22), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: nil, }, Value: "Not enough values to format: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(56), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "arr", "std", @@ -99212,19 +93474,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(56), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "std", }, @@ -99232,13 +93494,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(56), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(59), }, File: p1, @@ -99273,9 +93535,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -99285,19 +93546,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(67), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7601, + Ctx: p7564, FreeVars: ast.Identifiers{ "arr", }, @@ -99320,25 +93581,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(74), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(97), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: nil, }, Value: ", expected more than ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -99346,19 +93606,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(710), + Line: int(695), Column: int(100), }, End: ast.Location{ - Line: int(710), + Line: int(695), Column: int(102), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7557, + Ctx: p7520, FreeVars: ast.Identifiers{ "j2", }, @@ -99370,30 +93630,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(706), - Column: int(17), - }, - End: ast.Location{ - Line: int(710), - Column: int(103), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(711), + Line: int(696), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -99406,7 +93654,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -99431,13 +93679,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(712), + Line: int(697), Column: int(13), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(60), }, File: p1, @@ -99450,7 +93698,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7612, + Ctx: p7575, FreeVars: ast.Identifiers{ "code", "format_code", @@ -99463,19 +93711,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(712), + Line: int(697), Column: int(16), }, End: ast.Location{ - Line: int(712), + Line: int(697), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7612, + Ctx: p7575, FreeVars: ast.Identifiers{ "code", }, @@ -99483,19 +93731,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(712), + Line: int(697), Column: int(16), }, End: ast.Location{ - Line: int(712), + Line: int(697), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7612, + Ctx: p7575, FreeVars: ast.Identifiers{ "code", }, @@ -99503,13 +93751,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(712), + Line: int(697), Column: int(16), }, End: ast.Location{ - Line: int(712), + Line: int(697), Column: int(20), }, File: p1, @@ -99544,9 +93792,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -99554,38 +93801,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(712), + Line: int(697), Column: int(30), }, End: ast.Location{ - Line: int(712), + Line: int(697), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7612, + Ctx: p7575, FreeVars: nil, }, Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(713), + Line: int(698), Column: int(15), }, End: ast.Location{ - Line: int(713), + Line: int(698), Column: int(18), }, File: p1, @@ -99598,13 +93844,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7612, + Ctx: p7575, FreeVars: nil, }, Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -99617,19 +93862,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(15), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7612, + Ctx: p7575, FreeVars: ast.Identifiers{ "code", "format_code", @@ -99642,13 +93887,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(15), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(26), }, File: p1, @@ -99661,7 +93906,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7612, + Ctx: p7575, FreeVars: ast.Identifiers{ "format_code", }, @@ -99675,19 +93920,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(27), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7632, + Ctx: p7595, FreeVars: ast.Identifiers{ "val", }, @@ -99700,19 +93945,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(32), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7632, + Ctx: p7595, FreeVars: ast.Identifiers{ "code", }, @@ -99725,19 +93970,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(38), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7632, + Ctx: p7595, FreeVars: ast.Identifiers{ "tmp", }, @@ -99745,13 +93990,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(38), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(41), }, File: p1, @@ -99786,9 +94031,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -99797,19 +94041,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(46), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7632, + Ctx: p7595, FreeVars: ast.Identifiers{ "tmp2", }, @@ -99817,13 +94061,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(46), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(50), }, File: p1, @@ -99858,9 +94102,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -99869,19 +94112,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(715), + Line: int(700), Column: int(57), }, End: ast.Location{ - Line: int(715), + Line: int(700), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7632, + Ctx: p7595, FreeVars: ast.Identifiers{ "j2", }, @@ -99901,30 +94144,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(711), - Column: int(17), - }, - End: ast.Location{ - Line: int(715), - Column: int(60), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(716), + Line: int(701), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -99937,7 +94168,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -99960,13 +94191,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(717), + Line: int(702), Column: int(13), }, End: ast.Location{ - Line: int(720), + Line: int(705), Column: int(39), }, File: p1, @@ -99979,7 +94210,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7654, + Ctx: p7617, FreeVars: ast.Identifiers{ "code", "pad_left", @@ -99991,19 +94222,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(717), + Line: int(702), Column: int(16), }, End: ast.Location{ - Line: int(717), + Line: int(702), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7654, + Ctx: p7617, FreeVars: ast.Identifiers{ "code", }, @@ -100011,13 +94242,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(717), + Line: int(702), Column: int(16), }, End: ast.Location{ - Line: int(717), + Line: int(702), Column: int(27), }, File: p1, @@ -100031,13 +94262,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(717), + Line: int(702), Column: int(16), }, End: ast.Location{ - Line: int(717), + Line: int(702), Column: int(20), }, File: p1, @@ -100072,9 +94303,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "cflags", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, LeftBracketFodder: ast.Fodder{}, @@ -100099,28 +94329,27 @@ var _StdAst = &ast.DesugaredObject{ Value: "left", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(718), + Line: int(703), Column: int(15), }, End: ast.Location{ - Line: int(718), + Line: int(703), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7654, + Ctx: p7617, FreeVars: ast.Identifiers{ "pad_right", "s", @@ -100130,13 +94359,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(718), + Line: int(703), Column: int(15), }, End: ast.Location{ - Line: int(718), + Line: int(703), Column: int(24), }, File: p1, @@ -100149,7 +94378,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7654, + Ctx: p7617, FreeVars: ast.Identifiers{ "pad_right", }, @@ -100163,19 +94392,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(718), + Line: int(703), Column: int(25), }, End: ast.Location{ - Line: int(718), + Line: int(703), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7671, + Ctx: p7634, FreeVars: ast.Identifiers{ "s", }, @@ -100188,19 +94417,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(718), + Line: int(703), Column: int(28), }, End: ast.Location{ - Line: int(718), + Line: int(703), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7671, + Ctx: p7634, FreeVars: ast.Identifiers{ "tmp", }, @@ -100208,13 +94437,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(718), + Line: int(703), Column: int(28), }, End: ast.Location{ - Line: int(718), + Line: int(703), Column: int(31), }, File: p1, @@ -100249,9 +94478,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -100260,25 +94488,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(718), + Line: int(703), Column: int(36), }, End: ast.Location{ - Line: int(718), + Line: int(703), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7671, + Ctx: p7634, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -100301,19 +94528,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(720), + Line: int(705), Column: int(15), }, End: ast.Location{ - Line: int(720), + Line: int(705), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7654, + Ctx: p7617, FreeVars: ast.Identifiers{ "pad_left", "s", @@ -100323,13 +94550,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(720), + Line: int(705), Column: int(15), }, End: ast.Location{ - Line: int(720), + Line: int(705), Column: int(23), }, File: p1, @@ -100342,7 +94569,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7654, + Ctx: p7617, FreeVars: ast.Identifiers{ "pad_left", }, @@ -100356,19 +94583,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(720), + Line: int(705), Column: int(24), }, End: ast.Location{ - Line: int(720), + Line: int(705), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7687, + Ctx: p7650, FreeVars: ast.Identifiers{ "s", }, @@ -100381,19 +94608,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(720), + Line: int(705), Column: int(27), }, End: ast.Location{ - Line: int(720), + Line: int(705), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7687, + Ctx: p7650, FreeVars: ast.Identifiers{ "tmp", }, @@ -100401,13 +94628,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(720), + Line: int(705), Column: int(27), }, End: ast.Location{ - Line: int(720), + Line: int(705), Column: int(30), }, File: p1, @@ -100442,9 +94669,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: ast.Fodder{}, @@ -100453,25 +94679,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(720), + Line: int(705), Column: int(35), }, End: ast.Location{ - Line: int(720), + Line: int(705), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7687, + Ctx: p7650, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -100486,30 +94711,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(716), - Column: int(17), - }, - End: ast.Location{ - Line: int(720), - Column: int(39), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(721), + Line: int(706), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, @@ -100522,7 +94735,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "code", @@ -100542,13 +94755,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(722), + Line: int(707), Column: int(13), }, End: ast.Location{ - Line: int(725), + Line: int(710), Column: int(21), }, File: p1, @@ -100561,7 +94774,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7701, + Ctx: p7664, FreeVars: ast.Identifiers{ "code", "j2", @@ -100570,19 +94783,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(722), + Line: int(707), Column: int(16), }, End: ast.Location{ - Line: int(722), + Line: int(707), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7701, + Ctx: p7664, FreeVars: ast.Identifiers{ "code", }, @@ -100590,19 +94803,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(722), + Line: int(707), Column: int(16), }, End: ast.Location{ - Line: int(722), + Line: int(707), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7701, + Ctx: p7664, FreeVars: ast.Identifiers{ "code", }, @@ -100610,13 +94823,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(722), + Line: int(707), Column: int(16), }, End: ast.Location{ - Line: int(722), + Line: int(707), Column: int(20), }, File: p1, @@ -100651,9 +94864,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -100661,38 +94873,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(722), + Line: int(707), Column: int(30), }, End: ast.Location{ - Line: int(722), + Line: int(707), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7701, + Ctx: p7664, FreeVars: nil, }, Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(723), + Line: int(708), Column: int(15), }, End: ast.Location{ - Line: int(723), + Line: int(708), Column: int(17), }, File: p1, @@ -100705,7 +94916,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7701, + Ctx: p7664, FreeVars: ast.Identifiers{ "j2", }, @@ -100723,19 +94934,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(725), + Line: int(710), Column: int(15), }, End: ast.Location{ - Line: int(725), + Line: int(710), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7701, + Ctx: p7664, FreeVars: ast.Identifiers{ "j2", }, @@ -100743,13 +94954,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(725), + Line: int(710), Column: int(15), }, End: ast.Location{ - Line: int(725), + Line: int(710), Column: int(17), }, File: p1, @@ -100762,7 +94973,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7701, + Ctx: p7664, FreeVars: ast.Identifiers{ "j2", }, @@ -100774,57 +94985,46 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(725), + Line: int(710), Column: int(20), }, End: ast.Location{ - Line: int(725), + Line: int(710), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7701, + Ctx: p7664, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(721), - Column: int(17), - }, - End: ast.Location{ - Line: int(725), - Column: int(21), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "arr", "codes", @@ -100838,13 +95038,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(11), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(27), }, File: p1, @@ -100857,7 +95057,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7233, + Ctx: p7196, FreeVars: ast.Identifiers{ "format_codes_arr", }, @@ -100871,19 +95071,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(28), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "codes", }, @@ -100896,19 +95096,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(35), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "arr", }, @@ -100921,19 +95121,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(40), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "i", }, @@ -100941,19 +95141,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(40), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "i", }, @@ -100965,21 +95165,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(44), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -100989,19 +95190,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(47), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "j3", }, @@ -101014,19 +95215,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(51), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "s_padded", "v", @@ -101035,19 +95236,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(51), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "v", }, @@ -101059,19 +95260,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(726), + Line: int(711), Column: int(55), }, End: ast.Location{ - Line: int(726), + Line: int(711), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7728, + Ctx: p7691, FreeVars: ast.Identifiers{ "s_padded", }, @@ -101102,30 +95303,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(729), + Line: int(714), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -101138,7 +95327,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -101165,19 +95354,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(729), + Line: int(714), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, }, Fodder: nil, - Ctx: p7751, + Ctx: p7714, FreeVars: ast.Identifiers{ "format_code", "format_codes_obj", @@ -101187,96 +95376,43 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "codes", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(729), - Column: int(28), - }, - End: ast.Location{ - Line: int(729), - Column: int(33), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "codes", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "obj", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(729), - Column: int(35), - }, - End: ast.Location{ - Line: int(729), - Column: int(38), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "obj", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(729), - Column: int(40), - }, - End: ast.Location{ - Line: int(729), - Column: int(41), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(729), - Column: int(43), - }, - End: ast.Location{ - Line: int(729), - Column: int(44), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(730), + Line: int(715), Column: int(7), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -101289,7 +95425,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "codes", "format_code", @@ -101305,19 +95441,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(730), + Line: int(715), Column: int(10), }, End: ast.Location{ - Line: int(730), + Line: int(715), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "codes", "i", @@ -101327,19 +95463,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(730), + Line: int(715), Column: int(10), }, End: ast.Location{ - Line: int(730), + Line: int(715), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "i", }, @@ -101351,19 +95487,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(730), + Line: int(715), Column: int(15), }, End: ast.Location{ - Line: int(730), + Line: int(715), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "codes", "std", @@ -101372,19 +95508,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(730), + Line: int(715), Column: int(15), }, End: ast.Location{ - Line: int(730), + Line: int(715), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "std", }, @@ -101392,13 +95528,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(730), + Line: int(715), Column: int(15), }, End: ast.Location{ - Line: int(730), + Line: int(715), Column: int(18), }, File: p1, @@ -101433,9 +95569,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -101445,19 +95580,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(730), + Line: int(715), Column: int(26), }, End: ast.Location{ - Line: int(730), + Line: int(715), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7771, + Ctx: p7734, FreeVars: ast.Identifiers{ "codes", }, @@ -101479,13 +95614,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(731), + Line: int(716), Column: int(9), }, End: ast.Location{ - Line: int(731), + Line: int(716), Column: int(10), }, File: p1, @@ -101498,7 +95633,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "v", }, @@ -101516,13 +95651,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(733), + Line: int(718), Column: int(9), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -101535,7 +95670,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "codes", "format_code", @@ -101556,19 +95691,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(733), + Line: int(718), Column: int(22), }, End: ast.Location{ - Line: int(733), + Line: int(718), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7782, + Ctx: p7745, FreeVars: ast.Identifiers{ "codes", "i", @@ -101577,19 +95712,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(733), + Line: int(718), Column: int(22), }, End: ast.Location{ - Line: int(733), + Line: int(718), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7782, + Ctx: p7745, FreeVars: ast.Identifiers{ "codes", }, @@ -101600,19 +95735,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(733), + Line: int(718), Column: int(28), }, End: ast.Location{ - Line: int(733), + Line: int(718), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7782, + Ctx: p7745, FreeVars: ast.Identifiers{ "i", }, @@ -101624,30 +95759,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(733), - Column: int(15), - }, - End: ast.Location{ - Line: int(733), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(734), + Line: int(719), Column: int(9), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -101660,7 +95783,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -101677,19 +95800,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(734), + Line: int(719), Column: int(12), }, End: ast.Location{ - Line: int(734), + Line: int(719), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "std", @@ -101698,19 +95821,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(734), + Line: int(719), Column: int(12), }, End: ast.Location{ - Line: int(734), + Line: int(719), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "std", @@ -101719,19 +95842,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(734), + Line: int(719), Column: int(12), }, End: ast.Location{ - Line: int(734), + Line: int(719), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "std", }, @@ -101739,13 +95862,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(734), + Line: int(719), Column: int(12), }, End: ast.Location{ - Line: int(734), + Line: int(719), Column: int(15), }, File: p1, @@ -101780,9 +95903,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -101792,19 +95914,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(734), + Line: int(719), Column: int(21), }, End: ast.Location{ - Line: int(734), + Line: int(719), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7802, + Ctx: p7765, FreeVars: ast.Identifiers{ "code", }, @@ -101826,44 +95948,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(734), + Line: int(719), Column: int(30), }, End: ast.Location{ - Line: int(734), + Line: int(719), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: nil, }, Value: "string", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(11), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -101876,13 +95997,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(11), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(27), }, File: p1, @@ -101895,7 +96016,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "format_codes_obj", }, @@ -101909,19 +96030,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(28), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: ast.Identifiers{ "codes", }, @@ -101934,19 +96055,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(35), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: ast.Identifiers{ "obj", }, @@ -101959,19 +96080,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(40), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: ast.Identifiers{ "i", }, @@ -101979,19 +96100,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(40), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: ast.Identifiers{ "i", }, @@ -102003,21 +96124,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(44), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -102027,19 +96149,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(47), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: ast.Identifiers{ "code", "v", @@ -102048,19 +96170,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(47), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: ast.Identifiers{ "v", }, @@ -102072,19 +96194,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(735), + Line: int(720), Column: int(51), }, End: ast.Location{ - Line: int(735), + Line: int(720), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7812, + Ctx: p7775, FreeVars: ast.Identifiers{ "code", }, @@ -102113,13 +96235,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(737), + Line: int(722), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -102132,7 +96254,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -102154,13 +96276,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(738), + Line: int(723), Column: int(13), }, End: ast.Location{ - Line: int(741), + Line: int(726), Column: int(24), }, File: p1, @@ -102173,7 +96295,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7834, + Ctx: p7797, FreeVars: ast.Identifiers{ "code", }, @@ -102181,19 +96303,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(738), + Line: int(723), Column: int(16), }, End: ast.Location{ - Line: int(738), + Line: int(723), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7834, + Ctx: p7797, FreeVars: ast.Identifiers{ "code", }, @@ -102201,19 +96323,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(738), + Line: int(723), Column: int(16), }, End: ast.Location{ - Line: int(738), + Line: int(723), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7834, + Ctx: p7797, FreeVars: ast.Identifiers{ "code", }, @@ -102221,13 +96343,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(738), + Line: int(723), Column: int(16), }, End: ast.Location{ - Line: int(738), + Line: int(723), Column: int(20), }, File: p1, @@ -102262,9 +96384,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "mkey", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -102272,19 +96393,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(738), + Line: int(723), Column: int(29), }, End: ast.Location{ - Line: int(738), + Line: int(723), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7834, + Ctx: p7797, FreeVars: nil, }, }, @@ -102293,13 +96414,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(739), + Line: int(724), Column: int(15), }, End: ast.Location{ - Line: int(739), + Line: int(724), Column: int(45), }, File: p1, @@ -102312,31 +96433,30 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7834, + Ctx: p7797, FreeVars: nil, }, Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(739), + Line: int(724), Column: int(21), }, End: ast.Location{ - Line: int(739), + Line: int(724), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7834, + Ctx: p7797, FreeVars: nil, }, Value: "Mapping keys required.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{ @@ -102350,19 +96470,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(741), + Line: int(726), Column: int(15), }, End: ast.Location{ - Line: int(741), + Line: int(726), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7834, + Ctx: p7797, FreeVars: ast.Identifiers{ "code", }, @@ -102370,13 +96490,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(741), + Line: int(726), Column: int(15), }, End: ast.Location{ - Line: int(741), + Line: int(726), Column: int(19), }, File: p1, @@ -102418,38 +96538,25 @@ var _StdAst = &ast.DesugaredObject{ Value: "mkey", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(737), - Column: int(17), - }, - End: ast.Location{ - Line: int(741), - Column: int(24), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(742), + Line: int(727), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -102462,7 +96569,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -102485,13 +96592,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(743), + Line: int(728), Column: int(13), }, End: ast.Location{ - Line: int(746), + Line: int(731), Column: int(22), }, File: p1, @@ -102504,7 +96611,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7860, + Ctx: p7823, FreeVars: ast.Identifiers{ "code", }, @@ -102512,19 +96619,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(743), + Line: int(728), Column: int(16), }, End: ast.Location{ - Line: int(743), + Line: int(728), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7860, + Ctx: p7823, FreeVars: ast.Identifiers{ "code", }, @@ -102532,19 +96639,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(743), + Line: int(728), Column: int(16), }, End: ast.Location{ - Line: int(743), + Line: int(728), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7860, + Ctx: p7823, FreeVars: ast.Identifiers{ "code", }, @@ -102552,13 +96659,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(743), + Line: int(728), Column: int(16), }, End: ast.Location{ - Line: int(743), + Line: int(728), Column: int(20), }, File: p1, @@ -102593,9 +96700,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -102603,38 +96709,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(743), + Line: int(728), Column: int(27), }, End: ast.Location{ - Line: int(743), + Line: int(728), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7860, + Ctx: p7823, FreeVars: nil, }, Value: "*", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(744), + Line: int(729), Column: int(15), }, End: ast.Location{ - Line: int(744), + Line: int(729), Column: int(60), }, File: p1, @@ -102647,31 +96752,30 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7860, + Ctx: p7823, FreeVars: nil, }, Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(744), + Line: int(729), Column: int(21), }, End: ast.Location{ - Line: int(744), + Line: int(729), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7860, + Ctx: p7823, FreeVars: nil, }, Value: "Cannot use * field width with object.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{ @@ -102685,19 +96789,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(746), + Line: int(731), Column: int(15), }, End: ast.Location{ - Line: int(746), + Line: int(731), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7860, + Ctx: p7823, FreeVars: ast.Identifiers{ "code", }, @@ -102705,13 +96809,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(746), + Line: int(731), Column: int(15), }, End: ast.Location{ - Line: int(746), + Line: int(731), Column: int(19), }, File: p1, @@ -102753,38 +96857,25 @@ var _StdAst = &ast.DesugaredObject{ Value: "fw", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(742), - Column: int(17), - }, - End: ast.Location{ - Line: int(746), - Column: int(22), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(747), + Line: int(732), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -102797,7 +96888,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -102821,13 +96912,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(748), + Line: int(733), Column: int(13), }, End: ast.Location{ - Line: int(751), + Line: int(736), Column: int(24), }, File: p1, @@ -102840,7 +96931,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7886, + Ctx: p7849, FreeVars: ast.Identifiers{ "code", }, @@ -102848,19 +96939,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(748), + Line: int(733), Column: int(16), }, End: ast.Location{ - Line: int(748), + Line: int(733), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7886, + Ctx: p7849, FreeVars: ast.Identifiers{ "code", }, @@ -102868,19 +96959,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(748), + Line: int(733), Column: int(16), }, End: ast.Location{ - Line: int(748), + Line: int(733), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7886, + Ctx: p7849, FreeVars: ast.Identifiers{ "code", }, @@ -102888,13 +96979,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(748), + Line: int(733), Column: int(16), }, End: ast.Location{ - Line: int(748), + Line: int(733), Column: int(20), }, File: p1, @@ -102929,9 +97020,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -102939,38 +97029,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(748), + Line: int(733), Column: int(29), }, End: ast.Location{ - Line: int(748), + Line: int(733), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7886, + Ctx: p7849, FreeVars: nil, }, Value: "*", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(749), + Line: int(734), Column: int(15), }, End: ast.Location{ - Line: int(749), + Line: int(734), Column: int(58), }, File: p1, @@ -102983,31 +97072,30 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7886, + Ctx: p7849, FreeVars: nil, }, Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(749), + Line: int(734), Column: int(21), }, End: ast.Location{ - Line: int(749), + Line: int(734), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7886, + Ctx: p7849, FreeVars: nil, }, Value: "Cannot use * precision with object.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{ @@ -103021,19 +97109,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(751), + Line: int(736), Column: int(15), }, End: ast.Location{ - Line: int(751), + Line: int(736), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7886, + Ctx: p7849, FreeVars: ast.Identifiers{ "code", }, @@ -103041,13 +97129,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(751), + Line: int(736), Column: int(15), }, End: ast.Location{ - Line: int(751), + Line: int(736), Column: int(19), }, File: p1, @@ -103089,38 +97177,25 @@ var _StdAst = &ast.DesugaredObject{ Value: "prec", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(747), - Column: int(17), - }, - End: ast.Location{ - Line: int(751), - Column: int(24), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(752), + Line: int(737), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -103133,7 +97208,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -103158,13 +97233,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(753), + Line: int(738), Column: int(13), }, End: ast.Location{ - Line: int(756), + Line: int(741), Column: int(42), }, File: p1, @@ -103177,7 +97252,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "f", "obj", @@ -103187,19 +97262,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(753), + Line: int(738), Column: int(16), }, End: ast.Location{ - Line: int(753), + Line: int(738), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "f", "obj", @@ -103209,19 +97284,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(753), + Line: int(738), Column: int(16), }, End: ast.Location{ - Line: int(753), + Line: int(738), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "std", }, @@ -103229,13 +97304,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(753), + Line: int(738), Column: int(16), }, End: ast.Location{ - Line: int(753), + Line: int(738), Column: int(19), }, File: p1, @@ -103270,9 +97345,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHasAll", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -103282,19 +97356,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(753), + Line: int(738), Column: int(33), }, End: ast.Location{ - Line: int(753), + Line: int(738), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7923, + Ctx: p7886, FreeVars: ast.Identifiers{ "obj", }, @@ -103307,19 +97381,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(753), + Line: int(738), Column: int(38), }, End: ast.Location{ - Line: int(753), + Line: int(738), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7923, + Ctx: p7886, FreeVars: ast.Identifiers{ "f", }, @@ -103340,19 +97414,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(754), + Line: int(739), Column: int(15), }, End: ast.Location{ - Line: int(754), + Line: int(739), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "f", "obj", @@ -103361,13 +97435,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(754), + Line: int(739), Column: int(15), }, End: ast.Location{ - Line: int(754), + Line: int(739), Column: int(18), }, File: p1, @@ -103380,7 +97454,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "obj", }, @@ -103391,19 +97465,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(754), + Line: int(739), Column: int(19), }, End: ast.Location{ - Line: int(754), + Line: int(739), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "f", }, @@ -103424,13 +97498,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(756), + Line: int(741), Column: int(15), }, End: ast.Location{ - Line: int(756), + Line: int(741), Column: int(42), }, File: p1, @@ -103443,7 +97517,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "f", }, @@ -103451,19 +97525,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(756), + Line: int(741), Column: int(21), }, End: ast.Location{ - Line: int(756), + Line: int(741), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "f", }, @@ -103471,44 +97545,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(756), + Line: int(741), Column: int(21), }, End: ast.Location{ - Line: int(756), + Line: int(741), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7912, + Ctx: p7875, FreeVars: nil, }, Value: "No such field: ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(756), + Line: int(741), Column: int(41), }, End: ast.Location{ - Line: int(756), + Line: int(741), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7912, + Ctx: p7875, FreeVars: ast.Identifiers{ "f", }, @@ -103520,30 +97593,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(752), - Column: int(17), - }, - End: ast.Location{ - Line: int(756), - Column: int(42), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(757), + Line: int(742), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -103556,7 +97617,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -103581,13 +97642,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(758), + Line: int(743), Column: int(13), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(50), }, File: p1, @@ -103600,7 +97661,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7949, + Ctx: p7912, FreeVars: ast.Identifiers{ "code", "f", @@ -103613,19 +97674,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(758), + Line: int(743), Column: int(16), }, End: ast.Location{ - Line: int(758), + Line: int(743), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7949, + Ctx: p7912, FreeVars: ast.Identifiers{ "code", }, @@ -103633,19 +97694,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(758), + Line: int(743), Column: int(16), }, End: ast.Location{ - Line: int(758), + Line: int(743), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7949, + Ctx: p7912, FreeVars: ast.Identifiers{ "code", }, @@ -103653,13 +97714,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(758), + Line: int(743), Column: int(16), }, End: ast.Location{ - Line: int(758), + Line: int(743), Column: int(20), }, File: p1, @@ -103694,9 +97755,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "ctype", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, OpFodder: ast.Fodder{}, @@ -103704,38 +97764,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(758), + Line: int(743), Column: int(30), }, End: ast.Location{ - Line: int(758), + Line: int(743), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7949, + Ctx: p7912, FreeVars: nil, }, Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(759), + Line: int(744), Column: int(15), }, End: ast.Location{ - Line: int(759), + Line: int(744), Column: int(18), }, File: p1, @@ -103748,13 +97807,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7949, + Ctx: p7912, FreeVars: nil, }, Value: "%", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -103767,19 +97825,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(761), + Line: int(746), Column: int(15), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7949, + Ctx: p7912, FreeVars: ast.Identifiers{ "code", "f", @@ -103792,13 +97850,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(761), + Line: int(746), Column: int(15), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(26), }, File: p1, @@ -103811,7 +97869,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7949, + Ctx: p7912, FreeVars: ast.Identifiers{ "format_code", }, @@ -103825,19 +97883,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(761), + Line: int(746), Column: int(27), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7969, + Ctx: p7932, FreeVars: ast.Identifiers{ "val", }, @@ -103850,19 +97908,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(761), + Line: int(746), Column: int(32), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7969, + Ctx: p7932, FreeVars: ast.Identifiers{ "code", }, @@ -103875,19 +97933,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(761), + Line: int(746), Column: int(38), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7969, + Ctx: p7932, FreeVars: ast.Identifiers{ "fw", }, @@ -103900,19 +97958,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(761), + Line: int(746), Column: int(42), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7969, + Ctx: p7932, FreeVars: ast.Identifiers{ "prec", }, @@ -103925,19 +97983,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(761), + Line: int(746), Column: int(48), }, End: ast.Location{ - Line: int(761), + Line: int(746), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7969, + Ctx: p7932, FreeVars: ast.Identifiers{ "f", }, @@ -103957,30 +98015,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(757), - Column: int(17), - }, - End: ast.Location{ - Line: int(761), - Column: int(50), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(762), + Line: int(747), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, @@ -103993,7 +98039,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "code", "codes", @@ -104015,13 +98061,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(763), + Line: int(748), Column: int(13), }, End: ast.Location{ - Line: int(766), + Line: int(751), Column: int(35), }, File: p1, @@ -104034,7 +98080,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7985, + Ctx: p7948, FreeVars: ast.Identifiers{ "code", "fw", @@ -104046,19 +98092,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(763), + Line: int(748), Column: int(16), }, End: ast.Location{ - Line: int(763), + Line: int(748), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7985, + Ctx: p7948, FreeVars: ast.Identifiers{ "code", }, @@ -104066,13 +98112,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(763), + Line: int(748), Column: int(16), }, End: ast.Location{ - Line: int(763), + Line: int(748), Column: int(27), }, File: p1, @@ -104086,13 +98132,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(763), + Line: int(748), Column: int(16), }, End: ast.Location{ - Line: int(763), + Line: int(748), Column: int(20), }, File: p1, @@ -104127,9 +98173,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "cflags", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, LeftBracketFodder: ast.Fodder{}, @@ -104154,28 +98199,27 @@ var _StdAst = &ast.DesugaredObject{ Value: "left", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(764), + Line: int(749), Column: int(15), }, End: ast.Location{ - Line: int(764), + Line: int(749), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7985, + Ctx: p7948, FreeVars: ast.Identifiers{ "fw", "pad_right", @@ -104185,13 +98229,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(764), + Line: int(749), Column: int(15), }, End: ast.Location{ - Line: int(764), + Line: int(749), Column: int(24), }, File: p1, @@ -104204,7 +98248,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7985, + Ctx: p7948, FreeVars: ast.Identifiers{ "pad_right", }, @@ -104218,19 +98262,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(764), + Line: int(749), Column: int(25), }, End: ast.Location{ - Line: int(764), + Line: int(749), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8002, + Ctx: p7965, FreeVars: ast.Identifiers{ "s", }, @@ -104243,19 +98287,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(764), + Line: int(749), Column: int(28), }, End: ast.Location{ - Line: int(764), + Line: int(749), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8002, + Ctx: p7965, FreeVars: ast.Identifiers{ "fw", }, @@ -104268,25 +98312,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(764), + Line: int(749), Column: int(32), }, End: ast.Location{ - Line: int(764), + Line: int(749), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8002, + Ctx: p7965, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -104309,19 +98352,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(766), + Line: int(751), Column: int(15), }, End: ast.Location{ - Line: int(766), + Line: int(751), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7985, + Ctx: p7948, FreeVars: ast.Identifiers{ "fw", "pad_left", @@ -104331,13 +98374,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(766), + Line: int(751), Column: int(15), }, End: ast.Location{ - Line: int(766), + Line: int(751), Column: int(23), }, File: p1, @@ -104350,7 +98393,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7985, + Ctx: p7948, FreeVars: ast.Identifiers{ "pad_left", }, @@ -104364,19 +98407,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(766), + Line: int(751), Column: int(24), }, End: ast.Location{ - Line: int(766), + Line: int(751), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8015, + Ctx: p7978, FreeVars: ast.Identifiers{ "s", }, @@ -104389,19 +98432,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(766), + Line: int(751), Column: int(27), }, End: ast.Location{ - Line: int(766), + Line: int(751), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8015, + Ctx: p7978, FreeVars: ast.Identifiers{ "fw", }, @@ -104414,25 +98457,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(766), + Line: int(751), Column: int(31), }, End: ast.Location{ - Line: int(766), + Line: int(751), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8015, + Ctx: p7978, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -104447,36 +98489,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(762), - Column: int(17), - }, - End: ast.Location{ - Line: int(766), - Column: int(35), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "codes", "format_codes_obj", @@ -104489,13 +98519,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(11), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(27), }, File: p1, @@ -104508,7 +98538,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p7756, + Ctx: p7719, FreeVars: ast.Identifiers{ "format_codes_obj", }, @@ -104522,19 +98552,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(28), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: ast.Identifiers{ "codes", }, @@ -104547,19 +98577,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(35), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: ast.Identifiers{ "obj", }, @@ -104572,19 +98602,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(40), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: ast.Identifiers{ "i", }, @@ -104592,19 +98622,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(40), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: ast.Identifiers{ "i", }, @@ -104616,21 +98646,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(44), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -104640,19 +98671,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(47), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: ast.Identifiers{ "s_padded", "v", @@ -104661,19 +98692,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(47), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: ast.Identifiers{ "v", }, @@ -104685,19 +98716,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(767), + Line: int(752), Column: int(51), }, End: ast.Location{ - Line: int(767), + Line: int(752), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8027, + Ctx: p7990, FreeVars: ast.Identifiers{ "s_padded", }, @@ -104727,30 +98758,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(769), + Line: int(754), Column: int(5), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -104775,13 +98794,13 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(769), + Line: int(754), Column: int(8), }, End: ast.Location{ - Line: int(769), + Line: int(754), Column: int(25), }, File: p1, @@ -104796,13 +98815,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(769), + Line: int(754), Column: int(8), }, End: ast.Location{ - Line: int(769), + Line: int(754), Column: int(19), }, File: p1, @@ -104816,13 +98835,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(769), + Line: int(754), Column: int(8), }, End: ast.Location{ - Line: int(769), + Line: int(754), Column: int(11), }, File: p1, @@ -104857,9 +98876,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -104869,19 +98887,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(769), + Line: int(754), Column: int(20), }, End: ast.Location{ - Line: int(769), + Line: int(754), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8054, + Ctx: p8017, FreeVars: ast.Identifiers{ "vals", }, @@ -104902,13 +98920,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(770), + Line: int(755), Column: int(7), }, End: ast.Location{ - Line: int(770), + Line: int(755), Column: int(46), }, File: p1, @@ -104924,13 +98942,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(770), + Line: int(755), Column: int(7), }, End: ast.Location{ - Line: int(770), + Line: int(755), Column: int(23), }, File: p1, @@ -104957,19 +98975,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(770), + Line: int(755), Column: int(24), }, End: ast.Location{ - Line: int(770), + Line: int(755), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8063, + Ctx: p8026, FreeVars: ast.Identifiers{ "codes", }, @@ -104982,19 +99000,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(770), + Line: int(755), Column: int(31), }, End: ast.Location{ - Line: int(770), + Line: int(755), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8063, + Ctx: p8026, FreeVars: ast.Identifiers{ "vals", }, @@ -105007,21 +99025,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(770), + Line: int(755), Column: int(37), }, End: ast.Location{ - Line: int(770), + Line: int(755), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8063, + Ctx: p8026, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -105030,21 +99049,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(770), + Line: int(755), Column: int(40), }, End: ast.Location{ - Line: int(770), + Line: int(755), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8063, + Ctx: p8026, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -105053,25 +99073,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(770), + Line: int(755), Column: int(43), }, End: ast.Location{ - Line: int(770), + Line: int(755), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8063, + Ctx: p8026, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -105094,13 +99113,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(771), + Line: int(756), Column: int(10), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -105118,13 +99137,13 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(771), + Line: int(756), Column: int(13), }, End: ast.Location{ - Line: int(771), + Line: int(756), Column: int(31), }, File: p1, @@ -105139,13 +99158,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(771), + Line: int(756), Column: int(13), }, End: ast.Location{ - Line: int(771), + Line: int(756), Column: int(25), }, File: p1, @@ -105159,13 +99178,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(771), + Line: int(756), Column: int(13), }, End: ast.Location{ - Line: int(771), + Line: int(756), Column: int(16), }, File: p1, @@ -105200,9 +99219,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -105212,19 +99230,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(771), + Line: int(756), Column: int(26), }, End: ast.Location{ - Line: int(771), + Line: int(756), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8082, + Ctx: p8045, FreeVars: ast.Identifiers{ "vals", }, @@ -105245,13 +99263,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(772), + Line: int(757), Column: int(7), }, End: ast.Location{ - Line: int(772), + Line: int(757), Column: int(43), }, File: p1, @@ -105267,13 +99285,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(772), + Line: int(757), Column: int(7), }, End: ast.Location{ - Line: int(772), + Line: int(757), Column: int(23), }, File: p1, @@ -105300,19 +99318,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(772), + Line: int(757), Column: int(24), }, End: ast.Location{ - Line: int(772), + Line: int(757), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8091, + Ctx: p8054, FreeVars: ast.Identifiers{ "codes", }, @@ -105325,19 +99343,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(772), + Line: int(757), Column: int(31), }, End: ast.Location{ - Line: int(772), + Line: int(757), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8091, + Ctx: p8054, FreeVars: ast.Identifiers{ "vals", }, @@ -105350,21 +99368,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(772), + Line: int(757), Column: int(37), }, End: ast.Location{ - Line: int(772), + Line: int(757), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8091, + Ctx: p8054, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -105373,25 +99392,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(772), + Line: int(757), Column: int(40), }, End: ast.Location{ - Line: int(772), + Line: int(757), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8091, + Ctx: p8054, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -105414,13 +99432,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(7), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(48), }, File: p1, @@ -105436,13 +99454,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(7), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(23), }, File: p1, @@ -105469,19 +99487,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(24), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8105, + Ctx: p8068, FreeVars: ast.Identifiers{ "codes", }, @@ -105494,19 +99512,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(31), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8105, + Ctx: p8068, FreeVars: ast.Identifiers{ "vals", }, @@ -105516,19 +99534,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(32), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8111, + Ctx: p8074, FreeVars: ast.Identifiers{ "vals", }, @@ -105547,21 +99565,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(39), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8105, + Ctx: p8068, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -105570,21 +99589,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(42), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8105, + Ctx: p8068, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -105593,25 +99613,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(774), + Line: int(759), Column: int(45), }, End: ast.Location{ - Line: int(774), + Line: int(759), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8105, + Ctx: p8068, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -105647,18 +99666,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(307), - Column: int(3), - }, - End: ast.Location{ - Line: int(774), - Column: int(48), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -105683,7 +99690,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "foldr", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -105706,77 +99712,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(776), - Column: int(9), - }, - End: ast.Location{ - Line: int(776), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(776), - Column: int(15), - }, - End: ast.Location{ - Line: int(776), - Column: int(18), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "init", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(776), - Column: int(20), - }, - End: ast.Location{ - Line: int(776), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "init", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(777), + Line: int(762), Column: int(5), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(46), }, File: p1, @@ -105789,7 +99756,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8122, + Ctx: p8085, FreeVars: ast.Identifiers{ "arr", "func", @@ -105805,114 +99772,61 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(777), + Line: int(762), Column: int(11), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(57), }, File: p1, }, Fodder: nil, - Ctx: p8126, + Ctx: p8089, FreeVars: ast.Identifiers{ "aux", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(777), - Column: int(15), - }, - End: ast.Location{ - Line: int(777), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(777), - Column: int(21), - }, - End: ast.Location{ - Line: int(777), - Column: int(24), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "running", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(777), - Column: int(26), - }, - End: ast.Location{ - Line: int(777), - Column: int(33), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "running", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "idx", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(777), - Column: int(35), - }, - End: ast.Location{ - Line: int(777), - Column: int(38), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "idx", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(778), + Line: int(763), Column: int(7), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(57), }, File: p1, @@ -105925,7 +99839,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8131, + Ctx: p8094, FreeVars: ast.Identifiers{ "arr", "aux", @@ -105937,19 +99851,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(778), + Line: int(763), Column: int(10), }, End: ast.Location{ - Line: int(778), + Line: int(763), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8131, + Ctx: p8094, FreeVars: ast.Identifiers{ "idx", }, @@ -105957,19 +99871,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(778), + Line: int(763), Column: int(10), }, End: ast.Location{ - Line: int(778), + Line: int(763), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8131, + Ctx: p8094, FreeVars: ast.Identifiers{ "idx", }, @@ -105981,21 +99895,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(778), + Line: int(763), Column: int(16), }, End: ast.Location{ - Line: int(778), + Line: int(763), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8131, + Ctx: p8094, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -106003,13 +99918,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(779), + Line: int(764), Column: int(9), }, End: ast.Location{ - Line: int(779), + Line: int(764), Column: int(16), }, File: p1, @@ -106022,7 +99937,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8131, + Ctx: p8094, FreeVars: ast.Identifiers{ "running", }, @@ -106040,19 +99955,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(9), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8131, + Ctx: p8094, FreeVars: ast.Identifiers{ "arr", "aux", @@ -106064,13 +99979,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(9), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(12), }, File: p1, @@ -106083,7 +99998,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8131, + Ctx: p8094, FreeVars: ast.Identifiers{ "aux", }, @@ -106097,19 +100012,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(13), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8149, + Ctx: p8112, FreeVars: ast.Identifiers{ "func", }, @@ -106122,19 +100037,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(19), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8149, + Ctx: p8112, FreeVars: ast.Identifiers{ "arr", }, @@ -106147,19 +100062,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(24), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8149, + Ctx: p8112, FreeVars: ast.Identifiers{ "arr", "func", @@ -106170,19 +100085,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(24), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8149, + Ctx: p8112, FreeVars: ast.Identifiers{ "func", }, @@ -106196,19 +100111,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(29), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8159, + Ctx: p8122, FreeVars: ast.Identifiers{ "arr", "idx", @@ -106217,19 +100132,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(29), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8159, + Ctx: p8122, FreeVars: ast.Identifiers{ "arr", }, @@ -106240,19 +100155,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(33), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8159, + Ctx: p8122, FreeVars: ast.Identifiers{ "idx", }, @@ -106268,19 +100183,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(39), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8159, + Ctx: p8122, FreeVars: ast.Identifiers{ "running", }, @@ -106303,19 +100218,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(49), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8149, + Ctx: p8112, FreeVars: ast.Identifiers{ "idx", }, @@ -106323,19 +100238,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(49), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8149, + Ctx: p8112, FreeVars: ast.Identifiers{ "idx", }, @@ -106347,21 +100262,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(781), + Line: int(766), Column: int(55), }, End: ast.Location{ - Line: int(781), + Line: int(766), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8149, + Ctx: p8112, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -106379,36 +100295,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(5), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8122, + Ctx: p8085, FreeVars: ast.Identifiers{ "arr", "aux", @@ -106420,13 +100324,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(5), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(8), }, File: p1, @@ -106439,7 +100343,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8122, + Ctx: p8085, FreeVars: ast.Identifiers{ "aux", }, @@ -106453,19 +100357,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(9), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8179, + Ctx: p8142, FreeVars: ast.Identifiers{ "func", }, @@ -106478,19 +100382,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(15), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8179, + Ctx: p8142, FreeVars: ast.Identifiers{ "arr", }, @@ -106503,19 +100407,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(20), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8179, + Ctx: p8142, FreeVars: ast.Identifiers{ "init", }, @@ -106528,19 +100432,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(26), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8179, + Ctx: p8142, FreeVars: ast.Identifiers{ "arr", "std", @@ -106549,19 +100453,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(26), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8179, + Ctx: p8142, FreeVars: ast.Identifiers{ "arr", "std", @@ -106570,19 +100474,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(26), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8179, + Ctx: p8142, FreeVars: ast.Identifiers{ "std", }, @@ -106590,13 +100494,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(26), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(29), }, File: p1, @@ -106631,9 +100535,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -106643,19 +100546,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(37), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8196, + Ctx: p8159, FreeVars: ast.Identifiers{ "arr", }, @@ -106677,21 +100580,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(782), + Line: int(767), Column: int(44), }, End: ast.Location{ - Line: int(782), + Line: int(767), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8179, + Ctx: p8142, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -106708,18 +100612,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(776), - Column: int(3), - }, - End: ast.Location{ - Line: int(782), - Column: int(46), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -106744,7 +100636,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "foldl", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -106767,77 +100658,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(784), - Column: int(9), - }, - End: ast.Location{ - Line: int(784), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(784), - Column: int(15), - }, - End: ast.Location{ - Line: int(784), - Column: int(18), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "init", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(784), - Column: int(20), - }, - End: ast.Location{ - Line: int(784), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "init", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(785), + Line: int(770), Column: int(5), }, End: ast.Location{ - Line: int(790), + Line: int(775), Column: int(28), }, File: p1, @@ -106850,7 +100702,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8205, + Ctx: p8168, FreeVars: ast.Identifiers{ "arr", "func", @@ -106866,115 +100718,62 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(785), + Line: int(770), Column: int(11), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(57), }, File: p1, }, Fodder: nil, - Ctx: p8209, + Ctx: p8172, FreeVars: ast.Identifiers{ "aux", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(785), - Column: int(15), - }, - End: ast.Location{ - Line: int(785), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "func", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(785), - Column: int(21), - }, - End: ast.Location{ - Line: int(785), - Column: int(24), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "running", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(785), - Column: int(26), - }, - End: ast.Location{ - Line: int(785), - Column: int(33), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "running", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "idx", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(785), - Column: int(35), - }, - End: ast.Location{ - Line: int(785), - Column: int(38), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "idx", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(786), + Line: int(771), Column: int(7), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(57), }, File: p1, @@ -106987,7 +100786,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "arr", "aux", @@ -107000,19 +100799,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(786), + Line: int(771), Column: int(10), }, End: ast.Location{ - Line: int(786), + Line: int(771), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "arr", "idx", @@ -107022,19 +100821,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(786), + Line: int(771), Column: int(10), }, End: ast.Location{ - Line: int(786), + Line: int(771), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "idx", }, @@ -107046,19 +100845,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(786), + Line: int(771), Column: int(17), }, End: ast.Location{ - Line: int(786), + Line: int(771), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "arr", "std", @@ -107067,19 +100866,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(786), + Line: int(771), Column: int(17), }, End: ast.Location{ - Line: int(786), + Line: int(771), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "std", }, @@ -107087,13 +100886,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(786), + Line: int(771), Column: int(17), }, End: ast.Location{ - Line: int(786), + Line: int(771), Column: int(20), }, File: p1, @@ -107128,9 +100927,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -107140,19 +100938,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(786), + Line: int(771), Column: int(28), }, End: ast.Location{ - Line: int(786), + Line: int(771), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8229, + Ctx: p8192, FreeVars: ast.Identifiers{ "arr", }, @@ -107174,13 +100972,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(787), + Line: int(772), Column: int(9), }, End: ast.Location{ - Line: int(787), + Line: int(772), Column: int(16), }, File: p1, @@ -107193,7 +100991,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "running", }, @@ -107211,19 +101009,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(9), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "arr", "aux", @@ -107235,13 +101033,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(9), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(12), }, File: p1, @@ -107254,7 +101052,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8214, + Ctx: p8177, FreeVars: ast.Identifiers{ "aux", }, @@ -107268,19 +101066,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(13), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8242, + Ctx: p8205, FreeVars: ast.Identifiers{ "func", }, @@ -107293,19 +101091,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(19), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8242, + Ctx: p8205, FreeVars: ast.Identifiers{ "arr", }, @@ -107318,19 +101116,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(24), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8242, + Ctx: p8205, FreeVars: ast.Identifiers{ "arr", "func", @@ -107341,19 +101139,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(24), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8242, + Ctx: p8205, FreeVars: ast.Identifiers{ "func", }, @@ -107367,19 +101165,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(29), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8252, + Ctx: p8215, FreeVars: ast.Identifiers{ "running", }, @@ -107392,19 +101190,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(38), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8252, + Ctx: p8215, FreeVars: ast.Identifiers{ "arr", "idx", @@ -107413,19 +101211,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(38), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8252, + Ctx: p8215, FreeVars: ast.Identifiers{ "arr", }, @@ -107436,19 +101234,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(42), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8252, + Ctx: p8215, FreeVars: ast.Identifiers{ "idx", }, @@ -107474,19 +101272,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(49), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8242, + Ctx: p8205, FreeVars: ast.Identifiers{ "idx", }, @@ -107494,19 +101292,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(49), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8242, + Ctx: p8205, FreeVars: ast.Identifiers{ "idx", }, @@ -107518,21 +101316,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(789), + Line: int(774), Column: int(55), }, End: ast.Location{ - Line: int(789), + Line: int(774), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8242, + Ctx: p8205, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -107550,36 +101349,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(790), + Line: int(775), Column: int(5), }, End: ast.Location{ - Line: int(790), + Line: int(775), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8205, + Ctx: p8168, FreeVars: ast.Identifiers{ "arr", "aux", @@ -107590,13 +101377,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(790), + Line: int(775), Column: int(5), }, End: ast.Location{ - Line: int(790), + Line: int(775), Column: int(8), }, File: p1, @@ -107609,7 +101396,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8205, + Ctx: p8168, FreeVars: ast.Identifiers{ "aux", }, @@ -107623,19 +101410,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(790), + Line: int(775), Column: int(9), }, End: ast.Location{ - Line: int(790), + Line: int(775), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8272, + Ctx: p8235, FreeVars: ast.Identifiers{ "func", }, @@ -107648,19 +101435,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(790), + Line: int(775), Column: int(15), }, End: ast.Location{ - Line: int(790), + Line: int(775), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8272, + Ctx: p8235, FreeVars: ast.Identifiers{ "arr", }, @@ -107673,19 +101460,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(790), + Line: int(775), Column: int(20), }, End: ast.Location{ - Line: int(790), + Line: int(775), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8272, + Ctx: p8235, FreeVars: ast.Identifiers{ "init", }, @@ -107698,21 +101485,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(790), + Line: int(775), Column: int(26), }, End: ast.Location{ - Line: int(790), + Line: int(775), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8272, + Ctx: p8235, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: nil, @@ -107728,18 +101516,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(784), - Column: int(3), - }, - End: ast.Location{ - Line: int(790), - Column: int(28), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -107764,7 +101540,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "filterMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -107787,77 +101562,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "filter_func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(793), - Column: int(13), - }, - End: ast.Location{ - Line: int(793), - Column: int(24), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "map_func", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(793), - Column: int(26), - }, - End: ast.Location{ - Line: int(793), - Column: int(34), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(793), - Column: int(36), - }, - End: ast.Location{ - Line: int(793), - Column: int(39), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "filter_func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "map_func", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(794), + Line: int(779), Column: int(5), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(54), }, File: p1, @@ -107870,7 +101606,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "filter_func", @@ -107881,19 +101617,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(794), + Line: int(779), Column: int(8), }, End: ast.Location{ - Line: int(794), + Line: int(779), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "filter_func", "std", @@ -107903,19 +101639,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(794), + Line: int(779), Column: int(9), }, End: ast.Location{ - Line: int(794), + Line: int(779), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "filter_func", "std", @@ -107924,19 +101660,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(794), + Line: int(779), Column: int(9), }, End: ast.Location{ - Line: int(794), + Line: int(779), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "std", }, @@ -107944,13 +101680,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(794), + Line: int(779), Column: int(9), }, End: ast.Location{ - Line: int(794), + Line: int(779), Column: int(12), }, File: p1, @@ -107985,9 +101721,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -107997,19 +101732,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(794), + Line: int(779), Column: int(24), }, End: ast.Location{ - Line: int(794), + Line: int(779), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8298, + Ctx: p8261, FreeVars: ast.Identifiers{ "filter_func", }, @@ -108031,13 +101766,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(795), + Line: int(780), Column: int(7), }, End: ast.Location{ - Line: int(795), + Line: int(780), Column: int(89), }, File: p1, @@ -108050,7 +101785,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "filter_func", "std", @@ -108059,19 +101794,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(795), + Line: int(780), Column: int(14), }, End: ast.Location{ - Line: int(795), + Line: int(780), Column: int(88), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "filter_func", "std", @@ -108080,44 +101815,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(795), + Line: int(780), Column: int(14), }, End: ast.Location{ - Line: int(795), + Line: int(780), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: nil, }, Value: "std.filterMap first param must be function, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(795), + Line: int(780), Column: int(67), }, End: ast.Location{ - Line: int(795), + Line: int(780), Column: int(88), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "filter_func", "std", @@ -108126,19 +101860,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(795), + Line: int(780), Column: int(67), }, End: ast.Location{ - Line: int(795), + Line: int(780), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "std", }, @@ -108146,13 +101880,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(795), + Line: int(780), Column: int(67), }, End: ast.Location{ - Line: int(795), + Line: int(780), Column: int(70), }, File: p1, @@ -108187,9 +101921,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -108199,19 +101932,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(795), + Line: int(780), Column: int(76), }, End: ast.Location{ - Line: int(795), + Line: int(780), Column: int(87), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8315, + Ctx: p8278, FreeVars: ast.Identifiers{ "filter_func", }, @@ -108241,19 +101974,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(796), + Line: int(781), Column: int(10), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "filter_func", @@ -108264,19 +101997,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(796), + Line: int(781), Column: int(13), }, End: ast.Location{ - Line: int(796), + Line: int(781), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "map_func", "std", @@ -108286,19 +102019,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(796), + Line: int(781), Column: int(14), }, End: ast.Location{ - Line: int(796), + Line: int(781), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "map_func", "std", @@ -108307,19 +102040,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(796), + Line: int(781), Column: int(14), }, End: ast.Location{ - Line: int(796), + Line: int(781), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "std", }, @@ -108327,13 +102060,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(796), + Line: int(781), Column: int(14), }, End: ast.Location{ - Line: int(796), + Line: int(781), Column: int(17), }, File: p1, @@ -108368,9 +102101,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -108380,19 +102112,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(796), + Line: int(781), Column: int(29), }, End: ast.Location{ - Line: int(796), + Line: int(781), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8331, + Ctx: p8294, FreeVars: ast.Identifiers{ "map_func", }, @@ -108414,13 +102146,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(797), + Line: int(782), Column: int(7), }, End: ast.Location{ - Line: int(797), + Line: int(782), Column: int(87), }, File: p1, @@ -108433,7 +102165,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "map_func", "std", @@ -108442,19 +102174,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(797), + Line: int(782), Column: int(14), }, End: ast.Location{ - Line: int(797), + Line: int(782), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "map_func", "std", @@ -108463,44 +102195,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(797), + Line: int(782), Column: int(14), }, End: ast.Location{ - Line: int(797), + Line: int(782), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: nil, }, Value: "std.filterMap second param must be function, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(797), + Line: int(782), Column: int(68), }, End: ast.Location{ - Line: int(797), + Line: int(782), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "map_func", "std", @@ -108509,19 +102240,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(797), + Line: int(782), Column: int(68), }, End: ast.Location{ - Line: int(797), + Line: int(782), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "std", }, @@ -108529,13 +102260,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(797), + Line: int(782), Column: int(68), }, End: ast.Location{ - Line: int(797), + Line: int(782), Column: int(71), }, File: p1, @@ -108570,9 +102301,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -108582,19 +102312,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(797), + Line: int(782), Column: int(77), }, End: ast.Location{ - Line: int(797), + Line: int(782), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8348, + Ctx: p8311, FreeVars: ast.Identifiers{ "map_func", }, @@ -108624,19 +102354,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(798), + Line: int(783), Column: int(10), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "filter_func", @@ -108647,19 +102377,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(798), + Line: int(783), Column: int(13), }, End: ast.Location{ - Line: int(798), + Line: int(783), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "std", @@ -108669,19 +102399,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(798), + Line: int(783), Column: int(14), }, End: ast.Location{ - Line: int(798), + Line: int(783), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "std", @@ -108690,19 +102420,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(798), + Line: int(783), Column: int(14), }, End: ast.Location{ - Line: int(798), + Line: int(783), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "std", }, @@ -108710,13 +102440,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(798), + Line: int(783), Column: int(14), }, End: ast.Location{ - Line: int(798), + Line: int(783), Column: int(17), }, File: p1, @@ -108751,9 +102481,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -108763,19 +102492,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(798), + Line: int(783), Column: int(26), }, End: ast.Location{ - Line: int(798), + Line: int(783), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8364, + Ctx: p8327, FreeVars: ast.Identifiers{ "arr", }, @@ -108797,13 +102526,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(799), + Line: int(784), Column: int(7), }, End: ast.Location{ - Line: int(799), + Line: int(784), Column: int(78), }, File: p1, @@ -108816,7 +102545,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "std", @@ -108825,19 +102554,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(799), + Line: int(784), Column: int(14), }, End: ast.Location{ - Line: int(799), + Line: int(784), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "std", @@ -108846,44 +102575,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(799), + Line: int(784), Column: int(14), }, End: ast.Location{ - Line: int(799), + Line: int(784), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: nil, }, Value: "std.filterMap third param must be array, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(799), + Line: int(784), Column: int(64), }, End: ast.Location{ - Line: int(799), + Line: int(784), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "std", @@ -108892,19 +102620,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(799), + Line: int(784), Column: int(64), }, End: ast.Location{ - Line: int(799), + Line: int(784), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "std", }, @@ -108912,13 +102640,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(799), + Line: int(784), Column: int(64), }, End: ast.Location{ - Line: int(799), + Line: int(784), Column: int(67), }, File: p1, @@ -108953,9 +102681,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -108965,19 +102692,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(799), + Line: int(784), Column: int(73), }, End: ast.Location{ - Line: int(799), + Line: int(784), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8381, + Ctx: p8344, FreeVars: ast.Identifiers{ "arr", }, @@ -109007,19 +102734,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(7), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "arr", "filter_func", @@ -109030,19 +102757,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(7), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8285, + Ctx: p8248, FreeVars: ast.Identifiers{ "std", }, @@ -109050,13 +102777,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(7), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(10), }, File: p1, @@ -109098,9 +102825,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "map", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -109110,19 +102836,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(15), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8394, + Ctx: p8357, FreeVars: ast.Identifiers{ "map_func", }, @@ -109135,19 +102861,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(25), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8394, + Ctx: p8357, FreeVars: ast.Identifiers{ "arr", "filter_func", @@ -109157,19 +102883,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(25), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8394, + Ctx: p8357, FreeVars: ast.Identifiers{ "std", }, @@ -109177,13 +102903,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(25), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(28), }, File: p1, @@ -109218,9 +102944,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "filter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -109230,19 +102955,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(36), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8405, + Ctx: p8368, FreeVars: ast.Identifiers{ "filter_func", }, @@ -109255,19 +102980,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(801), + Line: int(786), Column: int(49), }, End: ast.Location{ - Line: int(801), + Line: int(786), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8405, + Ctx: p8368, FreeVars: ast.Identifiers{ "arr", }, @@ -109299,18 +103024,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(793), - Column: int(3), - }, - End: ast.Location{ - Line: int(801), - Column: int(54), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -109335,7 +103048,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "assertEqual", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -109356,58 +103068,33 @@ var _StdAst = &ast.DesugaredObject{ FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(803), - Column: int(15), - }, - End: ast.Location{ - Line: int(803), - Column: int(16), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(803), - Column: int(18), - }, - End: ast.Location{ - Line: int(803), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(804), + Line: int(789), Column: int(5), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(50), }, File: p1, @@ -109420,7 +103107,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", "b", @@ -109429,19 +103116,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(804), + Line: int(789), Column: int(8), }, End: ast.Location{ - Line: int(804), + Line: int(789), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", "b", @@ -109450,19 +103137,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(804), + Line: int(789), Column: int(8), }, End: ast.Location{ - Line: int(804), + Line: int(789), Column: int(9), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", }, @@ -109474,19 +103161,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(804), + Line: int(789), Column: int(13), }, End: ast.Location{ - Line: int(804), + Line: int(789), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "b", }, @@ -109498,13 +103185,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(805), + Line: int(790), Column: int(7), }, End: ast.Location{ - Line: int(805), + Line: int(790), Column: int(11), }, File: p1, @@ -109517,7 +103204,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8414, + Ctx: p8377, FreeVars: nil, }, Value: true, @@ -109533,13 +103220,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(7), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(50), }, File: p1, @@ -109552,7 +103239,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", "b", @@ -109561,19 +103248,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(13), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", "b", @@ -109582,19 +103269,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(13), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", }, @@ -109602,19 +103289,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(13), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", }, @@ -109622,44 +103309,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(13), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: nil, }, Value: "Assertion failed. ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(36), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "a", }, @@ -109672,25 +103358,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(40), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: nil, }, Value: " != ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -109698,19 +103383,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(807), + Line: int(792), Column: int(49), }, End: ast.Location{ - Line: int(807), + Line: int(792), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8414, + Ctx: p8377, FreeVars: ast.Identifiers{ "b", }, @@ -109722,18 +103407,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(803), - Column: int(3), - }, - End: ast.Location{ - Line: int(807), - Column: int(50), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -109758,7 +103431,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "abs", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -109781,39 +103453,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "n", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(809), - Column: int(7), - }, - End: ast.Location{ - Line: int(809), - Column: int(8), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(810), + Line: int(795), Column: int(5), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(30), }, File: p1, @@ -109826,7 +103487,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", "std", @@ -109835,19 +103496,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(810), + Line: int(795), Column: int(8), }, End: ast.Location{ - Line: int(810), + Line: int(795), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", "std", @@ -109857,19 +103518,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(810), + Line: int(795), Column: int(9), }, End: ast.Location{ - Line: int(810), + Line: int(795), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", "std", @@ -109878,19 +103539,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(810), + Line: int(795), Column: int(9), }, End: ast.Location{ - Line: int(810), + Line: int(795), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "std", }, @@ -109898,13 +103559,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(810), + Line: int(795), Column: int(9), }, End: ast.Location{ - Line: int(810), + Line: int(795), Column: int(12), }, File: p1, @@ -109939,9 +103600,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -109951,19 +103611,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(810), + Line: int(795), Column: int(22), }, End: ast.Location{ - Line: int(810), + Line: int(795), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8459, + Ctx: p8422, FreeVars: ast.Identifiers{ "n", }, @@ -109985,13 +103645,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(811), + Line: int(796), Column: int(7), }, End: ast.Location{ - Line: int(811), + Line: int(796), Column: int(58), }, File: p1, @@ -110004,7 +103664,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", "std", @@ -110013,19 +103673,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(811), + Line: int(796), Column: int(13), }, End: ast.Location{ - Line: int(811), + Line: int(796), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", "std", @@ -110034,44 +103694,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(811), + Line: int(796), Column: int(13), }, End: ast.Location{ - Line: int(811), + Line: int(796), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: nil, }, Value: "std.abs expected number, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(811), + Line: int(796), Column: int(47), }, End: ast.Location{ - Line: int(811), + Line: int(796), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", "std", @@ -110080,19 +103739,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(811), + Line: int(796), Column: int(47), }, End: ast.Location{ - Line: int(811), + Line: int(796), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "std", }, @@ -110100,13 +103759,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(811), + Line: int(796), Column: int(47), }, End: ast.Location{ - Line: int(811), + Line: int(796), Column: int(50), }, File: p1, @@ -110141,9 +103800,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -110153,19 +103811,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(811), + Line: int(796), Column: int(56), }, End: ast.Location{ - Line: int(811), + Line: int(796), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8476, + Ctx: p8439, FreeVars: ast.Identifiers{ "n", }, @@ -110195,13 +103853,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(813), + Line: int(798), Column: int(7), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(30), }, File: p1, @@ -110214,7 +103872,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", }, @@ -110222,19 +103880,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(813), + Line: int(798), Column: int(10), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", }, @@ -110242,19 +103900,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(813), + Line: int(798), Column: int(10), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", }, @@ -110266,21 +103924,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(813), + Line: int(798), Column: int(14), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -110288,19 +103947,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(813), + Line: int(798), Column: int(21), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", }, @@ -110311,19 +103970,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(813), + Line: int(798), Column: int(28), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", }, @@ -110332,19 +103991,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(813), + Line: int(798), Column: int(29), }, End: ast.Location{ - Line: int(813), + Line: int(798), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8446, + Ctx: p8409, FreeVars: ast.Identifiers{ "n", }, @@ -110356,18 +104015,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(809), - Column: int(3), - }, - End: ast.Location{ - Line: int(813), - Column: int(30), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -110392,7 +104039,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "sign", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -110415,39 +104061,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "n", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(815), - Column: int(8), - }, - End: ast.Location{ - Line: int(815), - Column: int(9), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "n", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(816), + Line: int(801), Column: int(5), }, End: ast.Location{ - Line: int(823), + Line: int(808), Column: int(13), }, File: p1, @@ -110460,7 +104095,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", "std", @@ -110469,19 +104104,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(816), + Line: int(801), Column: int(8), }, End: ast.Location{ - Line: int(816), + Line: int(801), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", "std", @@ -110491,19 +104126,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(816), + Line: int(801), Column: int(9), }, End: ast.Location{ - Line: int(816), + Line: int(801), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", "std", @@ -110512,19 +104147,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(816), + Line: int(801), Column: int(9), }, End: ast.Location{ - Line: int(816), + Line: int(801), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "std", }, @@ -110532,13 +104167,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(816), + Line: int(801), Column: int(9), }, End: ast.Location{ - Line: int(816), + Line: int(801), Column: int(12), }, File: p1, @@ -110573,9 +104208,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -110585,19 +104219,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(816), + Line: int(801), Column: int(22), }, End: ast.Location{ - Line: int(816), + Line: int(801), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8512, + Ctx: p8475, FreeVars: ast.Identifiers{ "n", }, @@ -110619,13 +104253,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(817), + Line: int(802), Column: int(7), }, End: ast.Location{ - Line: int(817), + Line: int(802), Column: int(59), }, File: p1, @@ -110638,7 +104272,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", "std", @@ -110647,19 +104281,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(817), + Line: int(802), Column: int(13), }, End: ast.Location{ - Line: int(817), + Line: int(802), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", "std", @@ -110668,44 +104302,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(817), + Line: int(802), Column: int(13), }, End: ast.Location{ - Line: int(817), + Line: int(802), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: nil, }, Value: "std.sign expected number, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(817), + Line: int(802), Column: int(48), }, End: ast.Location{ - Line: int(817), + Line: int(802), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", "std", @@ -110714,19 +104347,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(817), + Line: int(802), Column: int(48), }, End: ast.Location{ - Line: int(817), + Line: int(802), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "std", }, @@ -110734,13 +104367,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(817), + Line: int(802), Column: int(48), }, End: ast.Location{ - Line: int(817), + Line: int(802), Column: int(51), }, File: p1, @@ -110775,9 +104408,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -110787,19 +104419,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(817), + Line: int(802), Column: int(57), }, End: ast.Location{ - Line: int(817), + Line: int(802), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8529, + Ctx: p8492, FreeVars: ast.Identifiers{ "n", }, @@ -110829,13 +104461,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(819), + Line: int(804), Column: int(7), }, End: ast.Location{ - Line: int(823), + Line: int(808), Column: int(13), }, File: p1, @@ -110848,7 +104480,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", }, @@ -110856,19 +104488,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(819), + Line: int(804), Column: int(10), }, End: ast.Location{ - Line: int(819), + Line: int(804), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", }, @@ -110876,19 +104508,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(819), + Line: int(804), Column: int(10), }, End: ast.Location{ - Line: int(819), + Line: int(804), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", }, @@ -110900,21 +104532,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(819), + Line: int(804), Column: int(14), }, End: ast.Location{ - Line: int(819), + Line: int(804), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -110922,13 +104555,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(820), + Line: int(805), Column: int(9), }, End: ast.Location{ - Line: int(820), + Line: int(805), Column: int(10), }, File: p1, @@ -110941,9 +104574,10 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8499, + Ctx: p8462, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, ElseFodder: ast.Fodder{ @@ -110957,19 +104591,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(821), + Line: int(806), Column: int(12), }, End: ast.Location{ - Line: int(823), + Line: int(808), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", }, @@ -110977,19 +104611,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(821), + Line: int(806), Column: int(15), }, End: ast.Location{ - Line: int(821), + Line: int(806), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", }, @@ -110997,19 +104631,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(821), + Line: int(806), Column: int(15), }, End: ast.Location{ - Line: int(821), + Line: int(806), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: ast.Identifiers{ "n", }, @@ -111021,21 +104655,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(821), + Line: int(806), Column: int(19), }, End: ast.Location{ - Line: int(821), + Line: int(806), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -111043,13 +104678,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(822), + Line: int(807), Column: int(9), }, End: ast.Location{ - Line: int(822), + Line: int(807), Column: int(11), }, File: p1, @@ -111062,28 +104697,29 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8499, + Ctx: p8462, FreeVars: nil, }, Op: ast.UnaryOp(3), Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(822), + Line: int(807), Column: int(10), }, End: ast.Location{ - Line: int(822), + Line: int(807), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -111098,21 +104734,22 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(823), + Line: int(808), Column: int(12), }, End: ast.Location{ - Line: int(823), + Line: int(808), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8499, + Ctx: p8462, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -111120,18 +104757,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(815), - Column: int(3), - }, - End: ast.Location{ - Line: int(823), - Column: int(13), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -111156,7 +104781,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "max", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -111179,58 +104803,33 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(825), - Column: int(7), - }, - End: ast.Location{ - Line: int(825), - Column: int(8), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(825), - Column: int(10), - }, - End: ast.Location{ - Line: int(825), - Column: int(11), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(826), + Line: int(811), Column: int(5), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(29), }, File: p1, @@ -111243,7 +104842,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "b", @@ -111253,19 +104852,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(826), + Line: int(811), Column: int(8), }, End: ast.Location{ - Line: int(826), + Line: int(811), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "std", @@ -111275,19 +104874,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(826), + Line: int(811), Column: int(9), }, End: ast.Location{ - Line: int(826), + Line: int(811), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "std", @@ -111296,19 +104895,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(826), + Line: int(811), Column: int(9), }, End: ast.Location{ - Line: int(826), + Line: int(811), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "std", }, @@ -111316,13 +104915,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(826), + Line: int(811), Column: int(9), }, End: ast.Location{ - Line: int(826), + Line: int(811), Column: int(12), }, File: p1, @@ -111357,9 +104956,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -111369,19 +104967,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(826), + Line: int(811), Column: int(22), }, End: ast.Location{ - Line: int(826), + Line: int(811), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8574, + Ctx: p8537, FreeVars: ast.Identifiers{ "a", }, @@ -111403,13 +105001,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(827), + Line: int(812), Column: int(7), }, End: ast.Location{ - Line: int(827), + Line: int(812), Column: int(70), }, File: p1, @@ -111422,7 +105020,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "std", @@ -111431,19 +105029,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(827), + Line: int(812), Column: int(13), }, End: ast.Location{ - Line: int(827), + Line: int(812), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "std", @@ -111452,44 +105050,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(827), + Line: int(812), Column: int(13), }, End: ast.Location{ - Line: int(827), + Line: int(812), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: nil, }, Value: "std.max first param expected number, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(827), + Line: int(812), Column: int(59), }, End: ast.Location{ - Line: int(827), + Line: int(812), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "std", @@ -111498,19 +105095,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(827), + Line: int(812), Column: int(59), }, End: ast.Location{ - Line: int(827), + Line: int(812), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "std", }, @@ -111518,13 +105115,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(827), + Line: int(812), Column: int(59), }, End: ast.Location{ - Line: int(827), + Line: int(812), Column: int(62), }, File: p1, @@ -111559,9 +105156,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -111571,19 +105167,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(827), + Line: int(812), Column: int(68), }, End: ast.Location{ - Line: int(827), + Line: int(812), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8591, + Ctx: p8554, FreeVars: ast.Identifiers{ "a", }, @@ -111613,19 +105209,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(828), + Line: int(813), Column: int(10), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "b", @@ -111635,19 +105231,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(828), + Line: int(813), Column: int(13), }, End: ast.Location{ - Line: int(828), + Line: int(813), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "b", "std", @@ -111657,19 +105253,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(828), + Line: int(813), Column: int(14), }, End: ast.Location{ - Line: int(828), + Line: int(813), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "b", "std", @@ -111678,19 +105274,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(828), + Line: int(813), Column: int(14), }, End: ast.Location{ - Line: int(828), + Line: int(813), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "std", }, @@ -111698,13 +105294,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(828), + Line: int(813), Column: int(14), }, End: ast.Location{ - Line: int(828), + Line: int(813), Column: int(17), }, File: p1, @@ -111739,9 +105335,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -111751,19 +105346,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(828), + Line: int(813), Column: int(27), }, End: ast.Location{ - Line: int(828), + Line: int(813), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8607, + Ctx: p8570, FreeVars: ast.Identifiers{ "b", }, @@ -111785,13 +105380,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(829), + Line: int(814), Column: int(7), }, End: ast.Location{ - Line: int(829), + Line: int(814), Column: int(71), }, File: p1, @@ -111804,7 +105399,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "b", "std", @@ -111813,19 +105408,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(829), + Line: int(814), Column: int(13), }, End: ast.Location{ - Line: int(829), + Line: int(814), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "b", "std", @@ -111834,44 +105429,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(829), + Line: int(814), Column: int(13), }, End: ast.Location{ - Line: int(829), + Line: int(814), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: nil, }, Value: "std.max second param expected number, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(829), + Line: int(814), Column: int(60), }, End: ast.Location{ - Line: int(829), + Line: int(814), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "b", "std", @@ -111880,19 +105474,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(829), + Line: int(814), Column: int(60), }, End: ast.Location{ - Line: int(829), + Line: int(814), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "std", }, @@ -111900,13 +105494,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(829), + Line: int(814), Column: int(60), }, End: ast.Location{ - Line: int(829), + Line: int(814), Column: int(63), }, File: p1, @@ -111941,9 +105535,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -111953,19 +105546,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(829), + Line: int(814), Column: int(69), }, End: ast.Location{ - Line: int(829), + Line: int(814), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8624, + Ctx: p8587, FreeVars: ast.Identifiers{ "b", }, @@ -111995,13 +105588,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(831), + Line: int(816), Column: int(7), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(29), }, File: p1, @@ -112014,7 +105607,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "b", @@ -112023,19 +105616,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(831), + Line: int(816), Column: int(10), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", "b", @@ -112044,19 +105637,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(831), + Line: int(816), Column: int(10), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", }, @@ -112068,19 +105661,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(831), + Line: int(816), Column: int(14), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "b", }, @@ -112092,19 +105685,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(831), + Line: int(816), Column: int(21), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "a", }, @@ -112115,19 +105708,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(831), + Line: int(816), Column: int(28), }, End: ast.Location{ - Line: int(831), + Line: int(816), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8561, + Ctx: p8524, FreeVars: ast.Identifiers{ "b", }, @@ -112139,18 +105732,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(825), - Column: int(3), - }, - End: ast.Location{ - Line: int(831), - Column: int(29), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -112175,7 +105756,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "min", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -112198,58 +105778,33 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(833), - Column: int(7), - }, - End: ast.Location{ - Line: int(833), - Column: int(8), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(833), - Column: int(10), - }, - End: ast.Location{ - Line: int(833), - Column: int(11), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(834), + Line: int(819), Column: int(5), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(29), }, File: p1, @@ -112262,7 +105817,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "b", @@ -112272,19 +105827,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(834), + Line: int(819), Column: int(8), }, End: ast.Location{ - Line: int(834), + Line: int(819), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "std", @@ -112294,19 +105849,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(834), + Line: int(819), Column: int(9), }, End: ast.Location{ - Line: int(834), + Line: int(819), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "std", @@ -112315,19 +105870,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(834), + Line: int(819), Column: int(9), }, End: ast.Location{ - Line: int(834), + Line: int(819), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "std", }, @@ -112335,13 +105890,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(834), + Line: int(819), Column: int(9), }, End: ast.Location{ - Line: int(834), + Line: int(819), Column: int(12), }, File: p1, @@ -112376,9 +105931,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -112388,19 +105942,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(834), + Line: int(819), Column: int(22), }, End: ast.Location{ - Line: int(834), + Line: int(819), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8659, + Ctx: p8622, FreeVars: ast.Identifiers{ "a", }, @@ -112422,13 +105976,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(835), + Line: int(820), Column: int(7), }, End: ast.Location{ - Line: int(835), + Line: int(820), Column: int(70), }, File: p1, @@ -112441,7 +105995,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "std", @@ -112450,19 +106004,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(835), + Line: int(820), Column: int(13), }, End: ast.Location{ - Line: int(835), + Line: int(820), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "std", @@ -112471,44 +106025,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(835), + Line: int(820), Column: int(13), }, End: ast.Location{ - Line: int(835), + Line: int(820), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: nil, }, Value: "std.max first param expected number, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(835), + Line: int(820), Column: int(59), }, End: ast.Location{ - Line: int(835), + Line: int(820), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "std", @@ -112517,19 +106070,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(835), + Line: int(820), Column: int(59), }, End: ast.Location{ - Line: int(835), + Line: int(820), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "std", }, @@ -112537,13 +106090,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(835), + Line: int(820), Column: int(59), }, End: ast.Location{ - Line: int(835), + Line: int(820), Column: int(62), }, File: p1, @@ -112578,9 +106131,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -112590,19 +106142,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(835), + Line: int(820), Column: int(68), }, End: ast.Location{ - Line: int(835), + Line: int(820), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8676, + Ctx: p8639, FreeVars: ast.Identifiers{ "a", }, @@ -112632,19 +106184,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(836), + Line: int(821), Column: int(10), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "b", @@ -112654,19 +106206,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(836), + Line: int(821), Column: int(13), }, End: ast.Location{ - Line: int(836), + Line: int(821), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "b", "std", @@ -112676,19 +106228,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(836), + Line: int(821), Column: int(14), }, End: ast.Location{ - Line: int(836), + Line: int(821), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "b", "std", @@ -112697,19 +106249,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(836), + Line: int(821), Column: int(14), }, End: ast.Location{ - Line: int(836), + Line: int(821), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "std", }, @@ -112717,13 +106269,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(836), + Line: int(821), Column: int(14), }, End: ast.Location{ - Line: int(836), + Line: int(821), Column: int(17), }, File: p1, @@ -112758,9 +106310,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -112770,19 +106321,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(836), + Line: int(821), Column: int(27), }, End: ast.Location{ - Line: int(836), + Line: int(821), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8692, + Ctx: p8655, FreeVars: ast.Identifiers{ "b", }, @@ -112804,13 +106355,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(837), + Line: int(822), Column: int(7), }, End: ast.Location{ - Line: int(837), + Line: int(822), Column: int(71), }, File: p1, @@ -112823,7 +106374,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "b", "std", @@ -112832,19 +106383,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(837), + Line: int(822), Column: int(13), }, End: ast.Location{ - Line: int(837), + Line: int(822), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "b", "std", @@ -112853,44 +106404,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(837), + Line: int(822), Column: int(13), }, End: ast.Location{ - Line: int(837), + Line: int(822), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: nil, }, Value: "std.max second param expected number, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(837), + Line: int(822), Column: int(60), }, End: ast.Location{ - Line: int(837), + Line: int(822), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "b", "std", @@ -112899,19 +106449,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(837), + Line: int(822), Column: int(60), }, End: ast.Location{ - Line: int(837), + Line: int(822), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "std", }, @@ -112919,13 +106469,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(837), + Line: int(822), Column: int(60), }, End: ast.Location{ - Line: int(837), + Line: int(822), Column: int(63), }, File: p1, @@ -112960,9 +106510,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -112972,19 +106521,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(837), + Line: int(822), Column: int(69), }, End: ast.Location{ - Line: int(837), + Line: int(822), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8709, + Ctx: p8672, FreeVars: ast.Identifiers{ "b", }, @@ -113014,13 +106563,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(839), + Line: int(824), Column: int(7), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(29), }, File: p1, @@ -113033,7 +106582,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "b", @@ -113042,19 +106591,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(839), + Line: int(824), Column: int(10), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", "b", @@ -113063,19 +106612,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(839), + Line: int(824), Column: int(10), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", }, @@ -113087,19 +106636,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(839), + Line: int(824), Column: int(14), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "b", }, @@ -113111,19 +106660,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(839), + Line: int(824), Column: int(21), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "a", }, @@ -113134,19 +106683,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(839), + Line: int(824), Column: int(28), }, End: ast.Location{ - Line: int(839), + Line: int(824), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8646, + Ctx: p8609, FreeVars: ast.Identifiers{ "b", }, @@ -113158,18 +106707,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(833), - Column: int(3), - }, - End: ast.Location{ - Line: int(839), - Column: int(29), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -113194,7 +106731,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "clamp", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -113215,77 +106751,38 @@ var _StdAst = &ast.DesugaredObject{ FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(841), - Column: int(9), - }, - End: ast.Location{ - Line: int(841), - Column: int(10), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "minVal", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(841), - Column: int(12), - }, - End: ast.Location{ - Line: int(841), - Column: int(18), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "maxVal", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(841), - Column: int(20), - }, - End: ast.Location{ - Line: int(841), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "x", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "minVal", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "maxVal", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(842), + Line: int(827), Column: int(5), }, End: ast.Location{ - Line: int(844), + Line: int(829), Column: int(11), }, File: p1, @@ -113298,7 +106795,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "maxVal", "minVal", @@ -113308,19 +106805,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(842), + Line: int(827), Column: int(8), }, End: ast.Location{ - Line: int(842), - Column: int(18), + Line: int(827), + Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "minVal", "x", @@ -113329,19 +106826,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(842), + Line: int(827), Column: int(8), }, End: ast.Location{ - Line: int(842), + Line: int(827), Column: int(9), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "x", }, @@ -113353,19 +106850,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(842), - Column: int(12), + Line: int(827), + Column: int(13), }, End: ast.Location{ - Line: int(842), - Column: int(18), + Line: int(827), + Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "minVal", }, @@ -113377,19 +106874,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(842), - Column: int(24), + Line: int(827), + Column: int(25), }, End: ast.Location{ - Line: int(842), - Column: int(30), + Line: int(827), + Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "minVal", }, @@ -113407,19 +106904,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(843), + Line: int(828), Column: int(10), }, End: ast.Location{ - Line: int(844), + Line: int(829), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "maxVal", "x", @@ -113428,19 +106925,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(843), + Line: int(828), Column: int(13), }, End: ast.Location{ - Line: int(843), + Line: int(828), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "maxVal", "x", @@ -113449,19 +106946,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(843), + Line: int(828), Column: int(13), }, End: ast.Location{ - Line: int(843), + Line: int(828), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "x", }, @@ -113473,19 +106970,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(843), + Line: int(828), Column: int(17), }, End: ast.Location{ - Line: int(843), + Line: int(828), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "maxVal", }, @@ -113497,19 +106994,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(843), + Line: int(828), Column: int(29), }, End: ast.Location{ - Line: int(843), + Line: int(828), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "maxVal", }, @@ -113527,19 +107024,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(844), + Line: int(829), Column: int(10), }, End: ast.Location{ - Line: int(844), + Line: int(829), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8730, + Ctx: p8693, FreeVars: ast.Identifiers{ "x", }, @@ -113550,18 +107047,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(841), - Column: int(3), - }, - End: ast.Location{ - Line: int(844), - Column: int(11), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -113586,7 +107071,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flattenArrays", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -113609,45 +107093,34 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arrs", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(846), - Column: int(17), - }, - End: ast.Location{ - Line: int(846), - Column: int(21), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arrs", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(5), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8759, + Ctx: p8722, FreeVars: ast.Identifiers{ "arrs", "std", @@ -113656,19 +107129,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(5), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8759, + Ctx: p8722, FreeVars: ast.Identifiers{ "std", }, @@ -113676,13 +107149,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(5), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(8), }, File: p1, @@ -113724,9 +107197,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "foldl", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -113736,80 +107208,55 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(15), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8769, + Ctx: p8732, FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(847), - Column: int(24), - }, - End: ast.Location{ - Line: int(847), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(847), - Column: int(27), - }, - End: ast.Location{ - Line: int(847), - Column: int(28), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(30), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8772, + Ctx: p8735, FreeVars: ast.Identifiers{ "a", "b", @@ -113818,19 +107265,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(30), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8772, + Ctx: p8735, FreeVars: ast.Identifiers{ "a", }, @@ -113842,19 +107289,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(34), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8772, + Ctx: p8735, FreeVars: ast.Identifiers{ "b", }, @@ -113869,19 +107316,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(37), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8769, + Ctx: p8732, FreeVars: ast.Identifiers{ "arrs", }, @@ -113894,19 +107341,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(847), + Line: int(832), Column: int(43), }, End: ast.Location{ - Line: int(847), + Line: int(832), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8769, + Ctx: p8732, FreeVars: nil, }, Elements: nil, @@ -113925,18 +107372,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(846), - Column: int(3), - }, - End: ast.Location{ - Line: int(847), - Column: int(46), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -113961,7 +107396,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestIni", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -113984,39 +107418,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "ini", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(849), - Column: int(15), - }, - End: ast.Location{ - Line: int(849), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "ini", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(850), + Line: int(835), Column: int(5), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(71), }, File: p1, @@ -114029,7 +107452,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8787, + Ctx: p8750, FreeVars: ast.Identifiers{ "ini", "std", @@ -114043,63 +107466,52 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(850), + Line: int(835), Column: int(11), }, End: ast.Location{ - Line: int(859), + Line: int(844), Column: int(9), }, File: p1, }, Fodder: nil, - Ctx: p8791, + Ctx: p8754, FreeVars: ast.Identifiers{ "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "body", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(850), - Column: int(22), - }, - End: ast.Location{ - Line: int(850), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "body", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(851), + Line: int(836), Column: int(7), }, End: ast.Location{ - Line: int(859), + Line: int(844), Column: int(9), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8795, + Ctx: p8758, FreeVars: ast.Identifiers{ "body", "std", @@ -114108,19 +107520,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(851), + Line: int(836), Column: int(7), }, End: ast.Location{ - Line: int(851), + Line: int(836), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8795, + Ctx: p8758, FreeVars: ast.Identifiers{ "std", }, @@ -114128,13 +107540,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(851), + Line: int(836), Column: int(7), }, End: ast.Location{ - Line: int(851), + Line: int(836), Column: int(10), }, File: p1, @@ -114176,9 +107588,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -114188,19 +107599,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(851), + Line: int(836), Column: int(16), }, End: ast.Location{ - Line: int(851), + Line: int(836), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8805, + Ctx: p8768, FreeVars: nil, }, Elements: nil, @@ -114215,14 +107626,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(851), - Column: int(20), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(859), - Column: int(8), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -114295,7 +107706,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -114326,26 +107736,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -114376,13 +107775,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(852), + Line: int(837), Column: int(9), }, End: ast.Location{ - Line: int(856), + Line: int(841), Column: int(45), }, File: p1, @@ -114395,7 +107794,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8822, + Ctx: p8785, FreeVars: ast.Identifiers{ "body", "k", @@ -114410,19 +107809,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(852), + Line: int(837), Column: int(33), }, End: ast.Location{ - Line: int(852), + Line: int(837), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8826, + Ctx: p8789, FreeVars: ast.Identifiers{ "body", "k", @@ -114431,19 +107830,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(852), + Line: int(837), Column: int(33), }, End: ast.Location{ - Line: int(852), + Line: int(837), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8826, + Ctx: p8789, FreeVars: ast.Identifiers{ "body", }, @@ -114454,19 +107853,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(852), + Line: int(837), Column: int(38), }, End: ast.Location{ - Line: int(852), + Line: int(837), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8826, + Ctx: p8789, FreeVars: ast.Identifiers{ "k", }, @@ -114478,30 +107877,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(852), - Column: int(15), - }, - End: ast.Location{ - Line: int(852), - Column: int(40), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(853), + Line: int(838), Column: int(9), }, End: ast.Location{ - Line: int(856), + Line: int(841), Column: int(45), }, File: p1, @@ -114514,7 +107901,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8822, + Ctx: p8785, FreeVars: ast.Identifiers{ "k", "std", @@ -114524,19 +107911,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(853), + Line: int(838), Column: int(12), }, End: ast.Location{ - Line: int(853), + Line: int(838), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8822, + Ctx: p8785, FreeVars: ast.Identifiers{ "std", "value_or_values", @@ -114545,19 +107932,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(853), + Line: int(838), Column: int(12), }, End: ast.Location{ - Line: int(853), + Line: int(838), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8822, + Ctx: p8785, FreeVars: ast.Identifiers{ "std", }, @@ -114565,13 +107952,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(853), + Line: int(838), Column: int(12), }, End: ast.Location{ - Line: int(853), + Line: int(838), Column: int(15), }, File: p1, @@ -114606,9 +107993,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -114618,19 +108004,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(853), + Line: int(838), Column: int(24), }, End: ast.Location{ - Line: int(853), + Line: int(838), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8844, + Ctx: p8807, FreeVars: ast.Identifiers{ "value_or_values", }, @@ -114653,14 +108039,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(854), - Column: int(11), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(854), - Column: int(64), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -114734,7 +108120,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -114765,26 +108150,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "value", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -114817,14 +108191,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(854), - Column: int(12), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(854), - Column: int(34), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -114898,7 +108272,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -114910,25 +108283,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(854), + Line: int(839), Column: int(12), }, End: ast.Location{ - Line: int(854), + Line: int(839), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8869, + Ctx: p8832, FreeVars: nil, }, Value: "%s = %s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -114936,19 +108308,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(854), + Line: int(839), Column: int(24), }, End: ast.Location{ - Line: int(854), + Line: int(839), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8869, + Ctx: p8832, FreeVars: ast.Identifiers{ "k", "value", @@ -114959,19 +108331,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(854), + Line: int(839), Column: int(25), }, End: ast.Location{ - Line: int(854), + Line: int(839), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8874, + Ctx: p8837, FreeVars: ast.Identifiers{ "k", }, @@ -114984,19 +108356,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(854), + Line: int(839), Column: int(28), }, End: ast.Location{ - Line: int(854), + Line: int(839), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8874, + Ctx: p8837, FreeVars: ast.Identifiers{ "value", }, @@ -115032,19 +108404,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(854), + Line: int(839), Column: int(48), }, End: ast.Location{ - Line: int(854), + Line: int(839), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8822, + Ctx: p8785, FreeVars: ast.Identifiers{ "value_or_values", }, @@ -115072,13 +108444,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(856), + Line: int(841), Column: int(11), }, End: ast.Location{ - Line: int(856), + Line: int(841), Column: int(45), }, File: p1, @@ -115091,7 +108463,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8822, + Ctx: p8785, FreeVars: ast.Identifiers{ "k", "std", @@ -115105,14 +108477,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(856), - Column: int(12), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(856), - Column: int(44), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -115186,7 +108558,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -115198,25 +108569,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(856), + Line: int(841), Column: int(12), }, End: ast.Location{ - Line: int(856), + Line: int(841), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8894, + Ctx: p8857, FreeVars: nil, }, Value: "%s = %s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -115224,19 +108594,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(856), + Line: int(841), Column: int(24), }, End: ast.Location{ - Line: int(856), + Line: int(841), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8894, + Ctx: p8857, FreeVars: ast.Identifiers{ "k", "value_or_values", @@ -115247,19 +108617,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(856), + Line: int(841), Column: int(25), }, End: ast.Location{ - Line: int(856), + Line: int(841), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8899, + Ctx: p8862, FreeVars: ast.Identifiers{ "k", }, @@ -115272,19 +108642,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(856), + Line: int(841), Column: int(28), }, End: ast.Location{ - Line: int(856), + Line: int(841), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8899, + Ctx: p8862, FreeVars: ast.Identifiers{ "value_or_values", }, @@ -115328,19 +108698,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(858), + Line: int(843), Column: int(18), }, End: ast.Location{ - Line: int(858), + Line: int(843), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8805, + Ctx: p8768, FreeVars: ast.Identifiers{ "body", "std", @@ -115349,19 +108719,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(858), + Line: int(843), Column: int(18), }, End: ast.Location{ - Line: int(858), + Line: int(843), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8805, + Ctx: p8768, FreeVars: ast.Identifiers{ "std", }, @@ -115369,13 +108739,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(858), + Line: int(843), Column: int(18), }, End: ast.Location{ - Line: int(858), + Line: int(843), Column: int(21), }, File: p1, @@ -115410,9 +108780,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -115422,19 +108791,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(858), + Line: int(843), Column: int(35), }, End: ast.Location{ - Line: int(858), + Line: int(843), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8912, + Ctx: p8875, FreeVars: ast.Identifiers{ "body", }, @@ -115474,30 +108843,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(5), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(71), }, File: p1, @@ -115510,7 +108867,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8787, + Ctx: p8750, FreeVars: ast.Identifiers{ "body_lines", "ini", @@ -115525,83 +108882,58 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(11), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(79), }, File: p1, }, Fodder: nil, - Ctx: p8919, + Ctx: p8882, FreeVars: ast.Identifiers{ "body_lines", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "sname", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(861), - Column: int(25), - }, - End: ast.Location{ - Line: int(861), - Column: int(30), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "sname", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "sbody", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(861), - Column: int(32), - }, - End: ast.Location{ - Line: int(861), - Column: int(37), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "sbody", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(41), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8923, + Ctx: p8886, FreeVars: ast.Identifiers{ "body_lines", "sbody", @@ -115612,19 +108944,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(41), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8923, + Ctx: p8886, FreeVars: ast.Identifiers{ "sname", "std", @@ -115637,14 +108969,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(861), - Column: int(42), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(861), - Column: int(58), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -115717,7 +109049,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -115729,25 +109060,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(42), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8937, + Ctx: p8900, FreeVars: nil, }, Value: "[%s]", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -115755,19 +109085,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(51), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8937, + Ctx: p8900, FreeVars: ast.Identifiers{ "sname", }, @@ -115777,19 +109107,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(52), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8942, + Ctx: p8905, FreeVars: ast.Identifiers{ "sname", }, @@ -115823,19 +109153,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(62), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8923, + Ctx: p8886, FreeVars: ast.Identifiers{ "body_lines", "sbody", @@ -115844,19 +109174,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(62), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8923, + Ctx: p8886, FreeVars: ast.Identifiers{ "body_lines", }, @@ -115870,19 +109200,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(861), + Line: int(846), Column: int(73), }, End: ast.Location{ - Line: int(861), + Line: int(846), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8950, + Ctx: p8913, FreeVars: ast.Identifiers{ "sbody", }, @@ -115903,18 +109233,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, ast.LocalBind{ VarFodder: ast.Fodder{ @@ -115930,19 +109248,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(23), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8954, + Ctx: p8917, FreeVars: ast.Identifiers{ "body_lines", "ini", @@ -115952,19 +109270,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(26), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8954, + Ctx: p8917, FreeVars: ast.Identifiers{ "ini", "std", @@ -115973,19 +109291,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(26), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8954, + Ctx: p8917, FreeVars: ast.Identifiers{ "std", }, @@ -115993,13 +109311,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(26), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(29), }, File: p1, @@ -116034,9 +109352,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHas", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -116046,19 +109363,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(40), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8965, + Ctx: p8928, FreeVars: ast.Identifiers{ "ini", }, @@ -116071,25 +109388,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(45), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8965, + Ctx: p8928, FreeVars: nil, }, Value: "main", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -116105,19 +109421,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(58), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8954, + Ctx: p8917, FreeVars: ast.Identifiers{ "body_lines", "ini", @@ -116126,19 +109442,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(58), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8954, + Ctx: p8917, FreeVars: ast.Identifiers{ "body_lines", }, @@ -116152,19 +109468,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(69), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8974, + Ctx: p8937, FreeVars: ast.Identifiers{ "ini", }, @@ -116172,13 +109488,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(69), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(72), }, File: p1, @@ -116213,9 +109529,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "main", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -116232,19 +109547,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(862), + Line: int(847), Column: int(84), }, End: ast.Location{ - Line: int(862), + Line: int(847), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8954, + Ctx: p8917, FreeVars: nil, }, Elements: nil, @@ -116254,18 +109569,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(862), - Column: int(11), - }, - End: ast.Location{ - Line: int(862), - Column: int(86), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: ast.Fodder{ @@ -116283,14 +109586,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(863), - Column: int(26), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(866), - Column: int(6), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -116364,7 +109667,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -116395,26 +109697,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -116445,19 +109736,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(864), + Line: int(849), Column: int(7), }, End: ast.Location{ - Line: int(864), + Line: int(849), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8996, + Ctx: p8959, FreeVars: ast.Identifiers{ "ini", "k", @@ -116467,13 +109758,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(864), + Line: int(849), Column: int(7), }, End: ast.Location{ - Line: int(864), + Line: int(849), Column: int(20), }, File: p1, @@ -116486,7 +109777,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p8996, + Ctx: p8959, FreeVars: ast.Identifiers{ "section_lines", }, @@ -116500,19 +109791,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(864), + Line: int(849), Column: int(21), }, End: ast.Location{ - Line: int(864), + Line: int(849), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9003, + Ctx: p8966, FreeVars: ast.Identifiers{ "k", }, @@ -116525,19 +109816,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(864), + Line: int(849), Column: int(24), }, End: ast.Location{ - Line: int(864), + Line: int(849), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9003, + Ctx: p8966, FreeVars: ast.Identifiers{ "ini", "k", @@ -116546,19 +109837,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(864), + Line: int(849), Column: int(24), }, End: ast.Location{ - Line: int(864), + Line: int(849), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9003, + Ctx: p8966, FreeVars: ast.Identifiers{ "ini", }, @@ -116566,13 +109857,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(864), + Line: int(849), Column: int(24), }, End: ast.Location{ - Line: int(864), + Line: int(849), Column: int(27), }, File: p1, @@ -116607,28 +109898,27 @@ var _StdAst = &ast.DesugaredObject{ Value: "sections", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, LeftBracketFodder: ast.Fodder{}, Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(864), + Line: int(849), Column: int(37), }, End: ast.Location{ - Line: int(864), + Line: int(849), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9003, + Ctx: p8966, FreeVars: ast.Identifiers{ "k", }, @@ -116661,19 +109951,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(865), + Line: int(850), Column: int(16), }, End: ast.Location{ - Line: int(865), + Line: int(850), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9015, + Ctx: p8978, FreeVars: ast.Identifiers{ "ini", "std", @@ -116682,19 +109972,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(865), + Line: int(850), Column: int(16), }, End: ast.Location{ - Line: int(865), + Line: int(850), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9015, + Ctx: p8978, FreeVars: ast.Identifiers{ "std", }, @@ -116702,13 +109992,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(865), + Line: int(850), Column: int(16), }, End: ast.Location{ - Line: int(865), + Line: int(850), Column: int(19), }, File: p1, @@ -116743,9 +110033,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -116755,19 +110044,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(865), + Line: int(850), Column: int(33), }, End: ast.Location{ - Line: int(865), + Line: int(850), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9024, + Ctx: p8987, FreeVars: ast.Identifiers{ "ini", }, @@ -116775,13 +110064,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(865), + Line: int(850), Column: int(33), }, End: ast.Location{ - Line: int(865), + Line: int(850), Column: int(36), }, File: p1, @@ -116816,9 +110105,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "sections", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -116843,36 +110131,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(863), - Column: int(11), - }, - End: ast.Location{ - Line: int(866), - Column: int(6), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(5), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8787, + Ctx: p8750, FreeVars: ast.Identifiers{ "all_sections", "main_body", @@ -116882,19 +110158,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(5), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p8787, + Ctx: p8750, FreeVars: ast.Identifiers{ "std", }, @@ -116902,13 +110178,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(5), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(8), }, File: p1, @@ -116950,9 +110226,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -116962,25 +110237,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(14), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9039, + Ctx: p9002, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -116988,19 +110262,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(20), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9039, + Ctx: p9002, FreeVars: ast.Identifiers{ "all_sections", "main_body", @@ -117010,19 +110284,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(20), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9039, + Ctx: p9002, FreeVars: ast.Identifiers{ "all_sections", "main_body", @@ -117032,19 +110306,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(20), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9039, + Ctx: p9002, FreeVars: ast.Identifiers{ "main_body", }, @@ -117056,19 +110330,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(32), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9039, + Ctx: p9002, FreeVars: ast.Identifiers{ "all_sections", "std", @@ -117077,19 +110351,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(32), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9039, + Ctx: p9002, FreeVars: ast.Identifiers{ "std", }, @@ -117097,13 +110371,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(32), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(35), }, File: p1, @@ -117138,9 +110412,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "flattenArrays", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -117150,19 +110423,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(50), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9055, + Ctx: p9018, FreeVars: ast.Identifiers{ "all_sections", }, @@ -117185,19 +110458,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(66), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9039, + Ctx: p9002, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -117205,25 +110478,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(867), + Line: int(852), Column: int(67), }, End: ast.Location{ - Line: int(867), + Line: int(852), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9060, + Ctx: p9023, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -117246,18 +110518,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(849), - Column: int(3), - }, - End: ast.Location{ - Line: int(867), - Column: int(71), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -117282,7 +110542,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringJson", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -117305,39 +110564,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str_", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(869), - Column: int(20), - }, - End: ast.Location{ - Line: int(869), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str_", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(870), + Line: int(855), Column: int(5), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(70), }, File: p1, @@ -117350,7 +110598,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9067, + Ctx: p9030, FreeVars: ast.Identifiers{ "std", "str_", @@ -117364,19 +110612,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(870), + Line: int(855), Column: int(17), }, End: ast.Location{ - Line: int(870), + Line: int(855), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9071, + Ctx: p9034, FreeVars: ast.Identifiers{ "std", "str_", @@ -117385,19 +110633,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(870), + Line: int(855), Column: int(17), }, End: ast.Location{ - Line: int(870), + Line: int(855), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9071, + Ctx: p9034, FreeVars: ast.Identifiers{ "std", }, @@ -117405,13 +110653,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(870), + Line: int(855), Column: int(17), }, End: ast.Location{ - Line: int(870), + Line: int(855), Column: int(20), }, File: p1, @@ -117446,9 +110694,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "toString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -117458,19 +110705,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(870), + Line: int(855), Column: int(30), }, End: ast.Location{ - Line: int(870), + Line: int(855), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9080, + Ctx: p9043, FreeVars: ast.Identifiers{ "str_", }, @@ -117489,30 +110736,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(870), - Column: int(11), - }, - End: ast.Location{ - Line: int(870), - Column: int(35), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(871), + Line: int(856), Column: int(5), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(70), }, File: p1, @@ -117525,7 +110760,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9067, + Ctx: p9030, FreeVars: ast.Identifiers{ "std", "str", @@ -117539,57 +110774,46 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(871), + Line: int(856), Column: int(11), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, }, Fodder: nil, - Ctx: p9087, + Ctx: p9050, FreeVars: ast.Identifiers{ "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "ch", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(871), - Column: int(17), - }, - End: ast.Location{ - Line: int(871), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "ch", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(872), + Line: int(857), Column: int(7), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, @@ -117602,7 +110826,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -117611,19 +110835,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(872), + Line: int(857), Column: int(10), }, End: ast.Location{ - Line: int(872), + Line: int(857), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -117631,19 +110855,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(872), + Line: int(857), Column: int(10), }, End: ast.Location{ - Line: int(872), + Line: int(857), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -117655,38 +110879,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(872), + Line: int(857), Column: int(16), }, End: ast.Location{ - Line: int(872), + Line: int(857), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(873), + Line: int(858), Column: int(9), }, End: ast.Location{ - Line: int(873), + Line: int(858), Column: int(14), }, File: p1, @@ -117699,13 +110922,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -117718,19 +110940,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(874), + Line: int(859), Column: int(12), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -117739,19 +110961,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(874), + Line: int(859), Column: int(15), }, End: ast.Location{ - Line: int(874), + Line: int(859), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -117759,19 +110981,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(874), + Line: int(859), Column: int(15), }, End: ast.Location{ - Line: int(874), + Line: int(859), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -117783,38 +111005,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(874), + Line: int(859), Column: int(21), }, End: ast.Location{ - Line: int(874), + Line: int(859), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(875), + Line: int(860), Column: int(9), }, End: ast.Location{ - Line: int(875), + Line: int(860), Column: int(15), }, File: p1, @@ -117827,13 +111048,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\\\", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -117846,19 +111066,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(876), + Line: int(861), Column: int(12), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -117867,19 +111087,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(876), + Line: int(861), Column: int(15), }, End: ast.Location{ - Line: int(876), + Line: int(861), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -117887,19 +111107,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(876), + Line: int(861), Column: int(15), }, End: ast.Location{ - Line: int(876), + Line: int(861), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -117911,38 +111131,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(876), + Line: int(861), Column: int(21), }, End: ast.Location{ - Line: int(876), + Line: int(861), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\b", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(877), + Line: int(862), Column: int(9), }, End: ast.Location{ - Line: int(877), + Line: int(862), Column: int(14), }, File: p1, @@ -117955,13 +111174,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\b", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -117974,19 +111192,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(878), + Line: int(863), Column: int(12), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -117995,19 +111213,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(878), + Line: int(863), Column: int(15), }, End: ast.Location{ - Line: int(878), + Line: int(863), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118015,19 +111233,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(878), + Line: int(863), Column: int(15), }, End: ast.Location{ - Line: int(878), + Line: int(863), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118039,38 +111257,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(878), + Line: int(863), Column: int(21), }, End: ast.Location{ - Line: int(878), + Line: int(863), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\f", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(879), + Line: int(864), Column: int(9), }, End: ast.Location{ - Line: int(879), + Line: int(864), Column: int(14), }, File: p1, @@ -118083,13 +111300,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\f", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -118102,19 +111318,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(880), + Line: int(865), Column: int(12), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -118123,19 +111339,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(880), + Line: int(865), Column: int(15), }, End: ast.Location{ - Line: int(880), + Line: int(865), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118143,19 +111359,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(880), + Line: int(865), Column: int(15), }, End: ast.Location{ - Line: int(880), + Line: int(865), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118167,38 +111383,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(880), + Line: int(865), Column: int(21), }, End: ast.Location{ - Line: int(880), + Line: int(865), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(881), + Line: int(866), Column: int(9), }, End: ast.Location{ - Line: int(881), + Line: int(866), Column: int(14), }, File: p1, @@ -118211,13 +111426,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -118230,19 +111444,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(882), + Line: int(867), Column: int(12), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -118251,19 +111465,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(882), + Line: int(867), Column: int(15), }, End: ast.Location{ - Line: int(882), + Line: int(867), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118271,19 +111485,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(882), + Line: int(867), Column: int(15), }, End: ast.Location{ - Line: int(882), + Line: int(867), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118295,38 +111509,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(882), + Line: int(867), Column: int(21), }, End: ast.Location{ - Line: int(882), + Line: int(867), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\r", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(883), + Line: int(868), Column: int(9), }, End: ast.Location{ - Line: int(883), + Line: int(868), Column: int(14), }, File: p1, @@ -118339,13 +111552,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\r", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -118358,19 +111570,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(884), + Line: int(869), Column: int(12), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -118379,19 +111591,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(884), + Line: int(869), Column: int(15), }, End: ast.Location{ - Line: int(884), + Line: int(869), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118399,19 +111611,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(884), + Line: int(869), Column: int(15), }, End: ast.Location{ - Line: int(884), + Line: int(869), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -118423,38 +111635,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(884), + Line: int(869), Column: int(21), }, End: ast.Location{ - Line: int(884), + Line: int(869), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\t", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(885), + Line: int(870), Column: int(9), }, End: ast.Location{ - Line: int(885), + Line: int(870), Column: int(14), }, File: p1, @@ -118467,13 +111678,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\t", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -118486,13 +111696,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(887), + Line: int(872), Column: int(9), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, @@ -118505,7 +111715,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "std", @@ -118519,19 +111729,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(887), + Line: int(872), Column: int(20), }, End: ast.Location{ - Line: int(887), + Line: int(872), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9167, + Ctx: p9130, FreeVars: ast.Identifiers{ "ch", "std", @@ -118540,19 +111750,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(887), + Line: int(872), Column: int(20), }, End: ast.Location{ - Line: int(887), + Line: int(872), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9167, + Ctx: p9130, FreeVars: ast.Identifiers{ "std", }, @@ -118560,13 +111770,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(887), + Line: int(872), Column: int(20), }, End: ast.Location{ - Line: int(887), + Line: int(872), Column: int(23), }, File: p1, @@ -118601,9 +111811,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -118613,19 +111822,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(887), + Line: int(872), Column: int(34), }, End: ast.Location{ - Line: int(887), + Line: int(872), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9176, + Ctx: p9139, FreeVars: ast.Identifiers{ "ch", }, @@ -118644,30 +111853,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(887), - Column: int(15), - }, - End: ast.Location{ - Line: int(887), - Column: int(37), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(9), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, @@ -118680,7 +111877,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", "cp", @@ -118690,19 +111887,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(12), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118710,19 +111907,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(12), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118730,19 +111927,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(12), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118754,21 +111951,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(17), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, + Value: float64(32), OriginalString: "32", }, }, @@ -118777,19 +111975,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(24), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118797,19 +111995,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(24), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118817,19 +112015,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(24), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118841,21 +112039,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(30), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, + Value: float64(127), OriginalString: "127", }, }, @@ -118864,19 +112063,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(37), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118884,19 +112083,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(37), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -118908,21 +112107,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(888), + Line: int(873), Column: int(43), }, End: ast.Location{ - Line: int(888), + Line: int(873), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, + Value: float64(159), OriginalString: "159", }, }, @@ -118934,14 +112134,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(889), - Column: int(11), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(889), - Column: int(27), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -119014,7 +112214,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -119026,13 +112225,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(889), + Line: int(874), Column: int(11), }, End: ast.Location{ - Line: int(889), + Line: int(874), Column: int(20), }, File: p1, @@ -119045,13 +112244,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: nil, }, Value: "\\u%04x", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -119059,19 +112257,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(889), + Line: int(874), Column: int(23), }, End: ast.Location{ - Line: int(889), + Line: int(874), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "cp", }, @@ -119081,19 +112279,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(889), + Line: int(874), Column: int(24), }, End: ast.Location{ - Line: int(889), + Line: int(874), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9214, + Ctx: p9177, FreeVars: ast.Identifiers{ "cp", }, @@ -119127,13 +112325,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(891), + Line: int(876), Column: int(11), }, End: ast.Location{ - Line: int(891), + Line: int(876), Column: int(13), }, File: p1, @@ -119146,7 +112344,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9092, + Ctx: p9055, FreeVars: ast.Identifiers{ "ch", }, @@ -119165,6 +112363,10 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, + }, + }, + Body: &ast.Apply{ + NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ @@ -119177,22 +112379,6 @@ var _StdAst = &ast.DesugaredObject{ }, File: nil, }, - }, - }, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(892), - Column: int(5), - }, - End: ast.Location{ - Line: int(892), - Column: int(70), - }, - File: p1, - }, Fodder: nil, Ctx: nil, FreeVars: ast.Identifiers{ @@ -119265,7 +112451,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -119277,13 +112462,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(5), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(11), }, File: p1, @@ -119296,13 +112481,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9067, + Ctx: p9030, FreeVars: nil, }, Value: "\"%s\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -119310,19 +112494,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(14), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9067, + Ctx: p9030, FreeVars: ast.Identifiers{ "std", "str", @@ -119332,19 +112516,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(14), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9067, + Ctx: p9030, FreeVars: ast.Identifiers{ "std", }, @@ -119352,13 +112536,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(14), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(17), }, File: p1, @@ -119393,9 +112577,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -119405,25 +112588,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(23), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9239, + Ctx: p9202, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -119433,14 +112615,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(892), - Column: int(27), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(892), - Column: int(69), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -119514,7 +112696,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -119544,26 +112725,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "ch", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "ch", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -119593,19 +112763,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(28), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9255, + Ctx: p9218, FreeVars: ast.Identifiers{ "ch", "trans", @@ -119614,19 +112784,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(28), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9255, + Ctx: p9218, FreeVars: ast.Identifiers{ "trans", }, @@ -119640,19 +112810,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(34), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9261, + Ctx: p9224, FreeVars: ast.Identifiers{ "ch", }, @@ -119682,19 +112852,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(48), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9239, + Ctx: p9202, FreeVars: ast.Identifiers{ "std", "str", @@ -119703,19 +112873,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(48), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9239, + Ctx: p9202, FreeVars: ast.Identifiers{ "std", }, @@ -119723,13 +112893,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(48), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(51), }, File: p1, @@ -119764,9 +112934,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "stringChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -119776,19 +112945,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(892), + Line: int(877), Column: int(64), }, End: ast.Location{ - Line: int(892), + Line: int(877), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9272, + Ctx: p9235, FreeVars: ast.Identifiers{ "str", }, @@ -119839,18 +113008,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(869), - Column: int(3), - }, - End: ast.Location{ - Line: int(892), - Column: int(70), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -119875,7 +113032,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringPython", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -119898,45 +113054,34 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(894), - Column: int(22), - }, - End: ast.Location{ - Line: int(894), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(895), + Line: int(880), Column: int(5), }, End: ast.Location{ - Line: int(895), + Line: int(880), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9279, + Ctx: p9242, FreeVars: ast.Identifiers{ "std", "str", @@ -119945,19 +113090,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(895), + Line: int(880), Column: int(5), }, End: ast.Location{ - Line: int(895), + Line: int(880), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9279, + Ctx: p9242, FreeVars: ast.Identifiers{ "std", }, @@ -119965,13 +113110,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(895), + Line: int(880), Column: int(5), }, End: ast.Location{ - Line: int(895), + Line: int(880), Column: int(8), }, File: p1, @@ -120013,9 +113158,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringJson", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -120025,19 +113169,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(895), + Line: int(880), Column: int(26), }, End: ast.Location{ - Line: int(895), + Line: int(880), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9289, + Ctx: p9252, FreeVars: ast.Identifiers{ "str", }, @@ -120056,18 +113200,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(894), - Column: int(3), - }, - End: ast.Location{ - Line: int(895), - Column: int(30), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -120092,7 +113224,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringBash", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -120115,39 +113246,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str_", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(897), - Column: int(20), - }, - End: ast.Location{ - Line: int(897), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str_", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(898), + Line: int(883), Column: int(5), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(70), }, File: p1, @@ -120160,7 +113280,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9297, + Ctx: p9260, FreeVars: ast.Identifiers{ "std", "str_", @@ -120174,19 +113294,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(898), + Line: int(883), Column: int(17), }, End: ast.Location{ - Line: int(898), + Line: int(883), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9301, + Ctx: p9264, FreeVars: ast.Identifiers{ "std", "str_", @@ -120195,19 +113315,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(898), + Line: int(883), Column: int(17), }, End: ast.Location{ - Line: int(898), + Line: int(883), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9301, + Ctx: p9264, FreeVars: ast.Identifiers{ "std", }, @@ -120215,13 +113335,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(898), + Line: int(883), Column: int(17), }, End: ast.Location{ - Line: int(898), + Line: int(883), Column: int(20), }, File: p1, @@ -120256,9 +113376,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "toString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -120268,19 +113387,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(898), + Line: int(883), Column: int(30), }, End: ast.Location{ - Line: int(898), + Line: int(883), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9310, + Ctx: p9273, FreeVars: ast.Identifiers{ "str_", }, @@ -120299,30 +113418,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(898), - Column: int(11), - }, - End: ast.Location{ - Line: int(898), - Column: int(35), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(899), + Line: int(884), Column: int(5), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(70), }, File: p1, @@ -120335,7 +113442,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9297, + Ctx: p9260, FreeVars: ast.Identifiers{ "std", "str", @@ -120349,55 +113456,44 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(899), + Line: int(884), Column: int(11), }, End: ast.Location{ - Line: int(903), + Line: int(888), Column: int(11), }, File: p1, }, Fodder: nil, - Ctx: p9317, + Ctx: p9280, FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "ch", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(899), - Column: int(17), - }, - End: ast.Location{ - Line: int(899), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "ch", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(900), + Line: int(885), Column: int(7), }, End: ast.Location{ - Line: int(903), + Line: int(888), Column: int(11), }, File: p1, @@ -120410,7 +113506,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9321, + Ctx: p9284, FreeVars: ast.Identifiers{ "ch", }, @@ -120418,19 +113514,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(900), + Line: int(885), Column: int(10), }, End: ast.Location{ - Line: int(900), + Line: int(885), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9321, + Ctx: p9284, FreeVars: ast.Identifiers{ "ch", }, @@ -120438,19 +113534,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(900), + Line: int(885), Column: int(10), }, End: ast.Location{ - Line: int(900), + Line: int(885), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9321, + Ctx: p9284, FreeVars: ast.Identifiers{ "ch", }, @@ -120462,38 +113558,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(900), + Line: int(885), Column: int(16), }, End: ast.Location{ - Line: int(900), + Line: int(885), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9321, + Ctx: p9284, FreeVars: nil, }, Value: "'", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(901), + Line: int(886), Column: int(9), }, End: ast.Location{ - Line: int(901), + Line: int(886), Column: int(18), }, File: p1, @@ -120506,13 +113601,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9321, + Ctx: p9284, FreeVars: nil, }, Value: "'\"'\"'", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -120525,13 +113619,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(903), + Line: int(888), Column: int(9), }, End: ast.Location{ - Line: int(903), + Line: int(888), Column: int(11), }, File: p1, @@ -120544,7 +113638,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9321, + Ctx: p9284, FreeVars: ast.Identifiers{ "ch", }, @@ -120555,6 +113649,10 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, + }, + }, + Body: &ast.Apply{ + NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ @@ -120567,22 +113665,6 @@ var _StdAst = &ast.DesugaredObject{ }, File: nil, }, - }, - }, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(904), - Column: int(5), - }, - End: ast.Location{ - Line: int(904), - Column: int(70), - }, - File: p1, - }, Fodder: nil, Ctx: nil, FreeVars: ast.Identifiers{ @@ -120655,7 +113737,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -120667,13 +113748,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(5), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(11), }, File: p1, @@ -120686,13 +113767,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9297, + Ctx: p9260, FreeVars: nil, }, Value: "'%s'", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -120700,19 +113780,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(14), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9297, + Ctx: p9260, FreeVars: ast.Identifiers{ "std", "str", @@ -120722,19 +113802,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(14), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9297, + Ctx: p9260, FreeVars: ast.Identifiers{ "std", }, @@ -120742,13 +113822,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(14), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(17), }, File: p1, @@ -120783,9 +113863,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -120795,25 +113874,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(23), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9353, + Ctx: p9316, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -120823,14 +113901,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(904), - Column: int(27), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(904), - Column: int(69), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -120904,7 +113982,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -120934,26 +114011,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "ch", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "ch", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -120983,19 +114049,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(28), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9369, + Ctx: p9332, FreeVars: ast.Identifiers{ "ch", "trans", @@ -121004,19 +114070,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(28), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9369, + Ctx: p9332, FreeVars: ast.Identifiers{ "trans", }, @@ -121030,19 +114096,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(34), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9375, + Ctx: p9338, FreeVars: ast.Identifiers{ "ch", }, @@ -121072,19 +114138,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(48), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9353, + Ctx: p9316, FreeVars: ast.Identifiers{ "std", "str", @@ -121093,19 +114159,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(48), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9353, + Ctx: p9316, FreeVars: ast.Identifiers{ "std", }, @@ -121113,13 +114179,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(48), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(51), }, File: p1, @@ -121154,9 +114220,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "stringChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -121166,19 +114231,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(904), + Line: int(889), Column: int(64), }, End: ast.Location{ - Line: int(904), + Line: int(889), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9386, + Ctx: p9349, FreeVars: ast.Identifiers{ "str", }, @@ -121229,18 +114294,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(897), - Column: int(3), - }, - End: ast.Location{ - Line: int(904), - Column: int(70), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -121265,7 +114318,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringDollars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -121288,39 +114340,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str_", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(906), - Column: int(23), - }, - End: ast.Location{ - Line: int(906), - Column: int(27), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str_", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(907), + Line: int(892), Column: int(5), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(69), }, File: p1, @@ -121333,7 +114374,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9394, + Ctx: p9357, FreeVars: ast.Identifiers{ "std", "str_", @@ -121347,19 +114388,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(907), + Line: int(892), Column: int(17), }, End: ast.Location{ - Line: int(907), + Line: int(892), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9398, + Ctx: p9361, FreeVars: ast.Identifiers{ "std", "str_", @@ -121368,19 +114409,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(907), + Line: int(892), Column: int(17), }, End: ast.Location{ - Line: int(907), + Line: int(892), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9398, + Ctx: p9361, FreeVars: ast.Identifiers{ "std", }, @@ -121388,13 +114429,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(907), + Line: int(892), Column: int(17), }, End: ast.Location{ - Line: int(907), + Line: int(892), Column: int(20), }, File: p1, @@ -121429,9 +114470,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "toString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -121441,19 +114481,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(907), + Line: int(892), Column: int(30), }, End: ast.Location{ - Line: int(907), + Line: int(892), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9407, + Ctx: p9370, FreeVars: ast.Identifiers{ "str_", }, @@ -121472,30 +114512,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(907), - Column: int(11), - }, - End: ast.Location{ - Line: int(907), - Column: int(35), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(908), + Line: int(893), Column: int(5), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(69), }, File: p1, @@ -121508,7 +114536,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9394, + Ctx: p9357, FreeVars: ast.Identifiers{ "std", "str", @@ -121522,55 +114550,44 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(908), + Line: int(893), Column: int(11), }, End: ast.Location{ - Line: int(912), + Line: int(897), Column: int(11), }, File: p1, }, Fodder: nil, - Ctx: p9414, + Ctx: p9377, FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "ch", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(908), - Column: int(17), - }, - End: ast.Location{ - Line: int(908), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "ch", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(909), + Line: int(894), Column: int(7), }, End: ast.Location{ - Line: int(912), + Line: int(897), Column: int(11), }, File: p1, @@ -121583,7 +114600,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9418, + Ctx: p9381, FreeVars: ast.Identifiers{ "ch", }, @@ -121591,19 +114608,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(909), + Line: int(894), Column: int(10), }, End: ast.Location{ - Line: int(909), + Line: int(894), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9418, + Ctx: p9381, FreeVars: ast.Identifiers{ "ch", }, @@ -121611,19 +114628,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(909), + Line: int(894), Column: int(10), }, End: ast.Location{ - Line: int(909), + Line: int(894), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9418, + Ctx: p9381, FreeVars: ast.Identifiers{ "ch", }, @@ -121635,38 +114652,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(909), + Line: int(894), Column: int(16), }, End: ast.Location{ - Line: int(909), + Line: int(894), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9418, + Ctx: p9381, FreeVars: nil, }, Value: "$", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(910), + Line: int(895), Column: int(9), }, End: ast.Location{ - Line: int(910), + Line: int(895), Column: int(13), }, File: p1, @@ -121679,13 +114695,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9418, + Ctx: p9381, FreeVars: nil, }, Value: "$$", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -121698,13 +114713,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(912), + Line: int(897), Column: int(9), }, End: ast.Location{ - Line: int(912), + Line: int(897), Column: int(11), }, File: p1, @@ -121717,7 +114732,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9418, + Ctx: p9381, FreeVars: ast.Identifiers{ "ch", }, @@ -121728,36 +114743,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(5), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9394, + Ctx: p9357, FreeVars: ast.Identifiers{ "std", "str", @@ -121767,19 +114770,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(5), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9394, + Ctx: p9357, FreeVars: ast.Identifiers{ "std", }, @@ -121787,13 +114790,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(5), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(8), }, File: p1, @@ -121835,9 +114838,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "foldl", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -121847,82 +114849,57 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(15), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9441, + Ctx: p9404, FreeVars: ast.Identifiers{ "trans", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(913), - Column: int(24), - }, - End: ast.Location{ - Line: int(913), - Column: int(25), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(913), - Column: int(27), - }, - End: ast.Location{ - Line: int(913), - Column: int(28), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(30), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9445, + Ctx: p9408, FreeVars: ast.Identifiers{ "a", "b", @@ -121932,19 +114909,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(30), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9445, + Ctx: p9408, FreeVars: ast.Identifiers{ "a", }, @@ -121956,19 +114933,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(34), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9445, + Ctx: p9408, FreeVars: ast.Identifiers{ "b", "trans", @@ -121977,19 +114954,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(34), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9445, + Ctx: p9408, FreeVars: ast.Identifiers{ "trans", }, @@ -122003,19 +114980,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(40), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9455, + Ctx: p9418, FreeVars: ast.Identifiers{ "b", }, @@ -122040,19 +115017,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(44), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9441, + Ctx: p9404, FreeVars: ast.Identifiers{ "std", "str", @@ -122061,19 +115038,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(44), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9441, + Ctx: p9404, FreeVars: ast.Identifiers{ "std", }, @@ -122081,13 +115058,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(44), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(47), }, File: p1, @@ -122122,9 +115099,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "stringChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -122134,19 +115110,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(60), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9466, + Ctx: p9429, FreeVars: ast.Identifiers{ "str", }, @@ -122169,25 +115145,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(913), + Line: int(898), Column: int(66), }, End: ast.Location{ - Line: int(913), + Line: int(898), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9441, + Ctx: p9404, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -122203,18 +115178,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(906), - Column: int(3), - }, - End: ast.Location{ - Line: int(913), - Column: int(69), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -122239,7 +115202,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestJson", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -122262,45 +115224,34 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(915), - Column: int(16), - }, - End: ast.Location{ - Line: int(915), - Column: int(21), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(915), + Line: int(900), Column: int(25), }, End: ast.Location{ - Line: int(915), + Line: int(900), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9474, + Ctx: p9437, FreeVars: ast.Identifiers{ "std", "value", @@ -122309,19 +115260,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(915), + Line: int(900), Column: int(25), }, End: ast.Location{ - Line: int(915), + Line: int(900), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9474, + Ctx: p9437, FreeVars: ast.Identifiers{ "std", }, @@ -122329,13 +115280,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(915), + Line: int(900), Column: int(25), }, End: ast.Location{ - Line: int(915), + Line: int(900), Column: int(28), }, File: p1, @@ -122370,9 +115321,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestJsonEx", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -122382,19 +115332,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(915), + Line: int(900), Column: int(44), }, End: ast.Location{ - Line: int(915), + Line: int(900), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9483, + Ctx: p9446, FreeVars: ast.Identifiers{ "value", }, @@ -122407,25 +115357,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(915), + Line: int(900), Column: int(51), }, End: ast.Location{ - Line: int(915), + Line: int(900), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9483, + Ctx: p9446, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -122439,18 +115388,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(915), - Column: int(3), - }, - End: ast.Location{ - Line: int(915), - Column: int(58), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -122475,7 +115412,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestJsonEx", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -122498,58 +115434,33 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(917), - Column: int(18), - }, - End: ast.Location{ - Line: int(917), - Column: int(23), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "indent", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(917), - Column: int(25), - }, - End: ast.Location{ - Line: int(917), - Column: int(31), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "indent", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(918), + Line: int(903), Column: int(5), }, End: ast.Location{ - Line: int(952), + Line: int(937), Column: int(23), }, File: p1, @@ -122562,7 +115473,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9492, + Ctx: p9455, FreeVars: ast.Identifiers{ "indent", "std", @@ -122577,19 +115488,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(918), + Line: int(903), Column: int(11), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: nil, - Ctx: p9496, + Ctx: p9459, FreeVars: ast.Identifiers{ "aux", "indent", @@ -122597,77 +115508,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(918), - Column: int(15), - }, - End: ast.Location{ - Line: int(918), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "path", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(918), - Column: int(18), - }, - End: ast.Location{ - Line: int(918), - Column: int(22), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "path", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "cindent", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(918), - Column: int(24), - }, - End: ast.Location{ - Line: int(918), - Column: int(31), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "cindent", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(919), + Line: int(904), Column: int(7), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, @@ -122680,7 +115552,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -122693,19 +115565,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(919), + Line: int(904), Column: int(10), }, End: ast.Location{ - Line: int(919), + Line: int(904), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -122713,19 +115585,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(919), + Line: int(904), Column: int(10), }, End: ast.Location{ - Line: int(919), + Line: int(904), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -122737,19 +115609,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(919), + Line: int(904), Column: int(15), }, End: ast.Location{ - Line: int(919), + Line: int(904), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, Value: true, @@ -122759,13 +115631,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(920), + Line: int(905), Column: int(9), }, End: ast.Location{ - Line: int(920), + Line: int(905), Column: int(15), }, File: p1, @@ -122778,13 +115650,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, Value: "true", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -122797,19 +115668,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(921), + Line: int(906), Column: int(12), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -122822,19 +115693,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(921), + Line: int(906), Column: int(15), }, End: ast.Location{ - Line: int(921), + Line: int(906), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -122842,19 +115713,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(921), + Line: int(906), Column: int(15), }, End: ast.Location{ - Line: int(921), + Line: int(906), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -122866,19 +115737,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(921), + Line: int(906), Column: int(20), }, End: ast.Location{ - Line: int(921), + Line: int(906), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, Value: false, @@ -122888,13 +115759,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(922), + Line: int(907), Column: int(9), }, End: ast.Location{ - Line: int(922), + Line: int(907), Column: int(16), }, File: p1, @@ -122907,13 +115778,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, Value: "false", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -122926,19 +115796,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(923), + Line: int(908), Column: int(12), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -122951,19 +115821,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(923), + Line: int(908), Column: int(15), }, End: ast.Location{ - Line: int(923), + Line: int(908), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -122971,19 +115841,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(923), + Line: int(908), Column: int(15), }, End: ast.Location{ - Line: int(923), + Line: int(908), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -122995,19 +115865,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(923), + Line: int(908), Column: int(20), }, End: ast.Location{ - Line: int(923), + Line: int(908), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, }, @@ -123016,13 +115886,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(924), + Line: int(909), Column: int(9), }, End: ast.Location{ - Line: int(924), + Line: int(909), Column: int(15), }, File: p1, @@ -123035,13 +115905,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, Value: "null", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -123054,19 +115923,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(925), + Line: int(910), Column: int(12), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -123079,19 +115948,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(925), + Line: int(910), Column: int(15), }, End: ast.Location{ - Line: int(925), + Line: int(910), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", "v", @@ -123100,19 +115969,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(925), + Line: int(910), Column: int(15), }, End: ast.Location{ - Line: int(925), + Line: int(910), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -123120,13 +115989,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(925), + Line: int(910), Column: int(15), }, End: ast.Location{ - Line: int(925), + Line: int(910), Column: int(18), }, File: p1, @@ -123161,9 +116030,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -123173,19 +116041,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(925), + Line: int(910), Column: int(28), }, End: ast.Location{ - Line: int(925), + Line: int(910), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9542, + Ctx: p9505, FreeVars: ast.Identifiers{ "v", }, @@ -123206,19 +116074,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(926), + Line: int(911), Column: int(9), }, End: ast.Location{ - Line: int(926), + Line: int(911), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -123226,13 +116094,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(926), + Line: int(911), Column: int(9), }, End: ast.Location{ - Line: int(926), + Line: int(911), Column: int(11), }, File: p1, @@ -123245,32 +116113,31 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(926), + Line: int(911), Column: int(14), }, End: ast.Location{ - Line: int(926), + Line: int(911), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "v", }, @@ -123289,19 +116156,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(927), + Line: int(912), Column: int(12), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -123314,19 +116181,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(927), + Line: int(912), Column: int(15), }, End: ast.Location{ - Line: int(927), + Line: int(912), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", "v", @@ -123335,19 +116202,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(927), + Line: int(912), Column: int(15), }, End: ast.Location{ - Line: int(927), + Line: int(912), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -123355,13 +116222,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(927), + Line: int(912), Column: int(15), }, End: ast.Location{ - Line: int(927), + Line: int(912), Column: int(18), }, File: p1, @@ -123396,9 +116263,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -123408,19 +116274,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(927), + Line: int(912), Column: int(28), }, End: ast.Location{ - Line: int(927), + Line: int(912), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9562, + Ctx: p9525, FreeVars: ast.Identifiers{ "v", }, @@ -123441,19 +116307,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(928), + Line: int(913), Column: int(9), }, End: ast.Location{ - Line: int(928), + Line: int(913), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", "v", @@ -123462,19 +116328,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(928), + Line: int(913), Column: int(9), }, End: ast.Location{ - Line: int(928), + Line: int(913), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -123482,13 +116348,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(928), + Line: int(913), Column: int(9), }, End: ast.Location{ - Line: int(928), + Line: int(913), Column: int(12), }, File: p1, @@ -123530,9 +116396,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringJson", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -123542,19 +116407,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(928), + Line: int(913), Column: int(30), }, End: ast.Location{ - Line: int(928), + Line: int(913), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9574, + Ctx: p9537, FreeVars: ast.Identifiers{ "v", }, @@ -123582,19 +116447,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(929), + Line: int(914), Column: int(12), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -123607,19 +116472,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(929), + Line: int(914), Column: int(15), }, End: ast.Location{ - Line: int(929), + Line: int(914), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", "v", @@ -123628,19 +116493,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(929), + Line: int(914), Column: int(15), }, End: ast.Location{ - Line: int(929), + Line: int(914), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -123648,13 +116513,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(929), + Line: int(914), Column: int(15), }, End: ast.Location{ - Line: int(929), + Line: int(914), Column: int(18), }, File: p1, @@ -123689,9 +116554,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -123701,19 +116565,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(929), + Line: int(914), Column: int(30), }, End: ast.Location{ - Line: int(929), + Line: int(914), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9588, + Ctx: p9551, FreeVars: ast.Identifiers{ "v", }, @@ -123734,13 +116598,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(930), + Line: int(915), Column: int(9), }, End: ast.Location{ - Line: int(930), + Line: int(915), Column: int(54), }, File: p1, @@ -123753,7 +116617,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "path", }, @@ -123761,19 +116625,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(930), + Line: int(915), Column: int(15), }, End: ast.Location{ - Line: int(930), + Line: int(915), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "path", }, @@ -123781,44 +116645,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(930), + Line: int(915), Column: int(15), }, End: ast.Location{ - Line: int(930), + Line: int(915), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: nil, }, Value: "Tried to manifest function at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(930), + Line: int(915), Column: int(50), }, End: ast.Location{ - Line: int(930), + Line: int(915), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "path", }, @@ -123838,19 +116701,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(931), + Line: int(916), Column: int(12), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -123863,19 +116726,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(931), + Line: int(916), Column: int(15), }, End: ast.Location{ - Line: int(931), + Line: int(916), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", "v", @@ -123884,19 +116747,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(931), + Line: int(916), Column: int(15), }, End: ast.Location{ - Line: int(931), + Line: int(916), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -123904,13 +116767,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(931), + Line: int(916), Column: int(15), }, End: ast.Location{ - Line: int(931), + Line: int(916), Column: int(18), }, File: p1, @@ -123945,9 +116808,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -123957,19 +116819,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(931), + Line: int(916), Column: int(27), }, End: ast.Location{ - Line: int(931), + Line: int(916), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9610, + Ctx: p9573, FreeVars: ast.Identifiers{ "v", }, @@ -123990,13 +116852,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(9), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(28), }, File: p1, @@ -124009,7 +116871,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -124027,19 +116889,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(23), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9617, + Ctx: p9580, FreeVars: ast.Identifiers{ "std", "v", @@ -124048,19 +116910,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(23), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9617, + Ctx: p9580, FreeVars: ast.Identifiers{ "std", }, @@ -124068,13 +116930,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(23), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(26), }, File: p1, @@ -124109,9 +116971,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "range", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -124121,21 +116982,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(33), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9626, + Ctx: p9589, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -124144,19 +117006,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(36), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9626, + Ctx: p9589, FreeVars: ast.Identifiers{ "std", "v", @@ -124165,19 +117027,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(36), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9626, + Ctx: p9589, FreeVars: ast.Identifiers{ "std", "v", @@ -124186,19 +117048,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(36), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9626, + Ctx: p9589, FreeVars: ast.Identifiers{ "std", }, @@ -124206,13 +117068,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(36), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(39), }, File: p1, @@ -124247,9 +117109,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -124259,19 +117120,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(47), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9638, + Ctx: p9601, FreeVars: ast.Identifiers{ "v", }, @@ -124293,21 +117154,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(932), + Line: int(917), Column: int(52), }, End: ast.Location{ - Line: int(932), + Line: int(917), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9626, + Ctx: p9589, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -124323,30 +117185,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(932), - Column: int(15), - }, - End: ast.Location{ - Line: int(932), - Column: int(54), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(933), + Line: int(918), Column: int(9), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(28), }, File: p1, @@ -124359,7 +117209,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -124378,19 +117228,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(933), + Line: int(918), Column: int(28), }, End: ast.Location{ - Line: int(933), + Line: int(918), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9646, + Ctx: p9609, FreeVars: ast.Identifiers{ "cindent", "indent", @@ -124399,19 +117249,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(933), + Line: int(918), Column: int(28), }, End: ast.Location{ - Line: int(933), + Line: int(918), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9646, + Ctx: p9609, FreeVars: ast.Identifiers{ "cindent", }, @@ -124423,19 +117273,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(933), + Line: int(918), Column: int(38), }, End: ast.Location{ - Line: int(933), + Line: int(918), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9646, + Ctx: p9609, FreeVars: ast.Identifiers{ "indent", }, @@ -124445,30 +117295,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(933), - Column: int(15), - }, - End: ast.Location{ - Line: int(933), - Column: int(44), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(934), + Line: int(919), Column: int(9), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(28), }, File: p1, @@ -124481,7 +117319,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -124500,19 +117338,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(934), + Line: int(919), Column: int(23), }, End: ast.Location{ - Line: int(940), + Line: int(925), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9657, + Ctx: p9620, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -124526,19 +117364,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(934), + Line: int(919), Column: int(23), }, End: ast.Location{ - Line: int(939), + Line: int(924), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9657, + Ctx: p9620, FreeVars: ast.Identifiers{ "aux", "new_indent", @@ -124551,19 +117389,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(934), + Line: int(919), Column: int(23), }, End: ast.Location{ - Line: int(934), + Line: int(919), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9657, + Ctx: p9620, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -124571,25 +117409,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(934), + Line: int(919), Column: int(24), }, End: ast.Location{ - Line: int(934), + Line: int(919), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9664, + Ctx: p9627, FreeVars: nil, }, Value: "[\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -124609,19 +117446,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(935), + Line: int(920), Column: int(25), }, End: ast.Location{ - Line: int(939), + Line: int(924), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9657, + Ctx: p9620, FreeVars: ast.Identifiers{ "aux", "new_indent", @@ -124634,19 +117471,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(935), + Line: int(920), Column: int(25), }, End: ast.Location{ - Line: int(935), + Line: int(920), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9657, + Ctx: p9620, FreeVars: ast.Identifiers{ "std", }, @@ -124654,13 +117491,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(935), + Line: int(920), Column: int(25), }, End: ast.Location{ - Line: int(935), + Line: int(920), Column: int(28), }, File: p1, @@ -124695,9 +117532,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -124707,19 +117543,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(935), + Line: int(920), Column: int(34), }, End: ast.Location{ - Line: int(935), + Line: int(920), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9675, + Ctx: p9638, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -124727,25 +117563,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(935), + Line: int(920), Column: int(35), }, End: ast.Location{ - Line: int(935), + Line: int(920), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9678, + Ctx: p9641, FreeVars: nil, }, Value: ",\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -124761,14 +117596,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(936), - Column: int(34), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(939), - Column: int(35), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -124845,7 +117680,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -124878,26 +117712,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -124930,13 +117753,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(36), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(84), }, File: p1, @@ -124949,7 +117772,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9695, + Ctx: p9658, FreeVars: ast.Identifiers{ "aux", "i", @@ -124963,19 +117786,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(37), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(83), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9699, + Ctx: p9662, FreeVars: ast.Identifiers{ "aux", "i", @@ -124987,19 +117810,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(37), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9699, + Ctx: p9662, FreeVars: ast.Identifiers{ "new_indent", }, @@ -125011,19 +117834,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(50), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(83), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9699, + Ctx: p9662, FreeVars: ast.Identifiers{ "aux", "i", @@ -125035,19 +117858,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(50), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9699, + Ctx: p9662, FreeVars: ast.Identifiers{ "aux", }, @@ -125061,19 +117884,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(54), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9709, + Ctx: p9672, FreeVars: ast.Identifiers{ "i", "v", @@ -125082,19 +117905,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(54), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9709, + Ctx: p9672, FreeVars: ast.Identifiers{ "v", }, @@ -125105,19 +117928,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(56), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9709, + Ctx: p9672, FreeVars: ast.Identifiers{ "i", }, @@ -125133,19 +117956,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(60), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9709, + Ctx: p9672, FreeVars: ast.Identifiers{ "i", "path", @@ -125154,19 +117977,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(60), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9709, + Ctx: p9672, FreeVars: ast.Identifiers{ "path", }, @@ -125178,19 +118001,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(67), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9709, + Ctx: p9672, FreeVars: ast.Identifiers{ "i", }, @@ -125200,19 +118023,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(68), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9723, + Ctx: p9686, FreeVars: ast.Identifiers{ "i", }, @@ -125232,19 +118055,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(937), + Line: int(922), Column: int(72), }, End: ast.Location{ - Line: int(937), + Line: int(922), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9709, + Ctx: p9672, FreeVars: ast.Identifiers{ "new_indent", }, @@ -125281,19 +118104,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(938), + Line: int(923), Column: int(45), }, End: ast.Location{ - Line: int(938), + Line: int(923), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9675, + Ctx: p9638, FreeVars: ast.Identifiers{ "range", }, @@ -125333,19 +118156,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(940), + Line: int(925), Column: int(25), }, End: ast.Location{ - Line: int(940), + Line: int(925), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9657, + Ctx: p9620, FreeVars: ast.Identifiers{ "cindent", }, @@ -125355,19 +118178,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(940), + Line: int(925), Column: int(26), }, End: ast.Location{ - Line: int(940), + Line: int(925), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9734, + Ctx: p9697, FreeVars: ast.Identifiers{ "cindent", }, @@ -125375,19 +118198,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(940), + Line: int(925), Column: int(26), }, End: ast.Location{ - Line: int(940), + Line: int(925), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9734, + Ctx: p9697, FreeVars: ast.Identifiers{ "cindent", }, @@ -125395,44 +118218,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(940), + Line: int(925), Column: int(26), }, End: ast.Location{ - Line: int(940), + Line: int(925), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9734, + Ctx: p9697, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(940), + Line: int(925), Column: int(33), }, End: ast.Location{ - Line: int(940), + Line: int(925), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9734, + Ctx: p9697, FreeVars: ast.Identifiers{ "cindent", }, @@ -125445,25 +118267,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(940), + Line: int(925), Column: int(43), }, End: ast.Location{ - Line: int(940), + Line: int(925), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9734, + Ctx: p9697, FreeVars: nil, }, Value: "]", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, CommaFodder: nil, @@ -125475,36 +118296,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(934), - Column: int(15), - }, - End: ast.Location{ - Line: int(940), - Column: int(47), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(941), + Line: int(926), Column: int(9), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "lines", "std", @@ -125513,19 +118322,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(941), + Line: int(926), Column: int(9), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -125533,13 +118342,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(941), + Line: int(926), Column: int(9), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(12), }, File: p1, @@ -125581,9 +118390,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -125593,25 +118401,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(941), + Line: int(926), Column: int(18), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9752, + Ctx: p9715, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -125619,19 +118426,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(941), + Line: int(926), Column: int(22), }, End: ast.Location{ - Line: int(941), + Line: int(926), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9752, + Ctx: p9715, FreeVars: ast.Identifiers{ "lines", }, @@ -125662,19 +118469,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(942), + Line: int(927), Column: int(12), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -125687,19 +118494,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(942), + Line: int(927), Column: int(15), }, End: ast.Location{ - Line: int(942), + Line: int(927), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", "v", @@ -125708,19 +118515,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(942), + Line: int(927), Column: int(15), }, End: ast.Location{ - Line: int(942), + Line: int(927), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -125728,13 +118535,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(942), + Line: int(927), Column: int(15), }, End: ast.Location{ - Line: int(942), + Line: int(927), Column: int(18), }, File: p1, @@ -125769,9 +118576,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -125781,19 +118587,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(942), + Line: int(927), Column: int(28), }, End: ast.Location{ - Line: int(942), + Line: int(927), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9767, + Ctx: p9730, FreeVars: ast.Identifiers{ "v", }, @@ -125814,13 +118620,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(943), + Line: int(928), Column: int(9), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, @@ -125833,7 +118639,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -125851,19 +118657,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(943), + Line: int(928), Column: int(23), }, End: ast.Location{ - Line: int(950), + Line: int(935), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9774, + Ctx: p9737, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -125876,19 +118682,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(943), + Line: int(928), Column: int(23), }, End: ast.Location{ - Line: int(949), + Line: int(934), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9774, + Ctx: p9737, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -125901,19 +118707,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(943), + Line: int(928), Column: int(23), }, End: ast.Location{ - Line: int(943), + Line: int(928), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9774, + Ctx: p9737, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -125921,25 +118727,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(943), + Line: int(928), Column: int(24), }, End: ast.Location{ - Line: int(943), + Line: int(928), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9781, + Ctx: p9744, FreeVars: nil, }, Value: "{\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -125959,19 +118764,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(944), + Line: int(929), Column: int(25), }, End: ast.Location{ - Line: int(949), + Line: int(934), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9774, + Ctx: p9737, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -125984,19 +118789,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(944), + Line: int(929), Column: int(25), }, End: ast.Location{ - Line: int(944), + Line: int(929), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9774, + Ctx: p9737, FreeVars: ast.Identifiers{ "std", }, @@ -126004,13 +118809,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(944), + Line: int(929), Column: int(25), }, End: ast.Location{ - Line: int(944), + Line: int(929), Column: int(28), }, File: p1, @@ -126045,9 +118850,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -126057,19 +118861,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(944), + Line: int(929), Column: int(34), }, End: ast.Location{ - Line: int(944), + Line: int(929), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9792, + Ctx: p9755, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -126077,25 +118881,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(944), + Line: int(929), Column: int(35), }, End: ast.Location{ - Line: int(944), + Line: int(929), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9795, + Ctx: p9758, FreeVars: nil, }, Value: ",\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -126111,14 +118914,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(945), - Column: int(34), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(949), - Column: int(35), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -126195,7 +118998,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -126230,26 +119032,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -126284,13 +119075,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(36), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(79), }, File: p1, @@ -126303,7 +119094,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9812, + Ctx: p9775, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -126319,19 +119110,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(37), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -126345,19 +119136,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(37), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "cindent", "indent", @@ -126368,19 +119159,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(37), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "cindent", "indent", @@ -126391,19 +119182,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(37), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "cindent", "indent", @@ -126412,19 +119203,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(37), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "cindent", }, @@ -126436,19 +119227,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(47), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "indent", }, @@ -126461,19 +119252,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(56), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "k", "std", @@ -126482,19 +119273,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(56), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "std", }, @@ -126502,13 +119293,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(56), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(59), }, File: p1, @@ -126543,9 +119334,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringJson", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -126555,19 +119345,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(77), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9837, + Ctx: p9800, FreeVars: ast.Identifiers{ "k", }, @@ -126590,25 +119380,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(946), + Line: int(931), Column: int(82), }, End: ast.Location{ - Line: int(946), + Line: int(931), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: nil, }, Value: ": ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{ @@ -126623,19 +119412,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(39), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -126648,19 +119437,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(39), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9816, + Ctx: p9779, FreeVars: ast.Identifiers{ "aux", }, @@ -126674,19 +119463,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(43), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "k", "v", @@ -126695,19 +119484,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(43), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "v", }, @@ -126718,19 +119507,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(45), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "k", }, @@ -126746,19 +119535,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(49), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "k", "path", @@ -126767,19 +119556,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(49), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "path", }, @@ -126791,19 +119580,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(56), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "k", }, @@ -126813,19 +119602,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(57), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9861, + Ctx: p9824, FreeVars: ast.Identifiers{ "k", }, @@ -126845,19 +119634,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(61), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "cindent", "indent", @@ -126866,19 +119655,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(61), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "cindent", }, @@ -126890,19 +119679,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(947), + Line: int(932), Column: int(71), }, End: ast.Location{ - Line: int(947), + Line: int(932), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9847, + Ctx: p9810, FreeVars: ast.Identifiers{ "indent", }, @@ -126940,19 +119729,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(948), + Line: int(933), Column: int(45), }, End: ast.Location{ - Line: int(948), + Line: int(933), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9792, + Ctx: p9755, FreeVars: ast.Identifiers{ "std", "v", @@ -126961,19 +119750,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(948), + Line: int(933), Column: int(45), }, End: ast.Location{ - Line: int(948), + Line: int(933), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9792, + Ctx: p9755, FreeVars: ast.Identifiers{ "std", }, @@ -126981,13 +119770,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(948), + Line: int(933), Column: int(45), }, End: ast.Location{ - Line: int(948), + Line: int(933), Column: int(48), }, File: p1, @@ -127022,9 +119811,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -127034,19 +119822,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(948), + Line: int(933), Column: int(62), }, End: ast.Location{ - Line: int(948), + Line: int(933), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9878, + Ctx: p9841, FreeVars: ast.Identifiers{ "v", }, @@ -127096,19 +119884,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(950), + Line: int(935), Column: int(25), }, End: ast.Location{ - Line: int(950), + Line: int(935), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9774, + Ctx: p9737, FreeVars: ast.Identifiers{ "cindent", }, @@ -127118,19 +119906,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(950), + Line: int(935), Column: int(26), }, End: ast.Location{ - Line: int(950), + Line: int(935), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9885, + Ctx: p9848, FreeVars: ast.Identifiers{ "cindent", }, @@ -127138,19 +119926,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(950), + Line: int(935), Column: int(26), }, End: ast.Location{ - Line: int(950), + Line: int(935), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9885, + Ctx: p9848, FreeVars: ast.Identifiers{ "cindent", }, @@ -127158,44 +119946,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(950), + Line: int(935), Column: int(26), }, End: ast.Location{ - Line: int(950), + Line: int(935), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9885, + Ctx: p9848, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(950), + Line: int(935), Column: int(33), }, End: ast.Location{ - Line: int(950), + Line: int(935), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9885, + Ctx: p9848, FreeVars: ast.Identifiers{ "cindent", }, @@ -127208,25 +119995,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(950), + Line: int(935), Column: int(43), }, End: ast.Location{ - Line: int(950), + Line: int(935), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9885, + Ctx: p9848, FreeVars: nil, }, Value: "}", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, CommaFodder: nil, @@ -127238,36 +120024,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(943), - Column: int(15), - }, - End: ast.Location{ - Line: int(950), - Column: int(47), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(951), + Line: int(936), Column: int(9), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "lines", "std", @@ -127276,19 +120050,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(951), + Line: int(936), Column: int(9), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9501, + Ctx: p9464, FreeVars: ast.Identifiers{ "std", }, @@ -127296,13 +120070,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(951), + Line: int(936), Column: int(9), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(12), }, File: p1, @@ -127344,9 +120118,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -127356,25 +120129,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(951), + Line: int(936), Column: int(18), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9903, + Ctx: p9866, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -127382,19 +120154,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(951), + Line: int(936), Column: int(22), }, End: ast.Location{ - Line: int(951), + Line: int(936), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9903, + Ctx: p9866, FreeVars: ast.Identifiers{ "lines", }, @@ -127443,36 +120215,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(952), + Line: int(937), Column: int(5), }, End: ast.Location{ - Line: int(952), + Line: int(937), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9492, + Ctx: p9455, FreeVars: ast.Identifiers{ "aux", "value", @@ -127481,13 +120241,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(952), + Line: int(937), Column: int(5), }, End: ast.Location{ - Line: int(952), + Line: int(937), Column: int(8), }, File: p1, @@ -127500,7 +120260,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9492, + Ctx: p9455, FreeVars: ast.Identifiers{ "aux", }, @@ -127514,19 +120274,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(952), + Line: int(937), Column: int(9), }, End: ast.Location{ - Line: int(952), + Line: int(937), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9914, + Ctx: p9877, FreeVars: ast.Identifiers{ "value", }, @@ -127539,19 +120299,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(952), + Line: int(937), Column: int(16), }, End: ast.Location{ - Line: int(952), + Line: int(937), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9914, + Ctx: p9877, FreeVars: nil, }, Elements: nil, @@ -127564,25 +120324,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(952), + Line: int(937), Column: int(20), }, End: ast.Location{ - Line: int(952), + Line: int(937), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9914, + Ctx: p9877, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -127597,18 +120356,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(917), - Column: int(3), - }, - End: ast.Location{ - Line: int(952), - Column: int(23), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -127633,7 +120380,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestYamlDoc", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -127656,62 +120402,40 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(954), - Column: int(19), - }, - End: ast.Location{ - Line: int(954), - Column: int(24), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "indent_array_in_object", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.LiteralBoolean{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(954), - Column: int(49), - }, - End: ast.Location{ - Line: int(954), - Column: int(54), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "indent_array_in_object", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.LiteralBoolean{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(939), + Column: int(49), + }, + End: ast.Location{ + Line: int(939), + Column: int(54), + }, + File: p1, }, - File: p1, + Fodder: ast.Fodder{}, + Ctx: p9887, + FreeVars: nil, }, - Fodder: ast.Fodder{}, - Ctx: p9923, - FreeVars: nil, - }, - Value: false, - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(954), - Column: int(26), - }, - End: ast.Location{ - Line: int(954), - Column: int(54), + Value: false, }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -127720,13 +120444,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(955), + Line: int(940), Column: int(5), }, End: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(23), }, File: p1, @@ -127739,7 +120463,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9923, + Ctx: p9887, FreeVars: ast.Identifiers{ "indent_array_in_object", "std", @@ -127754,19 +120478,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(955), + Line: int(940), Column: int(11), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: nil, - Ctx: p9929, + Ctx: p9893, FreeVars: ast.Identifiers{ "aux", "indent_array_in_object", @@ -127774,77 +120498,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(955), - Column: int(15), - }, - End: ast.Location{ - Line: int(955), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "path", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(955), - Column: int(18), - }, - End: ast.Location{ - Line: int(955), - Column: int(22), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "path", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "cindent", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(955), - Column: int(24), - }, - End: ast.Location{ - Line: int(955), - Column: int(31), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "cindent", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(956), + Line: int(941), Column: int(7), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, @@ -127857,7 +120542,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -127870,19 +120555,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(956), + Line: int(941), Column: int(10), }, End: ast.Location{ - Line: int(956), + Line: int(941), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -127890,19 +120575,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(956), + Line: int(941), Column: int(10), }, End: ast.Location{ - Line: int(956), + Line: int(941), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -127914,19 +120599,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(956), + Line: int(941), Column: int(15), }, End: ast.Location{ - Line: int(956), + Line: int(941), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: true, @@ -127936,13 +120621,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(957), + Line: int(942), Column: int(9), }, End: ast.Location{ - Line: int(957), + Line: int(942), Column: int(15), }, File: p1, @@ -127955,13 +120640,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "true", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -127974,19 +120658,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(958), + Line: int(943), Column: int(12), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -127999,19 +120683,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(958), + Line: int(943), Column: int(15), }, End: ast.Location{ - Line: int(958), + Line: int(943), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -128019,19 +120703,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(958), + Line: int(943), Column: int(15), }, End: ast.Location{ - Line: int(958), + Line: int(943), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -128043,19 +120727,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(958), + Line: int(943), Column: int(20), }, End: ast.Location{ - Line: int(958), + Line: int(943), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: false, @@ -128065,13 +120749,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(959), + Line: int(944), Column: int(9), }, End: ast.Location{ - Line: int(959), + Line: int(944), Column: int(16), }, File: p1, @@ -128084,13 +120768,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "false", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -128103,19 +120786,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(960), + Line: int(945), Column: int(12), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -128128,19 +120811,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(960), + Line: int(945), Column: int(15), }, End: ast.Location{ - Line: int(960), + Line: int(945), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -128148,19 +120831,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(960), + Line: int(945), Column: int(15), }, End: ast.Location{ - Line: int(960), + Line: int(945), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -128172,19 +120855,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(960), + Line: int(945), Column: int(20), }, End: ast.Location{ - Line: int(960), + Line: int(945), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, }, @@ -128193,13 +120876,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(961), + Line: int(946), Column: int(9), }, End: ast.Location{ - Line: int(961), + Line: int(946), Column: int(15), }, File: p1, @@ -128212,13 +120895,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "null", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -128231,19 +120913,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(962), + Line: int(947), Column: int(12), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -128256,19 +120938,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(962), + Line: int(947), Column: int(15), }, End: ast.Location{ - Line: int(962), + Line: int(947), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -128277,19 +120959,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(962), + Line: int(947), Column: int(15), }, End: ast.Location{ - Line: int(962), + Line: int(947), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -128297,13 +120979,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(962), + Line: int(947), Column: int(15), }, End: ast.Location{ - Line: int(962), + Line: int(947), Column: int(18), }, File: p1, @@ -128338,9 +121020,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -128350,19 +121031,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(962), + Line: int(947), Column: int(28), }, End: ast.Location{ - Line: int(962), + Line: int(947), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9975, + Ctx: p9939, FreeVars: ast.Identifiers{ "v", }, @@ -128383,19 +121064,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(963), + Line: int(948), Column: int(9), }, End: ast.Location{ - Line: int(963), + Line: int(948), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -128403,13 +121084,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(963), + Line: int(948), Column: int(9), }, End: ast.Location{ - Line: int(963), + Line: int(948), Column: int(11), }, File: p1, @@ -128422,32 +121103,31 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(963), + Line: int(948), Column: int(14), }, End: ast.Location{ - Line: int(963), + Line: int(948), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -128466,19 +121146,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(964), + Line: int(949), Column: int(12), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -128491,19 +121171,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(964), + Line: int(949), Column: int(15), }, End: ast.Location{ - Line: int(964), + Line: int(949), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -128512,19 +121192,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(964), + Line: int(949), Column: int(15), }, End: ast.Location{ - Line: int(964), + Line: int(949), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -128532,13 +121212,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(964), + Line: int(949), Column: int(15), }, End: ast.Location{ - Line: int(964), + Line: int(949), Column: int(18), }, File: p1, @@ -128573,9 +121253,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -128585,19 +121264,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(964), + Line: int(949), Column: int(28), }, End: ast.Location{ - Line: int(964), + Line: int(949), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9995, + Ctx: p9959, FreeVars: ast.Identifiers{ "v", }, @@ -128618,13 +121297,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(965), + Line: int(950), Column: int(9), }, End: ast.Location{ - Line: int(972), + Line: int(957), Column: int(34), }, File: p1, @@ -128637,7 +121316,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "cindent", "std", @@ -128652,19 +121331,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(965), + Line: int(950), Column: int(21), }, End: ast.Location{ - Line: int(965), + Line: int(950), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10002, + Ctx: p9966, FreeVars: ast.Identifiers{ "std", "v", @@ -128673,19 +121352,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(965), + Line: int(950), Column: int(21), }, End: ast.Location{ - Line: int(965), + Line: int(950), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10002, + Ctx: p9966, FreeVars: ast.Identifiers{ "std", }, @@ -128693,13 +121372,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(965), + Line: int(950), Column: int(21), }, End: ast.Location{ - Line: int(965), + Line: int(950), Column: int(24), }, File: p1, @@ -128734,9 +121413,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -128746,19 +121424,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(965), + Line: int(950), Column: int(32), }, End: ast.Location{ - Line: int(965), + Line: int(950), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10011, + Ctx: p9975, FreeVars: ast.Identifiers{ "v", }, @@ -128777,30 +121455,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(965), - Column: int(15), - }, - End: ast.Location{ - Line: int(965), - Column: int(34), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(966), + Line: int(951), Column: int(9), }, End: ast.Location{ - Line: int(972), + Line: int(957), Column: int(34), }, File: p1, @@ -128813,7 +121479,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "cindent", "len", @@ -128824,19 +121490,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(966), + Line: int(951), Column: int(12), }, End: ast.Location{ - Line: int(966), + Line: int(951), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "len", }, @@ -128844,19 +121510,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(966), + Line: int(951), Column: int(12), }, End: ast.Location{ - Line: int(966), + Line: int(951), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "len", }, @@ -128868,21 +121534,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(966), + Line: int(951), Column: int(19), }, End: ast.Location{ - Line: int(966), + Line: int(951), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -128890,13 +121557,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(967), + Line: int(952), Column: int(11), }, End: ast.Location{ - Line: int(967), + Line: int(952), Column: int(15), }, File: p1, @@ -128909,13 +121576,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "\"\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -128928,19 +121594,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(14), }, End: ast.Location{ - Line: int(972), + Line: int(957), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "cindent", "len", @@ -128951,19 +121617,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(17), }, End: ast.Location{ - Line: int(968), + Line: int(953), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "len", "v", @@ -128972,19 +121638,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(17), }, End: ast.Location{ - Line: int(968), + Line: int(953), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "len", "v", @@ -128993,19 +121659,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(17), }, End: ast.Location{ - Line: int(968), + Line: int(953), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "v", }, @@ -129016,19 +121682,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(19), }, End: ast.Location{ - Line: int(968), + Line: int(953), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "len", }, @@ -129036,19 +121702,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(19), }, End: ast.Location{ - Line: int(968), + Line: int(953), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "len", }, @@ -129060,21 +121726,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(25), }, End: ast.Location{ - Line: int(968), + Line: int(953), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -129086,38 +121753,37 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(968), + Line: int(953), Column: int(31), }, End: ast.Location{ - Line: int(968), + Line: int(953), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(969), + Line: int(954), Column: int(11), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(82), }, File: p1, @@ -129130,7 +121796,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "cindent", "std", @@ -129145,19 +121811,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(969), + Line: int(954), Column: int(25), }, End: ast.Location{ - Line: int(969), + Line: int(954), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10043, + Ctx: p10007, FreeVars: ast.Identifiers{ "std", "v", @@ -129166,19 +121832,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(969), + Line: int(954), Column: int(25), }, End: ast.Location{ - Line: int(969), + Line: int(954), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10043, + Ctx: p10007, FreeVars: ast.Identifiers{ "std", }, @@ -129186,13 +121852,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(969), + Line: int(954), Column: int(25), }, End: ast.Location{ - Line: int(969), + Line: int(954), Column: int(28), }, File: p1, @@ -129227,9 +121893,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "split", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -129239,19 +121904,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(969), + Line: int(954), Column: int(35), }, End: ast.Location{ - Line: int(969), + Line: int(954), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10052, + Ctx: p10016, FreeVars: ast.Identifiers{ "v", }, @@ -129264,25 +121929,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(969), + Line: int(954), Column: int(38), }, End: ast.Location{ - Line: int(969), + Line: int(954), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10052, + Ctx: p10016, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -129296,36 +121960,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(969), - Column: int(17), - }, - End: ast.Location{ - Line: int(969), - Column: int(43), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(11), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "cindent", "split", @@ -129335,19 +121987,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(11), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -129355,13 +122007,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(11), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(14), }, File: p1, @@ -129403,9 +122055,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -129415,19 +122066,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(20), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "cindent", }, @@ -129435,19 +122086,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(20), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "cindent", }, @@ -129455,44 +122106,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(20), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(27), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "cindent", }, @@ -129505,25 +122155,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(37), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, CommaFodder: ast.Fodder{}, @@ -129532,19 +122181,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(43), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(81), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "split", "std", @@ -129553,19 +122202,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(43), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -129573,25 +122222,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(44), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10078, + Ctx: p10042, FreeVars: nil, }, Value: "|", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -129606,14 +122254,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(970), - Column: int(51), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(970), - Column: int(81), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -129686,7 +122334,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -129698,19 +122345,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(51), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "split", }, @@ -129723,21 +122370,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(57), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: nil, @@ -129746,19 +122394,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(59), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "split", "std", @@ -129767,19 +122415,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(59), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "split", "std", @@ -129788,19 +122436,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(59), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: ast.Identifiers{ "std", }, @@ -129808,13 +122456,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(59), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(62), }, File: p1, @@ -129849,9 +122497,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -129861,19 +122508,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(70), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10101, + Ctx: p10065, FreeVars: ast.Identifiers{ "split", }, @@ -129895,21 +122542,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(970), + Line: int(955), Column: int(79), }, End: ast.Location{ - Line: int(970), + Line: int(955), Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10065, + Ctx: p10029, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -129968,19 +122616,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(972), + Line: int(957), Column: int(11), }, End: ast.Location{ - Line: int(972), + Line: int(957), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -129989,19 +122637,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(972), + Line: int(957), Column: int(11), }, End: ast.Location{ - Line: int(972), + Line: int(957), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -130009,13 +122657,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(972), + Line: int(957), Column: int(11), }, End: ast.Location{ - Line: int(972), + Line: int(957), Column: int(14), }, File: p1, @@ -130057,9 +122705,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringJson", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -130069,19 +122716,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(972), + Line: int(957), Column: int(32), }, End: ast.Location{ - Line: int(972), + Line: int(957), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10116, + Ctx: p10080, FreeVars: ast.Identifiers{ "v", }, @@ -130112,19 +122759,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(973), + Line: int(958), Column: int(12), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -130137,19 +122784,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(973), + Line: int(958), Column: int(15), }, End: ast.Location{ - Line: int(973), + Line: int(958), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -130158,19 +122805,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(973), + Line: int(958), Column: int(15), }, End: ast.Location{ - Line: int(973), + Line: int(958), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -130178,13 +122825,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(973), + Line: int(958), Column: int(15), }, End: ast.Location{ - Line: int(973), + Line: int(958), Column: int(18), }, File: p1, @@ -130219,9 +122866,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -130231,19 +122877,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(973), + Line: int(958), Column: int(30), }, End: ast.Location{ - Line: int(973), + Line: int(958), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10130, + Ctx: p10094, FreeVars: ast.Identifiers{ "v", }, @@ -130264,13 +122910,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(974), + Line: int(959), Column: int(9), }, End: ast.Location{ - Line: int(974), + Line: int(959), Column: int(54), }, File: p1, @@ -130283,7 +122929,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "path", }, @@ -130291,19 +122937,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(974), + Line: int(959), Column: int(15), }, End: ast.Location{ - Line: int(974), + Line: int(959), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "path", }, @@ -130311,44 +122957,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(974), + Line: int(959), Column: int(15), }, End: ast.Location{ - Line: int(974), + Line: int(959), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "Tried to manifest function at ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(974), + Line: int(959), Column: int(50), }, End: ast.Location{ - Line: int(974), + Line: int(959), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "path", }, @@ -130368,19 +123013,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(975), + Line: int(960), Column: int(12), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -130393,19 +123038,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(975), + Line: int(960), Column: int(15), }, End: ast.Location{ - Line: int(975), + Line: int(960), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -130414,19 +123059,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(975), + Line: int(960), Column: int(15), }, End: ast.Location{ - Line: int(975), + Line: int(960), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -130434,13 +123079,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(975), + Line: int(960), Column: int(15), }, End: ast.Location{ - Line: int(975), + Line: int(960), Column: int(18), }, File: p1, @@ -130475,9 +123120,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -130487,19 +123131,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(975), + Line: int(960), Column: int(27), }, End: ast.Location{ - Line: int(975), + Line: int(960), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10152, + Ctx: p10116, FreeVars: ast.Identifiers{ "v", }, @@ -130520,13 +123164,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(976), + Line: int(961), Column: int(9), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(42), }, File: p1, @@ -130539,7 +123183,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -130551,19 +123195,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(976), + Line: int(961), Column: int(12), }, End: ast.Location{ - Line: int(976), + Line: int(961), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -130572,19 +123216,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(976), + Line: int(961), Column: int(12), }, End: ast.Location{ - Line: int(976), + Line: int(961), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -130593,19 +123237,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(976), + Line: int(961), Column: int(12), }, End: ast.Location{ - Line: int(976), + Line: int(961), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -130613,13 +123257,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(976), + Line: int(961), Column: int(12), }, End: ast.Location{ - Line: int(976), + Line: int(961), Column: int(15), }, File: p1, @@ -130654,9 +123298,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -130666,19 +123309,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(976), + Line: int(961), Column: int(23), }, End: ast.Location{ - Line: int(976), + Line: int(961), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10168, + Ctx: p10132, FreeVars: ast.Identifiers{ "v", }, @@ -130700,21 +123343,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(976), + Line: int(961), Column: int(29), }, End: ast.Location{ - Line: int(976), + Line: int(961), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -130722,13 +123366,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(977), + Line: int(962), Column: int(11), }, End: ast.Location{ - Line: int(977), + Line: int(962), Column: int(15), }, File: p1, @@ -130741,13 +123385,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "[]", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -130760,13 +123403,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(979), + Line: int(964), Column: int(11), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(42), }, File: p1, @@ -130779,7 +123422,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -130796,58 +123439,47 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(979), + Line: int(964), Column: int(17), }, End: ast.Location{ - Line: int(999), + Line: int(984), Column: int(14), }, File: p1, }, Fodder: nil, - Ctx: p10179, + Ctx: p10143, FreeVars: ast.Identifiers{ "cindent", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(979), - Column: int(24), - }, - End: ast.Location{ - Line: int(979), - Column: int(29), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(13), }, End: ast.Location{ - Line: int(999), + Line: int(984), Column: int(14), }, File: p1, @@ -130860,7 +123492,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "cindent", "std", @@ -130870,19 +123502,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(16), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -130891,19 +123523,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(16), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -130912,19 +123544,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(16), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", }, @@ -130932,13 +123564,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(16), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(19), }, File: p1, @@ -130973,9 +123605,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -130985,19 +123616,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(28), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10197, + Ctx: p10161, FreeVars: ast.Identifiers{ "value", }, @@ -131019,19 +123650,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(38), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -131040,19 +123671,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(38), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -131061,19 +123692,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(38), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", }, @@ -131081,13 +123712,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(38), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(41), }, File: p1, @@ -131122,9 +123753,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -131134,19 +123764,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(49), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10210, + Ctx: p10174, FreeVars: ast.Identifiers{ "value", }, @@ -131168,21 +123798,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(58), }, End: ast.Location{ - Line: int(980), + Line: int(965), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -131191,19 +123822,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(980), + Line: int(965), Column: int(65), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "cindent", }, @@ -131233,24 +123864,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(987), + Line: int(972), Column: int(27), }, End: ast.Location{ - Line: int(987), + Line: int(972), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10218, + Ctx: p10182, FreeVars: ast.Identifiers{ "cindent", }, @@ -131258,19 +123888,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(987), + Line: int(972), Column: int(27), }, End: ast.Location{ - Line: int(987), + Line: int(972), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10218, + Ctx: p10182, FreeVars: ast.Identifiers{ "cindent", }, @@ -131282,40 +123912,27 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(987), + Line: int(972), Column: int(37), }, End: ast.Location{ - Line: int(987), + Line: int(972), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10218, + Ctx: p10182, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(987), - Column: int(15), - }, - End: ast.Location{ - Line: int(987), - Column: int(41), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -131340,79 +123957,77 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(988), + Line: int(973), Column: int(22), }, End: ast.Location{ - Line: int(988), + Line: int(973), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10218, + Ctx: p10182, FreeVars: nil, }, Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(988), + Line: int(973), Column: int(22), }, End: ast.Location{ - Line: int(988), + Line: int(973), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10218, + Ctx: p10182, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(988), + Line: int(973), Column: int(29), }, End: ast.Location{ - Line: int(988), + Line: int(973), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10218, + Ctx: p10182, FreeVars: nil, }, Target: &ast.Self{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(988), + Line: int(973), Column: int(29), }, End: ast.Location{ - Line: int(988), + Line: int(973), Column: int(33), }, File: p1, @@ -131444,25 +124059,12 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(988), - Column: int(15), - }, - End: ast.Location{ - Line: int(988), - Column: int(44), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -131471,19 +124073,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(20), }, End: ast.Location{ - Line: int(999), + Line: int(984), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "cindent", "std", @@ -131493,19 +124095,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(23), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -131514,19 +124116,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(23), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -131535,19 +124137,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(23), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", }, @@ -131555,13 +124157,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(23), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(26), }, File: p1, @@ -131596,9 +124198,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -131608,19 +124209,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(36), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10242, + Ctx: p10206, FreeVars: ast.Identifiers{ "value", }, @@ -131642,19 +124243,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(46), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -131663,19 +124264,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(46), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", "value", @@ -131684,19 +124285,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(46), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "std", }, @@ -131704,13 +124305,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(46), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(49), }, File: p1, @@ -131745,9 +124346,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -131757,19 +124357,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(57), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10255, + Ctx: p10219, FreeVars: ast.Identifiers{ "value", }, @@ -131791,21 +124391,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(66), }, End: ast.Location{ - Line: int(989), + Line: int(974), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -131814,19 +124415,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(989), + Line: int(974), Column: int(73), }, End: ast.Location{ - Line: int(995), + Line: int(980), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "cindent", }, @@ -131856,24 +124457,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(990), + Line: int(975), Column: int(27), }, End: ast.Location{ - Line: int(990), + Line: int(975), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10263, + Ctx: p10227, FreeVars: ast.Identifiers{ "cindent", }, @@ -131881,19 +124481,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(990), + Line: int(975), Column: int(27), }, End: ast.Location{ - Line: int(990), + Line: int(975), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10263, + Ctx: p10227, FreeVars: ast.Identifiers{ "cindent", }, @@ -131905,40 +124505,27 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(990), + Line: int(975), Column: int(37), }, End: ast.Location{ - Line: int(990), + Line: int(975), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10263, + Ctx: p10227, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(990), - Column: int(15), - }, - End: ast.Location{ - Line: int(990), - Column: int(41), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -131963,44 +124550,30 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(994), + Line: int(979), Column: int(22), }, End: ast.Location{ - Line: int(994), + Line: int(979), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10263, + Ctx: p10227, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(994), - Column: int(15), - }, - End: ast.Location{ - Line: int(994), - Column: int(25), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -132009,19 +124582,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(995), + Line: int(980), Column: int(20), }, End: ast.Location{ - Line: int(999), + Line: int(984), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10184, + Ctx: p10148, FreeVars: ast.Identifiers{ "cindent", }, @@ -132051,24 +124624,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(997), + Line: int(982), Column: int(27), }, End: ast.Location{ - Line: int(997), + Line: int(982), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10275, + Ctx: p10239, FreeVars: ast.Identifiers{ "cindent", }, @@ -132076,18 +124648,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "cindent", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(997), - Column: int(15), - }, - End: ast.Location{ - Line: int(997), - Column: int(34), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -132112,44 +124672,30 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(998), + Line: int(983), Column: int(22), }, End: ast.Location{ - Line: int(998), + Line: int(983), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10275, + Ctx: p10239, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(998), - Column: int(15), - }, - End: ast.Location{ - Line: int(998), - Column: int(25), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -132159,30 +124705,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(11), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(42), }, File: p1, @@ -132195,7 +124729,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -132213,19 +124747,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(25), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10284, + Ctx: p10248, FreeVars: ast.Identifiers{ "std", "v", @@ -132234,19 +124768,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(25), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10284, + Ctx: p10248, FreeVars: ast.Identifiers{ "std", }, @@ -132254,13 +124788,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(25), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(28), }, File: p1, @@ -132295,9 +124829,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "range", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -132307,21 +124840,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(35), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10293, + Ctx: p10257, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -132330,19 +124864,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(38), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10293, + Ctx: p10257, FreeVars: ast.Identifiers{ "std", "v", @@ -132351,19 +124885,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(38), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10293, + Ctx: p10257, FreeVars: ast.Identifiers{ "std", "v", @@ -132372,19 +124906,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(38), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10293, + Ctx: p10257, FreeVars: ast.Identifiers{ "std", }, @@ -132392,13 +124926,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(38), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(41), }, File: p1, @@ -132433,9 +124967,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -132445,19 +124978,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(49), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10305, + Ctx: p10269, FreeVars: ast.Identifiers{ "v", }, @@ -132479,21 +125012,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(54), }, End: ast.Location{ - Line: int(1000), + Line: int(985), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10293, + Ctx: p10257, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -132509,30 +125043,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1000), - Column: int(17), - }, - End: ast.Location{ - Line: int(1000), - Column: int(56), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1001), + Line: int(986), Column: int(11), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(42), }, File: p1, @@ -132545,7 +125067,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -132566,14 +125088,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1001), - Column: int(25), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1005), - Column: int(12), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -132650,7 +125172,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -132684,26 +125205,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -132712,14 +125222,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1001), - Column: int(25), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1005), - Column: int(12), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -132796,7 +125306,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -132829,26 +125338,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "param", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "param", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -132881,19 +125379,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(13), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10338, + Ctx: p10302, FreeVars: ast.Identifiers{ "aux", "i", @@ -132905,19 +125403,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(13), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10338, + Ctx: p10302, FreeVars: ast.Identifiers{ "param", }, @@ -132925,13 +125423,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(13), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(16), }, File: p1, @@ -132944,32 +125442,31 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10338, + Ctx: p10302, FreeVars: nil, }, Value: "-", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(19), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10338, + Ctx: p10302, FreeVars: ast.Identifiers{ "param", }, @@ -132977,13 +125474,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(19), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(24), }, File: p1, @@ -133018,9 +125515,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -133029,19 +125525,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(33), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10338, + Ctx: p10302, FreeVars: ast.Identifiers{ "aux", "i", @@ -133053,19 +125549,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(33), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10338, + Ctx: p10302, FreeVars: ast.Identifiers{ "aux", }, @@ -133079,19 +125575,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(37), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10355, + Ctx: p10319, FreeVars: ast.Identifiers{ "i", "v", @@ -133100,19 +125596,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(37), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10355, + Ctx: p10319, FreeVars: ast.Identifiers{ "v", }, @@ -133123,19 +125619,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(39), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10355, + Ctx: p10319, FreeVars: ast.Identifiers{ "i", }, @@ -133151,19 +125647,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(43), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10355, + Ctx: p10319, FreeVars: ast.Identifiers{ "i", "path", @@ -133172,19 +125668,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(43), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10355, + Ctx: p10319, FreeVars: ast.Identifiers{ "path", }, @@ -133196,19 +125692,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(50), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10355, + Ctx: p10319, FreeVars: ast.Identifiers{ "i", }, @@ -133218,19 +125714,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(51), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10369, + Ctx: p10333, FreeVars: ast.Identifiers{ "i", }, @@ -133250,19 +125746,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(55), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10355, + Ctx: p10319, FreeVars: ast.Identifiers{ "param", }, @@ -133270,13 +125766,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(55), }, End: ast.Location{ - Line: int(1002), + Line: int(987), Column: int(60), }, File: p1, @@ -133311,9 +125807,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -133340,19 +125835,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(26), }, End: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10377, + Ctx: p10341, FreeVars: ast.Identifiers{ "i", "params", @@ -133364,19 +125859,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(27), }, End: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10381, + Ctx: p10345, FreeVars: ast.Identifiers{ "i", "params", @@ -133386,19 +125881,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(27), }, End: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10381, + Ctx: p10345, FreeVars: ast.Identifiers{ "params", }, @@ -133412,19 +125907,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(34), }, End: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10387, + Ctx: p10351, FreeVars: ast.Identifiers{ "i", "v", @@ -133433,19 +125928,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(34), }, End: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10387, + Ctx: p10351, FreeVars: ast.Identifiers{ "v", }, @@ -133456,19 +125951,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(36), }, End: ast.Location{ - Line: int(1004), + Line: int(989), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10387, + Ctx: p10351, FreeVars: ast.Identifiers{ "i", }, @@ -133511,19 +126006,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1003), + Line: int(988), Column: int(22), }, End: ast.Location{ - Line: int(1003), + Line: int(988), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10377, + Ctx: p10341, FreeVars: ast.Identifiers{ "range", }, @@ -133542,36 +126037,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1001), - Column: int(17), - }, - End: ast.Location{ - Line: int(1005), - Column: int(12), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(11), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "cindent", "parts", @@ -133581,19 +126064,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(11), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -133601,13 +126084,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(11), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(14), }, File: p1, @@ -133649,9 +126132,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -133661,19 +126143,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(20), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10405, + Ctx: p10369, FreeVars: ast.Identifiers{ "cindent", }, @@ -133681,44 +126163,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(20), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10405, + Ctx: p10369, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(27), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10405, + Ctx: p10369, FreeVars: ast.Identifiers{ "cindent", }, @@ -133732,19 +126213,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(36), }, End: ast.Location{ - Line: int(1006), + Line: int(991), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10405, + Ctx: p10369, FreeVars: ast.Identifiers{ "parts", }, @@ -133776,19 +126257,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(12), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -133801,19 +126282,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(15), }, End: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -133822,19 +126303,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(15), }, End: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -133842,13 +126323,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(15), }, End: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(18), }, File: p1, @@ -133883,9 +126364,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -133895,19 +126375,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(28), }, End: ast.Location{ - Line: int(1007), + Line: int(992), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10424, + Ctx: p10388, FreeVars: ast.Identifiers{ "v", }, @@ -133928,13 +126408,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(9), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, @@ -133947,7 +126427,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -133960,19 +126440,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(12), }, End: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -133981,19 +126461,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(12), }, End: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", "v", @@ -134002,19 +126482,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(12), }, End: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -134022,13 +126502,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(12), }, End: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(15), }, File: p1, @@ -134063,9 +126543,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -134075,19 +126554,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(23), }, End: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10440, + Ctx: p10404, FreeVars: ast.Identifiers{ "v", }, @@ -134109,21 +126588,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(29), }, End: ast.Location{ - Line: int(1008), + Line: int(993), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -134131,13 +126611,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1009), + Line: int(994), Column: int(11), }, End: ast.Location{ - Line: int(1009), + Line: int(994), Column: int(15), }, File: p1, @@ -134150,13 +126630,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: nil, }, Value: "{}", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -134169,13 +126648,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1011), + Line: int(996), Column: int(11), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, @@ -134188,7 +126667,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -134206,19 +126685,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1011), + Line: int(996), Column: int(17), }, End: ast.Location{ - Line: int(1028), + Line: int(1013), Column: int(14), }, File: p1, }, Fodder: nil, - Ctx: p10451, + Ctx: p10415, FreeVars: ast.Identifiers{ "cindent", "indent_array_in_object", @@ -134226,39 +126705,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1011), - Column: int(24), - }, - End: ast.Location{ - Line: int(1011), - Column: int(29), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(13), }, End: ast.Location{ - Line: int(1028), + Line: int(1013), Column: int(14), }, File: p1, @@ -134271,7 +126739,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "cindent", "indent_array_in_object", @@ -134282,19 +126750,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(16), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -134303,19 +126771,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(16), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -134324,19 +126792,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(16), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", }, @@ -134344,13 +126812,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(16), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(19), }, File: p1, @@ -134385,9 +126853,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -134397,19 +126864,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(28), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10469, + Ctx: p10433, FreeVars: ast.Identifiers{ "value", }, @@ -134431,19 +126898,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(38), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -134452,19 +126919,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(38), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -134473,19 +126940,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(38), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", }, @@ -134493,13 +126960,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(38), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(41), }, File: p1, @@ -134534,9 +127001,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -134546,19 +127012,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(49), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10482, + Ctx: p10446, FreeVars: ast.Identifiers{ "value", }, @@ -134580,21 +127046,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(58), }, End: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -134603,19 +127070,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1012), + Line: int(997), Column: int(65), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "cindent", "indent_array_in_object", @@ -134646,24 +127113,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(27), }, End: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: ast.Identifiers{ "cindent", "indent_array_in_object", @@ -134672,19 +127138,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(30), }, End: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: ast.Identifiers{ "indent_array_in_object", }, @@ -134695,19 +127161,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(58), }, End: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: ast.Identifiers{ "cindent", }, @@ -134715,19 +127181,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(58), }, End: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: ast.Identifiers{ "cindent", }, @@ -134739,44 +127205,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(68), }, End: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(78), }, End: ast.Location{ - Line: int(1019), + Line: int(1004), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: ast.Identifiers{ "cindent", }, @@ -134785,18 +127250,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1019), - Column: int(15), - }, - End: ast.Location{ - Line: int(1019), - Column: int(85), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -134821,79 +127274,77 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(22), }, End: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: nil, }, Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(22), }, End: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(29), }, End: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10490, + Ctx: p10454, FreeVars: nil, }, Target: &ast.Self{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(29), }, End: ast.Location{ - Line: int(1020), + Line: int(1005), Column: int(33), }, File: p1, @@ -134925,25 +127376,12 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1020), - Column: int(15), - }, - End: ast.Location{ - Line: int(1020), - Column: int(44), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -134952,19 +127390,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(20), }, End: ast.Location{ - Line: int(1028), + Line: int(1013), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "cindent", "std", @@ -134974,19 +127412,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(23), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -134995,19 +127433,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(23), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -135016,19 +127454,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(23), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", }, @@ -135036,13 +127474,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(23), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(26), }, File: p1, @@ -135077,9 +127515,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -135089,19 +127526,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(36), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10520, + Ctx: p10484, FreeVars: ast.Identifiers{ "value", }, @@ -135123,19 +127560,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(46), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -135144,19 +127581,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(46), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", "value", @@ -135165,19 +127602,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(46), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "std", }, @@ -135185,13 +127622,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(46), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(49), }, File: p1, @@ -135226,9 +127663,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -135238,19 +127674,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(57), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10533, + Ctx: p10497, FreeVars: ast.Identifiers{ "value", }, @@ -135272,21 +127708,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(66), }, End: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -135295,19 +127732,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1021), + Line: int(1006), Column: int(73), }, End: ast.Location{ - Line: int(1024), + Line: int(1009), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "cindent", }, @@ -135337,24 +127774,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1022), + Line: int(1007), Column: int(27), }, End: ast.Location{ - Line: int(1022), + Line: int(1007), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10541, + Ctx: p10505, FreeVars: ast.Identifiers{ "cindent", }, @@ -135362,19 +127798,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1022), + Line: int(1007), Column: int(27), }, End: ast.Location{ - Line: int(1022), + Line: int(1007), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10541, + Ctx: p10505, FreeVars: ast.Identifiers{ "cindent", }, @@ -135386,40 +127822,27 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1022), + Line: int(1007), Column: int(37), }, End: ast.Location{ - Line: int(1022), + Line: int(1007), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10541, + Ctx: p10505, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1022), - Column: int(15), - }, - End: ast.Location{ - Line: int(1022), - Column: int(41), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -135444,79 +127867,77 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(22), }, End: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10541, + Ctx: p10505, FreeVars: nil, }, Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(22), }, End: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10541, + Ctx: p10505, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(29), }, End: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10541, + Ctx: p10505, FreeVars: nil, }, Target: &ast.Self{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(29), }, End: ast.Location{ - Line: int(1023), + Line: int(1008), Column: int(33), }, File: p1, @@ -135548,25 +127969,12 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1023), - Column: int(15), - }, - End: ast.Location{ - Line: int(1023), - Column: int(44), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -135575,19 +127983,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1024), + Line: int(1009), Column: int(20), }, End: ast.Location{ - Line: int(1028), + Line: int(1013), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10456, + Ctx: p10420, FreeVars: ast.Identifiers{ "cindent", }, @@ -135617,24 +128025,23 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1026), + Line: int(1011), Column: int(27), }, End: ast.Location{ - Line: int(1026), + Line: int(1011), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10557, + Ctx: p10521, FreeVars: ast.Identifiers{ "cindent", }, @@ -135642,18 +128049,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "cindent", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1026), - Column: int(15), - }, - End: ast.Location{ - Line: int(1026), - Column: int(34), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(1), @@ -135678,44 +128073,30 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1027), + Line: int(1012), Column: int(22), }, End: ast.Location{ - Line: int(1027), + Line: int(1012), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10557, + Ctx: p10521, FreeVars: nil, }, Value: " ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1027), - Column: int(15), - }, - End: ast.Location{ - Line: int(1027), - Column: int(25), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -135725,30 +128106,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1029), + Line: int(1014), Column: int(11), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, @@ -135761,7 +128130,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "aux", "cindent", @@ -135781,14 +128150,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1029), - Column: int(25), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1033), - Column: int(12), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -135864,7 +128233,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -135898,26 +128266,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -135926,14 +128283,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1029), - Column: int(25), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1033), - Column: int(12), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -136010,7 +128367,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -136044,26 +128400,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "param", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "param", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -136097,19 +128442,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(13), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(98), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "aux", "k", @@ -136122,19 +128467,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(13), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "k", "param", @@ -136144,19 +128489,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(13), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "k", "std", @@ -136165,19 +128510,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(13), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "k", "std", @@ -136186,19 +128531,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(13), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "std", }, @@ -136206,13 +128551,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(13), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(16), }, File: p1, @@ -136254,9 +128599,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringJson", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -136266,19 +128610,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(34), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10607, + Ctx: p10571, FreeVars: ast.Identifiers{ "k", }, @@ -136300,25 +128644,24 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(39), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: nil, }, Value: ":", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, OpFodder: ast.Fodder{}, @@ -136326,19 +128669,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(45), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "param", }, @@ -136346,13 +128689,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(45), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(50), }, File: p1, @@ -136387,9 +128730,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "space", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, }, @@ -136398,19 +128740,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(59), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(98), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "aux", "k", @@ -136422,19 +128764,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(59), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10591, + Ctx: p10555, FreeVars: ast.Identifiers{ "aux", }, @@ -136448,19 +128790,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(63), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10621, + Ctx: p10585, FreeVars: ast.Identifiers{ "k", "v", @@ -136469,19 +128811,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(63), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10621, + Ctx: p10585, FreeVars: ast.Identifiers{ "v", }, @@ -136492,19 +128834,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(65), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10621, + Ctx: p10585, FreeVars: ast.Identifiers{ "k", }, @@ -136520,19 +128862,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(69), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10621, + Ctx: p10585, FreeVars: ast.Identifiers{ "k", "path", @@ -136541,19 +128883,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(69), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10621, + Ctx: p10585, FreeVars: ast.Identifiers{ "path", }, @@ -136565,19 +128907,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(76), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10621, + Ctx: p10585, FreeVars: ast.Identifiers{ "k", }, @@ -136587,19 +128929,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(77), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10635, + Ctx: p10599, FreeVars: ast.Identifiers{ "k", }, @@ -136619,19 +128961,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(81), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(97), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10621, + Ctx: p10585, FreeVars: ast.Identifiers{ "param", }, @@ -136639,13 +128981,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(81), }, End: ast.Location{ - Line: int(1030), + Line: int(1015), Column: int(86), }, File: p1, @@ -136680,9 +129022,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "new_indent", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, CommaFodder: nil, @@ -136709,19 +129050,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(26), }, End: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10643, + Ctx: p10607, FreeVars: ast.Identifiers{ "k", "params", @@ -136733,19 +129074,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(27), }, End: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10647, + Ctx: p10611, FreeVars: ast.Identifiers{ "k", "params", @@ -136755,19 +129096,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(27), }, End: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10647, + Ctx: p10611, FreeVars: ast.Identifiers{ "params", }, @@ -136781,19 +129122,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(34), }, End: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10653, + Ctx: p10617, FreeVars: ast.Identifiers{ "k", "v", @@ -136802,19 +129143,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(34), }, End: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10653, + Ctx: p10617, FreeVars: ast.Identifiers{ "v", }, @@ -136825,19 +129166,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(36), }, End: ast.Location{ - Line: int(1032), + Line: int(1017), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10653, + Ctx: p10617, FreeVars: ast.Identifiers{ "k", }, @@ -136880,19 +129221,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(22), }, End: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10643, + Ctx: p10607, FreeVars: ast.Identifiers{ "std", "v", @@ -136901,19 +129242,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(22), }, End: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10643, + Ctx: p10607, FreeVars: ast.Identifiers{ "std", }, @@ -136921,13 +129262,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(22), }, End: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(25), }, File: p1, @@ -136962,9 +129303,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -136974,19 +129314,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(39), }, End: ast.Location{ - Line: int(1031), + Line: int(1016), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10668, + Ctx: p10632, FreeVars: ast.Identifiers{ "v", }, @@ -137015,36 +129355,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1029), - Column: int(17), - }, - End: ast.Location{ - Line: int(1033), - Column: int(12), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(11), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "cindent", "lines", @@ -137054,19 +129382,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(11), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9934, + Ctx: p9898, FreeVars: ast.Identifiers{ "std", }, @@ -137074,13 +129402,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(11), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(14), }, File: p1, @@ -137122,9 +129450,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -137134,19 +129461,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(20), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10680, + Ctx: p10644, FreeVars: ast.Identifiers{ "cindent", }, @@ -137154,44 +129481,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(20), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10680, + Ctx: p10644, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(27), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10680, + Ctx: p10644, FreeVars: ast.Identifiers{ "cindent", }, @@ -137205,19 +129531,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(36), }, End: ast.Location{ - Line: int(1034), + Line: int(1019), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10680, + Ctx: p10644, FreeVars: ast.Identifiers{ "lines", }, @@ -137268,36 +129594,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(5), }, End: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p9923, + Ctx: p9887, FreeVars: ast.Identifiers{ "aux", "value", @@ -137306,13 +129620,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(5), }, End: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(8), }, File: p1, @@ -137325,7 +129639,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p9923, + Ctx: p9887, FreeVars: ast.Identifiers{ "aux", }, @@ -137339,19 +129653,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(9), }, End: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10695, + Ctx: p10659, FreeVars: ast.Identifiers{ "value", }, @@ -137364,19 +129678,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(16), }, End: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10695, + Ctx: p10659, FreeVars: nil, }, Elements: nil, @@ -137389,25 +129703,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(20), }, End: ast.Location{ - Line: int(1035), + Line: int(1020), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10695, + Ctx: p10659, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -137422,18 +129735,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(954), - Column: int(3), - }, - End: ast.Location{ - Line: int(1035), - Column: int(23), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -137458,7 +129759,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestYamlStream", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -137481,100 +129781,66 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1037), - Column: int(22), - }, - End: ast.Location{ - Line: int(1037), - Column: int(27), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "indent_array_in_object", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.LiteralBoolean{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1037), - Column: int(52), - }, - End: ast.Location{ - Line: int(1037), - Column: int(57), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "indent_array_in_object", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.LiteralBoolean{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1022), + Column: int(52), + }, + End: ast.Location{ + Line: int(1022), + Column: int(57), + }, + File: p1, }, - File: p1, + Fodder: ast.Fodder{}, + Ctx: p10669, + FreeVars: nil, }, - Fodder: ast.Fodder{}, - Ctx: p10704, - FreeVars: nil, - }, - Value: false, - }, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1037), - Column: int(29), - }, - End: ast.Location{ - Line: int(1037), - Column: int(57), + Value: false, }, - File: p1, + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "c_document_end", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.LiteralBoolean{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1037), - Column: int(74), - }, - End: ast.Location{ - Line: int(1037), - Column: int(78), + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "c_document_end", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.LiteralBoolean{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1022), + Column: int(74), + }, + End: ast.Location{ + Line: int(1022), + Column: int(78), + }, + File: p1, }, - File: p1, + Fodder: ast.Fodder{}, + Ctx: p10669, + FreeVars: nil, }, - Fodder: ast.Fodder{}, - Ctx: p10704, - FreeVars: nil, - }, - Value: true, - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1037), - Column: int(59), + Value: true, }, - End: ast.Location{ - Line: int(1037), - Column: int(78), - }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -137583,13 +129849,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(5), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(53), }, File: p1, @@ -137602,7 +129868,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "c_document_end", "indent_array_in_object", @@ -137613,19 +129879,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(8), }, End: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", "value", @@ -137635,19 +129901,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(9), }, End: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", "value", @@ -137656,19 +129922,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(9), }, End: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", }, @@ -137676,13 +129942,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(9), }, End: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(12), }, File: p1, @@ -137717,9 +129983,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -137729,19 +129994,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(21), }, End: ast.Location{ - Line: int(1038), + Line: int(1023), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10720, + Ctx: p10685, FreeVars: ast.Identifiers{ "value", }, @@ -137763,13 +130028,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(7), }, End: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(75), }, File: p1, @@ -137782,7 +130047,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", "value", @@ -137791,19 +130056,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(13), }, End: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", "value", @@ -137812,44 +130077,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(13), }, End: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: nil, }, Value: "manifestYamlStream only takes arrays, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(60), }, End: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", "value", @@ -137858,19 +130122,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(60), }, End: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", }, @@ -137878,13 +130142,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(60), }, End: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(63), }, File: p1, @@ -137919,9 +130183,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -137931,19 +130194,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(69), }, End: ast.Location{ - Line: int(1039), + Line: int(1024), Column: int(74), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10737, + Ctx: p10702, FreeVars: ast.Identifiers{ "value", }, @@ -137973,19 +130236,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(7), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "c_document_end", "indent_array_in_object", @@ -137996,19 +130259,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(7), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(8), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "indent_array_in_object", "std", @@ -138018,13 +130281,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(7), }, End: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(14), }, File: p1, @@ -138037,32 +130300,31 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10704, + Ctx: p10669, FreeVars: nil, }, Value: "---\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(17), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(8), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "indent_array_in_object", "std", @@ -138072,19 +130334,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(17), }, End: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "std", }, @@ -138092,13 +130354,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(17), }, End: ast.Location{ - Line: int(1041), + Line: int(1026), Column: int(20), }, File: p1, @@ -138133,9 +130395,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -138145,13 +130406,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(9), }, End: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(18), }, File: p1, @@ -138164,13 +130425,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10756, + Ctx: p10721, FreeVars: nil, }, Value: "\n---\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -138180,14 +130440,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1042), - Column: int(20), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1042), - Column: int(83), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -138261,7 +130521,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -138292,26 +130551,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "e", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "e", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -138342,19 +130590,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(21), }, End: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10772, + Ctx: p10737, FreeVars: ast.Identifiers{ "e", "indent_array_in_object", @@ -138364,19 +130612,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(21), }, End: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10772, + Ctx: p10737, FreeVars: ast.Identifiers{ "std", }, @@ -138384,13 +130632,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(21), }, End: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(24), }, File: p1, @@ -138425,9 +130673,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestYamlDoc", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -138437,19 +130684,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(41), }, End: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10781, + Ctx: p10746, FreeVars: ast.Identifiers{ "e", }, @@ -138462,19 +130709,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(44), }, End: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10781, + Ctx: p10746, FreeVars: ast.Identifiers{ "indent_array_in_object", }, @@ -138504,19 +130751,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(77), }, End: ast.Location{ - Line: int(1042), + Line: int(1027), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10756, + Ctx: p10721, FreeVars: ast.Identifiers{ "value", }, @@ -138556,19 +130803,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(11), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "c_document_end", }, @@ -138576,19 +130823,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(14), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: ast.Identifiers{ "c_document_end", }, @@ -138599,67 +130846,53 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(34), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: nil, }, Value: "\n...\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{}, BranchFalse: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(49), }, End: ast.Location{ - Line: int(1043), + Line: int(1028), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10704, + Ctx: p10669, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1037), - Column: int(3), - }, - End: ast.Location{ - Line: int(1043), - Column: int(53), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -138684,7 +130917,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestPython", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -138707,39 +130939,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1046), - Column: int(18), - }, - End: ast.Location{ - Line: int(1046), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(5), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, @@ -138752,7 +130973,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -138761,19 +130982,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(8), }, End: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -138782,19 +131003,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(8), }, End: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", }, @@ -138802,13 +131023,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(8), }, End: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(11), }, File: p1, @@ -138843,9 +131064,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -138855,19 +131075,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(21), }, End: ast.Location{ - Line: int(1047), + Line: int(1032), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10811, + Ctx: p10776, FreeVars: ast.Identifiers{ "v", }, @@ -138888,13 +131108,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1048), + Line: int(1033), Column: int(7), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(40), }, File: p1, @@ -138907,7 +131127,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -138923,14 +131143,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1048), - Column: int(22), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1051), - Column: int(8), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -139003,7 +131223,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -139034,26 +131253,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -139086,14 +131294,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1049), - Column: int(9), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1049), - Column: int(73), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -139167,7 +131375,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -139179,13 +131386,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(9), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(17), }, File: p1, @@ -139198,13 +131405,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10841, + Ctx: p10806, FreeVars: nil, }, Value: "%s: %s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -139212,19 +131418,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(20), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10841, + Ctx: p10806, FreeVars: ast.Identifiers{ "k", "std", @@ -139236,19 +131442,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(21), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10846, + Ctx: p10811, FreeVars: ast.Identifiers{ "k", "std", @@ -139257,19 +131463,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(21), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10846, + Ctx: p10811, FreeVars: ast.Identifiers{ "std", }, @@ -139277,13 +131483,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(21), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(24), }, File: p1, @@ -139318,9 +131524,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringPython", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -139330,19 +131535,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(44), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10855, + Ctx: p10820, FreeVars: ast.Identifiers{ "k", }, @@ -139365,19 +131570,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(48), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10846, + Ctx: p10811, FreeVars: ast.Identifiers{ "k", "std", @@ -139387,19 +131592,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(48), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10846, + Ctx: p10811, FreeVars: ast.Identifiers{ "std", }, @@ -139407,13 +131612,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(48), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(51), }, File: p1, @@ -139448,9 +131653,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestPython", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -139460,19 +131664,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(67), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10866, + Ctx: p10831, FreeVars: ast.Identifiers{ "k", "v", @@ -139481,19 +131685,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(67), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10866, + Ctx: p10831, FreeVars: ast.Identifiers{ "v", }, @@ -139504,19 +131708,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(69), }, End: ast.Location{ - Line: int(1049), + Line: int(1034), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10866, + Ctx: p10831, FreeVars: ast.Identifiers{ "k", }, @@ -139565,19 +131769,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(18), }, End: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10873, + Ctx: p10838, FreeVars: ast.Identifiers{ "std", "v", @@ -139586,19 +131790,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(18), }, End: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10873, + Ctx: p10838, FreeVars: ast.Identifiers{ "std", }, @@ -139606,13 +131810,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(18), }, End: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(21), }, File: p1, @@ -139647,9 +131851,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -139659,19 +131862,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(35), }, End: ast.Location{ - Line: int(1050), + Line: int(1035), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10882, + Ctx: p10847, FreeVars: ast.Identifiers{ "v", }, @@ -139700,18 +131903,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1048), - Column: int(13), - }, - End: ast.Location{ - Line: int(1051), - Column: int(8), - }, - File: p1, - }, }, }, Body: &ast.Apply{ @@ -139719,14 +131910,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1052), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1052), - Column: int(40), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -139799,7 +131990,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -139811,13 +132001,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(7), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(13), }, File: p1, @@ -139830,13 +132020,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: "{%s}", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -139844,19 +132033,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(16), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "fields", "std", @@ -139867,19 +132056,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(17), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10898, + Ctx: p10863, FreeVars: ast.Identifiers{ "fields", "std", @@ -139888,19 +132077,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(17), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10898, + Ctx: p10863, FreeVars: ast.Identifiers{ "std", }, @@ -139908,13 +132097,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(17), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(20), }, File: p1, @@ -139949,9 +132138,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -139961,25 +132149,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(26), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10907, + Ctx: p10872, FreeVars: nil, }, Value: ", ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -139987,19 +132174,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(32), }, End: ast.Location{ - Line: int(1052), + Line: int(1037), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10907, + Ctx: p10872, FreeVars: ast.Identifiers{ "fields", }, @@ -140044,19 +132231,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(10), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -140065,19 +132252,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(13), }, End: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -140086,19 +132273,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(13), }, End: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", }, @@ -140106,13 +132293,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(13), }, End: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(16), }, File: p1, @@ -140147,9 +132334,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -140159,19 +132345,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(25), }, End: ast.Location{ - Line: int(1053), + Line: int(1038), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10922, + Ctx: p10887, FreeVars: ast.Identifiers{ "v", }, @@ -140194,14 +132380,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1054), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1054), - Column: int(70), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -140274,7 +132460,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -140286,13 +132471,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(7), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(13), }, File: p1, @@ -140305,13 +132490,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: "[%s]", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -140319,19 +132503,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(16), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -140342,19 +132526,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(17), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10938, + Ctx: p10903, FreeVars: ast.Identifiers{ "std", "v", @@ -140363,19 +132547,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(17), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10938, + Ctx: p10903, FreeVars: ast.Identifiers{ "std", }, @@ -140383,13 +132567,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(17), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(20), }, File: p1, @@ -140424,9 +132608,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -140436,25 +132619,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(26), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10947, + Ctx: p10912, FreeVars: nil, }, Value: ", ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -140464,14 +132646,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1054), - Column: int(32), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1054), - Column: int(68), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -140544,7 +132726,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -140574,26 +132755,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "v2", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "v2", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -140623,19 +132793,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(33), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10963, + Ctx: p10928, FreeVars: ast.Identifiers{ "std", "v2", @@ -140644,19 +132814,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(33), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10963, + Ctx: p10928, FreeVars: ast.Identifiers{ "std", }, @@ -140664,13 +132834,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(33), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(36), }, File: p1, @@ -140705,9 +132875,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestPython", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -140717,19 +132886,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(52), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10972, + Ctx: p10937, FreeVars: ast.Identifiers{ "v2", }, @@ -140759,19 +132928,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(66), }, End: ast.Location{ - Line: int(1054), + Line: int(1039), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10947, + Ctx: p10912, FreeVars: ast.Identifiers{ "v", }, @@ -140825,19 +132994,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(10), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -140846,19 +133015,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(13), }, End: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -140867,19 +133036,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(13), }, End: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", }, @@ -140887,13 +133056,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(13), }, End: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(16), }, File: p1, @@ -140928,9 +133097,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -140940,19 +133108,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(26), }, End: ast.Location{ - Line: int(1055), + Line: int(1040), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10988, + Ctx: p10953, FreeVars: ast.Identifiers{ "v", }, @@ -140975,14 +133143,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1056), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1056), - Column: int(41), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -141055,7 +133223,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -141067,13 +133234,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(7), }, End: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(11), }, File: p1, @@ -141086,13 +133253,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: "%s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -141100,19 +133266,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(14), }, End: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -141123,19 +133289,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(15), }, End: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11004, + Ctx: p10969, FreeVars: ast.Identifiers{ "std", "v", @@ -141144,19 +133310,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(15), }, End: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11004, + Ctx: p10969, FreeVars: ast.Identifiers{ "std", }, @@ -141164,13 +133330,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(15), }, End: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(18), }, File: p1, @@ -141205,9 +133371,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "escapeStringPython", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -141217,19 +133382,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(38), }, End: ast.Location{ - Line: int(1056), + Line: int(1041), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11013, + Ctx: p10978, FreeVars: ast.Identifiers{ "v", }, @@ -141273,19 +133438,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(10), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -141294,19 +133459,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(13), }, End: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -141315,19 +133480,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(13), }, End: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", }, @@ -141335,13 +133500,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(13), }, End: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(16), }, File: p1, @@ -141376,9 +133541,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isFunction", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -141388,19 +133552,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(28), }, End: ast.Location{ - Line: int(1057), + Line: int(1042), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11027, + Ctx: p10992, FreeVars: ast.Identifiers{ "v", }, @@ -141421,13 +133585,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1058), + Line: int(1043), Column: int(7), }, End: ast.Location{ - Line: int(1058), + Line: int(1043), Column: int(39), }, File: p1, @@ -141440,31 +133604,30 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1058), + Line: int(1043), Column: int(13), }, End: ast.Location{ - Line: int(1058), + Line: int(1043), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: "cannot manifest function", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{ @@ -141478,19 +133641,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(10), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -141499,19 +133662,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(13), }, End: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -141520,19 +133683,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(13), }, End: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", }, @@ -141540,13 +133703,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(13), }, End: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(16), }, File: p1, @@ -141581,9 +133744,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isNumber", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -141593,19 +133755,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(26), }, End: ast.Location{ - Line: int(1059), + Line: int(1044), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11044, + Ctx: p11009, FreeVars: ast.Identifiers{ "v", }, @@ -141626,19 +133788,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(7), }, End: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", "v", @@ -141647,19 +133809,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(7), }, End: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "std", }, @@ -141667,13 +133829,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(7), }, End: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(10), }, File: p1, @@ -141715,9 +133877,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "toString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -141727,19 +133888,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(20), }, End: ast.Location{ - Line: int(1060), + Line: int(1045), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11056, + Ctx: p11021, FreeVars: ast.Identifiers{ "v", }, @@ -141767,19 +133928,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1061), + Line: int(1046), Column: int(10), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -141787,19 +133948,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1061), + Line: int(1046), Column: int(13), }, End: ast.Location{ - Line: int(1061), + Line: int(1046), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -141807,19 +133968,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1061), + Line: int(1046), Column: int(13), }, End: ast.Location{ - Line: int(1061), + Line: int(1046), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -141831,19 +133992,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1061), + Line: int(1046), Column: int(18), }, End: ast.Location{ - Line: int(1061), + Line: int(1046), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: true, @@ -141853,13 +134014,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1062), + Line: int(1047), Column: int(7), }, End: ast.Location{ - Line: int(1062), + Line: int(1047), Column: int(13), }, File: p1, @@ -141872,13 +134033,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: "True", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -141891,19 +134051,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1063), + Line: int(1048), Column: int(10), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -141911,19 +134071,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1063), + Line: int(1048), Column: int(13), }, End: ast.Location{ - Line: int(1063), + Line: int(1048), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -141931,19 +134091,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1063), + Line: int(1048), Column: int(13), }, End: ast.Location{ - Line: int(1063), + Line: int(1048), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -141955,19 +134115,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1063), + Line: int(1048), Column: int(18), }, End: ast.Location{ - Line: int(1063), + Line: int(1048), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: false, @@ -141977,13 +134137,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1064), + Line: int(1049), Column: int(7), }, End: ast.Location{ - Line: int(1064), + Line: int(1049), Column: int(14), }, File: p1, @@ -141996,13 +134156,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: "False", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: ast.Fodder{ ast.FodderElement{ @@ -142015,19 +134174,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1065), + Line: int(1050), Column: int(10), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -142035,19 +134194,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1065), + Line: int(1050), Column: int(13), }, End: ast.Location{ - Line: int(1065), + Line: int(1050), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -142055,19 +134214,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1065), + Line: int(1050), Column: int(13), }, End: ast.Location{ - Line: int(1065), + Line: int(1050), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: ast.Identifiers{ "v", }, @@ -142079,19 +134238,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1065), + Line: int(1050), Column: int(18), }, End: ast.Location{ - Line: int(1065), + Line: int(1050), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, }, @@ -142100,13 +134259,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(7), }, End: ast.Location{ - Line: int(1066), + Line: int(1051), Column: int(13), }, File: p1, @@ -142119,13 +134278,12 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p10800, + Ctx: p10765, FreeVars: nil, }, Value: "None", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, ElseFodder: nil, BranchFalse: &ast.LiteralNull{ @@ -142157,18 +134315,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1046), - Column: int(3), - }, - End: ast.Location{ - Line: int(1066), - Column: int(13), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -142193,7 +134339,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestPythonVars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -142216,39 +134361,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "conf", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1068), - Column: int(22), - }, - End: ast.Location{ - Line: int(1068), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "conf", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(5), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(32), }, File: p1, @@ -142261,7 +134395,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11095, + Ctx: p11060, FreeVars: ast.Identifiers{ "conf", "std", @@ -142277,14 +134411,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1069), - Column: int(18), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1069), - Column: int(96), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -142357,7 +134491,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -142388,26 +134521,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -142440,14 +134562,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1069), - Column: int(19), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1069), - Column: int(63), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -142521,7 +134643,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -142533,25 +134654,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(19), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11121, + Ctx: p11086, FreeVars: nil, }, Value: "%s = %s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -142559,19 +134679,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(31), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11121, + Ctx: p11086, FreeVars: ast.Identifiers{ "conf", "k", @@ -142583,19 +134703,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(32), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11126, + Ctx: p11091, FreeVars: ast.Identifiers{ "k", }, @@ -142608,19 +134728,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(35), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11126, + Ctx: p11091, FreeVars: ast.Identifiers{ "conf", "k", @@ -142630,19 +134750,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(35), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11126, + Ctx: p11091, FreeVars: ast.Identifiers{ "std", }, @@ -142650,13 +134770,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(35), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(38), }, File: p1, @@ -142691,9 +134811,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestPython", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -142703,19 +134822,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(54), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11137, + Ctx: p11102, FreeVars: ast.Identifiers{ "conf", "k", @@ -142724,19 +134843,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(54), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11137, + Ctx: p11102, FreeVars: ast.Identifiers{ "conf", }, @@ -142747,19 +134866,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(59), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11137, + Ctx: p11102, FreeVars: ast.Identifiers{ "k", }, @@ -142808,19 +134927,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(73), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(95), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11144, + Ctx: p11109, FreeVars: ast.Identifiers{ "conf", "std", @@ -142829,19 +134948,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(73), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(89), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11144, + Ctx: p11109, FreeVars: ast.Identifiers{ "std", }, @@ -142849,13 +134968,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(73), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(76), }, File: p1, @@ -142890,9 +135009,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -142902,19 +135020,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(90), }, End: ast.Location{ - Line: int(1069), + Line: int(1054), Column: int(94), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11153, + Ctx: p11118, FreeVars: ast.Identifiers{ "conf", }, @@ -142943,36 +135061,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1069), - Column: int(11), - }, - End: ast.Location{ - Line: int(1069), - Column: int(96), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(5), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11095, + Ctx: p11060, FreeVars: ast.Identifiers{ "std", "vars", @@ -142981,19 +135087,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(5), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11095, + Ctx: p11060, FreeVars: ast.Identifiers{ "std", }, @@ -143001,13 +135107,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(5), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(8), }, File: p1, @@ -143049,9 +135155,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -143061,25 +135166,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(14), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11165, + Ctx: p11130, FreeVars: nil, }, Value: "\n", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -143087,19 +135191,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(20), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11165, + Ctx: p11130, FreeVars: ast.Identifiers{ "vars", }, @@ -143107,19 +135211,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(20), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11165, + Ctx: p11130, FreeVars: ast.Identifiers{ "vars", }, @@ -143131,19 +135235,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(27), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11165, + Ctx: p11130, FreeVars: nil, }, Elements: []ast.CommaSeparatedExpr{ @@ -143151,25 +135255,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(28), }, End: ast.Location{ - Line: int(1070), + Line: int(1055), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11173, + Ctx: p11138, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -143191,18 +135294,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1068), - Column: int(3), - }, - End: ast.Location{ - Line: int(1070), - Column: int(32), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -143227,7 +135318,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "manifestXmlJsonml", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -143250,39 +135340,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1072), - Column: int(21), - }, - End: ast.Location{ - Line: int(1072), - Column: int(26), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(5), }, End: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(17), }, File: p1, @@ -143295,7 +135374,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", "value", @@ -143304,19 +135383,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(8), }, End: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", "value", @@ -143326,19 +135405,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(9), }, End: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", "value", @@ -143347,19 +135426,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(9), }, End: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", }, @@ -143367,13 +135446,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(9), }, End: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(12), }, File: p1, @@ -143408,9 +135487,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -143420,19 +135498,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(21), }, End: ast.Location{ - Line: int(1073), + Line: int(1058), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11193, + Ctx: p11158, FreeVars: ast.Identifiers{ "value", }, @@ -143454,13 +135532,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(7), }, End: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(75), }, File: p1, @@ -143473,7 +135551,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", "value", @@ -143484,14 +135562,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1074), - Column: int(13), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1074), - Column: int(75), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -143564,7 +135642,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -143576,25 +135653,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(13), }, End: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11180, + Ctx: p11145, FreeVars: nil, }, Value: "Expected a JSONML value (an array), got %s", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -143602,19 +135678,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(60), }, End: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", "value", @@ -143623,19 +135699,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(60), }, End: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", }, @@ -143643,13 +135719,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(60), }, End: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(63), }, File: p1, @@ -143684,9 +135760,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -143696,19 +135771,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(69), }, End: ast.Location{ - Line: int(1074), + Line: int(1059), Column: int(74), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11216, + Ctx: p11181, FreeVars: ast.Identifiers{ "value", }, @@ -143747,13 +135822,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1076), + Line: int(1061), Column: int(7), }, End: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(17), }, File: p1, @@ -143766,7 +135841,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "std", "value", @@ -143780,58 +135855,47 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1076), + Line: int(1061), Column: int(13), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, }, Fodder: nil, - Ctx: p11224, + Ctx: p11189, FreeVars: ast.Identifiers{ "aux", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1076), - Column: int(17), - }, - End: ast.Location{ - Line: int(1076), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "v", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(9), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, @@ -143844,7 +135908,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "aux", "std", @@ -143854,19 +135918,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(12), }, End: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "std", "v", @@ -143875,19 +135939,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(12), }, End: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "std", }, @@ -143895,13 +135959,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(12), }, End: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(15), }, File: p1, @@ -143936,9 +136000,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -143948,19 +136011,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(25), }, End: ast.Location{ - Line: int(1077), + Line: int(1062), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11240, + Ctx: p11205, FreeVars: ast.Identifiers{ "v", }, @@ -143981,13 +136044,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1078), + Line: int(1063), Column: int(11), }, End: ast.Location{ - Line: int(1078), + Line: int(1063), Column: int(12), }, File: p1, @@ -144000,7 +136063,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "v", }, @@ -144018,13 +136081,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1080), + Line: int(1065), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, @@ -144037,7 +136100,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "aux", "std", @@ -144052,19 +136115,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1080), + Line: int(1065), Column: int(23), }, End: ast.Location{ - Line: int(1080), + Line: int(1065), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11251, + Ctx: p11216, FreeVars: ast.Identifiers{ "v", }, @@ -144072,19 +136135,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1080), + Line: int(1065), Column: int(23), }, End: ast.Location{ - Line: int(1080), + Line: int(1065), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11251, + Ctx: p11216, FreeVars: ast.Identifiers{ "v", }, @@ -144095,21 +136158,22 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1080), + Line: int(1065), Column: int(25), }, End: ast.Location{ - Line: int(1080), + Line: int(1065), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11251, + Ctx: p11216, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, RightBracketFodder: ast.Fodder{}, @@ -144117,30 +136181,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1080), - Column: int(17), - }, - End: ast.Location{ - Line: int(1080), - Column: int(27), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, @@ -144153,7 +136205,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "aux", "std", @@ -144169,19 +136221,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(29), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11261, + Ctx: p11226, FreeVars: ast.Identifiers{ "std", "v", @@ -144190,19 +136242,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(29), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11261, + Ctx: p11226, FreeVars: ast.Identifiers{ "std", "v", @@ -144211,19 +136263,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(29), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11261, + Ctx: p11226, FreeVars: ast.Identifiers{ "std", "v", @@ -144232,19 +136284,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(29), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11261, + Ctx: p11226, FreeVars: ast.Identifiers{ "std", }, @@ -144252,13 +136304,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(29), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(32), }, File: p1, @@ -144293,9 +136345,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -144305,19 +136356,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(40), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11274, + Ctx: p11239, FreeVars: ast.Identifiers{ "v", }, @@ -144339,21 +136390,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(45), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11261, + Ctx: p11226, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -144362,19 +136414,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(50), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11261, + Ctx: p11226, FreeVars: ast.Identifiers{ "std", "v", @@ -144383,19 +136435,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(50), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11261, + Ctx: p11226, FreeVars: ast.Identifiers{ "std", }, @@ -144403,13 +136455,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(50), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(53), }, File: p1, @@ -144444,9 +136496,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -144456,19 +136507,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(63), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11286, + Ctx: p11251, FreeVars: ast.Identifiers{ "v", }, @@ -144476,19 +136527,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(63), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11286, + Ctx: p11251, FreeVars: ast.Identifiers{ "v", }, @@ -144499,21 +136550,22 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(65), }, End: ast.Location{ - Line: int(1081), + Line: int(1066), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11286, + Ctx: p11251, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, RightBracketFodder: ast.Fodder{}, @@ -144532,30 +136584,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1081), - Column: int(17), - }, - End: ast.Location{ - Line: int(1081), - Column: int(68), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, @@ -144568,7 +136608,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "aux", "has_attrs", @@ -144585,19 +136625,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(25), }, End: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11296, + Ctx: p11261, FreeVars: ast.Identifiers{ "has_attrs", "v", @@ -144606,19 +136646,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(28), }, End: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11296, + Ctx: p11261, FreeVars: ast.Identifiers{ "has_attrs", }, @@ -144629,19 +136669,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(43), }, End: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11296, + Ctx: p11261, FreeVars: ast.Identifiers{ "v", }, @@ -144649,19 +136689,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(43), }, End: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11296, + Ctx: p11261, FreeVars: ast.Identifiers{ "v", }, @@ -144672,21 +136712,22 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(45), }, End: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11296, + Ctx: p11261, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, RightBracketFodder: ast.Fodder{}, @@ -144696,19 +136737,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(53), }, End: ast.Location{ - Line: int(1082), + Line: int(1067), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11296, + Ctx: p11261, FreeVars: nil, }, Asserts: ast.Nodes{}, @@ -144718,30 +136759,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1082), - Column: int(17), - }, - End: ast.Location{ - Line: int(1082), - Column: int(55), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, @@ -144754,7 +136783,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "attrs", "aux", @@ -144772,19 +136801,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(28), }, End: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11311, + Ctx: p11276, FreeVars: ast.Identifiers{ "has_attrs", "std", @@ -144794,19 +136823,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(31), }, End: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11311, + Ctx: p11276, FreeVars: ast.Identifiers{ "has_attrs", }, @@ -144819,14 +136848,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1083), - Column: int(46), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1083), - Column: int(51), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -144899,7 +136928,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -144911,19 +136939,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(46), }, End: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11311, + Ctx: p11276, FreeVars: ast.Identifiers{ "v", }, @@ -144936,21 +136964,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(48), }, End: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11311, + Ctx: p11276, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, CommaFodder: nil, @@ -145013,14 +137042,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1083), - Column: int(57), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1083), - Column: int(62), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -145093,7 +137122,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -145105,19 +137133,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(57), }, End: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11311, + Ctx: p11276, FreeVars: ast.Identifiers{ "v", }, @@ -145130,21 +137158,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(59), }, End: ast.Location{ - Line: int(1083), + Line: int(1068), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11311, + Ctx: p11276, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, CommaFodder: nil, @@ -145204,30 +137233,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1083), - Column: int(17), - }, - End: ast.Location{ - Line: int(1083), - Column: int(62), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1084), + Line: int(1069), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, @@ -145240,7 +137257,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "attrs", "aux", @@ -145257,19 +137274,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(13), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(88), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11346, + Ctx: p11311, FreeVars: ast.Identifiers{ "attrs", "std", @@ -145278,19 +137295,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(13), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11346, + Ctx: p11311, FreeVars: ast.Identifiers{ "std", }, @@ -145298,13 +137315,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(13), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(16), }, File: p1, @@ -145346,9 +137363,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -145358,25 +137374,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(22), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11356, + Ctx: p11321, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -145386,14 +137401,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1085), - Column: int(26), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1085), - Column: int(87), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -145466,7 +137481,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -145497,26 +137511,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -145549,14 +137552,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1085), - Column: int(27), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1085), - Column: int(53), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -145630,7 +137633,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -145642,25 +137644,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(27), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11380, + Ctx: p11345, FreeVars: nil, }, Value: " %s=\"%s\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -145668,19 +137669,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(40), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11380, + Ctx: p11345, FreeVars: ast.Identifiers{ "attrs", "k", @@ -145691,19 +137692,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(41), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11385, + Ctx: p11350, FreeVars: ast.Identifiers{ "k", }, @@ -145716,19 +137717,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(44), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11385, + Ctx: p11350, FreeVars: ast.Identifiers{ "attrs", "k", @@ -145737,19 +137738,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(44), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11385, + Ctx: p11350, FreeVars: ast.Identifiers{ "attrs", }, @@ -145760,19 +137761,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(50), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11385, + Ctx: p11350, FreeVars: ast.Identifiers{ "k", }, @@ -145811,19 +137812,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(63), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11356, + Ctx: p11321, FreeVars: ast.Identifiers{ "attrs", "std", @@ -145832,19 +137833,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(63), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11356, + Ctx: p11321, FreeVars: ast.Identifiers{ "std", }, @@ -145852,13 +137853,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(63), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(66), }, File: p1, @@ -145893,9 +137894,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -145905,19 +137905,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(80), }, End: ast.Location{ - Line: int(1085), + Line: int(1070), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11402, + Ctx: p11367, FreeVars: ast.Identifiers{ "attrs", }, @@ -145956,36 +137956,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1084), - Column: int(17), - }, - End: ast.Location{ - Line: int(1085), - Column: int(88), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(95), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "attrs_str", "aux", @@ -145997,19 +137985,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11229, + Ctx: p11194, FreeVars: ast.Identifiers{ "std", }, @@ -146017,13 +138005,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(11), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(14), }, File: p1, @@ -146065,9 +138053,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "deepJoin", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -146077,19 +138064,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(24), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(94), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11414, + Ctx: p11379, FreeVars: ast.Identifiers{ "attrs_str", "aux", @@ -146103,25 +138090,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(25), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: nil, }, Value: "<", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -146129,19 +138115,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(30), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: ast.Identifiers{ "tag", }, @@ -146154,19 +138140,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(35), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: ast.Identifiers{ "attrs_str", }, @@ -146179,25 +138165,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(46), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: nil, }, Value: ">", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -146207,14 +138192,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1086), - Column: int(51), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1086), - Column: int(77), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -146288,7 +138273,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -146318,26 +138302,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -146367,19 +138340,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(52), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11439, + Ctx: p11404, FreeVars: ast.Identifiers{ "aux", "x", @@ -146388,19 +138361,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(52), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11439, + Ctx: p11404, FreeVars: ast.Identifiers{ "aux", }, @@ -146414,19 +138387,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(56), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11445, + Ctx: p11410, FreeVars: ast.Identifiers{ "x", }, @@ -146456,19 +138429,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(68), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: ast.Identifiers{ "children", }, @@ -146491,25 +138464,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(79), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(83), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: nil, }, Value: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(85), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(88), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: ast.Identifiers{ "tag", }, @@ -146542,25 +138514,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(90), }, End: ast.Location{ - Line: int(1086), + Line: int(1071), Column: int(93), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11418, + Ctx: p11383, FreeVars: nil, }, Value: ">", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -146587,36 +138558,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(7), }, End: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "aux", "value", @@ -146625,13 +138584,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(7), }, End: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(10), }, File: p1, @@ -146644,7 +138603,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11180, + Ctx: p11145, FreeVars: ast.Identifiers{ "aux", }, @@ -146658,19 +138617,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(11), }, End: ast.Location{ - Line: int(1088), + Line: int(1073), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11460, + Ctx: p11425, FreeVars: ast.Identifiers{ "value", }, @@ -146691,18 +138650,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1072), - Column: int(3), - }, - End: ast.Location{ - Line: int(1088), - Column: int(17), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -146727,7 +138674,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "base64", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -146751,39 +138697,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "input", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1093), - Column: int(10), - }, - End: ast.Location{ - Line: int(1093), - Column: int(15), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "input", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1094), + Line: int(1079), Column: int(5), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(24), }, File: p1, @@ -146796,7 +138731,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "base64_table", "input", @@ -146811,13 +138746,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(7), }, End: ast.Location{ - Line: int(1098), + Line: int(1083), Column: int(14), }, File: p1, @@ -146830,7 +138765,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11473, + Ctx: p11438, FreeVars: ast.Identifiers{ "input", "std", @@ -146839,19 +138774,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(10), }, End: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11473, + Ctx: p11438, FreeVars: ast.Identifiers{ "input", "std", @@ -146860,19 +138795,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(10), }, End: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11473, + Ctx: p11438, FreeVars: ast.Identifiers{ "std", }, @@ -146880,13 +138815,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(10), }, End: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(13), }, File: p1, @@ -146921,9 +138856,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -146933,19 +138867,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(23), }, End: ast.Location{ - Line: int(1095), + Line: int(1080), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11484, + Ctx: p11449, FreeVars: ast.Identifiers{ "input", }, @@ -146966,19 +138900,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(9), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11473, + Ctx: p11438, FreeVars: ast.Identifiers{ "input", "std", @@ -146987,19 +138921,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(9), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11473, + Ctx: p11438, FreeVars: ast.Identifiers{ "std", }, @@ -147007,13 +138941,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(9), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(12), }, File: p1, @@ -147055,9 +138989,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "map", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -147067,63 +139000,52 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(17), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11496, + Ctx: p11461, FreeVars: ast.Identifiers{ "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "c", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1096), - Column: int(26), - }, - End: ast.Location{ - Line: int(1096), - Column: int(27), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "c", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(29), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11500, + Ctx: p11465, FreeVars: ast.Identifiers{ "c", "std", @@ -147132,19 +139054,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(29), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11500, + Ctx: p11465, FreeVars: ast.Identifiers{ "std", }, @@ -147152,13 +139074,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(29), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(32), }, File: p1, @@ -147193,9 +139115,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -147205,19 +139126,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(43), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11509, + Ctx: p11474, FreeVars: ast.Identifiers{ "c", }, @@ -147241,19 +139162,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(47), }, End: ast.Location{ - Line: int(1096), + Line: int(1081), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11496, + Ctx: p11461, FreeVars: ast.Identifiers{ "input", }, @@ -147281,13 +139202,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1098), + Line: int(1083), Column: int(9), }, End: ast.Location{ - Line: int(1098), + Line: int(1083), Column: int(14), }, File: p1, @@ -147300,7 +139221,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11473, + Ctx: p11438, FreeVars: ast.Identifiers{ "input", }, @@ -147310,30 +139231,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1094), - Column: int(11), - }, - End: ast.Location{ - Line: int(1098), - Column: int(14), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1100), + Line: int(1085), Column: int(5), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(24), }, File: p1, @@ -147346,7 +139255,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "base64_table", "bytes", @@ -147361,19 +139270,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1100), + Line: int(1085), Column: int(11), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(33), }, File: p1, }, Fodder: nil, - Ctx: p11522, + Ctx: p11487, FreeVars: ast.Identifiers{ "aux", "base64_table", @@ -147381,77 +139290,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1100), - Column: int(15), - }, - End: ast.Location{ - Line: int(1100), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1100), - Column: int(20), - }, - End: ast.Location{ - Line: int(1100), - Column: int(21), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "r", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1100), - Column: int(23), - }, - End: ast.Location{ - Line: int(1100), - Column: int(24), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "r", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(7), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(33), }, File: p1, @@ -147464,7 +139334,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -147477,19 +139347,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(10), }, End: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "i", @@ -147499,19 +139369,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(10), }, End: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "i", }, @@ -147523,19 +139393,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(15), }, End: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "std", @@ -147544,19 +139414,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(15), }, End: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "std", }, @@ -147564,13 +139434,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(15), }, End: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(18), }, File: p1, @@ -147605,9 +139475,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -147617,19 +139486,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(26), }, End: ast.Location{ - Line: int(1101), + Line: int(1086), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11542, + Ctx: p11507, FreeVars: ast.Identifiers{ "arr", }, @@ -147651,13 +139520,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1102), + Line: int(1087), Column: int(9), }, End: ast.Location{ - Line: int(1102), + Line: int(1087), Column: int(10), }, File: p1, @@ -147670,7 +139539,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "r", }, @@ -147688,19 +139557,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(12), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -147713,19 +139582,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(15), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "i", @@ -147735,19 +139604,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(15), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "i", }, @@ -147755,19 +139624,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(15), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "i", }, @@ -147779,21 +139648,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(19), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -147802,19 +139672,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(24), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "std", @@ -147823,19 +139693,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(24), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "std", }, @@ -147843,13 +139713,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(24), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(27), }, File: p1, @@ -147884,9 +139754,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -147896,19 +139765,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(35), }, End: ast.Location{ - Line: int(1103), + Line: int(1088), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11566, + Ctx: p11531, FreeVars: ast.Identifiers{ "arr", }, @@ -147930,13 +139799,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1104), + Line: int(1089), Column: int(9), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(33), }, File: p1, @@ -147949,7 +139818,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -147966,19 +139835,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(11), }, End: ast.Location{ - Line: int(1109), + Line: int(1094), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -147988,19 +139857,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(11), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -148010,19 +139879,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(11), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -148032,13 +139901,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(11), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(23), }, File: p1, @@ -148051,7 +139920,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -148059,7 +139928,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "base64_table", }, @@ -148070,19 +139939,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(24), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "i", @@ -148091,19 +139960,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(25), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "i", @@ -148112,19 +139981,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(25), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "i", @@ -148133,19 +140002,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(25), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", }, @@ -148156,19 +140025,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(29), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "i", }, @@ -148183,21 +140052,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(34), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: nil, }, + Value: float64(252), OriginalString: "252", }, }, @@ -148206,21 +140076,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(42), }, End: ast.Location{ - Line: int(1106), + Line: int(1091), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -148232,19 +140103,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(11), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -148254,13 +140125,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(11), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(23), }, File: p1, @@ -148273,7 +140144,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -148281,7 +140152,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "base64_table", }, @@ -148292,19 +140163,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(24), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "i", @@ -148313,19 +140184,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(25), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "i", @@ -148334,19 +140205,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(25), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", "i", @@ -148355,19 +140226,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(25), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "arr", }, @@ -148378,19 +140249,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(29), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: ast.Identifiers{ "i", }, @@ -148405,21 +140276,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(34), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -148428,21 +140300,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(40), }, End: ast.Location{ - Line: int(1108), + Line: int(1093), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11573, + Ctx: p11538, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -148455,13 +140328,13 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1109), + Line: int(1094), Column: int(11), }, End: ast.Location{ - Line: int(1109), + Line: int(1094), Column: int(15), }, File: p1, @@ -148474,47 +140347,34 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11573, + Ctx: p11538, FreeVars: nil, }, Value: "==", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1104), - Column: int(15), - }, - End: ast.Location{ - Line: int(1109), - Column: int(15), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(9), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -148526,13 +140386,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(9), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(12), }, File: p1, @@ -148545,7 +140405,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "aux", }, @@ -148559,19 +140419,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(13), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11622, + Ctx: p11587, FreeVars: ast.Identifiers{ "arr", }, @@ -148584,19 +140444,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(18), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11622, + Ctx: p11587, FreeVars: ast.Identifiers{ "i", }, @@ -148604,19 +140464,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(18), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11622, + Ctx: p11587, FreeVars: ast.Identifiers{ "i", }, @@ -148628,21 +140488,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(22), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11622, + Ctx: p11587, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -148652,19 +140513,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(25), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11622, + Ctx: p11587, FreeVars: ast.Identifiers{ "r", "str", @@ -148673,19 +140534,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(25), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11622, + Ctx: p11587, FreeVars: ast.Identifiers{ "r", }, @@ -148697,19 +140558,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(29), }, End: ast.Location{ - Line: int(1110), + Line: int(1095), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11622, + Ctx: p11587, FreeVars: ast.Identifiers{ "str", }, @@ -148739,19 +140600,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(12), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -148764,19 +140625,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(15), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "i", @@ -148786,19 +140647,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(15), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "i", }, @@ -148806,19 +140667,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(15), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "i", }, @@ -148830,21 +140691,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(19), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -148853,19 +140715,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(24), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "std", @@ -148874,19 +140736,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(24), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "std", }, @@ -148894,13 +140756,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(24), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(27), }, File: p1, @@ -148935,9 +140797,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -148947,19 +140808,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(35), }, End: ast.Location{ - Line: int(1111), + Line: int(1096), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11654, + Ctx: p11619, FreeVars: ast.Identifiers{ "arr", }, @@ -148981,13 +140842,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1112), + Line: int(1097), Column: int(9), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(33), }, File: p1, @@ -149000,7 +140861,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -149017,19 +140878,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(11), }, End: ast.Location{ - Line: int(1119), + Line: int(1104), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -149039,19 +140900,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(11), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -149061,19 +140922,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(11), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -149083,19 +140944,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(11), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -149105,13 +140966,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(11), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(23), }, File: p1, @@ -149124,7 +140985,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -149132,7 +140993,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "base64_table", }, @@ -149143,19 +141004,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(24), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149164,19 +141025,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(25), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149185,19 +141046,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(25), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149206,19 +141067,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(25), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", }, @@ -149229,19 +141090,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(29), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "i", }, @@ -149256,21 +141117,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(34), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(252), OriginalString: "252", }, }, @@ -149279,21 +141141,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(42), }, End: ast.Location{ - Line: int(1114), + Line: int(1099), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -149305,19 +141168,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(11), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -149327,13 +141190,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(11), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(23), }, File: p1, @@ -149346,7 +141209,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -149354,7 +141217,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "base64_table", }, @@ -149365,19 +141228,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(24), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149386,19 +141249,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(24), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149407,19 +141270,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(25), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149428,19 +141291,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(25), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149449,19 +141312,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(25), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", }, @@ -149472,19 +141335,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(29), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "i", }, @@ -149499,21 +141362,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(34), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -149522,21 +141386,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(40), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -149545,19 +141410,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(44), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149566,19 +141431,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(45), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149587,19 +141452,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(45), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149608,19 +141473,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(45), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", }, @@ -149631,19 +141496,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(49), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "i", }, @@ -149651,19 +141516,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(49), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "i", }, @@ -149675,21 +141540,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(53), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -149701,21 +141567,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(58), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(240), OriginalString: "240", }, }, @@ -149724,21 +141591,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(66), }, End: ast.Location{ - Line: int(1116), + Line: int(1101), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -149752,19 +141620,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(11), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -149774,13 +141642,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(11), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(23), }, File: p1, @@ -149793,7 +141661,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -149801,7 +141669,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "base64_table", }, @@ -149812,19 +141680,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(24), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149833,19 +141701,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(25), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149854,19 +141722,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(25), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", "i", @@ -149875,19 +141743,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(25), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "arr", }, @@ -149898,19 +141766,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(29), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "i", }, @@ -149918,19 +141786,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(29), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: ast.Identifiers{ "i", }, @@ -149942,21 +141810,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(33), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -149968,21 +141837,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(38), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(15), OriginalString: "15", }, }, @@ -149991,21 +141861,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(45), }, End: ast.Location{ - Line: int(1118), + Line: int(1103), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -150018,13 +141889,13 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1119), + Line: int(1104), Column: int(11), }, End: ast.Location{ - Line: int(1119), + Line: int(1104), Column: int(14), }, File: p1, @@ -150037,47 +141908,34 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11661, + Ctx: p11626, FreeVars: nil, }, Value: "=", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1112), - Column: int(15), - }, - End: ast.Location{ - Line: int(1119), - Column: int(14), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(9), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -150089,13 +141947,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(9), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(12), }, File: p1, @@ -150108,7 +141966,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "aux", }, @@ -150122,19 +141980,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(13), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11750, + Ctx: p11715, FreeVars: ast.Identifiers{ "arr", }, @@ -150147,19 +142005,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(18), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11750, + Ctx: p11715, FreeVars: ast.Identifiers{ "i", }, @@ -150167,19 +142025,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(18), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11750, + Ctx: p11715, FreeVars: ast.Identifiers{ "i", }, @@ -150191,21 +142049,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(22), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11750, + Ctx: p11715, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -150215,19 +142074,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(25), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11750, + Ctx: p11715, FreeVars: ast.Identifiers{ "r", "str", @@ -150236,19 +142095,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(25), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11750, + Ctx: p11715, FreeVars: ast.Identifiers{ "r", }, @@ -150260,19 +142119,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(29), }, End: ast.Location{ - Line: int(1120), + Line: int(1105), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11750, + Ctx: p11715, FreeVars: ast.Identifiers{ "str", }, @@ -150302,13 +142161,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1122), + Line: int(1107), Column: int(9), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(33), }, File: p1, @@ -150321,7 +142180,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -150338,19 +142197,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(11), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -150360,19 +142219,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(11), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -150382,19 +142241,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(11), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -150404,19 +142263,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(11), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -150426,13 +142285,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(11), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(23), }, File: p1, @@ -150445,7 +142304,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -150453,7 +142312,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "base64_table", }, @@ -150464,19 +142323,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(24), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150485,19 +142344,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(25), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150506,19 +142365,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(25), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150527,19 +142386,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(25), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", }, @@ -150550,19 +142409,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(29), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -150577,21 +142436,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(34), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(252), OriginalString: "252", }, }, @@ -150600,21 +142460,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(42), }, End: ast.Location{ - Line: int(1124), + Line: int(1109), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -150626,19 +142487,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(11), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -150648,13 +142509,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(11), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(23), }, File: p1, @@ -150667,7 +142528,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -150675,7 +142536,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "base64_table", }, @@ -150686,19 +142547,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(24), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150707,19 +142568,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(24), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150728,19 +142589,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(25), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150749,19 +142610,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(25), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150770,19 +142631,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(25), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", }, @@ -150793,19 +142654,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(29), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -150820,21 +142681,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(34), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -150843,21 +142705,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(40), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -150866,19 +142729,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(44), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150887,19 +142750,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(45), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150908,19 +142771,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(45), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -150929,19 +142792,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(45), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", }, @@ -150952,19 +142815,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(49), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -150972,19 +142835,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(49), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -150996,21 +142859,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(53), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -151022,21 +142886,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(58), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(240), OriginalString: "240", }, }, @@ -151045,21 +142910,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(66), }, End: ast.Location{ - Line: int(1126), + Line: int(1111), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -151073,19 +142939,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(11), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -151095,13 +142961,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(11), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(23), }, File: p1, @@ -151114,7 +142980,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -151122,7 +142988,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "base64_table", }, @@ -151133,19 +142999,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(24), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151154,19 +143020,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(24), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151175,19 +143041,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(25), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151196,19 +143062,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(25), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151217,19 +143083,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(25), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", }, @@ -151240,19 +143106,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(29), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -151260,19 +143126,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(29), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -151284,21 +143150,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(33), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -151310,21 +143177,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(38), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(15), OriginalString: "15", }, }, @@ -151333,21 +143201,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(45), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -151356,19 +143225,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(49), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151377,19 +143246,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(50), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151398,19 +143267,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(50), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151419,19 +143288,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(50), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", }, @@ -151442,19 +143311,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(54), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -151462,19 +143331,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(54), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -151486,21 +143355,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(58), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -151512,21 +143382,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(63), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(192), OriginalString: "192", }, }, @@ -151535,21 +143406,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(71), }, End: ast.Location{ - Line: int(1128), + Line: int(1113), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(6), OriginalString: "6", }, }, @@ -151563,19 +143435,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(11), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "base64_table", @@ -151585,13 +143457,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(11), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(23), }, File: p1, @@ -151604,7 +143476,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -151612,7 +143484,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "base64_table", }, @@ -151623,19 +143495,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(25), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151644,19 +143516,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(25), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", "i", @@ -151665,19 +143537,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(25), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "arr", }, @@ -151688,19 +143560,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(29), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -151708,19 +143580,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(29), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: ast.Identifiers{ "i", }, @@ -151732,21 +143604,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(33), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -151758,21 +143631,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(38), }, End: ast.Location{ - Line: int(1130), + Line: int(1115), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11769, + Ctx: p11734, FreeVars: nil, }, + Value: float64(63), OriginalString: "63", }, }, @@ -151782,36 +143656,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1122), - Column: int(15), - }, - End: ast.Location{ - Line: int(1130), - Column: int(42), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(9), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "arr", "aux", @@ -151823,13 +143685,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(9), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(12), }, File: p1, @@ -151842,7 +143704,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11527, + Ctx: p11492, FreeVars: ast.Identifiers{ "aux", }, @@ -151856,19 +143718,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(13), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11891, + Ctx: p11856, FreeVars: ast.Identifiers{ "arr", }, @@ -151881,19 +143743,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(18), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11891, + Ctx: p11856, FreeVars: ast.Identifiers{ "i", }, @@ -151901,19 +143763,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(18), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11891, + Ctx: p11856, FreeVars: ast.Identifiers{ "i", }, @@ -151925,21 +143787,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(22), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11891, + Ctx: p11856, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -151949,19 +143812,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(25), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11891, + Ctx: p11856, FreeVars: ast.Identifiers{ "r", "str", @@ -151970,19 +143833,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(25), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11891, + Ctx: p11856, FreeVars: ast.Identifiers{ "r", }, @@ -151994,19 +143857,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(29), }, End: ast.Location{ - Line: int(1131), + Line: int(1116), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11891, + Ctx: p11856, FreeVars: ast.Identifiers{ "str", }, @@ -152031,30 +143894,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(5), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(24), }, File: p1, @@ -152067,7 +143918,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "aux", "bytes", @@ -152082,19 +143933,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(20), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11909, + Ctx: p11874, FreeVars: ast.Identifiers{ "bytes", "std", @@ -152103,19 +143954,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(20), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11909, + Ctx: p11874, FreeVars: ast.Identifiers{ "std", }, @@ -152123,13 +143974,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(20), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(23), }, File: p1, @@ -152164,9 +144015,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "foldl", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -152176,80 +144026,55 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(30), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11918, + Ctx: p11883, FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "r", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1133), - Column: int(39), - }, - End: ast.Location{ - Line: int(1133), - Column: int(40), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "r", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1133), - Column: int(42), - }, - End: ast.Location{ - Line: int(1133), - Column: int(43), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(45), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11921, + Ctx: p11886, FreeVars: ast.Identifiers{ "a", "r", @@ -152258,19 +144083,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(45), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11921, + Ctx: p11886, FreeVars: ast.Identifiers{ "r", }, @@ -152282,19 +144107,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(51), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11921, + Ctx: p11886, FreeVars: ast.Identifiers{ "a", }, @@ -152302,19 +144127,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(51), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11921, + Ctx: p11886, FreeVars: ast.Identifiers{ "a", }, @@ -152326,21 +144151,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(55), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11921, + Ctx: p11886, FreeVars: nil, }, + Value: float64(256), OriginalString: "256", }, }, @@ -152352,19 +144178,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(61), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11918, + Ctx: p11883, FreeVars: ast.Identifiers{ "bytes", }, @@ -152377,19 +144203,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(68), }, End: ast.Location{ - Line: int(1133), + Line: int(1118), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11918, + Ctx: p11883, FreeVars: nil, }, Value: true, @@ -152406,30 +144232,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1133), - Column: int(11), - }, - End: ast.Location{ - Line: int(1133), - Column: int(73), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1134), + Line: int(1119), Column: int(5), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(24), }, File: p1, @@ -152442,7 +144256,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "aux", "bytes", @@ -152452,19 +144266,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1134), + Line: int(1119), Column: int(8), }, End: ast.Location{ - Line: int(1134), + Line: int(1119), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "sanity", }, @@ -152473,19 +144287,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1134), + Line: int(1119), Column: int(9), }, End: ast.Location{ - Line: int(1134), + Line: int(1119), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "sanity", }, @@ -152497,13 +144311,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1135), + Line: int(1120), Column: int(7), }, End: ast.Location{ - Line: int(1135), + Line: int(1120), Column: int(71), }, File: p1, @@ -152516,31 +144330,30 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11468, + Ctx: p11433, FreeVars: nil, }, Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1135), + Line: int(1120), Column: int(13), }, End: ast.Location{ - Line: int(1135), + Line: int(1120), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11468, + Ctx: p11433, FreeVars: nil, }, Value: "Can only base64 encode strings / arrays of single bytes.", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ElseFodder: ast.Fodder{ @@ -152554,19 +144367,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(7), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "aux", "bytes", @@ -152575,13 +144388,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(7), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(10), }, File: p1, @@ -152594,7 +144407,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11468, + Ctx: p11433, FreeVars: ast.Identifiers{ "aux", }, @@ -152608,19 +144421,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(11), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11951, + Ctx: p11916, FreeVars: ast.Identifiers{ "bytes", }, @@ -152633,21 +144446,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(18), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11951, + Ctx: p11916, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -152656,25 +144470,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(21), }, End: ast.Location{ - Line: int(1137), + Line: int(1122), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11951, + Ctx: p11916, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -152692,18 +144505,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1093), - Column: int(3), - }, - End: ast.Location{ - Line: int(1137), - Column: int(24), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -152728,7 +144529,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "base64DecodeBytes", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -152752,39 +144552,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1140), - Column: int(21), - }, - End: ast.Location{ - Line: int(1140), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(5), }, End: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(22), }, File: p1, @@ -152797,7 +144586,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "base64_inv", "std", @@ -152807,19 +144596,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(8), }, End: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "std", "str", @@ -152830,14 +144619,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1141), - Column: int(8), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1141), - Column: int(27), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -152910,7 +144699,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -152922,19 +144710,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(8), }, End: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "std", "str", @@ -152943,19 +144731,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(8), }, End: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "std", }, @@ -152963,13 +144751,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(8), }, End: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(11), }, File: p1, @@ -153004,9 +144792,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -153016,19 +144803,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(19), }, End: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11982, + Ctx: p11947, FreeVars: ast.Identifiers{ "str", }, @@ -153051,21 +144838,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(26), }, End: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, CommaFodder: nil, @@ -153083,21 +144871,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(31), }, End: ast.Location{ - Line: int(1141), + Line: int(1126), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -153105,13 +144894,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1142), + Line: int(1127), Column: int(7), }, End: ast.Location{ - Line: int(1142), + Line: int(1127), Column: int(53), }, File: p1, @@ -153124,7 +144913,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "std", "str", @@ -153135,14 +144924,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1142), - Column: int(13), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1142), - Column: int(53), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -153215,7 +145004,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -153227,25 +145015,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1142), + Line: int(1127), Column: int(13), }, End: ast.Location{ - Line: int(1142), + Line: int(1127), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: nil, }, Value: "Not a base64 encoded string \"%s\"", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -153253,19 +145040,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1142), + Line: int(1127), Column: int(50), }, End: ast.Location{ - Line: int(1142), + Line: int(1127), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "str", }, @@ -153294,13 +145081,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1144), + Line: int(1129), Column: int(7), }, End: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(22), }, File: p1, @@ -153313,7 +145100,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "base64_inv", "std", @@ -153328,19 +145115,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1144), + Line: int(1129), Column: int(13), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(44), }, File: p1, }, Fodder: nil, - Ctx: p12006, + Ctx: p11971, FreeVars: ast.Identifiers{ "aux", "base64_inv", @@ -153348,77 +145135,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1144), - Column: int(17), - }, - End: ast.Location{ - Line: int(1144), - Column: int(20), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1144), - Column: int(22), - }, - End: ast.Location{ - Line: int(1144), - Column: int(23), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "r", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1144), - Column: int(25), - }, - End: ast.Location{ - Line: int(1144), - Column: int(26), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "r", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(9), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(44), }, File: p1, @@ -153431,7 +145179,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "aux", "base64_inv", @@ -153444,19 +145192,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(12), }, End: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "i", "std", @@ -153466,19 +145214,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(12), }, End: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "i", }, @@ -153490,19 +145238,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(17), }, End: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "std", "str", @@ -153511,19 +145259,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(17), }, End: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "std", }, @@ -153531,13 +145279,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(17), }, End: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(20), }, File: p1, @@ -153572,9 +145320,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -153584,19 +145331,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(28), }, End: ast.Location{ - Line: int(1145), + Line: int(1130), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12026, + Ctx: p11991, FreeVars: ast.Identifiers{ "str", }, @@ -153618,13 +145365,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1146), + Line: int(1131), Column: int(11), }, End: ast.Location{ - Line: int(1146), + Line: int(1131), Column: int(12), }, File: p1, @@ -153637,7 +145384,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "r", }, @@ -153655,13 +145402,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(11), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(44), }, File: p1, @@ -153674,7 +145421,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -153682,7 +145429,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "aux", "base64_inv", @@ -153699,19 +145446,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(22), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12038, + Ctx: p12003, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -153723,19 +145470,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(23), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -153745,19 +145492,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(23), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -153767,19 +145514,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(23), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -153789,19 +145536,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(23), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "base64_inv", }, @@ -153812,19 +145559,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(34), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "i", "str", @@ -153833,19 +145580,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(34), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "str", }, @@ -153856,19 +145603,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(38), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "i", }, @@ -153886,21 +145633,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(45), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -153909,19 +145657,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(50), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -153931,19 +145679,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(50), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -153953,19 +145701,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(50), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "base64_inv", }, @@ -153976,19 +145724,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(61), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "i", "str", @@ -153997,19 +145745,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(61), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(64), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "str", }, @@ -154020,19 +145768,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(65), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "i", }, @@ -154040,19 +145788,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(65), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: ast.Identifiers{ "i", }, @@ -154064,21 +145812,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(69), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -154093,21 +145842,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(76), }, End: ast.Location{ - Line: int(1149), + Line: int(1134), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12042, + Ctx: p12007, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -154120,30 +145870,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1149), - Column: int(17), - }, - End: ast.Location{ - Line: int(1149), - Column: int(79), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1151), + Line: int(1136), Column: int(11), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(44), }, File: p1, @@ -154156,7 +145894,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -154164,7 +145902,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "aux", "base64_inv", @@ -154182,13 +145920,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(13), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(86), }, File: p1, @@ -154201,7 +145939,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12080, + Ctx: p12045, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154211,19 +145949,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(16), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: ast.Identifiers{ "i", "str", @@ -154232,19 +145970,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(16), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: ast.Identifiers{ "i", "str", @@ -154253,19 +145991,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(16), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: ast.Identifiers{ "str", }, @@ -154276,19 +146014,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(20), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: ast.Identifiers{ "i", }, @@ -154296,19 +146034,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(20), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: ast.Identifiers{ "i", }, @@ -154320,21 +146058,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(24), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -154346,44 +146085,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(30), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: nil, }, Value: "=", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(39), }, End: ast.Location{ - Line: int(1152), + Line: int(1137), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: nil, }, Elements: nil, @@ -154401,19 +146139,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(18), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12080, + Ctx: p12045, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154425,19 +146163,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(19), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(85), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154447,19 +146185,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(19), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154469,19 +146207,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(20), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154491,19 +146229,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(20), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154513,19 +146251,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(20), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", }, @@ -154536,19 +146274,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(31), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "i", "str", @@ -154557,19 +146295,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(31), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "str", }, @@ -154580,19 +146318,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(35), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "i", }, @@ -154600,19 +146338,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(35), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "i", }, @@ -154624,21 +146362,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(39), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -154653,21 +146392,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(45), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: nil, }, + Value: float64(15), OriginalString: "15", }, }, @@ -154676,21 +146416,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(52), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -154699,19 +146440,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(57), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154721,19 +146462,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(57), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -154743,19 +146484,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(57), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "base64_inv", }, @@ -154766,19 +146507,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(68), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "i", "str", @@ -154787,19 +146528,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(68), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "str", }, @@ -154810,19 +146551,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(72), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "i", }, @@ -154830,19 +146571,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(72), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: ast.Identifiers{ "i", }, @@ -154854,21 +146595,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(76), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -154883,21 +146625,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(83), }, End: ast.Location{ - Line: int(1153), + Line: int(1138), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12100, + Ctx: p12065, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -154911,30 +146654,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1151), - Column: int(17), - }, - End: ast.Location{ - Line: int(1153), - Column: int(86), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1155), + Line: int(1140), Column: int(11), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(44), }, File: p1, @@ -154947,7 +146678,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(10), Comment: []string{ @@ -154955,7 +146686,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "aux", "base64_inv", @@ -154974,13 +146705,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(13), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(78), }, File: p1, @@ -154993,7 +146724,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12144, + Ctx: p12109, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -155003,19 +146734,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(16), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: ast.Identifiers{ "i", "str", @@ -155024,19 +146755,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(16), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: ast.Identifiers{ "i", "str", @@ -155045,19 +146776,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(16), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: ast.Identifiers{ "str", }, @@ -155068,19 +146799,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(20), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: ast.Identifiers{ "i", }, @@ -155088,19 +146819,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(20), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: ast.Identifiers{ "i", }, @@ -155112,21 +146843,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(24), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -155138,44 +146870,43 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(30), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: nil, }, Value: "=", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, }, ThenFodder: ast.Fodder{}, BranchTrue: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(39), }, End: ast.Location{ - Line: int(1156), + Line: int(1141), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: nil, }, Elements: nil, @@ -155193,19 +146924,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(18), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12144, + Ctx: p12109, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -155217,19 +146948,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(19), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -155239,19 +146970,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(19), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -155261,19 +146992,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(20), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -155283,19 +147014,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(20), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -155305,19 +147036,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(20), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "base64_inv", }, @@ -155328,19 +147059,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(31), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "i", "str", @@ -155349,19 +147080,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(31), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "str", }, @@ -155372,19 +147103,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(35), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "i", }, @@ -155392,19 +147123,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(35), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "i", }, @@ -155416,21 +147147,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(39), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -155445,21 +147177,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(45), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -155468,21 +147201,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(51), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: nil, }, + Value: float64(6), OriginalString: "6", }, }, @@ -155491,19 +147225,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(55), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "base64_inv", "i", @@ -155513,19 +147247,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(55), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "base64_inv", }, @@ -155536,19 +147270,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(66), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "i", "str", @@ -155557,19 +147291,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(66), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "str", }, @@ -155580,19 +147314,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(70), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "i", }, @@ -155600,19 +147334,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(70), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: ast.Identifiers{ "i", }, @@ -155624,21 +147358,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(74), }, End: ast.Location{ - Line: int(1157), + Line: int(1142), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12164, + Ctx: p12129, FreeVars: nil, }, + Value: float64(3), OriginalString: "3", }, }, @@ -155658,36 +147393,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1155), - Column: int(17), - }, - End: ast.Location{ - Line: int(1157), - Column: int(78), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(11), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "aux", "i", @@ -155701,13 +147424,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(11), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(14), }, File: p1, @@ -155720,7 +147443,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12011, + Ctx: p11976, FreeVars: ast.Identifiers{ "aux", }, @@ -155734,19 +147457,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(15), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "str", }, @@ -155759,19 +147482,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(20), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "i", }, @@ -155779,19 +147502,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(20), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "i", }, @@ -155803,21 +147526,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(24), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: nil, }, + Value: float64(4), OriginalString: "4", }, }, @@ -155827,19 +147551,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(27), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "n1", "n2", @@ -155850,19 +147574,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(27), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "n1", "n2", @@ -155872,19 +147596,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(27), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "n1", "r", @@ -155893,19 +147617,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(27), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "r", }, @@ -155917,19 +147641,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(31), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "n1", }, @@ -155942,19 +147666,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(36), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "n2", }, @@ -155967,19 +147691,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(41), }, End: ast.Location{ - Line: int(1158), + Line: int(1143), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12205, + Ctx: p12170, FreeVars: ast.Identifiers{ "n3", }, @@ -156004,36 +147728,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(7), }, End: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "aux", "str", @@ -156042,13 +147754,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(7), }, End: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(10), }, File: p1, @@ -156061,7 +147773,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p11961, + Ctx: p11926, FreeVars: ast.Identifiers{ "aux", }, @@ -156075,19 +147787,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(11), }, End: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12233, + Ctx: p12198, FreeVars: ast.Identifiers{ "str", }, @@ -156100,21 +147812,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(16), }, End: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12233, + Ctx: p12198, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -156123,19 +147836,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(19), }, End: ast.Location{ - Line: int(1159), + Line: int(1144), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12233, + Ctx: p12198, FreeVars: nil, }, Elements: nil, @@ -156156,18 +147869,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1140), - Column: int(3), - }, - End: ast.Location{ - Line: int(1159), - Column: int(22), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -156192,7 +147893,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "base64Decode", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -156215,39 +147915,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1161), - Column: int(16), - }, - End: ast.Location{ - Line: int(1161), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(5), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(58), }, File: p1, @@ -156260,7 +147949,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12243, + Ctx: p12208, FreeVars: ast.Identifiers{ "std", "str", @@ -156274,19 +147963,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(19), }, End: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12247, + Ctx: p12212, FreeVars: ast.Identifiers{ "std", "str", @@ -156295,19 +147984,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(19), }, End: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12247, + Ctx: p12212, FreeVars: ast.Identifiers{ "std", }, @@ -156315,13 +148004,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(19), }, End: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(22), }, File: p1, @@ -156356,9 +148045,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "base64DecodeBytes", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -156368,19 +148056,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(41), }, End: ast.Location{ - Line: int(1162), + Line: int(1147), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12256, + Ctx: p12221, FreeVars: ast.Identifiers{ "str", }, @@ -156399,36 +148087,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1162), - Column: int(11), - }, - End: ast.Location{ - Line: int(1162), - Column: int(45), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(5), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(58), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12243, + Ctx: p12208, FreeVars: ast.Identifiers{ "bytes", "std", @@ -156437,19 +148113,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(5), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12243, + Ctx: p12208, FreeVars: ast.Identifiers{ "std", }, @@ -156457,13 +148133,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(5), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(8), }, File: p1, @@ -156505,9 +148181,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -156517,25 +148192,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(14), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12268, + Ctx: p12233, FreeVars: nil, }, Value: "", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -156543,19 +148217,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(18), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12268, + Ctx: p12233, FreeVars: ast.Identifiers{ "bytes", "std", @@ -156564,19 +148238,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(18), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12268, + Ctx: p12233, FreeVars: ast.Identifiers{ "std", }, @@ -156584,13 +148258,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(18), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(21), }, File: p1, @@ -156625,9 +148299,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "map", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -156637,63 +148310,52 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(26), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12278, + Ctx: p12243, FreeVars: ast.Identifiers{ "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1163), - Column: int(35), - }, - End: ast.Location{ - Line: int(1163), - Column: int(36), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(38), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12282, + Ctx: p12247, FreeVars: ast.Identifiers{ "b", "std", @@ -156702,19 +148364,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(38), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12282, + Ctx: p12247, FreeVars: ast.Identifiers{ "std", }, @@ -156722,13 +148384,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(38), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(41), }, File: p1, @@ -156763,9 +148425,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "char", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -156775,19 +148436,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(47), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12291, + Ctx: p12256, FreeVars: ast.Identifiers{ "b", }, @@ -156811,19 +148472,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(51), }, End: ast.Location{ - Line: int(1163), + Line: int(1148), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12278, + Ctx: p12243, FreeVars: ast.Identifiers{ "bytes", }, @@ -156853,18 +148514,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1161), - Column: int(3), - }, - End: ast.Location{ - Line: int(1163), - Column: int(58), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -156889,7 +148538,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "reverse", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -156912,39 +148560,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1165), - Column: int(11), - }, - End: ast.Location{ - Line: int(1165), - Column: int(14), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(5), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(49), }, File: p1, @@ -156957,7 +148594,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12301, + Ctx: p12266, FreeVars: ast.Identifiers{ "arr", "std", @@ -156971,19 +148608,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(15), }, End: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12305, + Ctx: p12270, FreeVars: ast.Identifiers{ "arr", "std", @@ -156992,19 +148629,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(15), }, End: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12305, + Ctx: p12270, FreeVars: ast.Identifiers{ "std", }, @@ -157012,13 +148649,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(15), }, End: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(18), }, File: p1, @@ -157053,9 +148690,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -157065,19 +148701,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(26), }, End: ast.Location{ - Line: int(1166), + Line: int(1151), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12314, + Ctx: p12279, FreeVars: ast.Identifiers{ "arr", }, @@ -157096,36 +148732,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1166), - Column: int(11), - }, - End: ast.Location{ - Line: int(1166), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(5), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12301, + Ctx: p12266, FreeVars: ast.Identifiers{ "arr", "l", @@ -157135,19 +148759,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(5), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12301, + Ctx: p12266, FreeVars: ast.Identifiers{ "std", }, @@ -157155,13 +148779,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(5), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(8), }, File: p1, @@ -157203,9 +148827,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -157215,19 +148838,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(19), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12326, + Ctx: p12291, FreeVars: ast.Identifiers{ "l", }, @@ -157240,64 +148863,53 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(22), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12326, + Ctx: p12291, FreeVars: ast.Identifiers{ "arr", "l", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1167), - Column: int(31), - }, - End: ast.Location{ - Line: int(1167), - Column: int(32), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(34), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12332, + Ctx: p12297, FreeVars: ast.Identifiers{ "arr", "i", @@ -157307,19 +148919,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(34), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12332, + Ctx: p12297, FreeVars: ast.Identifiers{ "arr", }, @@ -157330,19 +148942,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(38), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12332, + Ctx: p12297, FreeVars: ast.Identifiers{ "i", "l", @@ -157351,19 +148963,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(38), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12332, + Ctx: p12297, FreeVars: ast.Identifiers{ "i", "l", @@ -157372,19 +148984,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(38), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12332, + Ctx: p12297, FreeVars: ast.Identifiers{ "l", }, @@ -157396,19 +149008,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(42), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12332, + Ctx: p12297, FreeVars: ast.Identifiers{ "i", }, @@ -157421,21 +149033,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(46), }, End: ast.Location{ - Line: int(1167), + Line: int(1152), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12332, + Ctx: p12297, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -157456,18 +149069,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1165), - Column: int(3), - }, - End: ast.Location{ - Line: int(1167), - Column: int(49), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -157492,7 +149093,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "sort", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -157516,64 +149116,42 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1170), - Column: int(8), - }, - End: ast.Location{ - Line: int(1170), - Column: int(11), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1170), - Column: int(18), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1155), + Column: int(18), + }, + End: ast.Location{ + Line: int(1155), + Column: int(20), + }, + File: p1, }, - End: ast.Location{ - Line: int(1170), - Column: int(20), + Fodder: ast.Fodder{}, + Ctx: p12316, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p12350, - FreeVars: ast.Identifiers{ - "id", - }, - }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1170), - Column: int(13), - }, - End: ast.Location{ - Line: int(1170), - Column: int(20), + Id: "id", }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -157582,13 +149160,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1171), + Line: int(1156), Column: int(5), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(67), }, File: p1, @@ -157601,7 +149179,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "id", @@ -157617,19 +149195,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1171), + Line: int(1156), Column: int(11), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, }, Fodder: nil, - Ctx: p12357, + Ctx: p12323, FreeVars: ast.Identifiers{ "id", "quickSort", @@ -157637,64 +149215,42 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1171), - Column: int(21), - }, - End: ast.Location{ - Line: int(1171), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, }, }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1171), - Column: int(31), + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1156), + Column: int(31), + }, + End: ast.Location{ + Line: int(1156), + Column: int(33), + }, + File: p1, }, - End: ast.Location{ - Line: int(1171), - Column: int(33), + Fodder: ast.Fodder{}, + Ctx: p12328, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p12361, - FreeVars: ast.Identifiers{ - "id", }, + Id: "id", }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1171), - Column: int(26), - }, - End: ast.Location{ - Line: int(1171), - Column: int(33), - }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -157703,13 +149259,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(7), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, @@ -157722,7 +149278,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -157738,19 +149294,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(17), }, End: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12368, + Ctx: p12335, FreeVars: ast.Identifiers{ "arr", "std", @@ -157759,19 +149315,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(17), }, End: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12368, + Ctx: p12335, FreeVars: ast.Identifiers{ "std", }, @@ -157779,13 +149335,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(17), }, End: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(20), }, File: p1, @@ -157820,9 +149376,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -157832,19 +149387,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(28), }, End: ast.Location{ - Line: int(1172), + Line: int(1157), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12377, + Ctx: p12344, FreeVars: ast.Identifiers{ "arr", }, @@ -157863,30 +149418,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1172), - Column: int(13), - }, - End: ast.Location{ - Line: int(1172), - Column: int(32), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(7), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, @@ -157899,7 +149442,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -157911,19 +149454,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(10), }, End: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "std", @@ -157932,19 +149475,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(10), }, End: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "std", @@ -157953,19 +149496,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(10), }, End: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "std", }, @@ -157973,13 +149516,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(10), }, End: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(13), }, File: p1, @@ -158014,9 +149557,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -158026,19 +149568,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(21), }, End: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12393, + Ctx: p12360, FreeVars: ast.Identifiers{ "arr", }, @@ -158060,21 +149602,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(29), }, End: ast.Location{ - Line: int(1173), + Line: int(1158), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -158082,13 +149625,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1174), + Line: int(1159), Column: int(9), }, End: ast.Location{ - Line: int(1174), + Line: int(1159), Column: int(12), }, File: p1, @@ -158101,7 +149644,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", }, @@ -158119,13 +149662,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1176), + Line: int(1161), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, @@ -158138,7 +149681,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -158155,49 +149698,38 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1176), + Line: int(1161), Column: int(21), }, End: ast.Location{ - Line: int(1176), + Line: int(1161), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12405, + Ctx: p12372, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1176), - Column: int(15), - }, - End: ast.Location{ - Line: int(1176), - Column: int(22), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, @@ -158210,7 +149742,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -158228,19 +149760,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(23), }, End: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12411, + Ctx: p12378, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -158250,19 +149782,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(23), }, End: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12411, + Ctx: p12378, FreeVars: ast.Identifiers{ "keyF", }, @@ -158276,19 +149808,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(28), }, End: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12417, + Ctx: p12384, FreeVars: ast.Identifiers{ "arr", "pos", @@ -158297,19 +149829,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(28), }, End: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12417, + Ctx: p12384, FreeVars: ast.Identifiers{ "arr", }, @@ -158320,19 +149852,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(32), }, End: ast.Location{ - Line: int(1177), + Line: int(1162), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12417, + Ctx: p12384, FreeVars: ast.Identifiers{ "pos", }, @@ -158354,30 +149886,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1177), - Column: int(15), - }, - End: ast.Location{ - Line: int(1177), - Column: int(37), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, @@ -158390,7 +149910,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -158409,19 +149929,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(22), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(94), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12428, + Ctx: p12395, FreeVars: ast.Identifiers{ "arr", "l", @@ -158432,19 +149952,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(22), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12428, + Ctx: p12395, FreeVars: ast.Identifiers{ "std", }, @@ -158452,13 +149972,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(22), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(25), }, File: p1, @@ -158493,9 +150013,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -158505,19 +150024,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(36), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12437, + Ctx: p12404, FreeVars: ast.Identifiers{ "l", }, @@ -158525,19 +150044,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(36), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12437, + Ctx: p12404, FreeVars: ast.Identifiers{ "l", }, @@ -158549,21 +150068,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(40), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12437, + Ctx: p12404, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -158573,64 +150093,53 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(43), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(93), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12437, + Ctx: p12404, FreeVars: ast.Identifiers{ "arr", "pos", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1178), - Column: int(52), - }, - End: ast.Location{ - Line: int(1178), - Column: int(53), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(55), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(93), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "arr", "i", @@ -158640,19 +150149,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(58), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "i", "pos", @@ -158661,19 +150170,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(58), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "i", }, @@ -158685,19 +150194,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(62), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "pos", }, @@ -158709,19 +150218,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(71), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "arr", "i", @@ -158730,19 +150239,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(71), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(74), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "arr", }, @@ -158753,19 +150262,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(75), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "i", }, @@ -158779,19 +150288,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(83), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(93), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "arr", "i", @@ -158800,19 +150309,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(83), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(86), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "arr", }, @@ -158823,19 +150332,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(87), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(92), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "i", }, @@ -158843,19 +150352,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(87), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(88), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: ast.Identifiers{ "i", }, @@ -158867,21 +150376,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(91), }, End: ast.Location{ - Line: int(1178), + Line: int(1163), Column: int(92), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12446, + Ctx: p12413, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -158902,30 +150412,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1178), - Column: int(15), - }, - End: ast.Location{ - Line: int(1178), - Column: int(94), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, @@ -158938,7 +150436,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -158957,19 +150455,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(22), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12474, + Ctx: p12441, FreeVars: ast.Identifiers{ "keyF", "pivot", @@ -158980,19 +150478,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(22), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12474, + Ctx: p12441, FreeVars: ast.Identifiers{ "std", }, @@ -159000,13 +150498,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(22), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(25), }, File: p1, @@ -159041,9 +150539,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "filter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -159053,64 +150550,53 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(33), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12483, + Ctx: p12450, FreeVars: ast.Identifiers{ "keyF", "pivot", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1179), - Column: int(42), - }, - End: ast.Location{ - Line: int(1179), - Column: int(43), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(45), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12487, + Ctx: p12454, FreeVars: ast.Identifiers{ "keyF", "pivot", @@ -159120,19 +150606,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(45), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12487, + Ctx: p12454, FreeVars: ast.Identifiers{ "keyF", "x", @@ -159141,19 +150627,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(45), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12487, + Ctx: p12454, FreeVars: ast.Identifiers{ "keyF", }, @@ -159167,19 +150653,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(50), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12495, + Ctx: p12462, FreeVars: ast.Identifiers{ "x", }, @@ -159201,19 +150687,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(55), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(60), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12487, + Ctx: p12454, FreeVars: ast.Identifiers{ "pivot", }, @@ -159228,19 +150714,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(62), }, End: ast.Location{ - Line: int(1179), + Line: int(1164), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12483, + Ctx: p12450, FreeVars: ast.Identifiers{ "rest", }, @@ -159259,30 +150745,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1179), - Column: int(15), - }, - End: ast.Location{ - Line: int(1179), - Column: int(67), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, @@ -159295,7 +150769,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -159315,19 +150789,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(23), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12506, + Ctx: p12473, FreeVars: ast.Identifiers{ "keyF", "pivot", @@ -159338,19 +150812,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(23), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12506, + Ctx: p12473, FreeVars: ast.Identifiers{ "std", }, @@ -159358,13 +150832,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(23), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(26), }, File: p1, @@ -159399,9 +150873,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "filter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -159411,64 +150884,53 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(34), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12515, + Ctx: p12482, FreeVars: ast.Identifiers{ "keyF", "pivot", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1180), - Column: int(43), - }, - End: ast.Location{ - Line: int(1180), - Column: int(44), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(46), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12519, + Ctx: p12486, FreeVars: ast.Identifiers{ "keyF", "pivot", @@ -159478,19 +150940,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(46), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12519, + Ctx: p12486, FreeVars: ast.Identifiers{ "keyF", "x", @@ -159499,19 +150961,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(46), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12519, + Ctx: p12486, FreeVars: ast.Identifiers{ "keyF", }, @@ -159525,19 +150987,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(51), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12527, + Ctx: p12494, FreeVars: ast.Identifiers{ "x", }, @@ -159559,19 +151021,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(57), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12519, + Ctx: p12486, FreeVars: ast.Identifiers{ "pivot", }, @@ -159586,19 +151048,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(64), }, End: ast.Location{ - Line: int(1180), + Line: int(1165), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12515, + Ctx: p12482, FreeVars: ast.Identifiers{ "rest", }, @@ -159617,36 +151079,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1180), - Column: int(15), - }, - End: ast.Location{ - Line: int(1180), - Column: int(69), - }, - File: p1, - }, }, }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -159659,19 +151109,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -159683,19 +151133,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "keyF", "left", @@ -159705,13 +151155,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(9), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(18), }, File: p1, @@ -159724,7 +151174,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "quickSort", }, @@ -159738,19 +151188,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(19), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12544, + Ctx: p12511, FreeVars: ast.Identifiers{ "left", }, @@ -159763,19 +151213,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(25), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12544, + Ctx: p12511, FreeVars: ast.Identifiers{ "keyF", }, @@ -159797,19 +151247,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(33), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "arr", "pos", @@ -159820,19 +151270,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(34), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12552, + Ctx: p12519, FreeVars: ast.Identifiers{ "arr", "pos", @@ -159841,19 +151291,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(34), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12552, + Ctx: p12519, FreeVars: ast.Identifiers{ "arr", }, @@ -159864,19 +151314,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(38), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12552, + Ctx: p12519, FreeVars: ast.Identifiers{ "pos", }, @@ -159898,19 +151348,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(46), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "keyF", "quickSort", @@ -159920,19 +151370,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(46), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12361, + Ctx: p12328, FreeVars: ast.Identifiers{ "quickSort", }, @@ -159946,19 +151396,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(56), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12564, + Ctx: p12531, FreeVars: ast.Identifiers{ "right", }, @@ -159971,19 +151421,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(63), }, End: ast.Location{ - Line: int(1181), + Line: int(1166), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12564, + Ctx: p12531, FreeVars: ast.Identifiers{ "keyF", }, @@ -160011,30 +151461,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1183), + Line: int(1168), Column: int(5), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(67), }, File: p1, @@ -160047,7 +151485,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -160063,77 +151501,52 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1183), + Line: int(1168), Column: int(11), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(20), }, File: p1, }, Fodder: nil, - Ctx: p12573, + Ctx: p12540, FreeVars: ast.Identifiers{ "keyF", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1183), - Column: int(17), - }, - End: ast.Location{ - Line: int(1183), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1183), - Column: int(20), - }, - End: ast.Location{ - Line: int(1183), - Column: int(21), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(7), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(20), }, File: p1, @@ -160146,7 +151559,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12578, + Ctx: p12545, FreeVars: ast.Identifiers{ "a", "b", @@ -160162,19 +151575,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(18), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12582, + Ctx: p12549, FreeVars: ast.Identifiers{ "a", "std", @@ -160183,19 +151596,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(18), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12582, + Ctx: p12549, FreeVars: ast.Identifiers{ "std", }, @@ -160203,13 +151616,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(18), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(21), }, File: p1, @@ -160244,9 +151657,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -160256,19 +151668,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(29), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12591, + Ctx: p12558, FreeVars: ast.Identifiers{ "a", }, @@ -160287,18 +151699,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1184), - Column: int(13), - }, - End: ast.Location{ - Line: int(1184), - Column: int(31), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: ast.Fodder{}, @@ -160307,19 +151707,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(38), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12594, + Ctx: p12561, FreeVars: ast.Identifiers{ "b", "std", @@ -160328,19 +151728,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(38), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12594, + Ctx: p12561, FreeVars: ast.Identifiers{ "std", }, @@ -160348,13 +151748,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(38), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(41), }, File: p1, @@ -160389,9 +151789,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -160401,19 +151800,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(49), }, End: ast.Location{ - Line: int(1184), + Line: int(1169), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12603, + Ctx: p12570, FreeVars: ast.Identifiers{ "b", }, @@ -160432,30 +151831,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1184), - Column: int(33), - }, - End: ast.Location{ - Line: int(1184), - Column: int(51), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1185), + Line: int(1170), Column: int(7), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(20), }, File: p1, @@ -160468,7 +151855,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12578, + Ctx: p12545, FreeVars: ast.Identifiers{ "a", "b", @@ -160486,19 +151873,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1185), + Line: int(1170), Column: int(13), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(43), }, File: p1, }, Fodder: nil, - Ctx: p12610, + Ctx: p12577, FreeVars: ast.Identifiers{ "a", "aux", @@ -160510,77 +151897,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1185), - Column: int(17), - }, - End: ast.Location{ - Line: int(1185), - Column: int(18), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1185), - Column: int(20), - }, - End: ast.Location{ - Line: int(1185), - Column: int(21), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "prefix", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1185), - Column: int(23), - }, - End: ast.Location{ - Line: int(1185), - Column: int(29), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "prefix", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1186), + Line: int(1171), Column: int(9), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(43), }, File: p1, @@ -160593,7 +151941,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", "aux", @@ -160610,19 +151958,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1186), + Line: int(1171), Column: int(12), }, End: ast.Location{ - Line: int(1186), + Line: int(1171), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "i", "la", @@ -160631,19 +151979,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1186), + Line: int(1171), Column: int(12), }, End: ast.Location{ - Line: int(1186), + Line: int(1171), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "i", }, @@ -160655,19 +152003,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1186), + Line: int(1171), Column: int(17), }, End: ast.Location{ - Line: int(1186), + Line: int(1171), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "la", }, @@ -160679,19 +152027,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(11), }, End: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "b", "j", @@ -160702,13 +152050,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(11), }, End: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(17), }, File: p1, @@ -160721,7 +152069,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "prefix", }, @@ -160735,14 +152083,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1187), - Column: int(20), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1187), - Column: int(25), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -160816,7 +152164,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -160828,19 +152175,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(20), }, End: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "b", }, @@ -160853,19 +152200,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(22), }, End: ast.Location{ - Line: int(1187), + Line: int(1172), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "j", }, @@ -160938,19 +152285,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1188), + Line: int(1173), Column: int(14), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", "aux", @@ -160966,19 +152313,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1188), + Line: int(1173), Column: int(17), }, End: ast.Location{ - Line: int(1188), + Line: int(1173), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "j", "lb", @@ -160987,19 +152334,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1188), + Line: int(1173), Column: int(17), }, End: ast.Location{ - Line: int(1188), + Line: int(1173), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "j", }, @@ -161011,19 +152358,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1188), + Line: int(1173), Column: int(22), }, End: ast.Location{ - Line: int(1188), + Line: int(1173), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "lb", }, @@ -161035,19 +152382,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(11), }, End: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", "i", @@ -161058,13 +152405,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(11), }, End: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(17), }, File: p1, @@ -161077,7 +152424,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "prefix", }, @@ -161091,14 +152438,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1189), - Column: int(20), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1189), - Column: int(25), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -161172,7 +152519,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -161184,19 +152530,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(20), }, End: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", }, @@ -161209,19 +152555,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(22), }, End: ast.Location{ - Line: int(1189), + Line: int(1174), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "i", }, @@ -161294,13 +152640,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(11), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(43), }, File: p1, @@ -161313,7 +152659,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", "aux", @@ -161327,19 +152673,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(14), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", "b", @@ -161351,19 +152697,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(14), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", "i", @@ -161373,19 +152719,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(14), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "keyF", }, @@ -161399,19 +152745,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(19), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12682, + Ctx: p12649, FreeVars: ast.Identifiers{ "a", "i", @@ -161420,19 +152766,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(19), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12682, + Ctx: p12649, FreeVars: ast.Identifiers{ "a", }, @@ -161443,19 +152789,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(21), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12682, + Ctx: p12649, FreeVars: ast.Identifiers{ "i", }, @@ -161480,19 +152826,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(28), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "b", "j", @@ -161502,19 +152848,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(28), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "keyF", }, @@ -161528,19 +152874,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(33), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12694, + Ctx: p12661, FreeVars: ast.Identifiers{ "b", "j", @@ -161549,19 +152895,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(33), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12694, + Ctx: p12661, FreeVars: ast.Identifiers{ "b", }, @@ -161572,19 +152918,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(35), }, End: ast.Location{ - Line: int(1191), + Line: int(1176), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12694, + Ctx: p12661, FreeVars: ast.Identifiers{ "j", }, @@ -161609,19 +152955,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(13), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "a", "aux", @@ -161633,13 +152979,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(13), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(16), }, File: p1, @@ -161652,7 +152998,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "aux", }, @@ -161666,19 +153012,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(17), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12707, + Ctx: p12674, FreeVars: ast.Identifiers{ "i", }, @@ -161686,19 +153032,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(17), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12707, + Ctx: p12674, FreeVars: ast.Identifiers{ "i", }, @@ -161710,21 +153056,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(21), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12707, + Ctx: p12674, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -161734,19 +153081,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(24), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12707, + Ctx: p12674, FreeVars: ast.Identifiers{ "j", }, @@ -161759,19 +153106,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(27), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12707, + Ctx: p12674, FreeVars: ast.Identifiers{ "a", "i", @@ -161781,19 +153128,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(27), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12707, + Ctx: p12674, FreeVars: ast.Identifiers{ "prefix", }, @@ -161805,19 +153152,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(36), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12707, + Ctx: p12674, FreeVars: ast.Identifiers{ "a", "i", @@ -161828,19 +153175,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(37), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12722, + Ctx: p12689, FreeVars: ast.Identifiers{ "a", "i", @@ -161849,19 +153196,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(37), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12722, + Ctx: p12689, FreeVars: ast.Identifiers{ "a", }, @@ -161872,19 +153219,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(39), }, End: ast.Location{ - Line: int(1192), + Line: int(1177), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12722, + Ctx: p12689, FreeVars: ast.Identifiers{ "i", }, @@ -161922,19 +153269,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(13), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "aux", "b", @@ -161946,13 +153293,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(13), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(16), }, File: p1, @@ -161965,7 +153312,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12615, + Ctx: p12582, FreeVars: ast.Identifiers{ "aux", }, @@ -161979,19 +153326,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(17), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12736, + Ctx: p12703, FreeVars: ast.Identifiers{ "i", }, @@ -162004,19 +153351,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(20), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12736, + Ctx: p12703, FreeVars: ast.Identifiers{ "j", }, @@ -162024,19 +153371,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(20), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12736, + Ctx: p12703, FreeVars: ast.Identifiers{ "j", }, @@ -162048,21 +153395,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(24), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12736, + Ctx: p12703, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -162072,19 +153420,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(27), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12736, + Ctx: p12703, FreeVars: ast.Identifiers{ "b", "j", @@ -162094,19 +153442,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(27), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12736, + Ctx: p12703, FreeVars: ast.Identifiers{ "prefix", }, @@ -162118,19 +153466,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(36), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12736, + Ctx: p12703, FreeVars: ast.Identifiers{ "b", "j", @@ -162141,19 +153489,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(37), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12751, + Ctx: p12718, FreeVars: ast.Identifiers{ "b", "j", @@ -162162,19 +153510,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(37), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12751, + Ctx: p12718, FreeVars: ast.Identifiers{ "b", }, @@ -162185,19 +153533,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(39), }, End: ast.Location{ - Line: int(1194), + Line: int(1179), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12751, + Ctx: p12718, FreeVars: ast.Identifiers{ "j", }, @@ -162230,36 +153578,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(7), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12578, + Ctx: p12545, FreeVars: ast.Identifiers{ "aux", }, @@ -162267,13 +153603,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(7), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(10), }, File: p1, @@ -162286,7 +153622,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12578, + Ctx: p12545, FreeVars: ast.Identifiers{ "aux", }, @@ -162300,21 +153636,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(11), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(12), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12764, + Ctx: p12731, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -162323,21 +153660,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(14), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12764, + Ctx: p12731, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -162346,19 +153684,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(17), }, End: ast.Location{ - Line: int(1195), + Line: int(1180), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12764, + Ctx: p12731, FreeVars: nil, }, Elements: nil, @@ -162380,30 +153718,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(5), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(67), }, File: p1, @@ -162416,7 +153742,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -162433,19 +153759,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(15), }, End: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12772, + Ctx: p12739, FreeVars: ast.Identifiers{ "arr", "std", @@ -162454,19 +153780,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(15), }, End: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12772, + Ctx: p12739, FreeVars: ast.Identifiers{ "std", }, @@ -162474,13 +153800,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(15), }, End: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(18), }, File: p1, @@ -162515,9 +153841,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -162527,19 +153852,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(26), }, End: ast.Location{ - Line: int(1197), + Line: int(1182), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12781, + Ctx: p12748, FreeVars: ast.Identifiers{ "arr", }, @@ -162558,30 +153883,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1197), - Column: int(11), - }, - End: ast.Location{ - Line: int(1197), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(5), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(67), }, File: p1, @@ -162594,7 +153907,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -162607,19 +153920,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(8), }, End: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "std", @@ -162628,19 +153941,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(8), }, End: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "std", @@ -162649,19 +153962,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(8), }, End: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "std", }, @@ -162669,13 +153982,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(8), }, End: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(11), }, File: p1, @@ -162710,9 +154023,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -162722,19 +154034,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(19), }, End: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12797, + Ctx: p12764, FreeVars: ast.Identifiers{ "arr", }, @@ -162756,21 +154068,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(27), }, End: ast.Location{ - Line: int(1198), + Line: int(1183), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12350, + Ctx: p12316, FreeVars: nil, }, + Value: float64(30), OriginalString: "30", }, }, @@ -162778,19 +154091,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(7), }, End: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -162800,13 +154113,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(7), }, End: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(16), }, File: p1, @@ -162819,7 +154132,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "quickSort", }, @@ -162833,19 +154146,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(17), }, End: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12807, + Ctx: p12774, FreeVars: ast.Identifiers{ "arr", }, @@ -162863,19 +154176,19 @@ var _StdAst = &ast.DesugaredObject{ Arg: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(27), }, End: ast.Location{ - Line: int(1199), + Line: int(1184), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12807, + Ctx: p12774, FreeVars: ast.Identifiers{ "keyF", }, @@ -162902,13 +154215,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(7), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(67), }, File: p1, @@ -162921,7 +154234,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -162938,19 +154251,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(19), }, End: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12818, + Ctx: p12785, FreeVars: ast.Identifiers{ "l", "std", @@ -162959,19 +154272,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(19), }, End: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12818, + Ctx: p12785, FreeVars: ast.Identifiers{ "std", }, @@ -162979,13 +154292,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(19), }, End: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(22), }, File: p1, @@ -163020,9 +154333,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "floor", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -163032,19 +154344,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(29), }, End: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12827, + Ctx: p12794, FreeVars: ast.Identifiers{ "l", }, @@ -163052,19 +154364,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(29), }, End: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12827, + Ctx: p12794, FreeVars: ast.Identifiers{ "l", }, @@ -163076,21 +154388,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(33), }, End: ast.Location{ - Line: int(1201), + Line: int(1186), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12827, + Ctx: p12794, FreeVars: nil, }, + Value: float64(2), OriginalString: "2", }, }, @@ -163106,30 +154419,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1201), - Column: int(13), - }, - End: ast.Location{ - Line: int(1201), - Column: int(35), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(7), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(67), }, File: p1, @@ -163142,7 +154443,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -163161,14 +154462,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1202), - Column: int(20), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1202), - Column: int(29), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -163242,7 +154543,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -163254,19 +154554,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(20), }, End: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12845, + Ctx: p12812, FreeVars: ast.Identifiers{ "arr", }, @@ -163301,19 +154601,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(25), }, End: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12845, + Ctx: p12812, FreeVars: ast.Identifiers{ "mid", }, @@ -163354,18 +154654,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1202), - Column: int(13), - }, - End: ast.Location{ - Line: int(1202), - Column: int(29), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: ast.Fodder{}, @@ -163376,14 +154664,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1202), - Column: int(39), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1202), - Column: int(48), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -163457,7 +154745,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -163469,19 +154756,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(39), }, End: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12860, + Ctx: p12827, FreeVars: ast.Identifiers{ "arr", }, @@ -163494,19 +154781,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(43), }, End: ast.Location{ - Line: int(1202), + Line: int(1187), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12860, + Ctx: p12827, FreeVars: ast.Identifiers{ "mid", }, @@ -163569,36 +154856,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1202), - Column: int(31), - }, - End: ast.Location{ - Line: int(1202), - Column: int(48), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(7), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "keyF", "left", @@ -163610,13 +154885,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(7), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(12), }, File: p1, @@ -163629,7 +154904,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12350, + Ctx: p12316, FreeVars: ast.Identifiers{ "merge", }, @@ -163643,19 +154918,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(13), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12873, + Ctx: p12840, FreeVars: ast.Identifiers{ "keyF", "left", @@ -163665,19 +154940,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(13), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12873, + Ctx: p12840, FreeVars: ast.Identifiers{ "std", }, @@ -163685,13 +154960,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(13), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(16), }, File: p1, @@ -163726,9 +155001,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "sort", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -163738,19 +155012,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(22), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12882, + Ctx: p12849, FreeVars: ast.Identifiers{ "left", }, @@ -163768,19 +155042,19 @@ var _StdAst = &ast.DesugaredObject{ Arg: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(33), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12882, + Ctx: p12849, FreeVars: ast.Identifiers{ "keyF", }, @@ -163802,19 +155076,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(40), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12873, + Ctx: p12840, FreeVars: ast.Identifiers{ "keyF", "right", @@ -163824,19 +155098,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(40), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12873, + Ctx: p12840, FreeVars: ast.Identifiers{ "std", }, @@ -163844,13 +155118,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(40), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(43), }, File: p1, @@ -163885,9 +155159,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "sort", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -163897,19 +155170,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(49), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12896, + Ctx: p12863, FreeVars: ast.Identifiers{ "right", }, @@ -163927,19 +155200,19 @@ var _StdAst = &ast.DesugaredObject{ Arg: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(61), }, End: ast.Location{ - Line: int(1203), + Line: int(1188), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12896, + Ctx: p12863, FreeVars: ast.Identifiers{ "keyF", }, @@ -163973,18 +155246,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1170), - Column: int(3), - }, - End: ast.Location{ - Line: int(1203), - Column: int(67), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -164009,7 +155270,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "uniq", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -164033,64 +155293,42 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1205), - Column: int(8), - }, - End: ast.Location{ - Line: int(1205), - Column: int(11), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1205), - Column: int(18), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1190), + Column: int(18), + }, + End: ast.Location{ + Line: int(1190), + Column: int(20), + }, + File: p1, }, - End: ast.Location{ - Line: int(1205), - Column: int(20), + Fodder: ast.Fodder{}, + Ctx: p12874, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p12906, - FreeVars: ast.Identifiers{ - "id", - }, - }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1205), - Column: int(13), - }, - End: ast.Location{ - Line: int(1205), - Column: int(20), + Id: "id", }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -164099,13 +155337,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1206), + Line: int(1191), Column: int(5), }, End: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(26), }, File: p1, @@ -164118,7 +155356,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12906, + Ctx: p12874, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -164133,77 +155371,52 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1206), + Line: int(1191), Column: int(11), }, End: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(16), }, File: p1, }, Fodder: nil, - Ctx: p12913, + Ctx: p12881, FreeVars: ast.Identifiers{ "keyF", "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1206), - Column: int(13), - }, - End: ast.Location{ - Line: int(1206), - Column: int(14), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1206), - Column: int(16), - }, - End: ast.Location{ - Line: int(1206), - Column: int(17), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(7), }, End: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(16), }, File: p1, @@ -164216,7 +155429,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", "b", @@ -164227,19 +155440,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(10), }, End: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", "std", @@ -164248,19 +155461,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(10), }, End: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", "std", @@ -164269,19 +155482,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(10), }, End: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "std", }, @@ -164289,13 +155502,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(10), }, End: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(13), }, File: p1, @@ -164330,9 +155543,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -164342,19 +155554,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(21), }, End: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12931, + Ctx: p12899, FreeVars: ast.Identifiers{ "a", }, @@ -164376,21 +155588,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(27), }, End: ast.Location{ - Line: int(1207), + Line: int(1192), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -164398,13 +155611,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1208), + Line: int(1193), Column: int(9), }, End: ast.Location{ - Line: int(1208), + Line: int(1193), Column: int(12), }, File: p1, @@ -164417,7 +155630,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "b", }, @@ -164427,19 +155640,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1208), + Line: int(1193), Column: int(10), }, End: ast.Location{ - Line: int(1208), + Line: int(1193), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12939, + Ctx: p12907, FreeVars: ast.Identifiers{ "b", }, @@ -164463,19 +155676,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(12), }, End: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", "b", @@ -164486,19 +155699,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(15), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", "b", @@ -164509,19 +155722,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(15), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", "keyF", @@ -164531,19 +155744,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(15), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "keyF", }, @@ -164557,19 +155770,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(20), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12952, + Ctx: p12920, FreeVars: ast.Identifiers{ "a", "std", @@ -164578,19 +155791,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(20), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12952, + Ctx: p12920, FreeVars: ast.Identifiers{ "a", }, @@ -164601,19 +155814,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(22), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12952, + Ctx: p12920, FreeVars: ast.Identifiers{ "a", "std", @@ -164622,19 +155835,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(22), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12952, + Ctx: p12920, FreeVars: ast.Identifiers{ "a", "std", @@ -164643,19 +155856,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(22), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12952, + Ctx: p12920, FreeVars: ast.Identifiers{ "std", }, @@ -164663,13 +155876,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(22), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(25), }, File: p1, @@ -164704,9 +155917,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -164716,19 +155928,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(33), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12967, + Ctx: p12935, FreeVars: ast.Identifiers{ "a", }, @@ -164750,21 +155962,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(38), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12952, + Ctx: p12920, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -164786,19 +155999,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(45), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "b", "keyF", @@ -164807,19 +156020,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(45), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "keyF", }, @@ -164833,19 +156046,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(50), }, End: ast.Location{ - Line: int(1209), + Line: int(1194), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12976, + Ctx: p12944, FreeVars: ast.Identifiers{ "b", }, @@ -164867,13 +156080,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1210), + Line: int(1195), Column: int(9), }, End: ast.Location{ - Line: int(1210), + Line: int(1195), Column: int(10), }, File: p1, @@ -164886,7 +156099,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", }, @@ -164904,19 +156117,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(9), }, End: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", "b", @@ -164925,13 +156138,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(9), }, End: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(10), }, File: p1, @@ -164944,7 +156157,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "a", }, @@ -164956,19 +156169,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(13), }, End: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12918, + Ctx: p12886, FreeVars: ast.Identifiers{ "b", }, @@ -164978,19 +156191,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(14), }, End: ast.Location{ - Line: int(1212), + Line: int(1197), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12991, + Ctx: p12959, FreeVars: ast.Identifiers{ "b", }, @@ -165009,36 +156222,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(5), }, End: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12906, + Ctx: p12874, FreeVars: ast.Identifiers{ "arr", "f", @@ -165048,19 +156249,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(5), }, End: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p12906, + Ctx: p12874, FreeVars: ast.Identifiers{ "std", }, @@ -165068,13 +156269,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(5), }, End: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(8), }, File: p1, @@ -165116,9 +156317,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "foldl", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -165128,19 +156328,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(15), }, End: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13003, + Ctx: p12971, FreeVars: ast.Identifiers{ "f", }, @@ -165153,19 +156353,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(18), }, End: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13003, + Ctx: p12971, FreeVars: ast.Identifiers{ "arr", }, @@ -165178,19 +156378,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(23), }, End: ast.Location{ - Line: int(1213), + Line: int(1198), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13003, + Ctx: p12971, FreeVars: nil, }, Elements: nil, @@ -165210,18 +156410,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1205), - Column: int(3), - }, - End: ast.Location{ - Line: int(1213), - Column: int(26), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -165246,7 +156434,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "set", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -165270,64 +156457,42 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1215), - Column: int(7), - }, - End: ast.Location{ - Line: int(1215), - Column: int(10), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1215), - Column: int(17), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1200), + Column: int(17), + }, + End: ast.Location{ + Line: int(1200), + Column: int(19), + }, + File: p1, }, - End: ast.Location{ - Line: int(1215), - Column: int(19), + Fodder: ast.Fodder{}, + Ctx: p12982, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p13013, - FreeVars: ast.Identifiers{ - "id", }, + Id: "id", }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1215), - Column: int(12), - }, - End: ast.Location{ - Line: int(1215), - Column: int(19), - }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -165336,19 +156501,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(5), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13013, + Ctx: p12982, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -165358,19 +156523,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(5), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13013, + Ctx: p12982, FreeVars: ast.Identifiers{ "std", }, @@ -165378,13 +156543,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(5), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(8), }, File: p1, @@ -165426,9 +156591,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "uniq", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -165438,19 +156602,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(14), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13025, + Ctx: p12994, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -165460,19 +156624,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(14), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13025, + Ctx: p12994, FreeVars: ast.Identifiers{ "std", }, @@ -165480,13 +156644,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(14), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(17), }, File: p1, @@ -165521,9 +156685,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "sort", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -165533,19 +156696,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(23), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13034, + Ctx: p13003, FreeVars: ast.Identifiers{ "arr", }, @@ -165558,19 +156721,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(28), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13034, + Ctx: p13003, FreeVars: ast.Identifiers{ "keyF", }, @@ -165593,19 +156756,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(35), }, End: ast.Location{ - Line: int(1216), + Line: int(1201), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13025, + Ctx: p12994, FreeVars: ast.Identifiers{ "keyF", }, @@ -165624,18 +156787,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1215), - Column: int(3), - }, - End: ast.Location{ - Line: int(1216), - Column: int(40), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -165660,7 +156811,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "setMember", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -165684,83 +156834,47 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1218), - Column: int(13), - }, - End: ast.Location{ - Line: int(1218), - Column: int(14), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1218), - Column: int(16), - }, - End: ast.Location{ - Line: int(1218), - Column: int(19), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1218), - Column: int(26), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "x", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1203), + Column: int(26), + }, + End: ast.Location{ + Line: int(1203), + Column: int(28), + }, + File: p1, }, - End: ast.Location{ - Line: int(1218), - Column: int(28), + Fodder: ast.Fodder{}, + Ctx: p13015, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p13045, - FreeVars: ast.Identifiers{ - "id", - }, - }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1218), - Column: int(21), - }, - End: ast.Location{ - Line: int(1218), - Column: int(28), + Id: "id", }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -165769,19 +156883,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(5), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13045, + Ctx: p13015, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -165792,19 +156906,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(5), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13045, + Ctx: p13015, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -165815,19 +156929,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(5), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(15), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13045, + Ctx: p13015, FreeVars: ast.Identifiers{ "std", }, @@ -165835,13 +156949,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(5), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(8), }, File: p1, @@ -165854,7 +156968,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -165891,9 +157005,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -165903,19 +157016,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(16), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13060, + Ctx: p13030, FreeVars: ast.Identifiers{ "arr", "keyF", @@ -165926,19 +157039,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(16), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13060, + Ctx: p13030, FreeVars: ast.Identifiers{ "std", }, @@ -165946,13 +157059,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(16), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(19), }, File: p1, @@ -165987,9 +157100,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "setInter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -165999,19 +157111,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(29), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13069, + Ctx: p13039, FreeVars: ast.Identifiers{ "x", }, @@ -166021,19 +157133,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(30), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13073, + Ctx: p13043, FreeVars: ast.Identifiers{ "x", }, @@ -166052,19 +157164,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(34), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13069, + Ctx: p13039, FreeVars: ast.Identifiers{ "arr", }, @@ -166077,19 +157189,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(39), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13069, + Ctx: p13039, FreeVars: ast.Identifiers{ "keyF", }, @@ -166121,38 +157233,27 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(48), }, End: ast.Location{ - Line: int(1220), + Line: int(1205), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13045, + Ctx: p13015, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1218), - Column: int(3), - }, - End: ast.Location{ - Line: int(1220), - Column: int(49), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -166177,7 +157278,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "setUnion", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -166201,83 +157301,47 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1222), - Column: int(12), - }, - End: ast.Location{ - Line: int(1222), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1222), - Column: int(15), - }, - End: ast.Location{ - Line: int(1222), - Column: int(16), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1222), - Column: int(23), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1207), + Column: int(23), + }, + End: ast.Location{ + Line: int(1207), + Column: int(25), + }, + File: p1, }, - End: ast.Location{ - Line: int(1222), - Column: int(25), + Fodder: ast.Fodder{}, + Ctx: p13056, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p13085, - FreeVars: ast.Identifiers{ - "id", }, + Id: "id", }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1222), - Column: int(18), - }, - End: ast.Location{ - Line: int(1222), - Column: int(25), - }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -166286,13 +157350,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1224), + Line: int(1209), Column: int(5), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(24), }, File: p1, @@ -166305,7 +157369,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -166313,7 +157377,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p13085, + Ctx: p13056, FreeVars: ast.Identifiers{ "a", "b", @@ -166329,19 +157393,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1224), + Line: int(1209), Column: int(11), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, }, Fodder: nil, - Ctx: p13093, + Ctx: p13064, FreeVars: ast.Identifiers{ "aux", "keyF", @@ -166349,115 +157413,48 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1224), - Column: int(15), - }, - End: ast.Location{ - Line: int(1224), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1224), - Column: int(18), - }, - End: ast.Location{ - Line: int(1224), - Column: int(19), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1224), - Column: int(21), - }, - End: ast.Location{ - Line: int(1224), - Column: int(22), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1224), - Column: int(24), - }, - End: ast.Location{ - Line: int(1224), - Column: int(25), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "acc", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1224), - Column: int(27), - }, - End: ast.Location{ - Line: int(1224), - Column: int(30), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "acc", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(7), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, @@ -166470,7 +157467,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -166485,19 +157482,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(10), }, End: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "i", @@ -166507,19 +157504,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(10), }, End: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "i", }, @@ -166531,19 +157528,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(15), }, End: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "std", @@ -166552,19 +157549,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(15), }, End: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "std", }, @@ -166572,13 +157569,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(15), }, End: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(18), }, File: p1, @@ -166613,9 +157610,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -166625,19 +157621,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(26), }, End: ast.Location{ - Line: int(1225), + Line: int(1210), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13113, + Ctx: p13084, FreeVars: ast.Identifiers{ "a", }, @@ -166659,19 +157655,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(9), }, End: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "acc", "b", @@ -166682,13 +157678,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(9), }, End: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(12), }, File: p1, @@ -166701,7 +157697,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "acc", }, @@ -166715,14 +157711,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1226), - Column: int(15), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1226), - Column: int(20), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -166796,7 +157792,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -166808,19 +157803,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(15), }, End: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "b", }, @@ -166833,19 +157828,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(17), }, End: ast.Location{ - Line: int(1226), + Line: int(1211), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "j", }, @@ -166918,19 +157913,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(12), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -166945,19 +157940,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(15), }, End: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "b", "j", @@ -166967,19 +157962,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(15), }, End: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "j", }, @@ -166991,19 +157986,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(20), }, End: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "b", "std", @@ -167012,19 +158007,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(20), }, End: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "std", }, @@ -167032,13 +158027,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(20), }, End: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(23), }, File: p1, @@ -167073,9 +158068,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -167085,19 +158079,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(31), }, End: ast.Location{ - Line: int(1227), + Line: int(1212), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13150, + Ctx: p13121, FreeVars: ast.Identifiers{ "b", }, @@ -167119,19 +158113,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(9), }, End: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -167142,13 +158136,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(9), }, End: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(12), }, File: p1, @@ -167161,7 +158155,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "acc", }, @@ -167175,14 +158169,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1228), - Column: int(15), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1228), - Column: int(20), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -167256,7 +158250,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -167268,19 +158261,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(15), }, End: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", }, @@ -167293,19 +158286,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(17), }, End: ast.Location{ - Line: int(1228), + Line: int(1213), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "i", }, @@ -167378,13 +158371,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(9), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, @@ -167397,7 +158390,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -167416,19 +158409,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(20), }, End: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13177, + Ctx: p13148, FreeVars: ast.Identifiers{ "a", "i", @@ -167438,19 +158431,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(20), }, End: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13177, + Ctx: p13148, FreeVars: ast.Identifiers{ "keyF", }, @@ -167464,19 +158457,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(25), }, End: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13183, + Ctx: p13154, FreeVars: ast.Identifiers{ "a", "i", @@ -167485,19 +158478,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(25), }, End: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13183, + Ctx: p13154, FreeVars: ast.Identifiers{ "a", }, @@ -167508,19 +158501,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(27), }, End: ast.Location{ - Line: int(1230), + Line: int(1215), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13183, + Ctx: p13154, FreeVars: ast.Identifiers{ "i", }, @@ -167542,30 +158535,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1230), - Column: int(15), - }, - End: ast.Location{ - Line: int(1230), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(9), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, @@ -167578,7 +158559,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -167598,19 +158579,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(20), }, End: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13194, + Ctx: p13165, FreeVars: ast.Identifiers{ "b", "j", @@ -167620,19 +158601,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(20), }, End: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13194, + Ctx: p13165, FreeVars: ast.Identifiers{ "keyF", }, @@ -167646,19 +158627,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(25), }, End: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13200, + Ctx: p13171, FreeVars: ast.Identifiers{ "b", "j", @@ -167667,19 +158648,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(25), }, End: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13200, + Ctx: p13171, FreeVars: ast.Identifiers{ "b", }, @@ -167690,19 +158671,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(27), }, End: ast.Location{ - Line: int(1231), + Line: int(1216), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13200, + Ctx: p13171, FreeVars: ast.Identifiers{ "j", }, @@ -167724,30 +158705,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1231), - Column: int(15), - }, - End: ast.Location{ - Line: int(1231), - Column: int(30), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1232), + Line: int(1217), Column: int(9), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, @@ -167760,7 +158729,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -167775,19 +158744,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1232), + Line: int(1217), Column: int(12), }, End: ast.Location{ - Line: int(1232), + Line: int(1217), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "ak", "bk", @@ -167796,19 +158765,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1232), + Line: int(1217), Column: int(12), }, End: ast.Location{ - Line: int(1232), + Line: int(1217), Column: int(14), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "ak", }, @@ -167820,19 +158789,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1232), + Line: int(1217), Column: int(18), }, End: ast.Location{ - Line: int(1232), + Line: int(1217), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "bk", }, @@ -167844,19 +158813,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(11), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -167869,13 +158838,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(11), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(14), }, File: p1, @@ -167888,7 +158857,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "aux", }, @@ -167902,19 +158871,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(15), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "a", }, @@ -167927,19 +158896,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(18), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "b", }, @@ -167952,19 +158921,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(21), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "i", }, @@ -167972,19 +158941,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(21), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "i", }, @@ -167996,21 +158965,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(25), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -168020,19 +158990,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(28), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "j", }, @@ -168040,19 +159010,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(28), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "j", }, @@ -168064,21 +159034,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(32), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -168088,19 +159059,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(35), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "a", "acc", @@ -168110,19 +159081,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(35), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "acc", }, @@ -168134,19 +159105,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(41), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13222, + Ctx: p13193, FreeVars: ast.Identifiers{ "a", "i", @@ -168157,19 +159128,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(42), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13244, + Ctx: p13215, FreeVars: ast.Identifiers{ "a", "i", @@ -168178,19 +159149,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(42), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13244, + Ctx: p13215, FreeVars: ast.Identifiers{ "a", }, @@ -168201,19 +159172,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(44), }, End: ast.Location{ - Line: int(1233), + Line: int(1218), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13244, + Ctx: p13215, FreeVars: ast.Identifiers{ "i", }, @@ -168251,19 +159222,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1234), + Line: int(1219), Column: int(14), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -168278,19 +159249,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1234), + Line: int(1219), Column: int(17), }, End: ast.Location{ - Line: int(1234), + Line: int(1219), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "ak", "bk", @@ -168299,19 +159270,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1234), + Line: int(1219), Column: int(17), }, End: ast.Location{ - Line: int(1234), + Line: int(1219), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "ak", }, @@ -168323,19 +159294,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1234), + Line: int(1219), Column: int(22), }, End: ast.Location{ - Line: int(1234), + Line: int(1219), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "bk", }, @@ -168347,19 +159318,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(11), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -168372,13 +159343,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(11), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(14), }, File: p1, @@ -168391,7 +159362,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "aux", }, @@ -168405,19 +159376,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(15), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "a", }, @@ -168430,19 +159401,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(18), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "b", }, @@ -168455,19 +159426,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(21), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "i", }, @@ -168475,19 +159446,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(21), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "i", }, @@ -168499,21 +159470,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(25), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -168523,19 +159495,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(28), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "j", }, @@ -168548,19 +159520,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(31), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "a", "acc", @@ -168570,19 +159542,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(31), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "acc", }, @@ -168594,19 +159566,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(37), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13266, + Ctx: p13237, FreeVars: ast.Identifiers{ "a", "i", @@ -168617,19 +159589,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(38), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13285, + Ctx: p13256, FreeVars: ast.Identifiers{ "a", "i", @@ -168638,19 +159610,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(38), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13285, + Ctx: p13256, FreeVars: ast.Identifiers{ "a", }, @@ -168661,19 +159633,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(40), }, End: ast.Location{ - Line: int(1235), + Line: int(1220), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13285, + Ctx: p13256, FreeVars: ast.Identifiers{ "i", }, @@ -168711,19 +159683,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(11), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "a", "acc", @@ -168736,13 +159708,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(11), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(14), }, File: p1, @@ -168755,7 +159727,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13098, + Ctx: p13069, FreeVars: ast.Identifiers{ "aux", }, @@ -168769,19 +159741,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(15), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "a", }, @@ -168794,19 +159766,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(18), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "b", }, @@ -168819,19 +159791,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(21), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "i", }, @@ -168844,19 +159816,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(24), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "j", }, @@ -168864,19 +159836,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(24), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "j", }, @@ -168888,21 +159860,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(28), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -168912,19 +159885,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(31), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "acc", "b", @@ -168934,19 +159907,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(31), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "acc", }, @@ -168958,19 +159931,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(37), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13299, + Ctx: p13270, FreeVars: ast.Identifiers{ "b", "j", @@ -168981,19 +159954,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(38), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13318, + Ctx: p13289, FreeVars: ast.Identifiers{ "b", "j", @@ -169002,19 +159975,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(38), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13318, + Ctx: p13289, FreeVars: ast.Identifiers{ "b", }, @@ -169025,19 +159998,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(40), }, End: ast.Location{ - Line: int(1237), + Line: int(1222), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13318, + Ctx: p13289, FreeVars: ast.Identifiers{ "j", }, @@ -169073,36 +160046,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(5), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13085, + Ctx: p13056, FreeVars: ast.Identifiers{ "a", "aux", @@ -169112,13 +160073,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(5), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(8), }, File: p1, @@ -169131,7 +160092,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13085, + Ctx: p13056, FreeVars: ast.Identifiers{ "aux", }, @@ -169145,19 +160106,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(9), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(10), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13331, + Ctx: p13302, FreeVars: ast.Identifiers{ "a", }, @@ -169170,19 +160131,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(12), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13331, + Ctx: p13302, FreeVars: ast.Identifiers{ "b", }, @@ -169195,21 +160156,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(15), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13331, + Ctx: p13302, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -169218,21 +160180,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(18), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13331, + Ctx: p13302, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -169241,19 +160204,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(21), }, End: ast.Location{ - Line: int(1238), + Line: int(1223), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13331, + Ctx: p13302, FreeVars: nil, }, Elements: nil, @@ -169273,18 +160236,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1222), - Column: int(3), - }, - End: ast.Location{ - Line: int(1238), - Column: int(24), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -169309,7 +160260,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "setInter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -169333,83 +160283,47 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1240), - Column: int(12), - }, - End: ast.Location{ - Line: int(1240), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1240), - Column: int(15), - }, - End: ast.Location{ - Line: int(1240), - Column: int(16), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1240), - Column: int(23), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1225), + Column: int(23), + }, + End: ast.Location{ + Line: int(1225), + Column: int(25), + }, + File: p1, }, - End: ast.Location{ - Line: int(1240), - Column: int(25), + Fodder: ast.Fodder{}, + Ctx: p13315, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p13343, - FreeVars: ast.Identifiers{ - "id", - }, - }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1240), - Column: int(18), + Id: "id", }, - End: ast.Location{ - Line: int(1240), - Column: int(25), - }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -169418,13 +160332,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1241), + Line: int(1226), Column: int(5), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(24), }, File: p1, @@ -169437,7 +160351,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13343, + Ctx: p13315, FreeVars: ast.Identifiers{ "a", "b", @@ -169453,19 +160367,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1241), + Line: int(1226), Column: int(11), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(35), }, File: p1, }, Fodder: nil, - Ctx: p13350, + Ctx: p13322, FreeVars: ast.Identifiers{ "aux", "keyF", @@ -169473,115 +160387,48 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1241), - Column: int(15), - }, - End: ast.Location{ - Line: int(1241), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1241), - Column: int(18), - }, - End: ast.Location{ - Line: int(1241), - Column: int(19), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1241), - Column: int(21), - }, - End: ast.Location{ - Line: int(1241), - Column: int(22), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1241), - Column: int(24), - }, - End: ast.Location{ - Line: int(1241), - Column: int(25), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "acc", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1241), - Column: int(27), - }, - End: ast.Location{ - Line: int(1241), - Column: int(30), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "acc", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(7), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(35), }, File: p1, @@ -169594,7 +160441,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "acc", @@ -169609,19 +160456,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(10), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "b", @@ -169633,19 +160480,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(10), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "i", @@ -169655,19 +160502,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(10), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "i", }, @@ -169679,19 +160526,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(15), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "std", @@ -169700,19 +160547,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(15), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "std", }, @@ -169720,13 +160567,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(15), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(18), }, File: p1, @@ -169761,9 +160608,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -169773,19 +160619,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(26), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13372, + Ctx: p13344, FreeVars: ast.Identifiers{ "a", }, @@ -169808,19 +160654,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(32), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "b", "j", @@ -169830,19 +160676,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(32), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "j", }, @@ -169854,19 +160700,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(37), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "b", "std", @@ -169875,19 +160721,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(37), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "std", }, @@ -169895,13 +160741,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(37), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(40), }, File: p1, @@ -169936,9 +160782,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -169948,19 +160793,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(48), }, End: ast.Location{ - Line: int(1242), + Line: int(1227), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13387, + Ctx: p13359, FreeVars: ast.Identifiers{ "b", }, @@ -169983,13 +160828,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1243), + Line: int(1228), Column: int(9), }, End: ast.Location{ - Line: int(1243), + Line: int(1228), Column: int(12), }, File: p1, @@ -170002,7 +160847,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "acc", }, @@ -170020,13 +160865,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(9), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(35), }, File: p1, @@ -170039,7 +160884,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "acc", @@ -170053,19 +160898,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(12), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "b", @@ -170077,19 +160922,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(12), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "i", @@ -170099,19 +160944,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(12), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "keyF", }, @@ -170125,19 +160970,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(17), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13404, + Ctx: p13376, FreeVars: ast.Identifiers{ "a", "i", @@ -170146,19 +160991,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(17), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13404, + Ctx: p13376, FreeVars: ast.Identifiers{ "a", }, @@ -170169,19 +161014,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(19), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13404, + Ctx: p13376, FreeVars: ast.Identifiers{ "i", }, @@ -170206,19 +161051,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(26), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "b", "j", @@ -170228,19 +161073,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(26), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "keyF", }, @@ -170254,19 +161099,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(31), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13416, + Ctx: p13388, FreeVars: ast.Identifiers{ "b", "j", @@ -170275,19 +161120,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(31), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13416, + Ctx: p13388, FreeVars: ast.Identifiers{ "b", }, @@ -170298,19 +161143,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(33), }, End: ast.Location{ - Line: int(1245), + Line: int(1230), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13416, + Ctx: p13388, FreeVars: ast.Identifiers{ "j", }, @@ -170335,19 +161180,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(11), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "acc", @@ -170360,13 +161205,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(11), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(14), }, File: p1, @@ -170379,7 +161224,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "aux", }, @@ -170393,19 +161238,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(15), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "a", }, @@ -170418,19 +161263,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(18), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "b", }, @@ -170443,19 +161288,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(21), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "i", }, @@ -170463,19 +161308,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(21), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "i", }, @@ -170487,21 +161332,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(25), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -170511,19 +161357,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(28), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "j", }, @@ -170531,19 +161377,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(28), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "j", }, @@ -170555,21 +161401,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(32), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -170579,19 +161426,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(35), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "a", "acc", @@ -170601,19 +161448,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(35), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "acc", }, @@ -170625,19 +161472,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(41), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13429, + Ctx: p13401, FreeVars: ast.Identifiers{ "a", "i", @@ -170648,19 +161495,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(42), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13451, + Ctx: p13423, FreeVars: ast.Identifiers{ "a", "i", @@ -170669,19 +161516,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(42), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13451, + Ctx: p13423, FreeVars: ast.Identifiers{ "a", }, @@ -170692,19 +161539,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(44), }, End: ast.Location{ - Line: int(1246), + Line: int(1231), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13451, + Ctx: p13423, FreeVars: ast.Identifiers{ "i", }, @@ -170742,19 +161589,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(14), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "acc", @@ -170768,19 +161615,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(17), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "b", @@ -170792,19 +161639,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(17), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "i", @@ -170814,19 +161661,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(17), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "keyF", }, @@ -170840,19 +161687,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(22), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13468, + Ctx: p13440, FreeVars: ast.Identifiers{ "a", "i", @@ -170861,19 +161708,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(22), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13468, + Ctx: p13440, FreeVars: ast.Identifiers{ "a", }, @@ -170884,19 +161731,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(24), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13468, + Ctx: p13440, FreeVars: ast.Identifiers{ "i", }, @@ -170921,19 +161768,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(30), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "b", "j", @@ -170943,19 +161790,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(30), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "keyF", }, @@ -170969,19 +161816,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(35), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13480, + Ctx: p13452, FreeVars: ast.Identifiers{ "b", "j", @@ -170990,19 +161837,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(35), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13480, + Ctx: p13452, FreeVars: ast.Identifiers{ "b", }, @@ -171013,19 +161860,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(37), }, End: ast.Location{ - Line: int(1247), + Line: int(1232), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13480, + Ctx: p13452, FreeVars: ast.Identifiers{ "j", }, @@ -171050,19 +161897,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(11), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "acc", @@ -171075,13 +161922,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(11), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(14), }, File: p1, @@ -171094,7 +161941,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "aux", }, @@ -171108,19 +161955,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(15), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13493, + Ctx: p13465, FreeVars: ast.Identifiers{ "a", }, @@ -171133,19 +161980,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(18), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13493, + Ctx: p13465, FreeVars: ast.Identifiers{ "b", }, @@ -171158,19 +162005,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(21), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13493, + Ctx: p13465, FreeVars: ast.Identifiers{ "i", }, @@ -171178,19 +162025,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(21), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13493, + Ctx: p13465, FreeVars: ast.Identifiers{ "i", }, @@ -171202,21 +162049,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(25), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13493, + Ctx: p13465, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -171226,19 +162074,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(28), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13493, + Ctx: p13465, FreeVars: ast.Identifiers{ "j", }, @@ -171251,19 +162099,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(31), }, End: ast.Location{ - Line: int(1248), + Line: int(1233), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13493, + Ctx: p13465, FreeVars: ast.Identifiers{ "acc", }, @@ -171291,19 +162139,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(11), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "a", "acc", @@ -171316,13 +162164,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(11), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(14), }, File: p1, @@ -171335,7 +162183,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13355, + Ctx: p13327, FreeVars: ast.Identifiers{ "aux", }, @@ -171349,19 +162197,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(15), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13514, + Ctx: p13486, FreeVars: ast.Identifiers{ "a", }, @@ -171374,19 +162222,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(18), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13514, + Ctx: p13486, FreeVars: ast.Identifiers{ "b", }, @@ -171399,19 +162247,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(21), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13514, + Ctx: p13486, FreeVars: ast.Identifiers{ "i", }, @@ -171424,19 +162272,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(24), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13514, + Ctx: p13486, FreeVars: ast.Identifiers{ "j", }, @@ -171444,19 +162292,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(24), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13514, + Ctx: p13486, FreeVars: ast.Identifiers{ "j", }, @@ -171468,21 +162316,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(28), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13514, + Ctx: p13486, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -171492,19 +162341,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(31), }, End: ast.Location{ - Line: int(1250), + Line: int(1235), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13514, + Ctx: p13486, FreeVars: ast.Identifiers{ "acc", }, @@ -171527,36 +162376,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(5), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13343, + Ctx: p13315, FreeVars: ast.Identifiers{ "a", "aux", @@ -171566,13 +162403,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(5), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(8), }, File: p1, @@ -171585,7 +162422,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13343, + Ctx: p13315, FreeVars: ast.Identifiers{ "aux", }, @@ -171599,19 +162436,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(9), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(10), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13534, + Ctx: p13506, FreeVars: ast.Identifiers{ "a", }, @@ -171624,19 +162461,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(12), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13534, + Ctx: p13506, FreeVars: ast.Identifiers{ "b", }, @@ -171649,21 +162486,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(15), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13534, + Ctx: p13506, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -171672,21 +162510,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(18), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13534, + Ctx: p13506, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -171695,19 +162534,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(21), }, End: ast.Location{ - Line: int(1251), + Line: int(1236), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13534, + Ctx: p13506, FreeVars: nil, }, Elements: nil, @@ -171727,18 +162566,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1240), - Column: int(3), - }, - End: ast.Location{ - Line: int(1251), - Column: int(24), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -171763,7 +162590,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "setDiff", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -171787,83 +162613,47 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1253), - Column: int(11), - }, - End: ast.Location{ - Line: int(1253), - Column: int(12), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1253), - Column: int(14), - }, - End: ast.Location{ - Line: int(1253), - Column: int(15), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "keyF", - EqFodder: ast.Fodder{}, - DefaultArg: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1253), - Column: int(22), + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, + }, + }, + Optional: []ast.NamedParameter{ + ast.NamedParameter{ + NameFodder: ast.Fodder{}, + Name: "keyF", + EqFodder: ast.Fodder{}, + DefaultArg: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1238), + Column: int(22), + }, + End: ast.Location{ + Line: int(1238), + Column: int(24), + }, + File: p1, }, - End: ast.Location{ - Line: int(1253), - Column: int(24), + Fodder: ast.Fodder{}, + Ctx: p13519, + FreeVars: ast.Identifiers{ + "id", }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p13546, - FreeVars: ast.Identifiers{ - "id", }, + Id: "id", }, - Id: "id", - }, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1253), - Column: int(17), - }, - End: ast.Location{ - Line: int(1253), - Column: int(24), - }, - File: p1, + CommaFodder: ast.Fodder{}, }, }, }, @@ -171872,13 +162662,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1254), + Line: int(1239), Column: int(5), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(24), }, File: p1, @@ -171891,7 +162681,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13546, + Ctx: p13519, FreeVars: ast.Identifiers{ "a", "b", @@ -171907,19 +162697,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1254), + Line: int(1239), Column: int(11), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(35), }, File: p1, }, Fodder: nil, - Ctx: p13553, + Ctx: p13526, FreeVars: ast.Identifiers{ "aux", "keyF", @@ -171927,115 +162717,48 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1254), - Column: int(15), - }, - End: ast.Location{ - Line: int(1254), - Column: int(16), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1254), - Column: int(18), - }, - End: ast.Location{ - Line: int(1254), - Column: int(19), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1254), - Column: int(21), - }, - End: ast.Location{ - Line: int(1254), - Column: int(22), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "j", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1254), - Column: int(24), - }, - End: ast.Location{ - Line: int(1254), - Column: int(25), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "j", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "acc", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1254), - Column: int(27), - }, - End: ast.Location{ - Line: int(1254), - Column: int(30), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "acc", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(7), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(35), }, File: p1, @@ -172048,7 +162771,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -172063,19 +162786,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(10), }, End: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "i", @@ -172085,19 +162808,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(10), }, End: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "i", }, @@ -172109,19 +162832,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(15), }, End: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "std", @@ -172130,19 +162853,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(15), }, End: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "std", }, @@ -172150,13 +162873,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(15), }, End: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(18), }, File: p1, @@ -172191,9 +162914,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -172203,19 +162925,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(26), }, End: ast.Location{ - Line: int(1255), + Line: int(1240), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13573, + Ctx: p13546, FreeVars: ast.Identifiers{ "a", }, @@ -172237,13 +162959,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1256), + Line: int(1241), Column: int(9), }, End: ast.Location{ - Line: int(1256), + Line: int(1241), Column: int(12), }, File: p1, @@ -172256,7 +162978,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "acc", }, @@ -172274,19 +162996,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(12), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -172301,19 +163023,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(15), }, End: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "b", "j", @@ -172323,19 +163045,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(15), }, End: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "j", }, @@ -172347,19 +163069,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(20), }, End: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "b", "std", @@ -172368,19 +163090,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(20), }, End: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "std", }, @@ -172388,13 +163110,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(20), }, End: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(23), }, File: p1, @@ -172429,9 +163151,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -172441,19 +163162,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(31), }, End: ast.Location{ - Line: int(1257), + Line: int(1242), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13594, + Ctx: p13567, FreeVars: ast.Identifiers{ "b", }, @@ -172475,19 +163196,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(9), }, End: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -172498,13 +163219,13 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(9), }, End: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(12), }, File: p1, @@ -172517,7 +163238,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "acc", }, @@ -172531,14 +163252,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1258), - Column: int(15), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1258), - Column: int(20), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -172612,7 +163333,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -172624,19 +163344,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(15), }, End: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", }, @@ -172649,19 +163369,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(17), }, End: ast.Location{ - Line: int(1258), + Line: int(1243), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "i", }, @@ -172734,13 +163454,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(9), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(35), }, File: p1, @@ -172753,7 +163473,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -172767,19 +163487,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(12), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "b", @@ -172791,19 +163511,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(12), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "i", @@ -172813,19 +163533,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(12), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "keyF", }, @@ -172839,19 +163559,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(17), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13627, + Ctx: p13600, FreeVars: ast.Identifiers{ "a", "i", @@ -172860,19 +163580,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(17), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13627, + Ctx: p13600, FreeVars: ast.Identifiers{ "a", }, @@ -172883,19 +163603,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(19), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13627, + Ctx: p13600, FreeVars: ast.Identifiers{ "i", }, @@ -172920,19 +163640,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(26), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "b", "j", @@ -172942,19 +163662,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(26), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "keyF", }, @@ -172968,19 +163688,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(31), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13639, + Ctx: p13612, FreeVars: ast.Identifiers{ "b", "j", @@ -172989,19 +163709,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(31), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13639, + Ctx: p13612, FreeVars: ast.Identifiers{ "b", }, @@ -173012,19 +163732,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(33), }, End: ast.Location{ - Line: int(1260), + Line: int(1245), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13639, + Ctx: p13612, FreeVars: ast.Identifiers{ "j", }, @@ -173049,19 +163769,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(11), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -173074,13 +163794,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(11), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(14), }, File: p1, @@ -173093,7 +163813,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "aux", }, @@ -173107,19 +163827,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(15), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: ast.Identifiers{ "a", }, @@ -173132,19 +163852,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(18), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: ast.Identifiers{ "b", }, @@ -173157,19 +163877,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(21), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: ast.Identifiers{ "i", }, @@ -173177,19 +163897,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(21), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: ast.Identifiers{ "i", }, @@ -173201,21 +163921,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(25), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -173225,19 +163946,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(28), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: ast.Identifiers{ "j", }, @@ -173245,19 +163966,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(28), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: ast.Identifiers{ "j", }, @@ -173269,21 +163990,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(32), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -173293,19 +164015,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(35), }, End: ast.Location{ - Line: int(1261), + Line: int(1246), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13652, + Ctx: p13625, FreeVars: ast.Identifiers{ "acc", }, @@ -173333,19 +164055,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(14), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -173359,19 +164081,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(17), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "b", @@ -173383,19 +164105,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(17), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "i", @@ -173405,19 +164127,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(17), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "keyF", }, @@ -173431,19 +164153,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(22), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13679, + Ctx: p13652, FreeVars: ast.Identifiers{ "a", "i", @@ -173452,19 +164174,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(22), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13679, + Ctx: p13652, FreeVars: ast.Identifiers{ "a", }, @@ -173475,19 +164197,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(24), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13679, + Ctx: p13652, FreeVars: ast.Identifiers{ "i", }, @@ -173512,19 +164234,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(30), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "b", "j", @@ -173534,19 +164256,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(30), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "keyF", }, @@ -173560,19 +164282,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(35), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13691, + Ctx: p13664, FreeVars: ast.Identifiers{ "b", "j", @@ -173581,19 +164303,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(35), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13691, + Ctx: p13664, FreeVars: ast.Identifiers{ "b", }, @@ -173604,19 +164326,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(37), }, End: ast.Location{ - Line: int(1262), + Line: int(1247), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13691, + Ctx: p13664, FreeVars: ast.Identifiers{ "j", }, @@ -173641,19 +164363,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(11), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -173666,13 +164388,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(11), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(14), }, File: p1, @@ -173685,7 +164407,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "aux", }, @@ -173699,19 +164421,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(15), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "a", }, @@ -173724,19 +164446,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(18), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "b", }, @@ -173749,19 +164471,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(21), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "i", }, @@ -173769,19 +164491,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(21), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "i", }, @@ -173793,21 +164515,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(25), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -173817,19 +164540,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(28), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "j", }, @@ -173842,19 +164565,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(31), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "a", "acc", @@ -173864,19 +164587,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(31), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "acc", }, @@ -173888,19 +164611,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(37), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13704, + Ctx: p13677, FreeVars: ast.Identifiers{ "a", "i", @@ -173911,19 +164634,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(38), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13723, + Ctx: p13696, FreeVars: ast.Identifiers{ "a", "i", @@ -173932,19 +164655,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(38), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13723, + Ctx: p13696, FreeVars: ast.Identifiers{ "a", }, @@ -173955,19 +164678,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(40), }, End: ast.Location{ - Line: int(1263), + Line: int(1248), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13723, + Ctx: p13696, FreeVars: ast.Identifiers{ "i", }, @@ -174005,19 +164728,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(11), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "a", "acc", @@ -174030,13 +164753,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(11), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(14), }, File: p1, @@ -174049,7 +164772,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13558, + Ctx: p13531, FreeVars: ast.Identifiers{ "aux", }, @@ -174063,19 +164786,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(15), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13737, + Ctx: p13710, FreeVars: ast.Identifiers{ "a", }, @@ -174088,19 +164811,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(18), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13737, + Ctx: p13710, FreeVars: ast.Identifiers{ "b", }, @@ -174113,19 +164836,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(21), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13737, + Ctx: p13710, FreeVars: ast.Identifiers{ "i", }, @@ -174138,19 +164861,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(24), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13737, + Ctx: p13710, FreeVars: ast.Identifiers{ "j", }, @@ -174158,19 +164881,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(24), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13737, + Ctx: p13710, FreeVars: ast.Identifiers{ "j", }, @@ -174182,21 +164905,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(28), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13737, + Ctx: p13710, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -174206,19 +164930,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(31), }, End: ast.Location{ - Line: int(1265), + Line: int(1250), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13737, + Ctx: p13710, FreeVars: ast.Identifiers{ "acc", }, @@ -174242,36 +164966,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(5), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13546, + Ctx: p13519, FreeVars: ast.Identifiers{ "a", "aux", @@ -174281,13 +164993,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(5), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(8), }, File: p1, @@ -174300,7 +165012,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13546, + Ctx: p13519, FreeVars: ast.Identifiers{ "aux", }, @@ -174314,19 +165026,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(9), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(10), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13757, + Ctx: p13730, FreeVars: ast.Identifiers{ "a", }, @@ -174339,19 +165051,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(12), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13757, + Ctx: p13730, FreeVars: ast.Identifiers{ "b", }, @@ -174364,21 +165076,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(15), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13757, + Ctx: p13730, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -174387,21 +165100,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(18), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13757, + Ctx: p13730, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -174410,19 +165124,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(21), }, End: ast.Location{ - Line: int(1266), + Line: int(1251), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13757, + Ctx: p13730, FreeVars: nil, }, Elements: nil, @@ -174442,18 +165156,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1253), - Column: int(3), - }, - End: ast.Location{ - Line: int(1266), - Column: int(24), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -174478,7 +165180,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mergePatch", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -174501,58 +165202,33 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "target", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1268), - Column: int(14), - }, - End: ast.Location{ - Line: int(1268), - Column: int(20), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "patch", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1268), - Column: int(22), - }, - End: ast.Location{ - Line: int(1268), - Column: int(27), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "target", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "patch", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(5), }, End: ast.Location{ - Line: int(1290), + Line: int(1275), Column: int(12), }, File: p1, @@ -174565,7 +165241,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "patch", "std", @@ -174575,19 +165251,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(8), }, End: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "patch", "std", @@ -174596,19 +165272,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(8), }, End: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "std", }, @@ -174616,13 +165292,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(8), }, End: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(11), }, File: p1, @@ -174657,9 +165333,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -174669,19 +165344,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(21), }, End: ast.Location{ - Line: int(1269), + Line: int(1254), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13781, + Ctx: p13754, FreeVars: ast.Identifiers{ "patch", }, @@ -174702,13 +165377,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1270), + Line: int(1255), Column: int(7), }, End: ast.Location{ - Line: int(1288), + Line: int(1273), Column: int(8), }, File: p1, @@ -174721,7 +165396,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "patch", "std", @@ -174736,13 +165411,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(9), }, End: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(52), }, File: p1, @@ -174755,7 +165430,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13789, + Ctx: p13762, FreeVars: ast.Identifiers{ "std", "target", @@ -174764,19 +165439,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(12), }, End: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13789, + Ctx: p13762, FreeVars: ast.Identifiers{ "std", "target", @@ -174785,19 +165460,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(12), }, End: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13789, + Ctx: p13762, FreeVars: ast.Identifiers{ "std", }, @@ -174805,13 +165480,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(12), }, End: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(15), }, File: p1, @@ -174846,9 +165521,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -174858,19 +165532,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(25), }, End: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13800, + Ctx: p13773, FreeVars: ast.Identifiers{ "target", }, @@ -174891,19 +165565,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(38), }, End: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13789, + Ctx: p13762, FreeVars: ast.Identifiers{ "target", }, @@ -174914,19 +165588,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(50), }, End: ast.Location{ - Line: int(1271), + Line: int(1256), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13789, + Ctx: p13762, FreeVars: nil, }, Asserts: ast.Nodes{}, @@ -174936,30 +165610,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1270), - Column: int(13), - }, - End: ast.Location{ - Line: int(1271), - Column: int(52), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1273), + Line: int(1258), Column: int(7), }, End: ast.Location{ - Line: int(1288), + Line: int(1273), Column: int(8), }, File: p1, @@ -174972,7 +165634,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "patch", "std", @@ -174987,13 +165649,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(9), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(84), }, File: p1, @@ -175006,7 +165668,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13811, + Ctx: p13784, FreeVars: ast.Identifiers{ "std", "target_object", @@ -175015,19 +165677,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(12), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13811, + Ctx: p13784, FreeVars: ast.Identifiers{ "std", "target_object", @@ -175036,19 +165698,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(12), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13811, + Ctx: p13784, FreeVars: ast.Identifiers{ "std", }, @@ -175056,13 +165718,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(12), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(15), }, File: p1, @@ -175097,9 +165759,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -175109,19 +165770,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(25), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13822, + Ctx: p13795, FreeVars: ast.Identifiers{ "target_object", }, @@ -175142,19 +165803,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(45), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(76), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13811, + Ctx: p13784, FreeVars: ast.Identifiers{ "std", "target_object", @@ -175163,19 +165824,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(45), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13811, + Ctx: p13784, FreeVars: ast.Identifiers{ "std", }, @@ -175183,13 +165844,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(45), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(48), }, File: p1, @@ -175224,9 +165885,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -175236,19 +165896,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(62), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(75), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13833, + Ctx: p13806, FreeVars: ast.Identifiers{ "target_object", }, @@ -175269,19 +165929,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(82), }, End: ast.Location{ - Line: int(1274), + Line: int(1259), Column: int(84), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13811, + Ctx: p13784, FreeVars: nil, }, Elements: nil, @@ -175291,30 +165951,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1273), - Column: int(13), - }, - End: ast.Location{ - Line: int(1274), - Column: int(84), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(7), }, End: ast.Location{ - Line: int(1288), + Line: int(1273), Column: int(8), }, File: p1, @@ -175327,7 +165975,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "patch", "std", @@ -175345,14 +165993,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1276), - Column: int(27), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1276), - Column: int(83), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -175425,7 +166073,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -175455,26 +166102,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -175502,19 +166138,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(66), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13854, + Ctx: p13827, FreeVars: ast.Identifiers{ "k", "patch", @@ -175523,19 +166159,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(66), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(74), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13854, + Ctx: p13827, FreeVars: ast.Identifiers{ "k", "patch", @@ -175544,19 +166180,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(66), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13854, + Ctx: p13827, FreeVars: ast.Identifiers{ "patch", }, @@ -175567,19 +166203,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(72), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13854, + Ctx: p13827, FreeVars: ast.Identifiers{ "k", }, @@ -175594,19 +166230,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(78), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13854, + Ctx: p13827, FreeVars: nil, }, }, @@ -175637,19 +166273,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(28), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13867, + Ctx: p13840, FreeVars: ast.Identifiers{ "k", }, @@ -175693,19 +166329,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(39), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(62), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13854, + Ctx: p13827, FreeVars: ast.Identifiers{ "patch", "std", @@ -175714,19 +166350,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(39), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13854, + Ctx: p13827, FreeVars: ast.Identifiers{ "std", }, @@ -175734,13 +166370,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(39), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(42), }, File: p1, @@ -175775,9 +166411,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -175787,19 +166422,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(56), }, End: ast.Location{ - Line: int(1276), + Line: int(1261), Column: int(61), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13879, + Ctx: p13852, FreeVars: ast.Identifiers{ "patch", }, @@ -175828,30 +166463,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1276), - Column: int(13), - }, - End: ast.Location{ - Line: int(1276), - Column: int(83), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(7), }, End: ast.Location{ - Line: int(1288), + Line: int(1273), Column: int(8), }, File: p1, @@ -175864,7 +166487,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "null_fields", "patch", @@ -175881,19 +166504,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(27), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13886, + Ctx: p13859, FreeVars: ast.Identifiers{ "patch", "std", @@ -175903,19 +166526,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(27), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13886, + Ctx: p13859, FreeVars: ast.Identifiers{ "std", }, @@ -175923,13 +166546,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(27), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(30), }, File: p1, @@ -175964,9 +166587,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "setUnion", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -175976,19 +166598,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(40), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13895, + Ctx: p13868, FreeVars: ast.Identifiers{ "target_fields", }, @@ -176001,19 +166623,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(55), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13895, + Ctx: p13868, FreeVars: ast.Identifiers{ "patch", "std", @@ -176022,19 +166644,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(55), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13895, + Ctx: p13868, FreeVars: ast.Identifiers{ "std", }, @@ -176042,13 +166664,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(55), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(58), }, File: p1, @@ -176083,9 +166705,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -176095,19 +166716,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(72), }, End: ast.Location{ - Line: int(1277), + Line: int(1262), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13906, + Ctx: p13879, FreeVars: ast.Identifiers{ "patch", }, @@ -176136,18 +166757,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1277), - Column: int(13), - }, - End: ast.Location{ - Line: int(1277), - Column: int(79), - }, - File: p1, - }, }, }, Body: &ast.Apply{ @@ -176155,14 +166764,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1279), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1288), - Column: int(8), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -176238,7 +166847,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "$objectFlatMerge", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -176252,14 +166860,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1279), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1288), - Column: int(8), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -176335,7 +166943,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -176367,26 +166974,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "k", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -176418,13 +167014,13 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1279), + Line: int(1264), Column: int(7), }, End: ast.Location{ - Line: int(1288), + Line: int(1273), Column: int(8), }, File: p1, @@ -176437,7 +167033,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "k", "patch", @@ -176452,19 +167048,19 @@ var _StdAst = &ast.DesugaredObject{ Name: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1280), + Line: int(1265), Column: int(10), }, End: ast.Location{ - Line: int(1280), + Line: int(1265), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "k", }, @@ -176474,13 +167070,13 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(11), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(55), }, File: p1, @@ -176493,7 +167089,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "patch", @@ -176504,19 +167100,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(14), }, End: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "patch", @@ -176527,19 +167123,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(15), }, End: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "patch", @@ -176549,19 +167145,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(15), }, End: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "std", }, @@ -176569,13 +167165,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(15), }, End: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(18), }, File: p1, @@ -176610,9 +167206,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHas", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -176622,19 +167217,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(29), }, End: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13951, + Ctx: p13924, FreeVars: ast.Identifiers{ "patch", }, @@ -176647,19 +167242,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(36), }, End: ast.Location{ - Line: int(1281), + Line: int(1266), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13951, + Ctx: p13924, FreeVars: ast.Identifiers{ "k", }, @@ -176681,19 +167276,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1282), + Line: int(1267), Column: int(13), }, End: ast.Location{ - Line: int(1282), + Line: int(1267), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "target_object", @@ -176702,13 +167297,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1282), + Line: int(1267), Column: int(13), }, End: ast.Location{ - Line: int(1282), + Line: int(1267), Column: int(26), }, File: p1, @@ -176721,7 +167316,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "target_object", }, @@ -176732,19 +167327,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1282), + Line: int(1267), Column: int(27), }, End: ast.Location{ - Line: int(1282), + Line: int(1267), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", }, @@ -176765,19 +167360,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(16), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "patch", @@ -176788,19 +167383,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(19), }, End: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "std", @@ -176811,19 +167406,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(20), }, End: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "std", @@ -176833,19 +167428,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(20), }, End: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "std", }, @@ -176853,13 +167448,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(20), }, End: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(23), }, File: p1, @@ -176894,9 +167489,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHas", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -176906,19 +167500,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(34), }, End: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13976, + Ctx: p13949, FreeVars: ast.Identifiers{ "target_object", }, @@ -176931,19 +167525,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(49), }, End: ast.Location{ - Line: int(1283), + Line: int(1268), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13976, + Ctx: p13949, FreeVars: ast.Identifiers{ "k", }, @@ -176965,19 +167559,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(13), }, End: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "patch", @@ -176987,19 +167581,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(13), }, End: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "std", }, @@ -177007,13 +167601,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(13), }, End: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(16), }, File: p1, @@ -177055,9 +167649,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "mergePatch", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -177067,19 +167660,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(28), }, End: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13990, + Ctx: p13963, FreeVars: nil, }, }, @@ -177089,19 +167682,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(34), }, End: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13990, + Ctx: p13963, FreeVars: ast.Identifiers{ "k", "patch", @@ -177110,19 +167703,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(34), }, End: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13990, + Ctx: p13963, FreeVars: ast.Identifiers{ "patch", }, @@ -177133,19 +167726,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(40), }, End: ast.Location{ - Line: int(1284), + Line: int(1269), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13990, + Ctx: p13963, FreeVars: ast.Identifiers{ "k", }, @@ -177176,19 +167769,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(13), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "k", "patch", @@ -177199,19 +167792,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(13), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13938, + Ctx: p13911, FreeVars: ast.Identifiers{ "std", }, @@ -177219,13 +167812,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(13), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(16), }, File: p1, @@ -177267,9 +167860,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "mergePatch", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -177279,19 +167871,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(28), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14008, + Ctx: p13981, FreeVars: ast.Identifiers{ "k", "target_object", @@ -177300,19 +167892,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(28), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14008, + Ctx: p13981, FreeVars: ast.Identifiers{ "target_object", }, @@ -177323,19 +167915,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(42), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14008, + Ctx: p13981, FreeVars: ast.Identifiers{ "k", }, @@ -177351,19 +167943,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(46), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14008, + Ctx: p13981, FreeVars: ast.Identifiers{ "k", "patch", @@ -177372,19 +167964,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(46), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14008, + Ctx: p13981, FreeVars: ast.Identifiers{ "patch", }, @@ -177395,19 +167987,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(52), }, End: ast.Location{ - Line: int(1286), + Line: int(1271), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14008, + Ctx: p13981, FreeVars: ast.Identifiers{ "k", }, @@ -177430,18 +168022,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1280), - Column: int(9), - }, - End: ast.Location{ - Line: int(1286), - Column: int(55), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -177459,19 +168039,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(18), }, End: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "both_fields", "null_fields", @@ -177481,19 +168061,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(18), }, End: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "std", }, @@ -177501,13 +168081,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(18), }, End: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(21), }, File: p1, @@ -177542,9 +168122,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "setDiff", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -177554,19 +168133,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(30), }, End: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14029, + Ctx: p14002, FreeVars: ast.Identifiers{ "both_fields", }, @@ -177579,19 +168158,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(43), }, End: ast.Location{ - Line: int(1287), + Line: int(1272), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14029, + Ctx: p14002, FreeVars: ast.Identifiers{ "null_fields", }, @@ -177643,13 +168222,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1290), + Line: int(1275), Column: int(7), }, End: ast.Location{ - Line: int(1290), + Line: int(1275), Column: int(12), }, File: p1, @@ -177662,7 +168241,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p13770, + Ctx: p13743, FreeVars: ast.Identifiers{ "patch", }, @@ -177672,18 +168251,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1268), - Column: int(3), - }, - End: ast.Location{ - Line: int(1290), - Column: int(12), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -177708,7 +168275,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -177731,45 +168297,34 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "o", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1292), - Column: int(16), - }, - End: ast.Location{ - Line: int(1292), - Column: int(17), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "o", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(5), }, End: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14042, + Ctx: p14015, FreeVars: ast.Identifiers{ "o", "std", @@ -177778,19 +168333,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(5), }, End: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14042, + Ctx: p14015, FreeVars: ast.Identifiers{ "std", }, @@ -177798,13 +168353,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(5), }, End: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(8), }, File: p1, @@ -177846,9 +168401,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFieldsEx", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -177858,19 +168412,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(24), }, End: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14052, + Ctx: p14025, FreeVars: ast.Identifiers{ "o", }, @@ -177883,19 +168437,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(27), }, End: ast.Location{ - Line: int(1293), + Line: int(1278), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14052, + Ctx: p14025, FreeVars: nil, }, Value: false, @@ -177912,18 +168466,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1292), - Column: int(3), - }, - End: ast.Location{ - Line: int(1293), - Column: int(33), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -177948,7 +168490,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFieldsAll", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -177971,45 +168512,34 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "o", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1295), - Column: int(19), - }, - End: ast.Location{ - Line: int(1295), - Column: int(20), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "o", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(5), }, End: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14060, + Ctx: p14033, FreeVars: ast.Identifiers{ "o", "std", @@ -178018,19 +168548,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(5), }, End: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14060, + Ctx: p14033, FreeVars: ast.Identifiers{ "std", }, @@ -178038,13 +168568,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(5), }, End: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(8), }, File: p1, @@ -178086,9 +168616,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFieldsEx", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -178098,19 +168627,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(24), }, End: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14070, + Ctx: p14043, FreeVars: ast.Identifiers{ "o", }, @@ -178123,19 +168652,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(27), }, End: ast.Location{ - Line: int(1296), + Line: int(1281), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14070, + Ctx: p14043, FreeVars: nil, }, Value: true, @@ -178152,18 +168681,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1295), - Column: int(3), - }, - End: ast.Location{ - Line: int(1296), - Column: int(32), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -178188,7 +168705,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHas", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -178211,64 +168727,39 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "o", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1298), - Column: int(13), - }, - End: ast.Location{ - Line: int(1298), - Column: int(14), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "f", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1298), - Column: int(16), - }, - End: ast.Location{ - Line: int(1298), - Column: int(17), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "o", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "f", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(5), }, End: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14078, + Ctx: p14051, FreeVars: ast.Identifiers{ "f", "o", @@ -178278,19 +168769,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(5), }, End: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14078, + Ctx: p14051, FreeVars: ast.Identifiers{ "std", }, @@ -178298,13 +168789,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(5), }, End: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(8), }, File: p1, @@ -178346,9 +168837,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHasEx", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -178358,19 +168848,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(21), }, End: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14088, + Ctx: p14061, FreeVars: ast.Identifiers{ "o", }, @@ -178383,19 +168873,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(24), }, End: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14088, + Ctx: p14061, FreeVars: ast.Identifiers{ "f", }, @@ -178408,19 +168898,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(27), }, End: ast.Location{ - Line: int(1299), + Line: int(1284), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14088, + Ctx: p14061, FreeVars: nil, }, Value: false, @@ -178437,18 +168927,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1298), - Column: int(3), - }, - End: ast.Location{ - Line: int(1299), - Column: int(33), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -178473,7 +168951,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHasAll", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -178496,64 +168973,39 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "o", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1301), - Column: int(16), - }, - End: ast.Location{ - Line: int(1301), - Column: int(17), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "f", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1301), - Column: int(19), - }, - End: ast.Location{ - Line: int(1301), - Column: int(20), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "o", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "f", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(5), }, End: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14098, + Ctx: p14071, FreeVars: ast.Identifiers{ "f", "o", @@ -178563,19 +169015,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(5), }, End: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14098, + Ctx: p14071, FreeVars: ast.Identifiers{ "std", }, @@ -178583,13 +169035,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(5), }, End: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(8), }, File: p1, @@ -178631,9 +169083,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectHasEx", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -178643,19 +169094,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(21), }, End: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14108, + Ctx: p14081, FreeVars: ast.Identifiers{ "o", }, @@ -178668,19 +169119,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(24), }, End: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14108, + Ctx: p14081, FreeVars: ast.Identifiers{ "f", }, @@ -178693,19 +169144,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(27), }, End: ast.Location{ - Line: int(1302), + Line: int(1287), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14108, + Ctx: p14081, FreeVars: nil, }, Value: true, @@ -178722,18 +169173,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1301), - Column: int(3), - }, - End: ast.Location{ - Line: int(1302), - Column: int(32), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -178755,10 +169194,9 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "objectValues", + Value: "equals", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -178781,168 +169219,122 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "o", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1304), - Column: int(16), - }, - End: ast.Location{ - Line: int(1304), - Column: int(17), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, - Body: &ast.Apply{ + Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1305), + Line: int(1290), Column: int(5), }, End: ast.Location{ - Line: int(1305), - Column: int(40), + Line: int(1323), + Column: int(34), }, File: p1, }, - Fodder: nil, - Ctx: nil, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(4), + Comment: []string{}, + }, + }, + Ctx: p14092, FreeVars: ast.Identifiers{ - "o", + "a", + "b", "std", }, }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: nil, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), + Binds: ast.LocalBinds{ + ast.LocalBind{ + VarFodder: ast.Fodder{}, + Variable: "ta", + EqFodder: ast.Fodder{}, + Body: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1290), + Column: int(16), + }, + End: ast.Location{ + Line: int(1290), + Column: int(27), + }, + File: p1, }, - End: ast.Location{ - Line: int(0), - Column: int(0), + Fodder: ast.Fodder{}, + Ctx: p14096, + FreeVars: ast.Identifiers{ + "a", + "std", }, - File: nil, }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "flatMap", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: nil, - Id: nil, - }, - FodderLeft: nil, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Function{ + Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(1290), + Column: int(16), }, End: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(1290), + Column: int(24), }, - File: nil, + File: p1, }, - Fodder: nil, - Ctx: nil, + Fodder: ast.Fodder{}, + Ctx: p14096, FreeVars: ast.Identifiers{ - "o", + "std", }, }, - ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(1290), + Column: int(16), }, End: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(1290), + Column: int(19), }, - File: nil, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", }, }, + Id: "std", }, - TrailingComma: false, - ParenRightFodder: nil, - Body: &ast.Array{ + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ FileName: "", @@ -178958,129 +169350,127 @@ var _StdAst = &ast.DesugaredObject{ }, Fodder: nil, Ctx: nil, - FreeVars: ast.Identifiers{ - "k", - "o", - }, + FreeVars: nil, }, - Elements: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1305), - Column: int(6), - }, - End: ast.Location{ - Line: int(1305), - Column: int(10), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14132, - FreeVars: ast.Identifiers{ - "k", - "o", + Value: "type", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, + }, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1290), + Column: int(25), }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1305), - Column: int(6), - }, - End: ast.Location{ - Line: int(1305), - Column: int(7), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14132, - FreeVars: ast.Identifiers{ - "o", - }, + End: ast.Location{ + Line: int(1290), + Column: int(26), }, - Id: "o", + File: p1, }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1305), - Column: int(8), - }, - End: ast.Location{ - Line: int(1305), - Column: int(9), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14132, - FreeVars: ast.Identifiers{ - "k", - }, - }, - Id: "k", + Fodder: ast.Fodder{}, + Ctx: p14105, + FreeVars: ast.Identifiers{ + "a", }, - RightBracketFodder: ast.Fodder{}, - Id: nil, }, - CommaFodder: nil, + Id: "a", }, + CommaFodder: nil, }, - TrailingComma: false, - CloseFodder: nil, }, + Named: nil, }, - CommaFodder: nil, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ + Fun: nil, + CloseFodder: ast.Fodder{}, + }, + }, + Body: &ast.Local{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1291), + Column: int(5), + }, + End: ast.Location{ + Line: int(1323), + Column: int(34), + }, + File: p1, + }, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(4), + Comment: []string{}, + }, + }, + Ctx: p14092, + FreeVars: ast.Identifiers{ + "a", + "b", + "std", + "ta", + }, + }, + Binds: ast.LocalBinds{ + ast.LocalBind{ + VarFodder: ast.Fodder{}, + Variable: "tb", + EqFodder: ast.Fodder{}, + Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1305), - Column: int(20), + Line: int(1291), + Column: int(16), }, End: ast.Location{ - Line: int(1305), - Column: int(39), + Line: int(1291), + Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14139, + Ctx: p14112, FreeVars: ast.Identifiers{ - "o", + "b", "std", }, }, Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1305), - Column: int(20), + Line: int(1291), + Column: int(16), }, End: ast.Location{ - Line: int(1305), - Column: int(36), + Line: int(1291), + Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14139, + Ctx: p14112, FreeVars: ast.Identifiers{ "std", }, @@ -179088,924 +169478,14 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1305), - Column: int(20), + Line: int(1291), + Column: int(16), }, End: ast.Location{ - Line: int(1305), - Column: int(23), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "objectFields", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1305), - Column: int(37), - }, - End: ast.Location{ - Line: int(1305), - Column: int(38), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14148, - FreeVars: ast.Identifiers{ - "o", - }, - }, - Id: "o", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: nil, - TailStrictFodder: nil, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1304), - Column: int(3), - }, - End: ast.Location{ - Line: int(1305), - Column: int(40), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "objectValuesAll", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "o", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1307), - Column: int(19), - }, - End: ast.Location{ - Line: int(1307), - Column: int(20), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(5), - }, - End: ast.Location{ - Line: int(1308), - Column: int(43), - }, - File: p1, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "o", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: nil, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "flatMap", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: nil, - Id: nil, - }, - FodderLeft: nil, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "o", - }, - }, - ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "k", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: nil, - Body: &ast.Array{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: ast.Identifiers{ - "k", - "o", - }, - }, - Elements: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(6), - }, - End: ast.Location{ - Line: int(1308), - Column: int(10), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14169, - FreeVars: ast.Identifiers{ - "k", - "o", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(6), - }, - End: ast.Location{ - Line: int(1308), - Column: int(7), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14169, - FreeVars: ast.Identifiers{ - "o", - }, - }, - Id: "o", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(8), - }, - End: ast.Location{ - Line: int(1308), - Column: int(9), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14169, - FreeVars: ast.Identifiers{ - "k", - }, - }, - Id: "k", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - CommaFodder: nil, - }, - }, - TrailingComma: false, - CloseFodder: nil, - }, - }, - CommaFodder: nil, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(20), - }, - End: ast.Location{ - Line: int(1308), - Column: int(42), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14176, - FreeVars: ast.Identifiers{ - "o", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(20), - }, - End: ast.Location{ - Line: int(1308), - Column: int(39), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14176, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(20), - }, - End: ast.Location{ - Line: int(1308), - Column: int(23), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "objectFieldsAll", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1308), - Column: int(40), - }, - End: ast.Location{ - Line: int(1308), - Column: int(41), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14185, - FreeVars: ast.Identifiers{ - "o", - }, - }, - Id: "o", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: nil, - TailStrictFodder: nil, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1307), - Column: int(3), - }, - End: ast.Location{ - Line: int(1308), - Column: int(43), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "equals", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1310), - Column: int(10), - }, - End: ast.Location{ - Line: int(1310), - Column: int(11), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1310), - Column: int(13), - }, - End: ast.Location{ - Line: int(1310), - Column: int(14), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1311), - Column: int(5), - }, - End: ast.Location{ - Line: int(1344), - Column: int(34), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - Ctx: p14193, - FreeVars: ast.Identifiers{ - "a", - "b", - "std", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "ta", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1311), - Column: int(16), - }, - End: ast.Location{ - Line: int(1311), - Column: int(27), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14197, - FreeVars: ast.Identifiers{ - "a", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1311), - Column: int(16), - }, - End: ast.Location{ - Line: int(1311), - Column: int(24), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14197, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1311), - Column: int(16), - }, - End: ast.Location{ - Line: int(1311), - Column: int(19), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "type", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1311), - Column: int(25), - }, - End: ast.Location{ - Line: int(1311), - Column: int(26), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14206, - FreeVars: ast.Identifiers{ - "a", - }, - }, - Id: "a", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1311), - Column: int(11), - }, - End: ast.Location{ - Line: int(1311), - Column: int(27), - }, - File: p1, - }, - }, - }, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1312), - Column: int(5), - }, - End: ast.Location{ - Line: int(1344), - Column: int(34), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - Ctx: p14193, - FreeVars: ast.Identifiers{ - "a", - "b", - "std", - "ta", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "tb", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1312), - Column: int(16), - }, - End: ast.Location{ - Line: int(1312), - Column: int(27), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14213, - FreeVars: ast.Identifiers{ - "b", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1312), - Column: int(16), - }, - End: ast.Location{ - Line: int(1312), - Column: int(24), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p14213, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1312), - Column: int(16), - }, - End: ast.Location{ - Line: int(1312), - Column: int(19), + Line: int(1291), + Column: int(19), }, File: p1, }, @@ -180039,9 +169519,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -180051,19 +169530,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1312), + Line: int(1291), Column: int(25), }, End: ast.Location{ - Line: int(1312), + Line: int(1291), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14222, + Ctx: p14121, FreeVars: ast.Identifiers{ "b", }, @@ -180082,30 +169561,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1312), - Column: int(11), - }, - End: ast.Location{ - Line: int(1312), - Column: int(27), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(5), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(34), }, File: p1, @@ -180118,7 +169585,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -180130,19 +169597,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(8), }, End: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", "ta", @@ -180153,19 +169620,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(9), }, End: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", "ta", @@ -180175,19 +169642,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(9), }, End: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", }, @@ -180195,13 +169662,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(9), }, End: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(12), }, File: p1, @@ -180236,9 +169703,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "primitiveEquals", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -180248,19 +169714,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(29), }, End: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14238, + Ctx: p14137, FreeVars: ast.Identifiers{ "ta", }, @@ -180273,19 +169739,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(33), }, End: ast.Location{ - Line: int(1313), + Line: int(1292), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14238, + Ctx: p14137, FreeVars: ast.Identifiers{ "tb", }, @@ -180307,13 +169773,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1314), + Line: int(1293), Column: int(7), }, End: ast.Location{ - Line: int(1314), + Line: int(1293), Column: int(12), }, File: p1, @@ -180326,7 +169792,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: nil, }, Value: false, @@ -180342,13 +169808,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(7), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(34), }, File: p1, @@ -180361,7 +169827,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -180372,19 +169838,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(10), }, End: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", "ta", @@ -180393,19 +169859,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(10), }, End: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", }, @@ -180413,13 +169879,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(10), }, End: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(13), }, File: p1, @@ -180454,9 +169920,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "primitiveEquals", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -180466,19 +169931,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(30), }, End: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14257, + Ctx: p14156, FreeVars: ast.Identifiers{ "ta", }, @@ -180491,25 +169956,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(34), }, End: ast.Location{ - Line: int(1316), + Line: int(1295), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14257, + Ctx: p14156, FreeVars: nil, }, Value: "array", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -180525,13 +169989,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(9), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(23), }, File: p1, @@ -180544,7 +170008,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -180559,19 +170023,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(20), }, End: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14265, + Ctx: p14164, FreeVars: ast.Identifiers{ "a", "std", @@ -180580,19 +170044,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(20), }, End: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14265, + Ctx: p14164, FreeVars: ast.Identifiers{ "std", }, @@ -180600,13 +170064,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(20), }, End: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(23), }, File: p1, @@ -180641,9 +170105,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -180653,19 +170116,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(31), }, End: ast.Location{ - Line: int(1317), + Line: int(1296), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14274, + Ctx: p14173, FreeVars: ast.Identifiers{ "a", }, @@ -180684,30 +170147,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1317), - Column: int(15), - }, - End: ast.Location{ - Line: int(1317), - Column: int(33), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(9), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(23), }, File: p1, @@ -180720,7 +170171,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -180731,19 +170182,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(12), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "b", "la", @@ -180754,19 +170205,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(13), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "b", "la", @@ -180776,19 +170227,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(13), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", }, @@ -180796,13 +170247,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(13), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(16), }, File: p1, @@ -180837,9 +170288,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "primitiveEquals", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -180849,19 +170299,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(33), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14290, + Ctx: p14189, FreeVars: ast.Identifiers{ "la", }, @@ -180874,19 +170324,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(37), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(50), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14290, + Ctx: p14189, FreeVars: ast.Identifiers{ "b", "std", @@ -180895,19 +170345,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(37), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14290, + Ctx: p14189, FreeVars: ast.Identifiers{ "std", }, @@ -180915,13 +170365,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(37), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(40), }, File: p1, @@ -180956,9 +170406,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -180968,19 +170417,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(48), }, End: ast.Location{ - Line: int(1318), + Line: int(1297), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14301, + Ctx: p14200, FreeVars: ast.Identifiers{ "b", }, @@ -181012,13 +170461,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1319), + Line: int(1298), Column: int(11), }, End: ast.Location{ - Line: int(1319), + Line: int(1298), Column: int(16), }, File: p1, @@ -181031,7 +170480,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: nil, }, Value: false, @@ -181047,13 +170496,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1321), + Line: int(1300), Column: int(11), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(23), }, File: p1, @@ -181066,7 +170515,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -181081,96 +170530,57 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1321), + Line: int(1300), Column: int(17), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(31), }, File: p1, }, Fodder: nil, - Ctx: p14311, + Ctx: p14210, FreeVars: ast.Identifiers{ "aux", "la", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1321), - Column: int(21), - }, - End: ast.Location{ - Line: int(1321), - Column: int(22), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1321), - Column: int(24), - }, - End: ast.Location{ - Line: int(1321), - Column: int(25), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1321), - Column: int(27), - }, - End: ast.Location{ - Line: int(1321), - Column: int(28), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1322), + Line: int(1301), Column: int(13), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(31), }, File: p1, @@ -181183,7 +170593,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "a", "aux", @@ -181195,19 +170605,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1322), + Line: int(1301), Column: int(16), }, End: ast.Location{ - Line: int(1322), + Line: int(1301), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "i", "la", @@ -181216,19 +170626,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1322), + Line: int(1301), Column: int(16), }, End: ast.Location{ - Line: int(1322), + Line: int(1301), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "i", }, @@ -181240,19 +170650,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1322), + Line: int(1301), Column: int(21), }, End: ast.Location{ - Line: int(1322), + Line: int(1301), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "la", }, @@ -181264,13 +170674,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1323), + Line: int(1302), Column: int(15), }, End: ast.Location{ - Line: int(1323), + Line: int(1302), Column: int(19), }, File: p1, @@ -181283,7 +170693,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14316, + Ctx: p14215, FreeVars: nil, }, Value: true, @@ -181299,19 +170709,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(18), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "a", "aux", @@ -181322,19 +170732,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(21), }, End: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "a", "b", @@ -181344,19 +170754,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(21), }, End: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "a", "i", @@ -181365,19 +170775,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(21), }, End: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "a", }, @@ -181388,19 +170798,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(23), }, End: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "i", }, @@ -181415,19 +170825,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(29), }, End: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "b", "i", @@ -181436,19 +170846,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(29), }, End: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "b", }, @@ -181459,19 +170869,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(31), }, End: ast.Location{ - Line: int(1324), + Line: int(1303), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "i", }, @@ -181486,13 +170896,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1325), + Line: int(1304), Column: int(15), }, End: ast.Location{ - Line: int(1325), + Line: int(1304), Column: int(20), }, File: p1, @@ -181505,7 +170915,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14316, + Ctx: p14215, FreeVars: nil, }, Value: false, @@ -181521,19 +170931,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(15), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "a", "aux", @@ -181544,13 +170954,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(15), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(18), }, File: p1, @@ -181563,7 +170973,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14316, + Ctx: p14215, FreeVars: ast.Identifiers{ "aux", }, @@ -181577,19 +170987,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(19), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14353, + Ctx: p14252, FreeVars: ast.Identifiers{ "a", }, @@ -181602,19 +171012,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(22), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14353, + Ctx: p14252, FreeVars: ast.Identifiers{ "b", }, @@ -181627,19 +171037,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(25), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14353, + Ctx: p14252, FreeVars: ast.Identifiers{ "i", }, @@ -181647,19 +171057,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(25), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14353, + Ctx: p14252, FreeVars: ast.Identifiers{ "i", }, @@ -181671,21 +171081,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(29), }, End: ast.Location{ - Line: int(1327), + Line: int(1306), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14353, + Ctx: p14252, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -181704,36 +171115,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(11), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "aux", @@ -181743,13 +171142,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(11), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(14), }, File: p1, @@ -181762,7 +171161,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "aux", }, @@ -181776,19 +171175,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(15), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14369, + Ctx: p14268, FreeVars: ast.Identifiers{ "a", }, @@ -181801,19 +171200,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(18), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14369, + Ctx: p14268, FreeVars: ast.Identifiers{ "b", }, @@ -181826,21 +171225,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(21), }, End: ast.Location{ - Line: int(1328), + Line: int(1307), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14369, + Ctx: p14268, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: nil, @@ -181867,19 +171267,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(12), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -181890,19 +171290,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(15), }, End: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", "ta", @@ -181911,19 +171311,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(15), }, End: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", }, @@ -181931,13 +171331,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(15), }, End: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(18), }, File: p1, @@ -181972,9 +171372,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "primitiveEquals", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -181984,19 +171383,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(35), }, End: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14386, + Ctx: p14285, FreeVars: ast.Identifiers{ "ta", }, @@ -182009,25 +171408,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(39), }, End: ast.Location{ - Line: int(1329), + Line: int(1308), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14386, + Ctx: p14285, FreeVars: nil, }, Value: "object", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -182043,13 +171441,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(9), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(23), }, File: p1, @@ -182062,7 +171460,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -182077,19 +171475,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(24), }, End: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14394, + Ctx: p14293, FreeVars: ast.Identifiers{ "a", "std", @@ -182098,19 +171496,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(24), }, End: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14394, + Ctx: p14293, FreeVars: ast.Identifiers{ "std", }, @@ -182118,13 +171516,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(24), }, End: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(27), }, File: p1, @@ -182159,9 +171557,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -182171,19 +171568,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(41), }, End: ast.Location{ - Line: int(1330), + Line: int(1309), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14403, + Ctx: p14302, FreeVars: ast.Identifiers{ "a", }, @@ -182202,30 +171599,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1330), - Column: int(15), - }, - End: ast.Location{ - Line: int(1330), - Column: int(43), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(9), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(23), }, File: p1, @@ -182238,7 +171623,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -182254,19 +171639,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(25), }, End: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14410, + Ctx: p14309, FreeVars: ast.Identifiers{ "fields", "std", @@ -182275,19 +171660,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(25), }, End: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14410, + Ctx: p14309, FreeVars: ast.Identifiers{ "std", }, @@ -182295,13 +171680,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(25), }, End: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(28), }, File: p1, @@ -182336,9 +171721,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -182348,19 +171732,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(36), }, End: ast.Location{ - Line: int(1331), + Line: int(1310), Column: int(42), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14419, + Ctx: p14318, FreeVars: ast.Identifiers{ "fields", }, @@ -182379,30 +171763,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1331), - Column: int(15), - }, - End: ast.Location{ - Line: int(1331), - Column: int(43), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(9), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(23), }, File: p1, @@ -182415,7 +171787,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -182427,19 +171799,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(12), }, End: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "b", "fields", @@ -182449,19 +171821,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(12), }, End: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(18), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "fields", }, @@ -182473,19 +171845,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(22), }, End: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(41), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "b", "std", @@ -182494,19 +171866,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(22), }, End: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", }, @@ -182514,13 +171886,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(22), }, End: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(25), }, File: p1, @@ -182555,9 +171927,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -182567,19 +171938,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(39), }, End: ast.Location{ - Line: int(1332), + Line: int(1311), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14437, + Ctx: p14336, FreeVars: ast.Identifiers{ "b", }, @@ -182601,13 +171972,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1333), + Line: int(1312), Column: int(11), }, End: ast.Location{ - Line: int(1333), + Line: int(1312), Column: int(16), }, File: p1, @@ -182620,7 +171991,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: nil, }, Value: false, @@ -182636,13 +172007,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1335), + Line: int(1314), Column: int(11), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(23), }, File: p1, @@ -182655,7 +172026,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -182671,19 +172042,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1335), + Line: int(1314), Column: int(17), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(31), }, File: p1, }, Fodder: nil, - Ctx: p14447, + Ctx: p14346, FreeVars: ast.Identifiers{ "aux", "fields", @@ -182691,77 +172062,38 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1335), - Column: int(21), - }, - End: ast.Location{ - Line: int(1335), - Column: int(22), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1335), - Column: int(24), - }, - End: ast.Location{ - Line: int(1335), - Column: int(25), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1335), - Column: int(27), - }, - End: ast.Location{ - Line: int(1335), - Column: int(28), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1336), + Line: int(1315), Column: int(13), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(31), }, File: p1, @@ -182774,7 +172106,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "a", "aux", @@ -182787,19 +172119,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1336), + Line: int(1315), Column: int(16), }, End: ast.Location{ - Line: int(1336), + Line: int(1315), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "i", "lfields", @@ -182808,19 +172140,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1336), + Line: int(1315), Column: int(16), }, End: ast.Location{ - Line: int(1336), + Line: int(1315), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "i", }, @@ -182832,19 +172164,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1336), + Line: int(1315), Column: int(21), }, End: ast.Location{ - Line: int(1336), + Line: int(1315), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "lfields", }, @@ -182856,13 +172188,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1337), + Line: int(1316), Column: int(15), }, End: ast.Location{ - Line: int(1337), + Line: int(1316), Column: int(19), }, File: p1, @@ -182875,7 +172207,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14452, + Ctx: p14351, FreeVars: nil, }, Value: true, @@ -182891,19 +172223,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(18), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "a", "aux", @@ -182915,19 +172247,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(21), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "a", "b", @@ -182943,19 +172275,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(31), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14469, + Ctx: p14368, FreeVars: ast.Identifiers{ "fields", "i", @@ -182964,19 +172296,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(31), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14469, + Ctx: p14368, FreeVars: ast.Identifiers{ "fields", }, @@ -182987,19 +172319,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(38), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14469, + Ctx: p14368, FreeVars: ast.Identifiers{ "i", }, @@ -183011,36 +172343,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1338), - Column: int(27), - }, - End: ast.Location{ - Line: int(1338), - Column: int(40), - }, - File: p1, - }, }, }, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(42), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "a", "b", @@ -183050,19 +172370,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(42), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(46), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "a", "f", @@ -183071,19 +172391,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(42), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "a", }, @@ -183094,19 +172414,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(44), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "f", }, @@ -183121,19 +172441,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(50), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "b", "f", @@ -183142,19 +172462,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(50), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(51), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "b", }, @@ -183165,19 +172485,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(52), }, End: ast.Location{ - Line: int(1338), + Line: int(1317), Column: int(53), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "f", }, @@ -183193,13 +172513,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1339), + Line: int(1318), Column: int(15), }, End: ast.Location{ - Line: int(1339), + Line: int(1318), Column: int(20), }, File: p1, @@ -183212,7 +172532,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14452, + Ctx: p14351, FreeVars: nil, }, Value: false, @@ -183228,19 +172548,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(15), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "a", "aux", @@ -183251,13 +172571,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(15), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(18), }, File: p1, @@ -183270,7 +172590,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14452, + Ctx: p14351, FreeVars: ast.Identifiers{ "aux", }, @@ -183284,19 +172604,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(19), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14499, + Ctx: p14398, FreeVars: ast.Identifiers{ "a", }, @@ -183309,19 +172629,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(22), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14499, + Ctx: p14398, FreeVars: ast.Identifiers{ "b", }, @@ -183334,19 +172654,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(25), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14499, + Ctx: p14398, FreeVars: ast.Identifiers{ "i", }, @@ -183354,19 +172674,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(25), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14499, + Ctx: p14398, FreeVars: ast.Identifiers{ "i", }, @@ -183378,21 +172698,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(29), }, End: ast.Location{ - Line: int(1341), + Line: int(1320), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14499, + Ctx: p14398, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -183411,36 +172732,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(11), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "aux", @@ -183450,13 +172759,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(11), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(14), }, File: p1, @@ -183469,7 +172778,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "aux", }, @@ -183483,19 +172792,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(15), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(16), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14515, + Ctx: p14414, FreeVars: ast.Identifiers{ "a", }, @@ -183508,19 +172817,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(18), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14515, + Ctx: p14414, FreeVars: ast.Identifiers{ "b", }, @@ -183533,21 +172842,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(21), }, End: ast.Location{ - Line: int(1342), + Line: int(1321), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14515, + Ctx: p14414, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: nil, @@ -183575,19 +172885,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(9), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "a", "b", @@ -183597,19 +172907,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(9), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14193, + Ctx: p14092, FreeVars: ast.Identifiers{ "std", }, @@ -183617,13 +172927,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(9), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(12), }, File: p1, @@ -183665,9 +172975,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "primitiveEquals", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -183677,19 +172986,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(29), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14531, + Ctx: p14430, FreeVars: ast.Identifiers{ "a", }, @@ -183702,19 +173011,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(32), }, End: ast.Location{ - Line: int(1344), + Line: int(1323), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14531, + Ctx: p14430, FreeVars: ast.Identifiers{ "b", }, @@ -183738,18 +173047,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1310), - Column: int(3), - }, - End: ast.Location{ - Line: int(1344), - Column: int(34), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -183774,7 +173071,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "resolvePath", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -183797,58 +173093,33 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "f", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1347), - Column: int(15), - }, - End: ast.Location{ - Line: int(1347), - Column: int(16), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "r", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1347), - Column: int(18), - }, - End: ast.Location{ - Line: int(1347), - Column: int(19), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "f", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "r", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(5), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(80), }, File: p1, @@ -183861,7 +173132,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14541, + Ctx: p14440, FreeVars: ast.Identifiers{ "f", "r", @@ -183876,19 +173147,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(17), }, End: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14545, + Ctx: p14444, FreeVars: ast.Identifiers{ "f", "std", @@ -183897,19 +173168,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(17), }, End: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14545, + Ctx: p14444, FreeVars: ast.Identifiers{ "std", }, @@ -183917,13 +173188,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(17), }, End: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(20), }, File: p1, @@ -183958,9 +173229,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "split", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -183970,19 +173240,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(27), }, End: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14554, + Ctx: p14453, FreeVars: ast.Identifiers{ "f", }, @@ -183995,25 +173265,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(30), }, End: ast.Location{ - Line: int(1348), + Line: int(1327), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14554, + Ctx: p14453, FreeVars: nil, }, Value: "/", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -184027,36 +173296,24 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1348), - Column: int(11), - }, - End: ast.Location{ - Line: int(1348), - Column: int(34), - }, - File: p1, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(5), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14541, + Ctx: p14440, FreeVars: ast.Identifiers{ "arr", "r", @@ -184066,19 +173323,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(5), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14541, + Ctx: p14440, FreeVars: ast.Identifiers{ "std", }, @@ -184086,13 +173343,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(5), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(8), }, File: p1, @@ -184134,9 +173391,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "join", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -184146,25 +173402,24 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(14), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14567, + Ctx: p14466, FreeVars: nil, }, Value: "/", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: ast.Fodder{}, }, @@ -184172,19 +173427,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(19), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14567, + Ctx: p14466, FreeVars: ast.Identifiers{ "arr", "r", @@ -184194,19 +173449,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(19), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(73), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14567, + Ctx: p14466, FreeVars: ast.Identifiers{ "arr", "std", @@ -184215,19 +173470,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(19), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14567, + Ctx: p14466, FreeVars: ast.Identifiers{ "std", }, @@ -184235,13 +173490,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(19), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(22), }, File: p1, @@ -184276,9 +173531,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "makeArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -184288,19 +173542,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(33), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14579, + Ctx: p14478, FreeVars: ast.Identifiers{ "arr", "std", @@ -184309,19 +173563,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(33), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(48), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14579, + Ctx: p14478, FreeVars: ast.Identifiers{ "arr", "std", @@ -184330,19 +173584,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(33), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(43), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14579, + Ctx: p14478, FreeVars: ast.Identifiers{ "std", }, @@ -184350,13 +173604,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(33), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(36), }, File: p1, @@ -184391,9 +173645,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -184403,19 +173656,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(44), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(47), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14590, + Ctx: p14489, FreeVars: ast.Identifiers{ "arr", }, @@ -184437,21 +173690,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(51), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14579, + Ctx: p14478, FreeVars: nil, }, + Value: float64(1), OriginalString: "1", }, }, @@ -184461,63 +173715,52 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(54), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14579, + Ctx: p14478, FreeVars: ast.Identifiers{ "arr", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1349), - Column: int(63), - }, - End: ast.Location{ - Line: int(1349), - Column: int(64), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(66), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14597, + Ctx: p14496, FreeVars: ast.Identifiers{ "arr", "i", @@ -184526,19 +173769,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(66), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14597, + Ctx: p14496, FreeVars: ast.Identifiers{ "arr", }, @@ -184549,19 +173792,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(70), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(71), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14597, + Ctx: p14496, FreeVars: ast.Identifiers{ "i", }, @@ -184587,19 +173830,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(76), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(79), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14567, + Ctx: p14466, FreeVars: ast.Identifiers{ "r", }, @@ -184609,19 +173852,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(77), }, End: ast.Location{ - Line: int(1349), + Line: int(1328), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14607, + Ctx: p14506, FreeVars: ast.Identifiers{ "r", }, @@ -184648,18 +173891,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1347), - Column: int(3), - }, - End: ast.Location{ - Line: int(1349), - Column: int(80), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -184684,7 +173915,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "prune", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -184708,39 +173938,28 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "a", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1351), - Column: int(9), - }, - End: ast.Location{ - Line: int(1351), - Column: int(10), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "a", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1352), + Line: int(1331), Column: int(5), }, End: ast.Location{ - Line: int(1368), + Line: int(1347), Column: int(8), }, File: p1, @@ -184753,7 +173972,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "$", "a", @@ -184768,57 +173987,46 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1352), + Line: int(1331), Column: int(11), }, End: ast.Location{ - Line: int(1360), + Line: int(1339), Column: int(13), }, File: p1, }, Fodder: nil, - Ctx: p14619, + Ctx: p14518, FreeVars: ast.Identifiers{ "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "b", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1352), - Column: int(21), - }, - End: ast.Location{ - Line: int(1352), - Column: int(22), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "b", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1353), + Line: int(1332), Column: int(7), }, End: ast.Location{ - Line: int(1360), + Line: int(1339), Column: int(13), }, File: p1, @@ -184831,7 +174039,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -184840,19 +174048,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1353), + Line: int(1332), Column: int(10), }, End: ast.Location{ - Line: int(1353), + Line: int(1332), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", }, @@ -184860,19 +174068,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1353), + Line: int(1332), Column: int(10), }, End: ast.Location{ - Line: int(1353), + Line: int(1332), Column: int(11), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", }, @@ -184884,19 +174092,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNull{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1353), + Line: int(1332), Column: int(15), }, End: ast.Location{ - Line: int(1353), + Line: int(1332), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: nil, }, }, @@ -184905,13 +174113,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1354), + Line: int(1333), Column: int(9), }, End: ast.Location{ - Line: int(1354), + Line: int(1333), Column: int(14), }, File: p1, @@ -184924,7 +174132,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14624, + Ctx: p14523, FreeVars: nil, }, Value: false, @@ -184940,19 +174148,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(12), }, End: ast.Location{ - Line: int(1360), + Line: int(1339), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -184961,19 +174169,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(15), }, End: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -184982,19 +174190,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(15), }, End: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "std", }, @@ -185002,13 +174210,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(15), }, End: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(18), }, File: p1, @@ -185043,9 +174251,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -185055,19 +174262,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(27), }, End: ast.Location{ - Line: int(1355), + Line: int(1334), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14645, + Ctx: p14544, FreeVars: ast.Identifiers{ "b", }, @@ -185088,19 +174295,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(9), }, End: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -185109,19 +174316,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(9), }, End: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -185130,19 +174337,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(9), }, End: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "std", }, @@ -185150,13 +174357,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(9), }, End: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(12), }, File: p1, @@ -185198,9 +174405,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -185210,19 +174416,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(20), }, End: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14659, + Ctx: p14558, FreeVars: ast.Identifiers{ "b", }, @@ -185244,21 +174450,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(25), }, End: ast.Location{ - Line: int(1356), + Line: int(1335), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -185273,19 +174480,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(12), }, End: ast.Location{ - Line: int(1360), + Line: int(1339), Column: int(13), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -185294,19 +174501,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(15), }, End: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -185315,19 +174522,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(15), }, End: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "std", }, @@ -185335,13 +174542,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(15), }, End: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(18), }, File: p1, @@ -185376,9 +174583,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -185388,19 +174594,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(28), }, End: ast.Location{ - Line: int(1357), + Line: int(1336), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14674, + Ctx: p14573, FreeVars: ast.Identifiers{ "b", }, @@ -185421,19 +174627,19 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(9), }, End: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -185442,19 +174648,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(9), }, End: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "b", "std", @@ -185463,19 +174669,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(9), }, End: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: ast.Identifiers{ "std", }, @@ -185483,13 +174689,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(9), }, End: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(12), }, File: p1, @@ -185531,9 +174737,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -185543,19 +174748,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(20), }, End: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14688, + Ctx: p14587, FreeVars: ast.Identifiers{ "b", }, @@ -185577,21 +174782,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(25), }, End: ast.Location{ - Line: int(1358), + Line: int(1337), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14624, + Ctx: p14523, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -185606,13 +174812,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.LiteralBoolean{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1360), + Line: int(1339), Column: int(9), }, End: ast.Location{ - Line: int(1360), + Line: int(1339), Column: int(13), }, File: p1, @@ -185625,7 +174831,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14624, + Ctx: p14523, FreeVars: nil, }, Value: true, @@ -185636,30 +174842,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(5), }, End: ast.Location{ - Line: int(1368), + Line: int(1347), Column: int(8), }, File: p1, @@ -185672,7 +174866,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "$", "a", @@ -185683,19 +174877,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(8), }, End: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "a", "std", @@ -185704,19 +174898,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(8), }, End: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "std", }, @@ -185724,13 +174918,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(8), }, End: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(11), }, File: p1, @@ -185765,9 +174959,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isArray", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -185777,19 +174970,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(20), }, End: ast.Location{ - Line: int(1361), + Line: int(1340), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14706, + Ctx: p14605, FreeVars: ast.Identifiers{ "a", }, @@ -185812,14 +175005,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1362), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1362), - Column: int(57), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -185894,7 +175087,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -185926,26 +175118,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -185975,19 +175156,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(35), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(56), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "$", "isContent", @@ -185997,19 +175178,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(35), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "isContent", }, @@ -186023,19 +175204,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(45), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(55), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14727, + Ctx: p14626, FreeVars: ast.Identifiers{ "$", "x", @@ -186044,19 +175225,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(45), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(52), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14727, + Ctx: p14626, FreeVars: ast.Identifiers{ "$", }, @@ -186064,13 +175245,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(45), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(46), }, File: p1, @@ -186105,9 +175286,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "prune", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -186117,19 +175297,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(53), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(54), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14736, + Ctx: p14635, FreeVars: ast.Identifiers{ "x", }, @@ -186183,19 +175363,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(8), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14742, + Ctx: p14641, FreeVars: ast.Identifiers{ "std", "x", @@ -186204,19 +175384,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(8), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14742, + Ctx: p14641, FreeVars: ast.Identifiers{ "std", }, @@ -186224,13 +175404,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(8), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(11), }, File: p1, @@ -186265,9 +175445,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "prune", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -186277,19 +175456,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(18), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14751, + Ctx: p14650, FreeVars: ast.Identifiers{ "x", }, @@ -186343,19 +175522,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(30), }, End: ast.Location{ - Line: int(1362), + Line: int(1341), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "a", }, @@ -186383,19 +175562,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(10), }, End: ast.Location{ - Line: int(1368), + Line: int(1347), Column: int(8), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "$", "a", @@ -186406,19 +175585,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(13), }, End: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(28), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "a", "std", @@ -186427,19 +175606,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(13), }, End: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "std", }, @@ -186447,13 +175626,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(13), }, End: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(16), }, File: p1, @@ -186488,9 +175667,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isObject", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -186500,19 +175678,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(26), }, End: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(27), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14768, + Ctx: p14667, FreeVars: ast.Identifiers{ "a", }, @@ -186535,14 +175713,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1363), - Column: int(34), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1367), - Column: int(6), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -186617,7 +175795,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "$objectFlatMerge", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -186631,14 +175808,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1363), - Column: int(34), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1367), - Column: int(6), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -186713,7 +175890,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -186746,26 +175922,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "x", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -186796,19 +175961,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(10), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(36), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "a", "isContent", @@ -186819,19 +175984,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(10), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "isContent", }, @@ -186845,19 +176010,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(20), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14797, + Ctx: p14696, FreeVars: ast.Identifiers{ "a", "std", @@ -186867,19 +176032,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(20), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(29), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14797, + Ctx: p14696, FreeVars: ast.Identifiers{ "std", }, @@ -186887,13 +176052,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(20), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(23), }, File: p1, @@ -186928,9 +176093,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "prune", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -186940,19 +176104,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(30), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14806, + Ctx: p14705, FreeVars: ast.Identifiers{ "a", "x", @@ -186961,19 +176125,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(30), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14806, + Ctx: p14705, FreeVars: ast.Identifiers{ "a", }, @@ -186984,19 +176148,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(32), }, End: ast.Location{ - Line: int(1366), + Line: int(1345), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14806, + Ctx: p14705, FreeVars: ast.Identifiers{ "x", }, @@ -187054,19 +176218,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1363), + Line: int(1342), Column: int(34), }, End: ast.Location{ - Line: int(1367), + Line: int(1346), Column: int(6), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "$", "a", @@ -187080,19 +176244,19 @@ var _StdAst = &ast.DesugaredObject{ Name: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(8), }, End: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(9), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "x", }, @@ -187102,19 +176266,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(12), }, End: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14821, + Ctx: p14720, FreeVars: ast.Identifiers{ "$", "a", @@ -187124,19 +176288,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(12), }, End: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14821, + Ctx: p14720, FreeVars: ast.Identifiers{ "$", }, @@ -187144,13 +176308,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(12), }, End: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(13), }, File: p1, @@ -187185,9 +176349,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "prune", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -187197,19 +176360,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(20), }, End: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(24), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14830, + Ctx: p14729, FreeVars: ast.Identifiers{ "a", "x", @@ -187218,19 +176381,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(20), }, End: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14830, + Ctx: p14729, FreeVars: ast.Identifiers{ "a", }, @@ -187241,19 +176404,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(22), }, End: ast.Location{ - Line: int(1364), + Line: int(1343), Column: int(23), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14830, + Ctx: p14729, FreeVars: ast.Identifiers{ "x", }, @@ -187274,18 +176437,6 @@ var _StdAst = &ast.DesugaredObject{ TailStrictFodder: nil, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1364), - Column: int(7), - }, - End: ast.Location{ - Line: int(1364), - Column: int(25), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -187327,19 +176478,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(16), }, End: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "a", "std", @@ -187348,19 +176499,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(16), }, End: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(32), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "std", }, @@ -187368,13 +176519,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(16), }, End: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(19), }, File: p1, @@ -187409,9 +176560,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "objectFields", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -187421,19 +176571,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(33), }, End: ast.Location{ - Line: int(1365), + Line: int(1344), Column: int(34), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14846, + Ctx: p14745, FreeVars: ast.Identifiers{ "a", }, @@ -187474,13 +176624,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1368), + Line: int(1347), Column: int(7), }, End: ast.Location{ - Line: int(1368), + Line: int(1347), Column: int(8), }, File: p1, @@ -187493,7 +176643,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14615, + Ctx: p14514, FreeVars: ast.Identifiers{ "a", }, @@ -187505,18 +176655,6 @@ var _StdAst = &ast.DesugaredObject{ }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1351), - Column: int(3), - }, - End: ast.Location{ - Line: int(1368), - Column: int(8), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -187541,7 +176679,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "findSubstr", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -187564,58 +176701,33 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "pat", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1370), - Column: int(14), - }, - End: ast.Location{ - Line: int(1370), - Column: int(17), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1370), - Column: int(19), - }, - End: ast.Location{ - Line: int(1370), - Column: int(22), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "pat", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(5), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(91), }, File: p1, @@ -187628,7 +176740,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -187638,19 +176750,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(8), }, End: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -187660,19 +176772,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(9), }, End: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -187681,19 +176793,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(9), }, End: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(21), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", }, @@ -187701,13 +176813,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(9), }, End: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(12), }, File: p1, @@ -187742,9 +176854,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -187754,19 +176865,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(22), }, End: ast.Location{ - Line: int(1371), + Line: int(1350), Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14870, + Ctx: p14769, FreeVars: ast.Identifiers{ "pat", }, @@ -187788,13 +176899,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(7), }, End: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(82), }, File: p1, @@ -187807,7 +176918,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -187816,19 +176927,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(13), }, End: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -187837,44 +176948,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(13), }, End: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(66), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: nil, }, Value: "findSubstr first parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(69), }, End: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -187883,19 +176993,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(69), }, End: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(77), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", }, @@ -187903,13 +177013,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(69), }, End: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(72), }, File: p1, @@ -187944,9 +177054,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -187956,19 +177065,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(78), }, End: ast.Location{ - Line: int(1372), + Line: int(1351), Column: int(81), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14887, + Ctx: p14786, FreeVars: ast.Identifiers{ "pat", }, @@ -187998,19 +177107,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(10), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(91), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -188020,19 +177129,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(13), }, End: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", "str", @@ -188042,19 +177151,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(14), }, End: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(31), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", "str", @@ -188063,19 +177172,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(14), }, End: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(26), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", }, @@ -188083,13 +177192,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(14), }, End: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(17), }, File: p1, @@ -188124,9 +177233,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "isString", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -188136,19 +177244,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(27), }, End: ast.Location{ - Line: int(1373), + Line: int(1352), Column: int(30), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14903, + Ctx: p14802, FreeVars: ast.Identifiers{ "str", }, @@ -188170,13 +177278,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Error{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(7), }, End: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(83), }, File: p1, @@ -188189,7 +177297,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", "str", @@ -188198,19 +177306,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(13), }, End: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(83), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", "str", @@ -188219,44 +177327,43 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(13), }, End: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: nil, }, Value: "findSubstr second parameter should be a string, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, OpFodder: ast.Fodder{}, Op: ast.BinaryOp(3), Right: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(70), }, End: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(83), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", "str", @@ -188265,19 +177372,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(70), }, End: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(78), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", }, @@ -188285,13 +177392,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(70), }, End: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(73), }, File: p1, @@ -188326,9 +177433,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "type", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -188338,19 +177444,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(79), }, End: ast.Location{ - Line: int(1374), + Line: int(1353), Column: int(82), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14920, + Ctx: p14819, FreeVars: ast.Identifiers{ "str", }, @@ -188380,13 +177486,13 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(7), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(91), }, File: p1, @@ -188399,7 +177505,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "std", @@ -188414,19 +177520,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(23), }, End: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14928, + Ctx: p14827, FreeVars: ast.Identifiers{ "pat", "std", @@ -188435,19 +177541,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(23), }, End: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14928, + Ctx: p14827, FreeVars: ast.Identifiers{ "std", }, @@ -188455,13 +177561,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(23), }, End: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(26), }, File: p1, @@ -188496,9 +177602,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -188508,19 +177613,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(34), }, End: ast.Location{ - Line: int(1376), + Line: int(1355), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14937, + Ctx: p14836, FreeVars: ast.Identifiers{ "pat", }, @@ -188539,30 +177644,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1376), - Column: int(13), - }, - End: ast.Location{ - Line: int(1376), - Column: int(38), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(7), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(91), }, File: p1, @@ -188575,7 +177668,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "pat_len", @@ -188591,19 +177684,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(23), }, End: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14944, + Ctx: p14843, FreeVars: ast.Identifiers{ "std", "str", @@ -188612,19 +177705,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(23), }, End: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14944, + Ctx: p14843, FreeVars: ast.Identifiers{ "std", }, @@ -188632,13 +177725,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(23), }, End: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(26), }, File: p1, @@ -188673,9 +177766,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "length", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -188685,19 +177777,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(34), }, End: ast.Location{ - Line: int(1377), + Line: int(1356), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14953, + Ctx: p14852, FreeVars: ast.Identifiers{ "str", }, @@ -188716,30 +177808,18 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1377), - Column: int(13), - }, - End: ast.Location{ - Line: int(1377), - Column: int(38), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(7), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(91), }, File: p1, @@ -188752,7 +177832,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "pat_len", @@ -188764,19 +177844,19 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(10), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat_len", "str_len", @@ -188785,19 +177865,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(10), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat_len", "str_len", @@ -188806,19 +177886,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(10), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat_len", }, @@ -188826,19 +177906,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(10), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat_len", }, @@ -188850,21 +177930,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(21), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(22), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -188873,19 +177954,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(26), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "str_len", }, @@ -188893,19 +177974,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(26), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(33), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "str_len", }, @@ -188917,21 +177998,22 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(37), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(38), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -188941,19 +178023,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(42), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat_len", "str_len", @@ -188962,19 +178044,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(42), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat_len", }, @@ -188986,19 +178068,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(52), }, End: ast.Location{ - Line: int(1378), + Line: int(1357), Column: int(59), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "str_len", }, @@ -189011,13 +178093,13 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1379), + Line: int(1358), Column: int(9), }, End: ast.Location{ - Line: int(1379), + Line: int(1358), Column: int(11), }, File: p1, @@ -189030,7 +178112,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p14857, + Ctx: p14756, FreeVars: nil, }, Elements: nil, @@ -189048,19 +178130,19 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(9), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(91), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "pat", "pat_len", @@ -189072,19 +178154,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(9), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(19), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14857, + Ctx: p14756, FreeVars: ast.Identifiers{ "std", }, @@ -189092,13 +178174,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(9), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(12), }, File: p1, @@ -189140,9 +178222,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "filter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -189152,19 +178233,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(20), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14991, + Ctx: p14890, FreeVars: ast.Identifiers{ "pat", "pat_len", @@ -189173,45 +178254,34 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1381), - Column: int(29), - }, - End: ast.Location{ - Line: int(1381), - Column: int(30), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(32), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14995, + Ctx: p14894, FreeVars: ast.Identifiers{ "i", "pat", @@ -189225,14 +178295,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1381), - Column: int(32), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1381), - Column: int(50), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -189307,7 +178377,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "slice", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -189319,19 +178388,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(32), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(35), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14995, + Ctx: p14894, FreeVars: ast.Identifiers{ "str", }, @@ -189344,19 +178413,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(36), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14995, + Ctx: p14894, FreeVars: ast.Identifiers{ "i", }, @@ -189369,19 +178438,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(38), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14995, + Ctx: p14894, FreeVars: ast.Identifiers{ "i", "pat_len", @@ -189390,19 +178459,19 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(38), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14995, + Ctx: p14894, FreeVars: ast.Identifiers{ "i", }, @@ -189414,19 +178483,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(42), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(49), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14995, + Ctx: p14894, FreeVars: ast.Identifiers{ "pat_len", }, @@ -189471,19 +178540,19 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(54), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(57), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14995, + Ctx: p14894, FreeVars: ast.Identifiers{ "pat", }, @@ -189498,19 +178567,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(59), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(90), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14991, + Ctx: p14890, FreeVars: ast.Identifiers{ "pat_len", "std", @@ -189520,19 +178589,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(59), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(68), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p14991, + Ctx: p14890, FreeVars: ast.Identifiers{ "std", }, @@ -189540,13 +178609,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(59), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(62), }, File: p1, @@ -189581,9 +178650,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "range", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -189593,21 +178661,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(69), }, End: ast.Location{ - Line: int(1381), + Line: int(1360), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15027, + Ctx: p14926, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -189616,5133 +178685,98 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1381), - Column: int(72), - }, - End: ast.Location{ - Line: int(1381), - Column: int(89), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15027, - FreeVars: ast.Identifiers{ - "pat_len", - "str_len", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1381), - Column: int(72), - }, - End: ast.Location{ - Line: int(1381), - Column: int(79), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15027, - FreeVars: ast.Identifiers{ - "str_len", - }, - }, - Id: "str_len", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(4), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1381), - Column: int(82), - }, - End: ast.Location{ - Line: int(1381), - Column: int(89), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15027, - FreeVars: ast.Identifiers{ - "pat_len", - }, - }, - Id: "pat_len", - }, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - }, - }, - }, - }, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1370), - Column: int(3), - }, - End: ast.Location{ - Line: int(1381), - Column: int(91), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "find", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "value", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1383), - Column: int(8), - }, - End: ast.Location{ - Line: int(1383), - Column: int(13), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1383), - Column: int(15), - }, - End: ast.Location{ - Line: int(1383), - Column: int(18), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1384), - Column: int(5), - }, - End: ast.Location{ - Line: int(1387), - Column: int(81), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "arr", - "std", - "value", - }, - }, - Cond: &ast.Unary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1384), - Column: int(8), - }, - End: ast.Location{ - Line: int(1384), - Column: int(25), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Op: ast.UnaryOp(0), - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1384), - Column: int(9), - }, - End: ast.Location{ - Line: int(1384), - Column: int(25), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1384), - Column: int(9), - }, - End: ast.Location{ - Line: int(1384), - Column: int(20), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1384), - Column: int(9), - }, - End: ast.Location{ - Line: int(1384), - Column: int(12), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "isArray", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1384), - Column: int(21), - }, - End: ast.Location{ - Line: int(1384), - Column: int(24), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15053, - FreeVars: ast.Identifiers{ - "arr", - }, - }, - Id: "arr", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Error{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1385), - Column: int(7), - }, - End: ast.Location{ - Line: int(1385), - Column: int(77), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Expr: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1385), - Column: int(13), - }, - End: ast.Location{ - Line: int(1385), - Column: int(77), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Left: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1385), - Column: int(13), - }, - End: ast.Location{ - Line: int(1385), - Column: int(61), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: nil, - }, - Value: "find second parameter should be an array, got ", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1385), - Column: int(64), - }, - End: ast.Location{ - Line: int(1385), - Column: int(77), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1385), - Column: int(64), - }, - End: ast.Location{ - Line: int(1385), - Column: int(72), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1385), - Column: int(64), - }, - End: ast.Location{ - Line: int(1385), - Column: int(67), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "type", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1385), - Column: int(73), - }, - End: ast.Location{ - Line: int(1385), - Column: int(76), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15070, - FreeVars: ast.Identifiers{ - "arr", - }, - }, - Id: "arr", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - }, - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - BranchFalse: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(7), - }, - End: ast.Location{ - Line: int(1387), - Column: int(81), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "arr", - "std", - "value", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(7), - }, - End: ast.Location{ - Line: int(1387), - Column: int(17), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15040, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(7), - }, - End: ast.Location{ - Line: int(1387), - Column: int(10), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "filter", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(18), - }, - End: ast.Location{ - Line: int(1387), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15083, - FreeVars: ast.Identifiers{ - "arr", - "value", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(27), - }, - End: ast.Location{ - Line: int(1387), - Column: int(28), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(30), - }, - End: ast.Location{ - Line: int(1387), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15087, - FreeVars: ast.Identifiers{ - "arr", - "i", - "value", - }, - }, - Left: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(30), - }, - End: ast.Location{ - Line: int(1387), - Column: int(36), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15087, - FreeVars: ast.Identifiers{ - "arr", - "i", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(30), - }, - End: ast.Location{ - Line: int(1387), - Column: int(33), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15087, - FreeVars: ast.Identifiers{ - "arr", - }, - }, - Id: "arr", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(34), - }, - End: ast.Location{ - Line: int(1387), - Column: int(35), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15087, - FreeVars: ast.Identifiers{ - "i", - }, - }, - Id: "i", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(40), - }, - End: ast.Location{ - Line: int(1387), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15087, - FreeVars: ast.Identifiers{ - "value", - }, - }, - Id: "value", - }, - }, - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(47), - }, - End: ast.Location{ - Line: int(1387), - Column: int(80), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15083, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(47), - }, - End: ast.Location{ - Line: int(1387), - Column: int(56), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15083, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(47), - }, - End: ast.Location{ - Line: int(1387), - Column: int(50), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "range", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(57), - }, - End: ast.Location{ - Line: int(1387), - Column: int(58), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15106, - FreeVars: nil, - }, - OriginalString: "0", - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(60), - }, - End: ast.Location{ - Line: int(1387), - Column: int(79), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15106, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Left: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(60), - }, - End: ast.Location{ - Line: int(1387), - Column: int(75), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15106, - FreeVars: ast.Identifiers{ - "arr", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(60), - }, - End: ast.Location{ - Line: int(1387), - Column: int(70), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15106, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(60), - }, - End: ast.Location{ - Line: int(1387), - Column: int(63), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "length", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(71), - }, - End: ast.Location{ - Line: int(1387), - Column: int(74), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15118, - FreeVars: ast.Identifiers{ - "arr", - }, - }, - Id: "arr", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(4), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1387), - Column: int(78), - }, - End: ast.Location{ - Line: int(1387), - Column: int(79), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15106, - FreeVars: nil, - }, - OriginalString: "1", - }, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1383), - Column: int(3), - }, - End: ast.Location{ - Line: int(1387), - Column: int(81), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__compare", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v1", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1391), - Column: int(13), - }, - End: ast.Location{ - Line: int(1391), - Column: int(15), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "v2", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1391), - Column: int(17), - }, - End: ast.Location{ - Line: int(1391), - Column: int(19), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(7), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "std", - "v1", - "v2", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "t1", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(18), - }, - End: ast.Location{ - Line: int(1392), - Column: int(30), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15131, - FreeVars: ast.Identifiers{ - "std", - "v1", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(18), - }, - End: ast.Location{ - Line: int(1392), - Column: int(26), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15131, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(18), - }, - End: ast.Location{ - Line: int(1392), - Column: int(21), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "type", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(27), - }, - End: ast.Location{ - Line: int(1392), - Column: int(29), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15140, - FreeVars: ast.Identifiers{ - "v1", - }, - }, - Id: "v1", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(13), - }, - End: ast.Location{ - Line: int(1392), - Column: int(30), - }, - File: p1, - }, - }, - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "t2", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(37), - }, - End: ast.Location{ - Line: int(1392), - Column: int(49), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15143, - FreeVars: ast.Identifiers{ - "std", - "v2", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(37), - }, - End: ast.Location{ - Line: int(1392), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15143, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(37), - }, - End: ast.Location{ - Line: int(1392), - Column: int(40), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "type", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(46), - }, - End: ast.Location{ - Line: int(1392), - Column: int(48), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15152, - FreeVars: ast.Identifiers{ - "v2", - }, - }, - Id: "v2", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1392), - Column: int(32), - }, - End: ast.Location{ - Line: int(1392), - Column: int(49), - }, - File: p1, - }, - }, - }, - Body: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1393), - Column: int(7), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "std", - "t1", - "t2", - "v1", - "v2", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1393), - Column: int(10), - }, - End: ast.Location{ - Line: int(1393), - Column: int(18), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - "t2", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1393), - Column: int(10), - }, - End: ast.Location{ - Line: int(1393), - Column: int(12), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Id: "t1", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(13), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1393), - Column: int(16), - }, - End: ast.Location{ - Line: int(1393), - Column: int(18), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t2", - }, - }, - Id: "t2", - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Error{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(9), - }, - End: ast.Location{ - Line: int(1394), - Column: int(77), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - "t2", - }, - }, - Expr: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(15), - }, - End: ast.Location{ - Line: int(1394), - Column: int(77), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - "t2", - }, - }, - Left: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(15), - }, - End: ast.Location{ - Line: int(1394), - Column: int(72), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(15), - }, - End: ast.Location{ - Line: int(1394), - Column: int(62), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(15), - }, - End: ast.Location{ - Line: int(1394), - Column: int(57), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: "Comparison requires matching types. Got ", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(60), - }, - End: ast.Location{ - Line: int(1394), - Column: int(62), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Id: "t1", - }, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(65), - }, - End: ast.Location{ - Line: int(1394), - Column: int(72), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: " and ", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1394), - Column: int(75), - }, - End: ast.Location{ - Line: int(1394), - Column: int(77), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t2", - }, - }, - Id: "t2", - }, - }, - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - BranchFalse: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1395), - Column: int(12), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "std", - "t1", - "v1", - "v2", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1395), - Column: int(15), - }, - End: ast.Location{ - Line: int(1395), - Column: int(28), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1395), - Column: int(15), - }, - End: ast.Location{ - Line: int(1395), - Column: int(17), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Id: "t1", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1395), - Column: int(21), - }, - End: ast.Location{ - Line: int(1395), - Column: int(28), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: "array", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1396), - Column: int(9), - }, - End: ast.Location{ - Line: int(1396), - Column: int(36), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "std", - "v1", - "v2", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1396), - Column: int(9), - }, - End: ast.Location{ - Line: int(1396), - Column: int(28), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1396), - Column: int(9), - }, - End: ast.Location{ - Line: int(1396), - Column: int(12), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__compare_array", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1396), - Column: int(29), - }, - End: ast.Location{ - Line: int(1396), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15196, - FreeVars: ast.Identifiers{ - "v1", - }, - }, - Id: "v1", - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1396), - Column: int(33), - }, - End: ast.Location{ - Line: int(1396), - Column: int(35), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15196, - FreeVars: ast.Identifiers{ - "v2", - }, - }, - Id: "v2", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - BranchFalse: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(12), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - "v1", - "v2", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(15), - }, - End: ast.Location{ - Line: int(1397), - Column: int(65), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(15), - }, - End: ast.Location{ - Line: int(1397), - Column: int(49), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(15), - }, - End: ast.Location{ - Line: int(1397), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(15), - }, - End: ast.Location{ - Line: int(1397), - Column: int(17), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Id: "t1", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(21), - }, - End: ast.Location{ - Line: int(1397), - Column: int(31), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: "function", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(18), - Right: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(35), - }, - End: ast.Location{ - Line: int(1397), - Column: int(49), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(35), - }, - End: ast.Location{ - Line: int(1397), - Column: int(37), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Id: "t1", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(41), - }, - End: ast.Location{ - Line: int(1397), - Column: int(49), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: "object", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - }, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(18), - Right: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(53), - }, - End: ast.Location{ - Line: int(1397), - Column: int(65), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(53), - }, - End: ast.Location{ - Line: int(1397), - Column: int(55), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Id: "t1", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1397), - Column: int(59), - }, - End: ast.Location{ - Line: int(1397), - Column: int(65), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: "bool", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Error{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1398), - Column: int(9), - }, - End: ast.Location{ - Line: int(1398), - Column: int(62), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Expr: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1398), - Column: int(15), - }, - End: ast.Location{ - Line: int(1398), - Column: int(62), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1398), - Column: int(15), - }, - End: ast.Location{ - Line: int(1398), - Column: int(37), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Left: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1398), - Column: int(15), - }, - End: ast.Location{ - Line: int(1398), - Column: int(32), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: "Values of type ", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1398), - Column: int(35), - }, - End: ast.Location{ - Line: int(1398), - Column: int(37), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "t1", - }, - }, - Id: "t1", - }, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1398), - Column: int(40), - }, - End: ast.Location{ - Line: int(1398), - Column: int(62), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Value: " are not comparable.", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - }, - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - BranchFalse: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1399), - Column: int(12), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v1", - "v2", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1399), - Column: int(15), - }, - End: ast.Location{ - Line: int(1399), - Column: int(22), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v1", - "v2", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1399), - Column: int(15), - }, - End: ast.Location{ - Line: int(1399), - Column: int(17), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v1", - }, - }, - Id: "v1", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(9), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1399), - Column: int(20), - }, - End: ast.Location{ - Line: int(1399), - Column: int(22), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v2", - }, - }, - Id: "v2", - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Unary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1399), - Column: int(28), - }, - End: ast.Location{ - Line: int(1399), - Column: int(30), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - Op: ast.UnaryOp(3), - Expr: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1399), - Column: int(29), - }, - End: ast.Location{ - Line: int(1399), - Column: int(30), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - OriginalString: "1", - }, - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - BranchFalse: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1400), - Column: int(12), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v1", - "v2", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1400), - Column: int(15), - }, - End: ast.Location{ - Line: int(1400), - Column: int(22), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v1", - "v2", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1400), - Column: int(15), - }, - End: ast.Location{ - Line: int(1400), - Column: int(17), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v1", - }, - }, - Id: "v1", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(7), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1400), - Column: int(20), - }, - End: ast.Location{ - Line: int(1400), - Column: int(22), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: ast.Identifiers{ - "v2", - }, - }, - Id: "v2", - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1400), - Column: int(28), - }, - End: ast.Location{ - Line: int(1400), - Column: int(29), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - OriginalString: "1", - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - BranchFalse: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1401), - Column: int(12), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15127, - FreeVars: nil, - }, - OriginalString: "0", - }, - }, - }, - }, - }, - }, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1391), - Column: int(3), - }, - End: ast.Location{ - Line: int(1401), - Column: int(13), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__compare_array", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr1", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1403), - Column: int(19), - }, - End: ast.Location{ - Line: int(1403), - Column: int(23), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr2", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1403), - Column: int(25), - }, - End: ast.Location{ - Line: int(1403), - Column: int(29), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(5), - }, - End: ast.Location{ - Line: int(1415), - Column: int(11), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - Ctx: p15262, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "std", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "len1", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(18), - }, - End: ast.Location{ - Line: int(1404), - Column: int(34), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15266, - FreeVars: ast.Identifiers{ - "arr1", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(18), - }, - End: ast.Location{ - Line: int(1404), - Column: int(28), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15266, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(18), - }, - End: ast.Location{ - Line: int(1404), - Column: int(21), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "length", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(29), - }, - End: ast.Location{ - Line: int(1404), - Column: int(33), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15275, - FreeVars: ast.Identifiers{ - "arr1", - }, - }, - Id: "arr1", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(11), - }, - End: ast.Location{ - Line: int(1404), - Column: int(34), - }, - File: p1, - }, - }, - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "len2", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(43), - }, - End: ast.Location{ - Line: int(1404), - Column: int(59), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15278, - FreeVars: ast.Identifiers{ - "arr2", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(43), - }, - End: ast.Location{ - Line: int(1404), - Column: int(53), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15278, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(43), - }, - End: ast.Location{ - Line: int(1404), - Column: int(46), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "length", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(54), - }, - End: ast.Location{ - Line: int(1404), - Column: int(58), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15287, - FreeVars: ast.Identifiers{ - "arr2", - }, - }, - Id: "arr2", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1404), - Column: int(36), - }, - End: ast.Location{ - Line: int(1404), - Column: int(59), - }, - File: p1, - }, - }, - }, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1405), - Column: int(5), - }, - End: ast.Location{ - Line: int(1415), - Column: int(11), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - Ctx: p15262, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "len1", - "len2", - "std", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "minLen", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1405), - Column: int(20), - }, - End: ast.Location{ - Line: int(1405), - Column: int(39), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15294, - FreeVars: ast.Identifiers{ - "len1", - "len2", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1405), - Column: int(20), - }, - End: ast.Location{ - Line: int(1405), - Column: int(27), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15294, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1405), - Column: int(20), - }, - End: ast.Location{ - Line: int(1405), - Column: int(23), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "min", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1405), - Column: int(28), - }, - End: ast.Location{ - Line: int(1405), - Column: int(32), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15303, - FreeVars: ast.Identifiers{ - "len1", - }, - }, - Id: "len1", - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1405), - Column: int(34), - }, - End: ast.Location{ - Line: int(1405), - Column: int(38), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15303, - FreeVars: ast.Identifiers{ - "len2", - }, - }, - Id: "len2", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1405), - Column: int(11), - }, - End: ast.Location{ - Line: int(1405), - Column: int(39), - }, - File: p1, - }, - }, - }, - Body: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1406), - Column: int(5), - }, - End: ast.Location{ - Line: int(1415), - Column: int(11), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - Ctx: p15262, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "len1", - "len2", - "minLen", - "std", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: nil, - Variable: "aux", - EqFodder: nil, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1406), - Column: int(11), - }, - End: ast.Location{ - Line: int(1414), - Column: int(34), - }, - File: p1, - }, - Fodder: nil, - Ctx: p15312, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "aux", - "len1", - "len2", - "minLen", - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1406), - Column: int(15), - }, - End: ast.Location{ - Line: int(1406), - Column: int(16), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1407), - Column: int(7), - }, - End: ast.Location{ - Line: int(1414), - Column: int(34), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "aux", - "i", - "len1", - "len2", - "minLen", - "std", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1407), - Column: int(10), - }, - End: ast.Location{ - Line: int(1407), - Column: int(20), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "i", - "minLen", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1407), - Column: int(10), - }, - End: ast.Location{ - Line: int(1407), - Column: int(11), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "i", - }, - }, - Id: "i", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(9), - Right: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1407), - Column: int(14), - }, - End: ast.Location{ - Line: int(1407), - Column: int(20), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "minLen", - }, - }, - Id: "minLen", - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Local{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(9), - }, - End: ast.Location{ - Line: int(1412), - Column: int(21), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "aux", - "i", - "std", - }, - }, - Binds: ast.LocalBinds{ - ast.LocalBind{ - VarFodder: ast.Fodder{}, - Variable: "cmpRes", - EqFodder: ast.Fodder{}, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(24), - }, - End: ast.Location{ - Line: int(1408), - Column: int(55), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15330, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "i", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(24), - }, - End: ast.Location{ - Line: int(1408), - Column: int(37), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15330, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(24), - }, - End: ast.Location{ - Line: int(1408), - Column: int(27), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__compare", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(38), - }, - End: ast.Location{ - Line: int(1408), - Column: int(45), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15339, - FreeVars: ast.Identifiers{ - "arr1", - "i", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(38), - }, - End: ast.Location{ - Line: int(1408), - Column: int(42), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15339, - FreeVars: ast.Identifiers{ - "arr1", - }, - }, - Id: "arr1", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(43), - }, - End: ast.Location{ - Line: int(1408), - Column: int(44), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15339, - FreeVars: ast.Identifiers{ - "i", - }, - }, - Id: "i", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(47), - }, - End: ast.Location{ - Line: int(1408), - Column: int(54), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15339, - FreeVars: ast.Identifiers{ - "arr2", - "i", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(47), - }, - End: ast.Location{ - Line: int(1408), - Column: int(51), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15339, - FreeVars: ast.Identifiers{ - "arr2", - }, - }, - Id: "arr2", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(52), - }, - End: ast.Location{ - Line: int(1408), - Column: int(53), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15339, - FreeVars: ast.Identifiers{ - "i", - }, - }, - Id: "i", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - Fun: nil, - CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1408), - Column: int(15), - }, - End: ast.Location{ - Line: int(1408), - Column: int(55), - }, - File: p1, - }, - }, - }, - Body: &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1409), - Column: int(9), - }, - End: ast.Location{ - Line: int(1412), - Column: int(21), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "aux", - "cmpRes", - "i", - }, - }, - Cond: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1409), - Column: int(12), - }, - End: ast.Location{ - Line: int(1409), - Column: int(23), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "cmpRes", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1409), - Column: int(12), - }, - End: ast.Location{ - Line: int(1409), - Column: int(18), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "cmpRes", - }, - }, - Id: "cmpRes", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(13), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1409), - Column: int(22), - }, - End: ast.Location{ - Line: int(1409), - Column: int(23), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: nil, - }, - OriginalString: "0", - }, - }, - ThenFodder: ast.Fodder{}, - BranchTrue: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1410), - Column: int(11), - }, - End: ast.Location{ - Line: int(1410), - Column: int(17), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(10), - Comment: []string{}, - }, - }, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "cmpRes", - }, - }, - Id: "cmpRes", - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - BranchFalse: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1412), - Column: int(11), - }, - End: ast.Location{ - Line: int(1412), - Column: int(21), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "aux", - "i", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1412), - Column: int(11), - }, - End: ast.Location{ - Line: int(1412), - Column: int(14), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(10), - Comment: []string{}, - }, - }, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "aux", - }, - }, - Id: "aux", - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1412), - Column: int(15), - }, - End: ast.Location{ - Line: int(1412), - Column: int(20), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15370, - FreeVars: ast.Identifiers{ - "i", - }, - }, - Left: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1412), - Column: int(15), - }, - End: ast.Location{ - Line: int(1412), - Column: int(16), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15370, - FreeVars: ast.Identifiers{ - "i", - }, - }, - Id: "i", - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(3), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1412), - Column: int(19), - }, - End: ast.Location{ - Line: int(1412), - Column: int(20), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15370, - FreeVars: nil, - }, - OriginalString: "1", - }, - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: true, - FodderRight: ast.Fodder{}, - TailStrictFodder: ast.Fodder{}, - }, - }, - }, - ElseFodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(6), - Comment: []string{}, - }, - }, - BranchFalse: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1414), - Column: int(9), - }, - End: ast.Location{ - Line: int(1414), - Column: int(34), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "len1", - "len2", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1414), - Column: int(9), - }, - End: ast.Location{ - Line: int(1414), - Column: int(22), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15317, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1414), - Column: int(9), - }, - End: ast.Location{ - Line: int(1414), - Column: int(12), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(8), - Comment: []string{}, - }, - }, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__compare", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1414), - Column: int(23), - }, - End: ast.Location{ - Line: int(1414), - Column: int(27), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15386, - FreeVars: ast.Identifiers{ - "len1", - }, - }, - Id: "len1", - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1414), - Column: int(29), - }, - End: ast.Location{ - Line: int(1414), - Column: int(33), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15386, - FreeVars: ast.Identifiers{ - "len2", - }, - }, - Id: "len2", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - }, - }, - Fun: nil, - CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - }, - }, - Body: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1415), - Column: int(5), - }, - End: ast.Location{ - Line: int(1415), - Column: int(11), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15262, - FreeVars: ast.Identifiers{ - "aux", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1415), - Column: int(5), - }, - End: ast.Location{ - Line: int(1415), - Column: int(8), - }, - File: p1, - }, - Fodder: ast.Fodder{ - ast.FodderElement{ - Kind: ast.FodderKind(0), - Blanks: int(0), - Indent: int(4), - Comment: []string{}, - }, - }, - Ctx: p15262, - FreeVars: ast.Identifiers{ - "aux", - }, - }, - Id: "aux", - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1415), - Column: int(9), - }, - End: ast.Location{ - Line: int(1415), - Column: int(10), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15397, - FreeVars: nil, - }, - OriginalString: "0", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - }, - }, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1403), - Column: int(3), - }, - End: ast.Location{ - Line: int(1415), - Column: int(11), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__array_less", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr1", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(16), - }, - End: ast.Location{ - Line: int(1417), - Column: int(20), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr2", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(22), - }, - End: ast.Location{ - Line: int(1417), - Column: int(26), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(30), - }, - End: ast.Location{ - Line: int(1417), - Column: int(67), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15403, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "std", - }, - }, - Left: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(30), - }, - End: ast.Location{ - Line: int(1417), - Column: int(61), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15403, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(30), - }, - End: ast.Location{ - Line: int(1417), - Column: int(49), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15403, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(30), - }, - End: ast.Location{ - Line: int(1417), - Column: int(33), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__compare_array", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(50), - }, - End: ast.Location{ - Line: int(1417), - Column: int(54), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15414, - FreeVars: ast.Identifiers{ - "arr1", - }, - }, - Id: "arr1", - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(56), - }, - End: ast.Location{ - Line: int(1417), - Column: int(60), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15414, - FreeVars: ast.Identifiers{ - "arr2", - }, - }, - Id: "arr2", - }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.Unary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(65), - }, - End: ast.Location{ - Line: int(1417), - Column: int(67), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15403, - FreeVars: nil, - }, - Op: ast.UnaryOp(3), - Expr: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(66), - }, - End: ast.Location{ - Line: int(1417), - Column: int(67), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15403, - FreeVars: nil, - }, - OriginalString: "1", - }, - }, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1417), - Column: int(3), - }, - End: ast.Location{ - Line: int(1417), - Column: int(67), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__array_greater", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr1", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(19), - }, - End: ast.Location{ - Line: int(1418), - Column: int(23), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr2", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(25), - }, - End: ast.Location{ - Line: int(1418), - Column: int(29), - }, - File: p1, - }, - }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(33), - }, - End: ast.Location{ - Line: int(1418), - Column: int(69), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15425, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "std", - }, - }, - Left: &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(33), - }, - End: ast.Location{ - Line: int(1418), - Column: int(64), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15425, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "std", - }, - }, - Target: &ast.Index{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(33), - }, - End: ast.Location{ - Line: int(1418), - Column: int(52), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15425, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Target: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(33), - }, - End: ast.Location{ - Line: int(1418), - Column: int(36), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: nil, - FreeVars: ast.Identifiers{ - "std", - }, - }, - Id: "std", - }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__compare_array", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(53), - }, - End: ast.Location{ - Line: int(1418), - Column: int(57), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15436, - FreeVars: ast.Identifiers{ - "arr1", - }, - }, - Id: "arr1", - }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(59), - }, - End: ast.Location{ - Line: int(1418), - Column: int(63), + FileName: "", + Begin: ast.Location{ + Line: int(1360), + Column: int(72), + }, + End: ast.Location{ + Line: int(1360), + Column: int(89), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14926, + FreeVars: ast.Identifiers{ + "pat_len", + "str_len", + }, + }, + Left: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1360), + Column: int(72), + }, + End: ast.Location{ + Line: int(1360), + Column: int(79), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14926, + FreeVars: ast.Identifiers{ + "str_len", + }, + }, + Id: "str_len", + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(4), + Right: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1360), + Column: int(82), + }, + End: ast.Location{ + Line: int(1360), + Column: int(89), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14926, + FreeVars: ast.Identifiers{ + "pat_len", + }, + }, + Id: "pat_len", + }, + }, + CommaFodder: nil, + }, + }, + Named: nil, + }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, + }, + CommaFodder: nil, }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15436, - FreeVars: ast.Identifiers{ - "arr2", }, + Named: nil, }, - Id: "arr2", + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(12), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(68), - }, - End: ast.Location{ - Line: int(1418), - Column: int(69), }, - File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p15425, - FreeVars: nil, }, - OriginalString: "1", }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1418), - Column: int(3), - }, - End: ast.Location{ - Line: int(1418), - Column: int(69), - }, - File: p1, - }, }, ast.DesugaredObjectField{ Hide: ast.ObjectFieldHide(0), @@ -194764,10 +178798,9 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "__array_less_or_equal", + Value: "find", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Body: &ast.Function{ NodeBase: ast.NodeBase{ @@ -194790,408 +178823,447 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr1", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1419), - Column: int(25), - }, - End: ast.Location{ - Line: int(1419), - Column: int(29), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr2", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1419), - Column: int(31), - }, - End: ast.Location{ - Line: int(1419), - Column: int(35), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "value", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "arr", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, - Body: &ast.Binary{ + Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1419), - Column: int(39), + Line: int(1363), + Column: int(5), }, End: ast.Location{ - Line: int(1419), - Column: int(75), + Line: int(1366), + Column: int(81), }, File: p1, }, - Fodder: ast.Fodder{}, - Ctx: p15446, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(4), + Comment: []string{}, + }, + }, + Ctx: p14939, FreeVars: ast.Identifiers{ - "arr1", - "arr2", + "arr", "std", + "value", }, }, - Left: &ast.Apply{ + Cond: &ast.Unary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1419), - Column: int(39), + Line: int(1363), + Column: int(8), }, End: ast.Location{ - Line: int(1419), - Column: int(70), + Line: int(1363), + Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15446, + Ctx: p14939, FreeVars: ast.Identifiers{ - "arr1", - "arr2", + "arr", "std", }, }, - Target: &ast.Index{ + Op: ast.UnaryOp(0), + Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1419), - Column: int(39), + Line: int(1363), + Column: int(9), }, End: ast.Location{ - Line: int(1419), - Column: int(58), + Line: int(1363), + Column: int(25), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15446, + Ctx: p14939, FreeVars: ast.Identifiers{ + "arr", "std", }, }, - Target: &ast.Var{ + Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1419), - Column: int(39), + Line: int(1363), + Column: int(9), }, End: ast.Location{ - Line: int(1419), - Column: int(42), + Line: int(1363), + Column: int(20), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: nil, + Ctx: p14939, FreeVars: ast.Identifiers{ "std", }, }, - Id: "std", + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1363), + Column: int(9), + }, + End: ast.Location{ + Line: int(1363), + Column: int(12), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "isArray", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, }, - LeftBracketFodder: ast.Fodder{}, - Index: &ast.LiteralString{ + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1363), + Column: int(21), + }, + End: ast.Location{ + Line: int(1363), + Column: int(24), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14952, + FreeVars: ast.Identifiers{ + "arr", + }, + }, + Id: "arr", + }, + CommaFodder: nil, + }, + }, + Named: nil, + }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, + }, + }, + ThenFodder: ast.Fodder{}, + BranchTrue: &ast.Error{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1364), + Column: int(7), + }, + End: ast.Location{ + Line: int(1364), + Column: int(77), + }, + File: p1, + }, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(6), + Comment: []string{}, + }, + }, + Ctx: p14939, + FreeVars: ast.Identifiers{ + "arr", + "std", + }, + }, + Expr: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1364), + Column: int(13), + }, + End: ast.Location{ + Line: int(1364), + Column: int(77), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14939, + FreeVars: ast.Identifiers{ + "arr", + "std", + }, + }, + Left: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(1364), + Column: int(13), }, End: ast.Location{ - Line: int(0), - Column: int(0), + Line: int(1364), + Column: int(61), }, - File: nil, + File: p1, }, - Fodder: nil, - Ctx: nil, + Fodder: ast.Fodder{}, + Ctx: p14939, FreeVars: nil, }, - Value: "__compare_array", + Value: "find second parameter should be an array, got ", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, - Id: nil, - }, - FodderLeft: ast.Fodder{}, - Arguments: ast.Arguments{ - Positional: []ast.CommaSeparatedExpr{ - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(3), + Right: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1364), + Column: int(64), + }, + End: ast.Location{ + Line: int(1364), + Column: int(77), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14939, + FreeVars: ast.Identifiers{ + "arr", + "std", + }, + }, + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1364), + Column: int(64), + }, + End: ast.Location{ + Line: int(1364), + Column: int(72), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14939, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1419), - Column: int(59), + Line: int(1364), + Column: int(64), }, End: ast.Location{ - Line: int(1419), - Column: int(63), + Line: int(1364), + Column: int(67), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15457, + Ctx: nil, FreeVars: ast.Identifiers{ - "arr1", + "std", }, }, - Id: "arr1", + Id: "std", }, - CommaFodder: ast.Fodder{}, - }, - ast.CommaSeparatedExpr{ - Expr: &ast.Var{ + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1419), - Column: int(65), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1419), - Column: int(69), + Line: int(0), + Column: int(0), }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15457, - FreeVars: ast.Identifiers{ - "arr2", + File: nil, }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, }, - Id: "arr2", + Value: "type", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", }, - CommaFodder: nil, - }, - }, - Named: nil, - }, - TrailingComma: false, - TailStrict: false, - FodderRight: ast.Fodder{}, - TailStrictFodder: nil, - }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(10), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1419), - Column: int(74), + RightBracketFodder: nil, + Id: nil, }, - End: ast.Location{ - Line: int(1419), - Column: int(75), + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1364), + Column: int(73), + }, + End: ast.Location{ + Line: int(1364), + Column: int(76), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14969, + FreeVars: ast.Identifiers{ + "arr", + }, + }, + Id: "arr", + }, + CommaFodder: nil, + }, + }, + Named: nil, }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15446, - FreeVars: nil, - }, - OriginalString: "0", - }, - }, - }, - PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1419), - Column: int(3), - }, - End: ast.Location{ - Line: int(1419), - Column: int(75), - }, - File: p1, - }, - }, - ast.DesugaredObjectField{ - Hide: ast.ObjectFieldHide(0), - Name: &ast.LiteralString{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: nil, - FreeVars: nil, - }, - Value: "__array_greater_or_equal", - Kind: ast.LiteralStringKind(1), - BlockIndent: "", - BlockTermIndent: "", - }, - Body: &ast.Function{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, - Fodder: nil, - Ctx: p11, - FreeVars: ast.Identifiers{ - "std", - }, - }, - ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr1", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1420), - Column: int(28), - }, - End: ast.Location{ - Line: int(1420), - Column: int(32), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "arr2", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1420), - Column: int(34), - }, - End: ast.Location{ - Line: int(1420), - Column: int(38), + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, - File: p1, }, }, - }, - TrailingComma: false, - ParenRightFodder: ast.Fodder{}, - Body: &ast.Binary{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1420), - Column: int(42), - }, - End: ast.Location{ - Line: int(1420), - Column: int(78), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15467, - FreeVars: ast.Identifiers{ - "arr1", - "arr2", - "std", + ElseFodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(4), + Comment: []string{}, }, }, - Left: &ast.Apply{ + BranchFalse: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1420), - Column: int(42), + Line: int(1366), + Column: int(7), }, End: ast.Location{ - Line: int(1420), - Column: int(73), + Line: int(1366), + Column: int(81), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15467, + Ctx: p14939, FreeVars: ast.Identifiers{ - "arr1", - "arr2", + "arr", "std", + "value", }, }, Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1420), - Column: int(42), + Line: int(1366), + Column: int(7), }, End: ast.Location{ - Line: int(1420), - Column: int(61), + Line: int(1366), + Column: int(17), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15467, + Ctx: p14939, FreeVars: ast.Identifiers{ "std", }, @@ -195199,18 +179271,25 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1420), - Column: int(42), + Line: int(1366), + Column: int(7), }, End: ast.Location{ - Line: int(1420), - Column: int(45), + Line: int(1366), + Column: int(10), }, File: p1, }, - Fodder: ast.Fodder{}, + Fodder: ast.Fodder{ + ast.FodderElement{ + Kind: ast.FodderKind(0), + Blanks: int(0), + Indent: int(6), + Comment: []string{}, + }, + }, Ctx: nil, FreeVars: ast.Identifiers{ "std", @@ -195237,64 +179316,467 @@ var _StdAst = &ast.DesugaredObject{ Ctx: nil, FreeVars: nil, }, - Value: "__compare_array", + Value: "filter", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, Arguments: ast.Arguments{ Positional: []ast.CommaSeparatedExpr{ ast.CommaSeparatedExpr{ - Expr: &ast.Var{ + Expr: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1420), - Column: int(62), + Line: int(1366), + Column: int(18), }, End: ast.Location{ - Line: int(1420), - Column: int(66), + Line: int(1366), + Column: int(45), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15478, + Ctx: p14982, FreeVars: ast.Identifiers{ - "arr1", + "arr", + "value", + }, + }, + ParenLeftFodder: ast.Fodder{}, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "i", + CommaFodder: nil, + }, + }, + Optional: nil, + }, + TrailingComma: false, + ParenRightFodder: ast.Fodder{}, + Body: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(30), + }, + End: ast.Location{ + Line: int(1366), + Column: int(45), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14986, + FreeVars: ast.Identifiers{ + "arr", + "i", + "value", + }, + }, + Left: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(30), + }, + End: ast.Location{ + Line: int(1366), + Column: int(36), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14986, + FreeVars: ast.Identifiers{ + "arr", + "i", + }, + }, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(30), + }, + End: ast.Location{ + Line: int(1366), + Column: int(33), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14986, + FreeVars: ast.Identifiers{ + "arr", + }, + }, + Id: "arr", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(34), + }, + End: ast.Location{ + Line: int(1366), + Column: int(35), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14986, + FreeVars: ast.Identifiers{ + "i", + }, + }, + Id: "i", + }, + RightBracketFodder: ast.Fodder{}, + Id: nil, + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(12), + Right: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(40), + }, + End: ast.Location{ + Line: int(1366), + Column: int(45), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14986, + FreeVars: ast.Identifiers{ + "value", + }, + }, + Id: "value", }, }, - Id: "arr1", }, CommaFodder: ast.Fodder{}, }, ast.CommaSeparatedExpr{ - Expr: &ast.Var{ + Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1420), - Column: int(68), + Line: int(1366), + Column: int(47), }, End: ast.Location{ - Line: int(1420), - Column: int(72), + Line: int(1366), + Column: int(80), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15478, + Ctx: p14982, FreeVars: ast.Identifiers{ - "arr2", + "arr", + "std", + }, + }, + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(47), + }, + End: ast.Location{ + Line: int(1366), + Column: int(56), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p14982, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(47), + }, + End: ast.Location{ + Line: int(1366), + Column: int(50), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "range", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, + }, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.LiteralNumber{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(57), + }, + End: ast.Location{ + Line: int(1366), + Column: int(58), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p15005, + FreeVars: nil, + }, + Value: float64(0), + OriginalString: "0", + }, + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedExpr{ + Expr: &ast.Binary{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(60), + }, + End: ast.Location{ + Line: int(1366), + Column: int(79), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p15005, + FreeVars: ast.Identifiers{ + "arr", + "std", + }, + }, + Left: &ast.Apply{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(60), + }, + End: ast.Location{ + Line: int(1366), + Column: int(75), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p15005, + FreeVars: ast.Identifiers{ + "arr", + "std", + }, + }, + Target: &ast.Index{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(60), + }, + End: ast.Location{ + Line: int(1366), + Column: int(70), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p15005, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Target: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(60), + }, + End: ast.Location{ + Line: int(1366), + Column: int(63), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: nil, + FreeVars: ast.Identifiers{ + "std", + }, + }, + Id: "std", + }, + LeftBracketFodder: ast.Fodder{}, + Index: &ast.LiteralString{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(0), + Column: int(0), + }, + End: ast.Location{ + Line: int(0), + Column: int(0), + }, + File: nil, + }, + Fodder: nil, + Ctx: nil, + FreeVars: nil, + }, + Value: "length", + Kind: ast.LiteralStringKind(1), + BlockIndent: "", + }, + RightBracketFodder: nil, + Id: nil, + }, + FodderLeft: ast.Fodder{}, + Arguments: ast.Arguments{ + Positional: []ast.CommaSeparatedExpr{ + ast.CommaSeparatedExpr{ + Expr: &ast.Var{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(71), + }, + End: ast.Location{ + Line: int(1366), + Column: int(74), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p15017, + FreeVars: ast.Identifiers{ + "arr", + }, + }, + Id: "arr", + }, + CommaFodder: nil, + }, + }, + Named: nil, + }, + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, + }, + OpFodder: ast.Fodder{}, + Op: ast.BinaryOp(4), + Right: &ast.LiteralNumber{ + NodeBase: ast.NodeBase{ + LocRange: ast.LocationRange{ + FileName: "", + Begin: ast.Location{ + Line: int(1366), + Column: int(78), + }, + End: ast.Location{ + Line: int(1366), + Column: int(79), + }, + File: p1, + }, + Fodder: ast.Fodder{}, + Ctx: p15005, + FreeVars: nil, + }, + Value: float64(1), + OriginalString: "1", + }, + }, + CommaFodder: nil, + }, }, + Named: nil, }, - Id: "arr2", + TrailingComma: false, + TailStrict: false, + FodderRight: ast.Fodder{}, + TailStrictFodder: nil, }, CommaFodder: nil, }, @@ -195306,43 +179788,9 @@ var _StdAst = &ast.DesugaredObject{ FodderRight: ast.Fodder{}, TailStrictFodder: nil, }, - OpFodder: ast.Fodder{}, - Op: ast.BinaryOp(8), - Right: &ast.LiteralNumber{ - NodeBase: ast.NodeBase{ - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1420), - Column: int(77), - }, - End: ast.Location{ - Line: int(1420), - Column: int(78), - }, - File: p1, - }, - Fodder: ast.Fodder{}, - Ctx: p15467, - FreeVars: nil, - }, - OriginalString: "0", - }, }, }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1420), - Column: int(3), - }, - End: ast.Location{ - Line: int(1420), - Column: int(78), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{ @@ -195353,7 +179801,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Self{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(25), Column: int(15), @@ -195365,24 +179813,12 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15485, + Ctx: p15022, FreeVars: nil, }, }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(25), - Column: int(9), - }, - End: ast.Location{ - Line: int(25), - Column: int(19), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: nil, @@ -195391,7 +179827,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(26), Column: int(14), @@ -195403,37 +179839,26 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15487, + Ctx: p15024, FreeVars: nil, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "x", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(26), - Column: int(23), - }, - End: ast.Location{ - Line: int(26), - Column: int(24), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "x", + CommaFodder: nil, }, }, + Optional: []ast.NamedParameter{}, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(26), Column: int(26), @@ -195445,7 +179870,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15490, + Ctx: p15027, FreeVars: ast.Identifiers{ "x", }, @@ -195455,18 +179880,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(26), - Column: int(9), - }, - End: ast.Location{ - Line: int(26), - Column: int(27), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: nil, @@ -195487,51 +179900,26 @@ var _StdAst = &ast.DesugaredObject{ File: nil, }, Fodder: nil, - Ctx: p15493, + Ctx: p15030, FreeVars: ast.Identifiers{ "std", }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "str", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(76), - Column: int(19), - }, - End: ast.Location{ - Line: int(76), - Column: int(22), - }, - File: p1, - }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "base", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(76), - Column: int(24), - }, - End: ast.Location{ - Line: int(76), - Column: int(28), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "str", + CommaFodder: ast.Fodder{}, + }, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "base", + CommaFodder: nil, }, }, + Optional: []ast.NamedParameter{}, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, @@ -195560,7 +179948,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(12), @@ -195572,7 +179960,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15499, + Ctx: p15036, FreeVars: ast.Identifiers{ "base", }, @@ -195580,7 +179968,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(12), @@ -195592,7 +179980,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15502, + Ctx: p15039, FreeVars: ast.Identifiers{ "base", }, @@ -195600,7 +179988,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(12), @@ -195612,7 +180000,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15505, + Ctx: p15042, FreeVars: ast.Identifiers{ "base", }, @@ -195624,7 +180012,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(19), @@ -195636,9 +180024,10 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15508, + Ctx: p15045, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -195647,7 +180036,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(24), @@ -195659,7 +180048,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15510, + Ctx: p15047, FreeVars: ast.Identifiers{ "base", }, @@ -195667,7 +180056,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(24), @@ -195679,7 +180068,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15513, + Ctx: p15050, FreeVars: ast.Identifiers{ "base", }, @@ -195691,7 +180080,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(32), @@ -195703,9 +180092,10 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15516, + Ctx: p15053, FreeVars: nil, }, + Value: float64(16), OriginalString: "16", }, }, @@ -195714,7 +180104,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(79), Column: int(5), @@ -195733,7 +180123,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, ast.FodderElement{ - Kind: ast.FodderKind(2), + Kind: ast.FodderKind(0), Blanks: int(0), Indent: int(4), Comment: []string{ @@ -195741,7 +180131,7 @@ var _StdAst = &ast.DesugaredObject{ }, }, }, - Ctx: p15520, + Ctx: p15057, FreeVars: ast.Identifiers{ "base", "std", @@ -195756,7 +180146,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(79), Column: int(23), @@ -195768,7 +180158,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15524, + Ctx: p15061, FreeVars: ast.Identifiers{ "std", }, @@ -195776,7 +180166,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(79), Column: int(23), @@ -195788,7 +180178,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15527, + Ctx: p15064, FreeVars: ast.Identifiers{ "std", }, @@ -195796,7 +180186,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(79), Column: int(23), @@ -195837,9 +180227,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -195849,7 +180238,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(79), Column: int(37), @@ -195861,13 +180250,12 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15534, + Ctx: p15071, FreeVars: nil, }, Value: "0", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -195881,24 +180269,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(79), - Column: int(11), - }, - End: ast.Location{ - Line: int(79), - Column: int(41), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(80), Column: int(5), @@ -195917,7 +180293,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15537, + Ctx: p15074, FreeVars: ast.Identifiers{ "base", "std", @@ -195933,7 +180309,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(80), Column: int(26), @@ -195945,7 +180321,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15541, + Ctx: p15078, FreeVars: ast.Identifiers{ "std", }, @@ -195953,7 +180329,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(80), Column: int(26), @@ -195965,7 +180341,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15544, + Ctx: p15081, FreeVars: ast.Identifiers{ "std", }, @@ -195973,7 +180349,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(80), Column: int(26), @@ -196014,9 +180390,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -196026,7 +180401,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(80), Column: int(40), @@ -196038,13 +180413,12 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15551, + Ctx: p15088, FreeVars: nil, }, Value: "A", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -196058,24 +180432,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(80), - Column: int(11), - }, - End: ast.Location{ - Line: int(80), - Column: int(44), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(81), Column: int(5), @@ -196094,7 +180456,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15554, + Ctx: p15091, FreeVars: ast.Identifiers{ "base", "std", @@ -196111,7 +180473,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(81), Column: int(26), @@ -196123,7 +180485,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15558, + Ctx: p15095, FreeVars: ast.Identifiers{ "std", }, @@ -196131,7 +180493,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(81), Column: int(26), @@ -196143,7 +180505,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15561, + Ctx: p15098, FreeVars: ast.Identifiers{ "std", }, @@ -196151,7 +180513,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(81), Column: int(26), @@ -196192,9 +180554,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -196204,7 +180565,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(81), Column: int(40), @@ -196216,13 +180577,12 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15568, + Ctx: p15105, FreeVars: nil, }, Value: "a", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -196236,24 +180596,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(81), - Column: int(11), - }, - End: ast.Location{ - Line: int(81), - Column: int(44), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(82), Column: int(5), @@ -196272,7 +180620,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15571, + Ctx: p15108, FreeVars: ast.Identifiers{ "base", "lower_a_code", @@ -196290,7 +180638,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Function{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(82), Column: int(11), @@ -196302,7 +180650,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: nil, - Ctx: p15575, + Ctx: p15112, FreeVars: ast.Identifiers{ "base", "lower_a_code", @@ -196313,52 +180661,27 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: ast.Fodder{}, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "aggregate", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(82), - Column: int(20), - }, - End: ast.Location{ - Line: int(82), - Column: int(29), - }, - File: p1, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "aggregate", + CommaFodder: ast.Fodder{}, }, - }, - ast.Parameter{ - NameFodder: ast.Fodder{}, - Name: "char", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(82), - Column: int(31), - }, - End: ast.Location{ - Line: int(82), - Column: int(35), - }, - File: p1, + ast.CommaSeparatedID{ + NameFodder: ast.Fodder{}, + Name: "char", + CommaFodder: nil, }, }, + Optional: []ast.NamedParameter{}, }, TrailingComma: false, ParenRightFodder: ast.Fodder{}, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(83), Column: int(7), @@ -196377,7 +180700,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15580, + Ctx: p15117, FreeVars: ast.Identifiers{ "aggregate", "base", @@ -196397,7 +180720,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(83), Column: int(20), @@ -196409,7 +180732,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15584, + Ctx: p15121, FreeVars: ast.Identifiers{ "char", "std", @@ -196418,7 +180741,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(83), Column: int(20), @@ -196430,7 +180753,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15587, + Ctx: p15124, FreeVars: ast.Identifiers{ "std", }, @@ -196438,7 +180761,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(83), Column: int(20), @@ -196479,9 +180802,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "codepoint", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -196491,7 +180813,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(83), Column: int(34), @@ -196503,7 +180825,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15594, + Ctx: p15131, FreeVars: ast.Identifiers{ "char", }, @@ -196522,24 +180844,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(83), - Column: int(13), - }, - End: ast.Location{ - Line: int(83), - Column: int(39), - }, - File: p1, - }, }, }, Body: &ast.Local{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(84), Column: int(7), @@ -196558,7 +180868,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15598, + Ctx: p15135, FreeVars: ast.Identifiers{ "aggregate", "base", @@ -196578,7 +180888,7 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(84), Column: int(21), @@ -196590,7 +180900,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15602, + Ctx: p15139, FreeVars: ast.Identifiers{ "code", "lower_a_code", @@ -196601,7 +180911,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(84), Column: int(24), @@ -196613,7 +180923,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15605, + Ctx: p15142, FreeVars: ast.Identifiers{ "code", "lower_a_code", @@ -196622,7 +180932,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(84), Column: int(24), @@ -196634,7 +180944,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15608, + Ctx: p15145, FreeVars: ast.Identifiers{ "code", }, @@ -196646,7 +180956,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(84), Column: int(32), @@ -196658,7 +180968,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15611, + Ctx: p15148, FreeVars: ast.Identifiers{ "lower_a_code", }, @@ -196670,7 +180980,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(85), Column: int(9), @@ -196682,7 +180992,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15614, + Ctx: p15151, FreeVars: ast.Identifiers{ "code", "lower_a_code", @@ -196691,7 +181001,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(85), Column: int(9), @@ -196703,7 +181013,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15617, + Ctx: p15154, FreeVars: ast.Identifiers{ "code", "lower_a_code", @@ -196712,7 +181022,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(85), Column: int(9), @@ -196731,7 +181041,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15621, + Ctx: p15158, FreeVars: ast.Identifiers{ "code", }, @@ -196743,7 +181053,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(85), Column: int(16), @@ -196755,7 +181065,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15624, + Ctx: p15161, FreeVars: ast.Identifiers{ "lower_a_code", }, @@ -196768,7 +181078,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(85), Column: int(31), @@ -196780,9 +181090,10 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15627, + Ctx: p15164, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -196797,7 +181108,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Conditional{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(86), Column: int(12), @@ -196809,7 +181120,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15630, + Ctx: p15167, FreeVars: ast.Identifiers{ "code", "upper_a_code", @@ -196819,7 +181130,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(86), Column: int(15), @@ -196831,7 +181142,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15633, + Ctx: p15170, FreeVars: ast.Identifiers{ "code", "upper_a_code", @@ -196840,7 +181151,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(86), Column: int(15), @@ -196852,7 +181163,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15636, + Ctx: p15173, FreeVars: ast.Identifiers{ "code", }, @@ -196864,7 +181175,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(86), Column: int(23), @@ -196876,7 +181187,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15639, + Ctx: p15176, FreeVars: ast.Identifiers{ "upper_a_code", }, @@ -196888,7 +181199,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(87), Column: int(9), @@ -196900,7 +181211,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15642, + Ctx: p15179, FreeVars: ast.Identifiers{ "code", "upper_a_code", @@ -196909,7 +181220,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(87), Column: int(9), @@ -196921,7 +181232,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15645, + Ctx: p15182, FreeVars: ast.Identifiers{ "code", "upper_a_code", @@ -196930,7 +181241,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(87), Column: int(9), @@ -196949,7 +181260,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15649, + Ctx: p15186, FreeVars: ast.Identifiers{ "code", }, @@ -196961,7 +181272,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(87), Column: int(16), @@ -196973,7 +181284,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15652, + Ctx: p15189, FreeVars: ast.Identifiers{ "upper_a_code", }, @@ -196986,7 +181297,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(87), Column: int(31), @@ -196998,9 +181309,10 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15655, + Ctx: p15192, FreeVars: nil, }, + Value: float64(10), OriginalString: "10", }, }, @@ -197015,7 +181327,7 @@ var _StdAst = &ast.DesugaredObject{ BranchFalse: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(89), Column: int(9), @@ -197027,7 +181339,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15658, + Ctx: p15195, FreeVars: ast.Identifiers{ "code", "zero_code", @@ -197036,7 +181348,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(89), Column: int(9), @@ -197055,7 +181367,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15662, + Ctx: p15199, FreeVars: ast.Identifiers{ "code", }, @@ -197067,7 +181379,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(89), Column: int(16), @@ -197079,7 +181391,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15665, + Ctx: p15202, FreeVars: ast.Identifiers{ "zero_code", }, @@ -197091,18 +181403,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: ast.Fodder{}, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(84), - Column: int(13), - }, - End: ast.Location{ - Line: int(89), - Column: int(25), - }, - File: p1, - }, }, }, Body: &ast.Conditional{ @@ -197132,7 +181432,7 @@ var _StdAst = &ast.DesugaredObject{ Cond: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(14), @@ -197144,7 +181444,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15670, + Ctx: p15207, FreeVars: ast.Identifiers{ "base", "digit", @@ -197153,7 +181453,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(14), @@ -197165,7 +181465,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15673, + Ctx: p15210, FreeVars: ast.Identifiers{ "digit", }, @@ -197173,7 +181473,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(14), @@ -197185,7 +181485,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15676, + Ctx: p15213, FreeVars: ast.Identifiers{ "digit", }, @@ -197197,7 +181497,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(23), @@ -197209,9 +181509,10 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15679, + Ctx: p15216, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, }, @@ -197220,7 +181521,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(28), @@ -197232,7 +181533,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15681, + Ctx: p15218, FreeVars: ast.Identifiers{ "base", "digit", @@ -197241,7 +181542,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(28), @@ -197253,7 +181554,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15684, + Ctx: p15221, FreeVars: ast.Identifiers{ "digit", }, @@ -197265,7 +181566,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(36), @@ -197277,7 +181578,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15687, + Ctx: p15224, FreeVars: ast.Identifiers{ "base", }, @@ -197290,7 +181591,7 @@ var _StdAst = &ast.DesugaredObject{ BranchTrue: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(91), Column: int(7), @@ -197302,7 +181603,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15690, + Ctx: p15227, FreeVars: ast.Identifiers{ "aggregate", "base", @@ -197312,7 +181613,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Binary{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(91), Column: int(7), @@ -197324,7 +181625,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15693, + Ctx: p15230, FreeVars: ast.Identifiers{ "aggregate", "base", @@ -197333,7 +181634,7 @@ var _StdAst = &ast.DesugaredObject{ Left: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(91), Column: int(7), @@ -197352,7 +181653,7 @@ var _StdAst = &ast.DesugaredObject{ Comment: []string{}, }, }, - Ctx: p15697, + Ctx: p15234, FreeVars: ast.Identifiers{ "base", }, @@ -197364,7 +181665,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(91), Column: int(14), @@ -197376,7 +181677,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15700, + Ctx: p15237, FreeVars: ast.Identifiers{ "aggregate", }, @@ -197389,7 +181690,7 @@ var _StdAst = &ast.DesugaredObject{ Right: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(91), Column: int(26), @@ -197401,7 +181702,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15703, + Ctx: p15240, FreeVars: ast.Identifiers{ "digit", }, @@ -197415,14 +181716,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(90), - Column: int(7), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(91), - Column: int(31), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -197437,14 +181738,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(90), - Column: int(43), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(90), - Column: int(86), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -197518,7 +181819,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -197530,7 +181830,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(43), @@ -197542,13 +181842,12 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15716, + Ctx: p15253, FreeVars: nil, }, Value: "%s is not a base %d integer", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -197556,7 +181855,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Array{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(75), @@ -197568,7 +181867,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15718, + Ctx: p15255, FreeVars: ast.Identifiers{ "base", "str", @@ -197579,7 +181878,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(76), @@ -197591,7 +181890,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15722, + Ctx: p15259, FreeVars: ast.Identifiers{ "str", }, @@ -197604,7 +181903,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(90), Column: int(81), @@ -197616,7 +181915,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15725, + Ctx: p15262, FreeVars: ast.Identifiers{ "base", }, @@ -197646,24 +181945,12 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, Body: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(5), @@ -197675,7 +181962,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15728, + Ctx: p15265, FreeVars: ast.Identifiers{ "addDigit", "std", @@ -197685,7 +181972,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(5), @@ -197697,7 +181984,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15731, + Ctx: p15268, FreeVars: ast.Identifiers{ "std", }, @@ -197705,7 +181992,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(5), @@ -197753,9 +182040,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "foldl", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -197765,7 +182051,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(15), @@ -197777,7 +182063,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15739, + Ctx: p15276, FreeVars: ast.Identifiers{ "addDigit", }, @@ -197790,7 +182076,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(25), @@ -197802,7 +182088,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15742, + Ctx: p15279, FreeVars: ast.Identifiers{ "std", "str", @@ -197811,7 +182097,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(25), @@ -197823,7 +182109,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15745, + Ctx: p15282, FreeVars: ast.Identifiers{ "std", }, @@ -197831,7 +182117,7 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(25), @@ -197872,9 +182158,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "stringChars", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -197884,7 +182169,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(41), @@ -197896,7 +182181,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15752, + Ctx: p15289, FreeVars: ast.Identifiers{ "str", }, @@ -197919,7 +182204,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(92), Column: int(47), @@ -197931,9 +182216,10 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15755, + Ctx: p15292, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: nil, @@ -197956,14 +182242,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(77), - Column: int(5), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(92), - Column: int(49), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -197977,14 +182263,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(77), - Column: int(37), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(77), - Column: int(69), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -198057,7 +182343,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "mod", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -198069,7 +182354,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(37), @@ -198081,13 +182366,12 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15767, + Ctx: p15304, FreeVars: nil, }, Value: "integer base %d invalid", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, CommaFodder: nil, }, @@ -198095,7 +182379,7 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ Line: int(77), Column: int(65), @@ -198107,7 +182391,7 @@ var _StdAst = &ast.DesugaredObject{ File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15769, + Ctx: p15306, FreeVars: ast.Identifiers{ "base", }, @@ -198129,18 +182413,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(76), - Column: int(9), - }, - End: ast.Location{ - Line: int(92), - Column: int(49), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: nil, @@ -198149,40 +182421,27 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.LiteralString{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1090), + Line: int(1075), Column: int(24), }, End: ast.Location{ - Line: int(1090), + Line: int(1075), Column: int(90), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15772, + Ctx: p15309, FreeVars: nil, }, Value: "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1090), - Column: int(9), - }, - End: ast.Location{ - Line: int(1090), - Column: int(90), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: nil, @@ -198193,14 +182452,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1091), - Column: int(22), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1091), - Column: int(72), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -198273,7 +182532,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "$objectFlatMerge", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -198287,14 +182545,14 @@ var _StdAst = &ast.DesugaredObject{ LocRange: ast.LocationRange{ FileName: "", Begin: ast.Location{ - Line: int(1091), - Column: int(22), + Line: int(0), + Column: int(0), }, End: ast.Location{ - Line: int(1091), - Column: int(72), + Line: int(0), + Column: int(0), }, - File: p1, + File: nil, }, Fodder: nil, Ctx: nil, @@ -198367,7 +182625,6 @@ var _StdAst = &ast.DesugaredObject{ Value: "flatMap", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, RightBracketFodder: nil, Id: nil, @@ -198397,26 +182654,15 @@ var _StdAst = &ast.DesugaredObject{ }, }, ParenLeftFodder: nil, - Parameters: []ast.Parameter{ - ast.Parameter{ - NameFodder: nil, - Name: "i", - EqFodder: nil, - DefaultArg: nil, - CommaFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{ + ast.CommaSeparatedID{ + NameFodder: nil, + Name: "i", + CommaFodder: nil, }, }, + Optional: nil, }, TrailingComma: false, ParenRightFodder: nil, @@ -198446,19 +182692,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.DesugaredObject{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(22), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(72), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15796, + Ctx: p15333, FreeVars: ast.Identifiers{ "base64_table", "i", @@ -198471,19 +182717,19 @@ var _StdAst = &ast.DesugaredObject{ Name: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(25), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(40), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15800, + Ctx: p15337, FreeVars: ast.Identifiers{ "base64_table", "i", @@ -198492,19 +182738,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(25), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(37), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15803, + Ctx: p15340, FreeVars: ast.Identifiers{ "base64_table", }, @@ -198515,19 +182761,19 @@ var _StdAst = &ast.DesugaredObject{ Index: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(38), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(39), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15806, + Ctx: p15343, FreeVars: ast.Identifiers{ "i", }, @@ -198540,19 +182786,19 @@ var _StdAst = &ast.DesugaredObject{ Body: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(43), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(44), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15809, + Ctx: p15346, FreeVars: ast.Identifiers{ "i", }, @@ -198560,18 +182806,6 @@ var _StdAst = &ast.DesugaredObject{ Id: "i", }, PlusSuper: false, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1091), - Column: int(24), - }, - End: ast.Location{ - Line: int(1091), - Column: int(44), - }, - File: p1, - }, }, }, Locals: ast.LocalBinds{}, @@ -198589,19 +182823,19 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.Apply{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(54), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(70), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15812, + Ctx: p15349, FreeVars: ast.Identifiers{ "std", }, @@ -198609,19 +182843,19 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Index{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(54), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(63), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15815, + Ctx: p15352, FreeVars: ast.Identifiers{ "std", }, @@ -198629,13 +182863,13 @@ var _StdAst = &ast.DesugaredObject{ Target: &ast.Var{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(54), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(57), }, File: p1, @@ -198670,9 +182904,8 @@ var _StdAst = &ast.DesugaredObject{ Value: "range", Kind: ast.LiteralStringKind(1), BlockIndent: "", - BlockTermIndent: "", }, - RightBracketFodder: ast.Fodder{}, + RightBracketFodder: nil, Id: nil, }, FodderLeft: ast.Fodder{}, @@ -198682,21 +182915,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(64), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(65), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15822, + Ctx: p15359, FreeVars: nil, }, + Value: float64(0), OriginalString: "0", }, CommaFodder: ast.Fodder{}, @@ -198705,21 +182939,22 @@ var _StdAst = &ast.DesugaredObject{ Expr: &ast.LiteralNumber{ NodeBase: ast.NodeBase{ LocRange: ast.LocationRange{ - FileName: "", + FileName: "", Begin: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(67), }, End: ast.Location{ - Line: int(1091), + Line: int(1076), Column: int(69), }, File: p1, }, Fodder: ast.Fodder{}, - Ctx: p15824, + Ctx: p15361, FreeVars: nil, }, + Value: float64(63), OriginalString: "63", }, CommaFodder: nil, @@ -198754,18 +182989,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(1091), - Column: int(9), - }, - End: ast.Location{ - Line: int(1091), - Column: int(72), - }, - File: p1, - }, }, ast.LocalBind{ VarFodder: nil, @@ -198792,18 +183015,6 @@ var _StdAst = &ast.DesugaredObject{ }, Fun: nil, CloseFodder: nil, - LocRange: ast.LocationRange{ - FileName: "", - Begin: ast.Location{ - Line: int(0), - Column: int(0), - }, - End: ast.Location{ - Line: int(0), - Column: int(0), - }, - File: nil, - }, }, }, } diff --git a/vendor/github.com/google/go-jsonnet/builtins.go b/vendor/github.com/google/go-jsonnet/builtins.go index f4bb5ebb..9d5d94b9 100644 --- a/vendor/github.com/google/go-jsonnet/builtins.go +++ b/vendor/github.com/google/go-jsonnet/builtins.go @@ -25,19 +25,17 @@ import ( "fmt" "math" "os" - "reflect" "sort" - "strconv" "strings" "github.com/google/go-jsonnet/ast" ) -func builtinPlus(i *interpreter, x, y value) (value, error) { +func builtinPlus(i *interpreter, trace traceElement, x, y value) (value, error) { // TODO(sbarzowski) perhaps a more elegant way to dispatch switch right := y.(type) { case valueString: - left, err := builtinToString(i, x) + left, err := builtinToString(i, trace, x) if err != nil { return nil, err } @@ -46,13 +44,13 @@ func builtinPlus(i *interpreter, x, y value) (value, error) { } switch left := x.(type) { case *valueNumber: - right, err := i.getNumber(y) + right, err := i.getNumber(y, trace) if err != nil { return nil, err } - return makeDoubleCheck(i, left.value+right.value) + return makeDoubleCheck(i, trace, left.value+right.value) case valueString: - right, err := builtinToString(i, y) + right, err := builtinToString(i, trace, y) if err != nil { return nil, err } @@ -62,153 +60,119 @@ func builtinPlus(i *interpreter, x, y value) (value, error) { case *valueObject: return makeValueExtendedObject(left, right), nil default: - return nil, i.typeErrorSpecific(y, &valueObject{}) + return nil, i.typeErrorSpecific(y, &valueObject{}, trace) } case *valueArray: - right, err := i.getArray(y) + right, err := i.getArray(y, trace) if err != nil { return nil, err } return concatArrays(left, right), nil default: - return nil, i.typeErrorGeneral(x) + return nil, i.typeErrorGeneral(x, trace) } } -func builtinMinus(i *interpreter, xv, yv value) (value, error) { - x, err := i.getNumber(xv) +func builtinMinus(i *interpreter, trace traceElement, xv, yv value) (value, error) { + x, err := i.getNumber(xv, trace) if err != nil { return nil, err } - y, err := i.getNumber(yv) + y, err := i.getNumber(yv, trace) if err != nil { return nil, err } - return makeDoubleCheck(i, x.value-y.value) + return makeDoubleCheck(i, trace, x.value-y.value) } -func builtinMult(i *interpreter, xv, yv value) (value, error) { - x, err := i.getNumber(xv) +func builtinMult(i *interpreter, trace traceElement, xv, yv value) (value, error) { + x, err := i.getNumber(xv, trace) if err != nil { return nil, err } - y, err := i.getNumber(yv) + y, err := i.getNumber(yv, trace) if err != nil { return nil, err } - return makeDoubleCheck(i, x.value*y.value) + return makeDoubleCheck(i, trace, x.value*y.value) } -func builtinDiv(i *interpreter, xv, yv value) (value, error) { - x, err := i.getNumber(xv) +func builtinDiv(i *interpreter, trace traceElement, xv, yv value) (value, error) { + x, err := i.getNumber(xv, trace) if err != nil { return nil, err } - y, err := i.getNumber(yv) + y, err := i.getNumber(yv, trace) if err != nil { return nil, err } if y.value == 0 { - return nil, i.Error("Division by zero.") + return nil, i.Error("Division by zero.", trace) } - return makeDoubleCheck(i, x.value/y.value) + return makeDoubleCheck(i, trace, x.value/y.value) } -func builtinModulo(i *interpreter, xv, yv value) (value, error) { - x, err := i.getNumber(xv) +func builtinModulo(i *interpreter, trace traceElement, xv, yv value) (value, error) { + x, err := i.getNumber(xv, trace) if err != nil { return nil, err } - y, err := i.getNumber(yv) + y, err := i.getNumber(yv, trace) if err != nil { return nil, err } if y.value == 0 { - return nil, i.Error("Division by zero.") + return nil, i.Error("Division by zero.", trace) } - return makeDoubleCheck(i, math.Mod(x.value, y.value)) + return makeDoubleCheck(i, trace, math.Mod(x.value, y.value)) } -func valueCmp(i *interpreter, x, y value) (int, error) { +func valueLess(i *interpreter, trace traceElement, x, yv value) (bool, error) { switch left := x.(type) { case *valueNumber: - right, err := i.getNumber(y) + right, err := i.getNumber(yv, trace) if err != nil { - return 0, err + return false, err } - return float64Cmp(left.value, right.value), nil + return left.value < right.value, nil case valueString: - right, err := i.getString(y) - if err != nil { - return 0, err - } - return stringCmp(left, right), nil - case *valueArray: - right, err := i.getArray(y) + right, err := i.getString(yv, trace) if err != nil { - return 0, err + return false, err } - return arrayCmp(i, left, right) + return stringLessThan(left, right), nil default: - return 0, i.typeErrorGeneral(x) - } -} - -func arrayCmp(i *interpreter, x, y *valueArray) (int, error) { - for index := 0; index < minInt(x.length(), y.length()); index++ { - left, err := x.index(i, index) - if err != nil { - return 0, err - } - right, err := y.index(i, index) - if err != nil { - return 0, err - } - cmp, err := valueCmp(i, left, right) - if err != nil { - return 0, err - } - if cmp != 0 { - return cmp, nil - } + return false, i.typeErrorGeneral(x, trace) } - return intCmp(x.length(), y.length()), nil } -func builtinLess(i *interpreter, x, y value) (value, error) { - r, err := valueCmp(i, x, y) - if err != nil { - return nil, err - } - return makeValueBoolean(r == -1), nil +func builtinLess(i *interpreter, trace traceElement, x, yv value) (value, error) { + b, err := valueLess(i, trace, x, yv) + return makeValueBoolean(b), err } -func builtinGreater(i *interpreter, x, y value) (value, error) { - r, err := valueCmp(i, x, y) - if err != nil { - return nil, err - } - return makeValueBoolean(r == 1), nil +func builtinGreater(i *interpreter, trace traceElement, x, y value) (value, error) { + return builtinLess(i, trace, y, x) } -func builtinGreaterEq(i *interpreter, x, y value) (value, error) { - r, err := valueCmp(i, x, y) +func builtinGreaterEq(i *interpreter, trace traceElement, x, y value) (value, error) { + res, err := builtinLess(i, trace, x, y) if err != nil { return nil, err } - return makeValueBoolean(r >= 0), nil + return res.(*valueBoolean).not(), nil } -func builtinLessEq(i *interpreter, x, y value) (value, error) { - r, err := valueCmp(i, x, y) +func builtinLessEq(i *interpreter, trace traceElement, x, y value) (value, error) { + res, err := builtinGreater(i, trace, x, y) if err != nil { return nil, err } - return makeValueBoolean(r <= 0), nil + return res.(*valueBoolean).not(), nil } -func builtinLength(i *interpreter, x value) (value, error) { +func builtinLength(i *interpreter, trace traceElement, x value) (value, error) { var num int switch x := x.(type) { case *valueObject: @@ -218,36 +182,31 @@ func builtinLength(i *interpreter, x value) (value, error) { case valueString: num = x.length() case *valueFunction: - for _, param := range x.parameters() { - if param.defaultArg == nil { - num++ - } - } + num = len(x.Parameters().required) default: - return nil, i.typeErrorGeneral(x) + return nil, i.typeErrorGeneral(x, trace) } return makeValueNumber(float64(num)), nil } -func builtinToString(i *interpreter, x value) (value, error) { +func builtinToString(i *interpreter, trace traceElement, x value) (value, error) { switch x := x.(type) { case valueString: return x, nil } var buf bytes.Buffer - err := i.manifestAndSerializeJSON(&buf, x, false, "") + err := i.manifestAndSerializeJSON(&buf, trace, x, false, "") if err != nil { return nil, err } return makeValueString(buf.String()), nil } -func builtinTrace(i *interpreter, x value, y value) (value, error) { - xStr, err := i.getString(x) +func builtinTrace(i *interpreter, trace traceElement, x value, y value) (value, error) { + xStr, err := i.getString(x, trace) if err != nil { return nil, err } - trace := i.stack.currentTrace filename := trace.loc.FileName line := trace.loc.Begin.Line fmt.Fprintf( @@ -265,12 +224,12 @@ type astMakeArrayElement struct { index int } -func builtinMakeArray(i *interpreter, szv, funcv value) (value, error) { - sz, err := i.getInt(szv) +func builtinMakeArray(i *interpreter, trace traceElement, szv, funcv value) (value, error) { + sz, err := i.getInt(szv, trace) if err != nil { return nil, err } - fun, err := i.getFunction(funcv) + fun, err := i.getFunction(funcv, trace) if err != nil { return nil, err } @@ -289,54 +248,41 @@ func builtinMakeArray(i *interpreter, szv, funcv value) (value, error) { return makeValueArray(elems), nil } -func builtinFlatMap(i *interpreter, funcv, arrv value) (value, error) { - fun, err := i.getFunction(funcv) +func builtinFlatMap(i *interpreter, trace traceElement, funcv, arrv value) (value, error) { + arr, err := i.getArray(arrv, trace) if err != nil { return nil, err } - switch arrv := arrv.(type) { - case *valueArray: - num := arrv.length() - // Start with capacity of the original array. - // This may spare us a few reallocations. - // TODO(sbarzowski) verify that it actually helps - elems := make([]*cachedThunk, 0, num) - for counter := 0; counter < num; counter++ { - returnedValue, err := fun.call(i, args(arrv.elements[counter])) - if err != nil { - return nil, err - } - returned, err := i.getArray(returnedValue) - if err != nil { - return nil, err - } - elems = append(elems, returned.elements...) + fun, err := i.getFunction(funcv, trace) + if err != nil { + return nil, err + } + num := arr.length() + // Start with capacity of the original array. + // This may spare us a few reallocations. + // TODO(sbarzowski) verify that it actually helps + elems := make([]*cachedThunk, 0, num) + for counter := 0; counter < num; counter++ { + returnedValue, err := fun.call(i, trace, args(arr.elements[counter])) + if err != nil { + return nil, err } - return makeValueArray(elems), nil - case valueString: - var str strings.Builder - for _, elem := range arrv.getRunes() { - returnedValue, err := fun.call(i, args(readyThunk(makeValueString(string(elem))))) - if err != nil { - return nil, err - } - returned, err := i.getString(returnedValue) - if err != nil { - return nil, err - } - str.WriteString(returned.getGoString()) + returned, err := i.getArray(returnedValue, trace) + if err != nil { + return nil, err + } + for _, elem := range returned.elements { + elems = append(elems, elem) } - return makeValueString(str.String()), nil - default: - return nil, i.Error("std.flatMap second param must be array / string, got " + arrv.getType().name) } + return makeValueArray(elems), nil } -func joinArrays(i *interpreter, sep *valueArray, arr *valueArray) (value, error) { +func joinArrays(i *interpreter, trace traceElement, sep *valueArray, arr *valueArray) (value, error) { result := make([]*cachedThunk, 0, arr.length()) first := true for _, elem := range arr.elements { - elemValue, err := i.evaluatePV(elem) + elemValue, err := i.evaluatePV(elem, trace) if err != nil { return nil, err } @@ -345,11 +291,15 @@ func joinArrays(i *interpreter, sep *valueArray, arr *valueArray) (value, error) continue case *valueArray: if !first { - result = append(result, sep.elements...) + for _, subElem := range sep.elements { + result = append(result, subElem) + } + } + for _, subElem := range v.elements { + result = append(result, subElem) } - result = append(result, v.elements...) default: - return nil, i.typeErrorSpecific(elemValue, &valueArray{}) + return nil, i.typeErrorSpecific(elemValue, &valueArray{}, trace) } first = false @@ -357,11 +307,11 @@ func joinArrays(i *interpreter, sep *valueArray, arr *valueArray) (value, error) return makeValueArray(result), nil } -func joinStrings(i *interpreter, sep valueString, arr *valueArray) (value, error) { +func joinStrings(i *interpreter, trace traceElement, sep valueString, arr *valueArray) (value, error) { result := make([]rune, 0, arr.length()) first := true for _, elem := range arr.elements { - elemValue, err := i.evaluatePV(elem) + elemValue, err := i.evaluatePV(elem, trace) if err != nil { return nil, err } @@ -374,51 +324,51 @@ func joinStrings(i *interpreter, sep valueString, arr *valueArray) (value, error } result = append(result, v.getRunes()...) default: - return nil, i.typeErrorSpecific(elemValue, emptyString()) + return nil, i.typeErrorSpecific(elemValue, emptyString(), trace) } first = false } return makeStringFromRunes(result), nil } -func builtinJoin(i *interpreter, sep, arrv value) (value, error) { - arr, err := i.getArray(arrv) +func builtinJoin(i *interpreter, trace traceElement, sep, arrv value) (value, error) { + arr, err := i.getArray(arrv, trace) if err != nil { return nil, err } switch sep := sep.(type) { case valueString: - return joinStrings(i, sep, arr) + return joinStrings(i, trace, sep, arr) case *valueArray: - return joinArrays(i, sep, arr) + return joinArrays(i, trace, sep, arr) default: - return nil, i.Error("join first parameter should be string or array, got " + sep.getType().name) + return nil, i.Error("join first parameter should be string or array, got "+sep.getType().name, trace) } } -func builtinReverse(i *interpreter, arrv value) (value, error) { - arr, err := i.getArray(arrv) +func builtinReverse(i *interpreter, trace traceElement, arrv value) (value, error) { + arr, err := i.getArray(arrv, trace) if err != nil { return nil, err } - lenArr := len(arr.elements) // lenx holds the original array length - reversedArray := make([]*cachedThunk, lenArr) // creates a slice that refer to a new array of length lenx + lenArr := len(arr.elements) // lenx holds the original array length + reversed_array := make([]*cachedThunk, lenArr) // creates a slice that refer to a new array of length lenx for i := 0; i < lenArr; i++ { j := lenArr - (i + 1) // j initially holds (lenx - 1) and decreases to 0 while i initially holds 0 and increase to (lenx - 1) - reversedArray[i] = arr.elements[j] + reversed_array[i] = arr.elements[j] } - return makeValueArray(reversedArray), nil + return makeValueArray(reversed_array), nil } -func builtinFilter(i *interpreter, funcv, arrv value) (value, error) { - arr, err := i.getArray(arrv) +func builtinFilter(i *interpreter, trace traceElement, funcv, arrv value) (value, error) { + arr, err := i.getArray(arrv, trace) if err != nil { return nil, err } - fun, err := i.getFunction(funcv) + fun, err := i.getFunction(funcv, trace) if err != nil { return nil, err } @@ -428,11 +378,11 @@ func builtinFilter(i *interpreter, funcv, arrv value) (value, error) { // TODO(sbarzowski) verify that it actually helps elems := make([]*cachedThunk, 0, num) for counter := 0; counter < num; counter++ { - includedValue, err := fun.call(i, args(arr.elements[counter])) + includedValue, err := fun.call(i, trace, args(arr.elements[counter])) if err != nil { return nil, err } - included, err := i.getBoolean(includedValue) + included, err := i.getBoolean(includedValue, trace) if err != nil { return nil, err } @@ -445,6 +395,7 @@ func builtinFilter(i *interpreter, funcv, arrv value) (value, error) { type sortData struct { i *interpreter + trace traceElement thunks []*cachedThunk keys []value err error @@ -455,12 +406,12 @@ func (d *sortData) Len() int { } func (d *sortData) Less(i, j int) bool { - r, err := valueCmp(d.i, d.keys[i], d.keys[j]) + b, err := valueLess(d.i, d.trace, d.keys[i], d.keys[j]) if err != nil { d.err = err panic("Error while comparing elements") } - return r == -1 + return b } func (d *sortData) Swap(i, j int) { @@ -480,45 +431,50 @@ func (d *sortData) Sort() (err error) { return } -func builtinSort(i *interpreter, arguments []value) (value, error) { +func arrayFromThunks(vs []value) *valueArray { + thunks := make([]*cachedThunk, len(vs)) + for i := range vs { + thunks[i] = readyThunk(vs[i]) + } + return makeValueArray(thunks) +} + +func builtinSort(i *interpreter, trace traceElement, arguments []value) (value, error) { arrv := arguments[0] keyFv := arguments[1] - arr, err := i.getArray(arrv) + arr, err := i.getArray(arrv, trace) if err != nil { return nil, err } - keyF, err := i.getFunction(keyFv) + keyF, err := i.getFunction(keyFv, trace) if err != nil { return nil, err } num := arr.length() - data := sortData{i: i, thunks: make([]*cachedThunk, num), keys: make([]value, num)} + data := sortData{i: i, trace: trace, thunks: make([]*cachedThunk, num), keys: make([]value, num)} for counter := 0; counter < num; counter++ { var err error data.thunks[counter] = arr.elements[counter] - data.keys[counter], err = keyF.call(i, args(arr.elements[counter])) + data.keys[counter], err = keyF.call(i, trace, args(arr.elements[counter])) if err != nil { return nil, err } } - err = data.Sort() - if err != nil { - return nil, err - } + data.Sort() return makeValueArray(data.thunks), nil } -func builtinRange(i *interpreter, fromv, tov value) (value, error) { - from, err := i.getInt(fromv) +func builtinRange(i *interpreter, trace traceElement, fromv, tov value) (value, error) { + from, err := i.getInt(fromv, trace) if err != nil { return nil, err } - to, err := i.getInt(tov) + to, err := i.getInt(tov, trace) if err != nil { return nil, err } @@ -529,16 +485,16 @@ func builtinRange(i *interpreter, fromv, tov value) (value, error) { return makeValueArray(elems), nil } -func builtinNegation(i *interpreter, x value) (value, error) { - b, err := i.getBoolean(x) +func builtinNegation(i *interpreter, trace traceElement, x value) (value, error) { + b, err := i.getBoolean(x, trace) if err != nil { return nil, err } return makeValueBoolean(!b.value), nil } -func builtinBitNeg(i *interpreter, x value) (value, error) { - n, err := i.getNumber(x) +func builtinBitNeg(i *interpreter, trace traceElement, x value) (value, error) { + n, err := i.getNumber(x, trace) if err != nil { return nil, err } @@ -546,12 +502,12 @@ func builtinBitNeg(i *interpreter, x value) (value, error) { return int64ToValue(^intValue), nil } -func builtinIdentity(i *interpreter, x value) (value, error) { +func builtinIdentity(i *interpreter, trace traceElement, x value) (value, error) { return x, nil } -func builtinUnaryPlus(i *interpreter, x value) (value, error) { - n, err := i.getNumber(x) +func builtinUnaryPlus(i *interpreter, trace traceElement, x value) (value, error) { + n, err := i.getNumber(x, trace) if err != nil { return nil, err } @@ -559,8 +515,8 @@ func builtinUnaryPlus(i *interpreter, x value) (value, error) { return makeValueNumber(n.value), nil } -func builtinUnaryMinus(i *interpreter, x value) (value, error) { - n, err := i.getNumber(x) +func builtinUnaryMinus(i *interpreter, trace traceElement, x value) (value, error) { + n, err := i.getNumber(x, trace) if err != nil { return nil, err } @@ -569,25 +525,25 @@ func builtinUnaryMinus(i *interpreter, x value) (value, error) { // TODO(sbarzowski) since we have a builtin implementation of equals it's no longer really // needed and we should deprecate it eventually -func primitiveEquals(i *interpreter, x, y value) (value, error) { +func primitiveEquals(i *interpreter, trace traceElement, x, y value) (value, error) { if x.getType() != y.getType() { return makeValueBoolean(false), nil } switch left := x.(type) { case *valueBoolean: - right, err := i.getBoolean(y) + right, err := i.getBoolean(y, trace) if err != nil { return nil, err } return makeValueBoolean(left.value == right.value), nil case *valueNumber: - right, err := i.getNumber(y) + right, err := i.getNumber(y, trace) if err != nil { return nil, err } return makeValueBoolean(left.value == right.value), nil case valueString: - right, err := i.getString(y) + right, err := i.getString(y, trace) if err != nil { return nil, err } @@ -595,33 +551,34 @@ func primitiveEquals(i *interpreter, x, y value) (value, error) { case *valueNull: return makeValueBoolean(true), nil case *valueFunction: - return nil, i.Error("Cannot test equality of functions") + return nil, i.Error("Cannot test equality of functions", trace) default: return nil, i.Error( - "primitiveEquals operates on primitive types, got " + x.getType().name, + "primitiveEquals operates on primitive types, got "+x.getType().name, + trace, ) } } -func rawEquals(i *interpreter, x, y value) (bool, error) { +func rawEquals(i *interpreter, trace traceElement, x, y value) (bool, error) { if x.getType() != y.getType() { return false, nil } switch left := x.(type) { case *valueBoolean: - right, err := i.getBoolean(y) + right, err := i.getBoolean(y, trace) if err != nil { return false, err } return left.value == right.value, nil case *valueNumber: - right, err := i.getNumber(y) + right, err := i.getNumber(y, trace) if err != nil { return false, err } return left.value == right.value, nil case valueString: - right, err := i.getString(y) + right, err := i.getString(y, trace) if err != nil { return false, err } @@ -629,7 +586,7 @@ func rawEquals(i *interpreter, x, y value) (bool, error) { case *valueNull: return true, nil case *valueArray: - right, err := i.getArray(y) + right, err := i.getArray(y, trace) if err != nil { return false, err } @@ -637,15 +594,15 @@ func rawEquals(i *interpreter, x, y value) (bool, error) { return false, nil } for j := range left.elements { - leftElem, err := i.evaluatePV(left.elements[j]) + leftElem, err := i.evaluatePV(left.elements[j], trace) if err != nil { return false, err } - rightElem, err := i.evaluatePV(right.elements[j]) + rightElem, err := i.evaluatePV(right.elements[j], trace) if err != nil { return false, err } - eq, err := rawEquals(i, leftElem, rightElem) + eq, err := rawEquals(i, trace, leftElem, rightElem) if err != nil { return false, err } @@ -655,7 +612,7 @@ func rawEquals(i *interpreter, x, y value) (bool, error) { } return true, nil case *valueObject: - right, err := i.getObject(y) + right, err := i.getObject(y, trace) if err != nil { return false, err } @@ -673,15 +630,15 @@ func rawEquals(i *interpreter, x, y value) (bool, error) { } for j := range leftFields { fieldName := leftFields[j] - leftField, err := left.index(i, fieldName) + leftField, err := left.index(i, trace, fieldName) if err != nil { return false, err } - rightField, err := right.index(i, fieldName) + rightField, err := right.index(i, trace, fieldName) if err != nil { return false, err } - eq, err := rawEquals(i, leftField, rightField) + eq, err := rawEquals(i, trace, leftField, rightField) if err != nil { return false, err } @@ -691,33 +648,33 @@ func rawEquals(i *interpreter, x, y value) (bool, error) { } return true, nil case *valueFunction: - return false, i.Error("Cannot test equality of functions") + return false, i.Error("Cannot test equality of functions", trace) } panic(fmt.Sprintf("Unhandled case in equals %#+v %#+v", x, y)) } -func builtinEquals(i *interpreter, x, y value) (value, error) { - eq, err := rawEquals(i, x, y) +func builtinEquals(i *interpreter, trace traceElement, x, y value) (value, error) { + eq, err := rawEquals(i, trace, x, y) if err != nil { return nil, err } return makeValueBoolean(eq), nil } -func builtinNotEquals(i *interpreter, x, y value) (value, error) { - eq, err := rawEquals(i, x, y) +func builtinNotEquals(i *interpreter, trace traceElement, x, y value) (value, error) { + eq, err := rawEquals(i, trace, x, y) if err != nil { return nil, err } return makeValueBoolean(!eq), nil } -func builtinType(i *interpreter, x value) (value, error) { +func builtinType(i *interpreter, trace traceElement, x value) (value, error) { return makeValueString(x.getType().name), nil } -func builtinMd5(i *interpreter, x value) (value, error) { - str, err := i.getString(x) +func builtinMd5(i *interpreter, trace traceElement, x value) (value, error) { + str, err := i.getString(x, trace) if err != nil { return nil, err } @@ -725,7 +682,7 @@ func builtinMd5(i *interpreter, x value) (value, error) { return makeValueString(hex.EncodeToString(hash[:])), nil } -func builtinBase64(i *interpreter, input value) (value, error) { +func builtinBase64(i *interpreter, trace traceElement, input value) (value, error) { var byteArr []byte var sanityCheck = func(v int) (string, bool) { @@ -739,57 +696,57 @@ func builtinBase64(i *interpreter, input value) (value, error) { switch input.(type) { case valueString: - vStr, err := i.getString(input) + vStr, err := i.getString(input, trace) if err != nil { return nil, err } - str := vStr.getGoString() - for _, r := range str { + runes := []rune(vStr.getGoString()) + for _, r := range runes { n := int(r) msg, ok := sanityCheck(n) if !ok { - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } } - byteArr = []byte(str) + byteArr = []byte(string(vStr.getGoString())) case *valueArray: - vArr, err := i.getArray(input) + vArr, err := i.getArray(input, trace) if err != nil { return nil, err } for _, cThunk := range vArr.elements { - cTv, err := cThunk.getValue(i) + cTv, err := cThunk.getValue(i, trace) if err != nil { return nil, err } - vInt, err := i.getInt(cTv) + vInt, err := i.getInt(cTv, trace) if err != nil { msg := fmt.Sprintf("base64 encountered a non-integer value in the array, got %s", cTv.getType().name) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } msg, ok := sanityCheck(vInt) if !ok { - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } byteArr = append(byteArr, byte(vInt)) } default: msg := fmt.Sprintf("base64 can only base64 encode strings / arrays of single bytes, got %s", input.getType().name) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } sEnc := base64.StdEncoding.EncodeToString(byteArr) return makeValueString(sEnc), nil } -func builtinEncodeUTF8(i *interpreter, x value) (value, error) { - str, err := i.getString(x) +func builtinEncodeUTF8(i *interpreter, trace traceElement, x value) (value, error) { + str, err := i.getString(x, trace) if err != nil { return nil, err } @@ -801,19 +758,19 @@ func builtinEncodeUTF8(i *interpreter, x value) (value, error) { return makeValueArray(elems), nil } -func builtinDecodeUTF8(i *interpreter, x value) (value, error) { - arr, err := i.getArray(x) +func builtinDecodeUTF8(i *interpreter, trace traceElement, x value) (value, error) { + arr, err := i.getArray(x, trace) if err != nil { return nil, err } bs := make([]byte, len(arr.elements)) // it will be longer if characters fall outside of ASCII for pos := range arr.elements { - v, err := i.evaluateInt(arr.elements[pos]) + v, err := i.evaluateInt(arr.elements[pos], trace) if err != nil { return nil, err } if v < 0 || v > 255 { - return nil, i.Error(fmt.Sprintf("Bytes must be integers in range [0, 255], got %d", v)) + return nil, i.Error(fmt.Sprintf("Bytes must be integers in range [0, 255], got %d", v), trace) } bs[pos] = byte(v) } @@ -824,47 +781,47 @@ func builtinDecodeUTF8(i *interpreter, x value) (value, error) { // https://en.wikipedia.org/wiki/Unicode#Architecture_and_terminology const codepointMax = 0x10FFFF -func builtinChar(i *interpreter, x value) (value, error) { - n, err := i.getNumber(x) +func builtinChar(i *interpreter, trace traceElement, x value) (value, error) { + n, err := i.getNumber(x, trace) if err != nil { return nil, err } if n.value > codepointMax { - return nil, i.Error(fmt.Sprintf("Invalid unicode codepoint, got %v", n.value)) + return nil, i.Error(fmt.Sprintf("Invalid unicode codepoint, got %v", n.value), trace) } else if n.value < 0 { - return nil, i.Error(fmt.Sprintf("Codepoints must be >= 0, got %v", n.value)) + return nil, i.Error(fmt.Sprintf("Codepoints must be >= 0, got %v", n.value), trace) } return makeValueString(string(rune(n.value))), nil } -func builtinCodepoint(i *interpreter, x value) (value, error) { - str, err := i.getString(x) +func builtinCodepoint(i *interpreter, trace traceElement, x value) (value, error) { + str, err := i.getString(x, trace) if err != nil { return nil, err } if str.length() != 1 { - return nil, i.Error(fmt.Sprintf("codepoint takes a string of length 1, got length %v", str.length())) + return nil, i.Error(fmt.Sprintf("codepoint takes a string of length 1, got length %v", str.length()), trace) } return makeValueNumber(float64(str.getRunes()[0])), nil } -func makeDoubleCheck(i *interpreter, x float64) (value, error) { +func makeDoubleCheck(i *interpreter, trace traceElement, x float64) (value, error) { if math.IsNaN(x) { - return nil, i.Error("Not a number") + return nil, i.Error("Not a number", trace) } if math.IsInf(x, 0) { - return nil, i.Error("Overflow") + return nil, i.Error("Overflow", trace) } return makeValueNumber(x), nil } -func liftNumeric(f func(float64) float64) func(*interpreter, value) (value, error) { - return func(i *interpreter, x value) (value, error) { - n, err := i.getNumber(x) +func liftNumeric(f func(float64) float64) func(*interpreter, traceElement, value) (value, error) { + return func(i *interpreter, trace traceElement, x value) (value, error) { + n, err := i.getNumber(x, trace) if err != nil { return nil, err } - return makeDoubleCheck(i, f(n.value)) + return makeDoubleCheck(i, trace, f(n.value)) } } @@ -894,43 +851,41 @@ var builtinExponent = liftNumeric(func(f float64) float64 { return float64(exponent) }) -func liftBitwise(f func(int64, int64) int64, positiveRightArg bool) func(*interpreter, value, value) (value, error) { - return func(i *interpreter, xv, yv value) (value, error) { - x, err := i.getNumber(xv) +func liftBitwise(f func(int64, int64) int64) func(*interpreter, traceElement, value, value) (value, error) { + return func(i *interpreter, trace traceElement, xv, yv value) (value, error) { + x, err := i.getNumber(xv, trace) if err != nil { return nil, err } - y, err := i.getNumber(yv) + y, err := i.getNumber(yv, trace) if err != nil { return nil, err } if x.value < math.MinInt64 || x.value > math.MaxInt64 { - msg := fmt.Sprintf("Bitwise operator argument %v outside of range [%v, %v]", x.value, int64(math.MinInt64), int64(math.MaxInt64)) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + msg := fmt.Sprintf("Bitwise operator argument %v outside of range [%v, %v]", x.value, math.MinInt64, math.MaxInt64) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } if y.value < math.MinInt64 || y.value > math.MaxInt64 { - msg := fmt.Sprintf("Bitwise operator argument %v outside of range [%v, %v]", y.value, int64(math.MinInt64), int64(math.MaxInt64)) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) - } - if positiveRightArg && y.value < 0 { - return nil, makeRuntimeError("Shift by negative exponent.", i.getCurrentStackTrace()) + msg := fmt.Sprintf("Bitwise operator argument %v outside of range [%v, %v]", y.value, math.MinInt64, math.MaxInt64) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } - return makeDoubleCheck(i, float64(f(int64(x.value), int64(y.value)))) + return makeDoubleCheck(i, trace, float64(f(int64(x.value), int64(y.value)))) } } -var builtinShiftL = liftBitwise(func(x, y int64) int64 { return x << uint(y%64) }, true) -var builtinShiftR = liftBitwise(func(x, y int64) int64 { return x >> uint(y%64) }, true) -var builtinBitwiseAnd = liftBitwise(func(x, y int64) int64 { return x & y }, false) -var builtinBitwiseOr = liftBitwise(func(x, y int64) int64 { return x | y }, false) -var builtinBitwiseXor = liftBitwise(func(x, y int64) int64 { return x ^ y }, false) +// TODO(sbarzowski) negative shifts +var builtinShiftL = liftBitwise(func(x, y int64) int64 { return x << uint(y) }) +var builtinShiftR = liftBitwise(func(x, y int64) int64 { return x >> uint(y) }) +var builtinBitwiseAnd = liftBitwise(func(x, y int64) int64 { return x & y }) +var builtinBitwiseOr = liftBitwise(func(x, y int64) int64 { return x | y }) +var builtinBitwiseXor = liftBitwise(func(x, y int64) int64 { return x ^ y }) -func builtinObjectFieldsEx(i *interpreter, objv, includeHiddenV value) (value, error) { - obj, err := i.getObject(objv) +func builtinObjectFieldsEx(i *interpreter, trace traceElement, objv, includeHiddenV value) (value, error) { + obj, err := i.getObject(objv, trace) if err != nil { return nil, err } - includeHidden, err := i.getBoolean(includeHiddenV) + includeHidden, err := i.getBoolean(includeHiddenV, trace) if err != nil { return nil, err } @@ -943,16 +898,16 @@ func builtinObjectFieldsEx(i *interpreter, objv, includeHiddenV value) (value, e return makeValueArray(elems), nil } -func builtinObjectHasEx(i *interpreter, objv value, fnamev value, includeHiddenV value) (value, error) { - obj, err := i.getObject(objv) +func builtinObjectHasEx(i *interpreter, trace traceElement, objv value, fnamev value, includeHiddenV value) (value, error) { + obj, err := i.getObject(objv, trace) if err != nil { return nil, err } - fname, err := i.getString(fnamev) + fname, err := i.getString(fnamev, trace) if err != nil { return nil, err } - includeHidden, err := i.getBoolean(includeHiddenV) + includeHidden, err := i.getBoolean(includeHiddenV, trace) if err != nil { return nil, err } @@ -961,52 +916,52 @@ func builtinObjectHasEx(i *interpreter, objv value, fnamev value, includeHiddenV return makeValueBoolean(hasField), nil } -func builtinPow(i *interpreter, basev value, expv value) (value, error) { - base, err := i.getNumber(basev) +func builtinPow(i *interpreter, trace traceElement, basev value, expv value) (value, error) { + base, err := i.getNumber(basev, trace) if err != nil { return nil, err } - exp, err := i.getNumber(expv) + exp, err := i.getNumber(expv, trace) if err != nil { return nil, err } - return makeDoubleCheck(i, math.Pow(base.value, exp.value)) + return makeDoubleCheck(i, trace, math.Pow(base.value, exp.value)) } -func builtinSubstr(i *interpreter, inputStr, inputFrom, inputLen value) (value, error) { - strV, err := i.getString(inputStr) +func builtinSubstr(i *interpreter, trace traceElement, inputStr, inputFrom, inputLen value) (value, error) { + strV, err := i.getString(inputStr, trace) if err != nil { msg := fmt.Sprintf("substr first parameter should be a string, got %s", inputStr.getType().name) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } - fromV, err := i.getNumber(inputFrom) + fromV, err := i.getNumber(inputFrom, trace) if err != nil { msg := fmt.Sprintf("substr second parameter should be a number, got %s", inputFrom.getType().name) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } if math.Mod(fromV.value, 1) != 0 { msg := fmt.Sprintf("substr second parameter should be an integer, got %f", fromV.value) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } - lenV, err := i.getNumber(inputLen) + lenV, err := i.getNumber(inputLen, trace) if err != nil { msg := fmt.Sprintf("substr third parameter should be a number, got %s", inputLen.getType().name) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } - lenInt, err := i.getInt(lenV) + lenInt, err := i.getInt(lenV, trace) if err != nil { msg := fmt.Sprintf("substr third parameter should be an integer, got %f", lenV.value) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } if lenInt < 0 { msg := fmt.Sprintf("substr third parameter should be greater than zero, got %d", lenInt) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } fromInt := int(fromV.value) @@ -1026,26 +981,26 @@ func builtinSubstr(i *interpreter, inputStr, inputFrom, inputLen value) (value, return makeValueString(string(runes[fromInt:endIndex])), nil } -func builtinSplitLimit(i *interpreter, strv, cv, maxSplitsV value) (value, error) { - str, err := i.getString(strv) +func builtinSplitLimit(i *interpreter, trace traceElement, strv, cv, maxSplitsV value) (value, error) { + str, err := i.getString(strv, trace) if err != nil { return nil, err } - c, err := i.getString(cv) + c, err := i.getString(cv, trace) if err != nil { return nil, err } - maxSplits, err := i.getInt(maxSplitsV) + maxSplits, err := i.getInt(maxSplitsV, trace) if err != nil { return nil, err } if maxSplits < -1 { - return nil, i.Error(fmt.Sprintf("std.splitLimit third parameter should be -1 or non-negative, got %v", maxSplits)) + return nil, i.Error(fmt.Sprintf("std.splitLimit third parameter should be -1 or non-negative, got %v", maxSplits), trace) } sStr := str.getGoString() sC := c.getGoString() if len(sC) != 1 { - return nil, i.Error(fmt.Sprintf("std.splitLimit second parameter should have length 1, got %v", len(sC))) + return nil, i.Error(fmt.Sprintf("std.splitLimit second parameter should have length 1, got %v", len(sC)), trace) } // the convention is slightly different from strings.splitN in Go (the meaning of non-negative values is shifted by one) @@ -1063,16 +1018,16 @@ func builtinSplitLimit(i *interpreter, strv, cv, maxSplitsV value) (value, error return makeValueArray(res), nil } -func builtinStrReplace(i *interpreter, strv, fromv, tov value) (value, error) { - str, err := i.getString(strv) +func builtinStrReplace(i *interpreter, trace traceElement, strv, fromv, tov value) (value, error) { + str, err := i.getString(strv, trace) if err != nil { return nil, err } - from, err := i.getString(fromv) + from, err := i.getString(fromv, trace) if err != nil { return nil, err } - to, err := i.getString(tov) + to, err := i.getString(tov, trace) if err != nil { return nil, err } @@ -1080,34 +1035,34 @@ func builtinStrReplace(i *interpreter, strv, fromv, tov value) (value, error) { sFrom := from.getGoString() sTo := to.getGoString() if len(sFrom) == 0 { - return nil, i.Error("'from' string must not be zero length.") + return nil, i.Error("'from' string must not be zero length.", trace) } return makeValueString(strings.Replace(sStr, sFrom, sTo, -1)), nil } -func base64DecodeGoBytes(i *interpreter, str string) ([]byte, error) { +func base64DecodeGoBytes(i *interpreter, trace traceElement, str string) ([]byte, error) { strLen := len(str) if strLen%4 != 0 { msg := fmt.Sprintf("input string appears not to be a base64 encoded string. Wrong length found (%d)", strLen) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } decodedBytes, err := base64.StdEncoding.DecodeString(str) if err != nil { - return nil, i.Error(fmt.Sprintf("failed to decode: %s", err)) + return nil, i.Error(fmt.Sprintf("failed to decode: %s", err), trace) } return decodedBytes, nil } -func builtinBase64DecodeBytes(i *interpreter, input value) (value, error) { - vStr, err := i.getString(input) +func builtinBase64DecodeBytes(i *interpreter, trace traceElement, input value) (value, error) { + vStr, err := i.getString(input, trace) if err != nil { msg := fmt.Sprintf("base64DecodeBytes requires a string, got %s", input.getType().name) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } - decodedBytes, err := base64DecodeGoBytes(i, vStr.getGoString()) + decodedBytes, err := base64DecodeGoBytes(i, trace, vStr.getGoString()) if err != nil { return nil, err } @@ -1120,14 +1075,14 @@ func builtinBase64DecodeBytes(i *interpreter, input value) (value, error) { return makeValueArray(res), nil } -func builtinBase64Decode(i *interpreter, input value) (value, error) { - vStr, err := i.getString(input) +func builtinBase64Decode(i *interpreter, trace traceElement, input value) (value, error) { + vStr, err := i.getString(input, trace) if err != nil { msg := fmt.Sprintf("base64DecodeBytes requires a string, got %s", input.getType().name) - return nil, makeRuntimeError(msg, i.getCurrentStackTrace()) + return nil, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } - decodedBytes, err := base64DecodeGoBytes(i, vStr.getGoString()) + decodedBytes, err := base64DecodeGoBytes(i, trace, vStr.getGoString()) if err != nil { return nil, err } @@ -1135,32 +1090,37 @@ func builtinBase64Decode(i *interpreter, input value) (value, error) { return makeValueString(string(decodedBytes)), nil } -func builtinUglyObjectFlatMerge(i *interpreter, x value) (value, error) { +func builtinUglyObjectFlatMerge(i *interpreter, trace traceElement, x value) (value, error) { // TODO(sbarzowski) consider keeping comprehensions in AST // It will probably be way less hacky, with better error messages and better performance - objarr, err := i.getArray(x) + objarr, err := i.getArray(x, trace) if err != nil { return nil, err } newFields := make(simpleObjectFieldMap) + var anyObj *simpleObject for _, elem := range objarr.elements { - obj, err := i.evaluateObject(elem) + obj, err := i.evaluateObject(elem, trace) if err != nil { return nil, err } - // starts getting ugly - we mess with object internals simpleObj := obj.uncached.(*simpleObject) - - if len(simpleObj.locals) > 0 { - panic("Locals should have been desugared in object comprehension.") - } - // there is only one field, really for fieldName, fieldVal := range simpleObj.fields { if _, alreadyExists := newFields[fieldName]; alreadyExists { - return nil, i.Error(duplicateFieldNameErrMsg(fieldName)) + return nil, i.Error(duplicateFieldNameErrMsg(fieldName), trace) + } + + // Here is the tricky part. Each field in a comprehension has different + // upValues, because for example in {[v]: v for v in ["x", "y", "z"] }, + // the v is different for each field. + // Yet, even though upValues are field-specific, they are shadowed by object locals, + // so we need to make holes to let them pass through + upValues := simpleObj.upValues + for _, l := range simpleObj.locals { + delete(upValues, l.name) } newFields[fieldName] = simpleObjectField{ @@ -1171,18 +1131,29 @@ func builtinUglyObjectFlatMerge(i *interpreter, x value) (value, error) { }, } } + anyObj = simpleObj + } + + var locals []objectLocal + var localUpValues bindingFrame + if len(objarr.elements) > 0 { + // another ugliness - we just take the locals of our last object, + // we assume that the locals are the same for each of merged objects + locals = anyObj.locals + // note that there are already holes for object locals + localUpValues = anyObj.upValues } return makeValueSimpleObject( - nil, + localUpValues, newFields, []unboundField{}, // No asserts allowed - nil, + locals, ), nil } -func builtinParseJSON(i *interpreter, str value) (value, error) { - sval, err := i.getString(str) +func builtinParseJSON(i *interpreter, trace traceElement, str value) (value, error) { + sval, err := i.getString(str, trace) if err != nil { return nil, err } @@ -1190,137 +1161,25 @@ func builtinParseJSON(i *interpreter, str value) (value, error) { var parsedJSON interface{} err = json.Unmarshal([]byte(s), &parsedJSON) if err != nil { - return nil, i.Error(fmt.Sprintf("failed to parse JSON: %v", err.Error())) - } - return jsonToValue(i, parsedJSON) -} - -func jsonEncode(v interface{}) (string, error) { - buf := new(bytes.Buffer) - enc := json.NewEncoder(buf) - enc.SetEscapeHTML(false) - err := enc.Encode(v) - if err != nil { - return "", err - } - - return strings.TrimRight(buf.String(), "\n"), nil -} - -// We have a very similar logic here /interpreter.go@v0.16.0#L695 and here: /interpreter.go@v0.16.0#L627 -// These should ideally be unified -// For backwards compatibility reasons, we are manually marshalling to json so we can control formatting -// In the future, it might be apt to use a library [pretty-printing] function -func builtinManifestJSONEx(i *interpreter, obj, indent value) (value, error) { - vindent, err := i.getString(indent) - if err != nil { - return nil, err - } - - sindent := vindent.getGoString() - - var path []string - - var aux func(ov value, path []string, cindent string) (string, error) - aux = func(ov value, path []string, cindent string) (string, error) { - if ov == nil { - fmt.Println("value is nil") - return "null", nil - } - - switch v := ov.(type) { - case *valueNull: - return "null", nil - case valueString: - jStr, err := jsonEncode(v.getGoString()) - if err != nil { - return "", i.Error(fmt.Sprintf("failed to marshal valueString to JSON: %v", err.Error())) - } - return jStr, nil - case *valueNumber: - return strconv.FormatFloat(v.value, 'f', -1, 64), nil - case *valueBoolean: - return fmt.Sprintf("%t", v.value), nil - case *valueFunction: - return "", i.Error(fmt.Sprintf("tried to manifest function at %s", path)) - case *valueArray: - newIndent := cindent + sindent - lines := []string{"[\n"} - - var arrayLines []string - for aI, cThunk := range v.elements { - cTv, err := cThunk.getValue(i) - if err != nil { - return "", err - } - - newPath := append(path, strconv.FormatInt(int64(aI), 10)) - s, err := aux(cTv, newPath, newIndent) - if err != nil { - return "", err - } - arrayLines = append(arrayLines, newIndent+s) - } - lines = append(lines, strings.Join(arrayLines, ",\n")) - lines = append(lines, "\n"+cindent+"]") - return strings.Join(lines, ""), nil - case *valueObject: - newIndent := cindent + sindent - lines := []string{"{\n"} - - fields := objectFields(v, withoutHidden) - sort.Strings(fields) - var objectLines []string - for _, fieldName := range fields { - fieldValue, err := v.index(i, fieldName) - if err != nil { - return "", err - } - - fieldNameMarshalled, err := jsonEncode(fieldName) - if err != nil { - return "", i.Error(fmt.Sprintf("failed to marshal object fieldname to JSON: %v", err.Error())) - } - - newPath := append(path, fieldName) - mvs, err := aux(fieldValue, newPath, newIndent) - if err != nil { - return "", err - } - - line := newIndent + string(fieldNameMarshalled) + ": " + mvs - objectLines = append(objectLines, line) - } - lines = append(lines, strings.Join(objectLines, ",\n")) - lines = append(lines, "\n"+cindent+"}") - return strings.Join(lines, ""), nil - default: - return "", i.Error(fmt.Sprintf("unknown type to marshal to JSON: %s", reflect.TypeOf(v))) - } - } - - finalString, err := aux(obj, path, "") - if err != nil { - return nil, err + return nil, i.Error(fmt.Sprintf("failed to parse JSON: %v", err.Error()), trace) } - - return makeValueString(finalString), nil + return jsonToValue(i, trace, parsedJSON) } -func builtinExtVar(i *interpreter, name value) (value, error) { - str, err := i.getString(name) +func builtinExtVar(i *interpreter, trace traceElement, name value) (value, error) { + str, err := i.getString(name, trace) if err != nil { return nil, err } index := str.getGoString() if pv, ok := i.extVars[index]; ok { - return i.evaluatePV(pv) + return i.evaluatePV(pv, trace) } - return nil, i.Error("Undefined external variable: " + string(index)) + return nil, i.Error("Undefined external variable: "+string(index), trace) } -func builtinNative(i *interpreter, name value) (value, error) { - str, err := i.getString(name) +func builtinNative(i *interpreter, trace traceElement, name value) (value, error) { + str, err := i.getString(name, trace) if err != nil { return nil, err } @@ -1338,224 +1197,200 @@ type builtin interface { Name() ast.Identifier } -func flattenArgs(args callArguments, params []namedParameter, defaults []value) []*cachedThunk { +func flattenArgs(args callArguments, params parameters, defaults []value) []*cachedThunk { positions := make(map[ast.Identifier]int) - for i, param := range params { - positions[param.name] = i + for i := 0; i < len(params.required); i++ { + positions[params.required[i]] = i + } + for i := 0; i < len(params.optional); i++ { + positions[params.optional[i].name] = i + len(params.required) } - flatArgs := make([]*cachedThunk, len(params)) + flatArgs := make([]*cachedThunk, len(params.required)+len(params.optional)) - // Bind positional arguments copy(flatArgs, args.positional) - // Bind named arguments for _, arg := range args.named { flatArgs[positions[arg.name]] = arg.pv } - // Bind defaults for unsatisfied named parameters - for i := range params { - if flatArgs[i] == nil { - flatArgs[i] = readyThunk(defaults[i]) + for i := 0; i < len(params.optional); i++ { + pos := len(params.required) + i + if flatArgs[pos] == nil { + flatArgs[pos] = readyThunk(defaults[i]) } } return flatArgs } -type unaryBuiltinFunc func(*interpreter, value) (value, error) +type unaryBuiltinFunc func(*interpreter, traceElement, value) (value, error) type unaryBuiltin struct { - name ast.Identifier - function unaryBuiltinFunc - params ast.Identifiers + name ast.Identifier + function unaryBuiltinFunc + parameters ast.Identifiers } -func (b *unaryBuiltin) evalCall(args callArguments, i *interpreter) (value, error) { - flatArgs := flattenArgs(args, b.parameters(), []value{}) +func getBuiltinTrace(trace traceElement, name ast.Identifier) traceElement { + context := "builtin function <" + string(name) + ">" + return traceElement{loc: trace.loc, context: &context} +} - x, err := flatArgs[0].getValue(i) +func (b *unaryBuiltin) evalCall(args callArguments, i *interpreter, trace traceElement) (value, error) { + flatArgs := flattenArgs(args, b.Parameters(), []value{}) + builtinTrace := getBuiltinTrace(trace, b.name) + x, err := flatArgs[0].getValue(i, trace) if err != nil { return nil, err } - return b.function(i, x) + return b.function(i, builtinTrace, x) } -func (b *unaryBuiltin) parameters() []namedParameter { - ret := make([]namedParameter, len(b.params)) - for i := range ret { - ret[i].name = b.params[i] - } - return ret +func (b *unaryBuiltin) Parameters() parameters { + return parameters{required: b.parameters} } func (b *unaryBuiltin) Name() ast.Identifier { return b.name } -type binaryBuiltinFunc func(*interpreter, value, value) (value, error) +type binaryBuiltinFunc func(*interpreter, traceElement, value, value) (value, error) type binaryBuiltin struct { - name ast.Identifier - function binaryBuiltinFunc - params ast.Identifiers + name ast.Identifier + function binaryBuiltinFunc + parameters ast.Identifiers } -func (b *binaryBuiltin) evalCall(args callArguments, i *interpreter) (value, error) { - flatArgs := flattenArgs(args, b.parameters(), []value{}) - - x, err := flatArgs[0].getValue(i) +func (b *binaryBuiltin) evalCall(args callArguments, i *interpreter, trace traceElement) (value, error) { + flatArgs := flattenArgs(args, b.Parameters(), []value{}) + builtinTrace := getBuiltinTrace(trace, b.name) + x, err := flatArgs[0].getValue(i, trace) if err != nil { return nil, err } - y, err := flatArgs[1].getValue(i) + y, err := flatArgs[1].getValue(i, trace) if err != nil { return nil, err } - return b.function(i, x, y) + return b.function(i, builtinTrace, x, y) } -func (b *binaryBuiltin) parameters() []namedParameter { - ret := make([]namedParameter, len(b.params)) - for i := range ret { - ret[i].name = b.params[i] - } - return ret +func (b *binaryBuiltin) Parameters() parameters { + return parameters{required: b.parameters} } func (b *binaryBuiltin) Name() ast.Identifier { return b.name } -type ternaryBuiltinFunc func(*interpreter, value, value, value) (value, error) +type ternaryBuiltinFunc func(*interpreter, traceElement, value, value, value) (value, error) type ternaryBuiltin struct { - name ast.Identifier - function ternaryBuiltinFunc - params ast.Identifiers + name ast.Identifier + function ternaryBuiltinFunc + parameters ast.Identifiers } -func (b *ternaryBuiltin) evalCall(args callArguments, i *interpreter) (value, error) { - flatArgs := flattenArgs(args, b.parameters(), []value{}) - - x, err := flatArgs[0].getValue(i) +func (b *ternaryBuiltin) evalCall(args callArguments, i *interpreter, trace traceElement) (value, error) { + flatArgs := flattenArgs(args, b.Parameters(), []value{}) + builtinTrace := getBuiltinTrace(trace, b.name) + x, err := flatArgs[0].getValue(i, trace) if err != nil { return nil, err } - y, err := flatArgs[1].getValue(i) + y, err := flatArgs[1].getValue(i, trace) if err != nil { return nil, err } - z, err := flatArgs[2].getValue(i) + z, err := flatArgs[2].getValue(i, trace) if err != nil { return nil, err } - return b.function(i, x, y, z) + return b.function(i, builtinTrace, x, y, z) } -func (b *ternaryBuiltin) parameters() []namedParameter { - ret := make([]namedParameter, len(b.params)) - for i := range ret { - ret[i].name = b.params[i] - } - return ret +func (b *ternaryBuiltin) Parameters() parameters { + return parameters{required: b.parameters} } func (b *ternaryBuiltin) Name() ast.Identifier { return b.name } -type generalBuiltinFunc func(*interpreter, []value) (value, error) - -type generalBuiltinParameter struct { - name ast.Identifier - // Note that the defaults are passed as values rather than AST nodes like in Parameters. - // This spares us unnecessary evaluation. - defaultValue value -} +type generalBuiltinFunc func(*interpreter, traceElement, []value) (value, error) // generalBuiltin covers cases that other builtin structures do not, // in particular it can have any number of parameters. It can also -// have optional parameters. The optional ones have non-nil defaultValues -// at the same index. +// have optional parameters. type generalBuiltin struct { name ast.Identifier - params []generalBuiltinParameter - function generalBuiltinFunc -} - -func (b *generalBuiltin) parameters() []namedParameter { - ret := make([]namedParameter, len(b.params)) - for i := range ret { - ret[i].name = b.params[i].name - if b.params[i].defaultValue != nil { - // This is not actually used because the defaultValue is used instead. - // The only reason we don't leave it nil is because the checkArguments - // function uses the non-nil status to indicate that the parameter - // is optional. - ret[i].defaultArg = &ast.LiteralNull{} - } - } - return ret + required ast.Identifiers + optional ast.Identifiers + // Note that the defaults are passed as values rather than AST nodes like in Parameters. + // This spares us unnecessary evaluation. + defaultValues []value + function generalBuiltinFunc } -func (b *generalBuiltin) defaultValues() []value { - ret := make([]value, len(b.params)) - for i := range ret { - ret[i] = b.params[i].defaultValue +func (b *generalBuiltin) Parameters() parameters { + optional := make([]namedParameter, len(b.optional)) + for i := range optional { + optional[i] = namedParameter{name: b.optional[i]} } - return ret + return parameters{required: b.required, optional: optional} } func (b *generalBuiltin) Name() ast.Identifier { return b.name } -func (b *generalBuiltin) evalCall(args callArguments, i *interpreter) (value, error) { - flatArgs := flattenArgs(args, b.parameters(), b.defaultValues()) +func (b *generalBuiltin) evalCall(args callArguments, i *interpreter, trace traceElement) (value, error) { + flatArgs := flattenArgs(args, b.Parameters(), b.defaultValues) + builtinTrace := getBuiltinTrace(trace, b.name) values := make([]value, len(flatArgs)) for j := 0; j < len(values); j++ { var err error - values[j], err = flatArgs[j].getValue(i) + values[j], err = flatArgs[j].getValue(i, trace) if err != nil { return nil, err } } - return b.function(i, values) + return b.function(i, builtinTrace, values) } // End of builtin utils -var builtinID = &unaryBuiltin{name: "id", function: builtinIdentity, params: ast.Identifiers{"x"}} +var builtinID = &unaryBuiltin{name: "id", function: builtinIdentity, parameters: ast.Identifiers{"x"}} var functionID = &valueFunction{ec: builtinID} var bopBuiltins = []*binaryBuiltin{ // Note that % and `in` are desugared instead of being handled here - ast.BopMult: &binaryBuiltin{name: "operator*", function: builtinMult, params: ast.Identifiers{"x", "y"}}, - ast.BopDiv: &binaryBuiltin{name: "operator/", function: builtinDiv, params: ast.Identifiers{"x", "y"}}, + ast.BopMult: &binaryBuiltin{name: "operator*", function: builtinMult, parameters: ast.Identifiers{"x", "y"}}, + ast.BopDiv: &binaryBuiltin{name: "operator/", function: builtinDiv, parameters: ast.Identifiers{"x", "y"}}, - ast.BopPlus: &binaryBuiltin{name: "operator+", function: builtinPlus, params: ast.Identifiers{"x", "y"}}, - ast.BopMinus: &binaryBuiltin{name: "operator-", function: builtinMinus, params: ast.Identifiers{"x", "y"}}, + ast.BopPlus: &binaryBuiltin{name: "operator+", function: builtinPlus, parameters: ast.Identifiers{"x", "y"}}, + ast.BopMinus: &binaryBuiltin{name: "operator-", function: builtinMinus, parameters: ast.Identifiers{"x", "y"}}, - ast.BopShiftL: &binaryBuiltin{name: "operator<<", function: builtinShiftL, params: ast.Identifiers{"x", "y"}}, - ast.BopShiftR: &binaryBuiltin{name: "operator>>", function: builtinShiftR, params: ast.Identifiers{"x", "y"}}, + ast.BopShiftL: &binaryBuiltin{name: "operator<<", function: builtinShiftL, parameters: ast.Identifiers{"x", "y"}}, + ast.BopShiftR: &binaryBuiltin{name: "operator>>", function: builtinShiftR, parameters: ast.Identifiers{"x", "y"}}, - ast.BopGreater: &binaryBuiltin{name: "operator>", function: builtinGreater, params: ast.Identifiers{"x", "y"}}, - ast.BopGreaterEq: &binaryBuiltin{name: "operator>=", function: builtinGreaterEq, params: ast.Identifiers{"x", "y"}}, - ast.BopLess: &binaryBuiltin{name: "operator<,", function: builtinLess, params: ast.Identifiers{"x", "y"}}, - ast.BopLessEq: &binaryBuiltin{name: "operator<=", function: builtinLessEq, params: ast.Identifiers{"x", "y"}}, + ast.BopGreater: &binaryBuiltin{name: "operator>", function: builtinGreater, parameters: ast.Identifiers{"x", "y"}}, + ast.BopGreaterEq: &binaryBuiltin{name: "operator>=", function: builtinGreaterEq, parameters: ast.Identifiers{"x", "y"}}, + ast.BopLess: &binaryBuiltin{name: "operator<,", function: builtinLess, parameters: ast.Identifiers{"x", "y"}}, + ast.BopLessEq: &binaryBuiltin{name: "operator<=", function: builtinLessEq, parameters: ast.Identifiers{"x", "y"}}, - ast.BopManifestEqual: &binaryBuiltin{name: "operator==", function: builtinEquals, params: ast.Identifiers{"x", "y"}}, - ast.BopManifestUnequal: &binaryBuiltin{name: "operator!=", function: builtinNotEquals, params: ast.Identifiers{"x", "y"}}, // Special case + ast.BopManifestEqual: &binaryBuiltin{name: "operator==", function: builtinEquals, parameters: ast.Identifiers{"x", "y"}}, + ast.BopManifestUnequal: &binaryBuiltin{name: "operator!=", function: builtinNotEquals, parameters: ast.Identifiers{"x", "y"}}, // Special case - ast.BopBitwiseAnd: &binaryBuiltin{name: "operator&", function: builtinBitwiseAnd, params: ast.Identifiers{"x", "y"}}, - ast.BopBitwiseXor: &binaryBuiltin{name: "operator^", function: builtinBitwiseXor, params: ast.Identifiers{"x", "y"}}, - ast.BopBitwiseOr: &binaryBuiltin{name: "operator|", function: builtinBitwiseOr, params: ast.Identifiers{"x", "y"}}, + ast.BopBitwiseAnd: &binaryBuiltin{name: "operator&", function: builtinBitwiseAnd, parameters: ast.Identifiers{"x", "y"}}, + ast.BopBitwiseXor: &binaryBuiltin{name: "operator^", function: builtinBitwiseXor, parameters: ast.Identifiers{"x", "y"}}, + ast.BopBitwiseOr: &binaryBuiltin{name: "operator|", function: builtinBitwiseOr, parameters: ast.Identifiers{"x", "y"}}, } var uopBuiltins = []*unaryBuiltin{ - ast.UopNot: &unaryBuiltin{name: "operator!", function: builtinNegation, params: ast.Identifiers{"x"}}, - ast.UopBitwiseNot: &unaryBuiltin{name: "operator~", function: builtinBitNeg, params: ast.Identifiers{"x"}}, - ast.UopPlus: &unaryBuiltin{name: "operator+ (unary)", function: builtinUnaryPlus, params: ast.Identifiers{"x"}}, - ast.UopMinus: &unaryBuiltin{name: "operator- (unary)", function: builtinUnaryMinus, params: ast.Identifiers{"x"}}, + ast.UopNot: &unaryBuiltin{name: "operator!", function: builtinNegation, parameters: ast.Identifiers{"x"}}, + ast.UopBitwiseNot: &unaryBuiltin{name: "operator~", function: builtinBitNeg, parameters: ast.Identifiers{"x"}}, + ast.UopPlus: &unaryBuiltin{name: "operator+ (unary)", function: builtinUnaryPlus, parameters: ast.Identifiers{"x"}}, + ast.UopMinus: &unaryBuiltin{name: "operator- (unary)", function: builtinUnaryMinus, parameters: ast.Identifiers{"x"}}, } func buildBuiltinMap(builtins []builtin) map[string]evalCallable { @@ -1568,52 +1403,51 @@ func buildBuiltinMap(builtins []builtin) map[string]evalCallable { var funcBuiltins = buildBuiltinMap([]builtin{ builtinID, - &unaryBuiltin{name: "extVar", function: builtinExtVar, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "length", function: builtinLength, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "toString", function: builtinToString, params: ast.Identifiers{"a"}}, - &binaryBuiltin{name: "trace", function: builtinTrace, params: ast.Identifiers{"str", "rest"}}, - &binaryBuiltin{name: "makeArray", function: builtinMakeArray, params: ast.Identifiers{"sz", "func"}}, - &binaryBuiltin{name: "flatMap", function: builtinFlatMap, params: ast.Identifiers{"func", "arr"}}, - &binaryBuiltin{name: "join", function: builtinJoin, params: ast.Identifiers{"sep", "arr"}}, - &unaryBuiltin{name: "reverse", function: builtinReverse, params: ast.Identifiers{"arr"}}, - &binaryBuiltin{name: "filter", function: builtinFilter, params: ast.Identifiers{"func", "arr"}}, - &binaryBuiltin{name: "range", function: builtinRange, params: ast.Identifiers{"from", "to"}}, - &binaryBuiltin{name: "primitiveEquals", function: primitiveEquals, params: ast.Identifiers{"x", "y"}}, - &binaryBuiltin{name: "equals", function: builtinEquals, params: ast.Identifiers{"x", "y"}}, - &binaryBuiltin{name: "objectFieldsEx", function: builtinObjectFieldsEx, params: ast.Identifiers{"obj", "hidden"}}, - &ternaryBuiltin{name: "objectHasEx", function: builtinObjectHasEx, params: ast.Identifiers{"obj", "fname", "hidden"}}, - &unaryBuiltin{name: "type", function: builtinType, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "char", function: builtinChar, params: ast.Identifiers{"n"}}, - &unaryBuiltin{name: "codepoint", function: builtinCodepoint, params: ast.Identifiers{"str"}}, - &unaryBuiltin{name: "ceil", function: builtinCeil, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "floor", function: builtinFloor, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "sqrt", function: builtinSqrt, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "sin", function: builtinSin, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "cos", function: builtinCos, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "tan", function: builtinTan, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "asin", function: builtinAsin, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "acos", function: builtinAcos, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "atan", function: builtinAtan, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "log", function: builtinLog, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "exp", function: builtinExp, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "mantissa", function: builtinMantissa, params: ast.Identifiers{"x"}}, - &unaryBuiltin{name: "exponent", function: builtinExponent, params: ast.Identifiers{"x"}}, - &binaryBuiltin{name: "pow", function: builtinPow, params: ast.Identifiers{"x", "n"}}, - &binaryBuiltin{name: "modulo", function: builtinModulo, params: ast.Identifiers{"x", "y"}}, - &unaryBuiltin{name: "md5", function: builtinMd5, params: ast.Identifiers{"s"}}, - &ternaryBuiltin{name: "substr", function: builtinSubstr, params: ast.Identifiers{"str", "from", "len"}}, - &ternaryBuiltin{name: "splitLimit", function: builtinSplitLimit, params: ast.Identifiers{"str", "c", "maxsplits"}}, - &ternaryBuiltin{name: "strReplace", function: builtinStrReplace, params: ast.Identifiers{"str", "from", "to"}}, - &unaryBuiltin{name: "base64Decode", function: builtinBase64Decode, params: ast.Identifiers{"str"}}, - &unaryBuiltin{name: "base64DecodeBytes", function: builtinBase64DecodeBytes, params: ast.Identifiers{"str"}}, - &unaryBuiltin{name: "parseJson", function: builtinParseJSON, params: ast.Identifiers{"str"}}, - &binaryBuiltin{name: "manifestJsonEx", function: builtinManifestJSONEx, params: ast.Identifiers{"value", "indent"}}, - &unaryBuiltin{name: "base64", function: builtinBase64, params: ast.Identifiers{"input"}}, - &unaryBuiltin{name: "encodeUTF8", function: builtinEncodeUTF8, params: ast.Identifiers{"str"}}, - &unaryBuiltin{name: "decodeUTF8", function: builtinDecodeUTF8, params: ast.Identifiers{"arr"}}, - &generalBuiltin{name: "sort", function: builtinSort, params: []generalBuiltinParameter{{name: "arr"}, {name: "keyF", defaultValue: functionID}}}, - &unaryBuiltin{name: "native", function: builtinNative, params: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "extVar", function: builtinExtVar, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "length", function: builtinLength, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "toString", function: builtinToString, parameters: ast.Identifiers{"a"}}, + &binaryBuiltin{name: "trace", function: builtinTrace, parameters: ast.Identifiers{"str", "rest"}}, + &binaryBuiltin{name: "makeArray", function: builtinMakeArray, parameters: ast.Identifiers{"sz", "func"}}, + &binaryBuiltin{name: "flatMap", function: builtinFlatMap, parameters: ast.Identifiers{"func", "arr"}}, + &binaryBuiltin{name: "join", function: builtinJoin, parameters: ast.Identifiers{"sep", "arr"}}, + &unaryBuiltin{name: "reverse", function: builtinReverse, parameters: ast.Identifiers{"arr"}}, + &binaryBuiltin{name: "filter", function: builtinFilter, parameters: ast.Identifiers{"func", "arr"}}, + &binaryBuiltin{name: "range", function: builtinRange, parameters: ast.Identifiers{"from", "to"}}, + &binaryBuiltin{name: "primitiveEquals", function: primitiveEquals, parameters: ast.Identifiers{"x", "y"}}, + &binaryBuiltin{name: "equals", function: builtinEquals, parameters: ast.Identifiers{"x", "y"}}, + &binaryBuiltin{name: "objectFieldsEx", function: builtinObjectFieldsEx, parameters: ast.Identifiers{"obj", "hidden"}}, + &ternaryBuiltin{name: "objectHasEx", function: builtinObjectHasEx, parameters: ast.Identifiers{"obj", "fname", "hidden"}}, + &unaryBuiltin{name: "type", function: builtinType, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "char", function: builtinChar, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "codepoint", function: builtinCodepoint, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "ceil", function: builtinCeil, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "floor", function: builtinFloor, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "sqrt", function: builtinSqrt, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "sin", function: builtinSin, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "cos", function: builtinCos, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "tan", function: builtinTan, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "asin", function: builtinAsin, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "acos", function: builtinAcos, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "atan", function: builtinAtan, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "log", function: builtinLog, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "exp", function: builtinExp, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "mantissa", function: builtinMantissa, parameters: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "exponent", function: builtinExponent, parameters: ast.Identifiers{"x"}}, + &binaryBuiltin{name: "pow", function: builtinPow, parameters: ast.Identifiers{"base", "exp"}}, + &binaryBuiltin{name: "modulo", function: builtinModulo, parameters: ast.Identifiers{"x", "y"}}, + &unaryBuiltin{name: "md5", function: builtinMd5, parameters: ast.Identifiers{"x"}}, + &ternaryBuiltin{name: "substr", function: builtinSubstr, parameters: ast.Identifiers{"str", "from", "len"}}, + &ternaryBuiltin{name: "splitLimit", function: builtinSplitLimit, parameters: ast.Identifiers{"str", "c", "maxsplits"}}, + &ternaryBuiltin{name: "strReplace", function: builtinStrReplace, parameters: ast.Identifiers{"str", "from", "to"}}, + &unaryBuiltin{name: "base64Decode", function: builtinBase64Decode, parameters: ast.Identifiers{"str"}}, + &unaryBuiltin{name: "base64DecodeBytes", function: builtinBase64DecodeBytes, parameters: ast.Identifiers{"str"}}, + &unaryBuiltin{name: "parseJson", function: builtinParseJSON, parameters: ast.Identifiers{"str"}}, + &unaryBuiltin{name: "base64", function: builtinBase64, parameters: ast.Identifiers{"input"}}, + &unaryBuiltin{name: "encodeUTF8", function: builtinEncodeUTF8, parameters: ast.Identifiers{"str"}}, + &unaryBuiltin{name: "decodeUTF8", function: builtinDecodeUTF8, parameters: ast.Identifiers{"arr"}}, + &generalBuiltin{name: "sort", function: builtinSort, required: ast.Identifiers{"arr"}, optional: ast.Identifiers{"keyF"}, defaultValues: []value{functionID}}, + &unaryBuiltin{name: "native", function: builtinNative, parameters: ast.Identifiers{"x"}}, // internal - &unaryBuiltin{name: "$objectFlatMerge", function: builtinUglyObjectFlatMerge, params: ast.Identifiers{"x"}}, + &unaryBuiltin{name: "$objectFlatMerge", function: builtinUglyObjectFlatMerge, parameters: ast.Identifiers{"x"}}, }) diff --git a/vendor/github.com/google/go-jsonnet/error_formatter.go b/vendor/github.com/google/go-jsonnet/error_formatter.go index f9a3976e..89c9d67e 100644 --- a/vendor/github.com/google/go-jsonnet/error_formatter.go +++ b/vendor/github.com/google/go-jsonnet/error_formatter.go @@ -69,7 +69,7 @@ func (ef *termErrorFormatter) Format(err error) string { case RuntimeError: return ef.formatRuntime(&err) case errors.StaticError: - return ef.formatStatic(err) + return ef.formatStatic(&err) default: return ef.formatInternal(err) } @@ -79,10 +79,10 @@ func (ef *termErrorFormatter) formatRuntime(err *RuntimeError) string { return err.Error() + "\n" + ef.buildStackTrace(err.StackTrace) } -func (ef *termErrorFormatter) formatStatic(err errors.StaticError) string { +func (ef *termErrorFormatter) formatStatic(err *errors.StaticError) string { var buf bytes.Buffer buf.WriteString(err.Error() + "\n") - ef.showCode(&buf, err.Loc()) + ef.showCode(&buf, err.Loc) return buf.String() } @@ -105,7 +105,7 @@ func (ef *termErrorFormatter) showCode(buf *bytes.Buffer, loc ast.LocationRange) beginning := ast.LineBeginning(&loc) ending := ast.LineEnding(&loc) fmt.Fprintf(buf, "%v", ef.sp.GetSnippet(beginning)) - errFprintf(buf, "%v", ef.sp.GetSnippet(loc)) //nolint:errcheck + errFprintf(buf, "%v", ef.sp.GetSnippet(loc)) fmt.Fprintf(buf, "%v", ef.sp.GetSnippet(ending)) buf.WriteByte('\n') } diff --git a/vendor/github.com/google/go-jsonnet/go.mod b/vendor/github.com/google/go-jsonnet/go.mod index ca5ff588..9907e94d 100644 --- a/vendor/github.com/google/go-jsonnet/go.mod +++ b/vendor/github.com/google/go-jsonnet/go.mod @@ -1,8 +1,12 @@ module github.com/google/go-jsonnet -go 1.13 - require ( - github.com/fatih/color v1.9.0 - github.com/sergi/go-diff v1.1.0 + github.com/fatih/color v1.7.0 + github.com/mattn/go-colorable v0.1.1 // indirect + github.com/mattn/go-isatty v0.0.7 // indirect + github.com/sergi/go-diff v1.0.0 + github.com/stretchr/testify v1.3.0 // indirect + golang.org/x/sys v0.0.0-20190531175056-4c3a928424d2 // indirect ) + +go 1.13 diff --git a/vendor/github.com/google/go-jsonnet/go.sum b/vendor/github.com/google/go-jsonnet/go.sum index ca45e232..3e679a81 100644 --- a/vendor/github.com/google/go-jsonnet/go.sum +++ b/vendor/github.com/google/go-jsonnet/go.sum @@ -1,31 +1,20 @@ +github.com/davecgh/go-spew v1.1.0 h1:ZDRjVQ15GmhC3fiQ8ni8+OwkZQO4DARzQgrnXU1Liz8= github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= -github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c= -github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= -github.com/fatih/color v1.9.0 h1:8xPHl4/q1VyqGIPif1F+1V3Y3lSmrq01EabUW3CoW5s= -github.com/fatih/color v1.9.0/go.mod h1:eQcE1qtQxscV5RaZvpXrrb8Drkc3/DdQ+uUYCNjL+zU= -github.com/kr/pretty v0.1.0 h1:L/CwN0zerZDmRFUapSPitk6f+Q3+0za1rQkzVuMiMFI= -github.com/kr/pretty v0.1.0/go.mod h1:dAy3ld7l9f0ibDNOQOHHMYYIIbhfbHSm3C4ZsoJORNo= -github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= -github.com/kr/text v0.1.0 h1:45sCR5RtlFHMR4UwH9sdQ5TC8v0qDQCHnXt+kaKSTVE= -github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI= -github.com/mattn/go-colorable v0.1.4 h1:snbPLB8fVfU9iwbbo30TPtbLRzwWu6aJS6Xh4eaaviA= -github.com/mattn/go-colorable v0.1.4/go.mod h1:U0ppj6V5qS13XJ6of8GYAs25YV2eR4EVcfRqFIhoBtE= -github.com/mattn/go-isatty v0.0.8/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= -github.com/mattn/go-isatty v0.0.11 h1:FxPOTFNqGkuDUGi3H/qkUbQO4ZiBa2brKq5r0l8TGeM= -github.com/mattn/go-isatty v0.0.11/go.mod h1:PhnuNfih5lzO57/f3n+odYbM4JtupLOxQOAqxQCu2WE= +github.com/fatih/color v1.7.0 h1:DkWD4oS2D8LGGgTQ6IvwJJXSL5Vp2ffcQg58nFV38Ys= +github.com/fatih/color v1.7.0/go.mod h1:Zm6kSWBoL9eyXnKyktHP6abPY2pDugNf5KwzbycvMj4= +github.com/mattn/go-colorable v0.1.1 h1:G1f5SKeVxmagw/IyvzvtZE4Gybcc4Tr1tf7I8z0XgOg= +github.com/mattn/go-colorable v0.1.1/go.mod h1:FuOcm+DKB9mbwrcAfNl7/TZVBZ6rcnceauSikq3lYCQ= +github.com/mattn/go-isatty v0.0.5/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= +github.com/mattn/go-isatty v0.0.7 h1:UvyT9uN+3r7yLEYSlJsbQGdsaB/a0DlgWP3pql6iwOc= +github.com/mattn/go-isatty v0.0.7/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= -github.com/sergi/go-diff v1.1.0 h1:we8PVUC3FE2uYfodKH/nBHMSetSfHDR6scGdBi+erh0= -github.com/sergi/go-diff v1.1.0/go.mod h1:STckp+ISIX8hZLjrqAeVduY0gWCT9IjLuqbuNXdaHfM= +github.com/sergi/go-diff v1.0.0 h1:Kpca3qRNrduNnOQeazBd0ysaKrUJiIuISHxogkT9RPQ= +github.com/sergi/go-diff v1.0.0/go.mod h1:0CfEIISq7TuYL3j771MWULgwwjU+GofnZX9QAmXWZgo= github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME= -github.com/stretchr/testify v1.4.0 h1:2E4SXV/wtOkTonXsotYi4li6zVWxYlZuYNCXe9XRJyk= -github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4= +github.com/stretchr/testify v1.3.0 h1:TivCn/peBQ7UY8ooIcPgZFpTNSz0Q2U6UrFlUfqbe0Q= +github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI= +golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223 h1:DH4skfRX4EBpamg7iV4ZlCpblAHI6s6TDM39bFZumv8= golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= -golang.org/x/sys v0.0.0-20191026070338-33540a1f6037 h1:YyJpGZS1sBuBCzLAR1VEpK193GlqGZbnPFnPV/5Rsb4= -golang.org/x/sys v0.0.0-20191026070338-33540a1f6037/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= -gopkg.in/check.v1 v1.0.0-20190902080502-41f04d3bba15 h1:YR8cESwS4TdDjEe65xsg0ogRM/Nc3DYOhEAlW+xobZo= -gopkg.in/check.v1 v1.0.0-20190902080502-41f04d3bba15/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= -gopkg.in/yaml.v2 v2.2.2/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= -gopkg.in/yaml.v2 v2.2.4 h1:/eiJrUcujPVeJ3xlSWaiNi3uSVmDGBK1pDHUHAnao1I= -gopkg.in/yaml.v2 v2.2.4/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= +golang.org/x/sys v0.0.0-20190531175056-4c3a928424d2 h1:T5DasATyLQfmbTpfEXx/IOL9vfjzW6up+ZDkmHvIf2s= +golang.org/x/sys v0.0.0-20190531175056-4c3a928424d2/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= diff --git a/vendor/github.com/google/go-jsonnet/imports.go b/vendor/github.com/google/go-jsonnet/imports.go index e994e4f0..dc254dbc 100644 --- a/vendor/github.com/google/go-jsonnet/imports.go +++ b/vendor/github.com/google/go-jsonnet/imports.go @@ -44,13 +44,6 @@ type Importer interface { // both nonexistence and contents of existing ones. // FileImporter may serve as an example. // - // IMPORTANT: The passed importedFrom might be "" (an empty string). - // It means that the import is coming from an ad-hoc snippet, e.g. - // code passed on the command line, read from stdin or passed as - // a snippet to execute. Importer may have a "default" path to use - // in such case or it may only allow absolute imports from such - // "anonymous locations". - // // Importing the same file multiple times must be a cheap operation // and shouldn't involve copying the whole file - the same buffer // should be returned. @@ -125,22 +118,22 @@ func (cache *importCache) importAST(importedFrom, importedPath string) (ast.Node if cachedNode, isCached := cache.astCache[foundAt]; isCached { return cachedNode, foundAt, nil } - node, err := program.SnippetToAST(ast.DiagnosticFileName(foundAt), foundAt, contents.String()) + node, err := program.SnippetToAST(foundAt, contents.String()) cache.astCache[foundAt] = node return node, foundAt, err } // ImportString imports a string, caches it and then returns it. -func (cache *importCache) importString(importedFrom, importedPath string, i *interpreter) (valueString, error) { +func (cache *importCache) importString(importedFrom, importedPath string, i *interpreter, trace traceElement) (valueString, error) { data, _, err := cache.importData(importedFrom, importedPath) if err != nil { - return nil, i.Error(err.Error()) + return nil, i.Error(err.Error(), trace) } return makeValueString(data.String()), nil } func codeToPV(i *interpreter, filename string, code string) *cachedThunk { - node, err := program.SnippetToAST(ast.DiagnosticFileName(filename), "", code) + node, err := program.SnippetToAST(filename, code) if err != nil { // TODO(sbarzowski) we should wrap (static) error here // within a RuntimeError. Because whether we get this error or not @@ -158,10 +151,10 @@ func codeToPV(i *interpreter, filename string, code string) *cachedThunk { } // ImportCode imports code from a path. -func (cache *importCache) importCode(importedFrom, importedPath string, i *interpreter) (value, error) { +func (cache *importCache) importCode(importedFrom, importedPath string, i *interpreter, trace traceElement) (value, error) { node, foundAt, err := cache.importAST(importedFrom, importedPath) if err != nil { - return nil, i.Error(err.Error()) + return nil, i.Error(err.Error(), trace) } var pv potentialValue if cachedPV, isCached := cache.codeCache[foundAt]; !isCached { @@ -176,7 +169,7 @@ func (cache *importCache) importCode(importedFrom, importedPath string, i *inter } else { pv = cachedPV } - return i.evaluatePV(pv) + return i.evaluatePV(pv, trace) } // Concrete importers @@ -229,11 +222,6 @@ func (importer *FileImporter) tryPath(dir, importedPath string) (found bool, con // Import imports file from the filesystem. func (importer *FileImporter) Import(importedFrom, importedPath string) (contents Contents, foundAt string, err error) { - // TODO(sbarzowski) Make sure that dir is absolute and resolving of "" - // is independent from current CWD. The default path should be saved - // in the importer. - // We need to relativize the paths in the error formatter, so that the stack traces - // don't have ugly absolute paths (less readable and messy with golden tests). dir, _ := path.Split(importedFrom) found, content, foundHere, err := importer.tryPath(dir, importedPath) if err != nil { diff --git a/vendor/github.com/google/go-jsonnet/internal/errors/static_error.go b/vendor/github.com/google/go-jsonnet/internal/errors/static_error.go index fdb8f0cc..62c526d8 100644 --- a/vendor/github.com/google/go-jsonnet/internal/errors/static_error.go +++ b/vendor/github.com/google/go-jsonnet/internal/errors/static_error.go @@ -27,45 +27,26 @@ import ( // StaticError represents an error during parsing/lexing or static analysis. // TODO(sbarzowski) Make it possible to have multiple static errors and warnings -type StaticError interface { - // WithContext returns a new StaticError with additional context before the error message. - WithContext(string) StaticError - // Error returns the string representation of a StaticError. - Error() string - // Loc returns the place in the source code that triggerred the error. - Loc() ast.LocationRange +type StaticError struct { + Loc ast.LocationRange + Msg string } -type staticError struct { - loc ast.LocationRange - msg string +// MakeStaticErrorMsg returns a StaticError with a message. +func MakeStaticErrorMsg(msg string) StaticError { + return StaticError{Msg: msg} } -func (err staticError) WithContext(context string) StaticError { - return staticError{ - loc: err.loc, - msg: fmt.Sprintf("%v while %s", err.msg, context), - } +// MakeStaticError returns a StaticError with a message and a LocationRange. +func MakeStaticError(msg string, lr ast.LocationRange) StaticError { + return StaticError{Msg: msg, Loc: lr} } -func (err staticError) Error() string { +// Error returns the string representation of a StaticError. +func (err StaticError) Error() string { loc := "" - if err.loc.IsSet() { - loc = err.loc.String() + if err.Loc.IsSet() { + loc = err.Loc.String() } - return fmt.Sprintf("%v %v", loc, err.msg) -} - -func (err staticError) Loc() ast.LocationRange { - return err.loc -} - -// MakeStaticErrorMsg returns a staticError with a message. -func MakeStaticErrorMsg(msg string) StaticError { - return staticError{msg: msg} -} - -// MakeStaticError returns a StaticError with a message and a LocationRange. -func MakeStaticError(msg string, lr ast.LocationRange) StaticError { - return staticError{msg: msg, loc: lr} + return fmt.Sprintf("%v %v", loc, err.Msg) } diff --git a/vendor/github.com/google/go-jsonnet/internal/parser/BUILD.bazel b/vendor/github.com/google/go-jsonnet/internal/parser/BUILD.bazel index 97e2d5d8..c69ce9bd 100644 --- a/vendor/github.com/google/go-jsonnet/internal/parser/BUILD.bazel +++ b/vendor/github.com/google/go-jsonnet/internal/parser/BUILD.bazel @@ -7,7 +7,6 @@ go_library( "lexer.go", "literalfield_set.go", "parser.go", - "string_util.go", ], importpath = "github.com/google/go-jsonnet/internal/parser", visibility = ["//:__subpackages__"], diff --git a/vendor/github.com/google/go-jsonnet/internal/parser/context.go b/vendor/github.com/google/go-jsonnet/internal/parser/context.go index 99528251..803b3015 100644 --- a/vendor/github.com/google/go-jsonnet/internal/parser/context.go +++ b/vendor/github.com/google/go-jsonnet/internal/parser/context.go @@ -30,14 +30,14 @@ const anonymous = "anonymous" // package or a separate internal astutils package. The only reason I'm not doing it // right now is that it's a pretty invasive change that deserves a separate PR. -// DirectChildren are children of AST node that are executed in the same context +// directChildren are children of AST node that are executed in the same context // and environment as their parent. It supports ASTs before and after desugaring. // // They must satisfy the following rules: // * (no-delayed-evaluation) They are evaluated when their parent is evaluated or never. // * (no-indirect-evaluation) They cannot be evaluated during evaluation of any non-direct children // * (same-environment) They must be evaluated in the same environment as their parent -func DirectChildren(node ast.Node) []ast.Node { +func directChildren(node ast.Node) []ast.Node { switch node := node.(type) { case *ast.Apply: return []ast.Node{node.Target} @@ -298,10 +298,8 @@ func specialChildren(node ast.Node) []ast.Node { return nil case *ast.Function: children := []ast.Node{node.Body} - for _, child := range node.Parameters { - if child.DefaultArg != nil { - children = append(children, child.DefaultArg) - } + for _, child := range node.Parameters.Optional { + children = append(children, child.DefaultArg) } return children case *ast.Import: @@ -354,7 +352,7 @@ func specialChildren(node ast.Node) []ast.Node { // Children returns all children of a node. It supports ASTs before and after desugaring. func Children(node ast.Node) []ast.Node { var result []ast.Node - result = append(result, DirectChildren(node)...) + result = append(result, directChildren(node)...) result = append(result, thunkChildren(node)...) result = append(result, specialChildren(node)...) return result @@ -391,16 +389,14 @@ func addContext(node ast.Node, context *string, bind string) { case *ast.Function: funContext := functionContext(bind) addContext(node.Body, funContext, anonymous) - for i := range node.Parameters { - if node.Parameters[i].DefaultArg != nil { - // Default arguments have the same context as the function body. - addContext(node.Parameters[i].DefaultArg, funContext, anonymous) - } + for i := range node.Parameters.Optional { + // Default arguments have the same context as the function body. + addContext(node.Parameters.Optional[i].DefaultArg, funContext, anonymous) } case *ast.Object: // TODO(sbarzowski) include fieldname, maybe even chains - outOfObject := DirectChildren(node) + outOfObject := directChildren(node) for _, f := range outOfObject { // This actually is evaluated outside of object addContext(f, context, anonymous) @@ -414,7 +410,7 @@ func addContext(node ast.Node, context *string, bind string) { } case *ast.ObjectComp: - outOfObject := DirectChildren(node) + outOfObject := directChildren(node) for _, f := range outOfObject { // This actually is evaluated outside of object addContext(f, context, anonymous) @@ -438,7 +434,7 @@ func addContext(node ast.Node, context *string, bind string) { } addContext(node.Body, context, bind) default: - for _, child := range DirectChildren(node) { + for _, child := range directChildren(node) { addContext(child, context, anonymous) } diff --git a/vendor/github.com/google/go-jsonnet/internal/parser/lexer.go b/vendor/github.com/google/go-jsonnet/internal/parser/lexer.go index 352b3df8..872d56eb 100644 --- a/vendor/github.com/google/go-jsonnet/internal/parser/lexer.go +++ b/vendor/github.com/google/go-jsonnet/internal/parser/lexer.go @@ -126,17 +126,6 @@ var tokenKindStrings = []string{ tokenEndOfFile: "end of file", } -var tokenHasContent = map[tokenKind]bool{ - tokenIdentifier: true, - tokenNumber: true, - tokenOperator: true, - tokenStringBlock: true, - tokenStringDouble: true, - tokenStringSingle: true, - tokenVerbatimStringDouble: true, - tokenVerbatimStringSingle: true, -} - func (tk tokenKind) String() string { if tk < 0 || int(tk) >= len(tokenKindStrings) { panic(fmt.Sprintf("INTERNAL ERROR: Unknown token kind:: %d", tk)) @@ -162,10 +151,10 @@ type Tokens []token func (t *token) String() string { if t.data == "" { return t.kind.String() - } else if tokenHasContent[t.kind] { - return fmt.Sprintf("(%v, \"%v\")", t.kind, t.data) - } else { + } else if t.kind == tokenOperator { return fmt.Sprintf("\"%v\"", t.data) + } else { + return fmt.Sprintf("(%v, \"%v\")", t.kind, t.data) } } @@ -276,12 +265,12 @@ type position struct { } type lexer struct { - diagnosticFilename ast.DiagnosticFileName // The file name being lexed, only used for errors - importedFilename string // Imported filename, used for resolving relative imports - input string // The input string - source *ast.Source + fileName string // The file name being lexed, only used for errors + input string // The input string + source *ast.Source - pos position // Current position in input + pos position // Current position in input + prev position // Previous position in input tokens Tokens // The tokens that we've generated so far @@ -296,24 +285,26 @@ type lexer struct { const lexEOF = -1 -func makeLexer(diagnosticFilename ast.DiagnosticFileName, importedFilename, input string) *lexer { +func makeLexer(fn string, input string) *lexer { return &lexer{ - input: input, - diagnosticFilename: diagnosticFilename, - importedFilename: importedFilename, - source: ast.BuildSource(diagnosticFilename, input), - pos: position{byteNo: 0, lineNo: 1, lineStart: 0}, - tokenStartLoc: ast.Location{Line: 1, Column: 1}, - freshLine: true, + fileName: fn, + input: input, + source: ast.BuildSource(input), + pos: position{byteNo: 0, lineNo: 1, lineStart: 0}, + prev: position{byteNo: lexEOF, lineNo: 0, lineStart: 0}, + tokenStartLoc: ast.Location{Line: 1, Column: 1}, + freshLine: true, } } // next returns the next rune in the input. func (l *lexer) next() rune { if int(l.pos.byteNo) >= len(l.input) { + l.prev = l.pos return lexEOF } r, w := utf8.DecodeRuneInString(l.input[l.pos.byteNo:]) + l.prev = l.pos l.pos.byteNo += w if r == '\n' { l.pos.lineStart = l.pos.byteNo @@ -335,13 +326,21 @@ func (l *lexer) acceptN(n int) { // peek returns but does not consume the next rune in the input. func (l *lexer) peek() rune { - if int(l.pos.byteNo) >= len(l.input) { - return lexEOF - } - r, _ := utf8.DecodeRuneInString(l.input[l.pos.byteNo:]) + r := l.next() + l.backup() return r } +// backup steps back one rune. Can only be called once per call of next. +// It also does not recover the previous value of freshLine. +func (l *lexer) backup() { + if l.prev.byteNo == lexEOF { + panic("backup called with no valid previous rune") + } + l.pos = l.prev + l.prev = position{byteNo: lexEOF} +} + func locationFromPosition(pos position) ast.Location { return ast.Location{Line: pos.lineNo, Column: pos.byteNo - pos.lineStart + 1} } @@ -350,6 +349,13 @@ func (l *lexer) location() ast.Location { return locationFromPosition(l.pos) } +func (l *lexer) prevLocation() ast.Location { + if l.prev.byteNo == lexEOF { + panic("prevLocation called with no valid previous rune") + } + return locationFromPosition(l.prev) +} + // Reset the current working token start to the current cursor position. This // may throw away some characters. This does not throw away any accumulated // fodder. @@ -365,7 +371,7 @@ func (l *lexer) emitFullToken(kind tokenKind, data, stringBlockIndent, stringBlo data: data, stringBlockIndent: stringBlockIndent, stringBlockTermIndent: stringBlockTermIndent, - loc: ast.MakeLocationRange(l.importedFilename, l.source, l.tokenStartLoc, l.location()), + loc: ast.MakeLocationRange(l.fileName, l.source, l.tokenStartLoc, l.location()), }) l.fodder = ast.Fodder{} } @@ -380,34 +386,31 @@ func (l *lexer) addFodder(kind ast.FodderKind, blanks int, indent int, comment [ l.fodder = append(l.fodder, elem) } -func (l *lexer) addFodderSafe(kind ast.FodderKind, blanks int, indent int, comment []string) { - elem := ast.MakeFodderElement(kind, blanks, indent, comment) - ast.FodderAppend(&l.fodder, elem) -} - func (l *lexer) makeStaticErrorPoint(msg string, loc ast.Location) errors.StaticError { - return errors.MakeStaticError(msg, ast.MakeLocationRange(l.importedFilename, l.source, loc, loc)) + return errors.StaticError{Msg: msg, Loc: ast.MakeLocationRange(l.fileName, l.source, loc, loc)} } // lexWhitespace consumes all whitespace and returns the number of \n and number of // spaces after last \n. It also converts \t to spaces. // The parameter 'r' is the rune that begins the whitespace. func (l *lexer) lexWhitespace() (int, int) { - r := l.peek() + r := l.next() indent := 0 newLines := 0 - for ; isWhitespace(r); r = l.peek() { - l.next() + for ; isWhitespace(r); r = l.next() { switch r { case '\r': // Ignore. + break case '\n': indent = 0 newLines++ + break case ' ': indent++ + break // This only works for \t at the beginning of lines, but we strip it everywhere else // anyway. The only case where this will cause a problem is spaces followed by \t @@ -415,8 +418,10 @@ func (l *lexer) lexWhitespace() (int, int) { // is enabled it will be fixed later. case '\t': indent += 8 + break } } + l.backup() return newLines, indent } @@ -426,13 +431,13 @@ func (l *lexer) lexUntilNewline() (string, int, int) { // Compute 'text'. var buf bytes.Buffer lastNonSpace := 0 - for r := l.peek(); r != lexEOF && r != '\n'; r = l.peek() { - l.next() + for r := l.next(); r != lexEOF && r != '\n'; r = l.next() { buf.WriteRune(r) if !isHorizontalWhitespace(r) { lastNonSpace = buf.Len() } } + l.backup() // Trim whitespace off the end. buf.Truncate(lastNonSpace) text := buf.String() @@ -473,8 +478,8 @@ func (l *lexer) lexNumber() error { state := numBegin outerLoop: - for { - r := l.peek() + for true { + r := l.next() switch state { case numBegin: switch { @@ -513,7 +518,7 @@ outerLoop: default: return l.makeStaticErrorPoint( fmt.Sprintf("Couldn't lex number, junk after decimal point: %v", strconv.QuoteRuneToASCII(r)), - l.location()) + l.prevLocation()) } case numAfterDigit: switch { @@ -533,7 +538,7 @@ outerLoop: default: return l.makeStaticErrorPoint( fmt.Sprintf("Couldn't lex number, junk after 'E': %v", strconv.QuoteRuneToASCII(r)), - l.location()) + l.prevLocation()) } case numAfterExpSign: if r >= '0' && r <= '9' { @@ -541,7 +546,7 @@ outerLoop: } else { return l.makeStaticErrorPoint( fmt.Sprintf("Couldn't lex number, junk after exponent sign: %v", strconv.QuoteRuneToASCII(r)), - l.location()) + l.prevLocation()) } case numAfterExpDigit: @@ -551,106 +556,80 @@ outerLoop: break outerLoop } } - l.next() } + l.backup() l.emitToken(tokenNumber) return nil } -// getTokenKindFromID will return a keyword if the identifier string is -// recognised as one, otherwise it will return tokenIdentifier. -func getTokenKindFromID(str string) tokenKind { - switch str { +// lexIdentifier will consume a identifer and emit a token. It is assumed +// that the next rune to be served by the lexer will be a leading digit. This +// may emit a keyword or an identifier. +func (l *lexer) lexIdentifier() { + r := l.next() + if !isIdentifierFirst(r) { + panic("Unexpected character in lexIdentifier") + } + for ; r != lexEOF; r = l.next() { + if !isIdentifier(r) { + break + } + } + l.backup() + + switch l.input[l.tokenStart:l.pos.byteNo] { case "assert": - return tokenAssert + l.emitToken(tokenAssert) case "else": - return tokenElse + l.emitToken(tokenElse) case "error": - return tokenError + l.emitToken(tokenError) case "false": - return tokenFalse + l.emitToken(tokenFalse) case "for": - return tokenFor + l.emitToken(tokenFor) case "function": - return tokenFunction + l.emitToken(tokenFunction) case "if": - return tokenIf + l.emitToken(tokenIf) case "import": - return tokenImport + l.emitToken(tokenImport) case "importstr": - return tokenImportStr + l.emitToken(tokenImportStr) case "in": - return tokenIn + l.emitToken(tokenIn) case "local": - return tokenLocal + l.emitToken(tokenLocal) case "null": - return tokenNullLit + l.emitToken(tokenNullLit) case "self": - return tokenSelf + l.emitToken(tokenSelf) case "super": - return tokenSuper + l.emitToken(tokenSuper) case "tailstrict": - return tokenTailStrict + l.emitToken(tokenTailStrict) case "then": - return tokenThen + l.emitToken(tokenThen) case "true": - return tokenTrue + l.emitToken(tokenTrue) default: // Not a keyword, assume it is an identifier - return tokenIdentifier - } -} - -// IsValidIdentifier is true if the string could be a valid identifier. -func IsValidIdentifier(str string) bool { - if len(str) == 0 { - return false - } - for i, r := range str { - if i == 0 { - if !isIdentifierFirst(r) { - return false - } - } else { - if !isIdentifier(r) { - return false - } - } - } - return getTokenKindFromID(str) == tokenIdentifier -} - -// lexIdentifier will consume an identifer and emit a token. It is assumed -// that the next rune to be served by the lexer will not be a leading digit. -// This may emit a keyword or an identifier. -func (l *lexer) lexIdentifier() { - r := l.peek() - if !isIdentifierFirst(r) { - panic("Unexpected character in lexIdentifier") - } - for ; r != lexEOF; r = l.peek() { - if !isIdentifier(r) { - break - } - l.next() + l.emitToken(tokenIdentifier) } - l.emitToken(getTokenKindFromID(l.input[l.tokenStart:l.pos.byteNo])) } // lexSymbol will lex a token that starts with a symbol. This could be a // C or C++ comment, block quote or an operator. This function assumes that the next // rune to be served by the lexer will be the first rune of the new token. func (l *lexer) lexSymbol() error { - // freshLine is reset by next() so cache it here. - freshLine := l.freshLine r := l.next() // Single line C++ style comment if r == '#' || (r == '/' && l.peek() == '/') { comment, blanks, indent := l.lexUntilNewline() var k ast.FodderKind - if freshLine { + if l.freshLine { k = ast.FodderParagraph } else { k = ast.FodderLineEnd @@ -661,10 +640,9 @@ func (l *lexer) lexSymbol() error { // C style comment (could be interstitial or paragraph comment) if r == '/' && l.peek() == '*' { - margin := l.pos.byteNo - l.pos.lineStart - 1 + margin := l.pos.byteNo - l.pos.lineStart commentStartLoc := l.tokenStartLoc - //nolint:ineffassign,staticcheck r := l.next() // consume the initial '*' for r = l.next(); r != '*' || l.peek() != '/'; r = l.next() { if r == lexEOF { @@ -698,9 +676,9 @@ func (l *lexer) lexSymbol() error { } } if allStar { - for i := range lines { - if lines[i][0] == '*' { - lines[i] = " " + lines[i] + for _, l := range lines { + if l[0] == '*' { + l = " " + l } } } @@ -709,7 +687,7 @@ func (l *lexer) lexSymbol() error { newLinesAfter = 1 indentAfter = 0 } - l.addFodderSafe(ast.FodderParagraph, newLinesAfter-1, indentAfter, lines) + l.addFodder(ast.FodderParagraph, newLinesAfter-1, indentAfter, lines) } return nil } @@ -730,10 +708,10 @@ func (l *lexer) lexSymbol() error { } // Process leading blank lines before calculating stringBlockIndent - for r = l.peek(); r == '\n'; r = l.peek() { - l.next() + for r = l.next(); r == '\n'; r = l.next() { cb.WriteRune(r) } + l.backup() numWhiteSpace := checkWhitespace(l.input[l.pos.byteNo:], l.input[l.pos.byteNo:]) stringBlockIndent := l.input[l.pos.byteNo : l.pos.byteNo+numWhiteSpace] if numWhiteSpace == 0 { @@ -755,20 +733,20 @@ func (l *lexer) lexSymbol() error { cb.WriteRune('\n') // Skip any blank lines - for r = l.peek(); r == '\n'; r = l.peek() { - l.next() + for r = l.next(); r == '\n'; r = l.next() { cb.WriteRune(r) } + l.backup() // Look at the next line numWhiteSpace = checkWhitespace(stringBlockIndent, l.input[l.pos.byteNo:]) if numWhiteSpace == 0 { // End of the text block var stringBlockTermIndent string - for r = l.peek(); r == ' ' || r == '\t'; r = l.peek() { - l.next() + for r = l.next(); r == ' ' || r == '\t'; r = l.next() { stringBlockTermIndent += string(r) } + l.backup() if !strings.HasPrefix(l.input[l.pos.byteNo:], "|||") { return l.makeStaticErrorPoint("Text block not terminated with |||", commentStartLoc) } @@ -782,7 +760,7 @@ func (l *lexer) lexSymbol() error { } // Assume any string of symbols is a single operator. - for r = l.peek(); isSymbol(r); r = l.peek() { + for r = l.next(); isSymbol(r); r = l.next() { // Not allowed // in operators if r == '/' && strings.HasPrefix(l.input[l.pos.byteNo:], "/") { break @@ -795,9 +773,10 @@ func (l *lexer) lexSymbol() error { if r == '|' && strings.HasPrefix(l.input[l.pos.byteNo:], "||") { break } - l.next() } + l.backup() + // Operators are not allowed to end with + - ~ ! unless they are one rune long. // So, wind it back if we need to, but stop at the first rune. // This relies on the hack that all operator symbols are ASCII and thus there is @@ -819,11 +798,11 @@ func (l *lexer) lexSymbol() error { } // Lex returns a slice of tokens recognised in input. -func Lex(diagnosticFilename ast.DiagnosticFileName, importedFilename, input string) (Tokens, error) { - l := makeLexer(diagnosticFilename, importedFilename, input) +func Lex(fn string, input string) (Tokens, error) { + l := makeLexer(fn, input) var err error - for { + for true { newLines, indent := l.lexWhitespace() // If it's the end of the file, discard final whitespace. if l.peek() == lexEOF { @@ -837,47 +816,37 @@ func Lex(diagnosticFilename ast.DiagnosticFileName, importedFilename, input stri l.addFodder(ast.FodderLineEnd, blanks, indent, []string{}) } l.resetTokenStart() // Don't include whitespace in actual token. - r := l.peek() + r := l.next() switch r { case '{': - l.next() l.emitToken(tokenBraceL) case '}': - l.next() l.emitToken(tokenBraceR) case '[': - l.next() l.emitToken(tokenBracketL) case ']': - l.next() l.emitToken(tokenBracketR) case ',': - l.next() l.emitToken(tokenComma) case '.': - l.next() l.emitToken(tokenDot) case '(': - l.next() l.emitToken(tokenParenL) case ')': - l.next() l.emitToken(tokenParenR) case ';': - l.next() l.emitToken(tokenSemicolon) case '0', '1', '2', '3', '4', '5', '6', '7', '8', '9': + l.backup() err = l.lexNumber() if err != nil { return nil, err } - // String literals - + // String literals case '"': - stringStartLoc := l.location() - l.next() + stringStartLoc := l.prevLocation() for r = l.next(); ; r = l.next() { if r == lexEOF { return nil, l.makeStaticErrorPoint("Unterminated String", stringStartLoc) @@ -889,13 +858,11 @@ func Lex(diagnosticFilename ast.DiagnosticFileName, importedFilename, input stri break } if r == '\\' && l.peek() != lexEOF { - //nolint:ineffassign,staticcheck r = l.next() } } case '\'': - stringStartLoc := l.location() - l.next() + stringStartLoc := l.prevLocation() for r = l.next(); ; r = l.next() { if r == lexEOF { return nil, l.makeStaticErrorPoint("Unterminated String", stringStartLoc) @@ -907,13 +874,10 @@ func Lex(diagnosticFilename ast.DiagnosticFileName, importedFilename, input stri break } if r == '\\' && l.peek() != lexEOF { - //nolint:ineffassign,staticcheck r = l.next() } } case '@': - stringStartLoc := l.location() - l.next() // Verbatim string literals. // ' and " quoting is interpreted here, unlike non-verbatim strings // where it is done later by jsonnet_string_unescape. This is OK @@ -921,6 +885,7 @@ func Lex(diagnosticFilename ast.DiagnosticFileName, importedFilename, input stri // repeated quote into a single quote, so we can go back to the // original form in the formatter. var data []rune + stringStartLoc := l.prevLocation() quot := l.next() var kind tokenKind if quot == '"' { @@ -952,8 +917,10 @@ func Lex(diagnosticFilename ast.DiagnosticFileName, importedFilename, input stri default: if isIdentifierFirst(r) { + l.backup() l.lexIdentifier() } else if isSymbol(r) || r == '#' { + l.backup() err = l.lexSymbol() if err != nil { return nil, err @@ -961,7 +928,7 @@ func Lex(diagnosticFilename ast.DiagnosticFileName, importedFilename, input stri } else { return nil, l.makeStaticErrorPoint( fmt.Sprintf("Could not lex the character %s", strconv.QuoteRuneToASCII(r)), - l.location()) + l.prevLocation()) } } diff --git a/vendor/github.com/google/go-jsonnet/internal/parser/parser.go b/vendor/github.com/google/go-jsonnet/internal/parser/parser.go index c576074c..b47d0711 100644 --- a/vendor/github.com/google/go-jsonnet/internal/parser/parser.go +++ b/vendor/github.com/google/go-jsonnet/internal/parser/parser.go @@ -19,6 +19,7 @@ package parser import ( "fmt" + "strconv" "github.com/google/go-jsonnet/ast" "github.com/google/go-jsonnet/internal/errors" @@ -56,9 +57,9 @@ var bopPrecedence = map[ast.BinaryOp]precedence{ // --------------------------------------------------------------------------- -func makeUnexpectedError(t *token, while string) errors.StaticError { +func makeUnexpectedError(t *token, while string) error { return errors.MakeStaticError( - fmt.Sprintf("Unexpected: %v", t), t.loc).WithContext(while) + fmt.Sprintf("Unexpected: %v while %v", t, while), t.loc) } func locFromTokens(begin, end *token) ast.LocationRange { @@ -88,14 +89,14 @@ func (p *parser) pop() *token { return t } -func (p *parser) unexpectedTokenError(tk tokenKind, t *token) errors.StaticError { +func (p *parser) unexpectedTokenError(tk tokenKind, t *token) error { if tk == t.kind { - panic("Unexpectedly expected token kind") + panic("Unexpectedly expected token kind.") } return errors.MakeStaticError(fmt.Sprintf("Expected token %v but got %v", tk, t), t.loc) } -func (p *parser) popExpect(tk tokenKind) (*token, errors.StaticError) { +func (p *parser) popExpect(tk tokenKind) (*token, error) { t := p.pop() if t.kind != tk { return nil, p.unexpectedTokenError(tk, t) @@ -103,7 +104,7 @@ func (p *parser) popExpect(tk tokenKind) (*token, errors.StaticError) { return t, nil } -func (p *parser) popExpectOp(op string) (*token, errors.StaticError) { +func (p *parser) popExpectOp(op string) (*token, error) { t := p.pop() if t.kind != tokenOperator || t.data != op { return nil, errors.MakeStaticError( @@ -120,10 +121,20 @@ func (p *parser) doublePeek() *token { return &p.t[p.currT+1] } +// in some cases it's convenient to parse something as an expression, and later +// decide that it should be just an identifer +func astVarToIdentifier(node ast.Node) (ast.Fodder, *ast.Identifier, bool) { + v, ok := node.(*ast.Var) + if ok { + return v.NodeBase.Fodder, &v.Id, true + } + return nil, nil, false +} + // parseArgument parses either id = expr or just expr. // It returns either (, id, , expr) or (nil, nil, nil, expr) // respectively. -func (p *parser) parseArgument() (ast.Fodder, *ast.Identifier, ast.Fodder, ast.Node, errors.StaticError) { +func (p *parser) parseArgument() (ast.Fodder, *ast.Identifier, ast.Fodder, ast.Node, error) { var idFodder ast.Fodder var id *ast.Identifier var eqFodder ast.Fodder @@ -143,13 +154,14 @@ func (p *parser) parseArgument() (ast.Fodder, *ast.Identifier, ast.Fodder, ast.N } // TODO(sbarzowski) - this returned bool is weird -func (p *parser) parseArguments(elementKind string) (*token, *ast.Arguments, bool, errors.StaticError) { +// TODO(sbarzowski) - name - it's also used for parameters +func (p *parser) parseArguments(elementKind string) (*token, *ast.Arguments, bool, error) { args := &ast.Arguments{} gotComma := false + commaFodder := ast.Fodder{} namedArgumentAdded := false first := true for { - commaFodder := ast.Fodder{} next := p.peek() if next.kind == tokenParenR { @@ -158,7 +170,7 @@ func (p *parser) parseArguments(elementKind string) (*token, *ast.Arguments, boo } if !first && !gotComma { - return nil, nil, false, errors.MakeStaticError(fmt.Sprintf("Expected a comma before next %s, got %s", elementKind, next), next.loc) + return nil, nil, false, errors.MakeStaticError(fmt.Sprintf("Expected a comma before next %s, got %s.", elementKind, next), next.loc) } idFodder, id, eqFodder, expr, err := p.parseArgument() @@ -198,75 +210,41 @@ func (p *parser) parseArguments(elementKind string) (*token, *ast.Arguments, boo } } -// parseParameter parses either id = expr or just id. -// It returns either (, id, , expr) or (, id, nil, nil) -// respectively. -func (p *parser) parseParameter() (ast.Parameter, errors.StaticError) { - ret := ast.Parameter{} - ident, err := p.popExpect(tokenIdentifier) +// TODO(sbarzowski) - this returned bool is weird +func (p *parser) parseParameters(elementKind string) (*token, *ast.Parameters, bool, error) { + parenR, args, trailingComma, err := p.parseArguments(elementKind) if err != nil { - return ret, err.WithContext("parsing parameter") + return nil, nil, false, err } - ret.Name = ast.Identifier(ident.data) - ret.NameFodder = ident.fodder - ret.LocRange = ident.loc - if p.peek().kind == tokenOperator && p.peek().data == "=" { - eq := p.pop() - ret.EqFodder = eq.fodder - ret.DefaultArg, err = p.parse(maxPrecedence) - if err != nil { - return ret, err - } - ret.LocRange = locFromTokenAST(ident, ret.DefaultArg) + var params ast.Parameters + for _, arg := range args.Positional { + idFodder, id, ok := astVarToIdentifier(arg.Expr) + if !ok { + return nil, nil, false, errors.MakeStaticError(fmt.Sprintf("Expected simple identifier but got a complex expression."), *arg.Expr.Loc()) + } + params.Required = append(params.Required, ast.CommaSeparatedID{ + NameFodder: idFodder, + Name: *id, + CommaFodder: arg.CommaFodder, + }) } - return ret, nil -} - -// TODO(sbarzowski) - this returned bool is weird -func (p *parser) parseParameters(elementKind string) (*token, []ast.Parameter, bool, errors.StaticError) { - - var parenR *token - var params []ast.Parameter - gotComma := false - first := true - for { - next := p.peek() - - if next.kind == tokenParenR { - // gotComma can be true or false here. - parenR = p.pop() - break - } - - if !first && !gotComma { - return nil, nil, false, errors.MakeStaticError(fmt.Sprintf("Expected a comma before next %s, got %s", elementKind, next), next.loc) - } - - param, err := p.parseParameter() - if err != nil { - return nil, nil, false, err - } - - if p.peek().kind == tokenComma { - comma := p.pop() - param.CommaFodder = comma.fodder - gotComma = true - } else { - gotComma = false - } - params = append(params, param) - - first = false + for _, arg := range args.Named { + params.Optional = append(params.Optional, ast.NamedParameter{ + NameFodder: arg.NameFodder, + Name: arg.Name, + EqFodder: arg.EqFodder, + DefaultArg: arg.Arg, + CommaFodder: arg.CommaFodder, + }) } - - return parenR, params, gotComma, nil + return parenR, ¶ms, trailingComma, nil } // TODO(sbarzowski) add location to all individual binds -func (p *parser) parseBind(binds *ast.LocalBinds) (*token, errors.StaticError) { - varID, popErr := p.popExpect(tokenIdentifier) - if popErr != nil { - return nil, popErr +func (p *parser) parseBind(binds *ast.LocalBinds) (*token, error) { + varID, err := p.popExpect(tokenIdentifier) + if err != nil { + return nil, err } for _, b := range *binds { if b.Variable == ast.Identifier(varID.data) { @@ -283,16 +261,16 @@ func (p *parser) parseBind(binds *ast.LocalBinds) (*token, errors.StaticError) { } fun = &ast.Function{ ParenLeftFodder: parenL.fodder, - Parameters: params, + Parameters: *params, TrailingComma: gotComma, ParenRightFodder: parenR.fodder, // Body gets filled in later. } } - eqToken, popErr := p.popExpectOp("=") - if popErr != nil { - return nil, popErr + eqToken, err := p.popExpectOp("=") + if err != nil { + return nil, err } body, err := p.parse(maxPrecedence) if err != nil { @@ -314,7 +292,6 @@ func (p *parser) parseBind(binds *ast.LocalBinds) (*token, errors.StaticError) { Body: body, Fun: fun, CloseFodder: delim.fodder, - LocRange: locFromTokenAST(varID, body), }) } else { *binds = append(*binds, ast.LocalBind{ @@ -323,14 +300,13 @@ func (p *parser) parseBind(binds *ast.LocalBinds) (*token, errors.StaticError) { EqFodder: eqToken.fodder, Body: body, CloseFodder: delim.fodder, - LocRange: locFromTokenAST(varID, body), }) } return delim, nil } -func (p *parser) parseObjectAssignmentOp() (opFodder ast.Fodder, plusSugar bool, hide ast.ObjectFieldHide, err errors.StaticError) { +func (p *parser) parseObjectAssignmentOp() (opFodder ast.Fodder, plusSugar bool, hide ast.ObjectFieldHide, err error) { op, err := p.popExpect(tokenOperator) if err != nil { return @@ -373,7 +349,7 @@ func (p *parser) parseObjectAssignmentOp() (opFodder ast.Fodder, plusSugar bool, // +gen set type LiteralField string -func (p *parser) parseObjectRemainderComp(fields ast.ObjectFields, gotComma bool, tok *token, next *token) (ast.Node, *token, errors.StaticError) { +func (p *parser) parseObjectRemainderComp(fields ast.ObjectFields, gotComma bool, tok *token, next *token) (ast.Node, *token, error) { numFields := 0 numAsserts := 0 var field ast.ObjectField @@ -390,16 +366,16 @@ func (p *parser) parseObjectRemainderComp(fields ast.ObjectFields, gotComma bool } if numAsserts > 0 { - return nil, nil, errors.MakeStaticError("Object comprehension cannot have asserts", next.loc) + return nil, nil, errors.MakeStaticError("Object comprehension cannot have asserts.", next.loc) } if numFields != 1 { - return nil, nil, errors.MakeStaticError("Object comprehension can only have one field", next.loc) + return nil, nil, errors.MakeStaticError("Object comprehension can only have one field.", next.loc) } if field.Hide != ast.ObjectFieldInherit { - return nil, nil, errors.MakeStaticError("Object comprehensions cannot have hidden fields", next.loc) + return nil, nil, errors.MakeStaticError("Object comprehensions cannot have hidden fields.", next.loc) } if field.Kind != ast.ObjectFieldExpr { - return nil, nil, errors.MakeStaticError("Object comprehensions can only have [e] fields", next.loc) + return nil, nil, errors.MakeStaticError("Object comprehensions can only have [e] fields.", next.loc) } spec, last, err := p.parseComprehensionSpecs(next, tokenBraceR) if err != nil { @@ -414,7 +390,7 @@ func (p *parser) parseObjectRemainderComp(fields ast.ObjectFields, gotComma bool }, last, nil } -func (p *parser) parseObjectRemainderField(literalFields *LiteralFieldSet, tok *token, next *token) (*ast.ObjectField, errors.StaticError) { +func (p *parser) parseObjectRemainderField(literalFields *LiteralFieldSet, tok *token, next *token) (*ast.ObjectField, error) { var kind ast.ObjectFieldKind var fodder1 ast.Fodder var expr1 ast.Node @@ -432,7 +408,7 @@ func (p *parser) parseObjectRemainderField(literalFields *LiteralFieldSet, tok * default: fodder1 = next.fodder kind = ast.ObjectFieldExpr - var err errors.StaticError + var err error expr1, err = p.parse(maxPrecedence) if err != nil { return nil, err @@ -448,10 +424,10 @@ func (p *parser) parseObjectRemainderField(literalFields *LiteralFieldSet, tok * methComma := false var parenL *token var parenR *token - var params []ast.Parameter + var params *ast.Parameters if p.peek().kind == tokenParenL { parenL = p.pop() - var err errors.StaticError + var err error parenR, params, methComma, err = p.parseParameters("method parameter") if err != nil { return nil, err @@ -485,7 +461,7 @@ func (p *parser) parseObjectRemainderField(literalFields *LiteralFieldSet, tok * if isMethod { method = &ast.Function{ ParenLeftFodder: parenL.fodder, - Parameters: params, + Parameters: *params, TrailingComma: methComma, ParenRightFodder: parenR.fodder, Body: body, @@ -509,14 +485,13 @@ func (p *parser) parseObjectRemainderField(literalFields *LiteralFieldSet, tok * OpFodder: opFodder, Expr2: body, CommaFodder: commaFodder, - LocRange: locFromTokenAST(next, body), }, nil } -func (p *parser) parseObjectRemainderLocal(binds *ast.IdentifierSet, tok *token, next *token) (*ast.ObjectField, errors.StaticError) { - varID, popErr := p.popExpect(tokenIdentifier) - if popErr != nil { - return nil, popErr +func (p *parser) parseObjectRemainderLocal(binds *ast.IdentifierSet, tok *token, next *token) (*ast.ObjectField, error) { + varID, err := p.popExpect(tokenIdentifier) + if err != nil { + return nil, err } id := ast.Identifier(varID.data) @@ -531,19 +506,18 @@ func (p *parser) parseObjectRemainderLocal(binds *ast.IdentifierSet, tok *token, funcComma := false var parenL *token var parenR *token - var params []ast.Parameter + var params *ast.Parameters if p.peek().kind == tokenParenL { parenL = p.pop() isMethod = true - var err errors.StaticError parenR, params, funcComma, err = p.parseParameters("function parameter") if err != nil { return nil, err } } - opToken, popErr := p.popExpectOp("=") - if popErr != nil { - return nil, popErr + opToken, err := p.popExpectOp("=") + if err != nil { + return nil, err } body, err := p.parse(maxPrecedence) @@ -555,7 +529,7 @@ func (p *parser) parseObjectRemainderLocal(binds *ast.IdentifierSet, tok *token, if isMethod { method = &ast.Function{ ParenLeftFodder: parenL.fodder, - Parameters: params, + Parameters: *params, ParenRightFodder: parenR.fodder, TrailingComma: funcComma, Body: body, @@ -580,16 +554,14 @@ func (p *parser) parseObjectRemainderLocal(binds *ast.IdentifierSet, tok *token, OpFodder: opToken.fodder, Expr2: body, CommaFodder: commaFodder, - LocRange: locFromTokenAST(varID, body), }, nil } -func (p *parser) parseObjectRemainderAssert(tok *token, next *token) (*ast.ObjectField, errors.StaticError) { +func (p *parser) parseObjectRemainderAssert(tok *token, next *token) (*ast.ObjectField, error) { cond, err := p.parse(maxPrecedence) if err != nil { return nil, err } - lastAST := cond // for determining location var msg ast.Node var colonFodder ast.Fodder if p.peek().kind == tokenOperator && p.peek().data == ":" { @@ -599,7 +571,6 @@ func (p *parser) parseObjectRemainderAssert(tok *token, next *token) (*ast.Objec if err != nil { return nil, err } - lastAST = msg } var commaFodder ast.Fodder @@ -615,12 +586,11 @@ func (p *parser) parseObjectRemainderAssert(tok *token, next *token) (*ast.Objec OpFodder: colonFodder, Expr3: msg, CommaFodder: commaFodder, - LocRange: locFromTokenAST(next, lastAST), }, nil } // Parse object or object comprehension without leading brace -func (p *parser) parseObjectRemainder(tok *token) (ast.Node, *token, errors.StaticError) { +func (p *parser) parseObjectRemainder(tok *token) (ast.Node, *token, error) { var fields ast.ObjectFields literalFields := make(LiteralFieldSet) binds := make(ast.IdentifierSet) @@ -647,11 +617,11 @@ func (p *parser) parseObjectRemainder(tok *token) (ast.Node, *token, errors.Stat } if !gotComma && !first { - return nil, nil, errors.MakeStaticError("Expected a comma before next field", next.loc) + return nil, nil, errors.MakeStaticError("Expected a comma before next field.", next.loc) } var field *ast.ObjectField - var err errors.StaticError + var err error switch next.kind { case tokenBracketL, tokenIdentifier, tokenStringDouble, tokenStringSingle, tokenStringBlock, tokenVerbatimStringDouble, tokenVerbatimStringSingle: @@ -690,14 +660,14 @@ func (p *parser) parseObjectRemainder(tok *token) (ast.Node, *token, errors.Stat } /* parses for x in expr for y in expr if expr for z in expr ... */ -func (p *parser) parseComprehensionSpecs(forToken *token, end tokenKind) (*ast.ForSpec, *token, errors.StaticError) { - var parseComprehensionSpecsHelper func(forToken *token, outer *ast.ForSpec) (*ast.ForSpec, *token, errors.StaticError) - parseComprehensionSpecsHelper = func(forToken *token, outer *ast.ForSpec) (*ast.ForSpec, *token, errors.StaticError) { +func (p *parser) parseComprehensionSpecs(forToken *token, end tokenKind) (*ast.ForSpec, *token, error) { + var parseComprehensionSpecsHelper func(forToken *token, outer *ast.ForSpec) (*ast.ForSpec, *token, error) + parseComprehensionSpecsHelper = func(forToken *token, outer *ast.ForSpec) (*ast.ForSpec, *token, error) { var ifSpecs []ast.IfSpec - varID, popErr := p.popExpect(tokenIdentifier) - if popErr != nil { - return nil, nil, popErr + varID, err := p.popExpect(tokenIdentifier) + if err != nil { + return nil, nil, err } id := ast.Identifier(varID.data) inToken, err := p.popExpect(tokenIn) @@ -745,7 +715,7 @@ func (p *parser) parseComprehensionSpecs(forToken *token, end tokenKind) (*ast.F // Assumes that the leading '[' has already been consumed and passed as tok. // Should read up to and consume the trailing ']' -func (p *parser) parseArray(tok *token) (ast.Node, errors.StaticError) { +func (p *parser) parseArray(tok *token) (ast.Node, error) { if p.peek().kind == tokenBracketR { bracketR := p.pop() return &ast.Array{ @@ -798,7 +768,7 @@ func (p *parser) parseArray(tok *token) (ast.Node, errors.StaticError) { break } if !gotComma { - return nil, errors.MakeStaticError("Expected a comma before next array element", next.loc) + return nil, errors.MakeStaticError("Expected a comma before next array element.", next.loc) } nextElem, err := p.parse(maxPrecedence) if err != nil { @@ -843,11 +813,10 @@ func tokenStringToAst(tok *token) *ast.LiteralString { } case tokenStringBlock: return &ast.LiteralString{ - NodeBase: ast.NewNodeBaseLoc(tok.loc, tok.fodder), - Value: tok.data, - Kind: ast.StringBlock, - BlockIndent: tok.stringBlockIndent, - BlockTermIndent: tok.stringBlockTermIndent, + NodeBase: ast.NewNodeBaseLoc(tok.loc, tok.fodder), + Value: tok.data, + Kind: ast.StringBlock, + BlockIndent: tok.stringBlockIndent, } case tokenVerbatimStringDouble: return &ast.LiteralString{ @@ -866,7 +835,7 @@ func tokenStringToAst(tok *token) *ast.LiteralString { } } -func (p *parser) parseTerminal() (ast.Node, errors.StaticError) { +func (p *parser) parseTerminal() (ast.Node, error) { tok := p.pop() switch tok.kind { case tokenAssert, tokenBraceR, tokenBracketR, tokenComma, tokenDot, tokenElse, @@ -875,7 +844,7 @@ func (p *parser) parseTerminal() (ast.Node, errors.StaticError) { return nil, makeUnexpectedError(tok, "parsing terminal") case tokenEndOfFile: - return nil, errors.MakeStaticError("Unexpected end of file", tok.loc) + return nil, errors.MakeStaticError("Unexpected end of file.", tok.loc) case tokenBraceL: obj, _, err := p.parseObjectRemainder(tok) @@ -901,8 +870,15 @@ func (p *parser) parseTerminal() (ast.Node, errors.StaticError) { // Literals case tokenNumber: + // This shouldn't fail as the lexer should make sure we have good input but + // we handle the error regardless. + num, err := strconv.ParseFloat(tok.data, 64) + if err != nil { + return nil, errors.MakeStaticError("Could not parse floating point number.", tok.loc) + } return &ast.LiteralNumber{ NodeBase: ast.NewNodeBaseLoc(tok.loc, tok.fodder), + Value: num, OriginalString: tok.data, }, nil case tokenStringDouble, tokenStringSingle, @@ -951,7 +927,7 @@ func (p *parser) parseTerminal() (ast.Node, errors.StaticError) { idFodder = fieldID.fodder id = (*ast.Identifier)(&fieldID.data) case tokenBracketL: - var err errors.StaticError + var err error index, err = p.parse(maxPrecedence) if err != nil { return nil, err @@ -962,7 +938,7 @@ func (p *parser) parseTerminal() (ast.Node, errors.StaticError) { } idFodder = bracketR.fodder default: - return nil, errors.MakeStaticError("Expected . or [ after super", tok.loc) + return nil, errors.MakeStaticError("Expected . or [ after super.", tok.loc) } return &ast.SuperIndex{ NodeBase: ast.NewNodeBaseLoc(tok.loc, tok.fodder), @@ -976,11 +952,11 @@ func (p *parser) parseTerminal() (ast.Node, errors.StaticError) { return nil, errors.MakeStaticError(fmt.Sprintf("INTERNAL ERROR: Unknown tok kind: %v", tok.kind), tok.loc) } -func (p *parser) parsingFailure(msg string, tok *token) (ast.Node, errors.StaticError) { +func (p *parser) parsingFailure(msg string, tok *token) (ast.Node, error) { return nil, errors.MakeStaticError(msg, tok.loc) } -func (p *parser) parse(prec precedence) (ast.Node, errors.StaticError) { +func (p *parser) parse(prec precedence) (ast.Node, error) { begin := p.peek() switch begin.kind { @@ -1082,7 +1058,7 @@ func (p *parser) parse(prec precedence) (ast.Node, errors.StaticError) { return &ast.Function{ NodeBase: ast.NewNodeBaseLoc(locFromTokenAST(begin, body), begin.fodder), ParenLeftFodder: next.fodder, - Parameters: params, + Parameters: *params, TrailingComma: gotComma, ParenRightFodder: parenR.fodder, Body: body, @@ -1296,11 +1272,10 @@ func (p *parser) parse(prec precedence) (ast.Node, errors.StaticError) { } id := ast.Identifier(fieldID.data) lhs = &ast.Index{ - NodeBase: ast.NewNodeBaseLoc(locFromTokens(begin, fieldID), ast.Fodder{}), - Target: lhs, - LeftBracketFodder: op.fodder, - Id: &id, - RightBracketFodder: fieldID.fodder, + NodeBase: ast.NewNodeBaseLoc(locFromTokens(begin, fieldID), ast.Fodder{}), + Target: lhs, + LeftBracketFodder: op.fodder, + Id: &id, } case tokenParenL: end, args, gotComma, err := p.parseArguments("function argument") @@ -1363,31 +1338,28 @@ func (p *parser) parse(prec precedence) (ast.Node, errors.StaticError) { // --------------------------------------------------------------------------- -// Parse parses a slice of tokens into a parse tree. Any fodder after the final token is -// returned as well. -func Parse(t Tokens) (ast.Node, ast.Fodder, errors.StaticError) { +// Parse parses a slice of tokens into a parse tree. +func Parse(t Tokens) (ast.Node, error) { p := makeParser(t) expr, err := p.parse(maxPrecedence) if err != nil { - return nil, nil, err + return nil, err } - eof := p.peek() - if eof.kind != tokenEndOfFile { - return nil, nil, errors.MakeStaticError(fmt.Sprintf("Did not expect: %v", eof), eof.loc) + if p.peek().kind != tokenEndOfFile { + return nil, errors.MakeStaticError(fmt.Sprintf("Did not expect: %v", p.peek()), p.peek().loc) } addContext(expr, &topLevelContext, anonymous) - return expr, eof.fodder, nil + return expr, nil } // SnippetToRawAST converts a Jsonnet code snippet to an AST (without any transformations). -// Any fodder after the final token is returned as well. -func SnippetToRawAST(diagnosticFilename ast.DiagnosticFileName, importedFilename, snippet string) (ast.Node, ast.Fodder, error) { - tokens, err := Lex(diagnosticFilename, importedFilename, snippet) +func SnippetToRawAST(filename string, snippet string) (ast.Node, error) { + tokens, err := Lex(filename, snippet) if err != nil { - return nil, nil, err + return nil, err } return Parse(tokens) } diff --git a/vendor/github.com/google/go-jsonnet/internal/parser/string_util.go b/vendor/github.com/google/go-jsonnet/internal/parser/string_util.go deleted file mode 100644 index dcbbc010..00000000 --- a/vendor/github.com/google/go-jsonnet/internal/parser/string_util.go +++ /dev/null @@ -1,134 +0,0 @@ -/* -Copyright 2016 Google Inc. All rights reserved. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -package parser - -import ( - "bytes" - "encoding/hex" - "fmt" - "unicode/utf8" - - "github.com/google/go-jsonnet/ast" - "github.com/google/go-jsonnet/internal/errors" -) - -// StringUnescape compiles out the escape codes in the string -func StringUnescape(loc *ast.LocationRange, s string) (string, error) { - var buf bytes.Buffer - // read one rune at a time - for i := 0; i < len(s); { - r, w := utf8.DecodeRuneInString(s[i:]) - i += w - switch r { - case '\\': - if i >= len(s) { - return "", errors.MakeStaticError("Truncated escape sequence in string literal.", *loc) - } - r2, w := utf8.DecodeRuneInString(s[i:]) - i += w - switch r2 { - case '"': - buf.WriteRune('"') - case '\'': - buf.WriteRune('\'') - case '\\': - buf.WriteRune('\\') - case '/': - buf.WriteRune('/') // See json.org, \/ is a valid escape. - case 'b': - buf.WriteRune('\b') - case 'f': - buf.WriteRune('\f') - case 'n': - buf.WriteRune('\n') - case 'r': - buf.WriteRune('\r') - case 't': - buf.WriteRune('\t') - case 'u': - if i+4 > len(s) { - return "", errors.MakeStaticError("Truncated unicode escape sequence in string literal.", *loc) - } - codeBytes, err := hex.DecodeString(s[i : i+4]) - if err != nil { - return "", errors.MakeStaticError(fmt.Sprintf("Unicode escape sequence was malformed: %s", s[0:4]), *loc) - } - code := int(codeBytes[0])*256 + int(codeBytes[1]) - buf.WriteRune(rune(code)) - i += 4 - default: - return "", errors.MakeStaticError(fmt.Sprintf("Unknown escape sequence in string literal: \\%c", r2), *loc) - } - - default: - buf.WriteRune(r) - } - } - return buf.String(), nil -} - -// StringEscape does the opposite of StringUnescape -func StringEscape(s string, single bool) string { - var buf bytes.Buffer - // read one rune at a time - for i := 0; i < len(s); { - r, w := utf8.DecodeRuneInString(s[i:]) - i += w - switch r { - case '"': - if !single { - buf.WriteRune('\\') - } - buf.WriteRune(r) - case '\'': - if single { - buf.WriteRune('\\') - } - buf.WriteRune(r) - case '\\': - buf.WriteRune('\\') - buf.WriteRune(r) - case '\b': - buf.WriteRune('\\') - buf.WriteRune('b') - case '\f': - buf.WriteRune('\\') - buf.WriteRune('f') - case '\n': - buf.WriteRune('\\') - buf.WriteRune('n') - case '\r': - buf.WriteRune('\\') - buf.WriteRune('r') - case '\t': - buf.WriteRune('\\') - buf.WriteRune('t') - case '\u0000': - buf.WriteString("\\u0000") - - default: - if r < 0x20 || (r >= 0x7f && r <= 0x9f) { - buf.WriteRune('\\') - buf.WriteRune('u') - buf.Write([]byte(fmt.Sprintf("%04x", int(r)))) - } else { - buf.WriteRune(r) - } - } - } - return buf.String() -} diff --git a/vendor/github.com/google/go-jsonnet/internal/program/desugarer.go b/vendor/github.com/google/go-jsonnet/internal/program/desugarer.go index 98c83f75..f79d4b93 100644 --- a/vendor/github.com/google/go-jsonnet/internal/program/desugarer.go +++ b/vendor/github.com/google/go-jsonnet/internal/program/desugarer.go @@ -17,12 +17,14 @@ limitations under the License. package program import ( + "bytes" + "encoding/hex" "fmt" "reflect" + "unicode/utf8" "github.com/google/go-jsonnet/ast" "github.com/google/go-jsonnet/internal/errors" - "github.com/google/go-jsonnet/internal/parser" ) var desugaredBop = map[ast.BinaryOp]ast.Identifier{ @@ -39,6 +41,60 @@ func makeStr(s string) *ast.LiteralString { } } +func stringUnescape(loc *ast.LocationRange, s string) (string, error) { + var buf bytes.Buffer + // read one rune at a time + for i := 0; i < len(s); { + r, w := utf8.DecodeRuneInString(s[i:]) + i += w + switch r { + case '\\': + if i >= len(s) { + return "", errors.MakeStaticError("Truncated escape sequence in string literal.", *loc) + } + r2, w := utf8.DecodeRuneInString(s[i:]) + i += w + switch r2 { + case '"': + buf.WriteRune('"') + case '\'': + buf.WriteRune('\'') + case '\\': + buf.WriteRune('\\') + case '/': + buf.WriteRune('/') // See json.org, \/ is a valid escape. + case 'b': + buf.WriteRune('\b') + case 'f': + buf.WriteRune('\f') + case 'n': + buf.WriteRune('\n') + case 'r': + buf.WriteRune('\r') + case 't': + buf.WriteRune('\t') + case 'u': + if i+4 > len(s) { + return "", errors.MakeStaticError("Truncated unicode escape sequence in string literal.", *loc) + } + codeBytes, err := hex.DecodeString(s[i : i+4]) + if err != nil { + return "", errors.MakeStaticError(fmt.Sprintf("Unicode escape sequence was malformed: %s", s[0:4]), *loc) + } + code := int(codeBytes[0])*256 + int(codeBytes[1]) + buf.WriteRune(rune(code)) + i += 4 + default: + return "", errors.MakeStaticError(fmt.Sprintf("Unknown escape sequence in string literal: \\%c", r2), *loc) + } + + default: + buf.WriteRune(r) + } + } + return buf.String(), nil +} + func desugarFields(nodeBase ast.NodeBase, fields *ast.ObjectFields, objLevel int) (*ast.DesugaredObject, error) { for i := range *fields { field := &((*fields)[i]) @@ -64,9 +120,6 @@ func desugarFields(nodeBase ast.NodeBase, fields *ast.ObjectFields, objLevel int } onFailure := &ast.Error{Expr: msg} asserts = append(asserts, &ast.Conditional{ - NodeBase: ast.NodeBase{ - LocRange: field.LocRange, - }, Cond: field.Expr2, BranchTrue: &ast.LiteralBoolean{Value: true}, // ignored anyway BranchFalse: onFailure, @@ -77,7 +130,6 @@ func desugarFields(nodeBase ast.NodeBase, fields *ast.ObjectFields, objLevel int Name: makeStr(string(*field.Id)), Body: field.Expr2, PlusSuper: field.SuperSugar, - LocRange: field.LocRange, }) case ast.ObjectFieldExpr, ast.ObjectFieldStr: @@ -86,14 +138,12 @@ func desugarFields(nodeBase ast.NodeBase, fields *ast.ObjectFields, objLevel int Name: field.Expr1, Body: field.Expr2, PlusSuper: field.SuperSugar, - LocRange: field.LocRange, }) case ast.ObjectLocal: locals = append(locals, ast.LocalBind{ Variable: *field.Id, Body: ast.Clone(field.Expr2), // TODO(sbarzowski) not sure if clone is needed - LocRange: field.LocRange, }) default: panic(fmt.Sprintf("Unexpected object field kind %v", field.Kind)) @@ -117,10 +167,7 @@ func desugarFields(nodeBase ast.NodeBase, fields *ast.ObjectFields, objLevel int return nil, err } } - err := desugarLocalBinds(locals, objLevel+1) - if err != nil { - return nil, err - } + desugarLocalBinds(locals, objLevel+1) for i := range desugaredFields { field := &(desugaredFields[i]) if field.Name != nil { @@ -145,8 +192,10 @@ func desugarFields(nodeBase ast.NodeBase, fields *ast.ObjectFields, objLevel int func simpleLambda(body ast.Node, paramName ast.Identifier) ast.Node { return &ast.Function{ - Body: body, - Parameters: []ast.Parameter{{Name: paramName}}, + Body: body, + Parameters: ast.Parameters{ + Required: []ast.CommaSeparatedID{{Name: paramName}}, + }, } } @@ -155,7 +204,7 @@ func buildAnd(left ast.Node, right ast.Node) ast.Node { } // inside is assumed to be already desugared (and cannot be desugared again) -func desugarForSpec(inside ast.Node, loc ast.LocationRange, forSpec *ast.ForSpec, objLevel int) (ast.Node, error) { +func desugarForSpec(inside ast.Node, forSpec *ast.ForSpec, objLevel int) (ast.Node, error) { var body ast.Node if len(forSpec.Conditions) > 0 { cond := forSpec.Conditions[0].Expr @@ -179,11 +228,11 @@ func desugarForSpec(inside ast.Node, loc ast.LocationRange, forSpec *ast.ForSpec if err != nil { return nil, err } - current := buildStdCall("flatMap", loc, function, forSpec.Expr) + current := buildStdCall("flatMap", function, forSpec.Expr) if forSpec.Outer == nil { return current, nil } - return desugarForSpec(current, loc, forSpec.Outer, objLevel) + return desugarForSpec(current, forSpec.Outer, objLevel) } func wrapInArray(inside ast.Node) ast.Node { @@ -195,7 +244,7 @@ func desugarArrayComp(comp *ast.ArrayComp, objLevel int) (ast.Node, error) { if err != nil { return nil, err } - return desugarForSpec(wrapInArray(comp.Body), *comp.Loc(), &comp.Spec, objLevel) + return desugarForSpec(wrapInArray(comp.Body), &comp.Spec, objLevel) } func desugarObjectComp(comp *ast.ObjectComp, objLevel int) (ast.Node, error) { @@ -204,30 +253,16 @@ func desugarObjectComp(comp *ast.ObjectComp, objLevel int) (ast.Node, error) { return nil, err } - // Magic merging which follows doesn't support object locals, so we need - // to desugar them completely, i.e. put them inside the fields. The locals - // can be different for each field in a comprehension (unlike locals in - // "normal" objects which have a fixed value), so it's not even too wasteful. - if len(obj.Locals) > 0 { - field := &obj.Fields[0] - field.Body = &ast.Local{ - Body: field.Body, - Binds: obj.Locals, - // TODO(sbarzowski) should I set some NodeBase stuff here? - } - obj.Locals = nil - } - if len(obj.Fields) != 1 { - panic("Wrong number of fields in object comprehension, it should have been caught during parsing") + panic("Too many fields in object comprehension, it should have been caught during parsing") } - desugaredArrayComp, err := desugarForSpec(wrapInArray(obj), *comp.Loc(), &comp.Spec, objLevel) + desugaredArrayComp, err := desugarForSpec(wrapInArray(obj), &comp.Spec, objLevel) if err != nil { return nil, err } - desugaredComp := buildStdCall("$objectFlatMerge", *comp.Loc(), desugaredArrayComp) + desugaredComp := buildStdCall("$objectFlatMerge", desugaredArrayComp) return desugaredComp, nil } @@ -245,7 +280,7 @@ func buildSimpleIndex(obj ast.Node, member ast.Identifier) ast.Node { } } -func buildStdCall(builtinName ast.Identifier, loc ast.LocationRange, args ...ast.Node) ast.Node { +func buildStdCall(builtinName ast.Identifier, args ...ast.Node) ast.Node { std := &ast.Var{Id: "std"} builtin := buildSimpleIndex(std, builtinName) positional := make([]ast.CommaSeparatedExpr, len(args)) @@ -253,9 +288,6 @@ func buildStdCall(builtinName ast.Identifier, loc ast.LocationRange, args ...ast positional[i].Expr = args[i] } return &ast.Apply{ - NodeBase: ast.NodeBase{ - LocRange: loc, - }, Target: builtin, Arguments: ast.Arguments{Positional: positional}, } @@ -339,14 +371,9 @@ func desugar(astPtr *ast.Node, objLevel int) (err error) { node.Message = buildLiteralString("Assertion failed") } *astPtr = &ast.Conditional{ - Cond: node.Cond, - BranchTrue: node.Rest, - BranchFalse: &ast.Error{ - NodeBase: ast.NodeBase{ - LocRange: *node.Loc(), - }, - Expr: node.Message, - }, + Cond: node.Cond, + BranchTrue: node.Rest, + BranchFalse: &ast.Error{Expr: node.Message}, } err = desugar(astPtr, objLevel) if err != nil { @@ -358,9 +385,9 @@ func desugar(astPtr *ast.Node, objLevel int) (err error) { if funcname, replaced := desugaredBop[node.Op]; replaced { if node.Op == ast.BopIn { // reversed order of arguments - *astPtr = buildStdCall(funcname, *node.Loc(), node.Right, node.Left) + *astPtr = buildStdCall(funcname, node.Right, node.Left) } else { - *astPtr = buildStdCall(funcname, *node.Loc(), node.Left, node.Right) + *astPtr = buildStdCall(funcname, node.Left, node.Right) } return desugar(astPtr, objLevel) } @@ -404,13 +431,11 @@ func desugar(astPtr *ast.Node, objLevel int) (err error) { } case *ast.Function: - for i := range node.Parameters { - param := &node.Parameters[i] - if param.DefaultArg != nil { - err = desugar(¶m.DefaultArg, objLevel) - if err != nil { - return - } + for i := range node.Parameters.Optional { + param := &node.Parameters.Optional[i] + err = desugar(¶m.DefaultArg, objLevel) + if err != nil { + return } } err = desugar(&node.Body, objLevel) @@ -463,7 +488,7 @@ func desugar(astPtr *ast.Node, objLevel int) (err error) { if node.Step == nil { node.Step = &ast.LiteralNull{} } - *astPtr = buildStdCall("slice", *node.Loc(), node.Target, node.BeginIndex, node.EndIndex, node.Step) + *astPtr = buildStdCall("slice", node.Target, node.BeginIndex, node.EndIndex, node.Step) err = desugar(astPtr, objLevel) if err != nil { return @@ -490,7 +515,7 @@ func desugar(astPtr *ast.Node, objLevel int) (err error) { case *ast.LiteralString: if node.Kind.FullyEscaped() { - unescaped, err := parser.StringUnescape(node.Loc(), node.Value) + unescaped, err := stringUnescape(node.Loc(), node.Value) if err != nil { return err } diff --git a/vendor/github.com/google/go-jsonnet/internal/program/program.go b/vendor/github.com/google/go-jsonnet/internal/program/program.go index 32014d37..91130d98 100644 --- a/vendor/github.com/google/go-jsonnet/internal/program/program.go +++ b/vendor/github.com/google/go-jsonnet/internal/program/program.go @@ -6,8 +6,8 @@ import ( ) // SnippetToAST converts a Jsonnet code snippet to a desugared and analyzed AST. -func SnippetToAST(diagnosticFilename ast.DiagnosticFileName, importedFilename, snippet string) (ast.Node, error) { - node, _, err := parser.SnippetToRawAST(diagnosticFilename, importedFilename, snippet) +func SnippetToAST(filename string, snippet string) (ast.Node, error) { + node, err := parser.SnippetToRawAST(filename, snippet) if err != nil { return nil, err } diff --git a/vendor/github.com/google/go-jsonnet/internal/program/static_analyzer.go b/vendor/github.com/google/go-jsonnet/internal/program/static_analyzer.go index 21c42166..53f004e4 100644 --- a/vendor/github.com/google/go-jsonnet/internal/program/static_analyzer.go +++ b/vendor/github.com/google/go-jsonnet/internal/program/static_analyzer.go @@ -76,17 +76,21 @@ func analyzeVisit(a ast.Node, inObject bool, vars ast.IdentifierSet) error { visitNext(a.Expr, inObject, vars, s) case *ast.Function: newVars := vars.Clone() - for _, param := range a.Parameters { + for _, param := range a.Parameters.Required { newVars.Add(param.Name) } - for _, param := range a.Parameters { - if param.DefaultArg != nil { - visitNext(param.DefaultArg, inObject, newVars, s) - } + for _, param := range a.Parameters.Optional { + newVars.Add(param.Name) + } + for _, param := range a.Parameters.Optional { + visitNext(param.DefaultArg, inObject, newVars, s) } visitNext(a.Body, inObject, newVars, s) // Parameters are free inside the body, but not visible here or outside - for _, param := range a.Parameters { + for _, param := range a.Parameters.Required { + s.freeVars.Remove(param.Name) + } + for _, param := range a.Parameters.Optional { s.freeVars.Remove(param.Name) } case *ast.Import: diff --git a/vendor/github.com/google/go-jsonnet/interpreter.go b/vendor/github.com/google/go-jsonnet/interpreter.go index 087223db..9986eda1 100644 --- a/vendor/github.com/google/go-jsonnet/interpreter.go +++ b/vendor/github.com/google/go-jsonnet/interpreter.go @@ -22,7 +22,6 @@ import ( "math" "reflect" "sort" - "strconv" "github.com/google/go-jsonnet/ast" "github.com/google/go-jsonnet/astgen" @@ -49,15 +48,15 @@ func makeEnvironment(upValues bindingFrame, sb selfBinding) environment { } } -func (i *interpreter) getCurrentStackTrace() []traceFrame { +func (i *interpreter) getCurrentStackTrace(additional traceElement) []traceFrame { var result []traceFrame for _, f := range i.stack.stack { if f.isCall { result = append(result, traceElementToTraceFrame(f.trace)) } } - if i.stack.currentTrace.loc != nil { - result = append(result, traceElementToTraceFrame(i.stack.currentTrace)) + if additional.loc != nil { + result = append(result, traceElementToTraceFrame(additional)) } return result } @@ -95,10 +94,9 @@ func dumpCallFrame(c *callFrame) string { } type callStack struct { - calls int - limit int - stack []*callFrame - currentTrace traceElement + calls int + limit int + stack []*callFrame } func dumpCallStack(c *callStack) string { @@ -124,7 +122,6 @@ func (s *callStack) popIfExists(whichFrame int) { if s.top().isCall { s.calls-- } - s.setCurrentTrace(s.stack[len(s.stack)-1].trace) s.stack = s.stack[:len(s.stack)-1] } } @@ -144,17 +141,6 @@ func (s *callStack) tailCallTrimStack() { } } -func (s *callStack) setCurrentTrace(trace traceElement) { - if s.currentTrace != (traceElement{}) { - panic("Tried to change the traceElement while the old one was still there.") - } - s.currentTrace = trace -} - -func (s *callStack) clearCurrentTrace() { - s.currentTrace = traceElement{} -} - type tailCallStatus int const ( @@ -162,26 +148,20 @@ const ( tailCall ) -func (s *callStack) newCall(env environment, trimmable bool) { - if s.currentTrace == (traceElement{}) { - panic("Saving empty traceElement on stack") - } +func (s *callStack) newCall(trace traceElement, env environment, trimmable bool) { s.stack = append(s.stack, &callFrame{ isCall: true, - trace: s.currentTrace, + trace: trace, env: env, trimmable: trimmable, }) - s.clearCurrentTrace() s.calls++ } func (s *callStack) newLocal(vars bindingFrame) { s.stack = append(s.stack, &callFrame{ - env: makeEnvironment(vars, selfBinding{}), - trace: s.currentTrace, + env: makeEnvironment(vars, selfBinding{}), }) - s.clearCurrentTrace() } // getSelfBinding resolves the self construct @@ -224,6 +204,14 @@ func (s *callStack) getCurrentEnv(ast ast.Node) environment { ) } +func (s *callStack) getTopEnv() environment { + top := s.stack[len(s.stack)-1] + if !top.isCall { + panic("getTopEnv is allowed only for artifical nodes which are called in new environment") + } + return top.env +} + // Build a binding frame containing specified variables. func (s *callStack) capture(freeVars ast.Identifiers) bindingFrame { env := make(bindingFrame) @@ -275,12 +263,12 @@ func addBindings(a, b bindingFrame) bindingFrame { return result } -func (i *interpreter) newCall(env environment, trimmable bool) error { +func (i *interpreter) newCall(trace traceElement, env environment, trimmable bool) error { s := &i.stack if s.calls >= s.limit { - return makeRuntimeError("max stack frames exceeded.", i.getCurrentStackTrace()) + return makeRuntimeError("max stack frames exceeded.", i.getCurrentStackTrace(trace)) } - s.newCall(env, trimmable) + s.newCall(trace, env, trimmable) return nil } @@ -289,10 +277,6 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { loc: a.Loc(), context: a.Context(), } - oldTrace := i.stack.currentTrace - i.stack.clearCurrentTrace() - i.stack.setCurrentTrace(trace) - defer func() { i.stack.clearCurrentTrace(); i.stack.setCurrentTrace(oldTrace) }() switch node := a.(type) { case *ast.Array: @@ -312,7 +296,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { if err != nil { return nil, err } - x, err := i.getBoolean(xv) + x, err := i.getBoolean(xv, trace) if err != nil { return nil, err } @@ -323,14 +307,14 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { if err != nil { return nil, err } - return i.getBoolean(yv) + return i.getBoolean(yv, trace) } else if node.Op == ast.BopOr { // Special case for shortcut semantics. xv, err := i.evaluate(node.Left, nonTailCall) if err != nil { return nil, err } - x, err := i.getBoolean(xv) + x, err := i.getBoolean(xv, trace) if err != nil { return nil, err } @@ -341,7 +325,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { if err != nil { return nil, err } - return i.getBoolean(yv) + return i.getBoolean(yv, trace) } else { left, err := i.evaluate(node.Left, nonTailCall) @@ -354,7 +338,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { } // TODO(dcunnin): The double dereference here is probably not necessary. builtin := bopBuiltins[node.Op] - return builtin.function(i, left, right) + return builtin.function(i, trace, left, right) } case *ast.Unary: @@ -365,7 +349,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { builtin := uopBuiltins[node.Op] - result, err := builtin.function(i, value) + result, err := builtin.function(i, trace, value) if err != nil { return nil, err } @@ -376,7 +360,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { if err != nil { return nil, err } - condBool, err := i.getBoolean(cond) + condBool, err := i.getBoolean(cond, trace) if err != nil { return nil, err } @@ -401,11 +385,11 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { // Omitted field. continue default: - return nil, i.Error(fmt.Sprintf("Field name must be string, got %v", fieldNameValue.getType().name)) + return nil, i.Error(fmt.Sprintf("Field name must be string, got %v", fieldNameValue.getType().name), trace) } if _, ok := fields[fieldName]; ok { - return nil, i.Error(duplicateFieldNameErrMsg(fieldName)) + return nil, i.Error(duplicateFieldNameErrMsg(fieldName), trace) } var f unboundField = &codeUnboundField{field.Body} if field.PlusSuper { @@ -431,16 +415,16 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { return nil, err } if msgVal.getType() != stringType { - msgVal, err = builtinToString(i, msgVal) + msgVal, err = builtinToString(i, trace, msgVal) if err != nil { return nil, err } } - msg, err := i.getString(msgVal) + msg, err := i.getString(msgVal, trace) if err != nil { return nil, err } - return nil, i.Error(msg.getGoString()) + return nil, i.Error(msg.getGoString(), trace) case *ast.Index: targetValue, err := i.evaluate(node.Target, nonTailCall) @@ -453,37 +437,37 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { } switch target := targetValue.(type) { case *valueObject: - indexString, err := i.getString(index) + indexString, err := i.getString(index, trace) if err != nil { return nil, err } - return target.index(i, indexString.getGoString()) + return target.index(i, trace, indexString.getGoString()) case *valueArray: - indexInt, err := i.getNumber(index) + indexInt, err := i.getNumber(index, trace) if err != nil { return nil, err } // TODO(https://github.com/google/jsonnet/issues/377): non-integer indexes should be an error - return target.index(i, int(indexInt.value)) + return target.index(i, trace, int(indexInt.value)) case valueString: - indexInt, err := i.getNumber(index) + indexInt, err := i.getNumber(index, trace) if err != nil { return nil, err } // TODO(https://github.com/google/jsonnet/issues/377): non-integer indexes should be an error - return target.index(i, int(indexInt.value)) + return target.index(i, trace, int(indexInt.value)) } - return nil, i.Error(fmt.Sprintf("Value non indexable: %v", reflect.TypeOf(targetValue))) + return nil, i.Error(fmt.Sprintf("Value non indexable: %v", reflect.TypeOf(targetValue)), trace) case *ast.Import: codePath := node.Loc().FileName - return i.importCache.importCode(codePath, node.File.Value, i) + return i.importCache.importCode(codePath, node.File.Value, i, trace) case *ast.ImportStr: codePath := node.Loc().FileName - return i.importCache.importString(codePath, node.File.Value, i) + return i.importCache.importString(codePath, node.File.Value, i, trace) case *ast.LiteralBoolean: return makeValueBoolean(node.Value), nil @@ -492,14 +476,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { return makeValueNull(), nil case *ast.LiteralNumber: - // Since the lexer ensures that OriginalString is of - // the right form, this will only fail if the number is - // too large to fit in a double. - num, err := strconv.ParseFloat(node.OriginalString, 64) - if err != nil { - return nil, i.Error("overflow") - } - return makeValueNumber(num), nil + return makeValueNumber(node.Value), nil case *ast.LiteralString: return makeValueString(node.Value), nil @@ -529,25 +506,25 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { case *ast.Var: foo := i.stack.lookUpVarOrPanic(node.Id) - return foo.getValue(i) + return foo.getValue(i, trace) case *ast.SuperIndex: index, err := i.evaluate(node.Index, nonTailCall) if err != nil { return nil, err } - indexStr, err := i.getString(index) + indexStr, err := i.getString(index, trace) if err != nil { return nil, err } - return objectIndex(i, i.stack.getSelfBinding().super(), indexStr.getGoString()) + return objectIndex(i, trace, i.stack.getSelfBinding().super(), indexStr.getGoString()) case *ast.InSuper: index, err := i.evaluate(node.Index, nonTailCall) if err != nil { return nil, err } - indexStr, err := i.getString(index) + indexStr, err := i.getString(index, trace) if err != nil { return nil, err } @@ -565,7 +542,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { if err != nil { return nil, err } - function, err := i.getFunction(target) + function, err := i.getFunction(target, trace) if err != nil { return nil, err } @@ -584,7 +561,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { for i, arg := range node.Arguments.Named { arguments.named[i] = namedCallArgument{name: arg.Name, pv: &cachedThunk{env: &argEnv, body: arg.Arg}} } - return i.evaluateTailCall(function, arguments, tc) + return i.evaluateTailCall(function, arguments, tc, trace) case *astMakeArrayElement: arguments := callArguments{ @@ -594,7 +571,7 @@ func (i *interpreter) evaluate(a ast.Node, tc tailCallStatus) (value, error) { }, }, } - return i.evaluateTailCall(node.function, arguments, tc) + return i.evaluateTailCall(node.function, arguments, tc, trace) default: panic(fmt.Sprintf("Executing this AST type not implemented: %v", reflect.TypeOf(a))) @@ -647,20 +624,14 @@ func unparseNumber(v float64) string { } // manifestJSON converts to standard JSON representation as in "encoding/json" package -func (i *interpreter) manifestJSON(v value) (interface{}, error) { - // TODO(sbarzowski) Add nice stack traces indicating the part of the code which - // evaluates to non-manifestable value (that might require passing context about - // the root value) - if i.stack.currentTrace == (traceElement{}) { - panic("manifesting JSON with empty traceElement") - } +func (i *interpreter) manifestJSON(trace traceElement, v value) (interface{}, error) { switch v := v.(type) { case *valueBoolean: return v.value, nil case *valueFunction: - return nil, makeRuntimeError("couldn't manifest function as JSON", i.getCurrentStackTrace()) + return nil, makeRuntimeError("couldn't manifest function in JSON output.", i.getCurrentStackTrace(trace)) case *valueNumber: return v.value, nil @@ -674,11 +645,11 @@ func (i *interpreter) manifestJSON(v value) (interface{}, error) { case *valueArray: result := make([]interface{}, 0, len(v.elements)) for _, th := range v.elements { - elVal, err := i.evaluatePV(th) + elVal, err := i.evaluatePV(th, trace) if err != nil { return nil, err } - elem, err := i.manifestJSON(elVal) + elem, err := i.manifestJSON(trace, elVal) if err != nil { return nil, err } @@ -690,7 +661,7 @@ func (i *interpreter) manifestJSON(v value) (interface{}, error) { fieldNames := objectFields(v, withoutHidden) sort.Strings(fieldNames) - err := checkAssertions(i, v) + err := checkAssertions(i, trace, v) if err != nil { return nil, err } @@ -698,12 +669,12 @@ func (i *interpreter) manifestJSON(v value) (interface{}, error) { result := make(map[string]interface{}) for _, fieldName := range fieldNames { - fieldVal, err := v.index(i, fieldName) + fieldVal, err := v.index(i, trace, fieldName) if err != nil { return nil, err } - field, err := i.manifestJSON(fieldVal) + field, err := i.manifestJSON(trace, fieldVal) if err != nil { return nil, err } @@ -715,7 +686,7 @@ func (i *interpreter) manifestJSON(v value) (interface{}, error) { default: return nil, makeRuntimeError( fmt.Sprintf("manifesting this value not implemented yet: %s", reflect.TypeOf(v)), - i.getCurrentStackTrace(), + i.getCurrentStackTrace(trace), ) } @@ -819,8 +790,8 @@ func serializeJSON(v interface{}, multiline bool, indent string, buf *bytes.Buff } func (i *interpreter) manifestAndSerializeJSON( - buf *bytes.Buffer, v value, multiline bool, indent string) error { - manifested, err := i.manifestJSON(v) + buf *bytes.Buffer, trace traceElement, v value, multiline bool, indent string) error { + manifested, err := i.manifestJSON(trace, v) if err != nil { return err } @@ -829,19 +800,19 @@ func (i *interpreter) manifestAndSerializeJSON( } // manifestString expects the value to be a string and returns it. -func (i *interpreter) manifestString(buf *bytes.Buffer, v value) error { +func (i *interpreter) manifestString(buf *bytes.Buffer, trace traceElement, v value) error { switch v := v.(type) { case valueString: buf.WriteString(v.getGoString()) return nil default: - return makeRuntimeError(fmt.Sprintf("expected string result, got: %s", v.getType().name), i.getCurrentStackTrace()) + return makeRuntimeError(fmt.Sprintf("expected string result, got: %s", v.getType().name), i.getCurrentStackTrace(trace)) } } -func (i *interpreter) manifestAndSerializeMulti(v value, stringOutputMode bool) (r map[string]string, err error) { +func (i *interpreter) manifestAndSerializeMulti(trace traceElement, v value, stringOutputMode bool) (r map[string]string, err error) { r = make(map[string]string) - json, err := i.manifestJSON(v) + json, err := i.manifestJSON(trace, v) if err != nil { return r, err } @@ -855,7 +826,7 @@ func (i *interpreter) manifestAndSerializeMulti(v value, stringOutputMode bool) default: msg := fmt.Sprintf("multi mode: top-level object's key %s has a value of type %T, "+ "should be a string", filename, val) - return r, makeRuntimeError(msg, i.getCurrentStackTrace()) + return r, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } } else { var buf bytes.Buffer @@ -868,14 +839,14 @@ func (i *interpreter) manifestAndSerializeMulti(v value, stringOutputMode bool) msg := fmt.Sprintf("multi mode: top-level object was a %s, "+ "should be an object whose keys are filenames and values hold "+ "the JSON for that file.", v.getType().name) - return r, makeRuntimeError(msg, i.getCurrentStackTrace()) + return r, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } return } -func (i *interpreter) manifestAndSerializeYAMLStream(v value) (r []string, err error) { +func (i *interpreter) manifestAndSerializeYAMLStream(trace traceElement, v value) (r []string, err error) { r = make([]string, 0) - json, err := i.manifestJSON(v) + json, err := i.manifestJSON(trace, v) if err != nil { return r, err } @@ -891,12 +862,12 @@ func (i *interpreter) manifestAndSerializeYAMLStream(v value) (r []string, err e msg := fmt.Sprintf("stream mode: top-level object was a %s, "+ "should be an array whose elements hold "+ "the JSON for each document in the stream.", v.getType().name) - return r, makeRuntimeError(msg, i.getCurrentStackTrace()) + return r, makeRuntimeError(msg, i.getCurrentStackTrace(trace)) } return } -func jsonToValue(i *interpreter, v interface{}) (value, error) { +func jsonToValue(i *interpreter, trace traceElement, v interface{}) (value, error) { switch v := v.(type) { case nil: return &nullValue, nil @@ -904,7 +875,7 @@ func jsonToValue(i *interpreter, v interface{}) (value, error) { case []interface{}: elems := make([]*cachedThunk, len(v)) for counter, elem := range v { - val, err := jsonToValue(i, elem) + val, err := jsonToValue(i, trace, elem) if err != nil { return nil, err } @@ -915,12 +886,12 @@ func jsonToValue(i *interpreter, v interface{}) (value, error) { case bool: return makeValueBoolean(v), nil case float64: - return makeDoubleCheck(i, v) + return makeDoubleCheck(i, trace, v) case map[string]interface{}: fieldMap := map[string]value{} for name, f := range v { - val, err := jsonToValue(i, f) + val, err := jsonToValue(i, trace, f) if err != nil { return nil, err } @@ -932,12 +903,12 @@ func jsonToValue(i *interpreter, v interface{}) (value, error) { return makeValueString(v), nil default: - return nil, i.Error(fmt.Sprintf("Not a json type: %#+v", v)) + return nil, i.Error(fmt.Sprintf("Not a json type: %#+v", v), trace) } } -func (i *interpreter) EvalInCleanEnv(env *environment, ast ast.Node, trimmable bool) (value, error) { - err := i.newCall(*env, trimmable) +func (i *interpreter) EvalInCleanEnv(fromWhere traceElement, env *environment, ast ast.Node, trimmable bool) (value, error) { + err := i.newCall(fromWhere, *env, trimmable) if err != nil { return nil, err } @@ -950,54 +921,55 @@ func (i *interpreter) EvalInCleanEnv(env *environment, ast ast.Node, trimmable b return val, err } -func (i *interpreter) evaluatePV(ph potentialValue) (value, error) { - return ph.getValue(i) +func (i *interpreter) evaluatePV(ph potentialValue, trace traceElement) (value, error) { + return ph.getValue(i, trace) } -func (i *interpreter) evaluateTailCall(function *valueFunction, args callArguments, tc tailCallStatus) (value, error) { +func (i *interpreter) evaluateTailCall(function *valueFunction, args callArguments, tc tailCallStatus, trace traceElement) (value, error) { if tc == tailCall { i.stack.tailCallTrimStack() } - return function.call(i, args) + return function.call(i, trace, args) } -func (i *interpreter) Error(s string) error { - err := makeRuntimeError(s, i.getCurrentStackTrace()) +func (i *interpreter) Error(s string, trace traceElement) error { + err := makeRuntimeError(s, i.getCurrentStackTrace(trace)) return err } -func (i *interpreter) typeErrorSpecific(bad value, good value) error { +func (i *interpreter) typeErrorSpecific(bad value, good value, trace traceElement) error { return i.Error( fmt.Sprintf("Unexpected type %v, expected %v", bad.getType().name, good.getType().name), + trace, ) } -func (i *interpreter) typeErrorGeneral(bad value) error { +func (i *interpreter) typeErrorGeneral(bad value, trace traceElement) error { return i.Error( fmt.Sprintf("Unexpected type %v", bad.getType().name), + trace, ) } -func (i *interpreter) getNumber(val value) (*valueNumber, error) { +func (i *interpreter) getNumber(val value, trace traceElement) (*valueNumber, error) { switch v := val.(type) { case *valueNumber: return v, nil default: - return nil, i.typeErrorSpecific(val, &valueNumber{}) + return nil, i.typeErrorSpecific(val, &valueNumber{}, trace) } } -//nolint:unused -func (i *interpreter) evaluateNumber(pv potentialValue) (*valueNumber, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateNumber(pv potentialValue, trace traceElement) (*valueNumber, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return nil, err } - return i.getNumber(v) + return i.getNumber(v, trace) } -func (i *interpreter) getInt(val value) (int, error) { - num, err := i.getNumber(val) +func (i *interpreter) getInt(val value, trace traceElement) (int, error) { + num, err := i.getNumber(val, trace) if err != nil { return 0, err } @@ -1005,128 +977,122 @@ func (i *interpreter) getInt(val value) (int, error) { // on any machine. And it's used only for indexing anyway. intNum := int(int32(num.value)) if float64(intNum) != num.value { - return 0, i.Error(fmt.Sprintf("Expected an integer, but got %v", num.value)) + return 0, i.Error(fmt.Sprintf("Expected an integer, but got %v", num.value), trace) } return intNum, nil } -func (i *interpreter) evaluateInt(pv potentialValue) (int, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateInt(pv potentialValue, trace traceElement) (int, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return 0, err } - return i.getInt(v) + return i.getInt(v, trace) } -//nolint:unused -func (i *interpreter) getInt64(val value) (int64, error) { - num, err := i.getNumber(val) +func (i *interpreter) getInt64(val value, trace traceElement) (int64, error) { + num, err := i.getNumber(val, trace) if err != nil { return 0, err } intNum := int64(num.value) if float64(intNum) != num.value { - return 0, i.Error(fmt.Sprintf("Expected an integer, but got %v", num.value)) + return 0, i.Error(fmt.Sprintf("Expected an integer, but got %v", num.value), trace) } return intNum, nil } -//nolint:unused -func (i *interpreter) evaluateInt64(pv potentialValue) (int64, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateInt64(pv potentialValue, trace traceElement) (int64, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return 0, err } - return i.getInt64(v) + return i.getInt64(v, trace) } -func (i *interpreter) getString(val value) (valueString, error) { +func (i *interpreter) getString(val value, trace traceElement) (valueString, error) { switch v := val.(type) { case valueString: return v, nil default: - return nil, i.typeErrorSpecific(val, emptyString()) + return nil, i.typeErrorSpecific(val, emptyString(), trace) } } -//nolint:unused -func (i *interpreter) evaluateString(pv potentialValue) (valueString, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateString(pv potentialValue, trace traceElement) (valueString, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return nil, err } - return i.getString(v) + return i.getString(v, trace) } -func (i *interpreter) getBoolean(val value) (*valueBoolean, error) { +func (i *interpreter) getBoolean(val value, trace traceElement) (*valueBoolean, error) { switch v := val.(type) { case *valueBoolean: return v, nil default: - return nil, i.typeErrorSpecific(val, &valueBoolean{}) + return nil, i.typeErrorSpecific(val, &valueBoolean{}, trace) } } -//nolint:unused -func (i *interpreter) evaluateBoolean(pv potentialValue) (*valueBoolean, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateBoolean(pv potentialValue, trace traceElement) (*valueBoolean, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return nil, err } - return i.getBoolean(v) + return i.getBoolean(v, trace) } -func (i *interpreter) getArray(val value) (*valueArray, error) { +func (i *interpreter) getArray(val value, trace traceElement) (*valueArray, error) { switch v := val.(type) { case *valueArray: return v, nil default: - return nil, i.typeErrorSpecific(val, &valueArray{}) + return nil, i.typeErrorSpecific(val, &valueArray{}, trace) } } -//nolint:unused -func (i *interpreter) evaluateArray(pv potentialValue) (*valueArray, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateArray(pv potentialValue, trace traceElement) (*valueArray, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return nil, err } - return i.getArray(v) + return i.getArray(v, trace) } -func (i *interpreter) getFunction(val value) (*valueFunction, error) { +func (i *interpreter) getFunction(val value, trace traceElement) (*valueFunction, error) { switch v := val.(type) { case *valueFunction: return v, nil default: - return nil, i.typeErrorSpecific(val, &valueFunction{}) + return nil, i.typeErrorSpecific(val, &valueFunction{}, trace) } } -//nolint:unused -func (i *interpreter) evaluateFunction(pv potentialValue) (*valueFunction, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateFunction(pv potentialValue, trace traceElement) (*valueFunction, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return nil, err } - return i.getFunction(v) + return i.getFunction(v, trace) } -func (i *interpreter) getObject(val value) (*valueObject, error) { +func (i *interpreter) getObject(val value, trace traceElement) (*valueObject, error) { switch v := val.(type) { case *valueObject: return v, nil default: - return nil, i.typeErrorSpecific(val, &valueObject{}) + return nil, i.typeErrorSpecific(val, &valueObject{}, trace) } } -func (i *interpreter) evaluateObject(pv potentialValue) (*valueObject, error) { - v, err := i.evaluatePV(pv) +func (i *interpreter) evaluateObject(pv potentialValue, trace traceElement) (*valueObject, error) { + v, err := i.evaluatePV(pv, trace) if err != nil { return nil, err } - return i.getObject(v) + return i.getObject(v, trace) } func buildStdObject(i *interpreter) (*valueObject, error) { @@ -1155,9 +1121,7 @@ func evaluateStd(i *interpreter) (value, error) { evalLoc := ast.MakeLocationRangeMessage("During evaluation of std") evalTrace := traceElement{loc: &evalLoc} node := astgen.StdAst - i.stack.setCurrentTrace(evalTrace) - defer i.stack.clearCurrentTrace() - return i.EvalInCleanEnv(&beforeStdEnv, node, false) + return i.EvalInCleanEnv(evalTrace, &beforeStdEnv, node, false) } func prepareExtVars(i *interpreter, ext vmExtMap, kind string) map[string]*cachedThunk { @@ -1217,9 +1181,7 @@ func evaluateAux(i *interpreter, node ast.Node, tla vmExtMap) (value, traceEleme loc: &evalLoc, } env := makeInitialEnv(node.Loc().FileName, i.baseStd) - i.stack.setCurrentTrace(evalTrace) - result, err := i.EvalInCleanEnv(&env, node, false) - i.stack.clearCurrentTrace() + result, err := i.EvalInCleanEnv(evalTrace, &env, node, false) if err != nil { return nil, traceElement{}, err } @@ -1230,13 +1192,11 @@ func evaluateAux(i *interpreter, node ast.Node, tla vmExtMap) (value, traceEleme for argName, pv := range toplevelArgMap { args.named = append(args.named, namedCallArgument{name: ast.Identifier(argName), pv: pv}) } - funcLoc := ast.MakeLocationRangeMessage("Top-level function call") + funcLoc := ast.MakeLocationRangeMessage("Top-level function") funcTrace := traceElement{ loc: &funcLoc, } - i.stack.setCurrentTrace(funcTrace) - result, err = f.call(i, args) - i.stack.clearCurrentTrace() + result, err = f.call(i, funcTrace, args) if err != nil { return nil, traceElement{}, err } @@ -1263,13 +1223,11 @@ func evaluate(node ast.Node, ext vmExtMap, tla vmExtMap, nativeFuncs map[string] } var buf bytes.Buffer - i.stack.setCurrentTrace(manifestationTrace) if stringOutputMode { - err = i.manifestString(&buf, result) + err = i.manifestString(&buf, manifestationTrace, result) } else { - err = i.manifestAndSerializeJSON(&buf, result, true, "") + err = i.manifestAndSerializeJSON(&buf, manifestationTrace, result, true, "") } - i.stack.clearCurrentTrace() if err != nil { return "", err } @@ -1291,10 +1249,7 @@ func evaluateMulti(node ast.Node, ext vmExtMap, tla vmExtMap, nativeFuncs map[st return nil, err } - i.stack.setCurrentTrace(manifestationTrace) - manifested, err := i.manifestAndSerializeMulti(result, stringOutputMode) - i.stack.clearCurrentTrace() - return manifested, err + return i.manifestAndSerializeMulti(manifestationTrace, result, stringOutputMode) } // TODO(sbarzowski) this function takes far too many arguments - build interpreter in vm instead @@ -1311,8 +1266,5 @@ func evaluateStream(node ast.Node, ext vmExtMap, tla vmExtMap, nativeFuncs map[s return nil, err } - i.stack.setCurrentTrace(manifestationTrace) - manifested, err := i.manifestAndSerializeYAMLStream(result) - i.stack.clearCurrentTrace() - return manifested, err + return i.manifestAndSerializeYAMLStream(manifestationTrace, result) } diff --git a/vendor/github.com/google/go-jsonnet/setup.py b/vendor/github.com/google/go-jsonnet/setup.py index 242c6ae7..0e184d32 100644 --- a/vendor/github.com/google/go-jsonnet/setup.py +++ b/vendor/github.com/google/go-jsonnet/setup.py @@ -16,7 +16,6 @@ from setuptools import setup from setuptools import Extension from setuptools.command.build_ext import build_ext as BuildExt -from setuptools.command.test import test as TestCommand from subprocess import Popen, PIPE DIR = os.path.abspath(os.path.dirname(__file__)) @@ -46,10 +45,6 @@ def run(self): BuildExt.run(self) -class NoopTestCommand(TestCommand): - def __init__(self, dist): - print("_gojsonnet does not support running tests with 'python setup.py test'. Please run 'pytest'.") - jsonnet_ext = Extension( '_gojsonnet', sources=MODULE_SOURCES, @@ -68,7 +63,7 @@ def __init__(self, dist): version=get_version(), cmdclass={ 'build_ext': BuildJsonnetExt, - 'test': NoopTestCommand, }, ext_modules=[jsonnet_ext], + test_suite="python._jsonnet_test", ) diff --git a/vendor/github.com/google/go-jsonnet/tests.sh b/vendor/github.com/google/go-jsonnet/tests.sh index a2818644..80e0bb0c 100644 --- a/vendor/github.com/google/go-jsonnet/tests.sh +++ b/vendor/github.com/google/go-jsonnet/tests.sh @@ -3,32 +3,27 @@ set -e PYTHON_COMMAND=${PYTHON_COMMAND:=python} -JSONNET_CPP_DIR=${JSONNET_CPP_DIR:=$PWD/cpp-jsonnet} set -x -[ "$SKIP_GO_TESTS" == 1 ] || go test ./... -if [ "$SKIP_PYTHON_BINDINGS_TESTS" == 1 ] -then - c-bindings-tests/build.sh -else - c-bindings-tests/run.sh +[ "$1" = "--skip-go-test" ] || go test ./... - $PYTHON_COMMAND setup.py build --build-platlib . - $PYTHON_COMMAND -m pytest python -fi + +c-bindings-tests/run.sh + +$PYTHON_COMMAND setup.py build +$PYTHON_COMMAND setup.py test export IMPLEMENTATION=golang -export OVERRIDE_DIR="$PWD/testdata/cpp-tests-override/" go build ./cmd/jsonnet -go build ./cmd/jsonnetfmt export DISABLE_LIB_TESTS=true +export DISABLE_FMT_TESTS=true export DISABLE_ERROR_TESTS=true -export JSONNETFMT_BIN="$PWD/jsonnetfmt" export JSONNET_BIN="$PWD/jsonnet" -cd "$JSONNET_CPP_DIR" +git submodule update --recursive cpp-jsonnet +cd cpp-jsonnet exec ./tests.sh diff --git a/vendor/github.com/google/go-jsonnet/thunks.go b/vendor/github.com/google/go-jsonnet/thunks.go index 1cd4175a..ed0618ec 100644 --- a/vendor/github.com/google/go-jsonnet/thunks.go +++ b/vendor/github.com/google/go-jsonnet/thunks.go @@ -16,9 +16,7 @@ limitations under the License. package jsonnet -import ( - "github.com/google/go-jsonnet/ast" -) +import "github.com/google/go-jsonnet/ast" // readyValue // ------------------------------------- @@ -32,13 +30,23 @@ type readyValue struct { content value } -func (rv *readyValue) evaluate(i *interpreter, sb selfBinding, origBinding bindingFrame, fieldName string) (value, error) { +func (rv *readyValue) evaluate(i *interpreter, trace traceElement, sb selfBinding, origBinding bindingFrame, fieldName string) (value, error) { return rv.content, nil } +func (rv *readyValue) aPotentialValue() {} + // potentialValues // ------------------------------------- +// evaluable is something that can be evaluated and the result is always the same +// It may require computation every time evaluation is requested (in contrast with +// potentialValue which guarantees that computation happens at most once). +type evaluable interface { + // fromWhere keeps the information from where the evaluation was requested. + getValue(i *interpreter, fromWhere traceElement) (value, error) +} + // cachedThunk is a wrapper that caches the value of a potentialValue after // the first evaluation. // Note: All potentialValues are required to provide the same value every time, @@ -58,14 +66,14 @@ func readyThunk(content value) *cachedThunk { return &cachedThunk{content: content} } -func (t *cachedThunk) getValue(i *interpreter) (value, error) { +func (t *cachedThunk) getValue(i *interpreter, trace traceElement) (value, error) { if t.content != nil { return t.content, nil } if t.err != nil { return nil, t.err } - v, err := i.EvalInCleanEnv(t.env, t.body, false) + v, err := i.EvalInCleanEnv(trace, t.env, t.body, false) if err != nil { // TODO(sbarzowski) perhaps cache errors as well // may be necessary if we allow handling them in any way @@ -87,9 +95,9 @@ type codeUnboundField struct { body ast.Node } -func (f *codeUnboundField) evaluate(i *interpreter, sb selfBinding, origBindings bindingFrame, fieldName string) (value, error) { +func (f *codeUnboundField) evaluate(i *interpreter, trace traceElement, sb selfBinding, origBindings bindingFrame, fieldName string) (value, error) { env := makeEnvironment(origBindings, sb) - return i.EvalInCleanEnv(&env, f.body, false) + return i.EvalInCleanEnv(trace, &env, f.body, false) } // Provide additional bindings for a field. It shadows bindings from the object. @@ -99,15 +107,16 @@ type bindingsUnboundField struct { bindings bindingFrame } -func (f *bindingsUnboundField) evaluate(i *interpreter, sb selfBinding, origBindings bindingFrame, fieldName string) (value, error) { - upValues := make(bindingFrame) +func (f *bindingsUnboundField) evaluate(i *interpreter, trace traceElement, sb selfBinding, origBindings bindingFrame, fieldName string) (value, error) { + var upValues bindingFrame + upValues = make(bindingFrame) for variable, pvalue := range origBindings { upValues[variable] = pvalue } for variable, pvalue := range f.bindings { upValues[variable] = pvalue } - return f.inner.evaluate(i, sb, upValues, fieldName) + return f.inner.evaluate(i, trace, sb, upValues, fieldName) } // plusSuperUnboundField represents a `field+: ...` that hasn't been bound to an object. @@ -115,19 +124,19 @@ type plusSuperUnboundField struct { inner unboundField } -func (f *plusSuperUnboundField) evaluate(i *interpreter, sb selfBinding, origBinding bindingFrame, fieldName string) (value, error) { - right, err := f.inner.evaluate(i, sb, origBinding, fieldName) +func (f *plusSuperUnboundField) evaluate(i *interpreter, trace traceElement, sb selfBinding, origBinding bindingFrame, fieldName string) (value, error) { + right, err := f.inner.evaluate(i, trace, sb, origBinding, fieldName) if err != nil { return nil, err } if !objectHasField(sb.super(), fieldName, withHidden) { return right, nil } - left, err := objectIndex(i, sb.super(), fieldName) + left, err := objectIndex(i, trace, sb.super(), fieldName) if err != nil { return nil, err } - return builtinPlus(i, left, right) + return builtinPlus(i, trace, left, right) } // evalCallables @@ -138,12 +147,12 @@ type closure struct { // arguments should be added to it, before executing it env environment function *ast.Function - params []namedParameter + params parameters } -func forceThunks(i *interpreter, args *bindingFrame) error { +func forceThunks(i *interpreter, trace traceElement, args *bindingFrame) error { for _, arg := range *args { - _, err := arg.getValue(i) + _, err := arg.getValue(i, trace) if err != nil { return err } @@ -151,11 +160,17 @@ func forceThunks(i *interpreter, args *bindingFrame) error { return nil } -func (closure *closure) evalCall(arguments callArguments, i *interpreter) (value, error) { +func (closure *closure) evalCall(arguments callArguments, i *interpreter, trace traceElement) (value, error) { argThunks := make(bindingFrame) - parameters := closure.parameters() + parameters := closure.Parameters() for i, arg := range arguments.positional { - argThunks[parameters[i].name] = arg + var name ast.Identifier + if i < len(parameters.required) { + name = parameters.required[i] + } else { + name = parameters.optional[i-len(parameters.required)].name + } + argThunks[name] = arg } for _, arg := range arguments.named { @@ -164,7 +179,8 @@ func (closure *closure) evalCall(arguments callArguments, i *interpreter) (value var calledEnvironment environment - for _, param := range parameters { + for i := range parameters.optional { + param := ¶meters.optional[i] if _, exists := argThunks[param.name]; !exists { argThunks[param.name] = &cachedThunk{ // Default arguments are evaluated in the same environment as function body @@ -175,7 +191,7 @@ func (closure *closure) evalCall(arguments callArguments, i *interpreter) (value } if arguments.tailstrict { - err := forceThunks(i, &argThunks) + err := forceThunks(i, trace, &argThunks) if err != nil { return nil, err } @@ -185,23 +201,30 @@ func (closure *closure) evalCall(arguments callArguments, i *interpreter) (value addBindings(closure.env.upValues, argThunks), closure.env.selfBinding, ) - return i.EvalInCleanEnv(&calledEnvironment, closure.function.Body, arguments.tailstrict) + return i.EvalInCleanEnv(trace, &calledEnvironment, closure.function.Body, arguments.tailstrict) } -func (closure *closure) parameters() []namedParameter { +func (closure *closure) Parameters() parameters { return closure.params } -func prepareClosureParameters(params []ast.Parameter, env environment) []namedParameter { - preparedParams := make([]namedParameter, 0, len(params)) - for _, named := range params { - preparedParams = append(preparedParams, namedParameter{ +func prepareClosureParameters(params ast.Parameters, env environment) parameters { + optionalParameters := make([]namedParameter, 0, len(params.Optional)) + for _, named := range params.Optional { + optionalParameters = append(optionalParameters, namedParameter{ name: named.Name, defaultArg: named.DefaultArg, }) } - return preparedParams + requiredParameters := make([]ast.Identifier, 0, len(params.Required)) + for _, required := range params.Required { + requiredParameters = append(requiredParameters, required.Name) + } + return parameters{ + required: requiredParameters, + optional: optionalParameters, + } } func makeClosure(env environment, function *ast.Function) *closure { @@ -220,15 +243,15 @@ type NativeFunction struct { } // evalCall evaluates a call to a NativeFunction and returns the result. -func (native *NativeFunction) evalCall(arguments callArguments, i *interpreter) (value, error) { - flatArgs := flattenArgs(arguments, native.parameters(), []value{}) +func (native *NativeFunction) evalCall(arguments callArguments, i *interpreter, trace traceElement) (value, error) { + flatArgs := flattenArgs(arguments, native.Parameters(), []value{}) nativeArgs := make([]interface{}, 0, len(flatArgs)) for _, arg := range flatArgs { - v, err := i.evaluatePV(arg) + v, err := i.evaluatePV(arg, trace) if err != nil { return nil, err } - json, err := i.manifestJSON(v) + json, err := i.manifestJSON(trace, v) if err != nil { return nil, err } @@ -236,16 +259,22 @@ func (native *NativeFunction) evalCall(arguments callArguments, i *interpreter) } resultJSON, err := native.Func(nativeArgs) if err != nil { - return nil, i.Error(err.Error()) + return nil, i.Error(err.Error(), trace) } - return jsonToValue(i, resultJSON) + return jsonToValue(i, trace, resultJSON) } // Parameters returns a NativeFunction's parameters. -func (native *NativeFunction) parameters() []namedParameter { - ret := make([]namedParameter, len(native.Params)) - for i := range ret { - ret[i].name = native.Params[i] - } - return ret +func (native *NativeFunction) Parameters() parameters { + return parameters{required: native.Params} +} + +// ------------------------------------- + +type defaultArgument struct { + body ast.Node +} + +func (da *defaultArgument) inEnv(env *environment) potentialValue { + return &cachedThunk{env: env, body: da.body} } diff --git a/vendor/github.com/google/go-jsonnet/travisBazel.sh b/vendor/github.com/google/go-jsonnet/travisBazel.sh index bd3b42aa..2bfc793c 100644 --- a/vendor/github.com/google/go-jsonnet/travisBazel.sh +++ b/vendor/github.com/google/go-jsonnet/travisBazel.sh @@ -8,7 +8,7 @@ bazel --host_jvm_args=-Xmx500m \ --spawn_strategy=standalone \ --genrule_strategy=standalone \ --test_strategy=standalone \ - --local_ram_resources=1536 \ + --local_resources=1536,1.5,0.5 \ --noshow_progress \ --verbose_failures \ --test_output=errors \ diff --git a/vendor/github.com/google/go-jsonnet/travisBuild.sh b/vendor/github.com/google/go-jsonnet/travisBuild.sh index b263c2a4..df355d76 100644 --- a/vendor/github.com/google/go-jsonnet/travisBuild.sh +++ b/vendor/github.com/google/go-jsonnet/travisBuild.sh @@ -1,11 +1,25 @@ #!/usr/bin/env bash -set -e - run_tests() { - golangci-lint run ./... $GOPATH/bin/goveralls -service=travis-ci - SKIP_GO_TESTS=1 ./tests.sh + ./tests.sh --skip-go-test +} + +release() { + env VERSION=$TRAVIS_TAG ./release.sh } -run_tests +if [ "$TRAVIS_PULL_REQUEST" != "false" ]; then + # Pull Requests. + echo -e "Build Pull Request #$TRAVIS_PULL_REQUEST => Branch [$TRAVIS_BRANCH]" + run_tests +elif [ "$TRAVIS_TAG" == "" ]; then + # Pushed branches. + echo -e "Build Branch $TRAVIS_BRANCH" + run_tests +else + # $TRAVIS_PULL_REQUEST == "false" and $TRAVIS_TAG != "" -> Releases. + echo -e 'Build Branch for Release => Branch ['$TRAVIS_BRANCH'] Tag ['$TRAVIS_TAG']' + release +fi + diff --git a/vendor/github.com/google/go-jsonnet/update_cpp_jsonnet.sh b/vendor/github.com/google/go-jsonnet/update_cpp_jsonnet.sh deleted file mode 100644 index 954d195d..00000000 --- a/vendor/github.com/google/go-jsonnet/update_cpp_jsonnet.sh +++ /dev/null @@ -1,16 +0,0 @@ -#!/bin/bash - -# Updates cpp-jsonnet repo and regenerates dependent files - -set -e -set -x - -cd cpp-jsonnet -git checkout master -git pull -cd .. -go run cmd/dumpstdlibast/dumpstdlibast.go cpp-jsonnet/stdlib/std.jsonnet > astgen/stdast.go - -set +x -echo -echo -e "\033[1mUpdate completed. Please check if any tests are broken and fix any encountered issues.\033[0m" diff --git a/vendor/github.com/google/go-jsonnet/util.go b/vendor/github.com/google/go-jsonnet/util.go deleted file mode 100644 index 91f24c74..00000000 --- a/vendor/github.com/google/go-jsonnet/util.go +++ /dev/null @@ -1,38 +0,0 @@ -package jsonnet - -func minInt(a, b int) int { - if a < b { - return a - } - return b -} - -func runeCmp(a, b rune) int { - if a < b { - return -1 - } else if a > b { - return 1 - } else { - return 0 - } -} - -func intCmp(a, b int) int { - if a < b { - return -1 - } else if a > b { - return 1 - } else { - return 0 - } -} - -func float64Cmp(a, b float64) int { - if a < b { - return -1 - } else if a > b { - return 1 - } else { - return 0 - } -} diff --git a/vendor/github.com/google/go-jsonnet/value.go b/vendor/github.com/google/go-jsonnet/value.go index 7598a7d7..faee5db9 100644 --- a/vendor/github.com/google/go-jsonnet/value.go +++ b/vendor/github.com/google/go-jsonnet/value.go @@ -58,7 +58,7 @@ var arrayType = &valueType{"array"} // TODO(sbarzowski) perhaps call it just "Thunk"? type potentialValue interface { // fromWhere keeps the information from where the evaluation was requested. - getValue(i *interpreter) (value, error) + getValue(i *interpreter, fromWhere traceElement) (value, error) aPotentialValue() } @@ -78,7 +78,7 @@ type valueString interface { length() int getRunes() []rune getGoString() string - index(i *interpreter, index int) (value, error) + index(i *interpreter, trace traceElement, index int) (value, error) } // valueFlatString represents a string value, internally using a []rune for quick @@ -89,11 +89,11 @@ type valueFlatString struct { value []rune } -func (s *valueFlatString) index(i *interpreter, index int) (value, error) { +func (s *valueFlatString) index(i *interpreter, trace traceElement, index int) (value, error) { if 0 <= index && index < s.length() { return makeValueString(string(s.value[index])), nil } - return nil, i.Error(fmt.Sprintf("Index %d out of bounds, not within [0, %v)", index, s.length())) + return nil, i.Error(fmt.Sprintf("Index %d out of bounds, not within [0, %v)", index, s.length()), trace) } func (s *valueFlatString) getRunes() []rune { @@ -142,12 +142,12 @@ func (s *valueStringTree) flattenToLeft() { } } -func (s *valueStringTree) index(i *interpreter, index int) (value, error) { +func (s *valueStringTree) index(i *interpreter, trace traceElement, index int) (value, error) { if 0 <= index && index < s.len { s.flattenToLeft() - return s.left.index(i, index) + return s.left.index(i, trace, index) } - return nil, i.Error(fmt.Sprintf("Index %d out of bounds, not within [0, %v)", index, s.length())) + return nil, i.Error(fmt.Sprintf("Index %d out of bounds, not within [0, %v)", index, s.length()), trace) } func (s *valueStringTree) getRunes() []rune { @@ -199,7 +199,7 @@ func concatStrings(a, b valueString) valueString { } } -func stringCmp(a, b valueString) int { +func stringLessThan(a, b valueString) bool { runesA := a.getRunes() runesB := b.getRunes() var length int @@ -210,10 +210,10 @@ func stringCmp(a, b valueString) int { } for i := 0; i < length; i++ { if runesA[i] != runesB[i] { - return runeCmp(runesA[i], runesB[i]) + return runesA[i] < runesB[i] } } - return intCmp(len(runesA), len(runesB)) + return len(runesA) < len(runesB) } func stringEqual(a, b valueString) bool { @@ -247,6 +247,10 @@ func makeValueBoolean(v bool) *valueBoolean { return &valueBoolean{value: v} } +func (b *valueBoolean) not() *valueBoolean { + return makeValueBoolean(!b.value) +} + type valueNumber struct { valueBase value float64 @@ -290,11 +294,11 @@ type valueArray struct { elements []*cachedThunk } -func (arr *valueArray) index(i *interpreter, index int) (value, error) { +func (arr *valueArray) index(i *interpreter, trace traceElement, index int) (value, error) { if 0 <= index && index < arr.length() { - return i.evaluatePV(arr.elements[index]) + return i.evaluatePV(arr.elements[index], trace) } - return nil, i.Error(fmt.Sprintf("Index %d out of bounds, not within [0, %v)", index, arr.length())) + return nil, i.Error(fmt.Sprintf("Index %d out of bounds, not within [0, %v)", index, arr.length()), trace) } func (arr *valueArray) length() int { @@ -309,7 +313,9 @@ func makeValueArray(elements []*cachedThunk) *valueArray { arrayElems = elements } else { arrayElems = make([]*cachedThunk, len(elements)) - copy(arrayElems, elements) + for i := range elements { + arrayElems[i] = elements[i] + } } return &valueArray{ elements: arrayElems, @@ -318,8 +324,12 @@ func makeValueArray(elements []*cachedThunk) *valueArray { func concatArrays(a, b *valueArray) *valueArray { result := make([]*cachedThunk, 0, len(a.elements)+len(b.elements)) - result = append(result, a.elements...) - result = append(result, b.elements...) + for _, r := range a.elements { + result = append(result, r) + } + for _, r := range b.elements { + result = append(result, r) + } return &valueArray{elements: result} } @@ -337,55 +347,62 @@ type valueFunction struct { // TODO(sbarzowski) better name? type evalCallable interface { - evalCall(args callArguments, i *interpreter) (value, error) - parameters() []namedParameter + evalCall(args callArguments, i *interpreter, trace traceElement) (value, error) + Parameters() parameters } -func (f *valueFunction) call(i *interpreter, args callArguments) (value, error) { - err := checkArguments(i, args, f.parameters()) +func (f *valueFunction) call(i *interpreter, trace traceElement, args callArguments) (value, error) { + err := checkArguments(i, trace, args, f.Parameters()) if err != nil { return nil, err } - return f.ec.evalCall(args, i) + return f.ec.evalCall(args, i, trace) } -func (f *valueFunction) parameters() []namedParameter { - return f.ec.parameters() +func (f *valueFunction) Parameters() parameters { + return f.ec.Parameters() } -func checkArguments(i *interpreter, args callArguments, params []namedParameter) error { +func checkArguments(i *interpreter, trace traceElement, args callArguments, params parameters) error { + received := make(map[ast.Identifier]bool) + accepted := make(map[ast.Identifier]bool) numPassed := len(args.positional) - maxExpected := len(params) + numExpected := len(params.required) + len(params.optional) - if numPassed > maxExpected { - return i.Error(fmt.Sprintf("function expected %v positional argument(s), but got %v", maxExpected, numPassed)) + if numPassed > numExpected { + return i.Error(fmt.Sprintf("function expected %v positional argument(s), but got %v", numExpected, numPassed), trace) } - // Parameter names the function will accept. - accepted := make(map[ast.Identifier]bool) - for _, param := range params { + for _, param := range params.required { + accepted[param] = true + } + + for _, param := range params.optional { accepted[param.name] = true } - // Parameter names the call will bind. - received := make(map[ast.Identifier]bool) for i := range args.positional { - received[params[i].name] = true + if i < len(params.required) { + received[params.required[i]] = true + } else { + received[params.optional[i-len(params.required)].name] = true + } } + for _, arg := range args.named { if _, present := received[arg.name]; present { - return i.Error(fmt.Sprintf("Argument %v already provided", arg.name)) + return i.Error(fmt.Sprintf("Argument %v already provided", arg.name), trace) } if _, present := accepted[arg.name]; !present { - return i.Error(fmt.Sprintf("function has no parameter %v", arg.name)) + return i.Error(fmt.Sprintf("function has no parameter %v", arg.name), trace) } received[arg.name] = true } - for _, param := range params { - if _, present := received[param.name]; !present && param.defaultArg == nil { - return i.Error(fmt.Sprintf("Missing argument: %v", param.name)) + for _, param := range params.required { + if _, present := received[param]; !present { + return i.Error(fmt.Sprintf("Missing argument: %v", param), trace) } } @@ -396,11 +413,22 @@ func (f *valueFunction) getType() *valueType { return functionType } +// parameters represents required position and optional named parameters for a +// function definition. +type parameters struct { + required ast.Identifiers + optional []namedParameter +} + type namedParameter struct { name ast.Identifier defaultArg ast.Node } +type potentialValueInEnv interface { + inEnv(env *environment) *cachedThunk +} + type callArguments struct { positional []*cachedThunk named []namedCallArgument @@ -492,8 +520,8 @@ func (*valueObject) getType() *valueType { return objectType } -func (obj *valueObject) index(i *interpreter, field string) (value, error) { - return objectIndex(i, objectBinding(obj), field) +func (obj *valueObject) index(i *interpreter, trace traceElement, field string) (value, error) { + return objectIndex(i, trace, objectBinding(obj), field) } func (obj *valueObject) assertionsChecked() bool { @@ -545,14 +573,14 @@ type simpleObject struct { locals []objectLocal } -func checkAssertionsHelper(i *interpreter, obj *valueObject, curr uncachedObject, superDepth int) error { +func checkAssertionsHelper(i *interpreter, trace traceElement, obj *valueObject, curr uncachedObject, superDepth int) error { switch curr := curr.(type) { case *extendedObject: - err := checkAssertionsHelper(i, obj, curr.right, superDepth) + err := checkAssertionsHelper(i, trace, obj, curr.right, superDepth) if err != nil { return err } - err = checkAssertionsHelper(i, obj, curr.left, superDepth+curr.right.inheritanceSize()) + err = checkAssertionsHelper(i, trace, obj, curr.left, superDepth+curr.right.inheritanceSize()) if err != nil { return err } @@ -561,7 +589,7 @@ func checkAssertionsHelper(i *interpreter, obj *valueObject, curr uncachedObject for _, assert := range curr.asserts { sb := selfBinding{self: obj, superDepth: superDepth} fieldUpValues := prepareFieldUpvalues(sb, curr.upValues, curr.locals) - _, err := assert.evaluate(i, sb, fieldUpValues, "") + _, err := assert.evaluate(i, trace, sb, fieldUpValues, "") if err != nil { return err } @@ -574,13 +602,13 @@ func checkAssertionsHelper(i *interpreter, obj *valueObject, curr uncachedObject } } -func checkAssertions(i *interpreter, obj *valueObject) error { +func checkAssertions(i *interpreter, trace traceElement, obj *valueObject) error { if !obj.assertionsChecked() { // Assertions may refer to the object that will normally // trigger checking of assertions, resulting in an endless recursion. // To avoid that, while we check them, we treat them as already passed. obj.setAssertionsCheckResult(errNoErrorInObjectInvariants) - obj.setAssertionsCheckResult(checkAssertionsHelper(i, obj, obj.uncached, 0)) + obj.setAssertionsCheckResult(checkAssertionsHelper(i, trace, obj, obj.uncached, 0)) } return obj.getAssertionsCheckResult() } @@ -610,7 +638,7 @@ type simpleObjectField struct { // unboundField is a field that doesn't know yet in which object it is. type unboundField interface { - evaluate(i *interpreter, sb selfBinding, origBinding bindingFrame, fieldName string) (value, error) + evaluate(i *interpreter, trace traceElement, sb selfBinding, origBinding bindingFrame, fieldName string) (value, error) } // extendedObject represents an object created through inheritance (left + right). @@ -699,18 +727,18 @@ func prepareFieldUpvalues(sb selfBinding, upValues bindingFrame, locals []object return newUpValues } -func objectIndex(i *interpreter, sb selfBinding, fieldName string) (value, error) { - err := checkAssertions(i, sb.self) +func objectIndex(i *interpreter, trace traceElement, sb selfBinding, fieldName string) (value, error) { + err := checkAssertions(i, trace, sb.self) if err != nil { return nil, err } if sb.superDepth >= sb.self.uncached.inheritanceSize() { - return nil, i.Error("Attempt to use super when there is no super class.") + return nil, i.Error("Attempt to use super when there is no super class.", trace) } found, field, upValues, locals, foundAt := findField(sb.self.uncached, sb.superDepth, fieldName) if !found { - return nil, i.Error(fmt.Sprintf("Field does not exist: %s", fieldName)) + return nil, i.Error(fmt.Sprintf("Field does not exist: %s", fieldName), trace) } if val, ok := sb.self.cache[objectCacheKey{field: fieldName, depth: foundAt}]; ok { @@ -720,7 +748,7 @@ func objectIndex(i *interpreter, sb selfBinding, fieldName string) (value, error fieldSelfBinding := selfBinding{self: sb.self, superDepth: foundAt} fieldUpValues := prepareFieldUpvalues(fieldSelfBinding, upValues, locals) - val, err := field.field.evaluate(i, fieldSelfBinding, fieldUpValues, fieldName) + val, err := field.field.evaluate(i, trace, fieldSelfBinding, fieldUpValues, fieldName) if err == nil { sb.self.cache[objectCacheKey{field: fieldName, depth: foundAt}] = val diff --git a/vendor/github.com/google/go-jsonnet/vm.go b/vendor/github.com/google/go-jsonnet/vm.go index 4ba94d4f..1b22c5a4 100644 --- a/vendor/github.com/google/go-jsonnet/vm.go +++ b/vendor/github.com/google/go-jsonnet/vm.go @@ -19,14 +19,9 @@ package jsonnet import ( "errors" "fmt" - "os" - "path/filepath" "runtime/debug" - "sort" - "strings" "github.com/google/go-jsonnet/ast" - "github.com/google/go-jsonnet/internal/parser" "github.com/google/go-jsonnet/internal/program" ) @@ -130,7 +125,7 @@ const ( ) // version is the current gojsonnet's version -const version = "v0.17.0" +const version = "v0.15.0" // Evaluate evaluates a Jsonnet program given by an Abstract Syntax Tree // and returns serialized JSON as string. @@ -146,7 +141,7 @@ func (vm *VM) Evaluate(node ast.Node) (val string, err error) { // EvaluateStream evaluates a Jsonnet program given by an Abstract Syntax Tree // and returns an array of JSON strings. -func (vm *VM) EvaluateStream(node ast.Node) (output []string, err error) { +func (vm *VM) EvaluateStream(node ast.Node) (output interface{}, err error) { defer func() { if r := recover(); r != nil { err = fmt.Errorf("(CRASH) %v\n%s", r, debug.Stack()) @@ -158,7 +153,7 @@ func (vm *VM) EvaluateStream(node ast.Node) (output []string, err error) { // EvaluateMulti evaluates a Jsonnet program given by an Abstract Syntax Tree // and returns key-value pairs. // The keys are strings and the values are JSON strigns (serialized JSON). -func (vm *VM) EvaluateMulti(node ast.Node) (output map[string]string, err error) { +func (vm *VM) EvaluateMulti(node ast.Node) (output interface{}, err error) { defer func() { if r := recover(); r != nil { err = fmt.Errorf("(CRASH) %v\n%s", r, debug.Stack()) @@ -167,13 +162,13 @@ func (vm *VM) EvaluateMulti(node ast.Node) (output map[string]string, err error) return evaluateMulti(node, vm.ext, vm.tla, vm.nativeFuncs, vm.MaxStack, vm.importCache, vm.StringOutput) } -func (vm *VM) evaluateSnippet(diagnosticFileName ast.DiagnosticFileName, filename string, snippet string, kind evalKind) (output interface{}, err error) { +func (vm *VM) evaluateSnippet(filename string, snippet string, kind evalKind) (output interface{}, err error) { defer func() { if r := recover(); r != nil { err = fmt.Errorf("(CRASH) %v\n%s", r, debug.Stack()) } }() - node, err := program.SnippetToAST(diagnosticFileName, filename, snippet) + node, err := SnippetToAST(filename, snippet) if err != nil { return "", err } @@ -191,85 +186,12 @@ func (vm *VM) evaluateSnippet(diagnosticFileName ast.DiagnosticFileName, filenam return output, nil } -func getAbsPath(path string) (string, error) { - var absPath string - if filepath.IsAbs(path) { - absPath = path - } else { - wd, err := os.Getwd() - if err != nil { - return "", nil - } - absPath = strings.Join([]string{wd, path}, string(filepath.Separator)) - } - cleanedAbsPath, err := filepath.EvalSymlinks(absPath) - if err != nil { - return "", err - } - return cleanedAbsPath, nil -} - -func (vm *VM) findDependencies(filePath string, node *ast.Node, dependencies map[string]struct{}, stackTrace *[]traceFrame) (err error) { - var cleanedAbsPath string - switch i := (*node).(type) { - case *ast.Import: - node, foundAt, err := vm.ImportAST(filePath, i.File.Value) - if err != nil { - *stackTrace = append([]traceFrame{{Loc: *i.Loc()}}, *stackTrace...) - return err - } - cleanedAbsPath = foundAt - if _, isFileImporter := vm.importer.(*FileImporter); isFileImporter { - cleanedAbsPath, err = getAbsPath(foundAt) - if err != nil { - *stackTrace = append([]traceFrame{{Loc: *i.Loc()}}, *stackTrace...) - return err - } - } - // Check that we haven't already parsed the imported file. - if _, alreadyParsed := dependencies[cleanedAbsPath]; alreadyParsed { - return nil - } - dependencies[cleanedAbsPath] = struct{}{} - err = vm.findDependencies(foundAt, &node, dependencies, stackTrace) - if err != nil { - *stackTrace = append([]traceFrame{{Loc: *i.Loc()}}, *stackTrace...) - return err - } - case *ast.ImportStr: - foundAt, err := vm.ResolveImport(filePath, i.File.Value) - if err != nil { - *stackTrace = append([]traceFrame{{Loc: *i.Loc()}}, *stackTrace...) - return err - } - cleanedAbsPath = foundAt - if _, isFileImporter := vm.importer.(*FileImporter); isFileImporter { - cleanedAbsPath, err = getAbsPath(foundAt) - if err != nil { - *stackTrace = append([]traceFrame{{Loc: *i.Loc()}}, *stackTrace...) - return err - } - } - dependencies[cleanedAbsPath] = struct{}{} - default: - for _, node := range parser.Children(i) { - err = vm.findDependencies(filePath, &node, dependencies, stackTrace) - if err != nil { - return err - } - } - } - return -} - // EvaluateSnippet evaluates a string containing Jsonnet code, return a JSON // string. // -// The filename parameter is used for resolving relative imports and for errors messages. -// -// Deprecated: Use EvaluateFile or EvaluateAnonymousSnippet instead. +// The filename parameter is only used for error messages. func (vm *VM) EvaluateSnippet(filename string, snippet string) (json string, formattedErr error) { - output, err := vm.evaluateSnippet(ast.DiagnosticFileName(filename), filename, snippet, evalKindRegular) + output, err := vm.evaluateSnippet(filename, snippet, evalKindRegular) if err != nil { return "", errors.New(vm.ErrorFormatter.Format(err)) } @@ -280,11 +202,9 @@ func (vm *VM) EvaluateSnippet(filename string, snippet string) (json string, for // EvaluateSnippetStream evaluates a string containing Jsonnet code to an array. // The array is returned as an array of JSON strings. // -// The filename parameter is used for resolving relative imports and for errors messages. -// -// Deprecated: Use EvaluateFileStream or EvaluateAnonymousSnippetStream instead. +// The filename parameter is only used for error messages. func (vm *VM) EvaluateSnippetStream(filename string, snippet string) (docs []string, formattedErr error) { - output, err := vm.evaluateSnippet(ast.DiagnosticFileName(filename), filename, snippet, evalKindStream) + output, err := vm.evaluateSnippet(filename, snippet, evalKindStream) if err != nil { return nil, errors.New(vm.ErrorFormatter.Format(err)) } @@ -295,50 +215,9 @@ func (vm *VM) EvaluateSnippetStream(filename string, snippet string) (docs []str // EvaluateSnippetMulti evaluates a string containing Jsonnet code to key-value // pairs. The keys are field name strings and the values are JSON strings. // -// The filename parameter is used for resolving relative imports and for errors messages. -// -// Deprecated: Use EvaluateFileMulti or EvaluateAnonymousSnippetMulti instead. -func (vm *VM) EvaluateSnippetMulti(filename string, snippet string) (files map[string]string, formattedErr error) { - output, err := vm.evaluateSnippet(ast.DiagnosticFileName(filename), filename, snippet, evalKindMulti) - if err != nil { - return nil, errors.New(vm.ErrorFormatter.Format(err)) - } - files = output.(map[string]string) - return -} - -// EvaluateAnonymousSnippet evaluates a string containing Jsonnet code, return a JSON -// string. -// // The filename parameter is only used for error messages. -func (vm *VM) EvaluateAnonymousSnippet(filename string, snippet string) (json string, formattedErr error) { - output, err := vm.evaluateSnippet(ast.DiagnosticFileName(filename), "", snippet, evalKindRegular) - if err != nil { - return "", errors.New(vm.ErrorFormatter.Format(err)) - } - json = output.(string) - return -} - -// EvaluateAnonymousSnippetStream evaluates a string containing Jsonnet code to an array. -// The array is returned as an array of JSON strings. -// -// The filename parameter is only used for error messages. -func (vm *VM) EvaluateAnonymousSnippetStream(filename string, snippet string) (docs []string, formattedErr error) { - output, err := vm.evaluateSnippet(ast.DiagnosticFileName(filename), "", snippet, evalKindStream) - if err != nil { - return nil, errors.New(vm.ErrorFormatter.Format(err)) - } - docs = output.([]string) - return -} - -// EvaluateAnonymousSnippetMulti evaluates a string containing Jsonnet code to key-value -// pairs. The keys are field name strings and the values are JSON strings. -// -// The filename parameter is only used for error messages. -func (vm *VM) EvaluateAnonymousSnippetMulti(filename string, snippet string) (files map[string]string, formattedErr error) { - output, err := vm.evaluateSnippet(ast.DiagnosticFileName(filename), "", snippet, evalKindMulti) +func (vm *VM) EvaluateSnippetMulti(filename string, snippet string) (files map[string]string, formattedErr error) { + output, err := vm.evaluateSnippet(filename, snippet, evalKindMulti) if err != nil { return nil, errors.New(vm.ErrorFormatter.Format(err)) } @@ -346,108 +225,6 @@ func (vm *VM) EvaluateAnonymousSnippetMulti(filename string, snippet string) (fi return } -// EvaluateFile evaluates Jsonnet code in a file and returns a JSON -// string. -// -// The importer is used to fetch the contents of the file. -func (vm *VM) EvaluateFile(filename string) (json string, formattedErr error) { - node, _, err := vm.ImportAST("", filename) - if err != nil { - return "", errors.New(vm.ErrorFormatter.Format(err)) - } - output, err := vm.Evaluate(node) - if err != nil { - return "", errors.New(vm.ErrorFormatter.Format(err)) - } - return output, nil -} - -// EvaluateFileStream evaluates Jsonnet code in a file to an array. -// The array is returned as an array of JSON strings. -// -// The importer is used to fetch the contents of the file. -func (vm *VM) EvaluateFileStream(filename string) (docs []string, formattedErr error) { - node, _, err := vm.ImportAST("", filename) - if err != nil { - return nil, errors.New(vm.ErrorFormatter.Format(err)) - } - output, err := vm.EvaluateStream(node) - if err != nil { - return nil, errors.New(vm.ErrorFormatter.Format(err)) - } - return output, nil -} - -// EvaluateFileMulti evaluates Jsonnet code in a file to key-value -// pairs. The keys are field name strings and the values are JSON strings. -// -// The importer is used to fetch the contents of the file. -func (vm *VM) EvaluateFileMulti(filename string) (files map[string]string, formattedErr error) { - node, _, err := vm.ImportAST("", filename) - if err != nil { - return nil, errors.New(vm.ErrorFormatter.Format(err)) - } - output, err := vm.EvaluateMulti(node) - if err != nil { - return nil, errors.New(vm.ErrorFormatter.Format(err)) - } - return output, nil -} - -// FindDependencies returns a sorted array of unique transitive dependencies (via import or importstr) -// from all the given `importedPaths` which are themselves excluded from the returned array. -// The `importedPaths` are parsed as if they were imported from a Jsonnet file located at `importedFrom`. -func (vm *VM) FindDependencies(importedFrom string, importedPaths []string) ([]string, error) { - var nodes []*ast.Node - var stackTrace []traceFrame - filePaths := make([]string, len(importedPaths)) - depsToExclude := make([]string, len(importedPaths)) - deps := make(map[string]struct{}) - - for i, filePath := range importedPaths { - node, foundAt, err := vm.ImportAST(importedFrom, filePath) - if err != nil { - return nil, err - } - cleanedAbsPath := foundAt - if _, isFileImporter := vm.importer.(*FileImporter); isFileImporter { - cleanedAbsPath, err = getAbsPath(foundAt) - if err != nil { - return nil, err - } - } - nodes = append(nodes, &node) - filePaths[i] = foundAt - - // Add `importedPaths` to the dependencies so that they are not parsed again. - // Will be removed before returning. - deps[cleanedAbsPath] = struct{}{} - depsToExclude[i] = cleanedAbsPath - } - - for i, filePath := range filePaths { - err := vm.findDependencies(filePath, nodes[i], deps, &stackTrace) - if err != nil { - err = makeRuntimeError(err.Error(), stackTrace) - return nil, errors.New(vm.ErrorFormatter.Format(err)) - } - } - - // Exclude `importedPaths` from the dependencies. - for _, dep := range depsToExclude { - delete(deps, dep) - } - - dependencies, i := make([]string, len(deps)), 0 - for key := range deps { - dependencies[i] = key - i++ - } - sort.Strings(dependencies) - - return dependencies, nil -} - // ResolveImport finds the actual path where the imported file can be found. // It will cache the contents of the file immediately as well, to avoid the possibility of the file // disappearing after being checked. @@ -474,7 +251,7 @@ func (vm *VM) ImportAST(importedFrom, importedPath string) (contents ast.Node, f // SnippetToAST parses a snippet and returns the resulting AST. func SnippetToAST(filename string, snippet string) (ast.Node, error) { - return program.SnippetToAST(ast.DiagnosticFileName(filename), filename, snippet) + return program.SnippetToAST(filename, snippet) } // Version returns the Jsonnet version number. diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go b/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go deleted file mode 100644 index 5b87dcd0..00000000 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go +++ /dev/null @@ -1,5225 +0,0 @@ -// Code generated by protoc-gen-go. DO NOT EDIT. -// source: OpenAPIv2/OpenAPIv2.proto - -package openapi_v2 - -import ( - fmt "fmt" - proto "github.com/golang/protobuf/proto" - any "github.com/golang/protobuf/ptypes/any" - math "math" -) - -// Reference imports to suppress errors if they are not otherwise used. -var _ = proto.Marshal -var _ = fmt.Errorf -var _ = math.Inf - -// This is a compile-time assertion to ensure that this generated file -// is compatible with the proto package it is being compiled against. -// A compilation error at this line likely means your copy of the -// proto package needs to be updated. -const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package - -type AdditionalPropertiesItem struct { - // Types that are valid to be assigned to Oneof: - // *AdditionalPropertiesItem_Schema - // *AdditionalPropertiesItem_Boolean - Oneof isAdditionalPropertiesItem_Oneof `protobuf_oneof:"oneof"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *AdditionalPropertiesItem) Reset() { *m = AdditionalPropertiesItem{} } -func (m *AdditionalPropertiesItem) String() string { return proto.CompactTextString(m) } -func (*AdditionalPropertiesItem) ProtoMessage() {} -func (*AdditionalPropertiesItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{0} -} - -func (m *AdditionalPropertiesItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_AdditionalPropertiesItem.Unmarshal(m, b) -} -func (m *AdditionalPropertiesItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_AdditionalPropertiesItem.Marshal(b, m, deterministic) -} -func (m *AdditionalPropertiesItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_AdditionalPropertiesItem.Merge(m, src) -} -func (m *AdditionalPropertiesItem) XXX_Size() int { - return xxx_messageInfo_AdditionalPropertiesItem.Size(m) -} -func (m *AdditionalPropertiesItem) XXX_DiscardUnknown() { - xxx_messageInfo_AdditionalPropertiesItem.DiscardUnknown(m) -} - -var xxx_messageInfo_AdditionalPropertiesItem proto.InternalMessageInfo - -type isAdditionalPropertiesItem_Oneof interface { - isAdditionalPropertiesItem_Oneof() -} - -type AdditionalPropertiesItem_Schema struct { - Schema *Schema `protobuf:"bytes,1,opt,name=schema,proto3,oneof"` -} - -type AdditionalPropertiesItem_Boolean struct { - Boolean bool `protobuf:"varint,2,opt,name=boolean,proto3,oneof"` -} - -func (*AdditionalPropertiesItem_Schema) isAdditionalPropertiesItem_Oneof() {} - -func (*AdditionalPropertiesItem_Boolean) isAdditionalPropertiesItem_Oneof() {} - -func (m *AdditionalPropertiesItem) GetOneof() isAdditionalPropertiesItem_Oneof { - if m != nil { - return m.Oneof - } - return nil -} - -func (m *AdditionalPropertiesItem) GetSchema() *Schema { - if x, ok := m.GetOneof().(*AdditionalPropertiesItem_Schema); ok { - return x.Schema - } - return nil -} - -func (m *AdditionalPropertiesItem) GetBoolean() bool { - if x, ok := m.GetOneof().(*AdditionalPropertiesItem_Boolean); ok { - return x.Boolean - } - return false -} - -// XXX_OneofWrappers is for the internal use of the proto package. -func (*AdditionalPropertiesItem) XXX_OneofWrappers() []interface{} { - return []interface{}{ - (*AdditionalPropertiesItem_Schema)(nil), - (*AdditionalPropertiesItem_Boolean)(nil), - } -} - -type Any struct { - Value *any.Any `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` - Yaml string `protobuf:"bytes,2,opt,name=yaml,proto3" json:"yaml,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Any) Reset() { *m = Any{} } -func (m *Any) String() string { return proto.CompactTextString(m) } -func (*Any) ProtoMessage() {} -func (*Any) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{1} -} - -func (m *Any) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Any.Unmarshal(m, b) -} -func (m *Any) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Any.Marshal(b, m, deterministic) -} -func (m *Any) XXX_Merge(src proto.Message) { - xxx_messageInfo_Any.Merge(m, src) -} -func (m *Any) XXX_Size() int { - return xxx_messageInfo_Any.Size(m) -} -func (m *Any) XXX_DiscardUnknown() { - xxx_messageInfo_Any.DiscardUnknown(m) -} - -var xxx_messageInfo_Any proto.InternalMessageInfo - -func (m *Any) GetValue() *any.Any { - if m != nil { - return m.Value - } - return nil -} - -func (m *Any) GetYaml() string { - if m != nil { - return m.Yaml - } - return "" -} - -type ApiKeySecurity struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` - In string `protobuf:"bytes,3,opt,name=in,proto3" json:"in,omitempty"` - Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,5,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ApiKeySecurity) Reset() { *m = ApiKeySecurity{} } -func (m *ApiKeySecurity) String() string { return proto.CompactTextString(m) } -func (*ApiKeySecurity) ProtoMessage() {} -func (*ApiKeySecurity) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{2} -} - -func (m *ApiKeySecurity) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ApiKeySecurity.Unmarshal(m, b) -} -func (m *ApiKeySecurity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ApiKeySecurity.Marshal(b, m, deterministic) -} -func (m *ApiKeySecurity) XXX_Merge(src proto.Message) { - xxx_messageInfo_ApiKeySecurity.Merge(m, src) -} -func (m *ApiKeySecurity) XXX_Size() int { - return xxx_messageInfo_ApiKeySecurity.Size(m) -} -func (m *ApiKeySecurity) XXX_DiscardUnknown() { - xxx_messageInfo_ApiKeySecurity.DiscardUnknown(m) -} - -var xxx_messageInfo_ApiKeySecurity proto.InternalMessageInfo - -func (m *ApiKeySecurity) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *ApiKeySecurity) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *ApiKeySecurity) GetIn() string { - if m != nil { - return m.In - } - return "" -} - -func (m *ApiKeySecurity) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *ApiKeySecurity) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type BasicAuthenticationSecurity struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,3,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *BasicAuthenticationSecurity) Reset() { *m = BasicAuthenticationSecurity{} } -func (m *BasicAuthenticationSecurity) String() string { return proto.CompactTextString(m) } -func (*BasicAuthenticationSecurity) ProtoMessage() {} -func (*BasicAuthenticationSecurity) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{3} -} - -func (m *BasicAuthenticationSecurity) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_BasicAuthenticationSecurity.Unmarshal(m, b) -} -func (m *BasicAuthenticationSecurity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_BasicAuthenticationSecurity.Marshal(b, m, deterministic) -} -func (m *BasicAuthenticationSecurity) XXX_Merge(src proto.Message) { - xxx_messageInfo_BasicAuthenticationSecurity.Merge(m, src) -} -func (m *BasicAuthenticationSecurity) XXX_Size() int { - return xxx_messageInfo_BasicAuthenticationSecurity.Size(m) -} -func (m *BasicAuthenticationSecurity) XXX_DiscardUnknown() { - xxx_messageInfo_BasicAuthenticationSecurity.DiscardUnknown(m) -} - -var xxx_messageInfo_BasicAuthenticationSecurity proto.InternalMessageInfo - -func (m *BasicAuthenticationSecurity) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *BasicAuthenticationSecurity) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *BasicAuthenticationSecurity) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type BodyParameter struct { - // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. - Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` - // The name of the parameter. - Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` - // Determines the location of the parameter. - In string `protobuf:"bytes,3,opt,name=in,proto3" json:"in,omitempty"` - // Determines whether or not this parameter is required or optional. - Required bool `protobuf:"varint,4,opt,name=required,proto3" json:"required,omitempty"` - Schema *Schema `protobuf:"bytes,5,opt,name=schema,proto3" json:"schema,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *BodyParameter) Reset() { *m = BodyParameter{} } -func (m *BodyParameter) String() string { return proto.CompactTextString(m) } -func (*BodyParameter) ProtoMessage() {} -func (*BodyParameter) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{4} -} - -func (m *BodyParameter) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_BodyParameter.Unmarshal(m, b) -} -func (m *BodyParameter) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_BodyParameter.Marshal(b, m, deterministic) -} -func (m *BodyParameter) XXX_Merge(src proto.Message) { - xxx_messageInfo_BodyParameter.Merge(m, src) -} -func (m *BodyParameter) XXX_Size() int { - return xxx_messageInfo_BodyParameter.Size(m) -} -func (m *BodyParameter) XXX_DiscardUnknown() { - xxx_messageInfo_BodyParameter.DiscardUnknown(m) -} - -var xxx_messageInfo_BodyParameter proto.InternalMessageInfo - -func (m *BodyParameter) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *BodyParameter) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *BodyParameter) GetIn() string { - if m != nil { - return m.In - } - return "" -} - -func (m *BodyParameter) GetRequired() bool { - if m != nil { - return m.Required - } - return false -} - -func (m *BodyParameter) GetSchema() *Schema { - if m != nil { - return m.Schema - } - return nil -} - -func (m *BodyParameter) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -// Contact information for the owners of the API. -type Contact struct { - // The identifying name of the contact person/organization. - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // The URL pointing to the contact information. - Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` - // The email address of the contact person/organization. - Email string `protobuf:"bytes,3,opt,name=email,proto3" json:"email,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,4,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Contact) Reset() { *m = Contact{} } -func (m *Contact) String() string { return proto.CompactTextString(m) } -func (*Contact) ProtoMessage() {} -func (*Contact) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{5} -} - -func (m *Contact) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Contact.Unmarshal(m, b) -} -func (m *Contact) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Contact.Marshal(b, m, deterministic) -} -func (m *Contact) XXX_Merge(src proto.Message) { - xxx_messageInfo_Contact.Merge(m, src) -} -func (m *Contact) XXX_Size() int { - return xxx_messageInfo_Contact.Size(m) -} -func (m *Contact) XXX_DiscardUnknown() { - xxx_messageInfo_Contact.DiscardUnknown(m) -} - -var xxx_messageInfo_Contact proto.InternalMessageInfo - -func (m *Contact) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *Contact) GetUrl() string { - if m != nil { - return m.Url - } - return "" -} - -func (m *Contact) GetEmail() string { - if m != nil { - return m.Email - } - return "" -} - -func (m *Contact) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Default struct { - AdditionalProperties []*NamedAny `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Default) Reset() { *m = Default{} } -func (m *Default) String() string { return proto.CompactTextString(m) } -func (*Default) ProtoMessage() {} -func (*Default) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{6} -} - -func (m *Default) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Default.Unmarshal(m, b) -} -func (m *Default) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Default.Marshal(b, m, deterministic) -} -func (m *Default) XXX_Merge(src proto.Message) { - xxx_messageInfo_Default.Merge(m, src) -} -func (m *Default) XXX_Size() int { - return xxx_messageInfo_Default.Size(m) -} -func (m *Default) XXX_DiscardUnknown() { - xxx_messageInfo_Default.DiscardUnknown(m) -} - -var xxx_messageInfo_Default proto.InternalMessageInfo - -func (m *Default) GetAdditionalProperties() []*NamedAny { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -// One or more JSON objects describing the schemas being consumed and produced by the API. -type Definitions struct { - AdditionalProperties []*NamedSchema `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Definitions) Reset() { *m = Definitions{} } -func (m *Definitions) String() string { return proto.CompactTextString(m) } -func (*Definitions) ProtoMessage() {} -func (*Definitions) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{7} -} - -func (m *Definitions) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Definitions.Unmarshal(m, b) -} -func (m *Definitions) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Definitions.Marshal(b, m, deterministic) -} -func (m *Definitions) XXX_Merge(src proto.Message) { - xxx_messageInfo_Definitions.Merge(m, src) -} -func (m *Definitions) XXX_Size() int { - return xxx_messageInfo_Definitions.Size(m) -} -func (m *Definitions) XXX_DiscardUnknown() { - xxx_messageInfo_Definitions.DiscardUnknown(m) -} - -var xxx_messageInfo_Definitions proto.InternalMessageInfo - -func (m *Definitions) GetAdditionalProperties() []*NamedSchema { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type Document struct { - // The Swagger version of this document. - Swagger string `protobuf:"bytes,1,opt,name=swagger,proto3" json:"swagger,omitempty"` - Info *Info `protobuf:"bytes,2,opt,name=info,proto3" json:"info,omitempty"` - // The host (name or ip) of the API. Example: 'swagger.io' - Host string `protobuf:"bytes,3,opt,name=host,proto3" json:"host,omitempty"` - // The base path to the API. Example: '/api'. - BasePath string `protobuf:"bytes,4,opt,name=base_path,json=basePath,proto3" json:"base_path,omitempty"` - // The transfer protocol of the API. - Schemes []string `protobuf:"bytes,5,rep,name=schemes,proto3" json:"schemes,omitempty"` - // A list of MIME types accepted by the API. - Consumes []string `protobuf:"bytes,6,rep,name=consumes,proto3" json:"consumes,omitempty"` - // A list of MIME types the API can produce. - Produces []string `protobuf:"bytes,7,rep,name=produces,proto3" json:"produces,omitempty"` - Paths *Paths `protobuf:"bytes,8,opt,name=paths,proto3" json:"paths,omitempty"` - Definitions *Definitions `protobuf:"bytes,9,opt,name=definitions,proto3" json:"definitions,omitempty"` - Parameters *ParameterDefinitions `protobuf:"bytes,10,opt,name=parameters,proto3" json:"parameters,omitempty"` - Responses *ResponseDefinitions `protobuf:"bytes,11,opt,name=responses,proto3" json:"responses,omitempty"` - Security []*SecurityRequirement `protobuf:"bytes,12,rep,name=security,proto3" json:"security,omitempty"` - SecurityDefinitions *SecurityDefinitions `protobuf:"bytes,13,opt,name=security_definitions,json=securityDefinitions,proto3" json:"security_definitions,omitempty"` - Tags []*Tag `protobuf:"bytes,14,rep,name=tags,proto3" json:"tags,omitempty"` - ExternalDocs *ExternalDocs `protobuf:"bytes,15,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,16,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Document) Reset() { *m = Document{} } -func (m *Document) String() string { return proto.CompactTextString(m) } -func (*Document) ProtoMessage() {} -func (*Document) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{8} -} - -func (m *Document) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Document.Unmarshal(m, b) -} -func (m *Document) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Document.Marshal(b, m, deterministic) -} -func (m *Document) XXX_Merge(src proto.Message) { - xxx_messageInfo_Document.Merge(m, src) -} -func (m *Document) XXX_Size() int { - return xxx_messageInfo_Document.Size(m) -} -func (m *Document) XXX_DiscardUnknown() { - xxx_messageInfo_Document.DiscardUnknown(m) -} - -var xxx_messageInfo_Document proto.InternalMessageInfo - -func (m *Document) GetSwagger() string { - if m != nil { - return m.Swagger - } - return "" -} - -func (m *Document) GetInfo() *Info { - if m != nil { - return m.Info - } - return nil -} - -func (m *Document) GetHost() string { - if m != nil { - return m.Host - } - return "" -} - -func (m *Document) GetBasePath() string { - if m != nil { - return m.BasePath - } - return "" -} - -func (m *Document) GetSchemes() []string { - if m != nil { - return m.Schemes - } - return nil -} - -func (m *Document) GetConsumes() []string { - if m != nil { - return m.Consumes - } - return nil -} - -func (m *Document) GetProduces() []string { - if m != nil { - return m.Produces - } - return nil -} - -func (m *Document) GetPaths() *Paths { - if m != nil { - return m.Paths - } - return nil -} - -func (m *Document) GetDefinitions() *Definitions { - if m != nil { - return m.Definitions - } - return nil -} - -func (m *Document) GetParameters() *ParameterDefinitions { - if m != nil { - return m.Parameters - } - return nil -} - -func (m *Document) GetResponses() *ResponseDefinitions { - if m != nil { - return m.Responses - } - return nil -} - -func (m *Document) GetSecurity() []*SecurityRequirement { - if m != nil { - return m.Security - } - return nil -} - -func (m *Document) GetSecurityDefinitions() *SecurityDefinitions { - if m != nil { - return m.SecurityDefinitions - } - return nil -} - -func (m *Document) GetTags() []*Tag { - if m != nil { - return m.Tags - } - return nil -} - -func (m *Document) GetExternalDocs() *ExternalDocs { - if m != nil { - return m.ExternalDocs - } - return nil -} - -func (m *Document) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Examples struct { - AdditionalProperties []*NamedAny `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Examples) Reset() { *m = Examples{} } -func (m *Examples) String() string { return proto.CompactTextString(m) } -func (*Examples) ProtoMessage() {} -func (*Examples) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{9} -} - -func (m *Examples) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Examples.Unmarshal(m, b) -} -func (m *Examples) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Examples.Marshal(b, m, deterministic) -} -func (m *Examples) XXX_Merge(src proto.Message) { - xxx_messageInfo_Examples.Merge(m, src) -} -func (m *Examples) XXX_Size() int { - return xxx_messageInfo_Examples.Size(m) -} -func (m *Examples) XXX_DiscardUnknown() { - xxx_messageInfo_Examples.DiscardUnknown(m) -} - -var xxx_messageInfo_Examples proto.InternalMessageInfo - -func (m *Examples) GetAdditionalProperties() []*NamedAny { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -// information about external documentation -type ExternalDocs struct { - Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` - Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,3,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ExternalDocs) Reset() { *m = ExternalDocs{} } -func (m *ExternalDocs) String() string { return proto.CompactTextString(m) } -func (*ExternalDocs) ProtoMessage() {} -func (*ExternalDocs) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{10} -} - -func (m *ExternalDocs) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ExternalDocs.Unmarshal(m, b) -} -func (m *ExternalDocs) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ExternalDocs.Marshal(b, m, deterministic) -} -func (m *ExternalDocs) XXX_Merge(src proto.Message) { - xxx_messageInfo_ExternalDocs.Merge(m, src) -} -func (m *ExternalDocs) XXX_Size() int { - return xxx_messageInfo_ExternalDocs.Size(m) -} -func (m *ExternalDocs) XXX_DiscardUnknown() { - xxx_messageInfo_ExternalDocs.DiscardUnknown(m) -} - -var xxx_messageInfo_ExternalDocs proto.InternalMessageInfo - -func (m *ExternalDocs) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *ExternalDocs) GetUrl() string { - if m != nil { - return m.Url - } - return "" -} - -func (m *ExternalDocs) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -// A deterministic version of a JSON Schema object. -type FileSchema struct { - Format string `protobuf:"bytes,1,opt,name=format,proto3" json:"format,omitempty"` - Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` - Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` - Default *Any `protobuf:"bytes,4,opt,name=default,proto3" json:"default,omitempty"` - Required []string `protobuf:"bytes,5,rep,name=required,proto3" json:"required,omitempty"` - Type string `protobuf:"bytes,6,opt,name=type,proto3" json:"type,omitempty"` - ReadOnly bool `protobuf:"varint,7,opt,name=read_only,json=readOnly,proto3" json:"read_only,omitempty"` - ExternalDocs *ExternalDocs `protobuf:"bytes,8,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` - Example *Any `protobuf:"bytes,9,opt,name=example,proto3" json:"example,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,10,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *FileSchema) Reset() { *m = FileSchema{} } -func (m *FileSchema) String() string { return proto.CompactTextString(m) } -func (*FileSchema) ProtoMessage() {} -func (*FileSchema) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{11} -} - -func (m *FileSchema) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_FileSchema.Unmarshal(m, b) -} -func (m *FileSchema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_FileSchema.Marshal(b, m, deterministic) -} -func (m *FileSchema) XXX_Merge(src proto.Message) { - xxx_messageInfo_FileSchema.Merge(m, src) -} -func (m *FileSchema) XXX_Size() int { - return xxx_messageInfo_FileSchema.Size(m) -} -func (m *FileSchema) XXX_DiscardUnknown() { - xxx_messageInfo_FileSchema.DiscardUnknown(m) -} - -var xxx_messageInfo_FileSchema proto.InternalMessageInfo - -func (m *FileSchema) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *FileSchema) GetTitle() string { - if m != nil { - return m.Title - } - return "" -} - -func (m *FileSchema) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *FileSchema) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *FileSchema) GetRequired() []string { - if m != nil { - return m.Required - } - return nil -} - -func (m *FileSchema) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *FileSchema) GetReadOnly() bool { - if m != nil { - return m.ReadOnly - } - return false -} - -func (m *FileSchema) GetExternalDocs() *ExternalDocs { - if m != nil { - return m.ExternalDocs - } - return nil -} - -func (m *FileSchema) GetExample() *Any { - if m != nil { - return m.Example - } - return nil -} - -func (m *FileSchema) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type FormDataParameterSubSchema struct { - // Determines whether or not this parameter is required or optional. - Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` - // Determines the location of the parameter. - In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` - // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. - Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` - // The name of the parameter. - Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` - // allows sending a parameter by name only or with an empty value. - AllowEmptyValue bool `protobuf:"varint,5,opt,name=allow_empty_value,json=allowEmptyValue,proto3" json:"allow_empty_value,omitempty"` - Type string `protobuf:"bytes,6,opt,name=type,proto3" json:"type,omitempty"` - Format string `protobuf:"bytes,7,opt,name=format,proto3" json:"format,omitempty"` - Items *PrimitivesItems `protobuf:"bytes,8,opt,name=items,proto3" json:"items,omitempty"` - CollectionFormat string `protobuf:"bytes,9,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` - Default *Any `protobuf:"bytes,10,opt,name=default,proto3" json:"default,omitempty"` - Maximum float64 `protobuf:"fixed64,11,opt,name=maximum,proto3" json:"maximum,omitempty"` - ExclusiveMaximum bool `protobuf:"varint,12,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` - Minimum float64 `protobuf:"fixed64,13,opt,name=minimum,proto3" json:"minimum,omitempty"` - ExclusiveMinimum bool `protobuf:"varint,14,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` - MaxLength int64 `protobuf:"varint,15,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` - MinLength int64 `protobuf:"varint,16,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` - Pattern string `protobuf:"bytes,17,opt,name=pattern,proto3" json:"pattern,omitempty"` - MaxItems int64 `protobuf:"varint,18,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` - MinItems int64 `protobuf:"varint,19,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` - UniqueItems bool `protobuf:"varint,20,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` - Enum []*Any `protobuf:"bytes,21,rep,name=enum,proto3" json:"enum,omitempty"` - MultipleOf float64 `protobuf:"fixed64,22,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,23,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *FormDataParameterSubSchema) Reset() { *m = FormDataParameterSubSchema{} } -func (m *FormDataParameterSubSchema) String() string { return proto.CompactTextString(m) } -func (*FormDataParameterSubSchema) ProtoMessage() {} -func (*FormDataParameterSubSchema) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{12} -} - -func (m *FormDataParameterSubSchema) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_FormDataParameterSubSchema.Unmarshal(m, b) -} -func (m *FormDataParameterSubSchema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_FormDataParameterSubSchema.Marshal(b, m, deterministic) -} -func (m *FormDataParameterSubSchema) XXX_Merge(src proto.Message) { - xxx_messageInfo_FormDataParameterSubSchema.Merge(m, src) -} -func (m *FormDataParameterSubSchema) XXX_Size() int { - return xxx_messageInfo_FormDataParameterSubSchema.Size(m) -} -func (m *FormDataParameterSubSchema) XXX_DiscardUnknown() { - xxx_messageInfo_FormDataParameterSubSchema.DiscardUnknown(m) -} - -var xxx_messageInfo_FormDataParameterSubSchema proto.InternalMessageInfo - -func (m *FormDataParameterSubSchema) GetRequired() bool { - if m != nil { - return m.Required - } - return false -} - -func (m *FormDataParameterSubSchema) GetIn() string { - if m != nil { - return m.In - } - return "" -} - -func (m *FormDataParameterSubSchema) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *FormDataParameterSubSchema) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *FormDataParameterSubSchema) GetAllowEmptyValue() bool { - if m != nil { - return m.AllowEmptyValue - } - return false -} - -func (m *FormDataParameterSubSchema) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *FormDataParameterSubSchema) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *FormDataParameterSubSchema) GetItems() *PrimitivesItems { - if m != nil { - return m.Items - } - return nil -} - -func (m *FormDataParameterSubSchema) GetCollectionFormat() string { - if m != nil { - return m.CollectionFormat - } - return "" -} - -func (m *FormDataParameterSubSchema) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *FormDataParameterSubSchema) GetMaximum() float64 { - if m != nil { - return m.Maximum - } - return 0 -} - -func (m *FormDataParameterSubSchema) GetExclusiveMaximum() bool { - if m != nil { - return m.ExclusiveMaximum - } - return false -} - -func (m *FormDataParameterSubSchema) GetMinimum() float64 { - if m != nil { - return m.Minimum - } - return 0 -} - -func (m *FormDataParameterSubSchema) GetExclusiveMinimum() bool { - if m != nil { - return m.ExclusiveMinimum - } - return false -} - -func (m *FormDataParameterSubSchema) GetMaxLength() int64 { - if m != nil { - return m.MaxLength - } - return 0 -} - -func (m *FormDataParameterSubSchema) GetMinLength() int64 { - if m != nil { - return m.MinLength - } - return 0 -} - -func (m *FormDataParameterSubSchema) GetPattern() string { - if m != nil { - return m.Pattern - } - return "" -} - -func (m *FormDataParameterSubSchema) GetMaxItems() int64 { - if m != nil { - return m.MaxItems - } - return 0 -} - -func (m *FormDataParameterSubSchema) GetMinItems() int64 { - if m != nil { - return m.MinItems - } - return 0 -} - -func (m *FormDataParameterSubSchema) GetUniqueItems() bool { - if m != nil { - return m.UniqueItems - } - return false -} - -func (m *FormDataParameterSubSchema) GetEnum() []*Any { - if m != nil { - return m.Enum - } - return nil -} - -func (m *FormDataParameterSubSchema) GetMultipleOf() float64 { - if m != nil { - return m.MultipleOf - } - return 0 -} - -func (m *FormDataParameterSubSchema) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Header struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Format string `protobuf:"bytes,2,opt,name=format,proto3" json:"format,omitempty"` - Items *PrimitivesItems `protobuf:"bytes,3,opt,name=items,proto3" json:"items,omitempty"` - CollectionFormat string `protobuf:"bytes,4,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` - Default *Any `protobuf:"bytes,5,opt,name=default,proto3" json:"default,omitempty"` - Maximum float64 `protobuf:"fixed64,6,opt,name=maximum,proto3" json:"maximum,omitempty"` - ExclusiveMaximum bool `protobuf:"varint,7,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` - Minimum float64 `protobuf:"fixed64,8,opt,name=minimum,proto3" json:"minimum,omitempty"` - ExclusiveMinimum bool `protobuf:"varint,9,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` - MaxLength int64 `protobuf:"varint,10,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` - MinLength int64 `protobuf:"varint,11,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` - Pattern string `protobuf:"bytes,12,opt,name=pattern,proto3" json:"pattern,omitempty"` - MaxItems int64 `protobuf:"varint,13,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` - MinItems int64 `protobuf:"varint,14,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` - UniqueItems bool `protobuf:"varint,15,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` - Enum []*Any `protobuf:"bytes,16,rep,name=enum,proto3" json:"enum,omitempty"` - MultipleOf float64 `protobuf:"fixed64,17,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` - Description string `protobuf:"bytes,18,opt,name=description,proto3" json:"description,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,19,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Header) Reset() { *m = Header{} } -func (m *Header) String() string { return proto.CompactTextString(m) } -func (*Header) ProtoMessage() {} -func (*Header) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{13} -} - -func (m *Header) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Header.Unmarshal(m, b) -} -func (m *Header) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Header.Marshal(b, m, deterministic) -} -func (m *Header) XXX_Merge(src proto.Message) { - xxx_messageInfo_Header.Merge(m, src) -} -func (m *Header) XXX_Size() int { - return xxx_messageInfo_Header.Size(m) -} -func (m *Header) XXX_DiscardUnknown() { - xxx_messageInfo_Header.DiscardUnknown(m) -} - -var xxx_messageInfo_Header proto.InternalMessageInfo - -func (m *Header) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *Header) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *Header) GetItems() *PrimitivesItems { - if m != nil { - return m.Items - } - return nil -} - -func (m *Header) GetCollectionFormat() string { - if m != nil { - return m.CollectionFormat - } - return "" -} - -func (m *Header) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *Header) GetMaximum() float64 { - if m != nil { - return m.Maximum - } - return 0 -} - -func (m *Header) GetExclusiveMaximum() bool { - if m != nil { - return m.ExclusiveMaximum - } - return false -} - -func (m *Header) GetMinimum() float64 { - if m != nil { - return m.Minimum - } - return 0 -} - -func (m *Header) GetExclusiveMinimum() bool { - if m != nil { - return m.ExclusiveMinimum - } - return false -} - -func (m *Header) GetMaxLength() int64 { - if m != nil { - return m.MaxLength - } - return 0 -} - -func (m *Header) GetMinLength() int64 { - if m != nil { - return m.MinLength - } - return 0 -} - -func (m *Header) GetPattern() string { - if m != nil { - return m.Pattern - } - return "" -} - -func (m *Header) GetMaxItems() int64 { - if m != nil { - return m.MaxItems - } - return 0 -} - -func (m *Header) GetMinItems() int64 { - if m != nil { - return m.MinItems - } - return 0 -} - -func (m *Header) GetUniqueItems() bool { - if m != nil { - return m.UniqueItems - } - return false -} - -func (m *Header) GetEnum() []*Any { - if m != nil { - return m.Enum - } - return nil -} - -func (m *Header) GetMultipleOf() float64 { - if m != nil { - return m.MultipleOf - } - return 0 -} - -func (m *Header) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Header) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type HeaderParameterSubSchema struct { - // Determines whether or not this parameter is required or optional. - Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` - // Determines the location of the parameter. - In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` - // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. - Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` - // The name of the parameter. - Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` - Type string `protobuf:"bytes,5,opt,name=type,proto3" json:"type,omitempty"` - Format string `protobuf:"bytes,6,opt,name=format,proto3" json:"format,omitempty"` - Items *PrimitivesItems `protobuf:"bytes,7,opt,name=items,proto3" json:"items,omitempty"` - CollectionFormat string `protobuf:"bytes,8,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` - Default *Any `protobuf:"bytes,9,opt,name=default,proto3" json:"default,omitempty"` - Maximum float64 `protobuf:"fixed64,10,opt,name=maximum,proto3" json:"maximum,omitempty"` - ExclusiveMaximum bool `protobuf:"varint,11,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` - Minimum float64 `protobuf:"fixed64,12,opt,name=minimum,proto3" json:"minimum,omitempty"` - ExclusiveMinimum bool `protobuf:"varint,13,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` - MaxLength int64 `protobuf:"varint,14,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` - MinLength int64 `protobuf:"varint,15,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` - Pattern string `protobuf:"bytes,16,opt,name=pattern,proto3" json:"pattern,omitempty"` - MaxItems int64 `protobuf:"varint,17,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` - MinItems int64 `protobuf:"varint,18,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` - UniqueItems bool `protobuf:"varint,19,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` - Enum []*Any `protobuf:"bytes,20,rep,name=enum,proto3" json:"enum,omitempty"` - MultipleOf float64 `protobuf:"fixed64,21,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,22,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *HeaderParameterSubSchema) Reset() { *m = HeaderParameterSubSchema{} } -func (m *HeaderParameterSubSchema) String() string { return proto.CompactTextString(m) } -func (*HeaderParameterSubSchema) ProtoMessage() {} -func (*HeaderParameterSubSchema) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{14} -} - -func (m *HeaderParameterSubSchema) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_HeaderParameterSubSchema.Unmarshal(m, b) -} -func (m *HeaderParameterSubSchema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_HeaderParameterSubSchema.Marshal(b, m, deterministic) -} -func (m *HeaderParameterSubSchema) XXX_Merge(src proto.Message) { - xxx_messageInfo_HeaderParameterSubSchema.Merge(m, src) -} -func (m *HeaderParameterSubSchema) XXX_Size() int { - return xxx_messageInfo_HeaderParameterSubSchema.Size(m) -} -func (m *HeaderParameterSubSchema) XXX_DiscardUnknown() { - xxx_messageInfo_HeaderParameterSubSchema.DiscardUnknown(m) -} - -var xxx_messageInfo_HeaderParameterSubSchema proto.InternalMessageInfo - -func (m *HeaderParameterSubSchema) GetRequired() bool { - if m != nil { - return m.Required - } - return false -} - -func (m *HeaderParameterSubSchema) GetIn() string { - if m != nil { - return m.In - } - return "" -} - -func (m *HeaderParameterSubSchema) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *HeaderParameterSubSchema) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *HeaderParameterSubSchema) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *HeaderParameterSubSchema) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *HeaderParameterSubSchema) GetItems() *PrimitivesItems { - if m != nil { - return m.Items - } - return nil -} - -func (m *HeaderParameterSubSchema) GetCollectionFormat() string { - if m != nil { - return m.CollectionFormat - } - return "" -} - -func (m *HeaderParameterSubSchema) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *HeaderParameterSubSchema) GetMaximum() float64 { - if m != nil { - return m.Maximum - } - return 0 -} - -func (m *HeaderParameterSubSchema) GetExclusiveMaximum() bool { - if m != nil { - return m.ExclusiveMaximum - } - return false -} - -func (m *HeaderParameterSubSchema) GetMinimum() float64 { - if m != nil { - return m.Minimum - } - return 0 -} - -func (m *HeaderParameterSubSchema) GetExclusiveMinimum() bool { - if m != nil { - return m.ExclusiveMinimum - } - return false -} - -func (m *HeaderParameterSubSchema) GetMaxLength() int64 { - if m != nil { - return m.MaxLength - } - return 0 -} - -func (m *HeaderParameterSubSchema) GetMinLength() int64 { - if m != nil { - return m.MinLength - } - return 0 -} - -func (m *HeaderParameterSubSchema) GetPattern() string { - if m != nil { - return m.Pattern - } - return "" -} - -func (m *HeaderParameterSubSchema) GetMaxItems() int64 { - if m != nil { - return m.MaxItems - } - return 0 -} - -func (m *HeaderParameterSubSchema) GetMinItems() int64 { - if m != nil { - return m.MinItems - } - return 0 -} - -func (m *HeaderParameterSubSchema) GetUniqueItems() bool { - if m != nil { - return m.UniqueItems - } - return false -} - -func (m *HeaderParameterSubSchema) GetEnum() []*Any { - if m != nil { - return m.Enum - } - return nil -} - -func (m *HeaderParameterSubSchema) GetMultipleOf() float64 { - if m != nil { - return m.MultipleOf - } - return 0 -} - -func (m *HeaderParameterSubSchema) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Headers struct { - AdditionalProperties []*NamedHeader `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Headers) Reset() { *m = Headers{} } -func (m *Headers) String() string { return proto.CompactTextString(m) } -func (*Headers) ProtoMessage() {} -func (*Headers) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{15} -} - -func (m *Headers) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Headers.Unmarshal(m, b) -} -func (m *Headers) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Headers.Marshal(b, m, deterministic) -} -func (m *Headers) XXX_Merge(src proto.Message) { - xxx_messageInfo_Headers.Merge(m, src) -} -func (m *Headers) XXX_Size() int { - return xxx_messageInfo_Headers.Size(m) -} -func (m *Headers) XXX_DiscardUnknown() { - xxx_messageInfo_Headers.DiscardUnknown(m) -} - -var xxx_messageInfo_Headers proto.InternalMessageInfo - -func (m *Headers) GetAdditionalProperties() []*NamedHeader { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -// General information about the API. -type Info struct { - // A unique and precise title of the API. - Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` - // A semantic version number of the API. - Version string `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` - // A longer description of the API. Should be different from the title. GitHub Flavored Markdown is allowed. - Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` - // The terms of service for the API. - TermsOfService string `protobuf:"bytes,4,opt,name=terms_of_service,json=termsOfService,proto3" json:"terms_of_service,omitempty"` - Contact *Contact `protobuf:"bytes,5,opt,name=contact,proto3" json:"contact,omitempty"` - License *License `protobuf:"bytes,6,opt,name=license,proto3" json:"license,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,7,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Info) Reset() { *m = Info{} } -func (m *Info) String() string { return proto.CompactTextString(m) } -func (*Info) ProtoMessage() {} -func (*Info) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{16} -} - -func (m *Info) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Info.Unmarshal(m, b) -} -func (m *Info) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Info.Marshal(b, m, deterministic) -} -func (m *Info) XXX_Merge(src proto.Message) { - xxx_messageInfo_Info.Merge(m, src) -} -func (m *Info) XXX_Size() int { - return xxx_messageInfo_Info.Size(m) -} -func (m *Info) XXX_DiscardUnknown() { - xxx_messageInfo_Info.DiscardUnknown(m) -} - -var xxx_messageInfo_Info proto.InternalMessageInfo - -func (m *Info) GetTitle() string { - if m != nil { - return m.Title - } - return "" -} - -func (m *Info) GetVersion() string { - if m != nil { - return m.Version - } - return "" -} - -func (m *Info) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Info) GetTermsOfService() string { - if m != nil { - return m.TermsOfService - } - return "" -} - -func (m *Info) GetContact() *Contact { - if m != nil { - return m.Contact - } - return nil -} - -func (m *Info) GetLicense() *License { - if m != nil { - return m.License - } - return nil -} - -func (m *Info) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type ItemsItem struct { - Schema []*Schema `protobuf:"bytes,1,rep,name=schema,proto3" json:"schema,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ItemsItem) Reset() { *m = ItemsItem{} } -func (m *ItemsItem) String() string { return proto.CompactTextString(m) } -func (*ItemsItem) ProtoMessage() {} -func (*ItemsItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{17} -} - -func (m *ItemsItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ItemsItem.Unmarshal(m, b) -} -func (m *ItemsItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ItemsItem.Marshal(b, m, deterministic) -} -func (m *ItemsItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_ItemsItem.Merge(m, src) -} -func (m *ItemsItem) XXX_Size() int { - return xxx_messageInfo_ItemsItem.Size(m) -} -func (m *ItemsItem) XXX_DiscardUnknown() { - xxx_messageInfo_ItemsItem.DiscardUnknown(m) -} - -var xxx_messageInfo_ItemsItem proto.InternalMessageInfo - -func (m *ItemsItem) GetSchema() []*Schema { - if m != nil { - return m.Schema - } - return nil -} - -type JsonReference struct { - XRef string `protobuf:"bytes,1,opt,name=_ref,json=Ref,proto3" json:"_ref,omitempty"` - Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *JsonReference) Reset() { *m = JsonReference{} } -func (m *JsonReference) String() string { return proto.CompactTextString(m) } -func (*JsonReference) ProtoMessage() {} -func (*JsonReference) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{18} -} - -func (m *JsonReference) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_JsonReference.Unmarshal(m, b) -} -func (m *JsonReference) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_JsonReference.Marshal(b, m, deterministic) -} -func (m *JsonReference) XXX_Merge(src proto.Message) { - xxx_messageInfo_JsonReference.Merge(m, src) -} -func (m *JsonReference) XXX_Size() int { - return xxx_messageInfo_JsonReference.Size(m) -} -func (m *JsonReference) XXX_DiscardUnknown() { - xxx_messageInfo_JsonReference.DiscardUnknown(m) -} - -var xxx_messageInfo_JsonReference proto.InternalMessageInfo - -func (m *JsonReference) GetXRef() string { - if m != nil { - return m.XRef - } - return "" -} - -func (m *JsonReference) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -type License struct { - // The name of the license type. It's encouraged to use an OSI compatible license. - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // The URL pointing to the license. - Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,3,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *License) Reset() { *m = License{} } -func (m *License) String() string { return proto.CompactTextString(m) } -func (*License) ProtoMessage() {} -func (*License) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{19} -} - -func (m *License) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_License.Unmarshal(m, b) -} -func (m *License) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_License.Marshal(b, m, deterministic) -} -func (m *License) XXX_Merge(src proto.Message) { - xxx_messageInfo_License.Merge(m, src) -} -func (m *License) XXX_Size() int { - return xxx_messageInfo_License.Size(m) -} -func (m *License) XXX_DiscardUnknown() { - xxx_messageInfo_License.DiscardUnknown(m) -} - -var xxx_messageInfo_License proto.InternalMessageInfo - -func (m *License) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *License) GetUrl() string { - if m != nil { - return m.Url - } - return "" -} - -func (m *License) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -// Automatically-generated message used to represent maps of Any as ordered (name,value) pairs. -type NamedAny struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *Any `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedAny) Reset() { *m = NamedAny{} } -func (m *NamedAny) String() string { return proto.CompactTextString(m) } -func (*NamedAny) ProtoMessage() {} -func (*NamedAny) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{20} -} - -func (m *NamedAny) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedAny.Unmarshal(m, b) -} -func (m *NamedAny) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedAny.Marshal(b, m, deterministic) -} -func (m *NamedAny) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedAny.Merge(m, src) -} -func (m *NamedAny) XXX_Size() int { - return xxx_messageInfo_NamedAny.Size(m) -} -func (m *NamedAny) XXX_DiscardUnknown() { - xxx_messageInfo_NamedAny.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedAny proto.InternalMessageInfo - -func (m *NamedAny) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedAny) GetValue() *Any { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of Header as ordered (name,value) pairs. -type NamedHeader struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *Header `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedHeader) Reset() { *m = NamedHeader{} } -func (m *NamedHeader) String() string { return proto.CompactTextString(m) } -func (*NamedHeader) ProtoMessage() {} -func (*NamedHeader) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{21} -} - -func (m *NamedHeader) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedHeader.Unmarshal(m, b) -} -func (m *NamedHeader) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedHeader.Marshal(b, m, deterministic) -} -func (m *NamedHeader) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedHeader.Merge(m, src) -} -func (m *NamedHeader) XXX_Size() int { - return xxx_messageInfo_NamedHeader.Size(m) -} -func (m *NamedHeader) XXX_DiscardUnknown() { - xxx_messageInfo_NamedHeader.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedHeader proto.InternalMessageInfo - -func (m *NamedHeader) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedHeader) GetValue() *Header { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of Parameter as ordered (name,value) pairs. -type NamedParameter struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *Parameter `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedParameter) Reset() { *m = NamedParameter{} } -func (m *NamedParameter) String() string { return proto.CompactTextString(m) } -func (*NamedParameter) ProtoMessage() {} -func (*NamedParameter) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{22} -} - -func (m *NamedParameter) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedParameter.Unmarshal(m, b) -} -func (m *NamedParameter) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedParameter.Marshal(b, m, deterministic) -} -func (m *NamedParameter) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedParameter.Merge(m, src) -} -func (m *NamedParameter) XXX_Size() int { - return xxx_messageInfo_NamedParameter.Size(m) -} -func (m *NamedParameter) XXX_DiscardUnknown() { - xxx_messageInfo_NamedParameter.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedParameter proto.InternalMessageInfo - -func (m *NamedParameter) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedParameter) GetValue() *Parameter { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of PathItem as ordered (name,value) pairs. -type NamedPathItem struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *PathItem `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedPathItem) Reset() { *m = NamedPathItem{} } -func (m *NamedPathItem) String() string { return proto.CompactTextString(m) } -func (*NamedPathItem) ProtoMessage() {} -func (*NamedPathItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{23} -} - -func (m *NamedPathItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedPathItem.Unmarshal(m, b) -} -func (m *NamedPathItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedPathItem.Marshal(b, m, deterministic) -} -func (m *NamedPathItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedPathItem.Merge(m, src) -} -func (m *NamedPathItem) XXX_Size() int { - return xxx_messageInfo_NamedPathItem.Size(m) -} -func (m *NamedPathItem) XXX_DiscardUnknown() { - xxx_messageInfo_NamedPathItem.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedPathItem proto.InternalMessageInfo - -func (m *NamedPathItem) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedPathItem) GetValue() *PathItem { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of Response as ordered (name,value) pairs. -type NamedResponse struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *Response `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedResponse) Reset() { *m = NamedResponse{} } -func (m *NamedResponse) String() string { return proto.CompactTextString(m) } -func (*NamedResponse) ProtoMessage() {} -func (*NamedResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{24} -} - -func (m *NamedResponse) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedResponse.Unmarshal(m, b) -} -func (m *NamedResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedResponse.Marshal(b, m, deterministic) -} -func (m *NamedResponse) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedResponse.Merge(m, src) -} -func (m *NamedResponse) XXX_Size() int { - return xxx_messageInfo_NamedResponse.Size(m) -} -func (m *NamedResponse) XXX_DiscardUnknown() { - xxx_messageInfo_NamedResponse.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedResponse proto.InternalMessageInfo - -func (m *NamedResponse) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedResponse) GetValue() *Response { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of ResponseValue as ordered (name,value) pairs. -type NamedResponseValue struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *ResponseValue `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedResponseValue) Reset() { *m = NamedResponseValue{} } -func (m *NamedResponseValue) String() string { return proto.CompactTextString(m) } -func (*NamedResponseValue) ProtoMessage() {} -func (*NamedResponseValue) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{25} -} - -func (m *NamedResponseValue) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedResponseValue.Unmarshal(m, b) -} -func (m *NamedResponseValue) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedResponseValue.Marshal(b, m, deterministic) -} -func (m *NamedResponseValue) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedResponseValue.Merge(m, src) -} -func (m *NamedResponseValue) XXX_Size() int { - return xxx_messageInfo_NamedResponseValue.Size(m) -} -func (m *NamedResponseValue) XXX_DiscardUnknown() { - xxx_messageInfo_NamedResponseValue.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedResponseValue proto.InternalMessageInfo - -func (m *NamedResponseValue) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedResponseValue) GetValue() *ResponseValue { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of Schema as ordered (name,value) pairs. -type NamedSchema struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *Schema `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedSchema) Reset() { *m = NamedSchema{} } -func (m *NamedSchema) String() string { return proto.CompactTextString(m) } -func (*NamedSchema) ProtoMessage() {} -func (*NamedSchema) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{26} -} - -func (m *NamedSchema) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedSchema.Unmarshal(m, b) -} -func (m *NamedSchema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedSchema.Marshal(b, m, deterministic) -} -func (m *NamedSchema) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedSchema.Merge(m, src) -} -func (m *NamedSchema) XXX_Size() int { - return xxx_messageInfo_NamedSchema.Size(m) -} -func (m *NamedSchema) XXX_DiscardUnknown() { - xxx_messageInfo_NamedSchema.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedSchema proto.InternalMessageInfo - -func (m *NamedSchema) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedSchema) GetValue() *Schema { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of SecurityDefinitionsItem as ordered (name,value) pairs. -type NamedSecurityDefinitionsItem struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *SecurityDefinitionsItem `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedSecurityDefinitionsItem) Reset() { *m = NamedSecurityDefinitionsItem{} } -func (m *NamedSecurityDefinitionsItem) String() string { return proto.CompactTextString(m) } -func (*NamedSecurityDefinitionsItem) ProtoMessage() {} -func (*NamedSecurityDefinitionsItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{27} -} - -func (m *NamedSecurityDefinitionsItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedSecurityDefinitionsItem.Unmarshal(m, b) -} -func (m *NamedSecurityDefinitionsItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedSecurityDefinitionsItem.Marshal(b, m, deterministic) -} -func (m *NamedSecurityDefinitionsItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedSecurityDefinitionsItem.Merge(m, src) -} -func (m *NamedSecurityDefinitionsItem) XXX_Size() int { - return xxx_messageInfo_NamedSecurityDefinitionsItem.Size(m) -} -func (m *NamedSecurityDefinitionsItem) XXX_DiscardUnknown() { - xxx_messageInfo_NamedSecurityDefinitionsItem.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedSecurityDefinitionsItem proto.InternalMessageInfo - -func (m *NamedSecurityDefinitionsItem) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedSecurityDefinitionsItem) GetValue() *SecurityDefinitionsItem { - if m != nil { - return m.Value - } - return nil -} - -// Automatically-generated message used to represent maps of string as ordered (name,value) pairs. -type NamedString struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedString) Reset() { *m = NamedString{} } -func (m *NamedString) String() string { return proto.CompactTextString(m) } -func (*NamedString) ProtoMessage() {} -func (*NamedString) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{28} -} - -func (m *NamedString) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedString.Unmarshal(m, b) -} -func (m *NamedString) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedString.Marshal(b, m, deterministic) -} -func (m *NamedString) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedString.Merge(m, src) -} -func (m *NamedString) XXX_Size() int { - return xxx_messageInfo_NamedString.Size(m) -} -func (m *NamedString) XXX_DiscardUnknown() { - xxx_messageInfo_NamedString.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedString proto.InternalMessageInfo - -func (m *NamedString) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedString) GetValue() string { - if m != nil { - return m.Value - } - return "" -} - -// Automatically-generated message used to represent maps of StringArray as ordered (name,value) pairs. -type NamedStringArray struct { - // Map key - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - // Mapped value - Value *StringArray `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NamedStringArray) Reset() { *m = NamedStringArray{} } -func (m *NamedStringArray) String() string { return proto.CompactTextString(m) } -func (*NamedStringArray) ProtoMessage() {} -func (*NamedStringArray) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{29} -} - -func (m *NamedStringArray) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NamedStringArray.Unmarshal(m, b) -} -func (m *NamedStringArray) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NamedStringArray.Marshal(b, m, deterministic) -} -func (m *NamedStringArray) XXX_Merge(src proto.Message) { - xxx_messageInfo_NamedStringArray.Merge(m, src) -} -func (m *NamedStringArray) XXX_Size() int { - return xxx_messageInfo_NamedStringArray.Size(m) -} -func (m *NamedStringArray) XXX_DiscardUnknown() { - xxx_messageInfo_NamedStringArray.DiscardUnknown(m) -} - -var xxx_messageInfo_NamedStringArray proto.InternalMessageInfo - -func (m *NamedStringArray) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *NamedStringArray) GetValue() *StringArray { - if m != nil { - return m.Value - } - return nil -} - -type NonBodyParameter struct { - // Types that are valid to be assigned to Oneof: - // *NonBodyParameter_HeaderParameterSubSchema - // *NonBodyParameter_FormDataParameterSubSchema - // *NonBodyParameter_QueryParameterSubSchema - // *NonBodyParameter_PathParameterSubSchema - Oneof isNonBodyParameter_Oneof `protobuf_oneof:"oneof"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NonBodyParameter) Reset() { *m = NonBodyParameter{} } -func (m *NonBodyParameter) String() string { return proto.CompactTextString(m) } -func (*NonBodyParameter) ProtoMessage() {} -func (*NonBodyParameter) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{30} -} - -func (m *NonBodyParameter) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NonBodyParameter.Unmarshal(m, b) -} -func (m *NonBodyParameter) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NonBodyParameter.Marshal(b, m, deterministic) -} -func (m *NonBodyParameter) XXX_Merge(src proto.Message) { - xxx_messageInfo_NonBodyParameter.Merge(m, src) -} -func (m *NonBodyParameter) XXX_Size() int { - return xxx_messageInfo_NonBodyParameter.Size(m) -} -func (m *NonBodyParameter) XXX_DiscardUnknown() { - xxx_messageInfo_NonBodyParameter.DiscardUnknown(m) -} - -var xxx_messageInfo_NonBodyParameter proto.InternalMessageInfo - -type isNonBodyParameter_Oneof interface { - isNonBodyParameter_Oneof() -} - -type NonBodyParameter_HeaderParameterSubSchema struct { - HeaderParameterSubSchema *HeaderParameterSubSchema `protobuf:"bytes,1,opt,name=header_parameter_sub_schema,json=headerParameterSubSchema,proto3,oneof"` -} - -type NonBodyParameter_FormDataParameterSubSchema struct { - FormDataParameterSubSchema *FormDataParameterSubSchema `protobuf:"bytes,2,opt,name=form_data_parameter_sub_schema,json=formDataParameterSubSchema,proto3,oneof"` -} - -type NonBodyParameter_QueryParameterSubSchema struct { - QueryParameterSubSchema *QueryParameterSubSchema `protobuf:"bytes,3,opt,name=query_parameter_sub_schema,json=queryParameterSubSchema,proto3,oneof"` -} - -type NonBodyParameter_PathParameterSubSchema struct { - PathParameterSubSchema *PathParameterSubSchema `protobuf:"bytes,4,opt,name=path_parameter_sub_schema,json=pathParameterSubSchema,proto3,oneof"` -} - -func (*NonBodyParameter_HeaderParameterSubSchema) isNonBodyParameter_Oneof() {} - -func (*NonBodyParameter_FormDataParameterSubSchema) isNonBodyParameter_Oneof() {} - -func (*NonBodyParameter_QueryParameterSubSchema) isNonBodyParameter_Oneof() {} - -func (*NonBodyParameter_PathParameterSubSchema) isNonBodyParameter_Oneof() {} - -func (m *NonBodyParameter) GetOneof() isNonBodyParameter_Oneof { - if m != nil { - return m.Oneof - } - return nil -} - -func (m *NonBodyParameter) GetHeaderParameterSubSchema() *HeaderParameterSubSchema { - if x, ok := m.GetOneof().(*NonBodyParameter_HeaderParameterSubSchema); ok { - return x.HeaderParameterSubSchema - } - return nil -} - -func (m *NonBodyParameter) GetFormDataParameterSubSchema() *FormDataParameterSubSchema { - if x, ok := m.GetOneof().(*NonBodyParameter_FormDataParameterSubSchema); ok { - return x.FormDataParameterSubSchema - } - return nil -} - -func (m *NonBodyParameter) GetQueryParameterSubSchema() *QueryParameterSubSchema { - if x, ok := m.GetOneof().(*NonBodyParameter_QueryParameterSubSchema); ok { - return x.QueryParameterSubSchema - } - return nil -} - -func (m *NonBodyParameter) GetPathParameterSubSchema() *PathParameterSubSchema { - if x, ok := m.GetOneof().(*NonBodyParameter_PathParameterSubSchema); ok { - return x.PathParameterSubSchema - } - return nil -} - -// XXX_OneofWrappers is for the internal use of the proto package. -func (*NonBodyParameter) XXX_OneofWrappers() []interface{} { - return []interface{}{ - (*NonBodyParameter_HeaderParameterSubSchema)(nil), - (*NonBodyParameter_FormDataParameterSubSchema)(nil), - (*NonBodyParameter_QueryParameterSubSchema)(nil), - (*NonBodyParameter_PathParameterSubSchema)(nil), - } -} - -type Oauth2AccessCodeSecurity struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` - Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` - AuthorizationUrl string `protobuf:"bytes,4,opt,name=authorization_url,json=authorizationUrl,proto3" json:"authorization_url,omitempty"` - TokenUrl string `protobuf:"bytes,5,opt,name=token_url,json=tokenUrl,proto3" json:"token_url,omitempty"` - Description string `protobuf:"bytes,6,opt,name=description,proto3" json:"description,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,7,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Oauth2AccessCodeSecurity) Reset() { *m = Oauth2AccessCodeSecurity{} } -func (m *Oauth2AccessCodeSecurity) String() string { return proto.CompactTextString(m) } -func (*Oauth2AccessCodeSecurity) ProtoMessage() {} -func (*Oauth2AccessCodeSecurity) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{31} -} - -func (m *Oauth2AccessCodeSecurity) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Oauth2AccessCodeSecurity.Unmarshal(m, b) -} -func (m *Oauth2AccessCodeSecurity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Oauth2AccessCodeSecurity.Marshal(b, m, deterministic) -} -func (m *Oauth2AccessCodeSecurity) XXX_Merge(src proto.Message) { - xxx_messageInfo_Oauth2AccessCodeSecurity.Merge(m, src) -} -func (m *Oauth2AccessCodeSecurity) XXX_Size() int { - return xxx_messageInfo_Oauth2AccessCodeSecurity.Size(m) -} -func (m *Oauth2AccessCodeSecurity) XXX_DiscardUnknown() { - xxx_messageInfo_Oauth2AccessCodeSecurity.DiscardUnknown(m) -} - -var xxx_messageInfo_Oauth2AccessCodeSecurity proto.InternalMessageInfo - -func (m *Oauth2AccessCodeSecurity) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *Oauth2AccessCodeSecurity) GetFlow() string { - if m != nil { - return m.Flow - } - return "" -} - -func (m *Oauth2AccessCodeSecurity) GetScopes() *Oauth2Scopes { - if m != nil { - return m.Scopes - } - return nil -} - -func (m *Oauth2AccessCodeSecurity) GetAuthorizationUrl() string { - if m != nil { - return m.AuthorizationUrl - } - return "" -} - -func (m *Oauth2AccessCodeSecurity) GetTokenUrl() string { - if m != nil { - return m.TokenUrl - } - return "" -} - -func (m *Oauth2AccessCodeSecurity) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Oauth2AccessCodeSecurity) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Oauth2ApplicationSecurity struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` - Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` - TokenUrl string `protobuf:"bytes,4,opt,name=token_url,json=tokenUrl,proto3" json:"token_url,omitempty"` - Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Oauth2ApplicationSecurity) Reset() { *m = Oauth2ApplicationSecurity{} } -func (m *Oauth2ApplicationSecurity) String() string { return proto.CompactTextString(m) } -func (*Oauth2ApplicationSecurity) ProtoMessage() {} -func (*Oauth2ApplicationSecurity) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{32} -} - -func (m *Oauth2ApplicationSecurity) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Oauth2ApplicationSecurity.Unmarshal(m, b) -} -func (m *Oauth2ApplicationSecurity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Oauth2ApplicationSecurity.Marshal(b, m, deterministic) -} -func (m *Oauth2ApplicationSecurity) XXX_Merge(src proto.Message) { - xxx_messageInfo_Oauth2ApplicationSecurity.Merge(m, src) -} -func (m *Oauth2ApplicationSecurity) XXX_Size() int { - return xxx_messageInfo_Oauth2ApplicationSecurity.Size(m) -} -func (m *Oauth2ApplicationSecurity) XXX_DiscardUnknown() { - xxx_messageInfo_Oauth2ApplicationSecurity.DiscardUnknown(m) -} - -var xxx_messageInfo_Oauth2ApplicationSecurity proto.InternalMessageInfo - -func (m *Oauth2ApplicationSecurity) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *Oauth2ApplicationSecurity) GetFlow() string { - if m != nil { - return m.Flow - } - return "" -} - -func (m *Oauth2ApplicationSecurity) GetScopes() *Oauth2Scopes { - if m != nil { - return m.Scopes - } - return nil -} - -func (m *Oauth2ApplicationSecurity) GetTokenUrl() string { - if m != nil { - return m.TokenUrl - } - return "" -} - -func (m *Oauth2ApplicationSecurity) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Oauth2ApplicationSecurity) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Oauth2ImplicitSecurity struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` - Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` - AuthorizationUrl string `protobuf:"bytes,4,opt,name=authorization_url,json=authorizationUrl,proto3" json:"authorization_url,omitempty"` - Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Oauth2ImplicitSecurity) Reset() { *m = Oauth2ImplicitSecurity{} } -func (m *Oauth2ImplicitSecurity) String() string { return proto.CompactTextString(m) } -func (*Oauth2ImplicitSecurity) ProtoMessage() {} -func (*Oauth2ImplicitSecurity) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{33} -} - -func (m *Oauth2ImplicitSecurity) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Oauth2ImplicitSecurity.Unmarshal(m, b) -} -func (m *Oauth2ImplicitSecurity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Oauth2ImplicitSecurity.Marshal(b, m, deterministic) -} -func (m *Oauth2ImplicitSecurity) XXX_Merge(src proto.Message) { - xxx_messageInfo_Oauth2ImplicitSecurity.Merge(m, src) -} -func (m *Oauth2ImplicitSecurity) XXX_Size() int { - return xxx_messageInfo_Oauth2ImplicitSecurity.Size(m) -} -func (m *Oauth2ImplicitSecurity) XXX_DiscardUnknown() { - xxx_messageInfo_Oauth2ImplicitSecurity.DiscardUnknown(m) -} - -var xxx_messageInfo_Oauth2ImplicitSecurity proto.InternalMessageInfo - -func (m *Oauth2ImplicitSecurity) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *Oauth2ImplicitSecurity) GetFlow() string { - if m != nil { - return m.Flow - } - return "" -} - -func (m *Oauth2ImplicitSecurity) GetScopes() *Oauth2Scopes { - if m != nil { - return m.Scopes - } - return nil -} - -func (m *Oauth2ImplicitSecurity) GetAuthorizationUrl() string { - if m != nil { - return m.AuthorizationUrl - } - return "" -} - -func (m *Oauth2ImplicitSecurity) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Oauth2ImplicitSecurity) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Oauth2PasswordSecurity struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` - Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` - TokenUrl string `protobuf:"bytes,4,opt,name=token_url,json=tokenUrl,proto3" json:"token_url,omitempty"` - Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Oauth2PasswordSecurity) Reset() { *m = Oauth2PasswordSecurity{} } -func (m *Oauth2PasswordSecurity) String() string { return proto.CompactTextString(m) } -func (*Oauth2PasswordSecurity) ProtoMessage() {} -func (*Oauth2PasswordSecurity) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{34} -} - -func (m *Oauth2PasswordSecurity) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Oauth2PasswordSecurity.Unmarshal(m, b) -} -func (m *Oauth2PasswordSecurity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Oauth2PasswordSecurity.Marshal(b, m, deterministic) -} -func (m *Oauth2PasswordSecurity) XXX_Merge(src proto.Message) { - xxx_messageInfo_Oauth2PasswordSecurity.Merge(m, src) -} -func (m *Oauth2PasswordSecurity) XXX_Size() int { - return xxx_messageInfo_Oauth2PasswordSecurity.Size(m) -} -func (m *Oauth2PasswordSecurity) XXX_DiscardUnknown() { - xxx_messageInfo_Oauth2PasswordSecurity.DiscardUnknown(m) -} - -var xxx_messageInfo_Oauth2PasswordSecurity proto.InternalMessageInfo - -func (m *Oauth2PasswordSecurity) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *Oauth2PasswordSecurity) GetFlow() string { - if m != nil { - return m.Flow - } - return "" -} - -func (m *Oauth2PasswordSecurity) GetScopes() *Oauth2Scopes { - if m != nil { - return m.Scopes - } - return nil -} - -func (m *Oauth2PasswordSecurity) GetTokenUrl() string { - if m != nil { - return m.TokenUrl - } - return "" -} - -func (m *Oauth2PasswordSecurity) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Oauth2PasswordSecurity) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Oauth2Scopes struct { - AdditionalProperties []*NamedString `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Oauth2Scopes) Reset() { *m = Oauth2Scopes{} } -func (m *Oauth2Scopes) String() string { return proto.CompactTextString(m) } -func (*Oauth2Scopes) ProtoMessage() {} -func (*Oauth2Scopes) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{35} -} - -func (m *Oauth2Scopes) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Oauth2Scopes.Unmarshal(m, b) -} -func (m *Oauth2Scopes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Oauth2Scopes.Marshal(b, m, deterministic) -} -func (m *Oauth2Scopes) XXX_Merge(src proto.Message) { - xxx_messageInfo_Oauth2Scopes.Merge(m, src) -} -func (m *Oauth2Scopes) XXX_Size() int { - return xxx_messageInfo_Oauth2Scopes.Size(m) -} -func (m *Oauth2Scopes) XXX_DiscardUnknown() { - xxx_messageInfo_Oauth2Scopes.DiscardUnknown(m) -} - -var xxx_messageInfo_Oauth2Scopes proto.InternalMessageInfo - -func (m *Oauth2Scopes) GetAdditionalProperties() []*NamedString { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type Operation struct { - Tags []string `protobuf:"bytes,1,rep,name=tags,proto3" json:"tags,omitempty"` - // A brief summary of the operation. - Summary string `protobuf:"bytes,2,opt,name=summary,proto3" json:"summary,omitempty"` - // A longer description of the operation, GitHub Flavored Markdown is allowed. - Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` - ExternalDocs *ExternalDocs `protobuf:"bytes,4,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` - // A unique identifier of the operation. - OperationId string `protobuf:"bytes,5,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` - // A list of MIME types the API can produce. - Produces []string `protobuf:"bytes,6,rep,name=produces,proto3" json:"produces,omitempty"` - // A list of MIME types the API can consume. - Consumes []string `protobuf:"bytes,7,rep,name=consumes,proto3" json:"consumes,omitempty"` - // The parameters needed to send a valid API call. - Parameters []*ParametersItem `protobuf:"bytes,8,rep,name=parameters,proto3" json:"parameters,omitempty"` - Responses *Responses `protobuf:"bytes,9,opt,name=responses,proto3" json:"responses,omitempty"` - // The transfer protocol of the API. - Schemes []string `protobuf:"bytes,10,rep,name=schemes,proto3" json:"schemes,omitempty"` - Deprecated bool `protobuf:"varint,11,opt,name=deprecated,proto3" json:"deprecated,omitempty"` - Security []*SecurityRequirement `protobuf:"bytes,12,rep,name=security,proto3" json:"security,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,13,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Operation) Reset() { *m = Operation{} } -func (m *Operation) String() string { return proto.CompactTextString(m) } -func (*Operation) ProtoMessage() {} -func (*Operation) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{36} -} - -func (m *Operation) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Operation.Unmarshal(m, b) -} -func (m *Operation) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Operation.Marshal(b, m, deterministic) -} -func (m *Operation) XXX_Merge(src proto.Message) { - xxx_messageInfo_Operation.Merge(m, src) -} -func (m *Operation) XXX_Size() int { - return xxx_messageInfo_Operation.Size(m) -} -func (m *Operation) XXX_DiscardUnknown() { - xxx_messageInfo_Operation.DiscardUnknown(m) -} - -var xxx_messageInfo_Operation proto.InternalMessageInfo - -func (m *Operation) GetTags() []string { - if m != nil { - return m.Tags - } - return nil -} - -func (m *Operation) GetSummary() string { - if m != nil { - return m.Summary - } - return "" -} - -func (m *Operation) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Operation) GetExternalDocs() *ExternalDocs { - if m != nil { - return m.ExternalDocs - } - return nil -} - -func (m *Operation) GetOperationId() string { - if m != nil { - return m.OperationId - } - return "" -} - -func (m *Operation) GetProduces() []string { - if m != nil { - return m.Produces - } - return nil -} - -func (m *Operation) GetConsumes() []string { - if m != nil { - return m.Consumes - } - return nil -} - -func (m *Operation) GetParameters() []*ParametersItem { - if m != nil { - return m.Parameters - } - return nil -} - -func (m *Operation) GetResponses() *Responses { - if m != nil { - return m.Responses - } - return nil -} - -func (m *Operation) GetSchemes() []string { - if m != nil { - return m.Schemes - } - return nil -} - -func (m *Operation) GetDeprecated() bool { - if m != nil { - return m.Deprecated - } - return false -} - -func (m *Operation) GetSecurity() []*SecurityRequirement { - if m != nil { - return m.Security - } - return nil -} - -func (m *Operation) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Parameter struct { - // Types that are valid to be assigned to Oneof: - // *Parameter_BodyParameter - // *Parameter_NonBodyParameter - Oneof isParameter_Oneof `protobuf_oneof:"oneof"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Parameter) Reset() { *m = Parameter{} } -func (m *Parameter) String() string { return proto.CompactTextString(m) } -func (*Parameter) ProtoMessage() {} -func (*Parameter) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{37} -} - -func (m *Parameter) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Parameter.Unmarshal(m, b) -} -func (m *Parameter) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Parameter.Marshal(b, m, deterministic) -} -func (m *Parameter) XXX_Merge(src proto.Message) { - xxx_messageInfo_Parameter.Merge(m, src) -} -func (m *Parameter) XXX_Size() int { - return xxx_messageInfo_Parameter.Size(m) -} -func (m *Parameter) XXX_DiscardUnknown() { - xxx_messageInfo_Parameter.DiscardUnknown(m) -} - -var xxx_messageInfo_Parameter proto.InternalMessageInfo - -type isParameter_Oneof interface { - isParameter_Oneof() -} - -type Parameter_BodyParameter struct { - BodyParameter *BodyParameter `protobuf:"bytes,1,opt,name=body_parameter,json=bodyParameter,proto3,oneof"` -} - -type Parameter_NonBodyParameter struct { - NonBodyParameter *NonBodyParameter `protobuf:"bytes,2,opt,name=non_body_parameter,json=nonBodyParameter,proto3,oneof"` -} - -func (*Parameter_BodyParameter) isParameter_Oneof() {} - -func (*Parameter_NonBodyParameter) isParameter_Oneof() {} - -func (m *Parameter) GetOneof() isParameter_Oneof { - if m != nil { - return m.Oneof - } - return nil -} - -func (m *Parameter) GetBodyParameter() *BodyParameter { - if x, ok := m.GetOneof().(*Parameter_BodyParameter); ok { - return x.BodyParameter - } - return nil -} - -func (m *Parameter) GetNonBodyParameter() *NonBodyParameter { - if x, ok := m.GetOneof().(*Parameter_NonBodyParameter); ok { - return x.NonBodyParameter - } - return nil -} - -// XXX_OneofWrappers is for the internal use of the proto package. -func (*Parameter) XXX_OneofWrappers() []interface{} { - return []interface{}{ - (*Parameter_BodyParameter)(nil), - (*Parameter_NonBodyParameter)(nil), - } -} - -// One or more JSON representations for parameters -type ParameterDefinitions struct { - AdditionalProperties []*NamedParameter `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ParameterDefinitions) Reset() { *m = ParameterDefinitions{} } -func (m *ParameterDefinitions) String() string { return proto.CompactTextString(m) } -func (*ParameterDefinitions) ProtoMessage() {} -func (*ParameterDefinitions) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{38} -} - -func (m *ParameterDefinitions) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ParameterDefinitions.Unmarshal(m, b) -} -func (m *ParameterDefinitions) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ParameterDefinitions.Marshal(b, m, deterministic) -} -func (m *ParameterDefinitions) XXX_Merge(src proto.Message) { - xxx_messageInfo_ParameterDefinitions.Merge(m, src) -} -func (m *ParameterDefinitions) XXX_Size() int { - return xxx_messageInfo_ParameterDefinitions.Size(m) -} -func (m *ParameterDefinitions) XXX_DiscardUnknown() { - xxx_messageInfo_ParameterDefinitions.DiscardUnknown(m) -} - -var xxx_messageInfo_ParameterDefinitions proto.InternalMessageInfo - -func (m *ParameterDefinitions) GetAdditionalProperties() []*NamedParameter { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type ParametersItem struct { - // Types that are valid to be assigned to Oneof: - // *ParametersItem_Parameter - // *ParametersItem_JsonReference - Oneof isParametersItem_Oneof `protobuf_oneof:"oneof"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ParametersItem) Reset() { *m = ParametersItem{} } -func (m *ParametersItem) String() string { return proto.CompactTextString(m) } -func (*ParametersItem) ProtoMessage() {} -func (*ParametersItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{39} -} - -func (m *ParametersItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ParametersItem.Unmarshal(m, b) -} -func (m *ParametersItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ParametersItem.Marshal(b, m, deterministic) -} -func (m *ParametersItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_ParametersItem.Merge(m, src) -} -func (m *ParametersItem) XXX_Size() int { - return xxx_messageInfo_ParametersItem.Size(m) -} -func (m *ParametersItem) XXX_DiscardUnknown() { - xxx_messageInfo_ParametersItem.DiscardUnknown(m) -} - -var xxx_messageInfo_ParametersItem proto.InternalMessageInfo - -type isParametersItem_Oneof interface { - isParametersItem_Oneof() -} - -type ParametersItem_Parameter struct { - Parameter *Parameter `protobuf:"bytes,1,opt,name=parameter,proto3,oneof"` -} - -type ParametersItem_JsonReference struct { - JsonReference *JsonReference `protobuf:"bytes,2,opt,name=json_reference,json=jsonReference,proto3,oneof"` -} - -func (*ParametersItem_Parameter) isParametersItem_Oneof() {} - -func (*ParametersItem_JsonReference) isParametersItem_Oneof() {} - -func (m *ParametersItem) GetOneof() isParametersItem_Oneof { - if m != nil { - return m.Oneof - } - return nil -} - -func (m *ParametersItem) GetParameter() *Parameter { - if x, ok := m.GetOneof().(*ParametersItem_Parameter); ok { - return x.Parameter - } - return nil -} - -func (m *ParametersItem) GetJsonReference() *JsonReference { - if x, ok := m.GetOneof().(*ParametersItem_JsonReference); ok { - return x.JsonReference - } - return nil -} - -// XXX_OneofWrappers is for the internal use of the proto package. -func (*ParametersItem) XXX_OneofWrappers() []interface{} { - return []interface{}{ - (*ParametersItem_Parameter)(nil), - (*ParametersItem_JsonReference)(nil), - } -} - -type PathItem struct { - XRef string `protobuf:"bytes,1,opt,name=_ref,json=Ref,proto3" json:"_ref,omitempty"` - Get *Operation `protobuf:"bytes,2,opt,name=get,proto3" json:"get,omitempty"` - Put *Operation `protobuf:"bytes,3,opt,name=put,proto3" json:"put,omitempty"` - Post *Operation `protobuf:"bytes,4,opt,name=post,proto3" json:"post,omitempty"` - Delete *Operation `protobuf:"bytes,5,opt,name=delete,proto3" json:"delete,omitempty"` - Options *Operation `protobuf:"bytes,6,opt,name=options,proto3" json:"options,omitempty"` - Head *Operation `protobuf:"bytes,7,opt,name=head,proto3" json:"head,omitempty"` - Patch *Operation `protobuf:"bytes,8,opt,name=patch,proto3" json:"patch,omitempty"` - // The parameters needed to send a valid API call. - Parameters []*ParametersItem `protobuf:"bytes,9,rep,name=parameters,proto3" json:"parameters,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,10,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *PathItem) Reset() { *m = PathItem{} } -func (m *PathItem) String() string { return proto.CompactTextString(m) } -func (*PathItem) ProtoMessage() {} -func (*PathItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{40} -} - -func (m *PathItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_PathItem.Unmarshal(m, b) -} -func (m *PathItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_PathItem.Marshal(b, m, deterministic) -} -func (m *PathItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_PathItem.Merge(m, src) -} -func (m *PathItem) XXX_Size() int { - return xxx_messageInfo_PathItem.Size(m) -} -func (m *PathItem) XXX_DiscardUnknown() { - xxx_messageInfo_PathItem.DiscardUnknown(m) -} - -var xxx_messageInfo_PathItem proto.InternalMessageInfo - -func (m *PathItem) GetXRef() string { - if m != nil { - return m.XRef - } - return "" -} - -func (m *PathItem) GetGet() *Operation { - if m != nil { - return m.Get - } - return nil -} - -func (m *PathItem) GetPut() *Operation { - if m != nil { - return m.Put - } - return nil -} - -func (m *PathItem) GetPost() *Operation { - if m != nil { - return m.Post - } - return nil -} - -func (m *PathItem) GetDelete() *Operation { - if m != nil { - return m.Delete - } - return nil -} - -func (m *PathItem) GetOptions() *Operation { - if m != nil { - return m.Options - } - return nil -} - -func (m *PathItem) GetHead() *Operation { - if m != nil { - return m.Head - } - return nil -} - -func (m *PathItem) GetPatch() *Operation { - if m != nil { - return m.Patch - } - return nil -} - -func (m *PathItem) GetParameters() []*ParametersItem { - if m != nil { - return m.Parameters - } - return nil -} - -func (m *PathItem) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type PathParameterSubSchema struct { - // Determines whether or not this parameter is required or optional. - Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` - // Determines the location of the parameter. - In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` - // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. - Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` - // The name of the parameter. - Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` - Type string `protobuf:"bytes,5,opt,name=type,proto3" json:"type,omitempty"` - Format string `protobuf:"bytes,6,opt,name=format,proto3" json:"format,omitempty"` - Items *PrimitivesItems `protobuf:"bytes,7,opt,name=items,proto3" json:"items,omitempty"` - CollectionFormat string `protobuf:"bytes,8,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` - Default *Any `protobuf:"bytes,9,opt,name=default,proto3" json:"default,omitempty"` - Maximum float64 `protobuf:"fixed64,10,opt,name=maximum,proto3" json:"maximum,omitempty"` - ExclusiveMaximum bool `protobuf:"varint,11,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` - Minimum float64 `protobuf:"fixed64,12,opt,name=minimum,proto3" json:"minimum,omitempty"` - ExclusiveMinimum bool `protobuf:"varint,13,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` - MaxLength int64 `protobuf:"varint,14,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` - MinLength int64 `protobuf:"varint,15,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` - Pattern string `protobuf:"bytes,16,opt,name=pattern,proto3" json:"pattern,omitempty"` - MaxItems int64 `protobuf:"varint,17,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` - MinItems int64 `protobuf:"varint,18,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` - UniqueItems bool `protobuf:"varint,19,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` - Enum []*Any `protobuf:"bytes,20,rep,name=enum,proto3" json:"enum,omitempty"` - MultipleOf float64 `protobuf:"fixed64,21,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,22,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *PathParameterSubSchema) Reset() { *m = PathParameterSubSchema{} } -func (m *PathParameterSubSchema) String() string { return proto.CompactTextString(m) } -func (*PathParameterSubSchema) ProtoMessage() {} -func (*PathParameterSubSchema) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{41} -} - -func (m *PathParameterSubSchema) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_PathParameterSubSchema.Unmarshal(m, b) -} -func (m *PathParameterSubSchema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_PathParameterSubSchema.Marshal(b, m, deterministic) -} -func (m *PathParameterSubSchema) XXX_Merge(src proto.Message) { - xxx_messageInfo_PathParameterSubSchema.Merge(m, src) -} -func (m *PathParameterSubSchema) XXX_Size() int { - return xxx_messageInfo_PathParameterSubSchema.Size(m) -} -func (m *PathParameterSubSchema) XXX_DiscardUnknown() { - xxx_messageInfo_PathParameterSubSchema.DiscardUnknown(m) -} - -var xxx_messageInfo_PathParameterSubSchema proto.InternalMessageInfo - -func (m *PathParameterSubSchema) GetRequired() bool { - if m != nil { - return m.Required - } - return false -} - -func (m *PathParameterSubSchema) GetIn() string { - if m != nil { - return m.In - } - return "" -} - -func (m *PathParameterSubSchema) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *PathParameterSubSchema) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *PathParameterSubSchema) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *PathParameterSubSchema) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *PathParameterSubSchema) GetItems() *PrimitivesItems { - if m != nil { - return m.Items - } - return nil -} - -func (m *PathParameterSubSchema) GetCollectionFormat() string { - if m != nil { - return m.CollectionFormat - } - return "" -} - -func (m *PathParameterSubSchema) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *PathParameterSubSchema) GetMaximum() float64 { - if m != nil { - return m.Maximum - } - return 0 -} - -func (m *PathParameterSubSchema) GetExclusiveMaximum() bool { - if m != nil { - return m.ExclusiveMaximum - } - return false -} - -func (m *PathParameterSubSchema) GetMinimum() float64 { - if m != nil { - return m.Minimum - } - return 0 -} - -func (m *PathParameterSubSchema) GetExclusiveMinimum() bool { - if m != nil { - return m.ExclusiveMinimum - } - return false -} - -func (m *PathParameterSubSchema) GetMaxLength() int64 { - if m != nil { - return m.MaxLength - } - return 0 -} - -func (m *PathParameterSubSchema) GetMinLength() int64 { - if m != nil { - return m.MinLength - } - return 0 -} - -func (m *PathParameterSubSchema) GetPattern() string { - if m != nil { - return m.Pattern - } - return "" -} - -func (m *PathParameterSubSchema) GetMaxItems() int64 { - if m != nil { - return m.MaxItems - } - return 0 -} - -func (m *PathParameterSubSchema) GetMinItems() int64 { - if m != nil { - return m.MinItems - } - return 0 -} - -func (m *PathParameterSubSchema) GetUniqueItems() bool { - if m != nil { - return m.UniqueItems - } - return false -} - -func (m *PathParameterSubSchema) GetEnum() []*Any { - if m != nil { - return m.Enum - } - return nil -} - -func (m *PathParameterSubSchema) GetMultipleOf() float64 { - if m != nil { - return m.MultipleOf - } - return 0 -} - -func (m *PathParameterSubSchema) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -// Relative paths to the individual endpoints. They must be relative to the 'basePath'. -type Paths struct { - VendorExtension []*NamedAny `protobuf:"bytes,1,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - Path []*NamedPathItem `protobuf:"bytes,2,rep,name=path,proto3" json:"path,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Paths) Reset() { *m = Paths{} } -func (m *Paths) String() string { return proto.CompactTextString(m) } -func (*Paths) ProtoMessage() {} -func (*Paths) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{42} -} - -func (m *Paths) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Paths.Unmarshal(m, b) -} -func (m *Paths) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Paths.Marshal(b, m, deterministic) -} -func (m *Paths) XXX_Merge(src proto.Message) { - xxx_messageInfo_Paths.Merge(m, src) -} -func (m *Paths) XXX_Size() int { - return xxx_messageInfo_Paths.Size(m) -} -func (m *Paths) XXX_DiscardUnknown() { - xxx_messageInfo_Paths.DiscardUnknown(m) -} - -var xxx_messageInfo_Paths proto.InternalMessageInfo - -func (m *Paths) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -func (m *Paths) GetPath() []*NamedPathItem { - if m != nil { - return m.Path - } - return nil -} - -type PrimitivesItems struct { - Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` - Format string `protobuf:"bytes,2,opt,name=format,proto3" json:"format,omitempty"` - Items *PrimitivesItems `protobuf:"bytes,3,opt,name=items,proto3" json:"items,omitempty"` - CollectionFormat string `protobuf:"bytes,4,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` - Default *Any `protobuf:"bytes,5,opt,name=default,proto3" json:"default,omitempty"` - Maximum float64 `protobuf:"fixed64,6,opt,name=maximum,proto3" json:"maximum,omitempty"` - ExclusiveMaximum bool `protobuf:"varint,7,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` - Minimum float64 `protobuf:"fixed64,8,opt,name=minimum,proto3" json:"minimum,omitempty"` - ExclusiveMinimum bool `protobuf:"varint,9,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` - MaxLength int64 `protobuf:"varint,10,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` - MinLength int64 `protobuf:"varint,11,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` - Pattern string `protobuf:"bytes,12,opt,name=pattern,proto3" json:"pattern,omitempty"` - MaxItems int64 `protobuf:"varint,13,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` - MinItems int64 `protobuf:"varint,14,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` - UniqueItems bool `protobuf:"varint,15,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` - Enum []*Any `protobuf:"bytes,16,rep,name=enum,proto3" json:"enum,omitempty"` - MultipleOf float64 `protobuf:"fixed64,17,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,18,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *PrimitivesItems) Reset() { *m = PrimitivesItems{} } -func (m *PrimitivesItems) String() string { return proto.CompactTextString(m) } -func (*PrimitivesItems) ProtoMessage() {} -func (*PrimitivesItems) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{43} -} - -func (m *PrimitivesItems) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_PrimitivesItems.Unmarshal(m, b) -} -func (m *PrimitivesItems) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_PrimitivesItems.Marshal(b, m, deterministic) -} -func (m *PrimitivesItems) XXX_Merge(src proto.Message) { - xxx_messageInfo_PrimitivesItems.Merge(m, src) -} -func (m *PrimitivesItems) XXX_Size() int { - return xxx_messageInfo_PrimitivesItems.Size(m) -} -func (m *PrimitivesItems) XXX_DiscardUnknown() { - xxx_messageInfo_PrimitivesItems.DiscardUnknown(m) -} - -var xxx_messageInfo_PrimitivesItems proto.InternalMessageInfo - -func (m *PrimitivesItems) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *PrimitivesItems) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *PrimitivesItems) GetItems() *PrimitivesItems { - if m != nil { - return m.Items - } - return nil -} - -func (m *PrimitivesItems) GetCollectionFormat() string { - if m != nil { - return m.CollectionFormat - } - return "" -} - -func (m *PrimitivesItems) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *PrimitivesItems) GetMaximum() float64 { - if m != nil { - return m.Maximum - } - return 0 -} - -func (m *PrimitivesItems) GetExclusiveMaximum() bool { - if m != nil { - return m.ExclusiveMaximum - } - return false -} - -func (m *PrimitivesItems) GetMinimum() float64 { - if m != nil { - return m.Minimum - } - return 0 -} - -func (m *PrimitivesItems) GetExclusiveMinimum() bool { - if m != nil { - return m.ExclusiveMinimum - } - return false -} - -func (m *PrimitivesItems) GetMaxLength() int64 { - if m != nil { - return m.MaxLength - } - return 0 -} - -func (m *PrimitivesItems) GetMinLength() int64 { - if m != nil { - return m.MinLength - } - return 0 -} - -func (m *PrimitivesItems) GetPattern() string { - if m != nil { - return m.Pattern - } - return "" -} - -func (m *PrimitivesItems) GetMaxItems() int64 { - if m != nil { - return m.MaxItems - } - return 0 -} - -func (m *PrimitivesItems) GetMinItems() int64 { - if m != nil { - return m.MinItems - } - return 0 -} - -func (m *PrimitivesItems) GetUniqueItems() bool { - if m != nil { - return m.UniqueItems - } - return false -} - -func (m *PrimitivesItems) GetEnum() []*Any { - if m != nil { - return m.Enum - } - return nil -} - -func (m *PrimitivesItems) GetMultipleOf() float64 { - if m != nil { - return m.MultipleOf - } - return 0 -} - -func (m *PrimitivesItems) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Properties struct { - AdditionalProperties []*NamedSchema `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Properties) Reset() { *m = Properties{} } -func (m *Properties) String() string { return proto.CompactTextString(m) } -func (*Properties) ProtoMessage() {} -func (*Properties) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{44} -} - -func (m *Properties) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Properties.Unmarshal(m, b) -} -func (m *Properties) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Properties.Marshal(b, m, deterministic) -} -func (m *Properties) XXX_Merge(src proto.Message) { - xxx_messageInfo_Properties.Merge(m, src) -} -func (m *Properties) XXX_Size() int { - return xxx_messageInfo_Properties.Size(m) -} -func (m *Properties) XXX_DiscardUnknown() { - xxx_messageInfo_Properties.DiscardUnknown(m) -} - -var xxx_messageInfo_Properties proto.InternalMessageInfo - -func (m *Properties) GetAdditionalProperties() []*NamedSchema { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type QueryParameterSubSchema struct { - // Determines whether or not this parameter is required or optional. - Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` - // Determines the location of the parameter. - In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` - // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. - Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` - // The name of the parameter. - Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` - // allows sending a parameter by name only or with an empty value. - AllowEmptyValue bool `protobuf:"varint,5,opt,name=allow_empty_value,json=allowEmptyValue,proto3" json:"allow_empty_value,omitempty"` - Type string `protobuf:"bytes,6,opt,name=type,proto3" json:"type,omitempty"` - Format string `protobuf:"bytes,7,opt,name=format,proto3" json:"format,omitempty"` - Items *PrimitivesItems `protobuf:"bytes,8,opt,name=items,proto3" json:"items,omitempty"` - CollectionFormat string `protobuf:"bytes,9,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` - Default *Any `protobuf:"bytes,10,opt,name=default,proto3" json:"default,omitempty"` - Maximum float64 `protobuf:"fixed64,11,opt,name=maximum,proto3" json:"maximum,omitempty"` - ExclusiveMaximum bool `protobuf:"varint,12,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` - Minimum float64 `protobuf:"fixed64,13,opt,name=minimum,proto3" json:"minimum,omitempty"` - ExclusiveMinimum bool `protobuf:"varint,14,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` - MaxLength int64 `protobuf:"varint,15,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` - MinLength int64 `protobuf:"varint,16,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` - Pattern string `protobuf:"bytes,17,opt,name=pattern,proto3" json:"pattern,omitempty"` - MaxItems int64 `protobuf:"varint,18,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` - MinItems int64 `protobuf:"varint,19,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` - UniqueItems bool `protobuf:"varint,20,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` - Enum []*Any `protobuf:"bytes,21,rep,name=enum,proto3" json:"enum,omitempty"` - MultipleOf float64 `protobuf:"fixed64,22,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,23,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *QueryParameterSubSchema) Reset() { *m = QueryParameterSubSchema{} } -func (m *QueryParameterSubSchema) String() string { return proto.CompactTextString(m) } -func (*QueryParameterSubSchema) ProtoMessage() {} -func (*QueryParameterSubSchema) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{45} -} - -func (m *QueryParameterSubSchema) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_QueryParameterSubSchema.Unmarshal(m, b) -} -func (m *QueryParameterSubSchema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_QueryParameterSubSchema.Marshal(b, m, deterministic) -} -func (m *QueryParameterSubSchema) XXX_Merge(src proto.Message) { - xxx_messageInfo_QueryParameterSubSchema.Merge(m, src) -} -func (m *QueryParameterSubSchema) XXX_Size() int { - return xxx_messageInfo_QueryParameterSubSchema.Size(m) -} -func (m *QueryParameterSubSchema) XXX_DiscardUnknown() { - xxx_messageInfo_QueryParameterSubSchema.DiscardUnknown(m) -} - -var xxx_messageInfo_QueryParameterSubSchema proto.InternalMessageInfo - -func (m *QueryParameterSubSchema) GetRequired() bool { - if m != nil { - return m.Required - } - return false -} - -func (m *QueryParameterSubSchema) GetIn() string { - if m != nil { - return m.In - } - return "" -} - -func (m *QueryParameterSubSchema) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *QueryParameterSubSchema) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *QueryParameterSubSchema) GetAllowEmptyValue() bool { - if m != nil { - return m.AllowEmptyValue - } - return false -} - -func (m *QueryParameterSubSchema) GetType() string { - if m != nil { - return m.Type - } - return "" -} - -func (m *QueryParameterSubSchema) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *QueryParameterSubSchema) GetItems() *PrimitivesItems { - if m != nil { - return m.Items - } - return nil -} - -func (m *QueryParameterSubSchema) GetCollectionFormat() string { - if m != nil { - return m.CollectionFormat - } - return "" -} - -func (m *QueryParameterSubSchema) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *QueryParameterSubSchema) GetMaximum() float64 { - if m != nil { - return m.Maximum - } - return 0 -} - -func (m *QueryParameterSubSchema) GetExclusiveMaximum() bool { - if m != nil { - return m.ExclusiveMaximum - } - return false -} - -func (m *QueryParameterSubSchema) GetMinimum() float64 { - if m != nil { - return m.Minimum - } - return 0 -} - -func (m *QueryParameterSubSchema) GetExclusiveMinimum() bool { - if m != nil { - return m.ExclusiveMinimum - } - return false -} - -func (m *QueryParameterSubSchema) GetMaxLength() int64 { - if m != nil { - return m.MaxLength - } - return 0 -} - -func (m *QueryParameterSubSchema) GetMinLength() int64 { - if m != nil { - return m.MinLength - } - return 0 -} - -func (m *QueryParameterSubSchema) GetPattern() string { - if m != nil { - return m.Pattern - } - return "" -} - -func (m *QueryParameterSubSchema) GetMaxItems() int64 { - if m != nil { - return m.MaxItems - } - return 0 -} - -func (m *QueryParameterSubSchema) GetMinItems() int64 { - if m != nil { - return m.MinItems - } - return 0 -} - -func (m *QueryParameterSubSchema) GetUniqueItems() bool { - if m != nil { - return m.UniqueItems - } - return false -} - -func (m *QueryParameterSubSchema) GetEnum() []*Any { - if m != nil { - return m.Enum - } - return nil -} - -func (m *QueryParameterSubSchema) GetMultipleOf() float64 { - if m != nil { - return m.MultipleOf - } - return 0 -} - -func (m *QueryParameterSubSchema) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type Response struct { - Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` - Schema *SchemaItem `protobuf:"bytes,2,opt,name=schema,proto3" json:"schema,omitempty"` - Headers *Headers `protobuf:"bytes,3,opt,name=headers,proto3" json:"headers,omitempty"` - Examples *Examples `protobuf:"bytes,4,opt,name=examples,proto3" json:"examples,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,5,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Response) Reset() { *m = Response{} } -func (m *Response) String() string { return proto.CompactTextString(m) } -func (*Response) ProtoMessage() {} -func (*Response) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{46} -} - -func (m *Response) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Response.Unmarshal(m, b) -} -func (m *Response) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Response.Marshal(b, m, deterministic) -} -func (m *Response) XXX_Merge(src proto.Message) { - xxx_messageInfo_Response.Merge(m, src) -} -func (m *Response) XXX_Size() int { - return xxx_messageInfo_Response.Size(m) -} -func (m *Response) XXX_DiscardUnknown() { - xxx_messageInfo_Response.DiscardUnknown(m) -} - -var xxx_messageInfo_Response proto.InternalMessageInfo - -func (m *Response) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Response) GetSchema() *SchemaItem { - if m != nil { - return m.Schema - } - return nil -} - -func (m *Response) GetHeaders() *Headers { - if m != nil { - return m.Headers - } - return nil -} - -func (m *Response) GetExamples() *Examples { - if m != nil { - return m.Examples - } - return nil -} - -func (m *Response) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -// One or more JSON representations for parameters -type ResponseDefinitions struct { - AdditionalProperties []*NamedResponse `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ResponseDefinitions) Reset() { *m = ResponseDefinitions{} } -func (m *ResponseDefinitions) String() string { return proto.CompactTextString(m) } -func (*ResponseDefinitions) ProtoMessage() {} -func (*ResponseDefinitions) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{47} -} - -func (m *ResponseDefinitions) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ResponseDefinitions.Unmarshal(m, b) -} -func (m *ResponseDefinitions) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ResponseDefinitions.Marshal(b, m, deterministic) -} -func (m *ResponseDefinitions) XXX_Merge(src proto.Message) { - xxx_messageInfo_ResponseDefinitions.Merge(m, src) -} -func (m *ResponseDefinitions) XXX_Size() int { - return xxx_messageInfo_ResponseDefinitions.Size(m) -} -func (m *ResponseDefinitions) XXX_DiscardUnknown() { - xxx_messageInfo_ResponseDefinitions.DiscardUnknown(m) -} - -var xxx_messageInfo_ResponseDefinitions proto.InternalMessageInfo - -func (m *ResponseDefinitions) GetAdditionalProperties() []*NamedResponse { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type ResponseValue struct { - // Types that are valid to be assigned to Oneof: - // *ResponseValue_Response - // *ResponseValue_JsonReference - Oneof isResponseValue_Oneof `protobuf_oneof:"oneof"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *ResponseValue) Reset() { *m = ResponseValue{} } -func (m *ResponseValue) String() string { return proto.CompactTextString(m) } -func (*ResponseValue) ProtoMessage() {} -func (*ResponseValue) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{48} -} - -func (m *ResponseValue) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ResponseValue.Unmarshal(m, b) -} -func (m *ResponseValue) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ResponseValue.Marshal(b, m, deterministic) -} -func (m *ResponseValue) XXX_Merge(src proto.Message) { - xxx_messageInfo_ResponseValue.Merge(m, src) -} -func (m *ResponseValue) XXX_Size() int { - return xxx_messageInfo_ResponseValue.Size(m) -} -func (m *ResponseValue) XXX_DiscardUnknown() { - xxx_messageInfo_ResponseValue.DiscardUnknown(m) -} - -var xxx_messageInfo_ResponseValue proto.InternalMessageInfo - -type isResponseValue_Oneof interface { - isResponseValue_Oneof() -} - -type ResponseValue_Response struct { - Response *Response `protobuf:"bytes,1,opt,name=response,proto3,oneof"` -} - -type ResponseValue_JsonReference struct { - JsonReference *JsonReference `protobuf:"bytes,2,opt,name=json_reference,json=jsonReference,proto3,oneof"` -} - -func (*ResponseValue_Response) isResponseValue_Oneof() {} - -func (*ResponseValue_JsonReference) isResponseValue_Oneof() {} - -func (m *ResponseValue) GetOneof() isResponseValue_Oneof { - if m != nil { - return m.Oneof - } - return nil -} - -func (m *ResponseValue) GetResponse() *Response { - if x, ok := m.GetOneof().(*ResponseValue_Response); ok { - return x.Response - } - return nil -} - -func (m *ResponseValue) GetJsonReference() *JsonReference { - if x, ok := m.GetOneof().(*ResponseValue_JsonReference); ok { - return x.JsonReference - } - return nil -} - -// XXX_OneofWrappers is for the internal use of the proto package. -func (*ResponseValue) XXX_OneofWrappers() []interface{} { - return []interface{}{ - (*ResponseValue_Response)(nil), - (*ResponseValue_JsonReference)(nil), - } -} - -// Response objects names can either be any valid HTTP status code or 'default'. -type Responses struct { - ResponseCode []*NamedResponseValue `protobuf:"bytes,1,rep,name=response_code,json=responseCode,proto3" json:"response_code,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,2,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Responses) Reset() { *m = Responses{} } -func (m *Responses) String() string { return proto.CompactTextString(m) } -func (*Responses) ProtoMessage() {} -func (*Responses) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{49} -} - -func (m *Responses) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Responses.Unmarshal(m, b) -} -func (m *Responses) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Responses.Marshal(b, m, deterministic) -} -func (m *Responses) XXX_Merge(src proto.Message) { - xxx_messageInfo_Responses.Merge(m, src) -} -func (m *Responses) XXX_Size() int { - return xxx_messageInfo_Responses.Size(m) -} -func (m *Responses) XXX_DiscardUnknown() { - xxx_messageInfo_Responses.DiscardUnknown(m) -} - -var xxx_messageInfo_Responses proto.InternalMessageInfo - -func (m *Responses) GetResponseCode() []*NamedResponseValue { - if m != nil { - return m.ResponseCode - } - return nil -} - -func (m *Responses) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -// A deterministic version of a JSON Schema object. -type Schema struct { - XRef string `protobuf:"bytes,1,opt,name=_ref,json=Ref,proto3" json:"_ref,omitempty"` - Format string `protobuf:"bytes,2,opt,name=format,proto3" json:"format,omitempty"` - Title string `protobuf:"bytes,3,opt,name=title,proto3" json:"title,omitempty"` - Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` - Default *Any `protobuf:"bytes,5,opt,name=default,proto3" json:"default,omitempty"` - MultipleOf float64 `protobuf:"fixed64,6,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` - Maximum float64 `protobuf:"fixed64,7,opt,name=maximum,proto3" json:"maximum,omitempty"` - ExclusiveMaximum bool `protobuf:"varint,8,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` - Minimum float64 `protobuf:"fixed64,9,opt,name=minimum,proto3" json:"minimum,omitempty"` - ExclusiveMinimum bool `protobuf:"varint,10,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` - MaxLength int64 `protobuf:"varint,11,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` - MinLength int64 `protobuf:"varint,12,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` - Pattern string `protobuf:"bytes,13,opt,name=pattern,proto3" json:"pattern,omitempty"` - MaxItems int64 `protobuf:"varint,14,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` - MinItems int64 `protobuf:"varint,15,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` - UniqueItems bool `protobuf:"varint,16,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` - MaxProperties int64 `protobuf:"varint,17,opt,name=max_properties,json=maxProperties,proto3" json:"max_properties,omitempty"` - MinProperties int64 `protobuf:"varint,18,opt,name=min_properties,json=minProperties,proto3" json:"min_properties,omitempty"` - Required []string `protobuf:"bytes,19,rep,name=required,proto3" json:"required,omitempty"` - Enum []*Any `protobuf:"bytes,20,rep,name=enum,proto3" json:"enum,omitempty"` - AdditionalProperties *AdditionalPropertiesItem `protobuf:"bytes,21,opt,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - Type *TypeItem `protobuf:"bytes,22,opt,name=type,proto3" json:"type,omitempty"` - Items *ItemsItem `protobuf:"bytes,23,opt,name=items,proto3" json:"items,omitempty"` - AllOf []*Schema `protobuf:"bytes,24,rep,name=all_of,json=allOf,proto3" json:"all_of,omitempty"` - Properties *Properties `protobuf:"bytes,25,opt,name=properties,proto3" json:"properties,omitempty"` - Discriminator string `protobuf:"bytes,26,opt,name=discriminator,proto3" json:"discriminator,omitempty"` - ReadOnly bool `protobuf:"varint,27,opt,name=read_only,json=readOnly,proto3" json:"read_only,omitempty"` - Xml *Xml `protobuf:"bytes,28,opt,name=xml,proto3" json:"xml,omitempty"` - ExternalDocs *ExternalDocs `protobuf:"bytes,29,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` - Example *Any `protobuf:"bytes,30,opt,name=example,proto3" json:"example,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,31,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Schema) Reset() { *m = Schema{} } -func (m *Schema) String() string { return proto.CompactTextString(m) } -func (*Schema) ProtoMessage() {} -func (*Schema) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{50} -} - -func (m *Schema) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Schema.Unmarshal(m, b) -} -func (m *Schema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Schema.Marshal(b, m, deterministic) -} -func (m *Schema) XXX_Merge(src proto.Message) { - xxx_messageInfo_Schema.Merge(m, src) -} -func (m *Schema) XXX_Size() int { - return xxx_messageInfo_Schema.Size(m) -} -func (m *Schema) XXX_DiscardUnknown() { - xxx_messageInfo_Schema.DiscardUnknown(m) -} - -var xxx_messageInfo_Schema proto.InternalMessageInfo - -func (m *Schema) GetXRef() string { - if m != nil { - return m.XRef - } - return "" -} - -func (m *Schema) GetFormat() string { - if m != nil { - return m.Format - } - return "" -} - -func (m *Schema) GetTitle() string { - if m != nil { - return m.Title - } - return "" -} - -func (m *Schema) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Schema) GetDefault() *Any { - if m != nil { - return m.Default - } - return nil -} - -func (m *Schema) GetMultipleOf() float64 { - if m != nil { - return m.MultipleOf - } - return 0 -} - -func (m *Schema) GetMaximum() float64 { - if m != nil { - return m.Maximum - } - return 0 -} - -func (m *Schema) GetExclusiveMaximum() bool { - if m != nil { - return m.ExclusiveMaximum - } - return false -} - -func (m *Schema) GetMinimum() float64 { - if m != nil { - return m.Minimum - } - return 0 -} - -func (m *Schema) GetExclusiveMinimum() bool { - if m != nil { - return m.ExclusiveMinimum - } - return false -} - -func (m *Schema) GetMaxLength() int64 { - if m != nil { - return m.MaxLength - } - return 0 -} - -func (m *Schema) GetMinLength() int64 { - if m != nil { - return m.MinLength - } - return 0 -} - -func (m *Schema) GetPattern() string { - if m != nil { - return m.Pattern - } - return "" -} - -func (m *Schema) GetMaxItems() int64 { - if m != nil { - return m.MaxItems - } - return 0 -} - -func (m *Schema) GetMinItems() int64 { - if m != nil { - return m.MinItems - } - return 0 -} - -func (m *Schema) GetUniqueItems() bool { - if m != nil { - return m.UniqueItems - } - return false -} - -func (m *Schema) GetMaxProperties() int64 { - if m != nil { - return m.MaxProperties - } - return 0 -} - -func (m *Schema) GetMinProperties() int64 { - if m != nil { - return m.MinProperties - } - return 0 -} - -func (m *Schema) GetRequired() []string { - if m != nil { - return m.Required - } - return nil -} - -func (m *Schema) GetEnum() []*Any { - if m != nil { - return m.Enum - } - return nil -} - -func (m *Schema) GetAdditionalProperties() *AdditionalPropertiesItem { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -func (m *Schema) GetType() *TypeItem { - if m != nil { - return m.Type - } - return nil -} - -func (m *Schema) GetItems() *ItemsItem { - if m != nil { - return m.Items - } - return nil -} - -func (m *Schema) GetAllOf() []*Schema { - if m != nil { - return m.AllOf - } - return nil -} - -func (m *Schema) GetProperties() *Properties { - if m != nil { - return m.Properties - } - return nil -} - -func (m *Schema) GetDiscriminator() string { - if m != nil { - return m.Discriminator - } - return "" -} - -func (m *Schema) GetReadOnly() bool { - if m != nil { - return m.ReadOnly - } - return false -} - -func (m *Schema) GetXml() *Xml { - if m != nil { - return m.Xml - } - return nil -} - -func (m *Schema) GetExternalDocs() *ExternalDocs { - if m != nil { - return m.ExternalDocs - } - return nil -} - -func (m *Schema) GetExample() *Any { - if m != nil { - return m.Example - } - return nil -} - -func (m *Schema) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type SchemaItem struct { - // Types that are valid to be assigned to Oneof: - // *SchemaItem_Schema - // *SchemaItem_FileSchema - Oneof isSchemaItem_Oneof `protobuf_oneof:"oneof"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *SchemaItem) Reset() { *m = SchemaItem{} } -func (m *SchemaItem) String() string { return proto.CompactTextString(m) } -func (*SchemaItem) ProtoMessage() {} -func (*SchemaItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{51} -} - -func (m *SchemaItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_SchemaItem.Unmarshal(m, b) -} -func (m *SchemaItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_SchemaItem.Marshal(b, m, deterministic) -} -func (m *SchemaItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_SchemaItem.Merge(m, src) -} -func (m *SchemaItem) XXX_Size() int { - return xxx_messageInfo_SchemaItem.Size(m) -} -func (m *SchemaItem) XXX_DiscardUnknown() { - xxx_messageInfo_SchemaItem.DiscardUnknown(m) -} - -var xxx_messageInfo_SchemaItem proto.InternalMessageInfo - -type isSchemaItem_Oneof interface { - isSchemaItem_Oneof() -} - -type SchemaItem_Schema struct { - Schema *Schema `protobuf:"bytes,1,opt,name=schema,proto3,oneof"` -} - -type SchemaItem_FileSchema struct { - FileSchema *FileSchema `protobuf:"bytes,2,opt,name=file_schema,json=fileSchema,proto3,oneof"` -} - -func (*SchemaItem_Schema) isSchemaItem_Oneof() {} - -func (*SchemaItem_FileSchema) isSchemaItem_Oneof() {} - -func (m *SchemaItem) GetOneof() isSchemaItem_Oneof { - if m != nil { - return m.Oneof - } - return nil -} - -func (m *SchemaItem) GetSchema() *Schema { - if x, ok := m.GetOneof().(*SchemaItem_Schema); ok { - return x.Schema - } - return nil -} - -func (m *SchemaItem) GetFileSchema() *FileSchema { - if x, ok := m.GetOneof().(*SchemaItem_FileSchema); ok { - return x.FileSchema - } - return nil -} - -// XXX_OneofWrappers is for the internal use of the proto package. -func (*SchemaItem) XXX_OneofWrappers() []interface{} { - return []interface{}{ - (*SchemaItem_Schema)(nil), - (*SchemaItem_FileSchema)(nil), - } -} - -type SecurityDefinitions struct { - AdditionalProperties []*NamedSecurityDefinitionsItem `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *SecurityDefinitions) Reset() { *m = SecurityDefinitions{} } -func (m *SecurityDefinitions) String() string { return proto.CompactTextString(m) } -func (*SecurityDefinitions) ProtoMessage() {} -func (*SecurityDefinitions) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{52} -} - -func (m *SecurityDefinitions) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_SecurityDefinitions.Unmarshal(m, b) -} -func (m *SecurityDefinitions) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_SecurityDefinitions.Marshal(b, m, deterministic) -} -func (m *SecurityDefinitions) XXX_Merge(src proto.Message) { - xxx_messageInfo_SecurityDefinitions.Merge(m, src) -} -func (m *SecurityDefinitions) XXX_Size() int { - return xxx_messageInfo_SecurityDefinitions.Size(m) -} -func (m *SecurityDefinitions) XXX_DiscardUnknown() { - xxx_messageInfo_SecurityDefinitions.DiscardUnknown(m) -} - -var xxx_messageInfo_SecurityDefinitions proto.InternalMessageInfo - -func (m *SecurityDefinitions) GetAdditionalProperties() []*NamedSecurityDefinitionsItem { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type SecurityDefinitionsItem struct { - // Types that are valid to be assigned to Oneof: - // *SecurityDefinitionsItem_BasicAuthenticationSecurity - // *SecurityDefinitionsItem_ApiKeySecurity - // *SecurityDefinitionsItem_Oauth2ImplicitSecurity - // *SecurityDefinitionsItem_Oauth2PasswordSecurity - // *SecurityDefinitionsItem_Oauth2ApplicationSecurity - // *SecurityDefinitionsItem_Oauth2AccessCodeSecurity - Oneof isSecurityDefinitionsItem_Oneof `protobuf_oneof:"oneof"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *SecurityDefinitionsItem) Reset() { *m = SecurityDefinitionsItem{} } -func (m *SecurityDefinitionsItem) String() string { return proto.CompactTextString(m) } -func (*SecurityDefinitionsItem) ProtoMessage() {} -func (*SecurityDefinitionsItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{53} -} - -func (m *SecurityDefinitionsItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_SecurityDefinitionsItem.Unmarshal(m, b) -} -func (m *SecurityDefinitionsItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_SecurityDefinitionsItem.Marshal(b, m, deterministic) -} -func (m *SecurityDefinitionsItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_SecurityDefinitionsItem.Merge(m, src) -} -func (m *SecurityDefinitionsItem) XXX_Size() int { - return xxx_messageInfo_SecurityDefinitionsItem.Size(m) -} -func (m *SecurityDefinitionsItem) XXX_DiscardUnknown() { - xxx_messageInfo_SecurityDefinitionsItem.DiscardUnknown(m) -} - -var xxx_messageInfo_SecurityDefinitionsItem proto.InternalMessageInfo - -type isSecurityDefinitionsItem_Oneof interface { - isSecurityDefinitionsItem_Oneof() -} - -type SecurityDefinitionsItem_BasicAuthenticationSecurity struct { - BasicAuthenticationSecurity *BasicAuthenticationSecurity `protobuf:"bytes,1,opt,name=basic_authentication_security,json=basicAuthenticationSecurity,proto3,oneof"` -} - -type SecurityDefinitionsItem_ApiKeySecurity struct { - ApiKeySecurity *ApiKeySecurity `protobuf:"bytes,2,opt,name=api_key_security,json=apiKeySecurity,proto3,oneof"` -} - -type SecurityDefinitionsItem_Oauth2ImplicitSecurity struct { - Oauth2ImplicitSecurity *Oauth2ImplicitSecurity `protobuf:"bytes,3,opt,name=oauth2_implicit_security,json=oauth2ImplicitSecurity,proto3,oneof"` -} - -type SecurityDefinitionsItem_Oauth2PasswordSecurity struct { - Oauth2PasswordSecurity *Oauth2PasswordSecurity `protobuf:"bytes,4,opt,name=oauth2_password_security,json=oauth2PasswordSecurity,proto3,oneof"` -} - -type SecurityDefinitionsItem_Oauth2ApplicationSecurity struct { - Oauth2ApplicationSecurity *Oauth2ApplicationSecurity `protobuf:"bytes,5,opt,name=oauth2_application_security,json=oauth2ApplicationSecurity,proto3,oneof"` -} - -type SecurityDefinitionsItem_Oauth2AccessCodeSecurity struct { - Oauth2AccessCodeSecurity *Oauth2AccessCodeSecurity `protobuf:"bytes,6,opt,name=oauth2_access_code_security,json=oauth2AccessCodeSecurity,proto3,oneof"` -} - -func (*SecurityDefinitionsItem_BasicAuthenticationSecurity) isSecurityDefinitionsItem_Oneof() {} - -func (*SecurityDefinitionsItem_ApiKeySecurity) isSecurityDefinitionsItem_Oneof() {} - -func (*SecurityDefinitionsItem_Oauth2ImplicitSecurity) isSecurityDefinitionsItem_Oneof() {} - -func (*SecurityDefinitionsItem_Oauth2PasswordSecurity) isSecurityDefinitionsItem_Oneof() {} - -func (*SecurityDefinitionsItem_Oauth2ApplicationSecurity) isSecurityDefinitionsItem_Oneof() {} - -func (*SecurityDefinitionsItem_Oauth2AccessCodeSecurity) isSecurityDefinitionsItem_Oneof() {} - -func (m *SecurityDefinitionsItem) GetOneof() isSecurityDefinitionsItem_Oneof { - if m != nil { - return m.Oneof - } - return nil -} - -func (m *SecurityDefinitionsItem) GetBasicAuthenticationSecurity() *BasicAuthenticationSecurity { - if x, ok := m.GetOneof().(*SecurityDefinitionsItem_BasicAuthenticationSecurity); ok { - return x.BasicAuthenticationSecurity - } - return nil -} - -func (m *SecurityDefinitionsItem) GetApiKeySecurity() *ApiKeySecurity { - if x, ok := m.GetOneof().(*SecurityDefinitionsItem_ApiKeySecurity); ok { - return x.ApiKeySecurity - } - return nil -} - -func (m *SecurityDefinitionsItem) GetOauth2ImplicitSecurity() *Oauth2ImplicitSecurity { - if x, ok := m.GetOneof().(*SecurityDefinitionsItem_Oauth2ImplicitSecurity); ok { - return x.Oauth2ImplicitSecurity - } - return nil -} - -func (m *SecurityDefinitionsItem) GetOauth2PasswordSecurity() *Oauth2PasswordSecurity { - if x, ok := m.GetOneof().(*SecurityDefinitionsItem_Oauth2PasswordSecurity); ok { - return x.Oauth2PasswordSecurity - } - return nil -} - -func (m *SecurityDefinitionsItem) GetOauth2ApplicationSecurity() *Oauth2ApplicationSecurity { - if x, ok := m.GetOneof().(*SecurityDefinitionsItem_Oauth2ApplicationSecurity); ok { - return x.Oauth2ApplicationSecurity - } - return nil -} - -func (m *SecurityDefinitionsItem) GetOauth2AccessCodeSecurity() *Oauth2AccessCodeSecurity { - if x, ok := m.GetOneof().(*SecurityDefinitionsItem_Oauth2AccessCodeSecurity); ok { - return x.Oauth2AccessCodeSecurity - } - return nil -} - -// XXX_OneofWrappers is for the internal use of the proto package. -func (*SecurityDefinitionsItem) XXX_OneofWrappers() []interface{} { - return []interface{}{ - (*SecurityDefinitionsItem_BasicAuthenticationSecurity)(nil), - (*SecurityDefinitionsItem_ApiKeySecurity)(nil), - (*SecurityDefinitionsItem_Oauth2ImplicitSecurity)(nil), - (*SecurityDefinitionsItem_Oauth2PasswordSecurity)(nil), - (*SecurityDefinitionsItem_Oauth2ApplicationSecurity)(nil), - (*SecurityDefinitionsItem_Oauth2AccessCodeSecurity)(nil), - } -} - -type SecurityRequirement struct { - AdditionalProperties []*NamedStringArray `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *SecurityRequirement) Reset() { *m = SecurityRequirement{} } -func (m *SecurityRequirement) String() string { return proto.CompactTextString(m) } -func (*SecurityRequirement) ProtoMessage() {} -func (*SecurityRequirement) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{54} -} - -func (m *SecurityRequirement) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_SecurityRequirement.Unmarshal(m, b) -} -func (m *SecurityRequirement) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_SecurityRequirement.Marshal(b, m, deterministic) -} -func (m *SecurityRequirement) XXX_Merge(src proto.Message) { - xxx_messageInfo_SecurityRequirement.Merge(m, src) -} -func (m *SecurityRequirement) XXX_Size() int { - return xxx_messageInfo_SecurityRequirement.Size(m) -} -func (m *SecurityRequirement) XXX_DiscardUnknown() { - xxx_messageInfo_SecurityRequirement.DiscardUnknown(m) -} - -var xxx_messageInfo_SecurityRequirement proto.InternalMessageInfo - -func (m *SecurityRequirement) GetAdditionalProperties() []*NamedStringArray { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type StringArray struct { - Value []string `protobuf:"bytes,1,rep,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *StringArray) Reset() { *m = StringArray{} } -func (m *StringArray) String() string { return proto.CompactTextString(m) } -func (*StringArray) ProtoMessage() {} -func (*StringArray) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{55} -} - -func (m *StringArray) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_StringArray.Unmarshal(m, b) -} -func (m *StringArray) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_StringArray.Marshal(b, m, deterministic) -} -func (m *StringArray) XXX_Merge(src proto.Message) { - xxx_messageInfo_StringArray.Merge(m, src) -} -func (m *StringArray) XXX_Size() int { - return xxx_messageInfo_StringArray.Size(m) -} -func (m *StringArray) XXX_DiscardUnknown() { - xxx_messageInfo_StringArray.DiscardUnknown(m) -} - -var xxx_messageInfo_StringArray proto.InternalMessageInfo - -func (m *StringArray) GetValue() []string { - if m != nil { - return m.Value - } - return nil -} - -type Tag struct { - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` - ExternalDocs *ExternalDocs `protobuf:"bytes,3,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,4,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Tag) Reset() { *m = Tag{} } -func (m *Tag) String() string { return proto.CompactTextString(m) } -func (*Tag) ProtoMessage() {} -func (*Tag) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{56} -} - -func (m *Tag) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Tag.Unmarshal(m, b) -} -func (m *Tag) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Tag.Marshal(b, m, deterministic) -} -func (m *Tag) XXX_Merge(src proto.Message) { - xxx_messageInfo_Tag.Merge(m, src) -} -func (m *Tag) XXX_Size() int { - return xxx_messageInfo_Tag.Size(m) -} -func (m *Tag) XXX_DiscardUnknown() { - xxx_messageInfo_Tag.DiscardUnknown(m) -} - -var xxx_messageInfo_Tag proto.InternalMessageInfo - -func (m *Tag) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *Tag) GetDescription() string { - if m != nil { - return m.Description - } - return "" -} - -func (m *Tag) GetExternalDocs() *ExternalDocs { - if m != nil { - return m.ExternalDocs - } - return nil -} - -func (m *Tag) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -type TypeItem struct { - Value []string `protobuf:"bytes,1,rep,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *TypeItem) Reset() { *m = TypeItem{} } -func (m *TypeItem) String() string { return proto.CompactTextString(m) } -func (*TypeItem) ProtoMessage() {} -func (*TypeItem) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{57} -} - -func (m *TypeItem) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_TypeItem.Unmarshal(m, b) -} -func (m *TypeItem) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_TypeItem.Marshal(b, m, deterministic) -} -func (m *TypeItem) XXX_Merge(src proto.Message) { - xxx_messageInfo_TypeItem.Merge(m, src) -} -func (m *TypeItem) XXX_Size() int { - return xxx_messageInfo_TypeItem.Size(m) -} -func (m *TypeItem) XXX_DiscardUnknown() { - xxx_messageInfo_TypeItem.DiscardUnknown(m) -} - -var xxx_messageInfo_TypeItem proto.InternalMessageInfo - -func (m *TypeItem) GetValue() []string { - if m != nil { - return m.Value - } - return nil -} - -// Any property starting with x- is valid. -type VendorExtension struct { - AdditionalProperties []*NamedAny `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *VendorExtension) Reset() { *m = VendorExtension{} } -func (m *VendorExtension) String() string { return proto.CompactTextString(m) } -func (*VendorExtension) ProtoMessage() {} -func (*VendorExtension) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{58} -} - -func (m *VendorExtension) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_VendorExtension.Unmarshal(m, b) -} -func (m *VendorExtension) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_VendorExtension.Marshal(b, m, deterministic) -} -func (m *VendorExtension) XXX_Merge(src proto.Message) { - xxx_messageInfo_VendorExtension.Merge(m, src) -} -func (m *VendorExtension) XXX_Size() int { - return xxx_messageInfo_VendorExtension.Size(m) -} -func (m *VendorExtension) XXX_DiscardUnknown() { - xxx_messageInfo_VendorExtension.DiscardUnknown(m) -} - -var xxx_messageInfo_VendorExtension proto.InternalMessageInfo - -func (m *VendorExtension) GetAdditionalProperties() []*NamedAny { - if m != nil { - return m.AdditionalProperties - } - return nil -} - -type Xml struct { - Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` - Namespace string `protobuf:"bytes,2,opt,name=namespace,proto3" json:"namespace,omitempty"` - Prefix string `protobuf:"bytes,3,opt,name=prefix,proto3" json:"prefix,omitempty"` - Attribute bool `protobuf:"varint,4,opt,name=attribute,proto3" json:"attribute,omitempty"` - Wrapped bool `protobuf:"varint,5,opt,name=wrapped,proto3" json:"wrapped,omitempty"` - VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Xml) Reset() { *m = Xml{} } -func (m *Xml) String() string { return proto.CompactTextString(m) } -func (*Xml) ProtoMessage() {} -func (*Xml) Descriptor() ([]byte, []int) { - return fileDescriptor_336adc04ae589d92, []int{59} -} - -func (m *Xml) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Xml.Unmarshal(m, b) -} -func (m *Xml) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Xml.Marshal(b, m, deterministic) -} -func (m *Xml) XXX_Merge(src proto.Message) { - xxx_messageInfo_Xml.Merge(m, src) -} -func (m *Xml) XXX_Size() int { - return xxx_messageInfo_Xml.Size(m) -} -func (m *Xml) XXX_DiscardUnknown() { - xxx_messageInfo_Xml.DiscardUnknown(m) -} - -var xxx_messageInfo_Xml proto.InternalMessageInfo - -func (m *Xml) GetName() string { - if m != nil { - return m.Name - } - return "" -} - -func (m *Xml) GetNamespace() string { - if m != nil { - return m.Namespace - } - return "" -} - -func (m *Xml) GetPrefix() string { - if m != nil { - return m.Prefix - } - return "" -} - -func (m *Xml) GetAttribute() bool { - if m != nil { - return m.Attribute - } - return false -} - -func (m *Xml) GetWrapped() bool { - if m != nil { - return m.Wrapped - } - return false -} - -func (m *Xml) GetVendorExtension() []*NamedAny { - if m != nil { - return m.VendorExtension - } - return nil -} - -func init() { - proto.RegisterType((*AdditionalPropertiesItem)(nil), "openapi.v2.AdditionalPropertiesItem") - proto.RegisterType((*Any)(nil), "openapi.v2.Any") - proto.RegisterType((*ApiKeySecurity)(nil), "openapi.v2.ApiKeySecurity") - proto.RegisterType((*BasicAuthenticationSecurity)(nil), "openapi.v2.BasicAuthenticationSecurity") - proto.RegisterType((*BodyParameter)(nil), "openapi.v2.BodyParameter") - proto.RegisterType((*Contact)(nil), "openapi.v2.Contact") - proto.RegisterType((*Default)(nil), "openapi.v2.Default") - proto.RegisterType((*Definitions)(nil), "openapi.v2.Definitions") - proto.RegisterType((*Document)(nil), "openapi.v2.Document") - proto.RegisterType((*Examples)(nil), "openapi.v2.Examples") - proto.RegisterType((*ExternalDocs)(nil), "openapi.v2.ExternalDocs") - proto.RegisterType((*FileSchema)(nil), "openapi.v2.FileSchema") - proto.RegisterType((*FormDataParameterSubSchema)(nil), "openapi.v2.FormDataParameterSubSchema") - proto.RegisterType((*Header)(nil), "openapi.v2.Header") - proto.RegisterType((*HeaderParameterSubSchema)(nil), "openapi.v2.HeaderParameterSubSchema") - proto.RegisterType((*Headers)(nil), "openapi.v2.Headers") - proto.RegisterType((*Info)(nil), "openapi.v2.Info") - proto.RegisterType((*ItemsItem)(nil), "openapi.v2.ItemsItem") - proto.RegisterType((*JsonReference)(nil), "openapi.v2.JsonReference") - proto.RegisterType((*License)(nil), "openapi.v2.License") - proto.RegisterType((*NamedAny)(nil), "openapi.v2.NamedAny") - proto.RegisterType((*NamedHeader)(nil), "openapi.v2.NamedHeader") - proto.RegisterType((*NamedParameter)(nil), "openapi.v2.NamedParameter") - proto.RegisterType((*NamedPathItem)(nil), "openapi.v2.NamedPathItem") - proto.RegisterType((*NamedResponse)(nil), "openapi.v2.NamedResponse") - proto.RegisterType((*NamedResponseValue)(nil), "openapi.v2.NamedResponseValue") - proto.RegisterType((*NamedSchema)(nil), "openapi.v2.NamedSchema") - proto.RegisterType((*NamedSecurityDefinitionsItem)(nil), "openapi.v2.NamedSecurityDefinitionsItem") - proto.RegisterType((*NamedString)(nil), "openapi.v2.NamedString") - proto.RegisterType((*NamedStringArray)(nil), "openapi.v2.NamedStringArray") - proto.RegisterType((*NonBodyParameter)(nil), "openapi.v2.NonBodyParameter") - proto.RegisterType((*Oauth2AccessCodeSecurity)(nil), "openapi.v2.Oauth2AccessCodeSecurity") - proto.RegisterType((*Oauth2ApplicationSecurity)(nil), "openapi.v2.Oauth2ApplicationSecurity") - proto.RegisterType((*Oauth2ImplicitSecurity)(nil), "openapi.v2.Oauth2ImplicitSecurity") - proto.RegisterType((*Oauth2PasswordSecurity)(nil), "openapi.v2.Oauth2PasswordSecurity") - proto.RegisterType((*Oauth2Scopes)(nil), "openapi.v2.Oauth2Scopes") - proto.RegisterType((*Operation)(nil), "openapi.v2.Operation") - proto.RegisterType((*Parameter)(nil), "openapi.v2.Parameter") - proto.RegisterType((*ParameterDefinitions)(nil), "openapi.v2.ParameterDefinitions") - proto.RegisterType((*ParametersItem)(nil), "openapi.v2.ParametersItem") - proto.RegisterType((*PathItem)(nil), "openapi.v2.PathItem") - proto.RegisterType((*PathParameterSubSchema)(nil), "openapi.v2.PathParameterSubSchema") - proto.RegisterType((*Paths)(nil), "openapi.v2.Paths") - proto.RegisterType((*PrimitivesItems)(nil), "openapi.v2.PrimitivesItems") - proto.RegisterType((*Properties)(nil), "openapi.v2.Properties") - proto.RegisterType((*QueryParameterSubSchema)(nil), "openapi.v2.QueryParameterSubSchema") - proto.RegisterType((*Response)(nil), "openapi.v2.Response") - proto.RegisterType((*ResponseDefinitions)(nil), "openapi.v2.ResponseDefinitions") - proto.RegisterType((*ResponseValue)(nil), "openapi.v2.ResponseValue") - proto.RegisterType((*Responses)(nil), "openapi.v2.Responses") - proto.RegisterType((*Schema)(nil), "openapi.v2.Schema") - proto.RegisterType((*SchemaItem)(nil), "openapi.v2.SchemaItem") - proto.RegisterType((*SecurityDefinitions)(nil), "openapi.v2.SecurityDefinitions") - proto.RegisterType((*SecurityDefinitionsItem)(nil), "openapi.v2.SecurityDefinitionsItem") - proto.RegisterType((*SecurityRequirement)(nil), "openapi.v2.SecurityRequirement") - proto.RegisterType((*StringArray)(nil), "openapi.v2.StringArray") - proto.RegisterType((*Tag)(nil), "openapi.v2.Tag") - proto.RegisterType((*TypeItem)(nil), "openapi.v2.TypeItem") - proto.RegisterType((*VendorExtension)(nil), "openapi.v2.VendorExtension") - proto.RegisterType((*Xml)(nil), "openapi.v2.Xml") -} - -func init() { proto.RegisterFile("OpenAPIv2/OpenAPIv2.proto", fileDescriptor_336adc04ae589d92) } - -var fileDescriptor_336adc04ae589d92 = []byte{ - // 3108 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x09, 0x6e, 0x88, 0x02, 0xff, 0xec, 0x1b, 0x4d, 0x73, 0x1c, 0x57, - 0x31, 0xab, 0xfd, 0xee, 0xd5, 0xae, 0x56, 0x23, 0x59, 0x5e, 0x49, 0x8e, 0xe3, 0x28, 0x5f, 0x8e, - 0x43, 0xe4, 0xa0, 0x14, 0xa4, 0x02, 0x45, 0x81, 0x1c, 0xdb, 0x15, 0x13, 0x13, 0x29, 0x23, 0x27, - 0x10, 0x08, 0x4c, 0x8d, 0x76, 0xdf, 0x4a, 0x93, 0xec, 0xce, 0xac, 0x67, 0x66, 0x65, 0x89, 0x03, - 0x07, 0xa8, 0xe2, 0x0c, 0x54, 0xce, 0x54, 0xc1, 0x81, 0xa2, 0x2a, 0x07, 0x4e, 0x9c, 0xf8, 0x03, - 0xdc, 0xf8, 0x07, 0x9c, 0xe1, 0x4a, 0x15, 0x27, 0x8a, 0x8f, 0x7e, 0x5f, 0xf3, 0xf9, 0x66, 0x76, - 0xc7, 0x72, 0x01, 0x05, 0x3a, 0xed, 0xce, 0xeb, 0x7e, 0xfd, 0xfa, 0xf5, 0x74, 0xf7, 0xeb, 0x8f, - 0x37, 0xb0, 0xbe, 0x37, 0x21, 0xf6, 0xee, 0xfe, 0xbd, 0x93, 0x9d, 0x9b, 0xc1, 0xbf, 0xed, 0x89, - 0xeb, 0xf8, 0x8e, 0x06, 0x0e, 0x0e, 0x98, 0x13, 0x6b, 0xfb, 0x64, 0x67, 0x63, 0xfd, 0xc8, 0x71, - 0x8e, 0x46, 0xe4, 0x26, 0x83, 0x1c, 0x4e, 0x87, 0x37, 0x4d, 0xfb, 0x8c, 0xa3, 0x6d, 0x8d, 0xa1, - 0xb7, 0x3b, 0x18, 0x58, 0xbe, 0xe5, 0xd8, 0xe6, 0x68, 0xdf, 0xc5, 0x49, 0xae, 0x6f, 0x11, 0xef, - 0x9e, 0x4f, 0xc6, 0xda, 0xe7, 0xa0, 0xe6, 0xf5, 0x8f, 0xc9, 0xd8, 0xec, 0x95, 0xae, 0x95, 0xae, - 0xb7, 0x76, 0xb4, 0xed, 0x90, 0xe6, 0xf6, 0x01, 0x83, 0xbc, 0xfd, 0x94, 0x2e, 0x70, 0xb4, 0x0d, - 0xa8, 0x1f, 0x3a, 0xce, 0x88, 0x98, 0x76, 0x6f, 0x01, 0xd1, 0x1b, 0x08, 0x92, 0x03, 0xb7, 0xea, - 0x50, 0x75, 0x6c, 0xe2, 0x0c, 0xb7, 0xee, 0x40, 0x79, 0xd7, 0x3e, 0xd3, 0x6e, 0x40, 0xf5, 0xc4, - 0x1c, 0x4d, 0x89, 0x20, 0xbc, 0xba, 0xcd, 0x19, 0xdc, 0x96, 0x0c, 0x6e, 0x23, 0x92, 0xce, 0x51, - 0x34, 0x0d, 0x2a, 0x67, 0xe6, 0x78, 0xc4, 0x88, 0x36, 0x75, 0xf6, 0x7f, 0xeb, 0xb3, 0x12, 0x74, - 0x76, 0x27, 0xd6, 0x3b, 0xe4, 0xec, 0x80, 0xf4, 0xa7, 0xae, 0xe5, 0x9f, 0x51, 0x34, 0xff, 0x6c, - 0xc2, 0x29, 0x22, 0x1a, 0xfd, 0x4f, 0xc7, 0x6c, 0x73, 0x4c, 0xe4, 0x54, 0xfa, 0x5f, 0xeb, 0xc0, - 0x82, 0x65, 0xf7, 0xca, 0x6c, 0x04, 0xff, 0x69, 0xd7, 0xa0, 0x35, 0x20, 0x5e, 0xdf, 0xb5, 0x26, - 0x54, 0x06, 0xbd, 0x0a, 0x03, 0x44, 0x87, 0xb4, 0xaf, 0x42, 0xf7, 0x84, 0xd8, 0x03, 0xc7, 0x35, - 0xc8, 0xa9, 0x4f, 0x6c, 0x8f, 0xa2, 0x55, 0xaf, 0x95, 0x19, 0xdf, 0x11, 0x81, 0xbc, 0x8b, 0xd4, - 0x07, 0x94, 0xef, 0x25, 0x8e, 0x7d, 0x47, 0x22, 0x6f, 0x7d, 0x5a, 0x82, 0xcd, 0x5b, 0xa6, 0x67, - 0xf5, 0x77, 0xa7, 0xfe, 0x31, 0xb1, 0x7d, 0xab, 0x6f, 0x52, 0xc2, 0xb9, 0xac, 0x27, 0xd8, 0x5a, - 0x98, 0x8f, 0xad, 0x72, 0x11, 0xb6, 0xfe, 0x58, 0x82, 0xf6, 0x2d, 0x67, 0x70, 0xb6, 0x6f, 0xba, - 0x88, 0xe3, 0x13, 0x37, 0xb9, 0x68, 0x29, 0xbd, 0xe8, 0x3c, 0x12, 0xdd, 0x80, 0x86, 0x4b, 0x1e, - 0x4e, 0x2d, 0x97, 0x0c, 0x98, 0x38, 0x1b, 0x7a, 0xf0, 0x8c, 0x2f, 0x5e, 0xaa, 0x54, 0x35, 0x4b, - 0xa5, 0x02, 0x85, 0x52, 0x6d, 0xb0, 0x56, 0x64, 0x83, 0x3f, 0x2e, 0x41, 0xfd, 0x2d, 0xc7, 0xf6, - 0xcd, 0xbe, 0x1f, 0x30, 0x5e, 0x8a, 0x30, 0xde, 0x85, 0xf2, 0xd4, 0x95, 0x8a, 0x45, 0xff, 0x6a, - 0xab, 0x50, 0xc5, 0x95, 0xad, 0x91, 0xd8, 0x0d, 0x7f, 0x50, 0x32, 0x52, 0x29, 0xc2, 0xc8, 0x03, - 0xa8, 0xdf, 0x26, 0x43, 0x73, 0x3a, 0xf2, 0xb5, 0x7b, 0x70, 0xc9, 0x0c, 0xec, 0xcd, 0x98, 0x04, - 0x06, 0x87, 0x8c, 0x65, 0x13, 0x5c, 0x35, 0x15, 0x26, 0xba, 0xf5, 0x1d, 0x68, 0x21, 0x55, 0xcb, - 0x66, 0x10, 0x4f, 0xbb, 0x9f, 0x4f, 0xf9, 0x72, 0x8a, 0xb2, 0x10, 0xb7, 0x9a, 0xf8, 0x9f, 0xaa, - 0xd0, 0xb8, 0xed, 0xf4, 0xa7, 0x63, 0xd4, 0x57, 0xad, 0x07, 0x75, 0xef, 0x91, 0x79, 0x74, 0x44, - 0x5c, 0x21, 0x3f, 0xf9, 0xa8, 0x3d, 0x0f, 0x15, 0xcb, 0x1e, 0x3a, 0x4c, 0x86, 0xad, 0x9d, 0x6e, - 0x74, 0x8d, 0x7b, 0x38, 0xae, 0x33, 0x28, 0x15, 0xfe, 0xb1, 0xe3, 0xf9, 0x42, 0xaa, 0xec, 0xbf, - 0xb6, 0x09, 0xcd, 0x43, 0xd3, 0x23, 0xc6, 0xc4, 0xf4, 0x8f, 0x85, 0xd5, 0x35, 0xe8, 0xc0, 0x3e, - 0x3e, 0xb3, 0x05, 0x29, 0x77, 0xc8, 0x3d, 0xb5, 0x34, 0xba, 0x20, 0x7f, 0xa4, 0xca, 0xd5, 0xc7, - 0xdd, 0x4e, 0x29, 0xa8, 0xc6, 0x40, 0xc1, 0x33, 0x85, 0xe1, 0xb6, 0x07, 0xd3, 0x3e, 0xc2, 0xea, - 0x1c, 0x26, 0x9f, 0xb5, 0x97, 0xa0, 0x4a, 0x57, 0xf2, 0x7a, 0x0d, 0xc6, 0xe9, 0x72, 0x94, 0x53, - 0xba, 0xa4, 0xa7, 0x73, 0xb8, 0xf6, 0x26, 0xb5, 0x81, 0x40, 0xaa, 0xbd, 0x26, 0x43, 0x8f, 0x09, - 0x2f, 0x22, 0x74, 0x3d, 0x8a, 0xab, 0x7d, 0x0d, 0x60, 0x22, 0x6d, 0xc9, 0xeb, 0x01, 0x9b, 0x79, - 0x2d, 0xbe, 0x90, 0x80, 0x46, 0x49, 0x44, 0xe6, 0x68, 0x5f, 0x81, 0xa6, 0x4b, 0xbc, 0x09, 0x0e, - 0xe3, 0x16, 0x5a, 0x8c, 0xc0, 0x33, 0x51, 0x02, 0xba, 0x00, 0x46, 0xe7, 0x87, 0x33, 0xb4, 0x2f, - 0x43, 0xc3, 0x13, 0x4e, 0xa5, 0xb7, 0xc8, 0xde, 0x7a, 0x6c, 0xb6, 0x74, 0x38, 0x3a, 0xb7, 0x46, - 0xfa, 0x6a, 0xf5, 0x60, 0x82, 0xa6, 0xc3, 0xaa, 0xfc, 0x6f, 0x44, 0x25, 0xd0, 0x4e, 0xb3, 0x21, - 0x09, 0x45, 0xd9, 0x58, 0xf1, 0xd2, 0x83, 0xda, 0x73, 0xe8, 0xd9, 0xcc, 0x23, 0xaf, 0xd7, 0x61, - 0xcc, 0x2c, 0x45, 0x69, 0x3c, 0x30, 0x8f, 0x74, 0x06, 0xc4, 0x4d, 0xb7, 0xa9, 0x5d, 0xb9, 0x54, - 0x6d, 0x07, 0x4e, 0xdf, 0xeb, 0x2d, 0xb1, 0x15, 0x7b, 0x51, 0xec, 0x3b, 0x02, 0x01, 0x55, 0xd2, - 0xd3, 0x17, 0x49, 0xe4, 0x49, 0x69, 0x9d, 0xdd, 0x22, 0xd6, 0xf9, 0x3e, 0x34, 0xee, 0x9c, 0x9a, - 0xe3, 0xc9, 0x08, 0x25, 0xf8, 0x04, 0xcd, 0xf3, 0x47, 0x25, 0x58, 0x8c, 0xb2, 0x3d, 0x87, 0x77, - 0x4d, 0x3b, 0xa4, 0x73, 0x3b, 0xf9, 0x7f, 0x2e, 0x00, 0xdc, 0xb5, 0x46, 0x84, 0x1b, 0xbb, 0xb6, - 0x06, 0xb5, 0xa1, 0xe3, 0x8e, 0x4d, 0x5f, 0x2c, 0x2f, 0x9e, 0xa8, 0xe3, 0xf3, 0x2d, 0x7f, 0x24, - 0x1d, 0x3b, 0x7f, 0x48, 0x72, 0x5c, 0x4e, 0x73, 0xfc, 0x32, 0xd4, 0x07, 0xdc, 0xb3, 0x31, 0x1b, - 0x4e, 0xbc, 0x63, 0xca, 0x91, 0x84, 0xc7, 0x8e, 0x05, 0x6e, 0xd4, 0xe1, 0xb1, 0x20, 0x4f, 0xc0, - 0x5a, 0xe4, 0x04, 0xdc, 0xa4, 0xb6, 0x60, 0x0e, 0x0c, 0xc7, 0x1e, 0x9d, 0xa1, 0x39, 0x8b, 0x73, - 0xc4, 0x1c, 0xec, 0xe1, 0x73, 0x5a, 0x67, 0x1a, 0x85, 0x74, 0x06, 0xd9, 0x26, 0xfc, 0x95, 0x0b, - 0x03, 0x4f, 0xb3, 0x2d, 0xe0, 0xca, 0x37, 0x00, 0x45, 0xde, 0xc0, 0x67, 0x35, 0xd8, 0xb8, 0x8b, - 0x52, 0xbe, 0x6d, 0xfa, 0x66, 0xe0, 0x00, 0x0e, 0xa6, 0x87, 0x07, 0x32, 0x6c, 0x0a, 0xc5, 0x52, - 0x4a, 0x9c, 0x96, 0xfc, 0x64, 0x5d, 0xc8, 0x8a, 0x55, 0xca, 0xd9, 0xe7, 0x73, 0x25, 0x72, 0xcc, - 0xdd, 0x80, 0x65, 0x73, 0x34, 0x72, 0x1e, 0x19, 0x64, 0x3c, 0x41, 0xdb, 0xe6, 0x81, 0x57, 0x95, - 0x2d, 0xb5, 0xc4, 0x00, 0x77, 0xe8, 0xf8, 0x07, 0x32, 0xd8, 0x4a, 0xbd, 0x88, 0x50, 0x67, 0xea, - 0x31, 0x9d, 0xf9, 0x3c, 0x54, 0x2d, 0x0c, 0x13, 0xa5, 0xec, 0x37, 0x63, 0x9e, 0xce, 0xb5, 0xc6, - 0x68, 0x12, 0x27, 0x3c, 0x92, 0x44, 0xe7, 0xca, 0x30, 0xb5, 0x57, 0x60, 0xb9, 0xef, 0x8c, 0x46, - 0xa4, 0x4f, 0x99, 0x35, 0x04, 0xd5, 0x26, 0xa3, 0xda, 0x0d, 0x01, 0x77, 0x39, 0xfd, 0x88, 0x6e, - 0xc1, 0x0c, 0xdd, 0xc2, 0xf3, 0x62, 0x6c, 0x9e, 0x5a, 0xe3, 0xe9, 0x98, 0x79, 0xcd, 0x92, 0x2e, - 0x1f, 0xe9, 0x8a, 0xe4, 0xb4, 0x3f, 0x9a, 0x7a, 0xc8, 0x8b, 0x21, 0x71, 0x16, 0xd9, 0xe6, 0xbb, - 0x01, 0xe0, 0x1b, 0x02, 0x99, 0x92, 0x41, 0xdf, 0x45, 0x51, 0xda, 0x82, 0x0c, 0x7f, 0x4c, 0x90, - 0x11, 0x38, 0x9d, 0x24, 0x19, 0x81, 0xfc, 0x34, 0x00, 0xae, 0x64, 0x8c, 0x88, 0x7d, 0x84, 0x67, - 0x1b, 0xf5, 0x66, 0x65, 0xbd, 0x89, 0x23, 0xf7, 0xd9, 0x00, 0x03, 0x5b, 0xb6, 0x04, 0x77, 0x05, - 0xd8, 0xb2, 0x05, 0x18, 0x99, 0xc0, 0x93, 0x88, 0x2a, 0x6b, 0x6f, 0x99, 0x1f, 0xb6, 0xe2, 0x91, - 0x5a, 0x04, 0xa5, 0xcb, 0x85, 0xae, 0xb1, 0x79, 0x0d, 0x1c, 0x60, 0x12, 0x66, 0x40, 0xa4, 0xca, - 0x81, 0x2b, 0x02, 0x68, 0xd9, 0x1c, 0xf8, 0x2c, 0x2c, 0x4e, 0x6d, 0xeb, 0xe1, 0x94, 0x08, 0xf8, - 0x2a, 0xe3, 0xbc, 0xc5, 0xc7, 0x38, 0x0a, 0xba, 0x6a, 0x62, 0xe3, 0xa6, 0x2e, 0xa5, 0x5d, 0x35, - 0x15, 0x35, 0x03, 0x6a, 0xcf, 0x40, 0x6b, 0x8c, 0xf2, 0xb6, 0xd0, 0x30, 0x0c, 0x67, 0xd8, 0x5b, - 0x63, 0x42, 0x02, 0x39, 0xb4, 0x37, 0x54, 0x5a, 0xcb, 0xe5, 0x42, 0xd6, 0x52, 0x85, 0xda, 0xdb, - 0x68, 0xe5, 0x18, 0x5b, 0xa8, 0xc2, 0xe2, 0x50, 0x17, 0x17, 0xd4, 0xba, 0x58, 0x3e, 0x9f, 0x2e, - 0x56, 0x66, 0xeb, 0x62, 0x75, 0x7e, 0x5d, 0xac, 0xcd, 0xa1, 0x8b, 0xf5, 0xd9, 0xba, 0xd8, 0x98, - 0x43, 0x17, 0x9b, 0x73, 0xe9, 0x22, 0xe4, 0xeb, 0x62, 0x2b, 0x47, 0x17, 0x17, 0x73, 0x74, 0xb1, - 0x9d, 0xa7, 0x8b, 0x9d, 0x19, 0xba, 0xb8, 0x94, 0xad, 0x8b, 0xdd, 0x02, 0xba, 0xb8, 0x9c, 0xd2, - 0xc5, 0x84, 0xb7, 0xd4, 0xe6, 0x4b, 0xa1, 0x56, 0x8a, 0x68, 0xeb, 0xdf, 0xab, 0xd0, 0xe3, 0xda, - 0xfa, 0x1f, 0xf1, 0xec, 0xd2, 0x42, 0xaa, 0x4a, 0x0b, 0xa9, 0xa9, 0x2d, 0xa4, 0x7e, 0x3e, 0x0b, - 0x69, 0xcc, 0xb6, 0x90, 0xe6, 0xfc, 0x16, 0x02, 0x73, 0x58, 0x48, 0x6b, 0xb6, 0x85, 0x2c, 0xce, - 0x61, 0x21, 0xed, 0xb9, 0x2c, 0xa4, 0x93, 0x6f, 0x21, 0x4b, 0x39, 0x16, 0xd2, 0xcd, 0xb1, 0x90, - 0xe5, 0x3c, 0x0b, 0xd1, 0x66, 0x58, 0xc8, 0x4a, 0xb6, 0x85, 0xac, 0x16, 0xb0, 0x90, 0x4b, 0x73, - 0x79, 0xeb, 0xb5, 0x22, 0xfa, 0xff, 0x4d, 0xa8, 0x73, 0xf5, 0x7f, 0x8c, 0xf4, 0x93, 0x4f, 0xcc, - 0x08, 0x9e, 0x7f, 0xb1, 0x00, 0x15, 0x9a, 0x40, 0x86, 0x81, 0x69, 0x29, 0x1a, 0x98, 0xa2, 0xd4, - 0x4f, 0x70, 0xd1, 0xb0, 0x32, 0x22, 0x1f, 0xe7, 0x30, 0xa4, 0xeb, 0xd0, 0xc5, 0xf7, 0x33, 0xf6, - 0x50, 0x24, 0x86, 0x47, 0xdc, 0x13, 0xab, 0x2f, 0x8d, 0xaa, 0xc3, 0xc6, 0xf7, 0x86, 0x07, 0x7c, - 0x54, 0x7b, 0x15, 0xea, 0x7d, 0x5e, 0x3e, 0x10, 0x4e, 0x7f, 0x25, 0xba, 0x09, 0x51, 0x59, 0xd0, - 0x25, 0x0e, 0x45, 0x1f, 0xe1, 0x34, 0xcc, 0xc4, 0x98, 0xe9, 0x25, 0xd0, 0xef, 0x73, 0x90, 0x2e, - 0x71, 0x94, 0xc2, 0xaf, 0x17, 0x11, 0xfe, 0x1b, 0xd0, 0x64, 0xca, 0xc0, 0x6a, 0x75, 0x37, 0x22, - 0xb5, 0xba, 0x72, 0x7e, 0x61, 0x65, 0xeb, 0x36, 0xb4, 0xbf, 0xee, 0x39, 0xb6, 0x4e, 0x86, 0xc4, - 0x25, 0x36, 0x6e, 0x74, 0x19, 0x2a, 0x86, 0x4b, 0x86, 0x42, 0xc6, 0x65, 0x04, 0xcc, 0xae, 0x3f, - 0x6d, 0x4d, 0xa0, 0x2e, 0xf6, 0x34, 0x67, 0x71, 0xe5, 0xdc, 0xb9, 0xcc, 0x1d, 0x68, 0x48, 0xa0, - 0x72, 0xc9, 0x17, 0x64, 0x55, 0x71, 0x41, 0xed, 0x80, 0x38, 0x74, 0xeb, 0x1d, 0x68, 0x45, 0x14, - 0x50, 0x49, 0xe9, 0x7a, 0x9c, 0x52, 0x4c, 0x98, 0x42, 0x6f, 0x05, 0xb1, 0xf7, 0xa0, 0xc3, 0x88, - 0x85, 0x45, 0x34, 0x15, 0xbd, 0x57, 0xe2, 0xf4, 0x2e, 0x29, 0x8b, 0x02, 0x92, 0xe4, 0x1e, 0xb4, - 0x05, 0x49, 0xff, 0x98, 0xbd, 0x5b, 0x15, 0xc5, 0x1b, 0x71, 0x8a, 0xab, 0xc9, 0x7a, 0x06, 0x9d, - 0x98, 0x24, 0x28, 0xab, 0x07, 0x85, 0x09, 0xca, 0x89, 0x92, 0xe0, 0x87, 0xa0, 0xc5, 0x08, 0x06, - 0xb9, 0x43, 0x8a, 0xea, 0xcd, 0x38, 0xd5, 0x75, 0x15, 0x55, 0x36, 0x3b, 0xf9, 0x72, 0xc4, 0x19, - 0x5a, 0xf4, 0xe5, 0x08, 0x4d, 0x17, 0xc4, 0xc6, 0x70, 0x85, 0x13, 0x4b, 0x97, 0x26, 0x32, 0x05, - 0xfb, 0x66, 0x9c, 0xfa, 0x73, 0x33, 0xea, 0x1e, 0x51, 0x39, 0xbf, 0x21, 0x79, 0xf7, 0x5d, 0xcb, - 0x3e, 0x52, 0x52, 0x5f, 0x8d, 0x52, 0x6f, 0xca, 0x89, 0xef, 0x43, 0x37, 0x32, 0x71, 0xd7, 0x75, - 0x4d, 0xb5, 0x82, 0xbf, 0x1a, 0xe7, 0x2d, 0xe6, 0x53, 0x23, 0x73, 0x25, 0xd9, 0xdf, 0x96, 0x91, - 0xae, 0x63, 0xc7, 0x6b, 0xbc, 0x04, 0x36, 0x8f, 0x99, 0x06, 0x1b, 0x41, 0xdd, 0xc9, 0xf0, 0xa6, - 0x87, 0x46, 0xac, 0xd2, 0xff, 0x7c, 0x5a, 0xe1, 0xd3, 0x01, 0xce, 0xdb, 0x4f, 0xe9, 0xbd, 0xe3, - 0xac, 0xe0, 0x67, 0x04, 0x57, 0x69, 0xc0, 0x60, 0x0c, 0x30, 0xeb, 0x55, 0xaf, 0xc4, 0xf7, 0xf0, - 0x62, 0x74, 0xa5, 0xec, 0x34, 0x19, 0xd7, 0xda, 0x18, 0x66, 0x27, 0xd1, 0x87, 0xb0, 0x81, 0x47, - 0xa3, 0x7b, 0xa6, 0x5e, 0xa9, 0x9c, 0x7e, 0x93, 0xef, 0x51, 0x6c, 0xe5, 0x32, 0x97, 0x1f, 0xaa, - 0x41, 0x9a, 0x01, 0xeb, 0xb4, 0x42, 0xa8, 0x5e, 0x82, 0x17, 0x3f, 0xb6, 0x92, 0x56, 0xa8, 0x5c, - 0x61, 0x6d, 0xa2, 0x84, 0x84, 0x4d, 0x12, 0x3c, 0xfc, 0x7a, 0x7b, 0xe6, 0xd4, 0x3f, 0xde, 0xd9, - 0xed, 0xf7, 0x89, 0xe7, 0xbd, 0xe5, 0x0c, 0xc8, 0xac, 0x3e, 0xc7, 0x10, 0xf3, 0x78, 0x59, 0x95, - 0xa7, 0xff, 0xb5, 0xd7, 0xe8, 0x81, 0x80, 0xec, 0xc8, 0x94, 0x28, 0x56, 0x1a, 0xe1, 0xd4, 0x0f, - 0x18, 0x5c, 0x17, 0x78, 0x34, 0x6a, 0xa2, 0xc3, 0x8e, 0x6b, 0x7d, 0x9f, 0xf5, 0x27, 0x0c, 0xea, - 0xbf, 0x45, 0x42, 0x14, 0x03, 0xbc, 0x8f, 0xce, 0x1c, 0x03, 0x18, 0xdf, 0xf9, 0x84, 0x70, 0x24, - 0x1e, 0x7f, 0x36, 0xd8, 0x00, 0x05, 0x26, 0x0e, 0x8f, 0xda, 0x7c, 0x91, 0x77, 0xa1, 0xc3, 0xef, - 0xaf, 0x25, 0x58, 0x17, 0x32, 0x9a, 0x4c, 0x46, 0xf3, 0x74, 0x54, 0x9e, 0x8c, 0x90, 0x62, 0xfb, - 0xae, 0xe4, 0xef, 0xbb, 0x3a, 0xdf, 0xbe, 0x0b, 0xf5, 0x34, 0x7e, 0xb8, 0x00, 0x6b, 0x9c, 0xb1, - 0x7b, 0x63, 0xba, 0x6f, 0xcb, 0xff, 0x6f, 0xd3, 0x8c, 0x7f, 0x83, 0x10, 0xfe, 0x52, 0x92, 0x42, - 0xd8, 0x37, 0x3d, 0xef, 0x91, 0xe3, 0x0e, 0xfe, 0x0f, 0xde, 0xfc, 0x47, 0xb0, 0x18, 0xe5, 0xeb, - 0x31, 0xfa, 0x3d, 0xec, 0x84, 0xc8, 0x08, 0xb8, 0x7f, 0x5e, 0x81, 0xe6, 0x1e, 0x3e, 0x98, 0x32, - 0xd9, 0x64, 0x75, 0xfb, 0x12, 0xab, 0xd3, 0xf2, 0x32, 0x3d, 0xed, 0xc9, 0x4c, 0xc7, 0x63, 0xd3, - 0x3d, 0x93, 0x31, 0xb7, 0x78, 0x9c, 0x23, 0xe6, 0x4e, 0x95, 0x6b, 0x2b, 0x85, 0xca, 0xb5, 0x98, - 0x10, 0x39, 0x92, 0x37, 0xc3, 0x1a, 0x48, 0xf1, 0x06, 0x63, 0xf7, 0x06, 0xb1, 0xde, 0x4f, 0x2d, - 0xd1, 0xfb, 0x89, 0xf6, 0x8c, 0xea, 0x89, 0x9e, 0xd1, 0x97, 0x62, 0x3d, 0x9b, 0x06, 0x13, 0xdd, - 0x86, 0x32, 0x3c, 0xe3, 0x47, 0x7d, 0xb4, 0x5b, 0xf3, 0x7a, 0xb4, 0x5b, 0xd3, 0x4c, 0x47, 0x76, - 0x32, 0xc0, 0x89, 0xf5, 0x68, 0x22, 0xad, 0x2d, 0x88, 0xb7, 0xb6, 0xae, 0x02, 0x0c, 0xc8, 0xc4, - 0x25, 0xe8, 0xcb, 0xc8, 0x40, 0x64, 0xbd, 0x91, 0x91, 0xf3, 0x75, 0x77, 0x54, 0xea, 0xd7, 0x2e, - 0xa2, 0x7e, 0xbf, 0x2a, 0x41, 0x33, 0x8c, 0x22, 0x6e, 0x41, 0xe7, 0x10, 0xc3, 0x8a, 0xf0, 0x30, - 0x14, 0x81, 0x43, 0x2c, 0xc0, 0x8b, 0x05, 0x1e, 0x78, 0xf0, 0xb5, 0x0f, 0x63, 0x91, 0xc8, 0x7d, - 0xd0, 0x6c, 0x7c, 0x9f, 0x09, 0x3a, 0x3c, 0x2c, 0xb8, 0x12, 0x63, 0x2a, 0x11, 0xc3, 0x20, 0xa9, - 0xae, 0x9d, 0x18, 0x0b, 0x4f, 0xcf, 0x23, 0x58, 0x55, 0xf5, 0xd9, 0xb4, 0xbd, 0x7c, 0x7b, 0xd9, - 0x48, 0x89, 0x21, 0x0c, 0xcc, 0xd5, 0x26, 0xf3, 0x69, 0x09, 0x3a, 0x71, 0xed, 0xd0, 0xbe, 0x00, - 0xcd, 0xa4, 0x44, 0xd4, 0xb1, 0x3e, 0x6e, 0x21, 0xc4, 0xa4, 0xd2, 0xfc, 0x18, 0x13, 0x32, 0x9a, - 0x83, 0xf1, 0x8c, 0x4c, 0x15, 0x2e, 0xc7, 0x52, 0x36, 0x2a, 0xcd, 0x8f, 0xa3, 0x03, 0xe1, 0xfe, - 0xff, 0x50, 0x86, 0x46, 0x90, 0x3a, 0x28, 0x32, 0xbb, 0x97, 0xa0, 0x7c, 0x44, 0x7c, 0x55, 0x26, - 0x12, 0xd8, 0xbf, 0x4e, 0x31, 0x28, 0xe2, 0x64, 0xea, 0x0b, 0xff, 0x98, 0x85, 0x88, 0x18, 0xda, - 0xcb, 0x50, 0x99, 0xd0, 0xf6, 0x6e, 0x25, 0x0f, 0x93, 0xa1, 0x60, 0x04, 0x5b, 0x1b, 0x90, 0x11, - 0x6e, 0x5a, 0x64, 0xd4, 0x19, 0xc8, 0x02, 0x09, 0xd3, 0x87, 0xba, 0x33, 0xe1, 0x6d, 0xc8, 0x5a, - 0x1e, 0xbe, 0xc4, 0xa2, 0xac, 0xd0, 0x90, 0x54, 0x14, 0xb9, 0xb2, 0x58, 0xa1, 0x28, 0x34, 0x27, - 0xc3, 0x40, 0xac, 0x7f, 0x2c, 0xda, 0x17, 0x19, 0xb8, 0x1c, 0x27, 0xe1, 0x26, 0x9a, 0x85, 0xdc, - 0xc4, 0xb9, 0x3b, 0x48, 0x7f, 0xab, 0xc2, 0x9a, 0x3a, 0x9a, 0xbc, 0xa8, 0x31, 0x5e, 0xd4, 0x18, - 0xff, 0xd7, 0x6b, 0x8c, 0x8f, 0xa0, 0xca, 0x2e, 0x68, 0x28, 0x29, 0x95, 0x0a, 0x50, 0x42, 0xe7, - 0x53, 0x61, 0xb7, 0x4d, 0x16, 0xd8, 0xa4, 0x75, 0x85, 0xc3, 0x17, 0x75, 0x13, 0x86, 0xb6, 0xf5, - 0xb3, 0x2a, 0x2c, 0x25, 0xb4, 0xf6, 0xa2, 0x27, 0x75, 0xd1, 0x93, 0x3a, 0x57, 0x4f, 0x4a, 0xa5, - 0xc3, 0x5a, 0x11, 0x6b, 0xf8, 0x36, 0x40, 0x18, 0x82, 0x3c, 0xe1, 0x3b, 0x5f, 0xbf, 0xae, 0xc1, - 0xe5, 0x8c, 0xc2, 0xc8, 0xc5, 0x35, 0x85, 0x8b, 0x6b, 0x0a, 0x17, 0xd7, 0x14, 0x42, 0x33, 0xfc, - 0x47, 0x09, 0x1a, 0x41, 0x39, 0x7d, 0xf6, 0xc5, 0xae, 0xed, 0xa0, 0x3b, 0xc3, 0xc3, 0xee, 0xb5, - 0x74, 0xcd, 0x9a, 0x1d, 0x3c, 0xf2, 0xea, 0xeb, 0xab, 0x50, 0xe7, 0x95, 0x55, 0x79, 0x78, 0xac, - 0xa4, 0x0b, 0xb2, 0x9e, 0x2e, 0x71, 0xb4, 0xd7, 0xa0, 0x21, 0xae, 0x2b, 0xc9, 0xcc, 0x7a, 0x35, - 0x9e, 0x59, 0x73, 0x98, 0x1e, 0x60, 0x9d, 0xff, 0x4e, 0x33, 0x81, 0x15, 0xc5, 0x65, 0x44, 0xed, - 0xdd, 0x7c, 0x87, 0x94, 0x3e, 0x73, 0x83, 0xd6, 0x82, 0xda, 0x25, 0xfd, 0xa4, 0x04, 0xed, 0x78, - 0x97, 0x61, 0x87, 0x3a, 0x22, 0x3e, 0x10, 0xdc, 0x1e, 0x57, 0xe4, 0xdc, 0x98, 0x20, 0x05, 0x78, - 0x4f, 0x36, 0xbf, 0xfa, 0x29, 0x26, 0xc2, 0x41, 0x66, 0xaf, 0xbd, 0x05, 0x6d, 0xb9, 0x8c, 0xd1, - 0x77, 0x06, 0x44, 0x6c, 0xf4, 0x6a, 0xe6, 0x46, 0x79, 0xb7, 0x63, 0x51, 0x4e, 0xa2, 0xb5, 0x5d, - 0xe5, 0xdb, 0x58, 0x28, 0xf2, 0x36, 0x7e, 0xd3, 0x84, 0x9a, 0x70, 0xd4, 0x8a, 0x8c, 0x2f, 0x2b, - 0x40, 0x09, 0x7a, 0xab, 0xe5, 0x9c, 0x4b, 0x7f, 0x95, 0xdc, 0x4b, 0x7f, 0xb3, 0x02, 0x8f, 0x84, - 0x25, 0xd6, 0x52, 0x96, 0x18, 0x71, 0x89, 0xf5, 0x39, 0x5c, 0x62, 0x63, 0xb6, 0x4b, 0x6c, 0xce, - 0xe1, 0x12, 0x61, 0x2e, 0x97, 0xd8, 0xca, 0x77, 0x89, 0x8b, 0x39, 0x2e, 0xb1, 0x9d, 0xe3, 0x12, - 0x3b, 0x79, 0x2e, 0x71, 0x69, 0x86, 0x4b, 0xec, 0xa6, 0x5d, 0xe2, 0x0b, 0xd0, 0xa1, 0xc4, 0x23, - 0xc6, 0xc6, 0x33, 0x81, 0x36, 0x8e, 0x46, 0x62, 0x05, 0x8a, 0x86, 0xcb, 0x44, 0xd0, 0x34, 0x81, - 0x66, 0xd9, 0x11, 0xb4, 0xe8, 0x41, 0xbf, 0x92, 0xb8, 0xa6, 0x39, 0x57, 0x46, 0xf0, 0x61, 0x96, - 0x0b, 0xb8, 0x94, 0x6e, 0x2d, 0x65, 0x7d, 0x7a, 0xa2, 0xf6, 0x06, 0xda, 0x75, 0x71, 0xec, 0xaf, - 0xa5, 0xed, 0xfe, 0x01, 0x8e, 0xf3, 0xd8, 0x9d, 0x05, 0x03, 0xaf, 0xc8, 0x43, 0xff, 0x72, 0x3a, - 0xb9, 0x0f, 0x9a, 0xe6, 0xf2, 0xb8, 0x7f, 0x19, 0x6a, 0x18, 0x60, 0x50, 0xfd, 0xec, 0x65, 0xf6, - 0xce, 0xab, 0x88, 0x81, 0xea, 0xfa, 0x45, 0x80, 0xc8, 0x8e, 0xd6, 0xd3, 0xce, 0x3c, 0xe4, 0x56, - 0x8f, 0x60, 0x6a, 0xcf, 0x43, 0x7b, 0x60, 0x51, 0x0b, 0x42, 0x61, 0x9b, 0xbe, 0xe3, 0xf6, 0x36, - 0x98, 0x82, 0xc4, 0x07, 0xe3, 0x57, 0x5e, 0x37, 0x13, 0x57, 0x5e, 0x9f, 0x85, 0xf2, 0xe9, 0x78, - 0xd4, 0xbb, 0x92, 0xb6, 0xb8, 0x6f, 0x8d, 0x47, 0x3a, 0x85, 0xa5, 0xcb, 0xac, 0x4f, 0x3f, 0xee, - 0xad, 0xd8, 0xab, 0x8f, 0x71, 0x2b, 0xf6, 0x99, 0x22, 0x1e, 0xeb, 0x07, 0x00, 0xe1, 0xb9, 0x57, - 0xf0, 0x4b, 0xa3, 0x37, 0xa1, 0x35, 0xb4, 0xd0, 0xa1, 0x64, 0x1f, 0xa9, 0xe1, 0x8d, 0x67, 0x9c, - 0x06, 0xc3, 0xe0, 0x29, 0xf4, 0xe2, 0x3e, 0xac, 0x28, 0xba, 0xb9, 0xda, 0x77, 0xf3, 0xcf, 0xaf, - 0xeb, 0xe9, 0x80, 0x3a, 0xa3, 0x25, 0xac, 0x3e, 0xce, 0xfe, 0x5c, 0x81, 0xcb, 0x59, 0xcd, 0xe8, - 0x31, 0x3c, 0x7d, 0x48, 0x3f, 0x12, 0x32, 0xcc, 0xd8, 0x57, 0x42, 0x46, 0x50, 0xf3, 0xe5, 0xa2, - 0x79, 0x29, 0x56, 0x61, 0xcd, 0xfe, 0xaa, 0x08, 0x37, 0xbe, 0x79, 0x98, 0xf3, 0xd1, 0xd1, 0x5d, - 0xe8, 0x22, 0x11, 0xe3, 0x13, 0x72, 0x16, 0xae, 0xc0, 0x25, 0x19, 0xab, 0x6b, 0xc5, 0xbf, 0xb2, - 0x42, 0xa2, 0x1d, 0x33, 0xfe, 0xdd, 0xd5, 0xf7, 0xa0, 0xe7, 0xb0, 0xb6, 0x84, 0x61, 0x89, 0x86, - 0x54, 0x48, 0xaf, 0x9c, 0xee, 0x8a, 0xaa, 0x7b, 0x57, 0xb4, 0x2b, 0xea, 0xa8, 0xbb, 0x5a, 0x21, - 0xfd, 0x89, 0xe8, 0xf5, 0x84, 0xf4, 0x2b, 0x59, 0xf4, 0x93, 0x6d, 0xa1, 0x90, 0x7e, 0xaa, 0x61, - 0x74, 0x04, 0x9b, 0x82, 0xbe, 0x19, 0x36, 0x12, 0xc3, 0x25, 0xf8, 0x01, 0xf7, 0x42, 0x7a, 0x09, - 0x45, 0xdb, 0x11, 0x57, 0x59, 0x77, 0x32, 0x7b, 0x92, 0x24, 0x5c, 0x88, 0x75, 0x75, 0x59, 0xb8, - 0x10, 0x2e, 0x54, 0x4b, 0x7b, 0xc7, 0xac, 0x1e, 0x30, 0x6d, 0xbc, 0x3b, 0x19, 0xb0, 0x50, 0xc3, - 0x8f, 0x43, 0x0d, 0x8f, 0xb4, 0x04, 0xb4, 0xf7, 0xf2, 0x35, 0xfc, 0x4a, 0x46, 0xdb, 0x88, 0x5f, - 0x2c, 0x50, 0x6b, 0xf5, 0x73, 0xd0, 0x8a, 0xde, 0x5c, 0x58, 0x0d, 0x3f, 0xee, 0x2b, 0x87, 0x77, - 0x1c, 0x7e, 0x57, 0x82, 0xf2, 0x03, 0x53, 0x7d, 0x2b, 0x62, 0xf6, 0xc7, 0x6e, 0x29, 0xcf, 0x56, - 0x3e, 0xf7, 0x37, 0x22, 0x85, 0xbe, 0xe0, 0xba, 0x06, 0x0d, 0x79, 0xc2, 0x64, 0xec, 0xef, 0x23, - 0x58, 0xfa, 0x20, 0x51, 0x6f, 0x7a, 0x82, 0x1f, 0x93, 0xfc, 0x1e, 0xa5, 0x87, 0x6e, 0x5e, 0x29, - 0xbd, 0x2b, 0xd0, 0xa4, 0xbf, 0xde, 0xc4, 0xec, 0xcb, 0x7b, 0x25, 0xe1, 0x00, 0x0d, 0xfe, 0x26, - 0x18, 0x0f, 0x5a, 0xa7, 0x22, 0xca, 0x13, 0x4f, 0x74, 0x16, 0x06, 0x27, 0xae, 0x75, 0x38, 0xf5, - 0x89, 0xf8, 0x4c, 0x2f, 0x1c, 0xa0, 0xa1, 0xcc, 0x23, 0x17, 0x0d, 0x82, 0x0c, 0x44, 0x0a, 0x2e, - 0x1f, 0xcf, 0xdd, 0xc7, 0xbc, 0xf5, 0x22, 0x74, 0x1c, 0xf7, 0x48, 0xe2, 0x1a, 0x27, 0x3b, 0xb7, - 0x16, 0xc5, 0xb7, 0xab, 0xfb, 0xf4, 0xeb, 0xcf, 0xfd, 0xd2, 0x2f, 0x17, 0xca, 0x7b, 0xbb, 0x07, - 0x87, 0x35, 0xf6, 0x31, 0xe8, 0xeb, 0xff, 0x0a, 0x00, 0x00, 0xff, 0xff, 0xd4, 0x0a, 0xef, 0xca, - 0xe4, 0x3a, 0x00, 0x00, -} diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv2/README.md b/vendor/github.com/googleapis/gnostic/OpenAPIv2/README.md deleted file mode 100644 index 836fb32a..00000000 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv2/README.md +++ /dev/null @@ -1,16 +0,0 @@ -# OpenAPI v2 Protocol Buffer Models - -This directory contains a Protocol Buffer-language model -and related code for supporting OpenAPI v2. - -Gnostic applications and plugins can use OpenAPIv2.proto -to generate Protocol Buffer support code for their preferred languages. - -OpenAPIv2.go is used by Gnostic to read JSON and YAML OpenAPI -descriptions into the Protocol Buffer-based datastructures -generated from OpenAPIv2.proto. - -OpenAPIv2.proto and OpenAPIv2.go are generated by the Gnostic -compiler generator, and OpenAPIv2.pb.go is generated by -protoc, the Protocol Buffer compiler, and protoc-gen-go, the -Protocol Buffer Go code generation plugin. diff --git a/vendor/github.com/googleapis/gnostic/compiler/README.md b/vendor/github.com/googleapis/gnostic/compiler/README.md index 848b16c6..ee9783d2 100644 --- a/vendor/github.com/googleapis/gnostic/compiler/README.md +++ b/vendor/github.com/googleapis/gnostic/compiler/README.md @@ -1,3 +1,4 @@ # Compiler support code -This directory contains compiler support code used by Gnostic and Gnostic extensions. \ No newline at end of file +This directory contains compiler support code used by Gnostic and Gnostic +extensions. diff --git a/vendor/github.com/googleapis/gnostic/compiler/context.go b/vendor/github.com/googleapis/gnostic/compiler/context.go index a64c1b75..1bfe9612 100644 --- a/vendor/github.com/googleapis/gnostic/compiler/context.go +++ b/vendor/github.com/googleapis/gnostic/compiler/context.go @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2017 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -14,30 +14,36 @@ package compiler +import ( + yaml "gopkg.in/yaml.v3" +) + // Context contains state of the compiler as it traverses a document. type Context struct { Parent *Context Name string + Node *yaml.Node ExtensionHandlers *[]ExtensionHandler } // NewContextWithExtensions returns a new object representing the compiler state -func NewContextWithExtensions(name string, parent *Context, extensionHandlers *[]ExtensionHandler) *Context { - return &Context{Name: name, Parent: parent, ExtensionHandlers: extensionHandlers} +func NewContextWithExtensions(name string, node *yaml.Node, parent *Context, extensionHandlers *[]ExtensionHandler) *Context { + return &Context{Name: name, Node: node, Parent: parent, ExtensionHandlers: extensionHandlers} } // NewContext returns a new object representing the compiler state -func NewContext(name string, parent *Context) *Context { +func NewContext(name string, node *yaml.Node, parent *Context) *Context { if parent != nil { - return &Context{Name: name, Parent: parent, ExtensionHandlers: parent.ExtensionHandlers} + return &Context{Name: name, Node: node, Parent: parent, ExtensionHandlers: parent.ExtensionHandlers} } return &Context{Name: name, Parent: parent, ExtensionHandlers: nil} } // Description returns a text description of the compiler state func (context *Context) Description() string { + name := context.Name if context.Parent != nil { - return context.Parent.Description() + "." + context.Name + name = context.Parent.Description() + "." + name } - return context.Name + return name } diff --git a/vendor/github.com/googleapis/gnostic/compiler/error.go b/vendor/github.com/googleapis/gnostic/compiler/error.go index d8672c10..6f40515d 100644 --- a/vendor/github.com/googleapis/gnostic/compiler/error.go +++ b/vendor/github.com/googleapis/gnostic/compiler/error.go @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2017 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -14,6 +14,8 @@ package compiler +import "fmt" + // Error represents compiler errors and their location in the document. type Error struct { Context *Context @@ -25,12 +27,19 @@ func NewError(context *Context, message string) *Error { return &Error{Context: context, Message: message} } +func (err *Error) locationDescription() string { + if err.Context.Node != nil { + return fmt.Sprintf("[%d,%d] %s", err.Context.Node.Line, err.Context.Node.Column, err.Context.Description()) + } + return err.Context.Description() +} + // Error returns the string value of an Error. func (err *Error) Error() string { if err.Context == nil { - return "ERROR " + err.Message + return err.Message } - return "ERROR " + err.Context.Description() + " " + err.Message + return err.locationDescription() + " " + err.Message } // ErrorGroup is a container for groups of Error values. diff --git a/vendor/github.com/googleapis/gnostic/compiler/extension-handler.go b/vendor/github.com/googleapis/gnostic/compiler/extension-handler.go deleted file mode 100644 index 1f85b650..00000000 --- a/vendor/github.com/googleapis/gnostic/compiler/extension-handler.go +++ /dev/null @@ -1,101 +0,0 @@ -// Copyright 2017 Google Inc. All Rights Reserved. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package compiler - -import ( - "bytes" - "fmt" - "os/exec" - - "strings" - - "errors" - - "github.com/golang/protobuf/proto" - "github.com/golang/protobuf/ptypes/any" - ext_plugin "github.com/googleapis/gnostic/extensions" - yaml "gopkg.in/yaml.v2" -) - -// ExtensionHandler describes a binary that is called by the compiler to handle specification extensions. -type ExtensionHandler struct { - Name string -} - -// HandleExtension calls a binary extension handler. -func HandleExtension(context *Context, in interface{}, extensionName string) (bool, *any.Any, error) { - handled := false - var errFromPlugin error - var outFromPlugin *any.Any - - if context != nil && context.ExtensionHandlers != nil && len(*(context.ExtensionHandlers)) != 0 { - for _, customAnyProtoGenerator := range *(context.ExtensionHandlers) { - outFromPlugin, errFromPlugin = customAnyProtoGenerator.handle(in, extensionName) - if outFromPlugin == nil { - continue - } else { - handled = true - break - } - } - } - return handled, outFromPlugin, errFromPlugin -} - -func (extensionHandlers *ExtensionHandler) handle(in interface{}, extensionName string) (*any.Any, error) { - if extensionHandlers.Name != "" { - binary, _ := yaml.Marshal(in) - - request := &ext_plugin.ExtensionHandlerRequest{} - - version := &ext_plugin.Version{} - version.Major = 0 - version.Minor = 1 - version.Patch = 0 - request.CompilerVersion = version - - request.Wrapper = &ext_plugin.Wrapper{} - - request.Wrapper.Version = "v2" - request.Wrapper.Yaml = string(binary) - request.Wrapper.ExtensionName = extensionName - - requestBytes, _ := proto.Marshal(request) - cmd := exec.Command(extensionHandlers.Name) - cmd.Stdin = bytes.NewReader(requestBytes) - output, err := cmd.Output() - - if err != nil { - fmt.Printf("Error: %+v\n", err) - return nil, err - } - response := &ext_plugin.ExtensionHandlerResponse{} - err = proto.Unmarshal(output, response) - if err != nil { - fmt.Printf("Error: %+v\n", err) - fmt.Printf("%s\n", string(output)) - return nil, err - } - if !response.Handled { - return nil, nil - } - if len(response.Error) != 0 { - message := fmt.Sprintf("Errors when parsing: %+v for field %s by vendor extension handler %s. Details %+v", in, extensionName, extensionHandlers.Name, strings.Join(response.Error, ",")) - return nil, errors.New(message) - } - return response.Value, nil - } - return nil, nil -} diff --git a/vendor/github.com/googleapis/gnostic/compiler/extensions.go b/vendor/github.com/googleapis/gnostic/compiler/extensions.go new file mode 100644 index 00000000..20848a0a --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/compiler/extensions.go @@ -0,0 +1,85 @@ +// Copyright 2017 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package compiler + +import ( + "bytes" + "fmt" + "os/exec" + "strings" + + "github.com/golang/protobuf/proto" + "github.com/golang/protobuf/ptypes/any" + extensions "github.com/googleapis/gnostic/extensions" + yaml "gopkg.in/yaml.v3" +) + +// ExtensionHandler describes a binary that is called by the compiler to handle specification extensions. +type ExtensionHandler struct { + Name string +} + +// CallExtension calls a binary extension handler. +func CallExtension(context *Context, in *yaml.Node, extensionName string) (handled bool, response *any.Any, err error) { + if context == nil || context.ExtensionHandlers == nil { + return false, nil, nil + } + handled = false + for _, handler := range *(context.ExtensionHandlers) { + response, err = handler.handle(in, extensionName) + if response == nil { + continue + } else { + handled = true + break + } + } + return handled, response, err +} + +func (extensionHandlers *ExtensionHandler) handle(in *yaml.Node, extensionName string) (*any.Any, error) { + if extensionHandlers.Name != "" { + yamlData, _ := yaml.Marshal(in) + request := &extensions.ExtensionHandlerRequest{ + CompilerVersion: &extensions.Version{ + Major: 0, + Minor: 1, + Patch: 0, + }, + Wrapper: &extensions.Wrapper{ + Version: "unknown", // TODO: set this to the type/version of spec being parsed. + Yaml: string(yamlData), + ExtensionName: extensionName, + }, + } + requestBytes, _ := proto.Marshal(request) + cmd := exec.Command(extensionHandlers.Name) + cmd.Stdin = bytes.NewReader(requestBytes) + output, err := cmd.Output() + if err != nil { + return nil, err + } + response := &extensions.ExtensionHandlerResponse{} + err = proto.Unmarshal(output, response) + if err != nil || !response.Handled { + return nil, err + } + if len(response.Errors) != 0 { + return nil, fmt.Errorf("Errors when parsing: %+v for field %s by vendor extension handler %s. Details %+v", in, extensionName, extensionHandlers.Name, strings.Join(response.Errors, ",")) + } + return response.Value, nil + } + return nil, nil +} diff --git a/vendor/github.com/googleapis/gnostic/compiler/helpers.go b/vendor/github.com/googleapis/gnostic/compiler/helpers.go index 76df635f..48f02f39 100644 --- a/vendor/github.com/googleapis/gnostic/compiler/helpers.go +++ b/vendor/github.com/googleapis/gnostic/compiler/helpers.go @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2017 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -16,56 +16,63 @@ package compiler import ( "fmt" - "gopkg.in/yaml.v2" "regexp" "sort" "strconv" + + "github.com/googleapis/gnostic/jsonschema" + "gopkg.in/yaml.v3" ) // compiler helper functions, usually called from generated code -// UnpackMap gets a yaml.MapSlice if possible. -func UnpackMap(in interface{}) (yaml.MapSlice, bool) { - m, ok := in.(yaml.MapSlice) - if ok { - return m, true - } - // do we have an empty array? - a, ok := in.([]interface{}) - if ok && len(a) == 0 { - // if so, return an empty map - return yaml.MapSlice{}, true +// UnpackMap gets a *yaml.Node if possible. +func UnpackMap(in *yaml.Node) (*yaml.Node, bool) { + if in == nil { + return nil, false } - return nil, false + return in, true } -// SortedKeysForMap returns the sorted keys of a yaml.MapSlice. -func SortedKeysForMap(m yaml.MapSlice) []string { +// SortedKeysForMap returns the sorted keys of a yamlv2.MapSlice. +func SortedKeysForMap(m *yaml.Node) []string { keys := make([]string, 0) - for _, item := range m { - keys = append(keys, item.Key.(string)) + if m.Kind == yaml.MappingNode { + for i := 0; i < len(m.Content); i += 2 { + keys = append(keys, m.Content[i].Value) + } } sort.Strings(keys) return keys } -// MapHasKey returns true if a yaml.MapSlice contains a specified key. -func MapHasKey(m yaml.MapSlice, key string) bool { - for _, item := range m { - itemKey, ok := item.Key.(string) - if ok && key == itemKey { - return true +// MapHasKey returns true if a yamlv2.MapSlice contains a specified key. +func MapHasKey(m *yaml.Node, key string) bool { + if m == nil { + return false + } + if m.Kind == yaml.MappingNode { + for i := 0; i < len(m.Content); i += 2 { + itemKey := m.Content[i].Value + if key == itemKey { + return true + } } } return false } // MapValueForKey gets the value of a map value for a specified key. -func MapValueForKey(m yaml.MapSlice, key string) interface{} { - for _, item := range m { - itemKey, ok := item.Key.(string) - if ok && key == itemKey { - return item.Value +func MapValueForKey(m *yaml.Node, key string) *yaml.Node { + if m == nil { + return nil + } + if m.Kind == yaml.MappingNode { + for i := 0; i < len(m.Content); i += 2 { + itemKey := m.Content[i].Value + if key == itemKey { + return m.Content[i+1] + } } } return nil @@ -83,8 +90,118 @@ func ConvertInterfaceArrayToStringArray(interfaceArray []interface{}) []string { return stringArray } +// SequenceNodeForNode returns a node if it is a SequenceNode. +func SequenceNodeForNode(node *yaml.Node) (*yaml.Node, bool) { + if node.Kind != yaml.SequenceNode { + return nil, false + } + return node, true +} + +// BoolForScalarNode returns the bool value of a node. +func BoolForScalarNode(node *yaml.Node) (bool, bool) { + if node == nil { + return false, false + } + if node.Kind == yaml.DocumentNode { + return BoolForScalarNode(node.Content[0]) + } + if node.Kind != yaml.ScalarNode { + return false, false + } + if node.Tag != "!!bool" { + return false, false + } + v, err := strconv.ParseBool(node.Value) + if err != nil { + return false, false + } + return v, true +} + +// IntForScalarNode returns the integer value of a node. +func IntForScalarNode(node *yaml.Node) (int64, bool) { + if node == nil { + return 0, false + } + if node.Kind == yaml.DocumentNode { + return IntForScalarNode(node.Content[0]) + } + if node.Kind != yaml.ScalarNode { + return 0, false + } + if node.Tag != "!!int" { + return 0, false + } + v, err := strconv.ParseInt(node.Value, 10, 64) + if err != nil { + return 0, false + } + return v, true +} + +// FloatForScalarNode returns the float value of a node. +func FloatForScalarNode(node *yaml.Node) (float64, bool) { + if node == nil { + return 0.0, false + } + if node.Kind == yaml.DocumentNode { + return FloatForScalarNode(node.Content[0]) + } + if node.Kind != yaml.ScalarNode { + return 0.0, false + } + if (node.Tag != "!!int") && (node.Tag != "!!float") { + return 0.0, false + } + v, err := strconv.ParseFloat(node.Value, 64) + if err != nil { + return 0.0, false + } + return v, true +} + +// StringForScalarNode returns the string value of a node. +func StringForScalarNode(node *yaml.Node) (string, bool) { + if node == nil { + return "", false + } + if node.Kind == yaml.DocumentNode { + return StringForScalarNode(node.Content[0]) + } + switch node.Kind { + case yaml.ScalarNode: + switch node.Tag { + case "!!int": + return node.Value, true + case "!!str": + return node.Value, true + case "!!timestamp": + return node.Value, true + case "!!null": + return "", true + default: + return "", false + } + default: + return "", false + } +} + +// StringArrayForSequenceNode converts a sequence node to an array of strings, if possible. +func StringArrayForSequenceNode(node *yaml.Node) []string { + stringArray := make([]string, 0) + for _, item := range node.Content { + v, ok := StringForScalarNode(item) + if ok { + stringArray = append(stringArray, v) + } + } + return stringArray +} + // MissingKeysInMap identifies which keys from a list of required keys are not in a map. -func MissingKeysInMap(m yaml.MapSlice, requiredKeys []string) []string { +func MissingKeysInMap(m *yaml.Node, requiredKeys []string) []string { missingKeys := make([]string, 0) for _, k := range requiredKeys { if !MapHasKey(m, k) { @@ -95,64 +212,109 @@ func MissingKeysInMap(m yaml.MapSlice, requiredKeys []string) []string { } // InvalidKeysInMap returns keys in a map that don't match a list of allowed keys and patterns. -func InvalidKeysInMap(m yaml.MapSlice, allowedKeys []string, allowedPatterns []*regexp.Regexp) []string { +func InvalidKeysInMap(m *yaml.Node, allowedKeys []string, allowedPatterns []*regexp.Regexp) []string { invalidKeys := make([]string, 0) - for _, item := range m { - itemKey, ok := item.Key.(string) - if ok { - key := itemKey - found := false - // does the key match an allowed key? - for _, allowedKey := range allowedKeys { - if key == allowedKey { + if m == nil || m.Kind != yaml.MappingNode { + return invalidKeys + } + for i := 0; i < len(m.Content); i += 2 { + key := m.Content[i].Value + found := false + // does the key match an allowed key? + for _, allowedKey := range allowedKeys { + if key == allowedKey { + found = true + break + } + } + if !found { + // does the key match an allowed pattern? + for _, allowedPattern := range allowedPatterns { + if allowedPattern.MatchString(key) { found = true break } } if !found { - // does the key match an allowed pattern? - for _, allowedPattern := range allowedPatterns { - if allowedPattern.MatchString(key) { - found = true - break - } - } - if !found { - invalidKeys = append(invalidKeys, key) - } + invalidKeys = append(invalidKeys, key) } } } return invalidKeys } -// DescribeMap describes a map (for debugging purposes). -func DescribeMap(in interface{}, indent string) string { - description := "" - m, ok := in.(map[string]interface{}) - if ok { - keys := make([]string, 0) - for k := range m { - keys = append(keys, k) - } - sort.Strings(keys) - for _, k := range keys { - v := m[k] - description += fmt.Sprintf("%s%s:\n", indent, k) - description += DescribeMap(v, indent+" ") - } - return description +// NewNullNode creates a new Null node. +func NewNullNode() *yaml.Node { + node := &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!null", } - a, ok := in.([]interface{}) - if ok { - for i, v := range a { - description += fmt.Sprintf("%s%d:\n", indent, i) - description += DescribeMap(v, indent+" ") - } - return description + return node +} + +// NewMappingNode creates a new Mapping node. +func NewMappingNode() *yaml.Node { + return &yaml.Node{ + Kind: yaml.MappingNode, + Content: make([]*yaml.Node, 0), + } +} + +// NewSequenceNode creates a new Sequence node. +func NewSequenceNode() *yaml.Node { + node := &yaml.Node{ + Kind: yaml.SequenceNode, + Content: make([]*yaml.Node, 0), + } + return node +} + +// NewScalarNodeForString creates a new node to hold a string. +func NewScalarNodeForString(s string) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!str", + Value: s, + } +} + +// NewSequenceNodeForStringArray creates a new node to hold an array of strings. +func NewSequenceNodeForStringArray(strings []string) *yaml.Node { + node := &yaml.Node{ + Kind: yaml.SequenceNode, + Content: make([]*yaml.Node, 0), + } + for _, s := range strings { + node.Content = append(node.Content, NewScalarNodeForString(s)) + } + return node +} + +// NewScalarNodeForBool creates a new node to hold a bool. +func NewScalarNodeForBool(b bool) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!bool", + Value: fmt.Sprintf("%t", b), + } +} + +// NewScalarNodeForFloat creates a new node to hold a float. +func NewScalarNodeForFloat(f float64) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!float", + Value: fmt.Sprintf("%g", f), + } +} + +// NewScalarNodeForInt creates a new node to hold an integer. +func NewScalarNodeForInt(i int64) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!int", + Value: fmt.Sprintf("%d", i), } - description += fmt.Sprintf("%s%+v\n", indent, in) - return description } // PluralProperties returns the string "properties" pluralized. @@ -195,3 +357,40 @@ func StringValue(item interface{}) (value string, ok bool) { } return "", false } + +// Description returns a human-readable represention of an item. +func Description(item interface{}) string { + value, ok := item.(*yaml.Node) + if ok { + return jsonschema.Render(value) + } + return fmt.Sprintf("%+v", item) +} + +// Display returns a description of a node for use in error messages. +func Display(node *yaml.Node) string { + switch node.Kind { + case yaml.ScalarNode: + switch node.Tag { + case "!!str": + return fmt.Sprintf("%s (string)", node.Value) + } + } + return fmt.Sprintf("%+v (%T)", node, node) +} + +// Marshal creates a yaml version of a structure in our preferred style +func Marshal(in *yaml.Node) []byte { + clearStyle(in) + //bytes, _ := yaml.Marshal(&yaml.Node{Kind: yaml.DocumentNode, Content: []*yaml.Node{in}}) + bytes, _ := yaml.Marshal(in) + + return bytes +} + +func clearStyle(node *yaml.Node) { + node.Style = 0 + for _, c := range node.Content { + clearStyle(c) + } +} diff --git a/vendor/github.com/googleapis/gnostic/compiler/main.go b/vendor/github.com/googleapis/gnostic/compiler/main.go index 9713a21c..ce9fcc45 100644 --- a/vendor/github.com/googleapis/gnostic/compiler/main.go +++ b/vendor/github.com/googleapis/gnostic/compiler/main.go @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2017 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. diff --git a/vendor/github.com/googleapis/gnostic/compiler/reader.go b/vendor/github.com/googleapis/gnostic/compiler/reader.go index 25affd06..be0e8b40 100644 --- a/vendor/github.com/googleapis/gnostic/compiler/reader.go +++ b/vendor/github.com/googleapis/gnostic/compiler/reader.go @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2017 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -15,7 +15,6 @@ package compiler import ( - "errors" "fmt" "io/ioutil" "log" @@ -23,18 +22,30 @@ import ( "net/url" "path/filepath" "strings" + "sync" - yaml "gopkg.in/yaml.v2" + yaml "gopkg.in/yaml.v3" ) +var verboseReader = false + var fileCache map[string][]byte -var infoCache map[string]interface{} -var count int64 +var infoCache map[string]*yaml.Node -var verboseReader = false var fileCacheEnable = true var infoCacheEnable = true +// These locks are used to synchronize accesses to the fileCache and infoCache +// maps (above). They are global state and can throw thread-related errors +// when modified from separate goroutines. The general strategy is to protect +// all public functions in this file with mutex Lock() calls. As a result, to +// avoid deadlock, these public functions should not call other public +// functions, so some public functions have private equivalents. +// In the future, we might consider replacing the maps with sync.Map and +// eliminating these mutexes. +var fileCacheMutex sync.Mutex +var infoCacheMutex sync.Mutex + func initializeFileCache() { if fileCache == nil { fileCache = make(map[string][]byte, 0) @@ -43,19 +54,42 @@ func initializeFileCache() { func initializeInfoCache() { if infoCache == nil { - infoCache = make(map[string]interface{}, 0) + infoCache = make(map[string]*yaml.Node, 0) } } +// EnableFileCache turns on file caching. +func EnableFileCache() { + fileCacheMutex.Lock() + defer fileCacheMutex.Unlock() + fileCacheEnable = true +} + +// EnableInfoCache turns on parsed info caching. +func EnableInfoCache() { + infoCacheMutex.Lock() + defer infoCacheMutex.Unlock() + infoCacheEnable = true +} + +// DisableFileCache turns off file caching. func DisableFileCache() { + fileCacheMutex.Lock() + defer fileCacheMutex.Unlock() fileCacheEnable = false } +// DisableInfoCache turns off parsed info caching. func DisableInfoCache() { + infoCacheMutex.Lock() + defer infoCacheMutex.Unlock() infoCacheEnable = false } +// RemoveFromFileCache removes an entry from the file cache. func RemoveFromFileCache(fileurl string) { + fileCacheMutex.Lock() + defer fileCacheMutex.Unlock() if !fileCacheEnable { return } @@ -63,7 +97,10 @@ func RemoveFromFileCache(fileurl string) { delete(fileCache, fileurl) } +// RemoveFromInfoCache removes an entry from the info cache. func RemoveFromInfoCache(filename string) { + infoCacheMutex.Lock() + defer infoCacheMutex.Unlock() if !infoCacheEnable { return } @@ -71,19 +108,44 @@ func RemoveFromInfoCache(filename string) { delete(infoCache, filename) } -func GetInfoCache() map[string]interface{} { +// GetInfoCache returns the info cache map. +func GetInfoCache() map[string]*yaml.Node { + infoCacheMutex.Lock() + defer infoCacheMutex.Unlock() if infoCache == nil { initializeInfoCache() } return infoCache } +// ClearFileCache clears the file cache. +func ClearFileCache() { + fileCacheMutex.Lock() + defer fileCacheMutex.Unlock() + fileCache = make(map[string][]byte, 0) +} + +// ClearInfoCache clears the info cache. func ClearInfoCache() { - infoCache = make(map[string]interface{}) + infoCacheMutex.Lock() + defer infoCacheMutex.Unlock() + infoCache = make(map[string]*yaml.Node) +} + +// ClearCaches clears all caches. +func ClearCaches() { + ClearFileCache() + ClearInfoCache() } // FetchFile gets a specified file from the local filesystem or a remote location. func FetchFile(fileurl string) ([]byte, error) { + fileCacheMutex.Lock() + defer fileCacheMutex.Unlock() + return fetchFile(fileurl) +} + +func fetchFile(fileurl string) ([]byte, error) { var bytes []byte initializeFileCache() if fileCacheEnable { @@ -104,7 +166,7 @@ func FetchFile(fileurl string) ([]byte, error) { } defer response.Body.Close() if response.StatusCode != 200 { - return nil, errors.New(fmt.Sprintf("Error downloading %s: %s", fileurl, response.Status)) + return nil, fmt.Errorf("Error downloading %s: %s", fileurl, response.Status) } bytes, err = ioutil.ReadAll(response.Body) if fileCacheEnable && err == nil { @@ -115,11 +177,17 @@ func FetchFile(fileurl string) ([]byte, error) { // ReadBytesForFile reads the bytes of a file. func ReadBytesForFile(filename string) ([]byte, error) { + fileCacheMutex.Lock() + defer fileCacheMutex.Unlock() + return readBytesForFile(filename) +} + +func readBytesForFile(filename string) ([]byte, error) { // is the filename a url? fileurl, _ := url.Parse(filename) if fileurl.Scheme != "" { // yes, fetch it - bytes, err := FetchFile(filename) + bytes, err := fetchFile(filename) if err != nil { return nil, err } @@ -133,8 +201,14 @@ func ReadBytesForFile(filename string) ([]byte, error) { return bytes, nil } -// ReadInfoFromBytes unmarshals a file as a yaml.MapSlice. -func ReadInfoFromBytes(filename string, bytes []byte) (interface{}, error) { +// ReadInfoFromBytes unmarshals a file as a *yaml.Node. +func ReadInfoFromBytes(filename string, bytes []byte) (*yaml.Node, error) { + infoCacheMutex.Lock() + defer infoCacheMutex.Unlock() + return readInfoFromBytes(filename, bytes) +} + +func readInfoFromBytes(filename string, bytes []byte) (*yaml.Node, error) { initializeInfoCache() if infoCacheEnable { cachedInfo, ok := infoCache[filename] @@ -148,19 +222,23 @@ func ReadInfoFromBytes(filename string, bytes []byte) (interface{}, error) { log.Printf("Reading info for file %s", filename) } } - var info yaml.MapSlice + var info yaml.Node err := yaml.Unmarshal(bytes, &info) if err != nil { return nil, err } if infoCacheEnable && len(filename) > 0 { - infoCache[filename] = info + infoCache[filename] = &info } - return info, nil + return &info, nil } // ReadInfoForRef reads a file and return the fragment needed to resolve a $ref. -func ReadInfoForRef(basefile string, ref string) (interface{}, error) { +func ReadInfoForRef(basefile string, ref string) (*yaml.Node, error) { + fileCacheMutex.Lock() + defer fileCacheMutex.Unlock() + infoCacheMutex.Lock() + defer infoCacheMutex.Unlock() initializeInfoCache() if infoCacheEnable { info, ok := infoCache[ref] @@ -174,7 +252,6 @@ func ReadInfoForRef(basefile string, ref string) (interface{}, error) { log.Printf("Reading info for ref %s#%s", basefile, ref) } } - count = count + 1 basedir, _ := filepath.Split(basefile) parts := strings.Split(ref, "#") var filename string @@ -187,24 +264,30 @@ func ReadInfoForRef(basefile string, ref string) (interface{}, error) { } else { filename = basefile } - bytes, err := ReadBytesForFile(filename) + bytes, err := readBytesForFile(filename) if err != nil { return nil, err } - info, err := ReadInfoFromBytes(filename, bytes) + info, err := readInfoFromBytes(filename, bytes) + if info != nil && info.Kind == yaml.DocumentNode { + info = info.Content[0] + } if err != nil { log.Printf("File error: %v\n", err) } else { + if info == nil { + return nil, NewError(nil, fmt.Sprintf("could not resolve %s", ref)) + } if len(parts) > 1 { path := strings.Split(parts[1], "/") for i, key := range path { if i > 0 { - m, ok := info.(yaml.MapSlice) - if ok { + m := info + if true { found := false - for _, section := range m { - if section.Key == key { - info = section.Value + for i := 0; i < len(m.Content); i += 2 { + if m.Content[i].Value == key { + info = m.Content[i+1] found = true } } diff --git a/vendor/github.com/googleapis/gnostic/extensions/README.md b/vendor/github.com/googleapis/gnostic/extensions/README.md index ff1c2eb1..4b5d63e5 100644 --- a/vendor/github.com/googleapis/gnostic/extensions/README.md +++ b/vendor/github.com/googleapis/gnostic/extensions/README.md @@ -1,5 +1,13 @@ # Extensions -This directory contains support code for building Gnostic extensions and associated examples. +**Extension Support is experimental.** -Extensions are used to compile vendor or specification extensions into protocol buffer structures. +This directory contains support code for building Gnostic extensio handlers and +associated examples. + +Extension handlers can be used to compile vendor or specification extensions +into protocol buffer structures. + +Like plugins, extension handlers are built as separate executables. Extension +bodies are written to extension handlers as serialized +ExtensionHandlerRequests. diff --git a/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go b/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go index 055a34e0..6b6a8e28 100644 --- a/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go +++ b/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go @@ -1,148 +1,186 @@ +// Copyright 2017 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + // Code generated by protoc-gen-go. DO NOT EDIT. +// versions: +// protoc-gen-go v1.24.0 +// protoc v3.12.0 // source: extensions/extension.proto -package openapiextension_v1 +package gnostic_extension_v1 import ( - fmt "fmt" proto "github.com/golang/protobuf/proto" any "github.com/golang/protobuf/ptypes/any" - math "math" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + reflect "reflect" + sync "sync" ) -// Reference imports to suppress errors if they are not otherwise used. -var _ = proto.Marshal -var _ = fmt.Errorf -var _ = math.Inf +const ( + // Verify that this generated code is sufficiently up-to-date. + _ = protoimpl.EnforceVersion(20 - protoimpl.MinVersion) + // Verify that runtime/protoimpl is sufficiently up-to-date. + _ = protoimpl.EnforceVersion(protoimpl.MaxVersion - 20) +) -// This is a compile-time assertion to ensure that this generated file -// is compatible with the proto package it is being compiled against. -// A compilation error at this line likely means your copy of the -// proto package needs to be updated. -const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package +// This is a compile-time assertion that a sufficiently up-to-date version +// of the legacy proto package is being used. +const _ = proto.ProtoPackageIsVersion4 -// The version number of OpenAPI compiler. +// The version number of Gnostic. type Version struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + Major int32 `protobuf:"varint,1,opt,name=major,proto3" json:"major,omitempty"` Minor int32 `protobuf:"varint,2,opt,name=minor,proto3" json:"minor,omitempty"` Patch int32 `protobuf:"varint,3,opt,name=patch,proto3" json:"patch,omitempty"` // A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should // be empty for mainline stable releases. - Suffix string `protobuf:"bytes,4,opt,name=suffix,proto3" json:"suffix,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + Suffix string `protobuf:"bytes,4,opt,name=suffix,proto3" json:"suffix,omitempty"` } -func (m *Version) Reset() { *m = Version{} } -func (m *Version) String() string { return proto.CompactTextString(m) } -func (*Version) ProtoMessage() {} -func (*Version) Descriptor() ([]byte, []int) { - return fileDescriptor_661e47e790f76671, []int{0} +func (x *Version) Reset() { + *x = Version{} + if protoimpl.UnsafeEnabled { + mi := &file_extensions_extension_proto_msgTypes[0] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } } -func (m *Version) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Version.Unmarshal(m, b) -} -func (m *Version) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Version.Marshal(b, m, deterministic) -} -func (m *Version) XXX_Merge(src proto.Message) { - xxx_messageInfo_Version.Merge(m, src) +func (x *Version) String() string { + return protoimpl.X.MessageStringOf(x) } -func (m *Version) XXX_Size() int { - return xxx_messageInfo_Version.Size(m) -} -func (m *Version) XXX_DiscardUnknown() { - xxx_messageInfo_Version.DiscardUnknown(m) + +func (*Version) ProtoMessage() {} + +func (x *Version) ProtoReflect() protoreflect.Message { + mi := &file_extensions_extension_proto_msgTypes[0] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) } -var xxx_messageInfo_Version proto.InternalMessageInfo +// Deprecated: Use Version.ProtoReflect.Descriptor instead. +func (*Version) Descriptor() ([]byte, []int) { + return file_extensions_extension_proto_rawDescGZIP(), []int{0} +} -func (m *Version) GetMajor() int32 { - if m != nil { - return m.Major +func (x *Version) GetMajor() int32 { + if x != nil { + return x.Major } return 0 } -func (m *Version) GetMinor() int32 { - if m != nil { - return m.Minor +func (x *Version) GetMinor() int32 { + if x != nil { + return x.Minor } return 0 } -func (m *Version) GetPatch() int32 { - if m != nil { - return m.Patch +func (x *Version) GetPatch() int32 { + if x != nil { + return x.Patch } return 0 } -func (m *Version) GetSuffix() string { - if m != nil { - return m.Suffix +func (x *Version) GetSuffix() string { + if x != nil { + return x.Suffix } return "" } // An encoded Request is written to the ExtensionHandler's stdin. type ExtensionHandlerRequest struct { - // The OpenAPI descriptions that were explicitly listed on the command line. - // The specifications will appear in the order they are specified to gnostic. + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // The extension to process. Wrapper *Wrapper `protobuf:"bytes,1,opt,name=wrapper,proto3" json:"wrapper,omitempty"` - // The version number of openapi compiler. - CompilerVersion *Version `protobuf:"bytes,3,opt,name=compiler_version,json=compilerVersion,proto3" json:"compiler_version,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + // The version number of Gnostic. + CompilerVersion *Version `protobuf:"bytes,2,opt,name=compiler_version,json=compilerVersion,proto3" json:"compiler_version,omitempty"` } -func (m *ExtensionHandlerRequest) Reset() { *m = ExtensionHandlerRequest{} } -func (m *ExtensionHandlerRequest) String() string { return proto.CompactTextString(m) } -func (*ExtensionHandlerRequest) ProtoMessage() {} -func (*ExtensionHandlerRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_661e47e790f76671, []int{1} +func (x *ExtensionHandlerRequest) Reset() { + *x = ExtensionHandlerRequest{} + if protoimpl.UnsafeEnabled { + mi := &file_extensions_extension_proto_msgTypes[1] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } } -func (m *ExtensionHandlerRequest) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ExtensionHandlerRequest.Unmarshal(m, b) -} -func (m *ExtensionHandlerRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ExtensionHandlerRequest.Marshal(b, m, deterministic) -} -func (m *ExtensionHandlerRequest) XXX_Merge(src proto.Message) { - xxx_messageInfo_ExtensionHandlerRequest.Merge(m, src) -} -func (m *ExtensionHandlerRequest) XXX_Size() int { - return xxx_messageInfo_ExtensionHandlerRequest.Size(m) +func (x *ExtensionHandlerRequest) String() string { + return protoimpl.X.MessageStringOf(x) } -func (m *ExtensionHandlerRequest) XXX_DiscardUnknown() { - xxx_messageInfo_ExtensionHandlerRequest.DiscardUnknown(m) + +func (*ExtensionHandlerRequest) ProtoMessage() {} + +func (x *ExtensionHandlerRequest) ProtoReflect() protoreflect.Message { + mi := &file_extensions_extension_proto_msgTypes[1] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) } -var xxx_messageInfo_ExtensionHandlerRequest proto.InternalMessageInfo +// Deprecated: Use ExtensionHandlerRequest.ProtoReflect.Descriptor instead. +func (*ExtensionHandlerRequest) Descriptor() ([]byte, []int) { + return file_extensions_extension_proto_rawDescGZIP(), []int{1} +} -func (m *ExtensionHandlerRequest) GetWrapper() *Wrapper { - if m != nil { - return m.Wrapper +func (x *ExtensionHandlerRequest) GetWrapper() *Wrapper { + if x != nil { + return x.Wrapper } return nil } -func (m *ExtensionHandlerRequest) GetCompilerVersion() *Version { - if m != nil { - return m.CompilerVersion +func (x *ExtensionHandlerRequest) GetCompilerVersion() *Version { + if x != nil { + return x.CompilerVersion } return nil } // The extensions writes an encoded ExtensionHandlerResponse to stdout. type ExtensionHandlerResponse struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + // true if the extension is handled by the extension handler; false otherwise Handled bool `protobuf:"varint,1,opt,name=handled,proto3" json:"handled,omitempty"` - // Error message. If non-empty, the extension handling failed. + // Error message(s). If non-empty, the extension handling failed. // The extension handler process should exit with status code zero // even if it reports an error in this way. // @@ -151,150 +189,277 @@ type ExtensionHandlerResponse struct { // itself -- such as the input Document being unparseable -- should be // reported by writing a message to stderr and exiting with a non-zero // status code. - Error []string `protobuf:"bytes,2,rep,name=error,proto3" json:"error,omitempty"` + Errors []string `protobuf:"bytes,2,rep,name=errors,proto3" json:"errors,omitempty"` // text output - Value *any.Any `protobuf:"bytes,3,opt,name=value,proto3" json:"value,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + Value *any.Any `protobuf:"bytes,3,opt,name=value,proto3" json:"value,omitempty"` } -func (m *ExtensionHandlerResponse) Reset() { *m = ExtensionHandlerResponse{} } -func (m *ExtensionHandlerResponse) String() string { return proto.CompactTextString(m) } -func (*ExtensionHandlerResponse) ProtoMessage() {} -func (*ExtensionHandlerResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_661e47e790f76671, []int{2} +func (x *ExtensionHandlerResponse) Reset() { + *x = ExtensionHandlerResponse{} + if protoimpl.UnsafeEnabled { + mi := &file_extensions_extension_proto_msgTypes[2] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } } -func (m *ExtensionHandlerResponse) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_ExtensionHandlerResponse.Unmarshal(m, b) -} -func (m *ExtensionHandlerResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_ExtensionHandlerResponse.Marshal(b, m, deterministic) -} -func (m *ExtensionHandlerResponse) XXX_Merge(src proto.Message) { - xxx_messageInfo_ExtensionHandlerResponse.Merge(m, src) +func (x *ExtensionHandlerResponse) String() string { + return protoimpl.X.MessageStringOf(x) } -func (m *ExtensionHandlerResponse) XXX_Size() int { - return xxx_messageInfo_ExtensionHandlerResponse.Size(m) -} -func (m *ExtensionHandlerResponse) XXX_DiscardUnknown() { - xxx_messageInfo_ExtensionHandlerResponse.DiscardUnknown(m) + +func (*ExtensionHandlerResponse) ProtoMessage() {} + +func (x *ExtensionHandlerResponse) ProtoReflect() protoreflect.Message { + mi := &file_extensions_extension_proto_msgTypes[2] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) } -var xxx_messageInfo_ExtensionHandlerResponse proto.InternalMessageInfo +// Deprecated: Use ExtensionHandlerResponse.ProtoReflect.Descriptor instead. +func (*ExtensionHandlerResponse) Descriptor() ([]byte, []int) { + return file_extensions_extension_proto_rawDescGZIP(), []int{2} +} -func (m *ExtensionHandlerResponse) GetHandled() bool { - if m != nil { - return m.Handled +func (x *ExtensionHandlerResponse) GetHandled() bool { + if x != nil { + return x.Handled } return false } -func (m *ExtensionHandlerResponse) GetError() []string { - if m != nil { - return m.Error +func (x *ExtensionHandlerResponse) GetErrors() []string { + if x != nil { + return x.Errors } return nil } -func (m *ExtensionHandlerResponse) GetValue() *any.Any { - if m != nil { - return m.Value +func (x *ExtensionHandlerResponse) GetValue() *any.Any { + if x != nil { + return x.Value } return nil } type Wrapper struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + // version of the OpenAPI specification in which this extension was written. Version string `protobuf:"bytes,1,opt,name=version,proto3" json:"version,omitempty"` - // Name of the extension + // Name of the extension. ExtensionName string `protobuf:"bytes,2,opt,name=extension_name,json=extensionName,proto3" json:"extension_name,omitempty"` - // Must be a valid yaml for the proto - Yaml string `protobuf:"bytes,3,opt,name=yaml,proto3" json:"yaml,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + // YAML-formatted extension value. + Yaml string `protobuf:"bytes,3,opt,name=yaml,proto3" json:"yaml,omitempty"` } -func (m *Wrapper) Reset() { *m = Wrapper{} } -func (m *Wrapper) String() string { return proto.CompactTextString(m) } -func (*Wrapper) ProtoMessage() {} -func (*Wrapper) Descriptor() ([]byte, []int) { - return fileDescriptor_661e47e790f76671, []int{3} +func (x *Wrapper) Reset() { + *x = Wrapper{} + if protoimpl.UnsafeEnabled { + mi := &file_extensions_extension_proto_msgTypes[3] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } } -func (m *Wrapper) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Wrapper.Unmarshal(m, b) +func (x *Wrapper) String() string { + return protoimpl.X.MessageStringOf(x) } -func (m *Wrapper) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Wrapper.Marshal(b, m, deterministic) -} -func (m *Wrapper) XXX_Merge(src proto.Message) { - xxx_messageInfo_Wrapper.Merge(m, src) -} -func (m *Wrapper) XXX_Size() int { - return xxx_messageInfo_Wrapper.Size(m) -} -func (m *Wrapper) XXX_DiscardUnknown() { - xxx_messageInfo_Wrapper.DiscardUnknown(m) + +func (*Wrapper) ProtoMessage() {} + +func (x *Wrapper) ProtoReflect() protoreflect.Message { + mi := &file_extensions_extension_proto_msgTypes[3] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) } -var xxx_messageInfo_Wrapper proto.InternalMessageInfo +// Deprecated: Use Wrapper.ProtoReflect.Descriptor instead. +func (*Wrapper) Descriptor() ([]byte, []int) { + return file_extensions_extension_proto_rawDescGZIP(), []int{3} +} -func (m *Wrapper) GetVersion() string { - if m != nil { - return m.Version +func (x *Wrapper) GetVersion() string { + if x != nil { + return x.Version } return "" } -func (m *Wrapper) GetExtensionName() string { - if m != nil { - return m.ExtensionName +func (x *Wrapper) GetExtensionName() string { + if x != nil { + return x.ExtensionName } return "" } -func (m *Wrapper) GetYaml() string { - if m != nil { - return m.Yaml +func (x *Wrapper) GetYaml() string { + if x != nil { + return x.Yaml } return "" } -func init() { - proto.RegisterType((*Version)(nil), "openapiextension.v1.Version") - proto.RegisterType((*ExtensionHandlerRequest)(nil), "openapiextension.v1.ExtensionHandlerRequest") - proto.RegisterType((*ExtensionHandlerResponse)(nil), "openapiextension.v1.ExtensionHandlerResponse") - proto.RegisterType((*Wrapper)(nil), "openapiextension.v1.Wrapper") -} - -func init() { proto.RegisterFile("extensions/extension.proto", fileDescriptor_661e47e790f76671) } - -var fileDescriptor_661e47e790f76671 = []byte{ - // 360 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x09, 0x6e, 0x88, 0x02, 0xff, 0x74, 0x91, 0xdf, 0x4b, 0xeb, 0x30, - 0x1c, 0xc5, 0xe9, 0x7e, 0xf5, 0xee, 0x7b, 0xb9, 0xbb, 0x12, 0x87, 0xd6, 0xe1, 0x83, 0x14, 0x04, - 0x11, 0xe9, 0x98, 0x82, 0xef, 0x1b, 0x0c, 0xf5, 0xc5, 0x8d, 0x3c, 0xcc, 0x37, 0x47, 0xd6, 0x65, - 0x5d, 0xa5, 0x4d, 0x62, 0xfa, 0xc3, 0xed, 0x5f, 0xf1, 0xd1, 0xbf, 0xd4, 0x34, 0x69, 0xeb, 0x83, - 0xfa, 0x96, 0xf3, 0xe1, 0x34, 0x39, 0xe7, 0x14, 0x06, 0x74, 0x97, 0x52, 0x96, 0x84, 0x9c, 0x25, - 0xc3, 0xfa, 0xe8, 0x09, 0xc9, 0x53, 0x8e, 0x0e, 0xb9, 0xa0, 0x8c, 0x88, 0xf0, 0x8b, 0xe7, 0xa3, - 0xc1, 0x49, 0xc0, 0x79, 0x10, 0xd1, 0xa1, 0xb6, 0xac, 0xb2, 0xcd, 0x90, 0xb0, 0xbd, 0xf1, 0xbb, - 0x3e, 0xd8, 0x0b, 0x2a, 0x0b, 0x23, 0xea, 0x43, 0x3b, 0x26, 0x2f, 0x5c, 0x3a, 0xd6, 0x99, 0x75, - 0xd1, 0xc6, 0x46, 0x68, 0x1a, 0x32, 0x45, 0x1b, 0x25, 0x2d, 0x44, 0x41, 0x05, 0x49, 0xfd, 0xad, - 0xd3, 0x34, 0x54, 0x0b, 0x74, 0x04, 0x9d, 0x24, 0xdb, 0x6c, 0xc2, 0x9d, 0xd3, 0x52, 0xb8, 0x8b, - 0x4b, 0xe5, 0xbe, 0x5b, 0x70, 0x3c, 0xad, 0x02, 0xdd, 0x13, 0xb6, 0x8e, 0xa8, 0xc4, 0xf4, 0x35, - 0xa3, 0x49, 0x8a, 0x6e, 0xc1, 0x7e, 0x93, 0x44, 0x08, 0x6a, 0xde, 0xfd, 0x7b, 0x7d, 0xea, 0xfd, - 0x50, 0xc1, 0x7b, 0x32, 0x1e, 0x5c, 0x99, 0xd1, 0x1d, 0x1c, 0xf8, 0x3c, 0x16, 0xa1, 0xba, 0x6a, - 0x99, 0x9b, 0x06, 0x3a, 0xcc, 0x6f, 0x17, 0x94, 0x2d, 0xf1, 0xff, 0xea, 0xab, 0x12, 0xb8, 0x39, - 0x38, 0xdf, 0xb3, 0x25, 0x42, 0x8d, 0x4b, 0x91, 0x03, 0xf6, 0x56, 0xa3, 0xb5, 0x0e, 0xf7, 0x07, - 0x57, 0xb2, 0x18, 0x80, 0x4a, 0xa9, 0x67, 0x69, 0xaa, 0xa6, 0x46, 0xa0, 0x4b, 0x68, 0xe7, 0x24, - 0xca, 0x68, 0x99, 0xa4, 0xef, 0x99, 0xe1, 0xbd, 0x6a, 0x78, 0x6f, 0xcc, 0xf6, 0xd8, 0x58, 0xdc, - 0x67, 0xb0, 0xcb, 0x52, 0xc5, 0x33, 0x55, 0x05, 0x4b, 0x0f, 0x57, 0x49, 0x74, 0x0e, 0xbd, 0xba, - 0xc5, 0x92, 0x91, 0x98, 0xea, 0xdf, 0xd0, 0xc5, 0xff, 0x6a, 0xfa, 0xa8, 0x20, 0x42, 0xd0, 0xda, - 0x93, 0x38, 0xd2, 0xcf, 0x76, 0xb1, 0x3e, 0x4f, 0xae, 0xa0, 0xc7, 0x65, 0xe0, 0x05, 0x8c, 0x27, - 0x69, 0xe8, 0xab, 0x09, 0x26, 0x68, 0xa6, 0x76, 0x19, 0xcf, 0x1f, 0xea, 0xba, 0x8b, 0xd1, 0xdc, - 0xfa, 0x68, 0x34, 0x67, 0xe3, 0xe9, 0xaa, 0xa3, 0x23, 0xde, 0x7c, 0x06, 0x00, 0x00, 0xff, 0xff, - 0xeb, 0xf3, 0xfa, 0x65, 0x5c, 0x02, 0x00, 0x00, +var File_extensions_extension_proto protoreflect.FileDescriptor + +var file_extensions_extension_proto_rawDesc = []byte{ + 0x0a, 0x1a, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x73, 0x2f, 0x65, 0x78, 0x74, + 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x14, 0x67, 0x6e, + 0x6f, 0x73, 0x74, 0x69, 0x63, 0x2e, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x2e, + 0x76, 0x31, 0x1a, 0x19, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, + 0x62, 0x75, 0x66, 0x2f, 0x61, 0x6e, 0x79, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x22, 0x63, 0x0a, + 0x07, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x14, 0x0a, 0x05, 0x6d, 0x61, 0x6a, 0x6f, + 0x72, 0x18, 0x01, 0x20, 0x01, 0x28, 0x05, 0x52, 0x05, 0x6d, 0x61, 0x6a, 0x6f, 0x72, 0x12, 0x14, + 0x0a, 0x05, 0x6d, 0x69, 0x6e, 0x6f, 0x72, 0x18, 0x02, 0x20, 0x01, 0x28, 0x05, 0x52, 0x05, 0x6d, + 0x69, 0x6e, 0x6f, 0x72, 0x12, 0x14, 0x0a, 0x05, 0x70, 0x61, 0x74, 0x63, 0x68, 0x18, 0x03, 0x20, + 0x01, 0x28, 0x05, 0x52, 0x05, 0x70, 0x61, 0x74, 0x63, 0x68, 0x12, 0x16, 0x0a, 0x06, 0x73, 0x75, + 0x66, 0x66, 0x69, 0x78, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x06, 0x73, 0x75, 0x66, 0x66, + 0x69, 0x78, 0x22, 0x9c, 0x01, 0x0a, 0x17, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, + 0x48, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x52, 0x65, 0x71, 0x75, 0x65, 0x73, 0x74, 0x12, 0x37, + 0x0a, 0x07, 0x77, 0x72, 0x61, 0x70, 0x70, 0x65, 0x72, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x1d, 0x2e, 0x67, 0x6e, 0x6f, 0x73, 0x74, 0x69, 0x63, 0x2e, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, + 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x31, 0x2e, 0x57, 0x72, 0x61, 0x70, 0x70, 0x65, 0x72, 0x52, 0x07, + 0x77, 0x72, 0x61, 0x70, 0x70, 0x65, 0x72, 0x12, 0x48, 0x0a, 0x10, 0x63, 0x6f, 0x6d, 0x70, 0x69, + 0x6c, 0x65, 0x72, 0x5f, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x1d, 0x2e, 0x67, 0x6e, 0x6f, 0x73, 0x74, 0x69, 0x63, 0x2e, 0x65, 0x78, 0x74, 0x65, + 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x31, 0x2e, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, + 0x52, 0x0f, 0x63, 0x6f, 0x6d, 0x70, 0x69, 0x6c, 0x65, 0x72, 0x56, 0x65, 0x72, 0x73, 0x69, 0x6f, + 0x6e, 0x22, 0x78, 0x0a, 0x18, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x48, 0x61, + 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x12, 0x18, 0x0a, + 0x07, 0x68, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x64, 0x18, 0x01, 0x20, 0x01, 0x28, 0x08, 0x52, 0x07, + 0x68, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x64, 0x12, 0x16, 0x0a, 0x06, 0x65, 0x72, 0x72, 0x6f, 0x72, + 0x73, 0x18, 0x02, 0x20, 0x03, 0x28, 0x09, 0x52, 0x06, 0x65, 0x72, 0x72, 0x6f, 0x72, 0x73, 0x12, + 0x2a, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x14, + 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, + 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x5e, 0x0a, 0x07, 0x57, + 0x72, 0x61, 0x70, 0x70, 0x65, 0x72, 0x12, 0x18, 0x0a, 0x07, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, + 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e, + 0x12, 0x25, 0x0a, 0x0e, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x5f, 0x6e, 0x61, + 0x6d, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0d, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, + 0x69, 0x6f, 0x6e, 0x4e, 0x61, 0x6d, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x79, 0x61, 0x6d, 0x6c, 0x18, + 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x79, 0x61, 0x6d, 0x6c, 0x42, 0x4b, 0x0a, 0x0e, 0x6f, + 0x72, 0x67, 0x2e, 0x67, 0x6e, 0x6f, 0x73, 0x74, 0x69, 0x63, 0x2e, 0x76, 0x31, 0x42, 0x10, 0x47, + 0x6e, 0x6f, 0x73, 0x74, 0x69, 0x63, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x50, + 0x01, 0x5a, 0x1f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x73, 0x3b, 0x67, 0x6e, + 0x6f, 0x73, 0x74, 0x69, 0x63, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x5f, + 0x76, 0x31, 0xa2, 0x02, 0x03, 0x47, 0x4e, 0x58, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33, +} + +var ( + file_extensions_extension_proto_rawDescOnce sync.Once + file_extensions_extension_proto_rawDescData = file_extensions_extension_proto_rawDesc +) + +func file_extensions_extension_proto_rawDescGZIP() []byte { + file_extensions_extension_proto_rawDescOnce.Do(func() { + file_extensions_extension_proto_rawDescData = protoimpl.X.CompressGZIP(file_extensions_extension_proto_rawDescData) + }) + return file_extensions_extension_proto_rawDescData +} + +var file_extensions_extension_proto_msgTypes = make([]protoimpl.MessageInfo, 4) +var file_extensions_extension_proto_goTypes = []interface{}{ + (*Version)(nil), // 0: gnostic.extension.v1.Version + (*ExtensionHandlerRequest)(nil), // 1: gnostic.extension.v1.ExtensionHandlerRequest + (*ExtensionHandlerResponse)(nil), // 2: gnostic.extension.v1.ExtensionHandlerResponse + (*Wrapper)(nil), // 3: gnostic.extension.v1.Wrapper + (*any.Any)(nil), // 4: google.protobuf.Any +} +var file_extensions_extension_proto_depIdxs = []int32{ + 3, // 0: gnostic.extension.v1.ExtensionHandlerRequest.wrapper:type_name -> gnostic.extension.v1.Wrapper + 0, // 1: gnostic.extension.v1.ExtensionHandlerRequest.compiler_version:type_name -> gnostic.extension.v1.Version + 4, // 2: gnostic.extension.v1.ExtensionHandlerResponse.value:type_name -> google.protobuf.Any + 3, // [3:3] is the sub-list for method output_type + 3, // [3:3] is the sub-list for method input_type + 3, // [3:3] is the sub-list for extension type_name + 3, // [3:3] is the sub-list for extension extendee + 0, // [0:3] is the sub-list for field type_name +} + +func init() { file_extensions_extension_proto_init() } +func file_extensions_extension_proto_init() { + if File_extensions_extension_proto != nil { + return + } + if !protoimpl.UnsafeEnabled { + file_extensions_extension_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Version); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_extensions_extension_proto_msgTypes[1].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ExtensionHandlerRequest); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_extensions_extension_proto_msgTypes[2].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ExtensionHandlerResponse); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_extensions_extension_proto_msgTypes[3].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Wrapper); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + } + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_extensions_extension_proto_rawDesc, + NumEnums: 0, + NumMessages: 4, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_extensions_extension_proto_goTypes, + DependencyIndexes: file_extensions_extension_proto_depIdxs, + MessageInfos: file_extensions_extension_proto_msgTypes, + }.Build() + File_extensions_extension_proto = out.File + file_extensions_extension_proto_rawDesc = nil + file_extensions_extension_proto_goTypes = nil + file_extensions_extension_proto_depIdxs = nil } diff --git a/vendor/github.com/googleapis/gnostic/extensions/extension.proto b/vendor/github.com/googleapis/gnostic/extensions/extension.proto index 04856f91..8ac1faff 100644 --- a/vendor/github.com/googleapis/gnostic/extensions/extension.proto +++ b/vendor/github.com/googleapis/gnostic/extensions/extension.proto @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2017 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -14,8 +14,9 @@ syntax = "proto3"; +package gnostic.extension.v1; + import "google/protobuf/any.proto"; -package openapiextension.v1; // This option lets the proto compiler generate Java code inside the package // name (see below) instead of inside an outer class. It creates a simpler @@ -26,7 +27,7 @@ option java_multiple_files = true; // The Java outer classname should be the filename in UpperCamelCase. This // class is only used to hold proto descriptor, so developers don't need to // work with it directly. -option java_outer_classname = "OpenAPIExtensionV1"; +option java_outer_classname = "GnosticExtension"; // The Java package name must be proto package name with proper prefix. option java_package = "org.gnostic.v1"; @@ -37,9 +38,12 @@ option java_package = "org.gnostic.v1"; // hopefully unique enough to not conflict with things that may come along in // the future. 'GPB' is reserved for the protocol buffer implementation itself. // -option objc_class_prefix = "OAE"; // "OpenAPI Extension" +option objc_class_prefix = "GNX"; // "Gnostic Extension" + +// The Go package name. +option go_package = "extensions;gnostic_extension_v1"; -// The version number of OpenAPI compiler. +// The version number of Gnostic. message Version { int32 major = 1; int32 minor = 2; @@ -52,12 +56,11 @@ message Version { // An encoded Request is written to the ExtensionHandler's stdin. message ExtensionHandlerRequest { - // The OpenAPI descriptions that were explicitly listed on the command line. - // The specifications will appear in the order they are specified to gnostic. + // The extension to process. Wrapper wrapper = 1; - // The version number of openapi compiler. - Version compiler_version = 3; + // The version number of Gnostic. + Version compiler_version = 2; } // The extensions writes an encoded ExtensionHandlerResponse to stdout. @@ -66,7 +69,7 @@ message ExtensionHandlerResponse { // true if the extension is handled by the extension handler; false otherwise bool handled = 1; - // Error message. If non-empty, the extension handling failed. + // Error message(s). If non-empty, the extension handling failed. // The extension handler process should exit with status code zero // even if it reports an error in this way. // @@ -75,7 +78,7 @@ message ExtensionHandlerResponse { // itself -- such as the input Document being unparseable -- should be // reported by writing a message to stderr and exiting with a non-zero // status code. - repeated string error = 2; + repeated string errors = 2; // text output google.protobuf.Any value = 3; @@ -85,9 +88,9 @@ message Wrapper { // version of the OpenAPI specification in which this extension was written. string version = 1; - // Name of the extension + // Name of the extension. string extension_name = 2; - // Must be a valid yaml for the proto + // YAML-formatted extension value. string yaml = 3; } diff --git a/vendor/github.com/googleapis/gnostic/extensions/extensions.go b/vendor/github.com/googleapis/gnostic/extensions/extensions.go index 94a8e62a..ec8afd00 100644 --- a/vendor/github.com/googleapis/gnostic/extensions/extensions.go +++ b/vendor/github.com/googleapis/gnostic/extensions/extensions.go @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2017 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -12,71 +12,53 @@ // See the License for the specific language governing permissions and // limitations under the License. -package openapiextension_v1 +package gnostic_extension_v1 import ( - "fmt" "io/ioutil" + "log" "os" "github.com/golang/protobuf/proto" "github.com/golang/protobuf/ptypes" ) -type documentHandler func(version string, extensionName string, document string) type extensionHandler func(name string, yamlInput string) (bool, proto.Message, error) -func forInputYamlFromOpenapic(handler documentHandler) { +// Main implements the main program of an extension handler. +func Main(handler extensionHandler) { + // unpack the request data, err := ioutil.ReadAll(os.Stdin) if err != nil { - fmt.Println("File error:", err.Error()) + log.Println("File error:", err.Error()) os.Exit(1) } if len(data) == 0 { - fmt.Println("No input data.") + log.Println("No input data.") os.Exit(1) } request := &ExtensionHandlerRequest{} err = proto.Unmarshal(data, request) if err != nil { - fmt.Println("Input error:", err.Error()) + log.Println("Input error:", err.Error()) os.Exit(1) } - handler(request.Wrapper.Version, request.Wrapper.ExtensionName, request.Wrapper.Yaml) -} - -// ProcessExtension calles the handler for a specified extension. -func ProcessExtension(handleExtension extensionHandler) { - response := &ExtensionHandlerResponse{} - forInputYamlFromOpenapic( - func(version string, extensionName string, yamlInput string) { - var newObject proto.Message - var err error - - handled, newObject, err := handleExtension(extensionName, yamlInput) - if !handled { - responseBytes, _ := proto.Marshal(response) - os.Stdout.Write(responseBytes) - os.Exit(0) - } - - // If we reach here, then the extension is handled - response.Handled = true - if err != nil { - response.Error = append(response.Error, err.Error()) - responseBytes, _ := proto.Marshal(response) - os.Stdout.Write(responseBytes) - os.Exit(0) - } - response.Value, err = ptypes.MarshalAny(newObject) - if err != nil { - response.Error = append(response.Error, err.Error()) - responseBytes, _ := proto.Marshal(response) - os.Stdout.Write(responseBytes) - os.Exit(0) - } - }) - + // call the handler + handled, output, err := handler(request.Wrapper.ExtensionName, request.Wrapper.Yaml) + // respond with the output of the handler + response := &ExtensionHandlerResponse{ + Handled: false, // default assumption + Errors: make([]string, 0), + } + if err != nil { + response.Errors = append(response.Errors, err.Error()) + } else if handled { + response.Handled = true + response.Value, err = ptypes.MarshalAny(output) + if err != nil { + response.Errors = append(response.Errors, err.Error()) + } + } responseBytes, _ := proto.Marshal(response) os.Stdout.Write(responseBytes) } diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/README.md b/vendor/github.com/googleapis/gnostic/jsonschema/README.md new file mode 100644 index 00000000..6793c517 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/README.md @@ -0,0 +1,4 @@ +# jsonschema + +This directory contains code for reading, writing, and manipulating JSON +schemas. diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/base.go b/vendor/github.com/googleapis/gnostic/jsonschema/base.go new file mode 100644 index 00000000..0af8b148 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/base.go @@ -0,0 +1,84 @@ + +// THIS FILE IS AUTOMATICALLY GENERATED. + +package jsonschema + +import ( + "encoding/base64" +) + +func baseSchemaBytes() ([]byte, error){ + return base64.StdEncoding.DecodeString( +`ewogICAgImlkIjogImh0dHA6Ly9qc29uLXNjaGVtYS5vcmcvZHJhZnQtMDQvc2NoZW1hIyIsCiAgICAi +JHNjaGVtYSI6ICJodHRwOi8vanNvbi1zY2hlbWEub3JnL2RyYWZ0LTA0L3NjaGVtYSMiLAogICAgImRl +c2NyaXB0aW9uIjogIkNvcmUgc2NoZW1hIG1ldGEtc2NoZW1hIiwKICAgICJkZWZpbml0aW9ucyI6IHsK +ICAgICAgICAic2NoZW1hQXJyYXkiOiB7CiAgICAgICAgICAgICJ0eXBlIjogImFycmF5IiwKICAgICAg +ICAgICAgIm1pbkl0ZW1zIjogMSwKICAgICAgICAgICAgIml0ZW1zIjogeyAiJHJlZiI6ICIjIiB9CiAg +ICAgICAgfSwKICAgICAgICAicG9zaXRpdmVJbnRlZ2VyIjogewogICAgICAgICAgICAidHlwZSI6ICJp +bnRlZ2VyIiwKICAgICAgICAgICAgIm1pbmltdW0iOiAwCiAgICAgICAgfSwKICAgICAgICAicG9zaXRp +dmVJbnRlZ2VyRGVmYXVsdDAiOiB7CiAgICAgICAgICAgICJhbGxPZiI6IFsgeyAiJHJlZiI6ICIjL2Rl +ZmluaXRpb25zL3Bvc2l0aXZlSW50ZWdlciIgfSwgeyAiZGVmYXVsdCI6IDAgfSBdCiAgICAgICAgfSwK +ICAgICAgICAic2ltcGxlVHlwZXMiOiB7CiAgICAgICAgICAgICJlbnVtIjogWyAiYXJyYXkiLCAiYm9v +bGVhbiIsICJpbnRlZ2VyIiwgIm51bGwiLCAibnVtYmVyIiwgIm9iamVjdCIsICJzdHJpbmciIF0KICAg +ICAgICB9LAogICAgICAgICJzdHJpbmdBcnJheSI6IHsKICAgICAgICAgICAgInR5cGUiOiAiYXJyYXki +LAogICAgICAgICAgICAiaXRlbXMiOiB7ICJ0eXBlIjogInN0cmluZyIgfSwKICAgICAgICAgICAgIm1p +bkl0ZW1zIjogMSwKICAgICAgICAgICAgInVuaXF1ZUl0ZW1zIjogdHJ1ZQogICAgICAgIH0KICAgIH0s +CiAgICAidHlwZSI6ICJvYmplY3QiLAogICAgInByb3BlcnRpZXMiOiB7CiAgICAgICAgImlkIjogewog +ICAgICAgICAgICAidHlwZSI6ICJzdHJpbmciLAogICAgICAgICAgICAiZm9ybWF0IjogInVyaSIKICAg +ICAgICB9LAogICAgICAgICIkc2NoZW1hIjogewogICAgICAgICAgICAidHlwZSI6ICJzdHJpbmciLAog +ICAgICAgICAgICAiZm9ybWF0IjogInVyaSIKICAgICAgICB9LAogICAgICAgICJ0aXRsZSI6IHsKICAg +ICAgICAgICAgInR5cGUiOiAic3RyaW5nIgogICAgICAgIH0sCiAgICAgICAgImRlc2NyaXB0aW9uIjog +ewogICAgICAgICAgICAidHlwZSI6ICJzdHJpbmciCiAgICAgICAgfSwKICAgICAgICAiZGVmYXVsdCI6 +IHt9LAogICAgICAgICJtdWx0aXBsZU9mIjogewogICAgICAgICAgICAidHlwZSI6ICJudW1iZXIiLAog +ICAgICAgICAgICAibWluaW11bSI6IDAsCiAgICAgICAgICAgICJleGNsdXNpdmVNaW5pbXVtIjogdHJ1 +ZQogICAgICAgIH0sCiAgICAgICAgIm1heGltdW0iOiB7CiAgICAgICAgICAgICJ0eXBlIjogIm51bWJl +ciIKICAgICAgICB9LAogICAgICAgICJleGNsdXNpdmVNYXhpbXVtIjogewogICAgICAgICAgICAidHlw +ZSI6ICJib29sZWFuIiwKICAgICAgICAgICAgImRlZmF1bHQiOiBmYWxzZQogICAgICAgIH0sCiAgICAg +ICAgIm1pbmltdW0iOiB7CiAgICAgICAgICAgICJ0eXBlIjogIm51bWJlciIKICAgICAgICB9LAogICAg +ICAgICJleGNsdXNpdmVNaW5pbXVtIjogewogICAgICAgICAgICAidHlwZSI6ICJib29sZWFuIiwKICAg +ICAgICAgICAgImRlZmF1bHQiOiBmYWxzZQogICAgICAgIH0sCiAgICAgICAgIm1heExlbmd0aCI6IHsg +IiRyZWYiOiAiIy9kZWZpbml0aW9ucy9wb3NpdGl2ZUludGVnZXIiIH0sCiAgICAgICAgIm1pbkxlbmd0 +aCI6IHsgIiRyZWYiOiAiIy9kZWZpbml0aW9ucy9wb3NpdGl2ZUludGVnZXJEZWZhdWx0MCIgfSwKICAg +ICAgICAicGF0dGVybiI6IHsKICAgICAgICAgICAgInR5cGUiOiAic3RyaW5nIiwKICAgICAgICAgICAg +ImZvcm1hdCI6ICJyZWdleCIKICAgICAgICB9LAogICAgICAgICJhZGRpdGlvbmFsSXRlbXMiOiB7CiAg +ICAgICAgICAgICJhbnlPZiI6IFsKICAgICAgICAgICAgICAgIHsgInR5cGUiOiAiYm9vbGVhbiIgfSwK +ICAgICAgICAgICAgICAgIHsgIiRyZWYiOiAiIyIgfQogICAgICAgICAgICBdLAogICAgICAgICAgICAi +ZGVmYXVsdCI6IHt9CiAgICAgICAgfSwKICAgICAgICAiaXRlbXMiOiB7CiAgICAgICAgICAgICJhbnlP +ZiI6IFsKICAgICAgICAgICAgICAgIHsgIiRyZWYiOiAiIyIgfSwKICAgICAgICAgICAgICAgIHsgIiRy +ZWYiOiAiIy9kZWZpbml0aW9ucy9zY2hlbWFBcnJheSIgfQogICAgICAgICAgICBdLAogICAgICAgICAg +ICAiZGVmYXVsdCI6IHt9CiAgICAgICAgfSwKICAgICAgICAibWF4SXRlbXMiOiB7ICIkcmVmIjogIiMv +ZGVmaW5pdGlvbnMvcG9zaXRpdmVJbnRlZ2VyIiB9LAogICAgICAgICJtaW5JdGVtcyI6IHsgIiRyZWYi +OiAiIy9kZWZpbml0aW9ucy9wb3NpdGl2ZUludGVnZXJEZWZhdWx0MCIgfSwKICAgICAgICAidW5pcXVl +SXRlbXMiOiB7CiAgICAgICAgICAgICJ0eXBlIjogImJvb2xlYW4iLAogICAgICAgICAgICAiZGVmYXVs +dCI6IGZhbHNlCiAgICAgICAgfSwKICAgICAgICAibWF4UHJvcGVydGllcyI6IHsgIiRyZWYiOiAiIy9k +ZWZpbml0aW9ucy9wb3NpdGl2ZUludGVnZXIiIH0sCiAgICAgICAgIm1pblByb3BlcnRpZXMiOiB7ICIk +cmVmIjogIiMvZGVmaW5pdGlvbnMvcG9zaXRpdmVJbnRlZ2VyRGVmYXVsdDAiIH0sCiAgICAgICAgInJl +cXVpcmVkIjogeyAiJHJlZiI6ICIjL2RlZmluaXRpb25zL3N0cmluZ0FycmF5IiB9LAogICAgICAgICJh +ZGRpdGlvbmFsUHJvcGVydGllcyI6IHsKICAgICAgICAgICAgImFueU9mIjogWwogICAgICAgICAgICAg +ICAgeyAidHlwZSI6ICJib29sZWFuIiB9LAogICAgICAgICAgICAgICAgeyAiJHJlZiI6ICIjIiB9CiAg +ICAgICAgICAgIF0sCiAgICAgICAgICAgICJkZWZhdWx0Ijoge30KICAgICAgICB9LAogICAgICAgICJk +ZWZpbml0aW9ucyI6IHsKICAgICAgICAgICAgInR5cGUiOiAib2JqZWN0IiwKICAgICAgICAgICAgImFk +ZGl0aW9uYWxQcm9wZXJ0aWVzIjogeyAiJHJlZiI6ICIjIiB9LAogICAgICAgICAgICAiZGVmYXVsdCI6 +IHt9CiAgICAgICAgfSwKICAgICAgICAicHJvcGVydGllcyI6IHsKICAgICAgICAgICAgInR5cGUiOiAi +b2JqZWN0IiwKICAgICAgICAgICAgImFkZGl0aW9uYWxQcm9wZXJ0aWVzIjogeyAiJHJlZiI6ICIjIiB9 +LAogICAgICAgICAgICAiZGVmYXVsdCI6IHt9CiAgICAgICAgfSwKICAgICAgICAicGF0dGVyblByb3Bl +cnRpZXMiOiB7CiAgICAgICAgICAgICJ0eXBlIjogIm9iamVjdCIsCiAgICAgICAgICAgICJhZGRpdGlv +bmFsUHJvcGVydGllcyI6IHsgIiRyZWYiOiAiIyIgfSwKICAgICAgICAgICAgImRlZmF1bHQiOiB7fQog +ICAgICAgIH0sCiAgICAgICAgImRlcGVuZGVuY2llcyI6IHsKICAgICAgICAgICAgInR5cGUiOiAib2Jq +ZWN0IiwKICAgICAgICAgICAgImFkZGl0aW9uYWxQcm9wZXJ0aWVzIjogewogICAgICAgICAgICAgICAg +ImFueU9mIjogWwogICAgICAgICAgICAgICAgICAgIHsgIiRyZWYiOiAiIyIgfSwKICAgICAgICAgICAg +ICAgICAgICB7ICIkcmVmIjogIiMvZGVmaW5pdGlvbnMvc3RyaW5nQXJyYXkiIH0KICAgICAgICAgICAg +ICAgIF0KICAgICAgICAgICAgfQogICAgICAgIH0sCiAgICAgICAgImVudW0iOiB7CiAgICAgICAgICAg +ICJ0eXBlIjogImFycmF5IiwKICAgICAgICAgICAgIm1pbkl0ZW1zIjogMSwKICAgICAgICAgICAgInVu +aXF1ZUl0ZW1zIjogdHJ1ZQogICAgICAgIH0sCiAgICAgICAgInR5cGUiOiB7CiAgICAgICAgICAgICJh +bnlPZiI6IFsKICAgICAgICAgICAgICAgIHsgIiRyZWYiOiAiIy9kZWZpbml0aW9ucy9zaW1wbGVUeXBl +cyIgfSwKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAidHlwZSI6ICJhcnJheSIs +CiAgICAgICAgICAgICAgICAgICAgIml0ZW1zIjogeyAiJHJlZiI6ICIjL2RlZmluaXRpb25zL3NpbXBs +ZVR5cGVzIiB9LAogICAgICAgICAgICAgICAgICAgICJtaW5JdGVtcyI6IDEsCiAgICAgICAgICAgICAg +ICAgICAgInVuaXF1ZUl0ZW1zIjogdHJ1ZQogICAgICAgICAgICAgICAgfQogICAgICAgICAgICBdCiAg +ICAgICAgfSwKICAgICAgICAiYWxsT2YiOiB7ICIkcmVmIjogIiMvZGVmaW5pdGlvbnMvc2NoZW1hQXJy +YXkiIH0sCiAgICAgICAgImFueU9mIjogeyAiJHJlZiI6ICIjL2RlZmluaXRpb25zL3NjaGVtYUFycmF5 +IiB9LAogICAgICAgICJvbmVPZiI6IHsgIiRyZWYiOiAiIy9kZWZpbml0aW9ucy9zY2hlbWFBcnJheSIg +fSwKICAgICAgICAibm90IjogeyAiJHJlZiI6ICIjIiB9CiAgICB9LAogICAgImRlcGVuZGVuY2llcyI6 +IHsKICAgICAgICAiZXhjbHVzaXZlTWF4aW11bSI6IFsgIm1heGltdW0iIF0sCiAgICAgICAgImV4Y2x1 +c2l2ZU1pbmltdW0iOiBbICJtaW5pbXVtIiBdCiAgICB9LAogICAgImRlZmF1bHQiOiB7fQp9Cg==`)} \ No newline at end of file diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/display.go b/vendor/github.com/googleapis/gnostic/jsonschema/display.go new file mode 100644 index 00000000..028a760a --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/display.go @@ -0,0 +1,229 @@ +// Copyright 2017 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package jsonschema + +import ( + "fmt" + "strings" +) + +// +// DISPLAY +// The following methods display Schemas. +// + +// Description returns a string representation of a string or string array. +func (s *StringOrStringArray) Description() string { + if s.String != nil { + return *s.String + } + if s.StringArray != nil { + return strings.Join(*s.StringArray, ", ") + } + return "" +} + +// Returns a string representation of a Schema. +func (schema *Schema) String() string { + return schema.describeSchema("") +} + +// Helper: Returns a string representation of a Schema indented by a specified string. +func (schema *Schema) describeSchema(indent string) string { + result := "" + if schema.Schema != nil { + result += indent + "$schema: " + *(schema.Schema) + "\n" + } + if schema.ID != nil { + result += indent + "id: " + *(schema.ID) + "\n" + } + if schema.MultipleOf != nil { + result += indent + fmt.Sprintf("multipleOf: %+v\n", *(schema.MultipleOf)) + } + if schema.Maximum != nil { + result += indent + fmt.Sprintf("maximum: %+v\n", *(schema.Maximum)) + } + if schema.ExclusiveMaximum != nil { + result += indent + fmt.Sprintf("exclusiveMaximum: %+v\n", *(schema.ExclusiveMaximum)) + } + if schema.Minimum != nil { + result += indent + fmt.Sprintf("minimum: %+v\n", *(schema.Minimum)) + } + if schema.ExclusiveMinimum != nil { + result += indent + fmt.Sprintf("exclusiveMinimum: %+v\n", *(schema.ExclusiveMinimum)) + } + if schema.MaxLength != nil { + result += indent + fmt.Sprintf("maxLength: %+v\n", *(schema.MaxLength)) + } + if schema.MinLength != nil { + result += indent + fmt.Sprintf("minLength: %+v\n", *(schema.MinLength)) + } + if schema.Pattern != nil { + result += indent + fmt.Sprintf("pattern: %+v\n", *(schema.Pattern)) + } + if schema.AdditionalItems != nil { + s := schema.AdditionalItems.Schema + if s != nil { + result += indent + "additionalItems:\n" + result += s.describeSchema(indent + " ") + } else { + b := *(schema.AdditionalItems.Boolean) + result += indent + fmt.Sprintf("additionalItems: %+v\n", b) + } + } + if schema.Items != nil { + result += indent + "items:\n" + items := schema.Items + if items.SchemaArray != nil { + for i, s := range *(items.SchemaArray) { + result += indent + " " + fmt.Sprintf("%d", i) + ":\n" + result += s.describeSchema(indent + " " + " ") + } + } else if items.Schema != nil { + result += items.Schema.describeSchema(indent + " " + " ") + } + } + if schema.MaxItems != nil { + result += indent + fmt.Sprintf("maxItems: %+v\n", *(schema.MaxItems)) + } + if schema.MinItems != nil { + result += indent + fmt.Sprintf("minItems: %+v\n", *(schema.MinItems)) + } + if schema.UniqueItems != nil { + result += indent + fmt.Sprintf("uniqueItems: %+v\n", *(schema.UniqueItems)) + } + if schema.MaxProperties != nil { + result += indent + fmt.Sprintf("maxProperties: %+v\n", *(schema.MaxProperties)) + } + if schema.MinProperties != nil { + result += indent + fmt.Sprintf("minProperties: %+v\n", *(schema.MinProperties)) + } + if schema.Required != nil { + result += indent + fmt.Sprintf("required: %+v\n", *(schema.Required)) + } + if schema.AdditionalProperties != nil { + s := schema.AdditionalProperties.Schema + if s != nil { + result += indent + "additionalProperties:\n" + result += s.describeSchema(indent + " ") + } else { + b := *(schema.AdditionalProperties.Boolean) + result += indent + fmt.Sprintf("additionalProperties: %+v\n", b) + } + } + if schema.Properties != nil { + result += indent + "properties:\n" + for _, pair := range *(schema.Properties) { + name := pair.Name + s := pair.Value + result += indent + " " + name + ":\n" + result += s.describeSchema(indent + " " + " ") + } + } + if schema.PatternProperties != nil { + result += indent + "patternProperties:\n" + for _, pair := range *(schema.PatternProperties) { + name := pair.Name + s := pair.Value + result += indent + " " + name + ":\n" + result += s.describeSchema(indent + " " + " ") + } + } + if schema.Dependencies != nil { + result += indent + "dependencies:\n" + for _, pair := range *(schema.Dependencies) { + name := pair.Name + schemaOrStringArray := pair.Value + s := schemaOrStringArray.Schema + if s != nil { + result += indent + " " + name + ":\n" + result += s.describeSchema(indent + " " + " ") + } else { + a := schemaOrStringArray.StringArray + if a != nil { + result += indent + " " + name + ":\n" + for _, s2 := range *a { + result += indent + " " + " " + s2 + "\n" + } + } + } + + } + } + if schema.Enumeration != nil { + result += indent + "enumeration:\n" + for _, value := range *(schema.Enumeration) { + if value.String != nil { + result += indent + " " + fmt.Sprintf("%+v\n", *value.String) + } else { + result += indent + " " + fmt.Sprintf("%+v\n", *value.Bool) + } + } + } + if schema.Type != nil { + result += indent + fmt.Sprintf("type: %+v\n", schema.Type.Description()) + } + if schema.AllOf != nil { + result += indent + "allOf:\n" + for _, s := range *(schema.AllOf) { + result += s.describeSchema(indent + " ") + result += indent + "-\n" + } + } + if schema.AnyOf != nil { + result += indent + "anyOf:\n" + for _, s := range *(schema.AnyOf) { + result += s.describeSchema(indent + " ") + result += indent + "-\n" + } + } + if schema.OneOf != nil { + result += indent + "oneOf:\n" + for _, s := range *(schema.OneOf) { + result += s.describeSchema(indent + " ") + result += indent + "-\n" + } + } + if schema.Not != nil { + result += indent + "not:\n" + result += schema.Not.describeSchema(indent + " ") + } + if schema.Definitions != nil { + result += indent + "definitions:\n" + for _, pair := range *(schema.Definitions) { + name := pair.Name + s := pair.Value + result += indent + " " + name + ":\n" + result += s.describeSchema(indent + " " + " ") + } + } + if schema.Title != nil { + result += indent + "title: " + *(schema.Title) + "\n" + } + if schema.Description != nil { + result += indent + "description: " + *(schema.Description) + "\n" + } + if schema.Default != nil { + result += indent + "default:\n" + result += indent + fmt.Sprintf(" %+v\n", *(schema.Default)) + } + if schema.Format != nil { + result += indent + "format: " + *(schema.Format) + "\n" + } + if schema.Ref != nil { + result += indent + "$ref: " + *(schema.Ref) + "\n" + } + return result +} diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/models.go b/vendor/github.com/googleapis/gnostic/jsonschema/models.go new file mode 100644 index 00000000..4781bdc5 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/models.go @@ -0,0 +1,228 @@ +// Copyright 2017 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Package jsonschema supports the reading, writing, and manipulation +// of JSON Schemas. +package jsonschema + +import "gopkg.in/yaml.v3" + +// The Schema struct models a JSON Schema and, because schemas are +// defined hierarchically, contains many references to itself. +// All fields are pointers and are nil if the associated values +// are not specified. +type Schema struct { + Schema *string // $schema + ID *string // id keyword used for $ref resolution scope + Ref *string // $ref, i.e. JSON Pointers + + // http://json-schema.org/latest/json-schema-validation.html + // 5.1. Validation keywords for numeric instances (number and integer) + MultipleOf *SchemaNumber + Maximum *SchemaNumber + ExclusiveMaximum *bool + Minimum *SchemaNumber + ExclusiveMinimum *bool + + // 5.2. Validation keywords for strings + MaxLength *int64 + MinLength *int64 + Pattern *string + + // 5.3. Validation keywords for arrays + AdditionalItems *SchemaOrBoolean + Items *SchemaOrSchemaArray + MaxItems *int64 + MinItems *int64 + UniqueItems *bool + + // 5.4. Validation keywords for objects + MaxProperties *int64 + MinProperties *int64 + Required *[]string + AdditionalProperties *SchemaOrBoolean + Properties *[]*NamedSchema + PatternProperties *[]*NamedSchema + Dependencies *[]*NamedSchemaOrStringArray + + // 5.5. Validation keywords for any instance type + Enumeration *[]SchemaEnumValue + Type *StringOrStringArray + AllOf *[]*Schema + AnyOf *[]*Schema + OneOf *[]*Schema + Not *Schema + Definitions *[]*NamedSchema + + // 6. Metadata keywords + Title *string + Description *string + Default *yaml.Node + + // 7. Semantic validation with "format" + Format *string +} + +// These helper structs represent "combination" types that generally can +// have values of one type or another. All are used to represent parts +// of Schemas. + +// SchemaNumber represents a value that can be either an Integer or a Float. +type SchemaNumber struct { + Integer *int64 + Float *float64 +} + +// NewSchemaNumberWithInteger creates and returns a new object +func NewSchemaNumberWithInteger(i int64) *SchemaNumber { + result := &SchemaNumber{} + result.Integer = &i + return result +} + +// NewSchemaNumberWithFloat creates and returns a new object +func NewSchemaNumberWithFloat(f float64) *SchemaNumber { + result := &SchemaNumber{} + result.Float = &f + return result +} + +// SchemaOrBoolean represents a value that can be either a Schema or a Boolean. +type SchemaOrBoolean struct { + Schema *Schema + Boolean *bool +} + +// NewSchemaOrBooleanWithSchema creates and returns a new object +func NewSchemaOrBooleanWithSchema(s *Schema) *SchemaOrBoolean { + result := &SchemaOrBoolean{} + result.Schema = s + return result +} + +// NewSchemaOrBooleanWithBoolean creates and returns a new object +func NewSchemaOrBooleanWithBoolean(b bool) *SchemaOrBoolean { + result := &SchemaOrBoolean{} + result.Boolean = &b + return result +} + +// StringOrStringArray represents a value that can be either +// a String or an Array of Strings. +type StringOrStringArray struct { + String *string + StringArray *[]string +} + +// NewStringOrStringArrayWithString creates and returns a new object +func NewStringOrStringArrayWithString(s string) *StringOrStringArray { + result := &StringOrStringArray{} + result.String = &s + return result +} + +// NewStringOrStringArrayWithStringArray creates and returns a new object +func NewStringOrStringArrayWithStringArray(a []string) *StringOrStringArray { + result := &StringOrStringArray{} + result.StringArray = &a + return result +} + +// SchemaOrStringArray represents a value that can be either +// a Schema or an Array of Strings. +type SchemaOrStringArray struct { + Schema *Schema + StringArray *[]string +} + +// SchemaOrSchemaArray represents a value that can be either +// a Schema or an Array of Schemas. +type SchemaOrSchemaArray struct { + Schema *Schema + SchemaArray *[]*Schema +} + +// NewSchemaOrSchemaArrayWithSchema creates and returns a new object +func NewSchemaOrSchemaArrayWithSchema(s *Schema) *SchemaOrSchemaArray { + result := &SchemaOrSchemaArray{} + result.Schema = s + return result +} + +// NewSchemaOrSchemaArrayWithSchemaArray creates and returns a new object +func NewSchemaOrSchemaArrayWithSchemaArray(a []*Schema) *SchemaOrSchemaArray { + result := &SchemaOrSchemaArray{} + result.SchemaArray = &a + return result +} + +// SchemaEnumValue represents a value that can be part of an +// enumeration in a Schema. +type SchemaEnumValue struct { + String *string + Bool *bool +} + +// NamedSchema is a name-value pair that is used to emulate maps +// with ordered keys. +type NamedSchema struct { + Name string + Value *Schema +} + +// NewNamedSchema creates and returns a new object +func NewNamedSchema(name string, value *Schema) *NamedSchema { + return &NamedSchema{Name: name, Value: value} +} + +// NamedSchemaOrStringArray is a name-value pair that is used +// to emulate maps with ordered keys. +type NamedSchemaOrStringArray struct { + Name string + Value *SchemaOrStringArray +} + +// Access named subschemas by name + +func namedSchemaArrayElementWithName(array *[]*NamedSchema, name string) *Schema { + if array == nil { + return nil + } + for _, pair := range *array { + if pair.Name == name { + return pair.Value + } + } + return nil +} + +// PropertyWithName returns the selected element. +func (s *Schema) PropertyWithName(name string) *Schema { + return namedSchemaArrayElementWithName(s.Properties, name) +} + +// PatternPropertyWithName returns the selected element. +func (s *Schema) PatternPropertyWithName(name string) *Schema { + return namedSchemaArrayElementWithName(s.PatternProperties, name) +} + +// DefinitionWithName returns the selected element. +func (s *Schema) DefinitionWithName(name string) *Schema { + return namedSchemaArrayElementWithName(s.Definitions, name) +} + +// AddProperty adds a named property. +func (s *Schema) AddProperty(name string, property *Schema) { + *s.Properties = append(*s.Properties, NewNamedSchema(name, property)) +} diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/operations.go b/vendor/github.com/googleapis/gnostic/jsonschema/operations.go new file mode 100644 index 00000000..ba8dd4a9 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/operations.go @@ -0,0 +1,394 @@ +// Copyright 2017 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package jsonschema + +import ( + "fmt" + "log" + "strings" +) + +// +// OPERATIONS +// The following methods perform operations on Schemas. +// + +// IsEmpty returns true if no members of the Schema are specified. +func (schema *Schema) IsEmpty() bool { + return (schema.Schema == nil) && + (schema.ID == nil) && + (schema.MultipleOf == nil) && + (schema.Maximum == nil) && + (schema.ExclusiveMaximum == nil) && + (schema.Minimum == nil) && + (schema.ExclusiveMinimum == nil) && + (schema.MaxLength == nil) && + (schema.MinLength == nil) && + (schema.Pattern == nil) && + (schema.AdditionalItems == nil) && + (schema.Items == nil) && + (schema.MaxItems == nil) && + (schema.MinItems == nil) && + (schema.UniqueItems == nil) && + (schema.MaxProperties == nil) && + (schema.MinProperties == nil) && + (schema.Required == nil) && + (schema.AdditionalProperties == nil) && + (schema.Properties == nil) && + (schema.PatternProperties == nil) && + (schema.Dependencies == nil) && + (schema.Enumeration == nil) && + (schema.Type == nil) && + (schema.AllOf == nil) && + (schema.AnyOf == nil) && + (schema.OneOf == nil) && + (schema.Not == nil) && + (schema.Definitions == nil) && + (schema.Title == nil) && + (schema.Description == nil) && + (schema.Default == nil) && + (schema.Format == nil) && + (schema.Ref == nil) +} + +// IsEqual returns true if two schemas are equal. +func (schema *Schema) IsEqual(schema2 *Schema) bool { + return schema.String() == schema2.String() +} + +// SchemaOperation represents a function that can be applied to a Schema. +type SchemaOperation func(schema *Schema, context string) + +// Applies a specified function to a Schema and all of the Schemas that it contains. +func (schema *Schema) applyToSchemas(operation SchemaOperation, context string) { + + if schema.AdditionalItems != nil { + s := schema.AdditionalItems.Schema + if s != nil { + s.applyToSchemas(operation, "AdditionalItems") + } + } + + if schema.Items != nil { + if schema.Items.SchemaArray != nil { + for _, s := range *(schema.Items.SchemaArray) { + s.applyToSchemas(operation, "Items.SchemaArray") + } + } else if schema.Items.Schema != nil { + schema.Items.Schema.applyToSchemas(operation, "Items.Schema") + } + } + + if schema.AdditionalProperties != nil { + s := schema.AdditionalProperties.Schema + if s != nil { + s.applyToSchemas(operation, "AdditionalProperties") + } + } + + if schema.Properties != nil { + for _, pair := range *(schema.Properties) { + s := pair.Value + s.applyToSchemas(operation, "Properties") + } + } + if schema.PatternProperties != nil { + for _, pair := range *(schema.PatternProperties) { + s := pair.Value + s.applyToSchemas(operation, "PatternProperties") + } + } + + if schema.Dependencies != nil { + for _, pair := range *(schema.Dependencies) { + schemaOrStringArray := pair.Value + s := schemaOrStringArray.Schema + if s != nil { + s.applyToSchemas(operation, "Dependencies") + } + } + } + + if schema.AllOf != nil { + for _, s := range *(schema.AllOf) { + s.applyToSchemas(operation, "AllOf") + } + } + if schema.AnyOf != nil { + for _, s := range *(schema.AnyOf) { + s.applyToSchemas(operation, "AnyOf") + } + } + if schema.OneOf != nil { + for _, s := range *(schema.OneOf) { + s.applyToSchemas(operation, "OneOf") + } + } + if schema.Not != nil { + schema.Not.applyToSchemas(operation, "Not") + } + + if schema.Definitions != nil { + for _, pair := range *(schema.Definitions) { + s := pair.Value + s.applyToSchemas(operation, "Definitions") + } + } + + operation(schema, context) +} + +// CopyProperties copies all non-nil properties from the source Schema to the schema Schema. +func (schema *Schema) CopyProperties(source *Schema) { + if source.Schema != nil { + schema.Schema = source.Schema + } + if source.ID != nil { + schema.ID = source.ID + } + if source.MultipleOf != nil { + schema.MultipleOf = source.MultipleOf + } + if source.Maximum != nil { + schema.Maximum = source.Maximum + } + if source.ExclusiveMaximum != nil { + schema.ExclusiveMaximum = source.ExclusiveMaximum + } + if source.Minimum != nil { + schema.Minimum = source.Minimum + } + if source.ExclusiveMinimum != nil { + schema.ExclusiveMinimum = source.ExclusiveMinimum + } + if source.MaxLength != nil { + schema.MaxLength = source.MaxLength + } + if source.MinLength != nil { + schema.MinLength = source.MinLength + } + if source.Pattern != nil { + schema.Pattern = source.Pattern + } + if source.AdditionalItems != nil { + schema.AdditionalItems = source.AdditionalItems + } + if source.Items != nil { + schema.Items = source.Items + } + if source.MaxItems != nil { + schema.MaxItems = source.MaxItems + } + if source.MinItems != nil { + schema.MinItems = source.MinItems + } + if source.UniqueItems != nil { + schema.UniqueItems = source.UniqueItems + } + if source.MaxProperties != nil { + schema.MaxProperties = source.MaxProperties + } + if source.MinProperties != nil { + schema.MinProperties = source.MinProperties + } + if source.Required != nil { + schema.Required = source.Required + } + if source.AdditionalProperties != nil { + schema.AdditionalProperties = source.AdditionalProperties + } + if source.Properties != nil { + schema.Properties = source.Properties + } + if source.PatternProperties != nil { + schema.PatternProperties = source.PatternProperties + } + if source.Dependencies != nil { + schema.Dependencies = source.Dependencies + } + if source.Enumeration != nil { + schema.Enumeration = source.Enumeration + } + if source.Type != nil { + schema.Type = source.Type + } + if source.AllOf != nil { + schema.AllOf = source.AllOf + } + if source.AnyOf != nil { + schema.AnyOf = source.AnyOf + } + if source.OneOf != nil { + schema.OneOf = source.OneOf + } + if source.Not != nil { + schema.Not = source.Not + } + if source.Definitions != nil { + schema.Definitions = source.Definitions + } + if source.Title != nil { + schema.Title = source.Title + } + if source.Description != nil { + schema.Description = source.Description + } + if source.Default != nil { + schema.Default = source.Default + } + if source.Format != nil { + schema.Format = source.Format + } + if source.Ref != nil { + schema.Ref = source.Ref + } +} + +// TypeIs returns true if the Type of a Schema includes the specified type +func (schema *Schema) TypeIs(typeName string) bool { + if schema.Type != nil { + // the schema Type is either a string or an array of strings + if schema.Type.String != nil { + return (*(schema.Type.String) == typeName) + } else if schema.Type.StringArray != nil { + for _, n := range *(schema.Type.StringArray) { + if n == typeName { + return true + } + } + } + } + return false +} + +// ResolveRefs resolves "$ref" elements in a Schema and its children. +// But if a reference refers to an object type, is inside a oneOf, or contains a oneOf, +// the reference is kept and we expect downstream tools to separately model these +// referenced schemas. +func (schema *Schema) ResolveRefs() { + rootSchema := schema + count := 1 + for count > 0 { + count = 0 + schema.applyToSchemas( + func(schema *Schema, context string) { + if schema.Ref != nil { + resolvedRef, err := rootSchema.resolveJSONPointer(*(schema.Ref)) + if err != nil { + log.Printf("%+v", err) + } else if resolvedRef.TypeIs("object") { + // don't substitute for objects, we'll model the referenced schema with a class + } else if context == "OneOf" { + // don't substitute for references inside oneOf declarations + } else if resolvedRef.OneOf != nil { + // don't substitute for references that contain oneOf declarations + } else if resolvedRef.AdditionalProperties != nil { + // don't substitute for references that look like objects + } else { + schema.Ref = nil + schema.CopyProperties(resolvedRef) + count++ + } + } + }, "") + } +} + +// resolveJSONPointer resolves JSON pointers. +// This current implementation is very crude and custom for OpenAPI 2.0 schemas. +// It panics for any pointer that it is unable to resolve. +func (schema *Schema) resolveJSONPointer(ref string) (result *Schema, err error) { + parts := strings.Split(ref, "#") + if len(parts) == 2 { + documentName := parts[0] + "#" + if documentName == "#" && schema.ID != nil { + documentName = *(schema.ID) + } + path := parts[1] + document := schemas[documentName] + pathParts := strings.Split(path, "/") + + // we currently do a very limited (hard-coded) resolution of certain paths and log errors for missed cases + if len(pathParts) == 1 { + return document, nil + } else if len(pathParts) == 3 { + switch pathParts[1] { + case "definitions": + dictionary := document.Definitions + for _, pair := range *dictionary { + if pair.Name == pathParts[2] { + result = pair.Value + } + } + case "properties": + dictionary := document.Properties + for _, pair := range *dictionary { + if pair.Name == pathParts[2] { + result = pair.Value + } + } + default: + break + } + } + } + if result == nil { + return nil, fmt.Errorf("unresolved pointer: %+v", ref) + } + return result, nil +} + +// ResolveAllOfs replaces "allOf" elements by merging their properties into the parent Schema. +func (schema *Schema) ResolveAllOfs() { + schema.applyToSchemas( + func(schema *Schema, context string) { + if schema.AllOf != nil { + for _, allOf := range *(schema.AllOf) { + schema.CopyProperties(allOf) + } + schema.AllOf = nil + } + }, "resolveAllOfs") +} + +// ResolveAnyOfs replaces all "anyOf" elements with "oneOf". +func (schema *Schema) ResolveAnyOfs() { + schema.applyToSchemas( + func(schema *Schema, context string) { + if schema.AnyOf != nil { + schema.OneOf = schema.AnyOf + schema.AnyOf = nil + } + }, "resolveAnyOfs") +} + +// return a pointer to a copy of a passed-in string +func stringptr(input string) (output *string) { + return &input +} + +// CopyOfficialSchemaProperty copies a named property from the official JSON Schema definition +func (schema *Schema) CopyOfficialSchemaProperty(name string) { + *schema.Properties = append(*schema.Properties, + NewNamedSchema(name, + &Schema{Ref: stringptr("http://json-schema.org/draft-04/schema#/properties/" + name)})) +} + +// CopyOfficialSchemaProperties copies named properties from the official JSON Schema definition +func (schema *Schema) CopyOfficialSchemaProperties(names []string) { + for _, name := range names { + schema.CopyOfficialSchemaProperty(name) + } +} diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/reader.go b/vendor/github.com/googleapis/gnostic/jsonschema/reader.go new file mode 100644 index 00000000..b8583d46 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/reader.go @@ -0,0 +1,442 @@ +// Copyright 2017 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +//go:generate go run generate-base.go + +package jsonschema + +import ( + "fmt" + "io/ioutil" + "strconv" + + "gopkg.in/yaml.v3" +) + +// This is a global map of all known Schemas. +// It is initialized when the first Schema is created and inserted. +var schemas map[string]*Schema + +// NewBaseSchema builds a schema object from an embedded json representation. +func NewBaseSchema() (schema *Schema, err error) { + b, err := baseSchemaBytes() + if err != nil { + return nil, err + } + var node yaml.Node + err = yaml.Unmarshal(b, &node) + if err != nil { + return nil, err + } + return NewSchemaFromObject(&node), nil +} + +// NewSchemaFromFile reads a schema from a file. +// Currently this assumes that schemas are stored in the source distribution of this project. +func NewSchemaFromFile(filename string) (schema *Schema, err error) { + file, err := ioutil.ReadFile(filename) + if err != nil { + return nil, err + } + var node yaml.Node + err = yaml.Unmarshal(file, &node) + if err != nil { + return nil, err + } + return NewSchemaFromObject(&node), nil +} + +// NewSchemaFromObject constructs a schema from a parsed JSON object. +// Due to the complexity of the schema representation, this is a +// custom reader and not the standard Go JSON reader (encoding/json). +func NewSchemaFromObject(jsonData *yaml.Node) *Schema { + switch jsonData.Kind { + case yaml.DocumentNode: + return NewSchemaFromObject(jsonData.Content[0]) + case yaml.MappingNode: + schema := &Schema{} + + for i := 0; i < len(jsonData.Content); i += 2 { + k := jsonData.Content[i].Value + v := jsonData.Content[i+1] + + switch k { + case "$schema": + schema.Schema = schema.stringValue(v) + case "id": + schema.ID = schema.stringValue(v) + + case "multipleOf": + schema.MultipleOf = schema.numberValue(v) + case "maximum": + schema.Maximum = schema.numberValue(v) + case "exclusiveMaximum": + schema.ExclusiveMaximum = schema.boolValue(v) + case "minimum": + schema.Minimum = schema.numberValue(v) + case "exclusiveMinimum": + schema.ExclusiveMinimum = schema.boolValue(v) + + case "maxLength": + schema.MaxLength = schema.intValue(v) + case "minLength": + schema.MinLength = schema.intValue(v) + case "pattern": + schema.Pattern = schema.stringValue(v) + + case "additionalItems": + schema.AdditionalItems = schema.schemaOrBooleanValue(v) + case "items": + schema.Items = schema.schemaOrSchemaArrayValue(v) + case "maxItems": + schema.MaxItems = schema.intValue(v) + case "minItems": + schema.MinItems = schema.intValue(v) + case "uniqueItems": + schema.UniqueItems = schema.boolValue(v) + + case "maxProperties": + schema.MaxProperties = schema.intValue(v) + case "minProperties": + schema.MinProperties = schema.intValue(v) + case "required": + schema.Required = schema.arrayOfStringsValue(v) + case "additionalProperties": + schema.AdditionalProperties = schema.schemaOrBooleanValue(v) + case "properties": + schema.Properties = schema.mapOfSchemasValue(v) + case "patternProperties": + schema.PatternProperties = schema.mapOfSchemasValue(v) + case "dependencies": + schema.Dependencies = schema.mapOfSchemasOrStringArraysValue(v) + + case "enum": + schema.Enumeration = schema.arrayOfEnumValuesValue(v) + + case "type": + schema.Type = schema.stringOrStringArrayValue(v) + case "allOf": + schema.AllOf = schema.arrayOfSchemasValue(v) + case "anyOf": + schema.AnyOf = schema.arrayOfSchemasValue(v) + case "oneOf": + schema.OneOf = schema.arrayOfSchemasValue(v) + case "not": + schema.Not = NewSchemaFromObject(v) + case "definitions": + schema.Definitions = schema.mapOfSchemasValue(v) + + case "title": + schema.Title = schema.stringValue(v) + case "description": + schema.Description = schema.stringValue(v) + + case "default": + schema.Default = v + + case "format": + schema.Format = schema.stringValue(v) + case "$ref": + schema.Ref = schema.stringValue(v) + default: + fmt.Printf("UNSUPPORTED (%s)\n", k) + } + } + + // insert schema in global map + if schema.ID != nil { + if schemas == nil { + schemas = make(map[string]*Schema, 0) + } + schemas[*(schema.ID)] = schema + } + return schema + + default: + fmt.Printf("schemaValue: unexpected node %+v\n", jsonData) + return nil + } + + return nil +} + +// +// BUILDERS +// The following methods build elements of Schemas from interface{} values. +// Each returns nil if it is unable to build the desired element. +// + +// Gets the string value of an interface{} value if possible. +func (schema *Schema) stringValue(v *yaml.Node) *string { + switch v.Kind { + case yaml.ScalarNode: + return &v.Value + default: + fmt.Printf("stringValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets the numeric value of an interface{} value if possible. +func (schema *Schema) numberValue(v *yaml.Node) *SchemaNumber { + number := &SchemaNumber{} + switch v.Kind { + case yaml.ScalarNode: + switch v.Tag { + case "!!float": + v2, _ := strconv.ParseFloat(v.Value, 64) + number.Float = &v2 + return number + case "!!int": + v2, _ := strconv.ParseInt(v.Value, 10, 64) + number.Integer = &v2 + return number + default: + fmt.Printf("stringValue: unexpected node %+v\n", v) + } + default: + fmt.Printf("stringValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets the integer value of an interface{} value if possible. +func (schema *Schema) intValue(v *yaml.Node) *int64 { + switch v.Kind { + case yaml.ScalarNode: + switch v.Tag { + case "!!float": + v2, _ := strconv.ParseFloat(v.Value, 64) + v3 := int64(v2) + return &v3 + case "!!int": + v2, _ := strconv.ParseInt(v.Value, 10, 64) + return &v2 + default: + fmt.Printf("intValue: unexpected node %+v\n", v) + } + default: + fmt.Printf("intValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets the bool value of an interface{} value if possible. +func (schema *Schema) boolValue(v *yaml.Node) *bool { + switch v.Kind { + case yaml.ScalarNode: + switch v.Tag { + case "!!bool": + v2, _ := strconv.ParseBool(v.Value) + return &v2 + default: + fmt.Printf("boolValue: unexpected node %+v\n", v) + } + default: + fmt.Printf("boolValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets a map of Schemas from an interface{} value if possible. +func (schema *Schema) mapOfSchemasValue(v *yaml.Node) *[]*NamedSchema { + switch v.Kind { + case yaml.MappingNode: + m := make([]*NamedSchema, 0) + for i := 0; i < len(v.Content); i += 2 { + k2 := v.Content[i].Value + v2 := v.Content[i+1] + pair := &NamedSchema{Name: k2, Value: NewSchemaFromObject(v2)} + m = append(m, pair) + } + return &m + default: + fmt.Printf("mapOfSchemasValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets an array of Schemas from an interface{} value if possible. +func (schema *Schema) arrayOfSchemasValue(v *yaml.Node) *[]*Schema { + switch v.Kind { + case yaml.SequenceNode: + m := make([]*Schema, 0) + for _, v2 := range v.Content { + switch v2.Kind { + case yaml.MappingNode: + s := NewSchemaFromObject(v2) + m = append(m, s) + default: + fmt.Printf("arrayOfSchemasValue: unexpected node %+v\n", v2) + } + } + return &m + case yaml.MappingNode: + m := make([]*Schema, 0) + s := NewSchemaFromObject(v) + m = append(m, s) + return &m + default: + fmt.Printf("arrayOfSchemasValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets a Schema or an array of Schemas from an interface{} value if possible. +func (schema *Schema) schemaOrSchemaArrayValue(v *yaml.Node) *SchemaOrSchemaArray { + switch v.Kind { + case yaml.SequenceNode: + m := make([]*Schema, 0) + for _, v2 := range v.Content { + switch v2.Kind { + case yaml.MappingNode: + s := NewSchemaFromObject(v2) + m = append(m, s) + default: + fmt.Printf("schemaOrSchemaArrayValue: unexpected node %+v\n", v2) + } + } + return &SchemaOrSchemaArray{SchemaArray: &m} + case yaml.MappingNode: + s := NewSchemaFromObject(v) + return &SchemaOrSchemaArray{Schema: s} + default: + fmt.Printf("schemaOrSchemaArrayValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets an array of strings from an interface{} value if possible. +func (schema *Schema) arrayOfStringsValue(v *yaml.Node) *[]string { + switch v.Kind { + case yaml.ScalarNode: + a := []string{v.Value} + return &a + case yaml.SequenceNode: + a := make([]string, 0) + for _, v2 := range v.Content { + switch v2.Kind { + case yaml.ScalarNode: + a = append(a, v2.Value) + default: + fmt.Printf("arrayOfStringsValue: unexpected node %+v\n", v2) + } + } + return &a + default: + fmt.Printf("arrayOfStringsValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets a string or an array of strings from an interface{} value if possible. +func (schema *Schema) stringOrStringArrayValue(v *yaml.Node) *StringOrStringArray { + switch v.Kind { + case yaml.ScalarNode: + s := &StringOrStringArray{} + s.String = &v.Value + return s + case yaml.SequenceNode: + a := make([]string, 0) + for _, v2 := range v.Content { + switch v2.Kind { + case yaml.ScalarNode: + a = append(a, v2.Value) + default: + fmt.Printf("arrayOfStringsValue: unexpected node %+v\n", v2) + } + } + s := &StringOrStringArray{} + s.StringArray = &a + return s + default: + fmt.Printf("arrayOfStringsValue: unexpected node %+v\n", v) + } + return nil +} + +// Gets an array of enum values from an interface{} value if possible. +func (schema *Schema) arrayOfEnumValuesValue(v *yaml.Node) *[]SchemaEnumValue { + a := make([]SchemaEnumValue, 0) + switch v.Kind { + case yaml.SequenceNode: + for _, v2 := range v.Content { + switch v2.Kind { + case yaml.ScalarNode: + switch v2.Tag { + case "!!str": + a = append(a, SchemaEnumValue{String: &v2.Value}) + case "!!bool": + v3, _ := strconv.ParseBool(v2.Value) + a = append(a, SchemaEnumValue{Bool: &v3}) + default: + fmt.Printf("arrayOfEnumValuesValue: unexpected type %s\n", v2.Tag) + } + default: + fmt.Printf("arrayOfEnumValuesValue: unexpected node %+v\n", v2) + } + } + default: + fmt.Printf("arrayOfEnumValuesValue: unexpected node %+v\n", v) + } + return &a +} + +// Gets a map of schemas or string arrays from an interface{} value if possible. +func (schema *Schema) mapOfSchemasOrStringArraysValue(v *yaml.Node) *[]*NamedSchemaOrStringArray { + m := make([]*NamedSchemaOrStringArray, 0) + switch v.Kind { + case yaml.MappingNode: + for i := 0; i < len(v.Content); i += 2 { + k2 := v.Content[i].Value + v2 := v.Content[i+1] + switch v2.Kind { + case yaml.SequenceNode: + a := make([]string, 0) + for _, v3 := range v2.Content { + switch v3.Kind { + case yaml.ScalarNode: + a = append(a, v3.Value) + default: + fmt.Printf("mapOfSchemasOrStringArraysValue: unexpected node %+v\n", v3) + } + } + s := &SchemaOrStringArray{} + s.StringArray = &a + pair := &NamedSchemaOrStringArray{Name: k2, Value: s} + m = append(m, pair) + default: + fmt.Printf("mapOfSchemasOrStringArraysValue: unexpected node %+v\n", v2) + } + } + default: + fmt.Printf("mapOfSchemasOrStringArraysValue: unexpected node %+v\n", v) + } + return &m +} + +// Gets a schema or a boolean value from an interface{} value if possible. +func (schema *Schema) schemaOrBooleanValue(v *yaml.Node) *SchemaOrBoolean { + schemaOrBoolean := &SchemaOrBoolean{} + switch v.Kind { + case yaml.ScalarNode: + v2, _ := strconv.ParseBool(v.Value) + schemaOrBoolean.Boolean = &v2 + case yaml.MappingNode: + schemaOrBoolean.Schema = NewSchemaFromObject(v) + default: + fmt.Printf("schemaOrBooleanValue: unexpected node %+v\n", v) + } + return schemaOrBoolean +} diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/schema.json b/vendor/github.com/googleapis/gnostic/jsonschema/schema.json new file mode 100644 index 00000000..85eb502a --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/schema.json @@ -0,0 +1,150 @@ +{ + "id": "http://json-schema.org/draft-04/schema#", + "$schema": "http://json-schema.org/draft-04/schema#", + "description": "Core schema meta-schema", + "definitions": { + "schemaArray": { + "type": "array", + "minItems": 1, + "items": { "$ref": "#" } + }, + "positiveInteger": { + "type": "integer", + "minimum": 0 + }, + "positiveIntegerDefault0": { + "allOf": [ { "$ref": "#/definitions/positiveInteger" }, { "default": 0 } ] + }, + "simpleTypes": { + "enum": [ "array", "boolean", "integer", "null", "number", "object", "string" ] + }, + "stringArray": { + "type": "array", + "items": { "type": "string" }, + "minItems": 1, + "uniqueItems": true + } + }, + "type": "object", + "properties": { + "id": { + "type": "string", + "format": "uri" + }, + "$schema": { + "type": "string", + "format": "uri" + }, + "title": { + "type": "string" + }, + "description": { + "type": "string" + }, + "default": {}, + "multipleOf": { + "type": "number", + "minimum": 0, + "exclusiveMinimum": true + }, + "maximum": { + "type": "number" + }, + "exclusiveMaximum": { + "type": "boolean", + "default": false + }, + "minimum": { + "type": "number" + }, + "exclusiveMinimum": { + "type": "boolean", + "default": false + }, + "maxLength": { "$ref": "#/definitions/positiveInteger" }, + "minLength": { "$ref": "#/definitions/positiveIntegerDefault0" }, + "pattern": { + "type": "string", + "format": "regex" + }, + "additionalItems": { + "anyOf": [ + { "type": "boolean" }, + { "$ref": "#" } + ], + "default": {} + }, + "items": { + "anyOf": [ + { "$ref": "#" }, + { "$ref": "#/definitions/schemaArray" } + ], + "default": {} + }, + "maxItems": { "$ref": "#/definitions/positiveInteger" }, + "minItems": { "$ref": "#/definitions/positiveIntegerDefault0" }, + "uniqueItems": { + "type": "boolean", + "default": false + }, + "maxProperties": { "$ref": "#/definitions/positiveInteger" }, + "minProperties": { "$ref": "#/definitions/positiveIntegerDefault0" }, + "required": { "$ref": "#/definitions/stringArray" }, + "additionalProperties": { + "anyOf": [ + { "type": "boolean" }, + { "$ref": "#" } + ], + "default": {} + }, + "definitions": { + "type": "object", + "additionalProperties": { "$ref": "#" }, + "default": {} + }, + "properties": { + "type": "object", + "additionalProperties": { "$ref": "#" }, + "default": {} + }, + "patternProperties": { + "type": "object", + "additionalProperties": { "$ref": "#" }, + "default": {} + }, + "dependencies": { + "type": "object", + "additionalProperties": { + "anyOf": [ + { "$ref": "#" }, + { "$ref": "#/definitions/stringArray" } + ] + } + }, + "enum": { + "type": "array", + "minItems": 1, + "uniqueItems": true + }, + "type": { + "anyOf": [ + { "$ref": "#/definitions/simpleTypes" }, + { + "type": "array", + "items": { "$ref": "#/definitions/simpleTypes" }, + "minItems": 1, + "uniqueItems": true + } + ] + }, + "allOf": { "$ref": "#/definitions/schemaArray" }, + "anyOf": { "$ref": "#/definitions/schemaArray" }, + "oneOf": { "$ref": "#/definitions/schemaArray" }, + "not": { "$ref": "#" } + }, + "dependencies": { + "exclusiveMaximum": [ "maximum" ], + "exclusiveMinimum": [ "minimum" ] + }, + "default": {} +} diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/writer.go b/vendor/github.com/googleapis/gnostic/jsonschema/writer.go new file mode 100644 index 00000000..340dc5f9 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/jsonschema/writer.go @@ -0,0 +1,369 @@ +// Copyright 2017 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package jsonschema + +import ( + "fmt" + + "gopkg.in/yaml.v3" +) + +const indentation = " " + +func renderMappingNode(node *yaml.Node, indent string) (result string) { + result = "{\n" + innerIndent := indent + indentation + for i := 0; i < len(node.Content); i += 2 { + // first print the key + key := node.Content[i].Value + result += fmt.Sprintf("%s\"%+v\": ", innerIndent, key) + // then the value + value := node.Content[i+1] + switch value.Kind { + case yaml.ScalarNode: + result += "\"" + value.Value + "\"" + case yaml.MappingNode: + result += renderMappingNode(value, innerIndent) + case yaml.SequenceNode: + result += renderSequenceNode(value, innerIndent) + default: + result += fmt.Sprintf("???MapItem(Key:%+v, Value:%T)", value, value) + } + if i < len(node.Content)-2 { + result += "," + } + result += "\n" + } + + result += indent + "}" + return result +} + +func renderSequenceNode(node *yaml.Node, indent string) (result string) { + result = "[\n" + innerIndent := indent + indentation + for i := 0; i < len(node.Content); i++ { + item := node.Content[i] + switch item.Kind { + case yaml.ScalarNode: + result += innerIndent + "\"" + item.Value + "\"" + case yaml.MappingNode: + result += innerIndent + renderMappingNode(item, innerIndent) + "" + default: + result += innerIndent + fmt.Sprintf("???ArrayItem(%+v)", item) + } + if i < len(node.Content)-1 { + result += "," + } + result += "\n" + } + result += indent + "]" + return result +} + +func renderStringArray(array []string, indent string) (result string) { + result = "[\n" + innerIndent := indent + indentation + for i, item := range array { + result += innerIndent + "\"" + item + "\"" + if i < len(array)-1 { + result += "," + } + result += "\n" + } + result += indent + "]" + return result +} + +// Render renders a yaml.Node as JSON +func Render(node *yaml.Node) string { + if node.Kind == yaml.DocumentNode { + if len(node.Content) == 1 { + return Render(node.Content[0]) + } + } else if node.Kind == yaml.MappingNode { + return renderMappingNode(node, "") + "\n" + } else if node.Kind == yaml.SequenceNode { + return renderSequenceNode(node, "") + "\n" + } + return "" +} + +func (object *SchemaNumber) nodeValue() *yaml.Node { + if object.Integer != nil { + return nodeForInt64(*object.Integer) + } else if object.Float != nil { + return nodeForFloat64(*object.Float) + } else { + return nil + } +} + +func (object *SchemaOrBoolean) nodeValue() *yaml.Node { + if object.Schema != nil { + return object.Schema.nodeValue() + } else if object.Boolean != nil { + return nodeForBoolean(*object.Boolean) + } else { + return nil + } +} + +func nodeForStringArray(array []string) *yaml.Node { + content := make([]*yaml.Node, 0) + for _, item := range array { + content = append(content, nodeForString(item)) + } + return nodeForSequence(content) +} + +func nodeForSchemaArray(array []*Schema) *yaml.Node { + content := make([]*yaml.Node, 0) + for _, item := range array { + content = append(content, item.nodeValue()) + } + return nodeForSequence(content) +} + +func (object *StringOrStringArray) nodeValue() *yaml.Node { + if object.String != nil { + return nodeForString(*object.String) + } else if object.StringArray != nil { + return nodeForStringArray(*(object.StringArray)) + } else { + return nil + } +} + +func (object *SchemaOrStringArray) nodeValue() *yaml.Node { + if object.Schema != nil { + return object.Schema.nodeValue() + } else if object.StringArray != nil { + return nodeForStringArray(*(object.StringArray)) + } else { + return nil + } +} + +func (object *SchemaOrSchemaArray) nodeValue() *yaml.Node { + if object.Schema != nil { + return object.Schema.nodeValue() + } else if object.SchemaArray != nil { + return nodeForSchemaArray(*(object.SchemaArray)) + } else { + return nil + } +} + +func (object *SchemaEnumValue) nodeValue() *yaml.Node { + if object.String != nil { + return nodeForString(*object.String) + } else if object.Bool != nil { + return nodeForBoolean(*object.Bool) + } else { + return nil + } +} + +func nodeForNamedSchemaArray(array *[]*NamedSchema) *yaml.Node { + content := make([]*yaml.Node, 0) + for _, pair := range *(array) { + content = appendPair(content, pair.Name, pair.Value.nodeValue()) + } + return nodeForMapping(content) +} + +func nodeForNamedSchemaOrStringArray(array *[]*NamedSchemaOrStringArray) *yaml.Node { + content := make([]*yaml.Node, 0) + for _, pair := range *(array) { + content = appendPair(content, pair.Name, pair.Value.nodeValue()) + } + return nodeForMapping(content) +} + +func nodeForSchemaEnumArray(array *[]SchemaEnumValue) *yaml.Node { + content := make([]*yaml.Node, 0) + for _, item := range *array { + content = append(content, item.nodeValue()) + } + return nodeForSequence(content) +} + +func nodeForMapping(content []*yaml.Node) *yaml.Node { + return &yaml.Node{ + Kind: yaml.MappingNode, + Content: content, + } +} + +func nodeForSequence(content []*yaml.Node) *yaml.Node { + return &yaml.Node{ + Kind: yaml.SequenceNode, + Content: content, + } +} + +func nodeForString(value string) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!str", + Value: value, + } +} + +func nodeForBoolean(value bool) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!bool", + Value: fmt.Sprintf("%t", value), + } +} + +func nodeForInt64(value int64) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!int", + Value: fmt.Sprintf("%d", value), + } +} + +func nodeForFloat64(value float64) *yaml.Node { + return &yaml.Node{ + Kind: yaml.ScalarNode, + Tag: "!!float", + Value: fmt.Sprintf("%f", value), + } +} + +func appendPair(nodes []*yaml.Node, name string, value *yaml.Node) []*yaml.Node { + nodes = append(nodes, nodeForString(name)) + nodes = append(nodes, value) + return nodes +} + +func (schema *Schema) nodeValue() *yaml.Node { + n := &yaml.Node{Kind: yaml.MappingNode} + content := make([]*yaml.Node, 0) + if schema.Title != nil { + content = appendPair(content, "title", nodeForString(*schema.Title)) + } + if schema.ID != nil { + content = appendPair(content, "id", nodeForString(*schema.ID)) + } + if schema.Schema != nil { + content = appendPair(content, "$schema", nodeForString(*schema.Schema)) + } + if schema.Type != nil { + content = appendPair(content, "type", schema.Type.nodeValue()) + } + if schema.Items != nil { + content = appendPair(content, "items", schema.Items.nodeValue()) + } + if schema.Description != nil { + content = appendPair(content, "description", nodeForString(*schema.Description)) + } + if schema.Required != nil { + content = appendPair(content, "required", nodeForStringArray(*schema.Required)) + } + if schema.AdditionalProperties != nil { + content = appendPair(content, "additionalProperties", schema.AdditionalProperties.nodeValue()) + } + if schema.PatternProperties != nil { + content = appendPair(content, "patternProperties", nodeForNamedSchemaArray(schema.PatternProperties)) + } + if schema.Properties != nil { + content = appendPair(content, "properties", nodeForNamedSchemaArray(schema.Properties)) + } + if schema.Dependencies != nil { + content = appendPair(content, "dependencies", nodeForNamedSchemaOrStringArray(schema.Dependencies)) + } + if schema.Ref != nil { + content = appendPair(content, "$ref", nodeForString(*schema.Ref)) + } + if schema.MultipleOf != nil { + content = appendPair(content, "multipleOf", schema.MultipleOf.nodeValue()) + } + if schema.Maximum != nil { + content = appendPair(content, "maximum", schema.Maximum.nodeValue()) + } + if schema.ExclusiveMaximum != nil { + content = appendPair(content, "exclusiveMaximum", nodeForBoolean(*schema.ExclusiveMaximum)) + } + if schema.Minimum != nil { + content = appendPair(content, "minimum", schema.Minimum.nodeValue()) + } + if schema.ExclusiveMinimum != nil { + content = appendPair(content, "exclusiveMinimum", nodeForBoolean(*schema.ExclusiveMinimum)) + } + if schema.MaxLength != nil { + content = appendPair(content, "maxLength", nodeForInt64(*schema.MaxLength)) + } + if schema.MinLength != nil { + content = appendPair(content, "minLength", nodeForInt64(*schema.MinLength)) + } + if schema.Pattern != nil { + content = appendPair(content, "pattern", nodeForString(*schema.Pattern)) + } + if schema.AdditionalItems != nil { + content = appendPair(content, "additionalItems", schema.AdditionalItems.nodeValue()) + } + if schema.MaxItems != nil { + content = appendPair(content, "maxItems", nodeForInt64(*schema.MaxItems)) + } + if schema.MinItems != nil { + content = appendPair(content, "minItems", nodeForInt64(*schema.MinItems)) + } + if schema.UniqueItems != nil { + content = appendPair(content, "uniqueItems", nodeForBoolean(*schema.UniqueItems)) + } + if schema.MaxProperties != nil { + content = appendPair(content, "maxProperties", nodeForInt64(*schema.MaxProperties)) + } + if schema.MinProperties != nil { + content = appendPair(content, "minProperties", nodeForInt64(*schema.MinProperties)) + } + if schema.Enumeration != nil { + content = appendPair(content, "enum", nodeForSchemaEnumArray(schema.Enumeration)) + } + if schema.AllOf != nil { + content = appendPair(content, "allOf", nodeForSchemaArray(*schema.AllOf)) + } + if schema.AnyOf != nil { + content = appendPair(content, "anyOf", nodeForSchemaArray(*schema.AnyOf)) + } + if schema.OneOf != nil { + content = appendPair(content, "oneOf", nodeForSchemaArray(*schema.OneOf)) + } + if schema.Not != nil { + content = appendPair(content, "not", schema.Not.nodeValue()) + } + if schema.Definitions != nil { + content = appendPair(content, "definitions", nodeForNamedSchemaArray(schema.Definitions)) + } + if schema.Default != nil { + // m = append(m, yaml.MapItem{Key: "default", Value: *schema.Default}) + } + if schema.Format != nil { + content = appendPair(content, "format", nodeForString(*schema.Format)) + } + n.Content = content + return n +} + +// JSONString returns a json representation of a schema. +func (schema *Schema) JSONString() string { + node := schema.nodeValue() + return Render(node) +} diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.go b/vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.go similarity index 65% rename from vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.go rename to vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.go index 4fd44c45..eb93b65e 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.go +++ b/vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.go @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2020 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -19,7 +19,7 @@ package openapi_v2 import ( "fmt" "github.com/googleapis/gnostic/compiler" - "gopkg.in/yaml.v2" + "gopkg.in/yaml.v3" "regexp" "strings" ) @@ -30,7 +30,7 @@ func Version() string { } // NewAdditionalPropertiesItem creates an object of type AdditionalPropertiesItem if possible, returning an error if not. -func NewAdditionalPropertiesItem(in interface{}, context *compiler.Context) (*AdditionalPropertiesItem, error) { +func NewAdditionalPropertiesItem(in *yaml.Node, context *compiler.Context) (*AdditionalPropertiesItem, error) { errors := make([]error, 0) x := &AdditionalPropertiesItem{} matched := false @@ -39,7 +39,7 @@ func NewAdditionalPropertiesItem(in interface{}, context *compiler.Context) (*Ad m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewSchema(m, compiler.NewContext("schema", context)) + t, matchingError := NewSchema(m, compiler.NewContext("schema", m, context)) if matchingError == nil { x.Oneof = &AdditionalPropertiesItem_Schema{Schema: t} matched = true @@ -49,28 +49,33 @@ func NewAdditionalPropertiesItem(in interface{}, context *compiler.Context) (*Ad } } // bool boolean = 2; - boolValue, ok := in.(bool) + boolValue, ok := compiler.BoolForScalarNode(in) if ok { x.Oneof = &AdditionalPropertiesItem_Boolean{Boolean: boolValue} + matched = true } if matched { // since the oneof matched one of its possibilities, discard any matching errors errors = make([]error, 0) + } else { + message := fmt.Sprintf("contains an invalid AdditionalPropertiesItem") + err := compiler.NewError(context, message) + errors = []error{err} } return x, compiler.NewErrorGroupOrNil(errors) } // NewAny creates an object of type Any if possible, returning an error if not. -func NewAny(in interface{}, context *compiler.Context) (*Any, error) { +func NewAny(in *yaml.Node, context *compiler.Context) (*Any, error) { errors := make([]error, 0) x := &Any{} - bytes, _ := yaml.Marshal(in) + bytes := compiler.Marshal(in) x.Yaml = string(bytes) return x, compiler.NewErrorGroupOrNil(errors) } // NewApiKeySecurity creates an object of type ApiKeySecurity if possible, returning an error if not. -func NewApiKeySecurity(in interface{}, context *compiler.Context) (*ApiKeySecurity, error) { +func NewApiKeySecurity(in *yaml.Node, context *compiler.Context) (*ApiKeySecurity, error) { errors := make([]error, 0) x := &ApiKeySecurity{} m, ok := compiler.UnpackMap(in) @@ -94,74 +99,74 @@ func NewApiKeySecurity(in interface{}, context *compiler.Context) (*ApiKeySecuri // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [apiKey] if ok && !compiler.StringArrayContainsValue([]string{"apiKey"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string name = 2; v2 := compiler.MapValueForKey(m, "name") if v2 != nil { - x.Name, ok = v2.(string) + x.Name, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string in = 3; v3 := compiler.MapValueForKey(m, "in") if v3 != nil { - x.In, ok = v3.(string) + x.In, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [header query] if ok && !compiler.StringArrayContainsValue([]string{"header", "query"}, x.In) { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 4; v4 := compiler.MapValueForKey(m, "description") if v4 != nil { - x.Description, ok = v4.(string) + x.Description, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 5; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -175,7 +180,7 @@ func NewApiKeySecurity(in interface{}, context *compiler.Context) (*ApiKeySecuri } // NewBasicAuthenticationSecurity creates an object of type BasicAuthenticationSecurity if possible, returning an error if not. -func NewBasicAuthenticationSecurity(in interface{}, context *compiler.Context) (*BasicAuthenticationSecurity, error) { +func NewBasicAuthenticationSecurity(in *yaml.Node, context *compiler.Context) (*BasicAuthenticationSecurity, error) { errors := make([]error, 0) x := &BasicAuthenticationSecurity{} m, ok := compiler.UnpackMap(in) @@ -199,50 +204,50 @@ func NewBasicAuthenticationSecurity(in interface{}, context *compiler.Context) ( // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [basic] if ok && !compiler.StringArrayContainsValue([]string{"basic"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 2; v2 := compiler.MapValueForKey(m, "description") if v2 != nil { - x.Description, ok = v2.(string) + x.Description, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 3; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -256,7 +261,7 @@ func NewBasicAuthenticationSecurity(in interface{}, context *compiler.Context) ( } // NewBodyParameter creates an object of type BodyParameter if possible, returning an error if not. -func NewBodyParameter(in interface{}, context *compiler.Context) (*BodyParameter, error) { +func NewBodyParameter(in *yaml.Node, context *compiler.Context) (*BodyParameter, error) { errors := make([]error, 0) x := &BodyParameter{} m, ok := compiler.UnpackMap(in) @@ -280,42 +285,42 @@ func NewBodyParameter(in interface{}, context *compiler.Context) (*BodyParameter // string description = 1; v1 := compiler.MapValueForKey(m, "description") if v1 != nil { - x.Description, ok = v1.(string) + x.Description, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string name = 2; v2 := compiler.MapValueForKey(m, "name") if v2 != nil { - x.Name, ok = v2.(string) + x.Name, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string in = 3; v3 := compiler.MapValueForKey(m, "in") if v3 != nil { - x.In, ok = v3.(string) + x.In, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [body] if ok && !compiler.StringArrayContainsValue([]string{"body"}, x.In) { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // bool required = 4; v4 := compiler.MapValueForKey(m, "required") if v4 != nil { - x.Required, ok = v4.(bool) + x.Required, ok = compiler.BoolForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for required: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for required: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } @@ -323,7 +328,7 @@ func NewBodyParameter(in interface{}, context *compiler.Context) (*BodyParameter v5 := compiler.MapValueForKey(m, "schema") if v5 != nil { var err error - x.Schema, err = NewSchema(v5, compiler.NewContext("schema", context)) + x.Schema, err = NewSchema(v5, compiler.NewContext("schema", v5, context)) if err != nil { errors = append(errors, err) } @@ -331,26 +336,26 @@ func NewBodyParameter(in interface{}, context *compiler.Context) (*BodyParameter // repeated NamedAny vendor_extension = 6; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -364,7 +369,7 @@ func NewBodyParameter(in interface{}, context *compiler.Context) (*BodyParameter } // NewContact creates an object of type Contact if possible, returning an error if not. -func NewContact(in interface{}, context *compiler.Context) (*Contact, error) { +func NewContact(in *yaml.Node, context *compiler.Context) (*Contact, error) { errors := make([]error, 0) x := &Contact{} m, ok := compiler.UnpackMap(in) @@ -382,53 +387,53 @@ func NewContact(in interface{}, context *compiler.Context) (*Contact, error) { // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string url = 2; v2 := compiler.MapValueForKey(m, "url") if v2 != nil { - x.Url, ok = v2.(string) + x.Url, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for url: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for url: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string email = 3; v3 := compiler.MapValueForKey(m, "email") if v3 != nil { - x.Email, ok = v3.(string) + x.Email, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for email: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for email: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 4; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -442,7 +447,7 @@ func NewContact(in interface{}, context *compiler.Context) (*Contact, error) { } // NewDefault creates an object of type Default if possible, returning an error if not. -func NewDefault(in interface{}, context *compiler.Context) (*Default, error) { +func NewDefault(in *yaml.Node, context *compiler.Context) (*Default, error) { errors := make([]error, 0) x := &Default{} m, ok := compiler.UnpackMap(in) @@ -453,25 +458,25 @@ func NewDefault(in interface{}, context *compiler.Context) (*Default, error) { // repeated NamedAny additional_properties = 1; // MAP: Any x.AdditionalProperties = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -484,7 +489,7 @@ func NewDefault(in interface{}, context *compiler.Context) (*Default, error) { } // NewDefinitions creates an object of type Definitions if possible, returning an error if not. -func NewDefinitions(in interface{}, context *compiler.Context) (*Definitions, error) { +func NewDefinitions(in *yaml.Node, context *compiler.Context) (*Definitions, error) { errors := make([]error, 0) x := &Definitions{} m, ok := compiler.UnpackMap(in) @@ -495,14 +500,14 @@ func NewDefinitions(in interface{}, context *compiler.Context) (*Definitions, er // repeated NamedSchema additional_properties = 1; // MAP: Schema x.AdditionalProperties = make([]*NamedSchema, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedSchema{} pair.Name = k var err error - pair.Value, err = NewSchema(v, compiler.NewContext(k, context)) + pair.Value, err = NewSchema(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -514,7 +519,7 @@ func NewDefinitions(in interface{}, context *compiler.Context) (*Definitions, er } // NewDocument creates an object of type Document if possible, returning an error if not. -func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { +func NewDocument(in *yaml.Node, context *compiler.Context) (*Document, error) { errors := make([]error, 0) x := &Document{} m, ok := compiler.UnpackMap(in) @@ -538,15 +543,15 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { // string swagger = 1; v1 := compiler.MapValueForKey(m, "swagger") if v1 != nil { - x.Swagger, ok = v1.(string) + x.Swagger, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for swagger: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for swagger: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [2.0] if ok && !compiler.StringArrayContainsValue([]string{"2.0"}, x.Swagger) { - message := fmt.Sprintf("has unexpected value for swagger: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for swagger: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -554,7 +559,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { v2 := compiler.MapValueForKey(m, "info") if v2 != nil { var err error - x.Info, err = NewInfo(v2, compiler.NewContext("info", context)) + x.Info, err = NewInfo(v2, compiler.NewContext("info", v2, context)) if err != nil { errors = append(errors, err) } @@ -562,57 +567,57 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { // string host = 3; v3 := compiler.MapValueForKey(m, "host") if v3 != nil { - x.Host, ok = v3.(string) + x.Host, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for host: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for host: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string base_path = 4; v4 := compiler.MapValueForKey(m, "basePath") if v4 != nil { - x.BasePath, ok = v4.(string) + x.BasePath, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for basePath: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for basePath: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // repeated string schemes = 5; v5 := compiler.MapValueForKey(m, "schemes") if v5 != nil { - v, ok := v5.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v5) if ok { - x.Schemes = compiler.ConvertInterfaceArrayToStringArray(v) + x.Schemes = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for schemes: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for schemes: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [http https ws wss] if ok && !compiler.StringArrayContainsValues([]string{"http", "https", "ws", "wss"}, x.Schemes) { - message := fmt.Sprintf("has unexpected value for schemes: %+v", v5) + message := fmt.Sprintf("has unexpected value for schemes: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // repeated string consumes = 6; v6 := compiler.MapValueForKey(m, "consumes") if v6 != nil { - v, ok := v6.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v6) if ok { - x.Consumes = compiler.ConvertInterfaceArrayToStringArray(v) + x.Consumes = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for consumes: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for consumes: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // repeated string produces = 7; v7 := compiler.MapValueForKey(m, "produces") if v7 != nil { - v, ok := v7.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v7) if ok { - x.Produces = compiler.ConvertInterfaceArrayToStringArray(v) + x.Produces = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for produces: %+v (%T)", v7, v7) + message := fmt.Sprintf("has unexpected value for produces: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } @@ -620,7 +625,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { v8 := compiler.MapValueForKey(m, "paths") if v8 != nil { var err error - x.Paths, err = NewPaths(v8, compiler.NewContext("paths", context)) + x.Paths, err = NewPaths(v8, compiler.NewContext("paths", v8, context)) if err != nil { errors = append(errors, err) } @@ -629,7 +634,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { v9 := compiler.MapValueForKey(m, "definitions") if v9 != nil { var err error - x.Definitions, err = NewDefinitions(v9, compiler.NewContext("definitions", context)) + x.Definitions, err = NewDefinitions(v9, compiler.NewContext("definitions", v9, context)) if err != nil { errors = append(errors, err) } @@ -638,7 +643,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { v10 := compiler.MapValueForKey(m, "parameters") if v10 != nil { var err error - x.Parameters, err = NewParameterDefinitions(v10, compiler.NewContext("parameters", context)) + x.Parameters, err = NewParameterDefinitions(v10, compiler.NewContext("parameters", v10, context)) if err != nil { errors = append(errors, err) } @@ -647,7 +652,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { v11 := compiler.MapValueForKey(m, "responses") if v11 != nil { var err error - x.Responses, err = NewResponseDefinitions(v11, compiler.NewContext("responses", context)) + x.Responses, err = NewResponseDefinitions(v11, compiler.NewContext("responses", v11, context)) if err != nil { errors = append(errors, err) } @@ -657,10 +662,10 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { if v12 != nil { // repeated SecurityRequirement x.Security = make([]*SecurityRequirement, 0) - a, ok := v12.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v12) if ok { - for _, item := range a { - y, err := NewSecurityRequirement(item, compiler.NewContext("security", context)) + for _, item := range a.Content { + y, err := NewSecurityRequirement(item, compiler.NewContext("security", item, context)) if err != nil { errors = append(errors, err) } @@ -672,7 +677,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { v13 := compiler.MapValueForKey(m, "securityDefinitions") if v13 != nil { var err error - x.SecurityDefinitions, err = NewSecurityDefinitions(v13, compiler.NewContext("securityDefinitions", context)) + x.SecurityDefinitions, err = NewSecurityDefinitions(v13, compiler.NewContext("securityDefinitions", v13, context)) if err != nil { errors = append(errors, err) } @@ -682,10 +687,10 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { if v14 != nil { // repeated Tag x.Tags = make([]*Tag, 0) - a, ok := v14.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v14) if ok { - for _, item := range a { - y, err := NewTag(item, compiler.NewContext("tags", context)) + for _, item := range a.Content { + y, err := NewTag(item, compiler.NewContext("tags", item, context)) if err != nil { errors = append(errors, err) } @@ -697,7 +702,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { v15 := compiler.MapValueForKey(m, "externalDocs") if v15 != nil { var err error - x.ExternalDocs, err = NewExternalDocs(v15, compiler.NewContext("externalDocs", context)) + x.ExternalDocs, err = NewExternalDocs(v15, compiler.NewContext("externalDocs", v15, context)) if err != nil { errors = append(errors, err) } @@ -705,26 +710,26 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { // repeated NamedAny vendor_extension = 16; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -738,7 +743,7 @@ func NewDocument(in interface{}, context *compiler.Context) (*Document, error) { } // NewExamples creates an object of type Examples if possible, returning an error if not. -func NewExamples(in interface{}, context *compiler.Context) (*Examples, error) { +func NewExamples(in *yaml.Node, context *compiler.Context) (*Examples, error) { errors := make([]error, 0) x := &Examples{} m, ok := compiler.UnpackMap(in) @@ -749,25 +754,25 @@ func NewExamples(in interface{}, context *compiler.Context) (*Examples, error) { // repeated NamedAny additional_properties = 1; // MAP: Any x.AdditionalProperties = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -780,7 +785,7 @@ func NewExamples(in interface{}, context *compiler.Context) (*Examples, error) { } // NewExternalDocs creates an object of type ExternalDocs if possible, returning an error if not. -func NewExternalDocs(in interface{}, context *compiler.Context) (*ExternalDocs, error) { +func NewExternalDocs(in *yaml.Node, context *compiler.Context) (*ExternalDocs, error) { errors := make([]error, 0) x := &ExternalDocs{} m, ok := compiler.UnpackMap(in) @@ -804,44 +809,44 @@ func NewExternalDocs(in interface{}, context *compiler.Context) (*ExternalDocs, // string description = 1; v1 := compiler.MapValueForKey(m, "description") if v1 != nil { - x.Description, ok = v1.(string) + x.Description, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string url = 2; v2 := compiler.MapValueForKey(m, "url") if v2 != nil { - x.Url, ok = v2.(string) + x.Url, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for url: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for url: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 3; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -855,7 +860,7 @@ func NewExternalDocs(in interface{}, context *compiler.Context) (*ExternalDocs, } // NewFileSchema creates an object of type FileSchema if possible, returning an error if not. -func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, error) { +func NewFileSchema(in *yaml.Node, context *compiler.Context) (*FileSchema, error) { errors := make([]error, 0) x := &FileSchema{} m, ok := compiler.UnpackMap(in) @@ -879,27 +884,27 @@ func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, erro // string format = 1; v1 := compiler.MapValueForKey(m, "format") if v1 != nil { - x.Format, ok = v1.(string) + x.Format, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string title = 2; v2 := compiler.MapValueForKey(m, "title") if v2 != nil { - x.Title, ok = v2.(string) + x.Title, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for title: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for title: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 3; v3 := compiler.MapValueForKey(m, "description") if v3 != nil { - x.Description, ok = v3.(string) + x.Description, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } @@ -907,7 +912,7 @@ func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, erro v4 := compiler.MapValueForKey(m, "default") if v4 != nil { var err error - x.Default, err = NewAny(v4, compiler.NewContext("default", context)) + x.Default, err = NewAny(v4, compiler.NewContext("default", v4, context)) if err != nil { errors = append(errors, err) } @@ -915,35 +920,35 @@ func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, erro // repeated string required = 5; v5 := compiler.MapValueForKey(m, "required") if v5 != nil { - v, ok := v5.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v5) if ok { - x.Required = compiler.ConvertInterfaceArrayToStringArray(v) + x.Required = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for required: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for required: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // string type = 6; v6 := compiler.MapValueForKey(m, "type") if v6 != nil { - x.Type, ok = v6.(string) + x.Type, ok = compiler.StringForScalarNode(v6) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [file] if ok && !compiler.StringArrayContainsValue([]string{"file"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // bool read_only = 7; v7 := compiler.MapValueForKey(m, "readOnly") if v7 != nil { - x.ReadOnly, ok = v7.(bool) + x.ReadOnly, ok = compiler.BoolForScalarNode(v7) if !ok { - message := fmt.Sprintf("has unexpected value for readOnly: %+v (%T)", v7, v7) + message := fmt.Sprintf("has unexpected value for readOnly: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } @@ -951,7 +956,7 @@ func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, erro v8 := compiler.MapValueForKey(m, "externalDocs") if v8 != nil { var err error - x.ExternalDocs, err = NewExternalDocs(v8, compiler.NewContext("externalDocs", context)) + x.ExternalDocs, err = NewExternalDocs(v8, compiler.NewContext("externalDocs", v8, context)) if err != nil { errors = append(errors, err) } @@ -960,7 +965,7 @@ func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, erro v9 := compiler.MapValueForKey(m, "example") if v9 != nil { var err error - x.Example, err = NewAny(v9, compiler.NewContext("example", context)) + x.Example, err = NewAny(v9, compiler.NewContext("example", v9, context)) if err != nil { errors = append(errors, err) } @@ -968,26 +973,26 @@ func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, erro // repeated NamedAny vendor_extension = 10; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -1001,7 +1006,7 @@ func NewFileSchema(in interface{}, context *compiler.Context) (*FileSchema, erro } // NewFormDataParameterSubSchema creates an object of type FormDataParameterSubSchema if possible, returning an error if not. -func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (*FormDataParameterSubSchema, error) { +func NewFormDataParameterSubSchema(in *yaml.Node, context *compiler.Context) (*FormDataParameterSubSchema, error) { errors := make([]error, 0) x := &FormDataParameterSubSchema{} m, ok := compiler.UnpackMap(in) @@ -1019,75 +1024,75 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* // bool required = 1; v1 := compiler.MapValueForKey(m, "required") if v1 != nil { - x.Required, ok = v1.(bool) + x.Required, ok = compiler.BoolForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for required: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for required: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string in = 2; v2 := compiler.MapValueForKey(m, "in") if v2 != nil { - x.In, ok = v2.(string) + x.In, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [formData] if ok && !compiler.StringArrayContainsValue([]string{"formData"}, x.In) { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 3; v3 := compiler.MapValueForKey(m, "description") if v3 != nil { - x.Description, ok = v3.(string) + x.Description, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string name = 4; v4 := compiler.MapValueForKey(m, "name") if v4 != nil { - x.Name, ok = v4.(string) + x.Name, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // bool allow_empty_value = 5; v5 := compiler.MapValueForKey(m, "allowEmptyValue") if v5 != nil { - x.AllowEmptyValue, ok = v5.(bool) + x.AllowEmptyValue, ok = compiler.BoolForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for allowEmptyValue: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for allowEmptyValue: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // string type = 6; v6 := compiler.MapValueForKey(m, "type") if v6 != nil { - x.Type, ok = v6.(string) + x.Type, ok = compiler.StringForScalarNode(v6) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [string number boolean integer array file] if ok && !compiler.StringArrayContainsValue([]string{"string", "number", "boolean", "integer", "array", "file"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // string format = 7; v7 := compiler.MapValueForKey(m, "format") if v7 != nil { - x.Format, ok = v7.(string) + x.Format, ok = compiler.StringForScalarNode(v7) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v7, v7) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1095,7 +1100,7 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* v8 := compiler.MapValueForKey(m, "items") if v8 != nil { var err error - x.Items, err = NewPrimitivesItems(v8, compiler.NewContext("items", context)) + x.Items, err = NewPrimitivesItems(v8, compiler.NewContext("items", v8, context)) if err != nil { errors = append(errors, err) } @@ -1103,15 +1108,15 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* // string collection_format = 9; v9 := compiler.MapValueForKey(m, "collectionFormat") if v9 != nil { - x.CollectionFormat, ok = v9.(string) + x.CollectionFormat, ok = compiler.StringForScalarNode(v9) if !ok { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v9, v9) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v9)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [csv ssv tsv pipes multi] if ok && !compiler.StringArrayContainsValue([]string{"csv", "ssv", "tsv", "pipes", "multi"}, x.CollectionFormat) { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v9, v9) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v9)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1119,7 +1124,7 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* v10 := compiler.MapValueForKey(m, "default") if v10 != nil { var err error - x.Default, err = NewAny(v10, compiler.NewContext("default", context)) + x.Default, err = NewAny(v10, compiler.NewContext("default", v10, context)) if err != nil { errors = append(errors, err) } @@ -1127,126 +1132,102 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* // float maximum = 11; v11 := compiler.MapValueForKey(m, "maximum") if v11 != nil { - switch v11 := v11.(type) { - case float64: - x.Maximum = v11 - case float32: - x.Maximum = float64(v11) - case uint64: - x.Maximum = float64(v11) - case uint32: - x.Maximum = float64(v11) - case int64: - x.Maximum = float64(v11) - case int32: - x.Maximum = float64(v11) - case int: - x.Maximum = float64(v11) - default: - message := fmt.Sprintf("has unexpected value for maximum: %+v (%T)", v11, v11) + v, ok := compiler.FloatForScalarNode(v11) + if ok { + x.Maximum = v + } else { + message := fmt.Sprintf("has unexpected value for maximum: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_maximum = 12; v12 := compiler.MapValueForKey(m, "exclusiveMaximum") if v12 != nil { - x.ExclusiveMaximum, ok = v12.(bool) + x.ExclusiveMaximum, ok = compiler.BoolForScalarNode(v12) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %+v (%T)", v12, v12) + message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %s", compiler.Display(v12)) errors = append(errors, compiler.NewError(context, message)) } } // float minimum = 13; v13 := compiler.MapValueForKey(m, "minimum") if v13 != nil { - switch v13 := v13.(type) { - case float64: - x.Minimum = v13 - case float32: - x.Minimum = float64(v13) - case uint64: - x.Minimum = float64(v13) - case uint32: - x.Minimum = float64(v13) - case int64: - x.Minimum = float64(v13) - case int32: - x.Minimum = float64(v13) - case int: - x.Minimum = float64(v13) - default: - message := fmt.Sprintf("has unexpected value for minimum: %+v (%T)", v13, v13) + v, ok := compiler.FloatForScalarNode(v13) + if ok { + x.Minimum = v + } else { + message := fmt.Sprintf("has unexpected value for minimum: %s", compiler.Display(v13)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_minimum = 14; v14 := compiler.MapValueForKey(m, "exclusiveMinimum") if v14 != nil { - x.ExclusiveMinimum, ok = v14.(bool) + x.ExclusiveMinimum, ok = compiler.BoolForScalarNode(v14) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %+v (%T)", v14, v14) + message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %s", compiler.Display(v14)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_length = 15; v15 := compiler.MapValueForKey(m, "maxLength") if v15 != nil { - t, ok := v15.(int) + t, ok := compiler.IntForScalarNode(v15) if ok { x.MaxLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxLength: %+v (%T)", v15, v15) + message := fmt.Sprintf("has unexpected value for maxLength: %s", compiler.Display(v15)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_length = 16; v16 := compiler.MapValueForKey(m, "minLength") if v16 != nil { - t, ok := v16.(int) + t, ok := compiler.IntForScalarNode(v16) if ok { x.MinLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minLength: %+v (%T)", v16, v16) + message := fmt.Sprintf("has unexpected value for minLength: %s", compiler.Display(v16)) errors = append(errors, compiler.NewError(context, message)) } } // string pattern = 17; v17 := compiler.MapValueForKey(m, "pattern") if v17 != nil { - x.Pattern, ok = v17.(string) + x.Pattern, ok = compiler.StringForScalarNode(v17) if !ok { - message := fmt.Sprintf("has unexpected value for pattern: %+v (%T)", v17, v17) + message := fmt.Sprintf("has unexpected value for pattern: %s", compiler.Display(v17)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_items = 18; v18 := compiler.MapValueForKey(m, "maxItems") if v18 != nil { - t, ok := v18.(int) + t, ok := compiler.IntForScalarNode(v18) if ok { x.MaxItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxItems: %+v (%T)", v18, v18) + message := fmt.Sprintf("has unexpected value for maxItems: %s", compiler.Display(v18)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_items = 19; v19 := compiler.MapValueForKey(m, "minItems") if v19 != nil { - t, ok := v19.(int) + t, ok := compiler.IntForScalarNode(v19) if ok { x.MinItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minItems: %+v (%T)", v19, v19) + message := fmt.Sprintf("has unexpected value for minItems: %s", compiler.Display(v19)) errors = append(errors, compiler.NewError(context, message)) } } // bool unique_items = 20; v20 := compiler.MapValueForKey(m, "uniqueItems") if v20 != nil { - x.UniqueItems, ok = v20.(bool) + x.UniqueItems, ok = compiler.BoolForScalarNode(v20) if !ok { - message := fmt.Sprintf("has unexpected value for uniqueItems: %+v (%T)", v20, v20) + message := fmt.Sprintf("has unexpected value for uniqueItems: %s", compiler.Display(v20)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1255,10 +1236,10 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* if v21 != nil { // repeated Any x.Enum = make([]*Any, 0) - a, ok := v21.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v21) if ok { - for _, item := range a { - y, err := NewAny(item, compiler.NewContext("enum", context)) + for _, item := range a.Content { + y, err := NewAny(item, compiler.NewContext("enum", item, context)) if err != nil { errors = append(errors, err) } @@ -1269,49 +1250,37 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* // float multiple_of = 22; v22 := compiler.MapValueForKey(m, "multipleOf") if v22 != nil { - switch v22 := v22.(type) { - case float64: - x.MultipleOf = v22 - case float32: - x.MultipleOf = float64(v22) - case uint64: - x.MultipleOf = float64(v22) - case uint32: - x.MultipleOf = float64(v22) - case int64: - x.MultipleOf = float64(v22) - case int32: - x.MultipleOf = float64(v22) - case int: - x.MultipleOf = float64(v22) - default: - message := fmt.Sprintf("has unexpected value for multipleOf: %+v (%T)", v22, v22) + v, ok := compiler.FloatForScalarNode(v22) + if ok { + x.MultipleOf = v + } else { + message := fmt.Sprintf("has unexpected value for multipleOf: %s", compiler.Display(v22)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 23; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -1325,7 +1294,7 @@ func NewFormDataParameterSubSchema(in interface{}, context *compiler.Context) (* } // NewHeader creates an object of type Header if possible, returning an error if not. -func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { +func NewHeader(in *yaml.Node, context *compiler.Context) (*Header, error) { errors := make([]error, 0) x := &Header{} m, ok := compiler.UnpackMap(in) @@ -1349,24 +1318,24 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [string number integer boolean array] if ok && !compiler.StringArrayContainsValue([]string{"string", "number", "integer", "boolean", "array"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string format = 2; v2 := compiler.MapValueForKey(m, "format") if v2 != nil { - x.Format, ok = v2.(string) + x.Format, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1374,7 +1343,7 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { v3 := compiler.MapValueForKey(m, "items") if v3 != nil { var err error - x.Items, err = NewPrimitivesItems(v3, compiler.NewContext("items", context)) + x.Items, err = NewPrimitivesItems(v3, compiler.NewContext("items", v3, context)) if err != nil { errors = append(errors, err) } @@ -1382,15 +1351,15 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { // string collection_format = 4; v4 := compiler.MapValueForKey(m, "collectionFormat") if v4 != nil { - x.CollectionFormat, ok = v4.(string) + x.CollectionFormat, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [csv ssv tsv pipes] if ok && !compiler.StringArrayContainsValue([]string{"csv", "ssv", "tsv", "pipes"}, x.CollectionFormat) { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1398,7 +1367,7 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { v5 := compiler.MapValueForKey(m, "default") if v5 != nil { var err error - x.Default, err = NewAny(v5, compiler.NewContext("default", context)) + x.Default, err = NewAny(v5, compiler.NewContext("default", v5, context)) if err != nil { errors = append(errors, err) } @@ -1406,126 +1375,102 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { // float maximum = 6; v6 := compiler.MapValueForKey(m, "maximum") if v6 != nil { - switch v6 := v6.(type) { - case float64: - x.Maximum = v6 - case float32: - x.Maximum = float64(v6) - case uint64: - x.Maximum = float64(v6) - case uint32: - x.Maximum = float64(v6) - case int64: - x.Maximum = float64(v6) - case int32: - x.Maximum = float64(v6) - case int: - x.Maximum = float64(v6) - default: - message := fmt.Sprintf("has unexpected value for maximum: %+v (%T)", v6, v6) + v, ok := compiler.FloatForScalarNode(v6) + if ok { + x.Maximum = v + } else { + message := fmt.Sprintf("has unexpected value for maximum: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_maximum = 7; v7 := compiler.MapValueForKey(m, "exclusiveMaximum") if v7 != nil { - x.ExclusiveMaximum, ok = v7.(bool) + x.ExclusiveMaximum, ok = compiler.BoolForScalarNode(v7) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %+v (%T)", v7, v7) + message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } // float minimum = 8; v8 := compiler.MapValueForKey(m, "minimum") if v8 != nil { - switch v8 := v8.(type) { - case float64: - x.Minimum = v8 - case float32: - x.Minimum = float64(v8) - case uint64: - x.Minimum = float64(v8) - case uint32: - x.Minimum = float64(v8) - case int64: - x.Minimum = float64(v8) - case int32: - x.Minimum = float64(v8) - case int: - x.Minimum = float64(v8) - default: - message := fmt.Sprintf("has unexpected value for minimum: %+v (%T)", v8, v8) + v, ok := compiler.FloatForScalarNode(v8) + if ok { + x.Minimum = v + } else { + message := fmt.Sprintf("has unexpected value for minimum: %s", compiler.Display(v8)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_minimum = 9; v9 := compiler.MapValueForKey(m, "exclusiveMinimum") if v9 != nil { - x.ExclusiveMinimum, ok = v9.(bool) + x.ExclusiveMinimum, ok = compiler.BoolForScalarNode(v9) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %+v (%T)", v9, v9) + message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %s", compiler.Display(v9)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_length = 10; v10 := compiler.MapValueForKey(m, "maxLength") if v10 != nil { - t, ok := v10.(int) + t, ok := compiler.IntForScalarNode(v10) if ok { x.MaxLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxLength: %+v (%T)", v10, v10) + message := fmt.Sprintf("has unexpected value for maxLength: %s", compiler.Display(v10)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_length = 11; v11 := compiler.MapValueForKey(m, "minLength") if v11 != nil { - t, ok := v11.(int) + t, ok := compiler.IntForScalarNode(v11) if ok { x.MinLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minLength: %+v (%T)", v11, v11) + message := fmt.Sprintf("has unexpected value for minLength: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } // string pattern = 12; v12 := compiler.MapValueForKey(m, "pattern") if v12 != nil { - x.Pattern, ok = v12.(string) + x.Pattern, ok = compiler.StringForScalarNode(v12) if !ok { - message := fmt.Sprintf("has unexpected value for pattern: %+v (%T)", v12, v12) + message := fmt.Sprintf("has unexpected value for pattern: %s", compiler.Display(v12)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_items = 13; v13 := compiler.MapValueForKey(m, "maxItems") if v13 != nil { - t, ok := v13.(int) + t, ok := compiler.IntForScalarNode(v13) if ok { x.MaxItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxItems: %+v (%T)", v13, v13) + message := fmt.Sprintf("has unexpected value for maxItems: %s", compiler.Display(v13)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_items = 14; v14 := compiler.MapValueForKey(m, "minItems") if v14 != nil { - t, ok := v14.(int) + t, ok := compiler.IntForScalarNode(v14) if ok { x.MinItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minItems: %+v (%T)", v14, v14) + message := fmt.Sprintf("has unexpected value for minItems: %s", compiler.Display(v14)) errors = append(errors, compiler.NewError(context, message)) } } // bool unique_items = 15; v15 := compiler.MapValueForKey(m, "uniqueItems") if v15 != nil { - x.UniqueItems, ok = v15.(bool) + x.UniqueItems, ok = compiler.BoolForScalarNode(v15) if !ok { - message := fmt.Sprintf("has unexpected value for uniqueItems: %+v (%T)", v15, v15) + message := fmt.Sprintf("has unexpected value for uniqueItems: %s", compiler.Display(v15)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1534,10 +1479,10 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { if v16 != nil { // repeated Any x.Enum = make([]*Any, 0) - a, ok := v16.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v16) if ok { - for _, item := range a { - y, err := NewAny(item, compiler.NewContext("enum", context)) + for _, item := range a.Content { + y, err := NewAny(item, compiler.NewContext("enum", item, context)) if err != nil { errors = append(errors, err) } @@ -1548,58 +1493,46 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { // float multiple_of = 17; v17 := compiler.MapValueForKey(m, "multipleOf") if v17 != nil { - switch v17 := v17.(type) { - case float64: - x.MultipleOf = v17 - case float32: - x.MultipleOf = float64(v17) - case uint64: - x.MultipleOf = float64(v17) - case uint32: - x.MultipleOf = float64(v17) - case int64: - x.MultipleOf = float64(v17) - case int32: - x.MultipleOf = float64(v17) - case int: - x.MultipleOf = float64(v17) - default: - message := fmt.Sprintf("has unexpected value for multipleOf: %+v (%T)", v17, v17) + v, ok := compiler.FloatForScalarNode(v17) + if ok { + x.MultipleOf = v + } else { + message := fmt.Sprintf("has unexpected value for multipleOf: %s", compiler.Display(v17)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 18; v18 := compiler.MapValueForKey(m, "description") if v18 != nil { - x.Description, ok = v18.(string) + x.Description, ok = compiler.StringForScalarNode(v18) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v18, v18) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v18)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 19; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -1613,7 +1546,7 @@ func NewHeader(in interface{}, context *compiler.Context) (*Header, error) { } // NewHeaderParameterSubSchema creates an object of type HeaderParameterSubSchema if possible, returning an error if not. -func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*HeaderParameterSubSchema, error) { +func NewHeaderParameterSubSchema(in *yaml.Node, context *compiler.Context) (*HeaderParameterSubSchema, error) { errors := make([]error, 0) x := &HeaderParameterSubSchema{} m, ok := compiler.UnpackMap(in) @@ -1631,66 +1564,66 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He // bool required = 1; v1 := compiler.MapValueForKey(m, "required") if v1 != nil { - x.Required, ok = v1.(bool) + x.Required, ok = compiler.BoolForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for required: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for required: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string in = 2; v2 := compiler.MapValueForKey(m, "in") if v2 != nil { - x.In, ok = v2.(string) + x.In, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [header] if ok && !compiler.StringArrayContainsValue([]string{"header"}, x.In) { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 3; v3 := compiler.MapValueForKey(m, "description") if v3 != nil { - x.Description, ok = v3.(string) + x.Description, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string name = 4; v4 := compiler.MapValueForKey(m, "name") if v4 != nil { - x.Name, ok = v4.(string) + x.Name, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // string type = 5; v5 := compiler.MapValueForKey(m, "type") if v5 != nil { - x.Type, ok = v5.(string) + x.Type, ok = compiler.StringForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [string number boolean integer array] if ok && !compiler.StringArrayContainsValue([]string{"string", "number", "boolean", "integer", "array"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // string format = 6; v6 := compiler.MapValueForKey(m, "format") if v6 != nil { - x.Format, ok = v6.(string) + x.Format, ok = compiler.StringForScalarNode(v6) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1698,7 +1631,7 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He v7 := compiler.MapValueForKey(m, "items") if v7 != nil { var err error - x.Items, err = NewPrimitivesItems(v7, compiler.NewContext("items", context)) + x.Items, err = NewPrimitivesItems(v7, compiler.NewContext("items", v7, context)) if err != nil { errors = append(errors, err) } @@ -1706,15 +1639,15 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He // string collection_format = 8; v8 := compiler.MapValueForKey(m, "collectionFormat") if v8 != nil { - x.CollectionFormat, ok = v8.(string) + x.CollectionFormat, ok = compiler.StringForScalarNode(v8) if !ok { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v8, v8) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v8)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [csv ssv tsv pipes] if ok && !compiler.StringArrayContainsValue([]string{"csv", "ssv", "tsv", "pipes"}, x.CollectionFormat) { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v8, v8) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v8)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1722,7 +1655,7 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He v9 := compiler.MapValueForKey(m, "default") if v9 != nil { var err error - x.Default, err = NewAny(v9, compiler.NewContext("default", context)) + x.Default, err = NewAny(v9, compiler.NewContext("default", v9, context)) if err != nil { errors = append(errors, err) } @@ -1730,126 +1663,102 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He // float maximum = 10; v10 := compiler.MapValueForKey(m, "maximum") if v10 != nil { - switch v10 := v10.(type) { - case float64: - x.Maximum = v10 - case float32: - x.Maximum = float64(v10) - case uint64: - x.Maximum = float64(v10) - case uint32: - x.Maximum = float64(v10) - case int64: - x.Maximum = float64(v10) - case int32: - x.Maximum = float64(v10) - case int: - x.Maximum = float64(v10) - default: - message := fmt.Sprintf("has unexpected value for maximum: %+v (%T)", v10, v10) + v, ok := compiler.FloatForScalarNode(v10) + if ok { + x.Maximum = v + } else { + message := fmt.Sprintf("has unexpected value for maximum: %s", compiler.Display(v10)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_maximum = 11; v11 := compiler.MapValueForKey(m, "exclusiveMaximum") if v11 != nil { - x.ExclusiveMaximum, ok = v11.(bool) + x.ExclusiveMaximum, ok = compiler.BoolForScalarNode(v11) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %+v (%T)", v11, v11) + message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } // float minimum = 12; v12 := compiler.MapValueForKey(m, "minimum") if v12 != nil { - switch v12 := v12.(type) { - case float64: - x.Minimum = v12 - case float32: - x.Minimum = float64(v12) - case uint64: - x.Minimum = float64(v12) - case uint32: - x.Minimum = float64(v12) - case int64: - x.Minimum = float64(v12) - case int32: - x.Minimum = float64(v12) - case int: - x.Minimum = float64(v12) - default: - message := fmt.Sprintf("has unexpected value for minimum: %+v (%T)", v12, v12) + v, ok := compiler.FloatForScalarNode(v12) + if ok { + x.Minimum = v + } else { + message := fmt.Sprintf("has unexpected value for minimum: %s", compiler.Display(v12)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_minimum = 13; v13 := compiler.MapValueForKey(m, "exclusiveMinimum") if v13 != nil { - x.ExclusiveMinimum, ok = v13.(bool) + x.ExclusiveMinimum, ok = compiler.BoolForScalarNode(v13) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %+v (%T)", v13, v13) + message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %s", compiler.Display(v13)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_length = 14; v14 := compiler.MapValueForKey(m, "maxLength") if v14 != nil { - t, ok := v14.(int) + t, ok := compiler.IntForScalarNode(v14) if ok { x.MaxLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxLength: %+v (%T)", v14, v14) + message := fmt.Sprintf("has unexpected value for maxLength: %s", compiler.Display(v14)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_length = 15; v15 := compiler.MapValueForKey(m, "minLength") if v15 != nil { - t, ok := v15.(int) + t, ok := compiler.IntForScalarNode(v15) if ok { x.MinLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minLength: %+v (%T)", v15, v15) + message := fmt.Sprintf("has unexpected value for minLength: %s", compiler.Display(v15)) errors = append(errors, compiler.NewError(context, message)) } } // string pattern = 16; v16 := compiler.MapValueForKey(m, "pattern") if v16 != nil { - x.Pattern, ok = v16.(string) + x.Pattern, ok = compiler.StringForScalarNode(v16) if !ok { - message := fmt.Sprintf("has unexpected value for pattern: %+v (%T)", v16, v16) + message := fmt.Sprintf("has unexpected value for pattern: %s", compiler.Display(v16)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_items = 17; v17 := compiler.MapValueForKey(m, "maxItems") if v17 != nil { - t, ok := v17.(int) + t, ok := compiler.IntForScalarNode(v17) if ok { x.MaxItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxItems: %+v (%T)", v17, v17) + message := fmt.Sprintf("has unexpected value for maxItems: %s", compiler.Display(v17)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_items = 18; v18 := compiler.MapValueForKey(m, "minItems") if v18 != nil { - t, ok := v18.(int) + t, ok := compiler.IntForScalarNode(v18) if ok { x.MinItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minItems: %+v (%T)", v18, v18) + message := fmt.Sprintf("has unexpected value for minItems: %s", compiler.Display(v18)) errors = append(errors, compiler.NewError(context, message)) } } // bool unique_items = 19; v19 := compiler.MapValueForKey(m, "uniqueItems") if v19 != nil { - x.UniqueItems, ok = v19.(bool) + x.UniqueItems, ok = compiler.BoolForScalarNode(v19) if !ok { - message := fmt.Sprintf("has unexpected value for uniqueItems: %+v (%T)", v19, v19) + message := fmt.Sprintf("has unexpected value for uniqueItems: %s", compiler.Display(v19)) errors = append(errors, compiler.NewError(context, message)) } } @@ -1858,10 +1767,10 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He if v20 != nil { // repeated Any x.Enum = make([]*Any, 0) - a, ok := v20.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v20) if ok { - for _, item := range a { - y, err := NewAny(item, compiler.NewContext("enum", context)) + for _, item := range a.Content { + y, err := NewAny(item, compiler.NewContext("enum", item, context)) if err != nil { errors = append(errors, err) } @@ -1872,49 +1781,37 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He // float multiple_of = 21; v21 := compiler.MapValueForKey(m, "multipleOf") if v21 != nil { - switch v21 := v21.(type) { - case float64: - x.MultipleOf = v21 - case float32: - x.MultipleOf = float64(v21) - case uint64: - x.MultipleOf = float64(v21) - case uint32: - x.MultipleOf = float64(v21) - case int64: - x.MultipleOf = float64(v21) - case int32: - x.MultipleOf = float64(v21) - case int: - x.MultipleOf = float64(v21) - default: - message := fmt.Sprintf("has unexpected value for multipleOf: %+v (%T)", v21, v21) + v, ok := compiler.FloatForScalarNode(v21) + if ok { + x.MultipleOf = v + } else { + message := fmt.Sprintf("has unexpected value for multipleOf: %s", compiler.Display(v21)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 22; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -1928,7 +1825,7 @@ func NewHeaderParameterSubSchema(in interface{}, context *compiler.Context) (*He } // NewHeaders creates an object of type Headers if possible, returning an error if not. -func NewHeaders(in interface{}, context *compiler.Context) (*Headers, error) { +func NewHeaders(in *yaml.Node, context *compiler.Context) (*Headers, error) { errors := make([]error, 0) x := &Headers{} m, ok := compiler.UnpackMap(in) @@ -1939,14 +1836,14 @@ func NewHeaders(in interface{}, context *compiler.Context) (*Headers, error) { // repeated NamedHeader additional_properties = 1; // MAP: Header x.AdditionalProperties = make([]*NamedHeader, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedHeader{} pair.Name = k var err error - pair.Value, err = NewHeader(v, compiler.NewContext(k, context)) + pair.Value, err = NewHeader(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -1958,7 +1855,7 @@ func NewHeaders(in interface{}, context *compiler.Context) (*Headers, error) { } // NewInfo creates an object of type Info if possible, returning an error if not. -func NewInfo(in interface{}, context *compiler.Context) (*Info, error) { +func NewInfo(in *yaml.Node, context *compiler.Context) (*Info, error) { errors := make([]error, 0) x := &Info{} m, ok := compiler.UnpackMap(in) @@ -1982,36 +1879,36 @@ func NewInfo(in interface{}, context *compiler.Context) (*Info, error) { // string title = 1; v1 := compiler.MapValueForKey(m, "title") if v1 != nil { - x.Title, ok = v1.(string) + x.Title, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for title: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for title: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string version = 2; v2 := compiler.MapValueForKey(m, "version") if v2 != nil { - x.Version, ok = v2.(string) + x.Version, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for version: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for version: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 3; v3 := compiler.MapValueForKey(m, "description") if v3 != nil { - x.Description, ok = v3.(string) + x.Description, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string terms_of_service = 4; v4 := compiler.MapValueForKey(m, "termsOfService") if v4 != nil { - x.TermsOfService, ok = v4.(string) + x.TermsOfService, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for termsOfService: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for termsOfService: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2019,7 +1916,7 @@ func NewInfo(in interface{}, context *compiler.Context) (*Info, error) { v5 := compiler.MapValueForKey(m, "contact") if v5 != nil { var err error - x.Contact, err = NewContact(v5, compiler.NewContext("contact", context)) + x.Contact, err = NewContact(v5, compiler.NewContext("contact", v5, context)) if err != nil { errors = append(errors, err) } @@ -2028,7 +1925,7 @@ func NewInfo(in interface{}, context *compiler.Context) (*Info, error) { v6 := compiler.MapValueForKey(m, "license") if v6 != nil { var err error - x.License, err = NewLicense(v6, compiler.NewContext("license", context)) + x.License, err = NewLicense(v6, compiler.NewContext("license", v6, context)) if err != nil { errors = append(errors, err) } @@ -2036,26 +1933,26 @@ func NewInfo(in interface{}, context *compiler.Context) (*Info, error) { // repeated NamedAny vendor_extension = 7; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -2069,7 +1966,7 @@ func NewInfo(in interface{}, context *compiler.Context) (*Info, error) { } // NewItemsItem creates an object of type ItemsItem if possible, returning an error if not. -func NewItemsItem(in interface{}, context *compiler.Context) (*ItemsItem, error) { +func NewItemsItem(in *yaml.Node, context *compiler.Context) (*ItemsItem, error) { errors := make([]error, 0) x := &ItemsItem{} m, ok := compiler.UnpackMap(in) @@ -2078,7 +1975,7 @@ func NewItemsItem(in interface{}, context *compiler.Context) (*ItemsItem, error) errors = append(errors, compiler.NewError(context, message)) } else { x.Schema = make([]*Schema, 0) - y, err := NewSchema(m, compiler.NewContext("", context)) + y, err := NewSchema(m, compiler.NewContext("", m, context)) if err != nil { return nil, err } @@ -2088,7 +1985,7 @@ func NewItemsItem(in interface{}, context *compiler.Context) (*ItemsItem, error) } // NewJsonReference creates an object of type JsonReference if possible, returning an error if not. -func NewJsonReference(in interface{}, context *compiler.Context) (*JsonReference, error) { +func NewJsonReference(in *yaml.Node, context *compiler.Context) (*JsonReference, error) { errors := make([]error, 0) x := &JsonReference{} m, ok := compiler.UnpackMap(in) @@ -2102,28 +1999,21 @@ func NewJsonReference(in interface{}, context *compiler.Context) (*JsonReference message := fmt.Sprintf("is missing required %s: %+v", compiler.PluralProperties(len(missingKeys)), strings.Join(missingKeys, ", ")) errors = append(errors, compiler.NewError(context, message)) } - allowedKeys := []string{"$ref", "description"} - var allowedPatterns []*regexp.Regexp - invalidKeys := compiler.InvalidKeysInMap(m, allowedKeys, allowedPatterns) - if len(invalidKeys) > 0 { - message := fmt.Sprintf("has invalid %s: %+v", compiler.PluralProperties(len(invalidKeys)), strings.Join(invalidKeys, ", ")) - errors = append(errors, compiler.NewError(context, message)) - } // string _ref = 1; v1 := compiler.MapValueForKey(m, "$ref") if v1 != nil { - x.XRef, ok = v1.(string) + x.XRef, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for $ref: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for $ref: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 2; v2 := compiler.MapValueForKey(m, "description") if v2 != nil { - x.Description, ok = v2.(string) + x.Description, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2132,7 +2022,7 @@ func NewJsonReference(in interface{}, context *compiler.Context) (*JsonReference } // NewLicense creates an object of type License if possible, returning an error if not. -func NewLicense(in interface{}, context *compiler.Context) (*License, error) { +func NewLicense(in *yaml.Node, context *compiler.Context) (*License, error) { errors := make([]error, 0) x := &License{} m, ok := compiler.UnpackMap(in) @@ -2156,44 +2046,44 @@ func NewLicense(in interface{}, context *compiler.Context) (*License, error) { // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string url = 2; v2 := compiler.MapValueForKey(m, "url") if v2 != nil { - x.Url, ok = v2.(string) + x.Url, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for url: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for url: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 3; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -2207,7 +2097,7 @@ func NewLicense(in interface{}, context *compiler.Context) (*License, error) { } // NewNamedAny creates an object of type NamedAny if possible, returning an error if not. -func NewNamedAny(in interface{}, context *compiler.Context) (*NamedAny, error) { +func NewNamedAny(in *yaml.Node, context *compiler.Context) (*NamedAny, error) { errors := make([]error, 0) x := &NamedAny{} m, ok := compiler.UnpackMap(in) @@ -2225,9 +2115,9 @@ func NewNamedAny(in interface{}, context *compiler.Context) (*NamedAny, error) { // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2235,7 +2125,7 @@ func NewNamedAny(in interface{}, context *compiler.Context) (*NamedAny, error) { v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewAny(v2, compiler.NewContext("value", context)) + x.Value, err = NewAny(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2245,7 +2135,7 @@ func NewNamedAny(in interface{}, context *compiler.Context) (*NamedAny, error) { } // NewNamedHeader creates an object of type NamedHeader if possible, returning an error if not. -func NewNamedHeader(in interface{}, context *compiler.Context) (*NamedHeader, error) { +func NewNamedHeader(in *yaml.Node, context *compiler.Context) (*NamedHeader, error) { errors := make([]error, 0) x := &NamedHeader{} m, ok := compiler.UnpackMap(in) @@ -2263,9 +2153,9 @@ func NewNamedHeader(in interface{}, context *compiler.Context) (*NamedHeader, er // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2273,7 +2163,7 @@ func NewNamedHeader(in interface{}, context *compiler.Context) (*NamedHeader, er v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewHeader(v2, compiler.NewContext("value", context)) + x.Value, err = NewHeader(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2283,7 +2173,7 @@ func NewNamedHeader(in interface{}, context *compiler.Context) (*NamedHeader, er } // NewNamedParameter creates an object of type NamedParameter if possible, returning an error if not. -func NewNamedParameter(in interface{}, context *compiler.Context) (*NamedParameter, error) { +func NewNamedParameter(in *yaml.Node, context *compiler.Context) (*NamedParameter, error) { errors := make([]error, 0) x := &NamedParameter{} m, ok := compiler.UnpackMap(in) @@ -2301,9 +2191,9 @@ func NewNamedParameter(in interface{}, context *compiler.Context) (*NamedParamet // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2311,7 +2201,7 @@ func NewNamedParameter(in interface{}, context *compiler.Context) (*NamedParamet v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewParameter(v2, compiler.NewContext("value", context)) + x.Value, err = NewParameter(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2321,7 +2211,7 @@ func NewNamedParameter(in interface{}, context *compiler.Context) (*NamedParamet } // NewNamedPathItem creates an object of type NamedPathItem if possible, returning an error if not. -func NewNamedPathItem(in interface{}, context *compiler.Context) (*NamedPathItem, error) { +func NewNamedPathItem(in *yaml.Node, context *compiler.Context) (*NamedPathItem, error) { errors := make([]error, 0) x := &NamedPathItem{} m, ok := compiler.UnpackMap(in) @@ -2339,9 +2229,9 @@ func NewNamedPathItem(in interface{}, context *compiler.Context) (*NamedPathItem // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2349,7 +2239,7 @@ func NewNamedPathItem(in interface{}, context *compiler.Context) (*NamedPathItem v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewPathItem(v2, compiler.NewContext("value", context)) + x.Value, err = NewPathItem(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2359,7 +2249,7 @@ func NewNamedPathItem(in interface{}, context *compiler.Context) (*NamedPathItem } // NewNamedResponse creates an object of type NamedResponse if possible, returning an error if not. -func NewNamedResponse(in interface{}, context *compiler.Context) (*NamedResponse, error) { +func NewNamedResponse(in *yaml.Node, context *compiler.Context) (*NamedResponse, error) { errors := make([]error, 0) x := &NamedResponse{} m, ok := compiler.UnpackMap(in) @@ -2377,9 +2267,9 @@ func NewNamedResponse(in interface{}, context *compiler.Context) (*NamedResponse // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2387,7 +2277,7 @@ func NewNamedResponse(in interface{}, context *compiler.Context) (*NamedResponse v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewResponse(v2, compiler.NewContext("value", context)) + x.Value, err = NewResponse(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2397,7 +2287,7 @@ func NewNamedResponse(in interface{}, context *compiler.Context) (*NamedResponse } // NewNamedResponseValue creates an object of type NamedResponseValue if possible, returning an error if not. -func NewNamedResponseValue(in interface{}, context *compiler.Context) (*NamedResponseValue, error) { +func NewNamedResponseValue(in *yaml.Node, context *compiler.Context) (*NamedResponseValue, error) { errors := make([]error, 0) x := &NamedResponseValue{} m, ok := compiler.UnpackMap(in) @@ -2415,9 +2305,9 @@ func NewNamedResponseValue(in interface{}, context *compiler.Context) (*NamedRes // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2425,7 +2315,7 @@ func NewNamedResponseValue(in interface{}, context *compiler.Context) (*NamedRes v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewResponseValue(v2, compiler.NewContext("value", context)) + x.Value, err = NewResponseValue(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2435,7 +2325,7 @@ func NewNamedResponseValue(in interface{}, context *compiler.Context) (*NamedRes } // NewNamedSchema creates an object of type NamedSchema if possible, returning an error if not. -func NewNamedSchema(in interface{}, context *compiler.Context) (*NamedSchema, error) { +func NewNamedSchema(in *yaml.Node, context *compiler.Context) (*NamedSchema, error) { errors := make([]error, 0) x := &NamedSchema{} m, ok := compiler.UnpackMap(in) @@ -2453,9 +2343,9 @@ func NewNamedSchema(in interface{}, context *compiler.Context) (*NamedSchema, er // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2463,7 +2353,7 @@ func NewNamedSchema(in interface{}, context *compiler.Context) (*NamedSchema, er v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewSchema(v2, compiler.NewContext("value", context)) + x.Value, err = NewSchema(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2473,7 +2363,7 @@ func NewNamedSchema(in interface{}, context *compiler.Context) (*NamedSchema, er } // NewNamedSecurityDefinitionsItem creates an object of type NamedSecurityDefinitionsItem if possible, returning an error if not. -func NewNamedSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*NamedSecurityDefinitionsItem, error) { +func NewNamedSecurityDefinitionsItem(in *yaml.Node, context *compiler.Context) (*NamedSecurityDefinitionsItem, error) { errors := make([]error, 0) x := &NamedSecurityDefinitionsItem{} m, ok := compiler.UnpackMap(in) @@ -2491,9 +2381,9 @@ func NewNamedSecurityDefinitionsItem(in interface{}, context *compiler.Context) // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2501,7 +2391,7 @@ func NewNamedSecurityDefinitionsItem(in interface{}, context *compiler.Context) v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewSecurityDefinitionsItem(v2, compiler.NewContext("value", context)) + x.Value, err = NewSecurityDefinitionsItem(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2511,7 +2401,7 @@ func NewNamedSecurityDefinitionsItem(in interface{}, context *compiler.Context) } // NewNamedString creates an object of type NamedString if possible, returning an error if not. -func NewNamedString(in interface{}, context *compiler.Context) (*NamedString, error) { +func NewNamedString(in *yaml.Node, context *compiler.Context) (*NamedString, error) { errors := make([]error, 0) x := &NamedString{} m, ok := compiler.UnpackMap(in) @@ -2529,18 +2419,18 @@ func NewNamedString(in interface{}, context *compiler.Context) (*NamedString, er // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string value = 2; v2 := compiler.MapValueForKey(m, "value") if v2 != nil { - x.Value, ok = v2.(string) + x.Value, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for value: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for value: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2549,7 +2439,7 @@ func NewNamedString(in interface{}, context *compiler.Context) (*NamedString, er } // NewNamedStringArray creates an object of type NamedStringArray if possible, returning an error if not. -func NewNamedStringArray(in interface{}, context *compiler.Context) (*NamedStringArray, error) { +func NewNamedStringArray(in *yaml.Node, context *compiler.Context) (*NamedStringArray, error) { errors := make([]error, 0) x := &NamedStringArray{} m, ok := compiler.UnpackMap(in) @@ -2567,9 +2457,9 @@ func NewNamedStringArray(in interface{}, context *compiler.Context) (*NamedStrin // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2577,7 +2467,7 @@ func NewNamedStringArray(in interface{}, context *compiler.Context) (*NamedStrin v2 := compiler.MapValueForKey(m, "value") if v2 != nil { var err error - x.Value, err = NewStringArray(v2, compiler.NewContext("value", context)) + x.Value, err = NewStringArray(v2, compiler.NewContext("value", v2, context)) if err != nil { errors = append(errors, err) } @@ -2587,7 +2477,7 @@ func NewNamedStringArray(in interface{}, context *compiler.Context) (*NamedStrin } // NewNonBodyParameter creates an object of type NonBodyParameter if possible, returning an error if not. -func NewNonBodyParameter(in interface{}, context *compiler.Context) (*NonBodyParameter, error) { +func NewNonBodyParameter(in *yaml.Node, context *compiler.Context) (*NonBodyParameter, error) { errors := make([]error, 0) x := &NonBodyParameter{} matched := false @@ -2605,7 +2495,7 @@ func NewNonBodyParameter(in interface{}, context *compiler.Context) (*NonBodyPar // HeaderParameterSubSchema header_parameter_sub_schema = 1; { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewHeaderParameterSubSchema(m, compiler.NewContext("headerParameterSubSchema", context)) + t, matchingError := NewHeaderParameterSubSchema(m, compiler.NewContext("headerParameterSubSchema", m, context)) if matchingError == nil { x.Oneof = &NonBodyParameter_HeaderParameterSubSchema{HeaderParameterSubSchema: t} matched = true @@ -2616,7 +2506,7 @@ func NewNonBodyParameter(in interface{}, context *compiler.Context) (*NonBodyPar // FormDataParameterSubSchema form_data_parameter_sub_schema = 2; { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewFormDataParameterSubSchema(m, compiler.NewContext("formDataParameterSubSchema", context)) + t, matchingError := NewFormDataParameterSubSchema(m, compiler.NewContext("formDataParameterSubSchema", m, context)) if matchingError == nil { x.Oneof = &NonBodyParameter_FormDataParameterSubSchema{FormDataParameterSubSchema: t} matched = true @@ -2627,7 +2517,7 @@ func NewNonBodyParameter(in interface{}, context *compiler.Context) (*NonBodyPar // QueryParameterSubSchema query_parameter_sub_schema = 3; { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewQueryParameterSubSchema(m, compiler.NewContext("queryParameterSubSchema", context)) + t, matchingError := NewQueryParameterSubSchema(m, compiler.NewContext("queryParameterSubSchema", m, context)) if matchingError == nil { x.Oneof = &NonBodyParameter_QueryParameterSubSchema{QueryParameterSubSchema: t} matched = true @@ -2638,7 +2528,7 @@ func NewNonBodyParameter(in interface{}, context *compiler.Context) (*NonBodyPar // PathParameterSubSchema path_parameter_sub_schema = 4; { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewPathParameterSubSchema(m, compiler.NewContext("pathParameterSubSchema", context)) + t, matchingError := NewPathParameterSubSchema(m, compiler.NewContext("pathParameterSubSchema", m, context)) if matchingError == nil { x.Oneof = &NonBodyParameter_PathParameterSubSchema{PathParameterSubSchema: t} matched = true @@ -2650,12 +2540,16 @@ func NewNonBodyParameter(in interface{}, context *compiler.Context) (*NonBodyPar if matched { // since the oneof matched one of its possibilities, discard any matching errors errors = make([]error, 0) + } else { + message := fmt.Sprintf("contains an invalid NonBodyParameter") + err := compiler.NewError(context, message) + errors = []error{err} } return x, compiler.NewErrorGroupOrNil(errors) } // NewOauth2AccessCodeSecurity creates an object of type Oauth2AccessCodeSecurity if possible, returning an error if not. -func NewOauth2AccessCodeSecurity(in interface{}, context *compiler.Context) (*Oauth2AccessCodeSecurity, error) { +func NewOauth2AccessCodeSecurity(in *yaml.Node, context *compiler.Context) (*Oauth2AccessCodeSecurity, error) { errors := make([]error, 0) x := &Oauth2AccessCodeSecurity{} m, ok := compiler.UnpackMap(in) @@ -2679,30 +2573,30 @@ func NewOauth2AccessCodeSecurity(in interface{}, context *compiler.Context) (*Oa // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [oauth2] if ok && !compiler.StringArrayContainsValue([]string{"oauth2"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string flow = 2; v2 := compiler.MapValueForKey(m, "flow") if v2 != nil { - x.Flow, ok = v2.(string) + x.Flow, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [accessCode] if ok && !compiler.StringArrayContainsValue([]string{"accessCode"}, x.Flow) { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2710,7 +2604,7 @@ func NewOauth2AccessCodeSecurity(in interface{}, context *compiler.Context) (*Oa v3 := compiler.MapValueForKey(m, "scopes") if v3 != nil { var err error - x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", context)) + x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", v3, context)) if err != nil { errors = append(errors, err) } @@ -2718,53 +2612,53 @@ func NewOauth2AccessCodeSecurity(in interface{}, context *compiler.Context) (*Oa // string authorization_url = 4; v4 := compiler.MapValueForKey(m, "authorizationUrl") if v4 != nil { - x.AuthorizationUrl, ok = v4.(string) + x.AuthorizationUrl, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for authorizationUrl: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for authorizationUrl: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // string token_url = 5; v5 := compiler.MapValueForKey(m, "tokenUrl") if v5 != nil { - x.TokenUrl, ok = v5.(string) + x.TokenUrl, ok = compiler.StringForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for tokenUrl: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for tokenUrl: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 6; v6 := compiler.MapValueForKey(m, "description") if v6 != nil { - x.Description, ok = v6.(string) + x.Description, ok = compiler.StringForScalarNode(v6) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 7; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -2778,7 +2672,7 @@ func NewOauth2AccessCodeSecurity(in interface{}, context *compiler.Context) (*Oa } // NewOauth2ApplicationSecurity creates an object of type Oauth2ApplicationSecurity if possible, returning an error if not. -func NewOauth2ApplicationSecurity(in interface{}, context *compiler.Context) (*Oauth2ApplicationSecurity, error) { +func NewOauth2ApplicationSecurity(in *yaml.Node, context *compiler.Context) (*Oauth2ApplicationSecurity, error) { errors := make([]error, 0) x := &Oauth2ApplicationSecurity{} m, ok := compiler.UnpackMap(in) @@ -2802,30 +2696,30 @@ func NewOauth2ApplicationSecurity(in interface{}, context *compiler.Context) (*O // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [oauth2] if ok && !compiler.StringArrayContainsValue([]string{"oauth2"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string flow = 2; v2 := compiler.MapValueForKey(m, "flow") if v2 != nil { - x.Flow, ok = v2.(string) + x.Flow, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [application] if ok && !compiler.StringArrayContainsValue([]string{"application"}, x.Flow) { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2833,7 +2727,7 @@ func NewOauth2ApplicationSecurity(in interface{}, context *compiler.Context) (*O v3 := compiler.MapValueForKey(m, "scopes") if v3 != nil { var err error - x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", context)) + x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", v3, context)) if err != nil { errors = append(errors, err) } @@ -2841,44 +2735,44 @@ func NewOauth2ApplicationSecurity(in interface{}, context *compiler.Context) (*O // string token_url = 4; v4 := compiler.MapValueForKey(m, "tokenUrl") if v4 != nil { - x.TokenUrl, ok = v4.(string) + x.TokenUrl, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for tokenUrl: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for tokenUrl: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 5; v5 := compiler.MapValueForKey(m, "description") if v5 != nil { - x.Description, ok = v5.(string) + x.Description, ok = compiler.StringForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 6; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -2892,7 +2786,7 @@ func NewOauth2ApplicationSecurity(in interface{}, context *compiler.Context) (*O } // NewOauth2ImplicitSecurity creates an object of type Oauth2ImplicitSecurity if possible, returning an error if not. -func NewOauth2ImplicitSecurity(in interface{}, context *compiler.Context) (*Oauth2ImplicitSecurity, error) { +func NewOauth2ImplicitSecurity(in *yaml.Node, context *compiler.Context) (*Oauth2ImplicitSecurity, error) { errors := make([]error, 0) x := &Oauth2ImplicitSecurity{} m, ok := compiler.UnpackMap(in) @@ -2916,30 +2810,30 @@ func NewOauth2ImplicitSecurity(in interface{}, context *compiler.Context) (*Oaut // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [oauth2] if ok && !compiler.StringArrayContainsValue([]string{"oauth2"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string flow = 2; v2 := compiler.MapValueForKey(m, "flow") if v2 != nil { - x.Flow, ok = v2.(string) + x.Flow, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [implicit] if ok && !compiler.StringArrayContainsValue([]string{"implicit"}, x.Flow) { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -2947,7 +2841,7 @@ func NewOauth2ImplicitSecurity(in interface{}, context *compiler.Context) (*Oaut v3 := compiler.MapValueForKey(m, "scopes") if v3 != nil { var err error - x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", context)) + x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", v3, context)) if err != nil { errors = append(errors, err) } @@ -2955,44 +2849,44 @@ func NewOauth2ImplicitSecurity(in interface{}, context *compiler.Context) (*Oaut // string authorization_url = 4; v4 := compiler.MapValueForKey(m, "authorizationUrl") if v4 != nil { - x.AuthorizationUrl, ok = v4.(string) + x.AuthorizationUrl, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for authorizationUrl: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for authorizationUrl: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 5; v5 := compiler.MapValueForKey(m, "description") if v5 != nil { - x.Description, ok = v5.(string) + x.Description, ok = compiler.StringForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 6; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3006,7 +2900,7 @@ func NewOauth2ImplicitSecurity(in interface{}, context *compiler.Context) (*Oaut } // NewOauth2PasswordSecurity creates an object of type Oauth2PasswordSecurity if possible, returning an error if not. -func NewOauth2PasswordSecurity(in interface{}, context *compiler.Context) (*Oauth2PasswordSecurity, error) { +func NewOauth2PasswordSecurity(in *yaml.Node, context *compiler.Context) (*Oauth2PasswordSecurity, error) { errors := make([]error, 0) x := &Oauth2PasswordSecurity{} m, ok := compiler.UnpackMap(in) @@ -3030,30 +2924,30 @@ func NewOauth2PasswordSecurity(in interface{}, context *compiler.Context) (*Oaut // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [oauth2] if ok && !compiler.StringArrayContainsValue([]string{"oauth2"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string flow = 2; v2 := compiler.MapValueForKey(m, "flow") if v2 != nil { - x.Flow, ok = v2.(string) + x.Flow, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [password] if ok && !compiler.StringArrayContainsValue([]string{"password"}, x.Flow) { - message := fmt.Sprintf("has unexpected value for flow: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for flow: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3061,7 +2955,7 @@ func NewOauth2PasswordSecurity(in interface{}, context *compiler.Context) (*Oaut v3 := compiler.MapValueForKey(m, "scopes") if v3 != nil { var err error - x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", context)) + x.Scopes, err = NewOauth2Scopes(v3, compiler.NewContext("scopes", v3, context)) if err != nil { errors = append(errors, err) } @@ -3069,44 +2963,44 @@ func NewOauth2PasswordSecurity(in interface{}, context *compiler.Context) (*Oaut // string token_url = 4; v4 := compiler.MapValueForKey(m, "tokenUrl") if v4 != nil { - x.TokenUrl, ok = v4.(string) + x.TokenUrl, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for tokenUrl: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for tokenUrl: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 5; v5 := compiler.MapValueForKey(m, "description") if v5 != nil { - x.Description, ok = v5.(string) + x.Description, ok = compiler.StringForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 6; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3120,7 +3014,7 @@ func NewOauth2PasswordSecurity(in interface{}, context *compiler.Context) (*Oaut } // NewOauth2Scopes creates an object of type Oauth2Scopes if possible, returning an error if not. -func NewOauth2Scopes(in interface{}, context *compiler.Context) (*Oauth2Scopes, error) { +func NewOauth2Scopes(in *yaml.Node, context *compiler.Context) (*Oauth2Scopes, error) { errors := make([]error, 0) x := &Oauth2Scopes{} m, ok := compiler.UnpackMap(in) @@ -3131,13 +3025,13 @@ func NewOauth2Scopes(in interface{}, context *compiler.Context) (*Oauth2Scopes, // repeated NamedString additional_properties = 1; // MAP: string x.AdditionalProperties = make([]*NamedString, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedString{} pair.Name = k - pair.Value = v.(string) + pair.Value, _ = compiler.StringForScalarNode(v) x.AdditionalProperties = append(x.AdditionalProperties, pair) } } @@ -3146,7 +3040,7 @@ func NewOauth2Scopes(in interface{}, context *compiler.Context) (*Oauth2Scopes, } // NewOperation creates an object of type Operation if possible, returning an error if not. -func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) { +func NewOperation(in *yaml.Node, context *compiler.Context) (*Operation, error) { errors := make([]error, 0) x := &Operation{} m, ok := compiler.UnpackMap(in) @@ -3170,29 +3064,29 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) // repeated string tags = 1; v1 := compiler.MapValueForKey(m, "tags") if v1 != nil { - v, ok := v1.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v1) if ok { - x.Tags = compiler.ConvertInterfaceArrayToStringArray(v) + x.Tags = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for tags: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for tags: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string summary = 2; v2 := compiler.MapValueForKey(m, "summary") if v2 != nil { - x.Summary, ok = v2.(string) + x.Summary, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for summary: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for summary: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 3; v3 := compiler.MapValueForKey(m, "description") if v3 != nil { - x.Description, ok = v3.(string) + x.Description, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3200,7 +3094,7 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) v4 := compiler.MapValueForKey(m, "externalDocs") if v4 != nil { var err error - x.ExternalDocs, err = NewExternalDocs(v4, compiler.NewContext("externalDocs", context)) + x.ExternalDocs, err = NewExternalDocs(v4, compiler.NewContext("externalDocs", v4, context)) if err != nil { errors = append(errors, err) } @@ -3208,31 +3102,31 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) // string operation_id = 5; v5 := compiler.MapValueForKey(m, "operationId") if v5 != nil { - x.OperationId, ok = v5.(string) + x.OperationId, ok = compiler.StringForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for operationId: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for operationId: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // repeated string produces = 6; v6 := compiler.MapValueForKey(m, "produces") if v6 != nil { - v, ok := v6.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v6) if ok { - x.Produces = compiler.ConvertInterfaceArrayToStringArray(v) + x.Produces = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for produces: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for produces: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // repeated string consumes = 7; v7 := compiler.MapValueForKey(m, "consumes") if v7 != nil { - v, ok := v7.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v7) if ok { - x.Consumes = compiler.ConvertInterfaceArrayToStringArray(v) + x.Consumes = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for consumes: %+v (%T)", v7, v7) + message := fmt.Sprintf("has unexpected value for consumes: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3241,10 +3135,10 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) if v8 != nil { // repeated ParametersItem x.Parameters = make([]*ParametersItem, 0) - a, ok := v8.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v8) if ok { - for _, item := range a { - y, err := NewParametersItem(item, compiler.NewContext("parameters", context)) + for _, item := range a.Content { + y, err := NewParametersItem(item, compiler.NewContext("parameters", item, context)) if err != nil { errors = append(errors, err) } @@ -3256,7 +3150,7 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) v9 := compiler.MapValueForKey(m, "responses") if v9 != nil { var err error - x.Responses, err = NewResponses(v9, compiler.NewContext("responses", context)) + x.Responses, err = NewResponses(v9, compiler.NewContext("responses", v9, context)) if err != nil { errors = append(errors, err) } @@ -3264,26 +3158,26 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) // repeated string schemes = 10; v10 := compiler.MapValueForKey(m, "schemes") if v10 != nil { - v, ok := v10.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v10) if ok { - x.Schemes = compiler.ConvertInterfaceArrayToStringArray(v) + x.Schemes = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for schemes: %+v (%T)", v10, v10) + message := fmt.Sprintf("has unexpected value for schemes: %s", compiler.Display(v10)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [http https ws wss] if ok && !compiler.StringArrayContainsValues([]string{"http", "https", "ws", "wss"}, x.Schemes) { - message := fmt.Sprintf("has unexpected value for schemes: %+v", v10) + message := fmt.Sprintf("has unexpected value for schemes: %s", compiler.Display(v10)) errors = append(errors, compiler.NewError(context, message)) } } // bool deprecated = 11; v11 := compiler.MapValueForKey(m, "deprecated") if v11 != nil { - x.Deprecated, ok = v11.(bool) + x.Deprecated, ok = compiler.BoolForScalarNode(v11) if !ok { - message := fmt.Sprintf("has unexpected value for deprecated: %+v (%T)", v11, v11) + message := fmt.Sprintf("has unexpected value for deprecated: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3292,10 +3186,10 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) if v12 != nil { // repeated SecurityRequirement x.Security = make([]*SecurityRequirement, 0) - a, ok := v12.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v12) if ok { - for _, item := range a { - y, err := NewSecurityRequirement(item, compiler.NewContext("security", context)) + for _, item := range a.Content { + y, err := NewSecurityRequirement(item, compiler.NewContext("security", item, context)) if err != nil { errors = append(errors, err) } @@ -3306,26 +3200,26 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) // repeated NamedAny vendor_extension = 13; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3339,7 +3233,7 @@ func NewOperation(in interface{}, context *compiler.Context) (*Operation, error) } // NewParameter creates an object of type Parameter if possible, returning an error if not. -func NewParameter(in interface{}, context *compiler.Context) (*Parameter, error) { +func NewParameter(in *yaml.Node, context *compiler.Context) (*Parameter, error) { errors := make([]error, 0) x := &Parameter{} matched := false @@ -3348,7 +3242,7 @@ func NewParameter(in interface{}, context *compiler.Context) (*Parameter, error) m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewBodyParameter(m, compiler.NewContext("bodyParameter", context)) + t, matchingError := NewBodyParameter(m, compiler.NewContext("bodyParameter", m, context)) if matchingError == nil { x.Oneof = &Parameter_BodyParameter{BodyParameter: t} matched = true @@ -3362,7 +3256,7 @@ func NewParameter(in interface{}, context *compiler.Context) (*Parameter, error) m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewNonBodyParameter(m, compiler.NewContext("nonBodyParameter", context)) + t, matchingError := NewNonBodyParameter(m, compiler.NewContext("nonBodyParameter", m, context)) if matchingError == nil { x.Oneof = &Parameter_NonBodyParameter{NonBodyParameter: t} matched = true @@ -3374,12 +3268,16 @@ func NewParameter(in interface{}, context *compiler.Context) (*Parameter, error) if matched { // since the oneof matched one of its possibilities, discard any matching errors errors = make([]error, 0) + } else { + message := fmt.Sprintf("contains an invalid Parameter") + err := compiler.NewError(context, message) + errors = []error{err} } return x, compiler.NewErrorGroupOrNil(errors) } // NewParameterDefinitions creates an object of type ParameterDefinitions if possible, returning an error if not. -func NewParameterDefinitions(in interface{}, context *compiler.Context) (*ParameterDefinitions, error) { +func NewParameterDefinitions(in *yaml.Node, context *compiler.Context) (*ParameterDefinitions, error) { errors := make([]error, 0) x := &ParameterDefinitions{} m, ok := compiler.UnpackMap(in) @@ -3390,14 +3288,14 @@ func NewParameterDefinitions(in interface{}, context *compiler.Context) (*Parame // repeated NamedParameter additional_properties = 1; // MAP: Parameter x.AdditionalProperties = make([]*NamedParameter, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedParameter{} pair.Name = k var err error - pair.Value, err = NewParameter(v, compiler.NewContext(k, context)) + pair.Value, err = NewParameter(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3409,7 +3307,7 @@ func NewParameterDefinitions(in interface{}, context *compiler.Context) (*Parame } // NewParametersItem creates an object of type ParametersItem if possible, returning an error if not. -func NewParametersItem(in interface{}, context *compiler.Context) (*ParametersItem, error) { +func NewParametersItem(in *yaml.Node, context *compiler.Context) (*ParametersItem, error) { errors := make([]error, 0) x := &ParametersItem{} matched := false @@ -3418,7 +3316,7 @@ func NewParametersItem(in interface{}, context *compiler.Context) (*ParametersIt m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewParameter(m, compiler.NewContext("parameter", context)) + t, matchingError := NewParameter(m, compiler.NewContext("parameter", m, context)) if matchingError == nil { x.Oneof = &ParametersItem_Parameter{Parameter: t} matched = true @@ -3432,7 +3330,7 @@ func NewParametersItem(in interface{}, context *compiler.Context) (*ParametersIt m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewJsonReference(m, compiler.NewContext("jsonReference", context)) + t, matchingError := NewJsonReference(m, compiler.NewContext("jsonReference", m, context)) if matchingError == nil { x.Oneof = &ParametersItem_JsonReference{JsonReference: t} matched = true @@ -3444,12 +3342,16 @@ func NewParametersItem(in interface{}, context *compiler.Context) (*ParametersIt if matched { // since the oneof matched one of its possibilities, discard any matching errors errors = make([]error, 0) + } else { + message := fmt.Sprintf("contains an invalid ParametersItem") + err := compiler.NewError(context, message) + errors = []error{err} } return x, compiler.NewErrorGroupOrNil(errors) } // NewPathItem creates an object of type PathItem if possible, returning an error if not. -func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { +func NewPathItem(in *yaml.Node, context *compiler.Context) (*PathItem, error) { errors := make([]error, 0) x := &PathItem{} m, ok := compiler.UnpackMap(in) @@ -3467,9 +3369,9 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { // string _ref = 1; v1 := compiler.MapValueForKey(m, "$ref") if v1 != nil { - x.XRef, ok = v1.(string) + x.XRef, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for $ref: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for $ref: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3477,7 +3379,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { v2 := compiler.MapValueForKey(m, "get") if v2 != nil { var err error - x.Get, err = NewOperation(v2, compiler.NewContext("get", context)) + x.Get, err = NewOperation(v2, compiler.NewContext("get", v2, context)) if err != nil { errors = append(errors, err) } @@ -3486,7 +3388,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { v3 := compiler.MapValueForKey(m, "put") if v3 != nil { var err error - x.Put, err = NewOperation(v3, compiler.NewContext("put", context)) + x.Put, err = NewOperation(v3, compiler.NewContext("put", v3, context)) if err != nil { errors = append(errors, err) } @@ -3495,7 +3397,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { v4 := compiler.MapValueForKey(m, "post") if v4 != nil { var err error - x.Post, err = NewOperation(v4, compiler.NewContext("post", context)) + x.Post, err = NewOperation(v4, compiler.NewContext("post", v4, context)) if err != nil { errors = append(errors, err) } @@ -3504,7 +3406,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { v5 := compiler.MapValueForKey(m, "delete") if v5 != nil { var err error - x.Delete, err = NewOperation(v5, compiler.NewContext("delete", context)) + x.Delete, err = NewOperation(v5, compiler.NewContext("delete", v5, context)) if err != nil { errors = append(errors, err) } @@ -3513,7 +3415,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { v6 := compiler.MapValueForKey(m, "options") if v6 != nil { var err error - x.Options, err = NewOperation(v6, compiler.NewContext("options", context)) + x.Options, err = NewOperation(v6, compiler.NewContext("options", v6, context)) if err != nil { errors = append(errors, err) } @@ -3522,7 +3424,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { v7 := compiler.MapValueForKey(m, "head") if v7 != nil { var err error - x.Head, err = NewOperation(v7, compiler.NewContext("head", context)) + x.Head, err = NewOperation(v7, compiler.NewContext("head", v7, context)) if err != nil { errors = append(errors, err) } @@ -3531,7 +3433,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { v8 := compiler.MapValueForKey(m, "patch") if v8 != nil { var err error - x.Patch, err = NewOperation(v8, compiler.NewContext("patch", context)) + x.Patch, err = NewOperation(v8, compiler.NewContext("patch", v8, context)) if err != nil { errors = append(errors, err) } @@ -3541,10 +3443,10 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { if v9 != nil { // repeated ParametersItem x.Parameters = make([]*ParametersItem, 0) - a, ok := v9.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v9) if ok { - for _, item := range a { - y, err := NewParametersItem(item, compiler.NewContext("parameters", context)) + for _, item := range a.Content { + y, err := NewParametersItem(item, compiler.NewContext("parameters", item, context)) if err != nil { errors = append(errors, err) } @@ -3555,26 +3457,26 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { // repeated NamedAny vendor_extension = 10; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3588,7 +3490,7 @@ func NewPathItem(in interface{}, context *compiler.Context) (*PathItem, error) { } // NewPathParameterSubSchema creates an object of type PathParameterSubSchema if possible, returning an error if not. -func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*PathParameterSubSchema, error) { +func NewPathParameterSubSchema(in *yaml.Node, context *compiler.Context) (*PathParameterSubSchema, error) { errors := make([]error, 0) x := &PathParameterSubSchema{} m, ok := compiler.UnpackMap(in) @@ -3612,66 +3514,66 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path // bool required = 1; v1 := compiler.MapValueForKey(m, "required") if v1 != nil { - x.Required, ok = v1.(bool) + x.Required, ok = compiler.BoolForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for required: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for required: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string in = 2; v2 := compiler.MapValueForKey(m, "in") if v2 != nil { - x.In, ok = v2.(string) + x.In, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [path] if ok && !compiler.StringArrayContainsValue([]string{"path"}, x.In) { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 3; v3 := compiler.MapValueForKey(m, "description") if v3 != nil { - x.Description, ok = v3.(string) + x.Description, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string name = 4; v4 := compiler.MapValueForKey(m, "name") if v4 != nil { - x.Name, ok = v4.(string) + x.Name, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // string type = 5; v5 := compiler.MapValueForKey(m, "type") if v5 != nil { - x.Type, ok = v5.(string) + x.Type, ok = compiler.StringForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [string number boolean integer array] if ok && !compiler.StringArrayContainsValue([]string{"string", "number", "boolean", "integer", "array"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // string format = 6; v6 := compiler.MapValueForKey(m, "format") if v6 != nil { - x.Format, ok = v6.(string) + x.Format, ok = compiler.StringForScalarNode(v6) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3679,7 +3581,7 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path v7 := compiler.MapValueForKey(m, "items") if v7 != nil { var err error - x.Items, err = NewPrimitivesItems(v7, compiler.NewContext("items", context)) + x.Items, err = NewPrimitivesItems(v7, compiler.NewContext("items", v7, context)) if err != nil { errors = append(errors, err) } @@ -3687,15 +3589,15 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path // string collection_format = 8; v8 := compiler.MapValueForKey(m, "collectionFormat") if v8 != nil { - x.CollectionFormat, ok = v8.(string) + x.CollectionFormat, ok = compiler.StringForScalarNode(v8) if !ok { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v8, v8) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v8)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [csv ssv tsv pipes] if ok && !compiler.StringArrayContainsValue([]string{"csv", "ssv", "tsv", "pipes"}, x.CollectionFormat) { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v8, v8) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v8)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3703,7 +3605,7 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path v9 := compiler.MapValueForKey(m, "default") if v9 != nil { var err error - x.Default, err = NewAny(v9, compiler.NewContext("default", context)) + x.Default, err = NewAny(v9, compiler.NewContext("default", v9, context)) if err != nil { errors = append(errors, err) } @@ -3711,126 +3613,102 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path // float maximum = 10; v10 := compiler.MapValueForKey(m, "maximum") if v10 != nil { - switch v10 := v10.(type) { - case float64: - x.Maximum = v10 - case float32: - x.Maximum = float64(v10) - case uint64: - x.Maximum = float64(v10) - case uint32: - x.Maximum = float64(v10) - case int64: - x.Maximum = float64(v10) - case int32: - x.Maximum = float64(v10) - case int: - x.Maximum = float64(v10) - default: - message := fmt.Sprintf("has unexpected value for maximum: %+v (%T)", v10, v10) + v, ok := compiler.FloatForScalarNode(v10) + if ok { + x.Maximum = v + } else { + message := fmt.Sprintf("has unexpected value for maximum: %s", compiler.Display(v10)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_maximum = 11; v11 := compiler.MapValueForKey(m, "exclusiveMaximum") if v11 != nil { - x.ExclusiveMaximum, ok = v11.(bool) + x.ExclusiveMaximum, ok = compiler.BoolForScalarNode(v11) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %+v (%T)", v11, v11) + message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } // float minimum = 12; v12 := compiler.MapValueForKey(m, "minimum") if v12 != nil { - switch v12 := v12.(type) { - case float64: - x.Minimum = v12 - case float32: - x.Minimum = float64(v12) - case uint64: - x.Minimum = float64(v12) - case uint32: - x.Minimum = float64(v12) - case int64: - x.Minimum = float64(v12) - case int32: - x.Minimum = float64(v12) - case int: - x.Minimum = float64(v12) - default: - message := fmt.Sprintf("has unexpected value for minimum: %+v (%T)", v12, v12) + v, ok := compiler.FloatForScalarNode(v12) + if ok { + x.Minimum = v + } else { + message := fmt.Sprintf("has unexpected value for minimum: %s", compiler.Display(v12)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_minimum = 13; v13 := compiler.MapValueForKey(m, "exclusiveMinimum") if v13 != nil { - x.ExclusiveMinimum, ok = v13.(bool) + x.ExclusiveMinimum, ok = compiler.BoolForScalarNode(v13) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %+v (%T)", v13, v13) + message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %s", compiler.Display(v13)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_length = 14; v14 := compiler.MapValueForKey(m, "maxLength") if v14 != nil { - t, ok := v14.(int) + t, ok := compiler.IntForScalarNode(v14) if ok { x.MaxLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxLength: %+v (%T)", v14, v14) + message := fmt.Sprintf("has unexpected value for maxLength: %s", compiler.Display(v14)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_length = 15; v15 := compiler.MapValueForKey(m, "minLength") if v15 != nil { - t, ok := v15.(int) + t, ok := compiler.IntForScalarNode(v15) if ok { x.MinLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minLength: %+v (%T)", v15, v15) + message := fmt.Sprintf("has unexpected value for minLength: %s", compiler.Display(v15)) errors = append(errors, compiler.NewError(context, message)) } } // string pattern = 16; v16 := compiler.MapValueForKey(m, "pattern") if v16 != nil { - x.Pattern, ok = v16.(string) + x.Pattern, ok = compiler.StringForScalarNode(v16) if !ok { - message := fmt.Sprintf("has unexpected value for pattern: %+v (%T)", v16, v16) + message := fmt.Sprintf("has unexpected value for pattern: %s", compiler.Display(v16)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_items = 17; v17 := compiler.MapValueForKey(m, "maxItems") if v17 != nil { - t, ok := v17.(int) + t, ok := compiler.IntForScalarNode(v17) if ok { x.MaxItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxItems: %+v (%T)", v17, v17) + message := fmt.Sprintf("has unexpected value for maxItems: %s", compiler.Display(v17)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_items = 18; v18 := compiler.MapValueForKey(m, "minItems") if v18 != nil { - t, ok := v18.(int) + t, ok := compiler.IntForScalarNode(v18) if ok { x.MinItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minItems: %+v (%T)", v18, v18) + message := fmt.Sprintf("has unexpected value for minItems: %s", compiler.Display(v18)) errors = append(errors, compiler.NewError(context, message)) } } // bool unique_items = 19; v19 := compiler.MapValueForKey(m, "uniqueItems") if v19 != nil { - x.UniqueItems, ok = v19.(bool) + x.UniqueItems, ok = compiler.BoolForScalarNode(v19) if !ok { - message := fmt.Sprintf("has unexpected value for uniqueItems: %+v (%T)", v19, v19) + message := fmt.Sprintf("has unexpected value for uniqueItems: %s", compiler.Display(v19)) errors = append(errors, compiler.NewError(context, message)) } } @@ -3839,10 +3717,10 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path if v20 != nil { // repeated Any x.Enum = make([]*Any, 0) - a, ok := v20.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v20) if ok { - for _, item := range a { - y, err := NewAny(item, compiler.NewContext("enum", context)) + for _, item := range a.Content { + y, err := NewAny(item, compiler.NewContext("enum", item, context)) if err != nil { errors = append(errors, err) } @@ -3853,49 +3731,37 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path // float multiple_of = 21; v21 := compiler.MapValueForKey(m, "multipleOf") if v21 != nil { - switch v21 := v21.(type) { - case float64: - x.MultipleOf = v21 - case float32: - x.MultipleOf = float64(v21) - case uint64: - x.MultipleOf = float64(v21) - case uint32: - x.MultipleOf = float64(v21) - case int64: - x.MultipleOf = float64(v21) - case int32: - x.MultipleOf = float64(v21) - case int: - x.MultipleOf = float64(v21) - default: - message := fmt.Sprintf("has unexpected value for multipleOf: %+v (%T)", v21, v21) + v, ok := compiler.FloatForScalarNode(v21) + if ok { + x.MultipleOf = v + } else { + message := fmt.Sprintf("has unexpected value for multipleOf: %s", compiler.Display(v21)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 22; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3909,7 +3775,7 @@ func NewPathParameterSubSchema(in interface{}, context *compiler.Context) (*Path } // NewPaths creates an object of type Paths if possible, returning an error if not. -func NewPaths(in interface{}, context *compiler.Context) (*Paths, error) { +func NewPaths(in *yaml.Node, context *compiler.Context) (*Paths, error) { errors := make([]error, 0) x := &Paths{} m, ok := compiler.UnpackMap(in) @@ -3927,26 +3793,26 @@ func NewPaths(in interface{}, context *compiler.Context) (*Paths, error) { // repeated NamedAny vendor_extension = 1; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3958,15 +3824,15 @@ func NewPaths(in interface{}, context *compiler.Context) (*Paths, error) { // repeated NamedPathItem path = 2; // MAP: PathItem ^/ x.Path = make([]*NamedPathItem, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "/") { pair := &NamedPathItem{} pair.Name = k var err error - pair.Value, err = NewPathItem(v, compiler.NewContext(k, context)) + pair.Value, err = NewPathItem(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -3979,7 +3845,7 @@ func NewPaths(in interface{}, context *compiler.Context) (*Paths, error) { } // NewPrimitivesItems creates an object of type PrimitivesItems if possible, returning an error if not. -func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesItems, error) { +func NewPrimitivesItems(in *yaml.Node, context *compiler.Context) (*PrimitivesItems, error) { errors := make([]error, 0) x := &PrimitivesItems{} m, ok := compiler.UnpackMap(in) @@ -3997,24 +3863,24 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI // string type = 1; v1 := compiler.MapValueForKey(m, "type") if v1 != nil { - x.Type, ok = v1.(string) + x.Type, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [string number integer boolean array] if ok && !compiler.StringArrayContainsValue([]string{"string", "number", "integer", "boolean", "array"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string format = 2; v2 := compiler.MapValueForKey(m, "format") if v2 != nil { - x.Format, ok = v2.(string) + x.Format, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4022,7 +3888,7 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI v3 := compiler.MapValueForKey(m, "items") if v3 != nil { var err error - x.Items, err = NewPrimitivesItems(v3, compiler.NewContext("items", context)) + x.Items, err = NewPrimitivesItems(v3, compiler.NewContext("items", v3, context)) if err != nil { errors = append(errors, err) } @@ -4030,15 +3896,15 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI // string collection_format = 4; v4 := compiler.MapValueForKey(m, "collectionFormat") if v4 != nil { - x.CollectionFormat, ok = v4.(string) + x.CollectionFormat, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [csv ssv tsv pipes] if ok && !compiler.StringArrayContainsValue([]string{"csv", "ssv", "tsv", "pipes"}, x.CollectionFormat) { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4046,7 +3912,7 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI v5 := compiler.MapValueForKey(m, "default") if v5 != nil { var err error - x.Default, err = NewAny(v5, compiler.NewContext("default", context)) + x.Default, err = NewAny(v5, compiler.NewContext("default", v5, context)) if err != nil { errors = append(errors, err) } @@ -4054,126 +3920,102 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI // float maximum = 6; v6 := compiler.MapValueForKey(m, "maximum") if v6 != nil { - switch v6 := v6.(type) { - case float64: - x.Maximum = v6 - case float32: - x.Maximum = float64(v6) - case uint64: - x.Maximum = float64(v6) - case uint32: - x.Maximum = float64(v6) - case int64: - x.Maximum = float64(v6) - case int32: - x.Maximum = float64(v6) - case int: - x.Maximum = float64(v6) - default: - message := fmt.Sprintf("has unexpected value for maximum: %+v (%T)", v6, v6) + v, ok := compiler.FloatForScalarNode(v6) + if ok { + x.Maximum = v + } else { + message := fmt.Sprintf("has unexpected value for maximum: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_maximum = 7; v7 := compiler.MapValueForKey(m, "exclusiveMaximum") if v7 != nil { - x.ExclusiveMaximum, ok = v7.(bool) + x.ExclusiveMaximum, ok = compiler.BoolForScalarNode(v7) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %+v (%T)", v7, v7) + message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } // float minimum = 8; v8 := compiler.MapValueForKey(m, "minimum") if v8 != nil { - switch v8 := v8.(type) { - case float64: - x.Minimum = v8 - case float32: - x.Minimum = float64(v8) - case uint64: - x.Minimum = float64(v8) - case uint32: - x.Minimum = float64(v8) - case int64: - x.Minimum = float64(v8) - case int32: - x.Minimum = float64(v8) - case int: - x.Minimum = float64(v8) - default: - message := fmt.Sprintf("has unexpected value for minimum: %+v (%T)", v8, v8) + v, ok := compiler.FloatForScalarNode(v8) + if ok { + x.Minimum = v + } else { + message := fmt.Sprintf("has unexpected value for minimum: %s", compiler.Display(v8)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_minimum = 9; v9 := compiler.MapValueForKey(m, "exclusiveMinimum") if v9 != nil { - x.ExclusiveMinimum, ok = v9.(bool) + x.ExclusiveMinimum, ok = compiler.BoolForScalarNode(v9) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %+v (%T)", v9, v9) + message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %s", compiler.Display(v9)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_length = 10; v10 := compiler.MapValueForKey(m, "maxLength") if v10 != nil { - t, ok := v10.(int) + t, ok := compiler.IntForScalarNode(v10) if ok { x.MaxLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxLength: %+v (%T)", v10, v10) + message := fmt.Sprintf("has unexpected value for maxLength: %s", compiler.Display(v10)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_length = 11; v11 := compiler.MapValueForKey(m, "minLength") if v11 != nil { - t, ok := v11.(int) + t, ok := compiler.IntForScalarNode(v11) if ok { x.MinLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minLength: %+v (%T)", v11, v11) + message := fmt.Sprintf("has unexpected value for minLength: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } // string pattern = 12; v12 := compiler.MapValueForKey(m, "pattern") if v12 != nil { - x.Pattern, ok = v12.(string) + x.Pattern, ok = compiler.StringForScalarNode(v12) if !ok { - message := fmt.Sprintf("has unexpected value for pattern: %+v (%T)", v12, v12) + message := fmt.Sprintf("has unexpected value for pattern: %s", compiler.Display(v12)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_items = 13; v13 := compiler.MapValueForKey(m, "maxItems") if v13 != nil { - t, ok := v13.(int) + t, ok := compiler.IntForScalarNode(v13) if ok { x.MaxItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxItems: %+v (%T)", v13, v13) + message := fmt.Sprintf("has unexpected value for maxItems: %s", compiler.Display(v13)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_items = 14; v14 := compiler.MapValueForKey(m, "minItems") if v14 != nil { - t, ok := v14.(int) + t, ok := compiler.IntForScalarNode(v14) if ok { x.MinItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minItems: %+v (%T)", v14, v14) + message := fmt.Sprintf("has unexpected value for minItems: %s", compiler.Display(v14)) errors = append(errors, compiler.NewError(context, message)) } } // bool unique_items = 15; v15 := compiler.MapValueForKey(m, "uniqueItems") if v15 != nil { - x.UniqueItems, ok = v15.(bool) + x.UniqueItems, ok = compiler.BoolForScalarNode(v15) if !ok { - message := fmt.Sprintf("has unexpected value for uniqueItems: %+v (%T)", v15, v15) + message := fmt.Sprintf("has unexpected value for uniqueItems: %s", compiler.Display(v15)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4182,10 +4024,10 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI if v16 != nil { // repeated Any x.Enum = make([]*Any, 0) - a, ok := v16.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v16) if ok { - for _, item := range a { - y, err := NewAny(item, compiler.NewContext("enum", context)) + for _, item := range a.Content { + y, err := NewAny(item, compiler.NewContext("enum", item, context)) if err != nil { errors = append(errors, err) } @@ -4196,49 +4038,37 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI // float multiple_of = 17; v17 := compiler.MapValueForKey(m, "multipleOf") if v17 != nil { - switch v17 := v17.(type) { - case float64: - x.MultipleOf = v17 - case float32: - x.MultipleOf = float64(v17) - case uint64: - x.MultipleOf = float64(v17) - case uint32: - x.MultipleOf = float64(v17) - case int64: - x.MultipleOf = float64(v17) - case int32: - x.MultipleOf = float64(v17) - case int: - x.MultipleOf = float64(v17) - default: - message := fmt.Sprintf("has unexpected value for multipleOf: %+v (%T)", v17, v17) + v, ok := compiler.FloatForScalarNode(v17) + if ok { + x.MultipleOf = v + } else { + message := fmt.Sprintf("has unexpected value for multipleOf: %s", compiler.Display(v17)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 18; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -4252,7 +4082,7 @@ func NewPrimitivesItems(in interface{}, context *compiler.Context) (*PrimitivesI } // NewProperties creates an object of type Properties if possible, returning an error if not. -func NewProperties(in interface{}, context *compiler.Context) (*Properties, error) { +func NewProperties(in *yaml.Node, context *compiler.Context) (*Properties, error) { errors := make([]error, 0) x := &Properties{} m, ok := compiler.UnpackMap(in) @@ -4263,14 +4093,14 @@ func NewProperties(in interface{}, context *compiler.Context) (*Properties, erro // repeated NamedSchema additional_properties = 1; // MAP: Schema x.AdditionalProperties = make([]*NamedSchema, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedSchema{} pair.Name = k var err error - pair.Value, err = NewSchema(v, compiler.NewContext(k, context)) + pair.Value, err = NewSchema(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -4282,7 +4112,7 @@ func NewProperties(in interface{}, context *compiler.Context) (*Properties, erro } // NewQueryParameterSubSchema creates an object of type QueryParameterSubSchema if possible, returning an error if not. -func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*QueryParameterSubSchema, error) { +func NewQueryParameterSubSchema(in *yaml.Node, context *compiler.Context) (*QueryParameterSubSchema, error) { errors := make([]error, 0) x := &QueryParameterSubSchema{} m, ok := compiler.UnpackMap(in) @@ -4300,75 +4130,75 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que // bool required = 1; v1 := compiler.MapValueForKey(m, "required") if v1 != nil { - x.Required, ok = v1.(bool) + x.Required, ok = compiler.BoolForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for required: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for required: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string in = 2; v2 := compiler.MapValueForKey(m, "in") if v2 != nil { - x.In, ok = v2.(string) + x.In, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [query] if ok && !compiler.StringArrayContainsValue([]string{"query"}, x.In) { - message := fmt.Sprintf("has unexpected value for in: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for in: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 3; v3 := compiler.MapValueForKey(m, "description") if v3 != nil { - x.Description, ok = v3.(string) + x.Description, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string name = 4; v4 := compiler.MapValueForKey(m, "name") if v4 != nil { - x.Name, ok = v4.(string) + x.Name, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // bool allow_empty_value = 5; v5 := compiler.MapValueForKey(m, "allowEmptyValue") if v5 != nil { - x.AllowEmptyValue, ok = v5.(bool) + x.AllowEmptyValue, ok = compiler.BoolForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for allowEmptyValue: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for allowEmptyValue: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // string type = 6; v6 := compiler.MapValueForKey(m, "type") if v6 != nil { - x.Type, ok = v6.(string) + x.Type, ok = compiler.StringForScalarNode(v6) if !ok { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [string number boolean integer array] if ok && !compiler.StringArrayContainsValue([]string{"string", "number", "boolean", "integer", "array"}, x.Type) { - message := fmt.Sprintf("has unexpected value for type: %+v (%T)", v6, v6) + message := fmt.Sprintf("has unexpected value for type: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // string format = 7; v7 := compiler.MapValueForKey(m, "format") if v7 != nil { - x.Format, ok = v7.(string) + x.Format, ok = compiler.StringForScalarNode(v7) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v7, v7) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4376,7 +4206,7 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que v8 := compiler.MapValueForKey(m, "items") if v8 != nil { var err error - x.Items, err = NewPrimitivesItems(v8, compiler.NewContext("items", context)) + x.Items, err = NewPrimitivesItems(v8, compiler.NewContext("items", v8, context)) if err != nil { errors = append(errors, err) } @@ -4384,15 +4214,15 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que // string collection_format = 9; v9 := compiler.MapValueForKey(m, "collectionFormat") if v9 != nil { - x.CollectionFormat, ok = v9.(string) + x.CollectionFormat, ok = compiler.StringForScalarNode(v9) if !ok { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v9, v9) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v9)) errors = append(errors, compiler.NewError(context, message)) } // check for valid enum values // [csv ssv tsv pipes multi] if ok && !compiler.StringArrayContainsValue([]string{"csv", "ssv", "tsv", "pipes", "multi"}, x.CollectionFormat) { - message := fmt.Sprintf("has unexpected value for collectionFormat: %+v (%T)", v9, v9) + message := fmt.Sprintf("has unexpected value for collectionFormat: %s", compiler.Display(v9)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4400,7 +4230,7 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que v10 := compiler.MapValueForKey(m, "default") if v10 != nil { var err error - x.Default, err = NewAny(v10, compiler.NewContext("default", context)) + x.Default, err = NewAny(v10, compiler.NewContext("default", v10, context)) if err != nil { errors = append(errors, err) } @@ -4408,126 +4238,102 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que // float maximum = 11; v11 := compiler.MapValueForKey(m, "maximum") if v11 != nil { - switch v11 := v11.(type) { - case float64: - x.Maximum = v11 - case float32: - x.Maximum = float64(v11) - case uint64: - x.Maximum = float64(v11) - case uint32: - x.Maximum = float64(v11) - case int64: - x.Maximum = float64(v11) - case int32: - x.Maximum = float64(v11) - case int: - x.Maximum = float64(v11) - default: - message := fmt.Sprintf("has unexpected value for maximum: %+v (%T)", v11, v11) + v, ok := compiler.FloatForScalarNode(v11) + if ok { + x.Maximum = v + } else { + message := fmt.Sprintf("has unexpected value for maximum: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_maximum = 12; v12 := compiler.MapValueForKey(m, "exclusiveMaximum") if v12 != nil { - x.ExclusiveMaximum, ok = v12.(bool) + x.ExclusiveMaximum, ok = compiler.BoolForScalarNode(v12) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %+v (%T)", v12, v12) + message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %s", compiler.Display(v12)) errors = append(errors, compiler.NewError(context, message)) } } // float minimum = 13; v13 := compiler.MapValueForKey(m, "minimum") if v13 != nil { - switch v13 := v13.(type) { - case float64: - x.Minimum = v13 - case float32: - x.Minimum = float64(v13) - case uint64: - x.Minimum = float64(v13) - case uint32: - x.Minimum = float64(v13) - case int64: - x.Minimum = float64(v13) - case int32: - x.Minimum = float64(v13) - case int: - x.Minimum = float64(v13) - default: - message := fmt.Sprintf("has unexpected value for minimum: %+v (%T)", v13, v13) + v, ok := compiler.FloatForScalarNode(v13) + if ok { + x.Minimum = v + } else { + message := fmt.Sprintf("has unexpected value for minimum: %s", compiler.Display(v13)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_minimum = 14; v14 := compiler.MapValueForKey(m, "exclusiveMinimum") if v14 != nil { - x.ExclusiveMinimum, ok = v14.(bool) + x.ExclusiveMinimum, ok = compiler.BoolForScalarNode(v14) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %+v (%T)", v14, v14) + message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %s", compiler.Display(v14)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_length = 15; v15 := compiler.MapValueForKey(m, "maxLength") if v15 != nil { - t, ok := v15.(int) + t, ok := compiler.IntForScalarNode(v15) if ok { x.MaxLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxLength: %+v (%T)", v15, v15) + message := fmt.Sprintf("has unexpected value for maxLength: %s", compiler.Display(v15)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_length = 16; v16 := compiler.MapValueForKey(m, "minLength") if v16 != nil { - t, ok := v16.(int) + t, ok := compiler.IntForScalarNode(v16) if ok { x.MinLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minLength: %+v (%T)", v16, v16) + message := fmt.Sprintf("has unexpected value for minLength: %s", compiler.Display(v16)) errors = append(errors, compiler.NewError(context, message)) } } // string pattern = 17; v17 := compiler.MapValueForKey(m, "pattern") if v17 != nil { - x.Pattern, ok = v17.(string) + x.Pattern, ok = compiler.StringForScalarNode(v17) if !ok { - message := fmt.Sprintf("has unexpected value for pattern: %+v (%T)", v17, v17) + message := fmt.Sprintf("has unexpected value for pattern: %s", compiler.Display(v17)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_items = 18; v18 := compiler.MapValueForKey(m, "maxItems") if v18 != nil { - t, ok := v18.(int) + t, ok := compiler.IntForScalarNode(v18) if ok { x.MaxItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxItems: %+v (%T)", v18, v18) + message := fmt.Sprintf("has unexpected value for maxItems: %s", compiler.Display(v18)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_items = 19; v19 := compiler.MapValueForKey(m, "minItems") if v19 != nil { - t, ok := v19.(int) + t, ok := compiler.IntForScalarNode(v19) if ok { x.MinItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minItems: %+v (%T)", v19, v19) + message := fmt.Sprintf("has unexpected value for minItems: %s", compiler.Display(v19)) errors = append(errors, compiler.NewError(context, message)) } } // bool unique_items = 20; v20 := compiler.MapValueForKey(m, "uniqueItems") if v20 != nil { - x.UniqueItems, ok = v20.(bool) + x.UniqueItems, ok = compiler.BoolForScalarNode(v20) if !ok { - message := fmt.Sprintf("has unexpected value for uniqueItems: %+v (%T)", v20, v20) + message := fmt.Sprintf("has unexpected value for uniqueItems: %s", compiler.Display(v20)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4536,10 +4342,10 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que if v21 != nil { // repeated Any x.Enum = make([]*Any, 0) - a, ok := v21.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v21) if ok { - for _, item := range a { - y, err := NewAny(item, compiler.NewContext("enum", context)) + for _, item := range a.Content { + y, err := NewAny(item, compiler.NewContext("enum", item, context)) if err != nil { errors = append(errors, err) } @@ -4550,49 +4356,37 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que // float multiple_of = 22; v22 := compiler.MapValueForKey(m, "multipleOf") if v22 != nil { - switch v22 := v22.(type) { - case float64: - x.MultipleOf = v22 - case float32: - x.MultipleOf = float64(v22) - case uint64: - x.MultipleOf = float64(v22) - case uint32: - x.MultipleOf = float64(v22) - case int64: - x.MultipleOf = float64(v22) - case int32: - x.MultipleOf = float64(v22) - case int: - x.MultipleOf = float64(v22) - default: - message := fmt.Sprintf("has unexpected value for multipleOf: %+v (%T)", v22, v22) + v, ok := compiler.FloatForScalarNode(v22) + if ok { + x.MultipleOf = v + } else { + message := fmt.Sprintf("has unexpected value for multipleOf: %s", compiler.Display(v22)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 23; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -4606,7 +4400,7 @@ func NewQueryParameterSubSchema(in interface{}, context *compiler.Context) (*Que } // NewResponse creates an object of type Response if possible, returning an error if not. -func NewResponse(in interface{}, context *compiler.Context) (*Response, error) { +func NewResponse(in *yaml.Node, context *compiler.Context) (*Response, error) { errors := make([]error, 0) x := &Response{} m, ok := compiler.UnpackMap(in) @@ -4630,9 +4424,9 @@ func NewResponse(in interface{}, context *compiler.Context) (*Response, error) { // string description = 1; v1 := compiler.MapValueForKey(m, "description") if v1 != nil { - x.Description, ok = v1.(string) + x.Description, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4640,7 +4434,7 @@ func NewResponse(in interface{}, context *compiler.Context) (*Response, error) { v2 := compiler.MapValueForKey(m, "schema") if v2 != nil { var err error - x.Schema, err = NewSchemaItem(v2, compiler.NewContext("schema", context)) + x.Schema, err = NewSchemaItem(v2, compiler.NewContext("schema", v2, context)) if err != nil { errors = append(errors, err) } @@ -4649,7 +4443,7 @@ func NewResponse(in interface{}, context *compiler.Context) (*Response, error) { v3 := compiler.MapValueForKey(m, "headers") if v3 != nil { var err error - x.Headers, err = NewHeaders(v3, compiler.NewContext("headers", context)) + x.Headers, err = NewHeaders(v3, compiler.NewContext("headers", v3, context)) if err != nil { errors = append(errors, err) } @@ -4658,7 +4452,7 @@ func NewResponse(in interface{}, context *compiler.Context) (*Response, error) { v4 := compiler.MapValueForKey(m, "examples") if v4 != nil { var err error - x.Examples, err = NewExamples(v4, compiler.NewContext("examples", context)) + x.Examples, err = NewExamples(v4, compiler.NewContext("examples", v4, context)) if err != nil { errors = append(errors, err) } @@ -4666,26 +4460,26 @@ func NewResponse(in interface{}, context *compiler.Context) (*Response, error) { // repeated NamedAny vendor_extension = 5; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -4699,7 +4493,7 @@ func NewResponse(in interface{}, context *compiler.Context) (*Response, error) { } // NewResponseDefinitions creates an object of type ResponseDefinitions if possible, returning an error if not. -func NewResponseDefinitions(in interface{}, context *compiler.Context) (*ResponseDefinitions, error) { +func NewResponseDefinitions(in *yaml.Node, context *compiler.Context) (*ResponseDefinitions, error) { errors := make([]error, 0) x := &ResponseDefinitions{} m, ok := compiler.UnpackMap(in) @@ -4710,14 +4504,14 @@ func NewResponseDefinitions(in interface{}, context *compiler.Context) (*Respons // repeated NamedResponse additional_properties = 1; // MAP: Response x.AdditionalProperties = make([]*NamedResponse, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedResponse{} pair.Name = k var err error - pair.Value, err = NewResponse(v, compiler.NewContext(k, context)) + pair.Value, err = NewResponse(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -4729,7 +4523,7 @@ func NewResponseDefinitions(in interface{}, context *compiler.Context) (*Respons } // NewResponseValue creates an object of type ResponseValue if possible, returning an error if not. -func NewResponseValue(in interface{}, context *compiler.Context) (*ResponseValue, error) { +func NewResponseValue(in *yaml.Node, context *compiler.Context) (*ResponseValue, error) { errors := make([]error, 0) x := &ResponseValue{} matched := false @@ -4738,7 +4532,7 @@ func NewResponseValue(in interface{}, context *compiler.Context) (*ResponseValue m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewResponse(m, compiler.NewContext("response", context)) + t, matchingError := NewResponse(m, compiler.NewContext("response", m, context)) if matchingError == nil { x.Oneof = &ResponseValue_Response{Response: t} matched = true @@ -4752,7 +4546,7 @@ func NewResponseValue(in interface{}, context *compiler.Context) (*ResponseValue m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewJsonReference(m, compiler.NewContext("jsonReference", context)) + t, matchingError := NewJsonReference(m, compiler.NewContext("jsonReference", m, context)) if matchingError == nil { x.Oneof = &ResponseValue_JsonReference{JsonReference: t} matched = true @@ -4764,12 +4558,16 @@ func NewResponseValue(in interface{}, context *compiler.Context) (*ResponseValue if matched { // since the oneof matched one of its possibilities, discard any matching errors errors = make([]error, 0) + } else { + message := fmt.Sprintf("contains an invalid ResponseValue") + err := compiler.NewError(context, message) + errors = []error{err} } return x, compiler.NewErrorGroupOrNil(errors) } // NewResponses creates an object of type Responses if possible, returning an error if not. -func NewResponses(in interface{}, context *compiler.Context) (*Responses, error) { +func NewResponses(in *yaml.Node, context *compiler.Context) (*Responses, error) { errors := make([]error, 0) x := &Responses{} m, ok := compiler.UnpackMap(in) @@ -4787,15 +4585,15 @@ func NewResponses(in interface{}, context *compiler.Context) (*Responses, error) // repeated NamedResponseValue response_code = 1; // MAP: ResponseValue ^([0-9]{3})$|^(default)$ x.ResponseCode = make([]*NamedResponseValue, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if pattern2.MatchString(k) { pair := &NamedResponseValue{} pair.Name = k var err error - pair.Value, err = NewResponseValue(v, compiler.NewContext(k, context)) + pair.Value, err = NewResponseValue(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -4806,26 +4604,26 @@ func NewResponses(in interface{}, context *compiler.Context) (*Responses, error) // repeated NamedAny vendor_extension = 2; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -4839,7 +4637,7 @@ func NewResponses(in interface{}, context *compiler.Context) (*Responses, error) } // NewSchema creates an object of type Schema if possible, returning an error if not. -func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { +func NewSchema(in *yaml.Node, context *compiler.Context) (*Schema, error) { errors := make([]error, 0) x := &Schema{} m, ok := compiler.UnpackMap(in) @@ -4857,36 +4655,36 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { // string _ref = 1; v1 := compiler.MapValueForKey(m, "$ref") if v1 != nil { - x.XRef, ok = v1.(string) + x.XRef, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for $ref: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for $ref: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string format = 2; v2 := compiler.MapValueForKey(m, "format") if v2 != nil { - x.Format, ok = v2.(string) + x.Format, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for format: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for format: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string title = 3; v3 := compiler.MapValueForKey(m, "title") if v3 != nil { - x.Title, ok = v3.(string) + x.Title, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for title: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for title: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 4; v4 := compiler.MapValueForKey(m, "description") if v4 != nil { - x.Description, ok = v4.(string) + x.Description, ok = compiler.StringForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } @@ -4894,7 +4692,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v5 := compiler.MapValueForKey(m, "default") if v5 != nil { var err error - x.Default, err = NewAny(v5, compiler.NewContext("default", context)) + x.Default, err = NewAny(v5, compiler.NewContext("default", v5, context)) if err != nil { errors = append(errors, err) } @@ -4902,182 +4700,146 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { // float multiple_of = 6; v6 := compiler.MapValueForKey(m, "multipleOf") if v6 != nil { - switch v6 := v6.(type) { - case float64: - x.MultipleOf = v6 - case float32: - x.MultipleOf = float64(v6) - case uint64: - x.MultipleOf = float64(v6) - case uint32: - x.MultipleOf = float64(v6) - case int64: - x.MultipleOf = float64(v6) - case int32: - x.MultipleOf = float64(v6) - case int: - x.MultipleOf = float64(v6) - default: - message := fmt.Sprintf("has unexpected value for multipleOf: %+v (%T)", v6, v6) + v, ok := compiler.FloatForScalarNode(v6) + if ok { + x.MultipleOf = v + } else { + message := fmt.Sprintf("has unexpected value for multipleOf: %s", compiler.Display(v6)) errors = append(errors, compiler.NewError(context, message)) } } // float maximum = 7; v7 := compiler.MapValueForKey(m, "maximum") if v7 != nil { - switch v7 := v7.(type) { - case float64: - x.Maximum = v7 - case float32: - x.Maximum = float64(v7) - case uint64: - x.Maximum = float64(v7) - case uint32: - x.Maximum = float64(v7) - case int64: - x.Maximum = float64(v7) - case int32: - x.Maximum = float64(v7) - case int: - x.Maximum = float64(v7) - default: - message := fmt.Sprintf("has unexpected value for maximum: %+v (%T)", v7, v7) + v, ok := compiler.FloatForScalarNode(v7) + if ok { + x.Maximum = v + } else { + message := fmt.Sprintf("has unexpected value for maximum: %s", compiler.Display(v7)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_maximum = 8; v8 := compiler.MapValueForKey(m, "exclusiveMaximum") if v8 != nil { - x.ExclusiveMaximum, ok = v8.(bool) + x.ExclusiveMaximum, ok = compiler.BoolForScalarNode(v8) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %+v (%T)", v8, v8) + message := fmt.Sprintf("has unexpected value for exclusiveMaximum: %s", compiler.Display(v8)) errors = append(errors, compiler.NewError(context, message)) } } // float minimum = 9; v9 := compiler.MapValueForKey(m, "minimum") if v9 != nil { - switch v9 := v9.(type) { - case float64: - x.Minimum = v9 - case float32: - x.Minimum = float64(v9) - case uint64: - x.Minimum = float64(v9) - case uint32: - x.Minimum = float64(v9) - case int64: - x.Minimum = float64(v9) - case int32: - x.Minimum = float64(v9) - case int: - x.Minimum = float64(v9) - default: - message := fmt.Sprintf("has unexpected value for minimum: %+v (%T)", v9, v9) + v, ok := compiler.FloatForScalarNode(v9) + if ok { + x.Minimum = v + } else { + message := fmt.Sprintf("has unexpected value for minimum: %s", compiler.Display(v9)) errors = append(errors, compiler.NewError(context, message)) } } // bool exclusive_minimum = 10; v10 := compiler.MapValueForKey(m, "exclusiveMinimum") if v10 != nil { - x.ExclusiveMinimum, ok = v10.(bool) + x.ExclusiveMinimum, ok = compiler.BoolForScalarNode(v10) if !ok { - message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %+v (%T)", v10, v10) + message := fmt.Sprintf("has unexpected value for exclusiveMinimum: %s", compiler.Display(v10)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_length = 11; v11 := compiler.MapValueForKey(m, "maxLength") if v11 != nil { - t, ok := v11.(int) + t, ok := compiler.IntForScalarNode(v11) if ok { x.MaxLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxLength: %+v (%T)", v11, v11) + message := fmt.Sprintf("has unexpected value for maxLength: %s", compiler.Display(v11)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_length = 12; v12 := compiler.MapValueForKey(m, "minLength") if v12 != nil { - t, ok := v12.(int) + t, ok := compiler.IntForScalarNode(v12) if ok { x.MinLength = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minLength: %+v (%T)", v12, v12) + message := fmt.Sprintf("has unexpected value for minLength: %s", compiler.Display(v12)) errors = append(errors, compiler.NewError(context, message)) } } // string pattern = 13; v13 := compiler.MapValueForKey(m, "pattern") if v13 != nil { - x.Pattern, ok = v13.(string) + x.Pattern, ok = compiler.StringForScalarNode(v13) if !ok { - message := fmt.Sprintf("has unexpected value for pattern: %+v (%T)", v13, v13) + message := fmt.Sprintf("has unexpected value for pattern: %s", compiler.Display(v13)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_items = 14; v14 := compiler.MapValueForKey(m, "maxItems") if v14 != nil { - t, ok := v14.(int) + t, ok := compiler.IntForScalarNode(v14) if ok { x.MaxItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxItems: %+v (%T)", v14, v14) + message := fmt.Sprintf("has unexpected value for maxItems: %s", compiler.Display(v14)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_items = 15; v15 := compiler.MapValueForKey(m, "minItems") if v15 != nil { - t, ok := v15.(int) + t, ok := compiler.IntForScalarNode(v15) if ok { x.MinItems = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minItems: %+v (%T)", v15, v15) + message := fmt.Sprintf("has unexpected value for minItems: %s", compiler.Display(v15)) errors = append(errors, compiler.NewError(context, message)) } } // bool unique_items = 16; v16 := compiler.MapValueForKey(m, "uniqueItems") if v16 != nil { - x.UniqueItems, ok = v16.(bool) + x.UniqueItems, ok = compiler.BoolForScalarNode(v16) if !ok { - message := fmt.Sprintf("has unexpected value for uniqueItems: %+v (%T)", v16, v16) + message := fmt.Sprintf("has unexpected value for uniqueItems: %s", compiler.Display(v16)) errors = append(errors, compiler.NewError(context, message)) } } // int64 max_properties = 17; v17 := compiler.MapValueForKey(m, "maxProperties") if v17 != nil { - t, ok := v17.(int) + t, ok := compiler.IntForScalarNode(v17) if ok { x.MaxProperties = int64(t) } else { - message := fmt.Sprintf("has unexpected value for maxProperties: %+v (%T)", v17, v17) + message := fmt.Sprintf("has unexpected value for maxProperties: %s", compiler.Display(v17)) errors = append(errors, compiler.NewError(context, message)) } } // int64 min_properties = 18; v18 := compiler.MapValueForKey(m, "minProperties") if v18 != nil { - t, ok := v18.(int) + t, ok := compiler.IntForScalarNode(v18) if ok { x.MinProperties = int64(t) } else { - message := fmt.Sprintf("has unexpected value for minProperties: %+v (%T)", v18, v18) + message := fmt.Sprintf("has unexpected value for minProperties: %s", compiler.Display(v18)) errors = append(errors, compiler.NewError(context, message)) } } // repeated string required = 19; v19 := compiler.MapValueForKey(m, "required") if v19 != nil { - v, ok := v19.([]interface{}) + v, ok := compiler.SequenceNodeForNode(v19) if ok { - x.Required = compiler.ConvertInterfaceArrayToStringArray(v) + x.Required = compiler.StringArrayForSequenceNode(v) } else { - message := fmt.Sprintf("has unexpected value for required: %+v (%T)", v19, v19) + message := fmt.Sprintf("has unexpected value for required: %s", compiler.Display(v19)) errors = append(errors, compiler.NewError(context, message)) } } @@ -5086,10 +4848,10 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { if v20 != nil { // repeated Any x.Enum = make([]*Any, 0) - a, ok := v20.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v20) if ok { - for _, item := range a { - y, err := NewAny(item, compiler.NewContext("enum", context)) + for _, item := range a.Content { + y, err := NewAny(item, compiler.NewContext("enum", item, context)) if err != nil { errors = append(errors, err) } @@ -5101,7 +4863,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v21 := compiler.MapValueForKey(m, "additionalProperties") if v21 != nil { var err error - x.AdditionalProperties, err = NewAdditionalPropertiesItem(v21, compiler.NewContext("additionalProperties", context)) + x.AdditionalProperties, err = NewAdditionalPropertiesItem(v21, compiler.NewContext("additionalProperties", v21, context)) if err != nil { errors = append(errors, err) } @@ -5110,7 +4872,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v22 := compiler.MapValueForKey(m, "type") if v22 != nil { var err error - x.Type, err = NewTypeItem(v22, compiler.NewContext("type", context)) + x.Type, err = NewTypeItem(v22, compiler.NewContext("type", v22, context)) if err != nil { errors = append(errors, err) } @@ -5119,7 +4881,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v23 := compiler.MapValueForKey(m, "items") if v23 != nil { var err error - x.Items, err = NewItemsItem(v23, compiler.NewContext("items", context)) + x.Items, err = NewItemsItem(v23, compiler.NewContext("items", v23, context)) if err != nil { errors = append(errors, err) } @@ -5129,10 +4891,10 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { if v24 != nil { // repeated Schema x.AllOf = make([]*Schema, 0) - a, ok := v24.([]interface{}) + a, ok := compiler.SequenceNodeForNode(v24) if ok { - for _, item := range a { - y, err := NewSchema(item, compiler.NewContext("allOf", context)) + for _, item := range a.Content { + y, err := NewSchema(item, compiler.NewContext("allOf", item, context)) if err != nil { errors = append(errors, err) } @@ -5144,7 +4906,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v25 := compiler.MapValueForKey(m, "properties") if v25 != nil { var err error - x.Properties, err = NewProperties(v25, compiler.NewContext("properties", context)) + x.Properties, err = NewProperties(v25, compiler.NewContext("properties", v25, context)) if err != nil { errors = append(errors, err) } @@ -5152,18 +4914,18 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { // string discriminator = 26; v26 := compiler.MapValueForKey(m, "discriminator") if v26 != nil { - x.Discriminator, ok = v26.(string) + x.Discriminator, ok = compiler.StringForScalarNode(v26) if !ok { - message := fmt.Sprintf("has unexpected value for discriminator: %+v (%T)", v26, v26) + message := fmt.Sprintf("has unexpected value for discriminator: %s", compiler.Display(v26)) errors = append(errors, compiler.NewError(context, message)) } } // bool read_only = 27; v27 := compiler.MapValueForKey(m, "readOnly") if v27 != nil { - x.ReadOnly, ok = v27.(bool) + x.ReadOnly, ok = compiler.BoolForScalarNode(v27) if !ok { - message := fmt.Sprintf("has unexpected value for readOnly: %+v (%T)", v27, v27) + message := fmt.Sprintf("has unexpected value for readOnly: %s", compiler.Display(v27)) errors = append(errors, compiler.NewError(context, message)) } } @@ -5171,7 +4933,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v28 := compiler.MapValueForKey(m, "xml") if v28 != nil { var err error - x.Xml, err = NewXml(v28, compiler.NewContext("xml", context)) + x.Xml, err = NewXml(v28, compiler.NewContext("xml", v28, context)) if err != nil { errors = append(errors, err) } @@ -5180,7 +4942,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v29 := compiler.MapValueForKey(m, "externalDocs") if v29 != nil { var err error - x.ExternalDocs, err = NewExternalDocs(v29, compiler.NewContext("externalDocs", context)) + x.ExternalDocs, err = NewExternalDocs(v29, compiler.NewContext("externalDocs", v29, context)) if err != nil { errors = append(errors, err) } @@ -5189,7 +4951,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { v30 := compiler.MapValueForKey(m, "example") if v30 != nil { var err error - x.Example, err = NewAny(v30, compiler.NewContext("example", context)) + x.Example, err = NewAny(v30, compiler.NewContext("example", v30, context)) if err != nil { errors = append(errors, err) } @@ -5197,26 +4959,26 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { // repeated NamedAny vendor_extension = 31; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -5230,7 +4992,7 @@ func NewSchema(in interface{}, context *compiler.Context) (*Schema, error) { } // NewSchemaItem creates an object of type SchemaItem if possible, returning an error if not. -func NewSchemaItem(in interface{}, context *compiler.Context) (*SchemaItem, error) { +func NewSchemaItem(in *yaml.Node, context *compiler.Context) (*SchemaItem, error) { errors := make([]error, 0) x := &SchemaItem{} matched := false @@ -5239,7 +5001,7 @@ func NewSchemaItem(in interface{}, context *compiler.Context) (*SchemaItem, erro m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewSchema(m, compiler.NewContext("schema", context)) + t, matchingError := NewSchema(m, compiler.NewContext("schema", m, context)) if matchingError == nil { x.Oneof = &SchemaItem_Schema{Schema: t} matched = true @@ -5253,7 +5015,7 @@ func NewSchemaItem(in interface{}, context *compiler.Context) (*SchemaItem, erro m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewFileSchema(m, compiler.NewContext("fileSchema", context)) + t, matchingError := NewFileSchema(m, compiler.NewContext("fileSchema", m, context)) if matchingError == nil { x.Oneof = &SchemaItem_FileSchema{FileSchema: t} matched = true @@ -5265,12 +5027,16 @@ func NewSchemaItem(in interface{}, context *compiler.Context) (*SchemaItem, erro if matched { // since the oneof matched one of its possibilities, discard any matching errors errors = make([]error, 0) + } else { + message := fmt.Sprintf("contains an invalid SchemaItem") + err := compiler.NewError(context, message) + errors = []error{err} } return x, compiler.NewErrorGroupOrNil(errors) } // NewSecurityDefinitions creates an object of type SecurityDefinitions if possible, returning an error if not. -func NewSecurityDefinitions(in interface{}, context *compiler.Context) (*SecurityDefinitions, error) { +func NewSecurityDefinitions(in *yaml.Node, context *compiler.Context) (*SecurityDefinitions, error) { errors := make([]error, 0) x := &SecurityDefinitions{} m, ok := compiler.UnpackMap(in) @@ -5281,14 +5047,14 @@ func NewSecurityDefinitions(in interface{}, context *compiler.Context) (*Securit // repeated NamedSecurityDefinitionsItem additional_properties = 1; // MAP: SecurityDefinitionsItem x.AdditionalProperties = make([]*NamedSecurityDefinitionsItem, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedSecurityDefinitionsItem{} pair.Name = k var err error - pair.Value, err = NewSecurityDefinitionsItem(v, compiler.NewContext(k, context)) + pair.Value, err = NewSecurityDefinitionsItem(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -5300,7 +5066,7 @@ func NewSecurityDefinitions(in interface{}, context *compiler.Context) (*Securit } // NewSecurityDefinitionsItem creates an object of type SecurityDefinitionsItem if possible, returning an error if not. -func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*SecurityDefinitionsItem, error) { +func NewSecurityDefinitionsItem(in *yaml.Node, context *compiler.Context) (*SecurityDefinitionsItem, error) { errors := make([]error, 0) x := &SecurityDefinitionsItem{} matched := false @@ -5309,7 +5075,7 @@ func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*Sec m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewBasicAuthenticationSecurity(m, compiler.NewContext("basicAuthenticationSecurity", context)) + t, matchingError := NewBasicAuthenticationSecurity(m, compiler.NewContext("basicAuthenticationSecurity", m, context)) if matchingError == nil { x.Oneof = &SecurityDefinitionsItem_BasicAuthenticationSecurity{BasicAuthenticationSecurity: t} matched = true @@ -5323,7 +5089,7 @@ func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*Sec m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewApiKeySecurity(m, compiler.NewContext("apiKeySecurity", context)) + t, matchingError := NewApiKeySecurity(m, compiler.NewContext("apiKeySecurity", m, context)) if matchingError == nil { x.Oneof = &SecurityDefinitionsItem_ApiKeySecurity{ApiKeySecurity: t} matched = true @@ -5337,7 +5103,7 @@ func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*Sec m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewOauth2ImplicitSecurity(m, compiler.NewContext("oauth2ImplicitSecurity", context)) + t, matchingError := NewOauth2ImplicitSecurity(m, compiler.NewContext("oauth2ImplicitSecurity", m, context)) if matchingError == nil { x.Oneof = &SecurityDefinitionsItem_Oauth2ImplicitSecurity{Oauth2ImplicitSecurity: t} matched = true @@ -5351,7 +5117,7 @@ func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*Sec m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewOauth2PasswordSecurity(m, compiler.NewContext("oauth2PasswordSecurity", context)) + t, matchingError := NewOauth2PasswordSecurity(m, compiler.NewContext("oauth2PasswordSecurity", m, context)) if matchingError == nil { x.Oneof = &SecurityDefinitionsItem_Oauth2PasswordSecurity{Oauth2PasswordSecurity: t} matched = true @@ -5365,7 +5131,7 @@ func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*Sec m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewOauth2ApplicationSecurity(m, compiler.NewContext("oauth2ApplicationSecurity", context)) + t, matchingError := NewOauth2ApplicationSecurity(m, compiler.NewContext("oauth2ApplicationSecurity", m, context)) if matchingError == nil { x.Oneof = &SecurityDefinitionsItem_Oauth2ApplicationSecurity{Oauth2ApplicationSecurity: t} matched = true @@ -5379,7 +5145,7 @@ func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*Sec m, ok := compiler.UnpackMap(in) if ok { // errors might be ok here, they mean we just don't have the right subtype - t, matchingError := NewOauth2AccessCodeSecurity(m, compiler.NewContext("oauth2AccessCodeSecurity", context)) + t, matchingError := NewOauth2AccessCodeSecurity(m, compiler.NewContext("oauth2AccessCodeSecurity", m, context)) if matchingError == nil { x.Oneof = &SecurityDefinitionsItem_Oauth2AccessCodeSecurity{Oauth2AccessCodeSecurity: t} matched = true @@ -5391,12 +5157,16 @@ func NewSecurityDefinitionsItem(in interface{}, context *compiler.Context) (*Sec if matched { // since the oneof matched one of its possibilities, discard any matching errors errors = make([]error, 0) + } else { + message := fmt.Sprintf("contains an invalid SecurityDefinitionsItem") + err := compiler.NewError(context, message) + errors = []error{err} } return x, compiler.NewErrorGroupOrNil(errors) } // NewSecurityRequirement creates an object of type SecurityRequirement if possible, returning an error if not. -func NewSecurityRequirement(in interface{}, context *compiler.Context) (*SecurityRequirement, error) { +func NewSecurityRequirement(in *yaml.Node, context *compiler.Context) (*SecurityRequirement, error) { errors := make([]error, 0) x := &SecurityRequirement{} m, ok := compiler.UnpackMap(in) @@ -5407,14 +5177,14 @@ func NewSecurityRequirement(in interface{}, context *compiler.Context) (*Securit // repeated NamedStringArray additional_properties = 1; // MAP: StringArray x.AdditionalProperties = make([]*NamedStringArray, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedStringArray{} pair.Name = k var err error - pair.Value, err = NewStringArray(v, compiler.NewContext(k, context)) + pair.Value, err = NewStringArray(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -5426,24 +5196,19 @@ func NewSecurityRequirement(in interface{}, context *compiler.Context) (*Securit } // NewStringArray creates an object of type StringArray if possible, returning an error if not. -func NewStringArray(in interface{}, context *compiler.Context) (*StringArray, error) { +func NewStringArray(in *yaml.Node, context *compiler.Context) (*StringArray, error) { errors := make([]error, 0) x := &StringArray{} - a, ok := in.([]interface{}) - if !ok { - message := fmt.Sprintf("has unexpected value for StringArray: %+v (%T)", in, in) - errors = append(errors, compiler.NewError(context, message)) - } else { - x.Value = make([]string, 0) - for _, s := range a { - x.Value = append(x.Value, s.(string)) - } + x.Value = make([]string, 0) + for _, node := range in.Content { + s, _ := compiler.StringForScalarNode(node) + x.Value = append(x.Value, s) } return x, compiler.NewErrorGroupOrNil(errors) } // NewTag creates an object of type Tag if possible, returning an error if not. -func NewTag(in interface{}, context *compiler.Context) (*Tag, error) { +func NewTag(in *yaml.Node, context *compiler.Context) (*Tag, error) { errors := make([]error, 0) x := &Tag{} m, ok := compiler.UnpackMap(in) @@ -5467,18 +5232,18 @@ func NewTag(in interface{}, context *compiler.Context) (*Tag, error) { // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string description = 2; v2 := compiler.MapValueForKey(m, "description") if v2 != nil { - x.Description, ok = v2.(string) + x.Description, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for description: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for description: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } @@ -5486,7 +5251,7 @@ func NewTag(in interface{}, context *compiler.Context) (*Tag, error) { v3 := compiler.MapValueForKey(m, "externalDocs") if v3 != nil { var err error - x.ExternalDocs, err = NewExternalDocs(v3, compiler.NewContext("externalDocs", context)) + x.ExternalDocs, err = NewExternalDocs(v3, compiler.NewContext("externalDocs", v3, context)) if err != nil { errors = append(errors, err) } @@ -5494,26 +5259,26 @@ func NewTag(in interface{}, context *compiler.Context) (*Tag, error) { // repeated NamedAny vendor_extension = 4; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -5527,17 +5292,19 @@ func NewTag(in interface{}, context *compiler.Context) (*Tag, error) { } // NewTypeItem creates an object of type TypeItem if possible, returning an error if not. -func NewTypeItem(in interface{}, context *compiler.Context) (*TypeItem, error) { +func NewTypeItem(in *yaml.Node, context *compiler.Context) (*TypeItem, error) { errors := make([]error, 0) x := &TypeItem{} - switch in := in.(type) { - case string: + v1 := in + switch v1.Kind { + case yaml.ScalarNode: x.Value = make([]string, 0) - x.Value = append(x.Value, in) - case []interface{}: + x.Value = append(x.Value, v1.Value) + case yaml.SequenceNode: x.Value = make([]string, 0) - for _, v := range in { - value, ok := v.(string) + for _, v := range v1.Content { + value := v.Value + ok := v.Kind == yaml.ScalarNode if ok { x.Value = append(x.Value, value) } else { @@ -5553,7 +5320,7 @@ func NewTypeItem(in interface{}, context *compiler.Context) (*TypeItem, error) { } // NewVendorExtension creates an object of type VendorExtension if possible, returning an error if not. -func NewVendorExtension(in interface{}, context *compiler.Context) (*VendorExtension, error) { +func NewVendorExtension(in *yaml.Node, context *compiler.Context) (*VendorExtension, error) { errors := make([]error, 0) x := &VendorExtension{} m, ok := compiler.UnpackMap(in) @@ -5564,25 +5331,25 @@ func NewVendorExtension(in interface{}, context *compiler.Context) (*VendorExten // repeated NamedAny additional_properties = 1; // MAP: Any x.AdditionalProperties = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -5595,7 +5362,7 @@ func NewVendorExtension(in interface{}, context *compiler.Context) (*VendorExten } // NewXml creates an object of type Xml if possible, returning an error if not. -func NewXml(in interface{}, context *compiler.Context) (*Xml, error) { +func NewXml(in *yaml.Node, context *compiler.Context) (*Xml, error) { errors := make([]error, 0) x := &Xml{} m, ok := compiler.UnpackMap(in) @@ -5613,71 +5380,71 @@ func NewXml(in interface{}, context *compiler.Context) (*Xml, error) { // string name = 1; v1 := compiler.MapValueForKey(m, "name") if v1 != nil { - x.Name, ok = v1.(string) + x.Name, ok = compiler.StringForScalarNode(v1) if !ok { - message := fmt.Sprintf("has unexpected value for name: %+v (%T)", v1, v1) + message := fmt.Sprintf("has unexpected value for name: %s", compiler.Display(v1)) errors = append(errors, compiler.NewError(context, message)) } } // string namespace = 2; v2 := compiler.MapValueForKey(m, "namespace") if v2 != nil { - x.Namespace, ok = v2.(string) + x.Namespace, ok = compiler.StringForScalarNode(v2) if !ok { - message := fmt.Sprintf("has unexpected value for namespace: %+v (%T)", v2, v2) + message := fmt.Sprintf("has unexpected value for namespace: %s", compiler.Display(v2)) errors = append(errors, compiler.NewError(context, message)) } } // string prefix = 3; v3 := compiler.MapValueForKey(m, "prefix") if v3 != nil { - x.Prefix, ok = v3.(string) + x.Prefix, ok = compiler.StringForScalarNode(v3) if !ok { - message := fmt.Sprintf("has unexpected value for prefix: %+v (%T)", v3, v3) + message := fmt.Sprintf("has unexpected value for prefix: %s", compiler.Display(v3)) errors = append(errors, compiler.NewError(context, message)) } } // bool attribute = 4; v4 := compiler.MapValueForKey(m, "attribute") if v4 != nil { - x.Attribute, ok = v4.(bool) + x.Attribute, ok = compiler.BoolForScalarNode(v4) if !ok { - message := fmt.Sprintf("has unexpected value for attribute: %+v (%T)", v4, v4) + message := fmt.Sprintf("has unexpected value for attribute: %s", compiler.Display(v4)) errors = append(errors, compiler.NewError(context, message)) } } // bool wrapped = 5; v5 := compiler.MapValueForKey(m, "wrapped") if v5 != nil { - x.Wrapped, ok = v5.(bool) + x.Wrapped, ok = compiler.BoolForScalarNode(v5) if !ok { - message := fmt.Sprintf("has unexpected value for wrapped: %+v (%T)", v5, v5) + message := fmt.Sprintf("has unexpected value for wrapped: %s", compiler.Display(v5)) errors = append(errors, compiler.NewError(context, message)) } } // repeated NamedAny vendor_extension = 6; // MAP: Any ^x- x.VendorExtension = make([]*NamedAny, 0) - for _, item := range m { - k, ok := compiler.StringValue(item.Key) + for i := 0; i < len(m.Content); i += 2 { + k, ok := compiler.StringForScalarNode(m.Content[i]) if ok { - v := item.Value + v := m.Content[i+1] if strings.HasPrefix(k, "x-") { pair := &NamedAny{} pair.Name = k result := &Any{} - handled, resultFromExt, err := compiler.HandleExtension(context, v, k) + handled, resultFromExt, err := compiler.CallExtension(context, v, k) if handled { if err != nil { errors = append(errors, err) } else { - bytes, _ := yaml.Marshal(v) + bytes := compiler.Marshal(v) result.Yaml = string(bytes) result.Value = resultFromExt pair.Value = result } } else { - pair.Value, err = NewAny(v, compiler.NewContext(k, context)) + pair.Value, err = NewAny(v, compiler.NewContext(k, v, context)) if err != nil { errors = append(errors, err) } @@ -5691,7 +5458,7 @@ func NewXml(in interface{}, context *compiler.Context) (*Xml, error) { } // ResolveReferences resolves references found inside AdditionalPropertiesItem objects. -func (m *AdditionalPropertiesItem) ResolveReferences(root string) (interface{}, error) { +func (m *AdditionalPropertiesItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) { p, ok := m.Oneof.(*AdditionalPropertiesItem_Schema) @@ -5706,13 +5473,13 @@ func (m *AdditionalPropertiesItem) ResolveReferences(root string) (interface{}, } // ResolveReferences resolves references found inside Any objects. -func (m *Any) ResolveReferences(root string) (interface{}, error) { +func (m *Any) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) return nil, compiler.NewErrorGroupOrNil(errors) } // ResolveReferences resolves references found inside ApiKeySecurity objects. -func (m *ApiKeySecurity) ResolveReferences(root string) (interface{}, error) { +func (m *ApiKeySecurity) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.VendorExtension { if item != nil { @@ -5726,7 +5493,7 @@ func (m *ApiKeySecurity) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside BasicAuthenticationSecurity objects. -func (m *BasicAuthenticationSecurity) ResolveReferences(root string) (interface{}, error) { +func (m *BasicAuthenticationSecurity) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.VendorExtension { if item != nil { @@ -5740,7 +5507,7 @@ func (m *BasicAuthenticationSecurity) ResolveReferences(root string) (interface{ } // ResolveReferences resolves references found inside BodyParameter objects. -func (m *BodyParameter) ResolveReferences(root string) (interface{}, error) { +func (m *BodyParameter) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Schema != nil { _, err := m.Schema.ResolveReferences(root) @@ -5760,7 +5527,7 @@ func (m *BodyParameter) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Contact objects. -func (m *Contact) ResolveReferences(root string) (interface{}, error) { +func (m *Contact) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.VendorExtension { if item != nil { @@ -5774,7 +5541,7 @@ func (m *Contact) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Default objects. -func (m *Default) ResolveReferences(root string) (interface{}, error) { +func (m *Default) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -5788,7 +5555,7 @@ func (m *Default) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Definitions objects. -func (m *Definitions) ResolveReferences(root string) (interface{}, error) { +func (m *Definitions) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -5802,7 +5569,7 @@ func (m *Definitions) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Document objects. -func (m *Document) ResolveReferences(root string) (interface{}, error) { +func (m *Document) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Info != nil { _, err := m.Info.ResolveReferences(root) @@ -5874,7 +5641,7 @@ func (m *Document) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Examples objects. -func (m *Examples) ResolveReferences(root string) (interface{}, error) { +func (m *Examples) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -5888,7 +5655,7 @@ func (m *Examples) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside ExternalDocs objects. -func (m *ExternalDocs) ResolveReferences(root string) (interface{}, error) { +func (m *ExternalDocs) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.VendorExtension { if item != nil { @@ -5902,7 +5669,7 @@ func (m *ExternalDocs) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside FileSchema objects. -func (m *FileSchema) ResolveReferences(root string) (interface{}, error) { +func (m *FileSchema) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Default != nil { _, err := m.Default.ResolveReferences(root) @@ -5934,7 +5701,7 @@ func (m *FileSchema) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside FormDataParameterSubSchema objects. -func (m *FormDataParameterSubSchema) ResolveReferences(root string) (interface{}, error) { +func (m *FormDataParameterSubSchema) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Items != nil { _, err := m.Items.ResolveReferences(root) @@ -5968,7 +5735,7 @@ func (m *FormDataParameterSubSchema) ResolveReferences(root string) (interface{} } // ResolveReferences resolves references found inside Header objects. -func (m *Header) ResolveReferences(root string) (interface{}, error) { +func (m *Header) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Items != nil { _, err := m.Items.ResolveReferences(root) @@ -6002,7 +5769,7 @@ func (m *Header) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside HeaderParameterSubSchema objects. -func (m *HeaderParameterSubSchema) ResolveReferences(root string) (interface{}, error) { +func (m *HeaderParameterSubSchema) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Items != nil { _, err := m.Items.ResolveReferences(root) @@ -6036,7 +5803,7 @@ func (m *HeaderParameterSubSchema) ResolveReferences(root string) (interface{}, } // ResolveReferences resolves references found inside Headers objects. -func (m *Headers) ResolveReferences(root string) (interface{}, error) { +func (m *Headers) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -6050,7 +5817,7 @@ func (m *Headers) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Info objects. -func (m *Info) ResolveReferences(root string) (interface{}, error) { +func (m *Info) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Contact != nil { _, err := m.Contact.ResolveReferences(root) @@ -6076,7 +5843,7 @@ func (m *Info) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside ItemsItem objects. -func (m *ItemsItem) ResolveReferences(root string) (interface{}, error) { +func (m *ItemsItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.Schema { if item != nil { @@ -6090,7 +5857,7 @@ func (m *ItemsItem) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside JsonReference objects. -func (m *JsonReference) ResolveReferences(root string) (interface{}, error) { +func (m *JsonReference) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.XRef != "" { info, err := compiler.ReadInfoForRef(root, m.XRef) @@ -6110,7 +5877,7 @@ func (m *JsonReference) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside License objects. -func (m *License) ResolveReferences(root string) (interface{}, error) { +func (m *License) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.VendorExtension { if item != nil { @@ -6124,7 +5891,7 @@ func (m *License) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NamedAny objects. -func (m *NamedAny) ResolveReferences(root string) (interface{}, error) { +func (m *NamedAny) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6136,7 +5903,7 @@ func (m *NamedAny) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NamedHeader objects. -func (m *NamedHeader) ResolveReferences(root string) (interface{}, error) { +func (m *NamedHeader) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6148,7 +5915,7 @@ func (m *NamedHeader) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NamedParameter objects. -func (m *NamedParameter) ResolveReferences(root string) (interface{}, error) { +func (m *NamedParameter) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6160,7 +5927,7 @@ func (m *NamedParameter) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NamedPathItem objects. -func (m *NamedPathItem) ResolveReferences(root string) (interface{}, error) { +func (m *NamedPathItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6172,7 +5939,7 @@ func (m *NamedPathItem) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NamedResponse objects. -func (m *NamedResponse) ResolveReferences(root string) (interface{}, error) { +func (m *NamedResponse) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6184,7 +5951,7 @@ func (m *NamedResponse) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NamedResponseValue objects. -func (m *NamedResponseValue) ResolveReferences(root string) (interface{}, error) { +func (m *NamedResponseValue) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6196,7 +5963,7 @@ func (m *NamedResponseValue) ResolveReferences(root string) (interface{}, error) } // ResolveReferences resolves references found inside NamedSchema objects. -func (m *NamedSchema) ResolveReferences(root string) (interface{}, error) { +func (m *NamedSchema) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6208,7 +5975,7 @@ func (m *NamedSchema) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NamedSecurityDefinitionsItem objects. -func (m *NamedSecurityDefinitionsItem) ResolveReferences(root string) (interface{}, error) { +func (m *NamedSecurityDefinitionsItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6220,13 +5987,13 @@ func (m *NamedSecurityDefinitionsItem) ResolveReferences(root string) (interface } // ResolveReferences resolves references found inside NamedString objects. -func (m *NamedString) ResolveReferences(root string) (interface{}, error) { +func (m *NamedString) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) return nil, compiler.NewErrorGroupOrNil(errors) } // ResolveReferences resolves references found inside NamedStringArray objects. -func (m *NamedStringArray) ResolveReferences(root string) (interface{}, error) { +func (m *NamedStringArray) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Value != nil { _, err := m.Value.ResolveReferences(root) @@ -6238,7 +6005,7 @@ func (m *NamedStringArray) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside NonBodyParameter objects. -func (m *NonBodyParameter) ResolveReferences(root string) (interface{}, error) { +func (m *NonBodyParameter) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) { p, ok := m.Oneof.(*NonBodyParameter_HeaderParameterSubSchema) @@ -6280,7 +6047,7 @@ func (m *NonBodyParameter) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Oauth2AccessCodeSecurity objects. -func (m *Oauth2AccessCodeSecurity) ResolveReferences(root string) (interface{}, error) { +func (m *Oauth2AccessCodeSecurity) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Scopes != nil { _, err := m.Scopes.ResolveReferences(root) @@ -6300,7 +6067,7 @@ func (m *Oauth2AccessCodeSecurity) ResolveReferences(root string) (interface{}, } // ResolveReferences resolves references found inside Oauth2ApplicationSecurity objects. -func (m *Oauth2ApplicationSecurity) ResolveReferences(root string) (interface{}, error) { +func (m *Oauth2ApplicationSecurity) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Scopes != nil { _, err := m.Scopes.ResolveReferences(root) @@ -6320,7 +6087,7 @@ func (m *Oauth2ApplicationSecurity) ResolveReferences(root string) (interface{}, } // ResolveReferences resolves references found inside Oauth2ImplicitSecurity objects. -func (m *Oauth2ImplicitSecurity) ResolveReferences(root string) (interface{}, error) { +func (m *Oauth2ImplicitSecurity) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Scopes != nil { _, err := m.Scopes.ResolveReferences(root) @@ -6340,7 +6107,7 @@ func (m *Oauth2ImplicitSecurity) ResolveReferences(root string) (interface{}, er } // ResolveReferences resolves references found inside Oauth2PasswordSecurity objects. -func (m *Oauth2PasswordSecurity) ResolveReferences(root string) (interface{}, error) { +func (m *Oauth2PasswordSecurity) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Scopes != nil { _, err := m.Scopes.ResolveReferences(root) @@ -6360,7 +6127,7 @@ func (m *Oauth2PasswordSecurity) ResolveReferences(root string) (interface{}, er } // ResolveReferences resolves references found inside Oauth2Scopes objects. -func (m *Oauth2Scopes) ResolveReferences(root string) (interface{}, error) { +func (m *Oauth2Scopes) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -6374,7 +6141,7 @@ func (m *Oauth2Scopes) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Operation objects. -func (m *Operation) ResolveReferences(root string) (interface{}, error) { +func (m *Operation) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.ExternalDocs != nil { _, err := m.ExternalDocs.ResolveReferences(root) @@ -6416,7 +6183,7 @@ func (m *Operation) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Parameter objects. -func (m *Parameter) ResolveReferences(root string) (interface{}, error) { +func (m *Parameter) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) { p, ok := m.Oneof.(*Parameter_BodyParameter) @@ -6440,7 +6207,7 @@ func (m *Parameter) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside ParameterDefinitions objects. -func (m *ParameterDefinitions) ResolveReferences(root string) (interface{}, error) { +func (m *ParameterDefinitions) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -6454,7 +6221,7 @@ func (m *ParameterDefinitions) ResolveReferences(root string) (interface{}, erro } // ResolveReferences resolves references found inside ParametersItem objects. -func (m *ParametersItem) ResolveReferences(root string) (interface{}, error) { +func (m *ParametersItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) { p, ok := m.Oneof.(*ParametersItem_Parameter) @@ -6486,7 +6253,7 @@ func (m *ParametersItem) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside PathItem objects. -func (m *PathItem) ResolveReferences(root string) (interface{}, error) { +func (m *PathItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.XRef != "" { info, err := compiler.ReadInfoForRef(root, m.XRef) @@ -6564,7 +6331,7 @@ func (m *PathItem) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside PathParameterSubSchema objects. -func (m *PathParameterSubSchema) ResolveReferences(root string) (interface{}, error) { +func (m *PathParameterSubSchema) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Items != nil { _, err := m.Items.ResolveReferences(root) @@ -6598,7 +6365,7 @@ func (m *PathParameterSubSchema) ResolveReferences(root string) (interface{}, er } // ResolveReferences resolves references found inside Paths objects. -func (m *Paths) ResolveReferences(root string) (interface{}, error) { +func (m *Paths) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.VendorExtension { if item != nil { @@ -6620,7 +6387,7 @@ func (m *Paths) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside PrimitivesItems objects. -func (m *PrimitivesItems) ResolveReferences(root string) (interface{}, error) { +func (m *PrimitivesItems) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Items != nil { _, err := m.Items.ResolveReferences(root) @@ -6654,7 +6421,7 @@ func (m *PrimitivesItems) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Properties objects. -func (m *Properties) ResolveReferences(root string) (interface{}, error) { +func (m *Properties) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -6668,7 +6435,7 @@ func (m *Properties) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside QueryParameterSubSchema objects. -func (m *QueryParameterSubSchema) ResolveReferences(root string) (interface{}, error) { +func (m *QueryParameterSubSchema) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Items != nil { _, err := m.Items.ResolveReferences(root) @@ -6702,7 +6469,7 @@ func (m *QueryParameterSubSchema) ResolveReferences(root string) (interface{}, e } // ResolveReferences resolves references found inside Response objects. -func (m *Response) ResolveReferences(root string) (interface{}, error) { +func (m *Response) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.Schema != nil { _, err := m.Schema.ResolveReferences(root) @@ -6734,7 +6501,7 @@ func (m *Response) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside ResponseDefinitions objects. -func (m *ResponseDefinitions) ResolveReferences(root string) (interface{}, error) { +func (m *ResponseDefinitions) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -6748,7 +6515,7 @@ func (m *ResponseDefinitions) ResolveReferences(root string) (interface{}, error } // ResolveReferences resolves references found inside ResponseValue objects. -func (m *ResponseValue) ResolveReferences(root string) (interface{}, error) { +func (m *ResponseValue) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) { p, ok := m.Oneof.(*ResponseValue_Response) @@ -6780,7 +6547,7 @@ func (m *ResponseValue) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Responses objects. -func (m *Responses) ResolveReferences(root string) (interface{}, error) { +func (m *Responses) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.ResponseCode { if item != nil { @@ -6802,7 +6569,7 @@ func (m *Responses) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Schema objects. -func (m *Schema) ResolveReferences(root string) (interface{}, error) { +func (m *Schema) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.XRef != "" { info, err := compiler.ReadInfoForRef(root, m.XRef) @@ -6894,7 +6661,7 @@ func (m *Schema) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside SchemaItem objects. -func (m *SchemaItem) ResolveReferences(root string) (interface{}, error) { +func (m *SchemaItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) { p, ok := m.Oneof.(*SchemaItem_Schema) @@ -6918,7 +6685,7 @@ func (m *SchemaItem) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside SecurityDefinitions objects. -func (m *SecurityDefinitions) ResolveReferences(root string) (interface{}, error) { +func (m *SecurityDefinitions) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -6932,7 +6699,7 @@ func (m *SecurityDefinitions) ResolveReferences(root string) (interface{}, error } // ResolveReferences resolves references found inside SecurityDefinitionsItem objects. -func (m *SecurityDefinitionsItem) ResolveReferences(root string) (interface{}, error) { +func (m *SecurityDefinitionsItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) { p, ok := m.Oneof.(*SecurityDefinitionsItem_BasicAuthenticationSecurity) @@ -6992,7 +6759,7 @@ func (m *SecurityDefinitionsItem) ResolveReferences(root string) (interface{}, e } // ResolveReferences resolves references found inside SecurityRequirement objects. -func (m *SecurityRequirement) ResolveReferences(root string) (interface{}, error) { +func (m *SecurityRequirement) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -7006,13 +6773,13 @@ func (m *SecurityRequirement) ResolveReferences(root string) (interface{}, error } // ResolveReferences resolves references found inside StringArray objects. -func (m *StringArray) ResolveReferences(root string) (interface{}, error) { +func (m *StringArray) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) return nil, compiler.NewErrorGroupOrNil(errors) } // ResolveReferences resolves references found inside Tag objects. -func (m *Tag) ResolveReferences(root string) (interface{}, error) { +func (m *Tag) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) if m.ExternalDocs != nil { _, err := m.ExternalDocs.ResolveReferences(root) @@ -7032,13 +6799,13 @@ func (m *Tag) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside TypeItem objects. -func (m *TypeItem) ResolveReferences(root string) (interface{}, error) { +func (m *TypeItem) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) return nil, compiler.NewErrorGroupOrNil(errors) } // ResolveReferences resolves references found inside VendorExtension objects. -func (m *VendorExtension) ResolveReferences(root string) (interface{}, error) { +func (m *VendorExtension) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.AdditionalProperties { if item != nil { @@ -7052,7 +6819,7 @@ func (m *VendorExtension) ResolveReferences(root string) (interface{}, error) { } // ResolveReferences resolves references found inside Xml objects. -func (m *Xml) ResolveReferences(root string) (interface{}, error) { +func (m *Xml) ResolveReferences(root string) (*yaml.Node, error) { errors := make([]error, 0) for _, item := range m.VendorExtension { if item != nil { @@ -7066,7 +6833,7 @@ func (m *Xml) ResolveReferences(root string) (interface{}, error) { } // ToRawInfo returns a description of AdditionalPropertiesItem suitable for JSON or YAML export. -func (m *AdditionalPropertiesItem) ToRawInfo() interface{} { +func (m *AdditionalPropertiesItem) ToRawInfo() *yaml.Node { // ONE OF WRAPPER // AdditionalPropertiesItem // {Name:schema Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -7076,792 +6843,883 @@ func (m *AdditionalPropertiesItem) ToRawInfo() interface{} { } // {Name:boolean Type:bool StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if v1, ok := m.GetOneof().(*AdditionalPropertiesItem_Boolean); ok { - return v1.Boolean + return compiler.NewScalarNodeForBool(v1.Boolean) } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of Any suitable for JSON or YAML export. -func (m *Any) ToRawInfo() interface{} { +func (m *Any) ToRawInfo() *yaml.Node { var err error - var info1 []yaml.MapSlice - err = yaml.Unmarshal([]byte(m.Yaml), &info1) - if err == nil { - return info1 - } - var info2 yaml.MapSlice - err = yaml.Unmarshal([]byte(m.Yaml), &info2) + var node yaml.Node + err = yaml.Unmarshal([]byte(m.Yaml), &node) if err == nil { - return info2 - } - var info3 interface{} - err = yaml.Unmarshal([]byte(m.Yaml), &info3) - if err == nil { - return info3 + if node.Kind == yaml.DocumentNode { + return node.Content[0] + } + return &node } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of ApiKeySecurity suitable for JSON or YAML export. -func (m *ApiKeySecurity) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *ApiKeySecurity) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) // always include this required field. - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) // always include this required field. - info = append(info, yaml.MapItem{Key: "in", Value: m.In}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("in")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.In)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of BasicAuthenticationSecurity suitable for JSON or YAML export. -func (m *BasicAuthenticationSecurity) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *BasicAuthenticationSecurity) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of BodyParameter suitable for JSON or YAML export. -func (m *BodyParameter) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *BodyParameter) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } // always include this required field. - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) // always include this required field. - info = append(info, yaml.MapItem{Key: "in", Value: m.In}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("in")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.In)) if m.Required != false { - info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("required")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Required)) } // always include this required field. - info = append(info, yaml.MapItem{Key: "schema", Value: m.Schema.ToRawInfo()}) - // &{Name:schema Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} + info.Content = append(info.Content, compiler.NewScalarNodeForString("schema")) + info.Content = append(info.Content, m.Schema.ToRawInfo()) if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Contact suitable for JSON or YAML export. -func (m *Contact) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Contact) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } if m.Url != "" { - info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("url")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Url)) } if m.Email != "" { - info = append(info, yaml.MapItem{Key: "email", Value: m.Email}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("email")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Email)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Default suitable for JSON or YAML export. -func (m *Default) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Default) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:false Description:} return info } // ToRawInfo returns a description of Definitions suitable for JSON or YAML export. -func (m *Definitions) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Definitions) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedSchema StringEnumValues:[] MapType:Schema Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of Document suitable for JSON or YAML export. -func (m *Document) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Document) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "swagger", Value: m.Swagger}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("swagger")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Swagger)) // always include this required field. - info = append(info, yaml.MapItem{Key: "info", Value: m.Info.ToRawInfo()}) - // &{Name:info Type:Info StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} + info.Content = append(info.Content, compiler.NewScalarNodeForString("info")) + info.Content = append(info.Content, m.Info.ToRawInfo()) if m.Host != "" { - info = append(info, yaml.MapItem{Key: "host", Value: m.Host}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("host")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Host)) } if m.BasePath != "" { - info = append(info, yaml.MapItem{Key: "basePath", Value: m.BasePath}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("basePath")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.BasePath)) } if len(m.Schemes) != 0 { - info = append(info, yaml.MapItem{Key: "schemes", Value: m.Schemes}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("schemes")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Schemes)) } if len(m.Consumes) != 0 { - info = append(info, yaml.MapItem{Key: "consumes", Value: m.Consumes}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("consumes")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Consumes)) } if len(m.Produces) != 0 { - info = append(info, yaml.MapItem{Key: "produces", Value: m.Produces}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("produces")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Produces)) } // always include this required field. - info = append(info, yaml.MapItem{Key: "paths", Value: m.Paths.ToRawInfo()}) - // &{Name:paths Type:Paths StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} + info.Content = append(info.Content, compiler.NewScalarNodeForString("paths")) + info.Content = append(info.Content, m.Paths.ToRawInfo()) if m.Definitions != nil { - info = append(info, yaml.MapItem{Key: "definitions", Value: m.Definitions.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("definitions")) + info.Content = append(info.Content, m.Definitions.ToRawInfo()) } - // &{Name:definitions Type:Definitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Parameters != nil { - info = append(info, yaml.MapItem{Key: "parameters", Value: m.Parameters.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("parameters")) + info.Content = append(info.Content, m.Parameters.ToRawInfo()) } - // &{Name:parameters Type:ParameterDefinitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Responses != nil { - info = append(info, yaml.MapItem{Key: "responses", Value: m.Responses.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("responses")) + info.Content = append(info.Content, m.Responses.ToRawInfo()) } - // &{Name:responses Type:ResponseDefinitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Security) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Security { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "security", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("security")) + info.Content = append(info.Content, items) } - // &{Name:security Type:SecurityRequirement StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.SecurityDefinitions != nil { - info = append(info, yaml.MapItem{Key: "securityDefinitions", Value: m.SecurityDefinitions.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("securityDefinitions")) + info.Content = append(info.Content, m.SecurityDefinitions.ToRawInfo()) } - // &{Name:securityDefinitions Type:SecurityDefinitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Tags) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Tags { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "tags", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("tags")) + info.Content = append(info.Content, items) } - // &{Name:tags Type:Tag StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("externalDocs")) + info.Content = append(info.Content, m.ExternalDocs.ToRawInfo()) } - // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Examples suitable for JSON or YAML export. -func (m *Examples) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Examples) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of ExternalDocs suitable for JSON or YAML export. -func (m *ExternalDocs) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *ExternalDocs) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } // always include this required field. - info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("url")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Url)) if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of FileSchema suitable for JSON or YAML export. -func (m *FileSchema) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *FileSchema) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Title != "" { - info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("title")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Title)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Required) != 0 { - info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("required")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Required)) } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) if m.ReadOnly != false { - info = append(info, yaml.MapItem{Key: "readOnly", Value: m.ReadOnly}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("readOnly")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ReadOnly)) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("externalDocs")) + info.Content = append(info.Content, m.ExternalDocs.ToRawInfo()) } - // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("example")) + info.Content = append(info.Content, m.Example.ToRawInfo()) } - // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of FormDataParameterSubSchema suitable for JSON or YAML export. -func (m *FormDataParameterSubSchema) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *FormDataParameterSubSchema) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Required != false { - info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("required")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Required)) } if m.In != "" { - info = append(info, yaml.MapItem{Key: "in", Value: m.In}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("in")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.In)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } if m.AllowEmptyValue != false { - info = append(info, yaml.MapItem{Key: "allowEmptyValue", Value: m.AllowEmptyValue}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("allowEmptyValue")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.AllowEmptyValue)) } if m.Type != "" { - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) } if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Items != nil { - info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("items")) + info.Content = append(info.Content, m.Items.ToRawInfo()) } - // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("collectionFormat")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.CollectionFormat)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Maximum)) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMaximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMaximum)) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Minimum)) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMinimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMinimum)) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxLength)) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinLength)) } if m.Pattern != "" { - info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("pattern")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Pattern)) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxItems)) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinItems)) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("uniqueItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.UniqueItems)) } if len(m.Enum) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Enum { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "enum", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("enum")) + info.Content = append(info.Content, items) } - // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("multipleOf")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.MultipleOf)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Header suitable for JSON or YAML export. -func (m *Header) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Header) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Items != nil { - info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("items")) + info.Content = append(info.Content, m.Items.ToRawInfo()) } - // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("collectionFormat")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.CollectionFormat)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Maximum)) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMaximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMaximum)) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Minimum)) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMinimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMinimum)) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxLength)) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinLength)) } if m.Pattern != "" { - info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("pattern")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Pattern)) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxItems)) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinItems)) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("uniqueItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.UniqueItems)) } if len(m.Enum) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Enum { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "enum", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("enum")) + info.Content = append(info.Content, items) } - // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("multipleOf")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.MultipleOf)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of HeaderParameterSubSchema suitable for JSON or YAML export. -func (m *HeaderParameterSubSchema) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *HeaderParameterSubSchema) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Required != false { - info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("required")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Required)) } if m.In != "" { - info = append(info, yaml.MapItem{Key: "in", Value: m.In}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("in")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.In)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } if m.Type != "" { - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) } if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Items != nil { - info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("items")) + info.Content = append(info.Content, m.Items.ToRawInfo()) } - // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("collectionFormat")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.CollectionFormat)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Maximum)) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMaximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMaximum)) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Minimum)) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMinimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMinimum)) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxLength)) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinLength)) } if m.Pattern != "" { - info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("pattern")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Pattern)) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxItems)) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinItems)) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("uniqueItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.UniqueItems)) } if len(m.Enum) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Enum { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "enum", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("enum")) + info.Content = append(info.Content, items) } - // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("multipleOf")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.MultipleOf)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Headers suitable for JSON or YAML export. -func (m *Headers) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Headers) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedHeader StringEnumValues:[] MapType:Header Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of Info suitable for JSON or YAML export. -func (m *Info) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Info) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("title")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Title)) // always include this required field. - info = append(info, yaml.MapItem{Key: "version", Value: m.Version}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("version")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Version)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.TermsOfService != "" { - info = append(info, yaml.MapItem{Key: "termsOfService", Value: m.TermsOfService}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("termsOfService")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.TermsOfService)) } if m.Contact != nil { - info = append(info, yaml.MapItem{Key: "contact", Value: m.Contact.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("contact")) + info.Content = append(info.Content, m.Contact.ToRawInfo()) } - // &{Name:contact Type:Contact StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.License != nil { - info = append(info, yaml.MapItem{Key: "license", Value: m.License.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("license")) + info.Content = append(info.Content, m.License.ToRawInfo()) } - // &{Name:license Type:License StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of ItemsItem suitable for JSON or YAML export. -func (m *ItemsItem) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *ItemsItem) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if len(m.Schema) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Schema { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "schema", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("schema")) + info.Content = append(info.Content, items) } - // &{Name:schema Type:Schema StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} return info } // ToRawInfo returns a description of JsonReference suitable for JSON or YAML export. -func (m *JsonReference) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *JsonReference) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("$ref")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.XRef)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } return info } // ToRawInfo returns a description of License suitable for JSON or YAML export. -func (m *License) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *License) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) if m.Url != "" { - info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("url")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Url)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of NamedAny suitable for JSON or YAML export. -func (m *NamedAny) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedAny) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedHeader suitable for JSON or YAML export. -func (m *NamedHeader) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedHeader) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:Header StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedParameter suitable for JSON or YAML export. -func (m *NamedParameter) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedParameter) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:Parameter StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedPathItem suitable for JSON or YAML export. -func (m *NamedPathItem) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedPathItem) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:PathItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedResponse suitable for JSON or YAML export. -func (m *NamedResponse) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedResponse) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:Response StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedResponseValue suitable for JSON or YAML export. -func (m *NamedResponseValue) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedResponseValue) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:ResponseValue StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedSchema suitable for JSON or YAML export. -func (m *NamedSchema) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedSchema) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedSecurityDefinitionsItem suitable for JSON or YAML export. -func (m *NamedSecurityDefinitionsItem) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedSecurityDefinitionsItem) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:SecurityDefinitionsItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NamedString suitable for JSON or YAML export. -func (m *NamedString) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedString) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } if m.Value != "" { - info = append(info, yaml.MapItem{Key: "value", Value: m.Value}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("value")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Value)) } return info } // ToRawInfo returns a description of NamedStringArray suitable for JSON or YAML export. -func (m *NamedStringArray) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *NamedStringArray) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } // &{Name:value Type:StringArray StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info } // ToRawInfo returns a description of NonBodyParameter suitable for JSON or YAML export. -func (m *NonBodyParameter) ToRawInfo() interface{} { +func (m *NonBodyParameter) ToRawInfo() *yaml.Node { // ONE OF WRAPPER // NonBodyParameter // {Name:headerParameterSubSchema Type:HeaderParameterSubSchema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -7884,126 +7742,143 @@ func (m *NonBodyParameter) ToRawInfo() interface{} { if v3 != nil { return v3.ToRawInfo() } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of Oauth2AccessCodeSecurity suitable for JSON or YAML export. -func (m *Oauth2AccessCodeSecurity) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Oauth2AccessCodeSecurity) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) // always include this required field. - info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("flow")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Flow)) if m.Scopes != nil { - info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("scopes")) + info.Content = append(info.Content, m.Scopes.ToRawInfo()) } - // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} // always include this required field. - info = append(info, yaml.MapItem{Key: "authorizationUrl", Value: m.AuthorizationUrl}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("authorizationUrl")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.AuthorizationUrl)) // always include this required field. - info = append(info, yaml.MapItem{Key: "tokenUrl", Value: m.TokenUrl}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("tokenUrl")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.TokenUrl)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Oauth2ApplicationSecurity suitable for JSON or YAML export. -func (m *Oauth2ApplicationSecurity) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Oauth2ApplicationSecurity) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) // always include this required field. - info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("flow")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Flow)) if m.Scopes != nil { - info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("scopes")) + info.Content = append(info.Content, m.Scopes.ToRawInfo()) } - // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} // always include this required field. - info = append(info, yaml.MapItem{Key: "tokenUrl", Value: m.TokenUrl}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("tokenUrl")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.TokenUrl)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Oauth2ImplicitSecurity suitable for JSON or YAML export. -func (m *Oauth2ImplicitSecurity) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Oauth2ImplicitSecurity) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) // always include this required field. - info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("flow")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Flow)) if m.Scopes != nil { - info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("scopes")) + info.Content = append(info.Content, m.Scopes.ToRawInfo()) } - // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} // always include this required field. - info = append(info, yaml.MapItem{Key: "authorizationUrl", Value: m.AuthorizationUrl}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("authorizationUrl")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.AuthorizationUrl)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Oauth2PasswordSecurity suitable for JSON or YAML export. -func (m *Oauth2PasswordSecurity) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Oauth2PasswordSecurity) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) // always include this required field. - info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("flow")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Flow)) if m.Scopes != nil { - info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("scopes")) + info.Content = append(info.Content, m.Scopes.ToRawInfo()) } - // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} // always include this required field. - info = append(info, yaml.MapItem{Key: "tokenUrl", Value: m.TokenUrl}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("tokenUrl")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.TokenUrl)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Oauth2Scopes suitable for JSON or YAML export. -func (m *Oauth2Scopes) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Oauth2Scopes) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } @@ -8012,69 +7887,77 @@ func (m *Oauth2Scopes) ToRawInfo() interface{} { } // ToRawInfo returns a description of Operation suitable for JSON or YAML export. -func (m *Operation) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Operation) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if len(m.Tags) != 0 { - info = append(info, yaml.MapItem{Key: "tags", Value: m.Tags}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("tags")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Tags)) } if m.Summary != "" { - info = append(info, yaml.MapItem{Key: "summary", Value: m.Summary}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("summary")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Summary)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("externalDocs")) + info.Content = append(info.Content, m.ExternalDocs.ToRawInfo()) } - // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.OperationId != "" { - info = append(info, yaml.MapItem{Key: "operationId", Value: m.OperationId}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("operationId")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.OperationId)) } if len(m.Produces) != 0 { - info = append(info, yaml.MapItem{Key: "produces", Value: m.Produces}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("produces")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Produces)) } if len(m.Consumes) != 0 { - info = append(info, yaml.MapItem{Key: "consumes", Value: m.Consumes}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("consumes")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Consumes)) } if len(m.Parameters) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Parameters { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "parameters", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("parameters")) + info.Content = append(info.Content, items) } - // &{Name:parameters Type:ParametersItem StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:The parameters needed to send a valid API call.} // always include this required field. - info = append(info, yaml.MapItem{Key: "responses", Value: m.Responses.ToRawInfo()}) - // &{Name:responses Type:Responses StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} + info.Content = append(info.Content, compiler.NewScalarNodeForString("responses")) + info.Content = append(info.Content, m.Responses.ToRawInfo()) if len(m.Schemes) != 0 { - info = append(info, yaml.MapItem{Key: "schemes", Value: m.Schemes}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("schemes")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Schemes)) } if m.Deprecated != false { - info = append(info, yaml.MapItem{Key: "deprecated", Value: m.Deprecated}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("deprecated")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Deprecated)) } if len(m.Security) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Security { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "security", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("security")) + info.Content = append(info.Content, items) } - // &{Name:security Type:SecurityRequirement StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Parameter suitable for JSON or YAML export. -func (m *Parameter) ToRawInfo() interface{} { +func (m *Parameter) ToRawInfo() *yaml.Node { // ONE OF WRAPPER // Parameter // {Name:bodyParameter Type:BodyParameter StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -8087,26 +7970,26 @@ func (m *Parameter) ToRawInfo() interface{} { if v1 != nil { return v1.ToRawInfo() } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of ParameterDefinitions suitable for JSON or YAML export. -func (m *ParameterDefinitions) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *ParameterDefinitions) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedParameter StringEnumValues:[] MapType:Parameter Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of ParametersItem suitable for JSON or YAML export. -func (m *ParametersItem) ToRawInfo() interface{} { +func (m *ParametersItem) ToRawInfo() *yaml.Node { // ONE OF WRAPPER // ParametersItem // {Name:parameter Type:Parameter StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -8119,390 +8002,443 @@ func (m *ParametersItem) ToRawInfo() interface{} { if v1 != nil { return v1.ToRawInfo() } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of PathItem suitable for JSON or YAML export. -func (m *PathItem) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *PathItem) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.XRef != "" { - info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("$ref")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.XRef)) } if m.Get != nil { - info = append(info, yaml.MapItem{Key: "get", Value: m.Get.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("get")) + info.Content = append(info.Content, m.Get.ToRawInfo()) } - // &{Name:get Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Put != nil { - info = append(info, yaml.MapItem{Key: "put", Value: m.Put.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("put")) + info.Content = append(info.Content, m.Put.ToRawInfo()) } - // &{Name:put Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Post != nil { - info = append(info, yaml.MapItem{Key: "post", Value: m.Post.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("post")) + info.Content = append(info.Content, m.Post.ToRawInfo()) } - // &{Name:post Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Delete != nil { - info = append(info, yaml.MapItem{Key: "delete", Value: m.Delete.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("delete")) + info.Content = append(info.Content, m.Delete.ToRawInfo()) } - // &{Name:delete Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Options != nil { - info = append(info, yaml.MapItem{Key: "options", Value: m.Options.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("options")) + info.Content = append(info.Content, m.Options.ToRawInfo()) } - // &{Name:options Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Head != nil { - info = append(info, yaml.MapItem{Key: "head", Value: m.Head.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("head")) + info.Content = append(info.Content, m.Head.ToRawInfo()) } - // &{Name:head Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Patch != nil { - info = append(info, yaml.MapItem{Key: "patch", Value: m.Patch.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("patch")) + info.Content = append(info.Content, m.Patch.ToRawInfo()) } - // &{Name:patch Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Parameters) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Parameters { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "parameters", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("parameters")) + info.Content = append(info.Content, items) } - // &{Name:parameters Type:ParametersItem StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:The parameters needed to send a valid API call.} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of PathParameterSubSchema suitable for JSON or YAML export. -func (m *PathParameterSubSchema) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *PathParameterSubSchema) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("required")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Required)) if m.In != "" { - info = append(info, yaml.MapItem{Key: "in", Value: m.In}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("in")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.In)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } if m.Type != "" { - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) } if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Items != nil { - info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("items")) + info.Content = append(info.Content, m.Items.ToRawInfo()) } - // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("collectionFormat")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.CollectionFormat)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Maximum)) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMaximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMaximum)) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Minimum)) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMinimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMinimum)) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxLength)) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinLength)) } if m.Pattern != "" { - info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("pattern")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Pattern)) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxItems)) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinItems)) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("uniqueItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.UniqueItems)) } if len(m.Enum) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Enum { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "enum", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("enum")) + info.Content = append(info.Content, items) } - // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("multipleOf")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.MultipleOf)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Paths suitable for JSON or YAML export. -func (m *Paths) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Paths) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} if m.Path != nil { for _, item := range m.Path { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:Path Type:NamedPathItem StringEnumValues:[] MapType:PathItem Repeated:true Pattern:^/ Implicit:true Description:} return info } // ToRawInfo returns a description of PrimitivesItems suitable for JSON or YAML export. -func (m *PrimitivesItems) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *PrimitivesItems) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Type != "" { - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) } if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Items != nil { - info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("items")) + info.Content = append(info.Content, m.Items.ToRawInfo()) } - // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("collectionFormat")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.CollectionFormat)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Maximum)) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMaximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMaximum)) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Minimum)) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMinimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMinimum)) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxLength)) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinLength)) } if m.Pattern != "" { - info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("pattern")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Pattern)) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxItems)) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinItems)) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("uniqueItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.UniqueItems)) } if len(m.Enum) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Enum { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "enum", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("enum")) + info.Content = append(info.Content, items) } - // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("multipleOf")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.MultipleOf)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Properties suitable for JSON or YAML export. -func (m *Properties) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Properties) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedSchema StringEnumValues:[] MapType:Schema Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of QueryParameterSubSchema suitable for JSON or YAML export. -func (m *QueryParameterSubSchema) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *QueryParameterSubSchema) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Required != false { - info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("required")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Required)) } if m.In != "" { - info = append(info, yaml.MapItem{Key: "in", Value: m.In}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("in")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.In)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } if m.AllowEmptyValue != false { - info = append(info, yaml.MapItem{Key: "allowEmptyValue", Value: m.AllowEmptyValue}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("allowEmptyValue")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.AllowEmptyValue)) } if m.Type != "" { - info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type)) } if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Items != nil { - info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("items")) + info.Content = append(info.Content, m.Items.ToRawInfo()) } - // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("collectionFormat")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.CollectionFormat)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Maximum)) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMaximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMaximum)) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Minimum)) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMinimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMinimum)) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxLength)) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinLength)) } if m.Pattern != "" { - info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("pattern")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Pattern)) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxItems)) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinItems)) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("uniqueItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.UniqueItems)) } if len(m.Enum) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Enum { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "enum", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("enum")) + info.Content = append(info.Content, items) } - // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("multipleOf")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.MultipleOf)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Response suitable for JSON or YAML export. -func (m *Response) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Response) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) if m.Schema != nil { - info = append(info, yaml.MapItem{Key: "schema", Value: m.Schema.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("schema")) + info.Content = append(info.Content, m.Schema.ToRawInfo()) } - // &{Name:schema Type:SchemaItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Headers != nil { - info = append(info, yaml.MapItem{Key: "headers", Value: m.Headers.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("headers")) + info.Content = append(info.Content, m.Headers.ToRawInfo()) } - // &{Name:headers Type:Headers StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Examples != nil { - info = append(info, yaml.MapItem{Key: "examples", Value: m.Examples.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("examples")) + info.Content = append(info.Content, m.Examples.ToRawInfo()) } - // &{Name:examples Type:Examples StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of ResponseDefinitions suitable for JSON or YAML export. -func (m *ResponseDefinitions) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *ResponseDefinitions) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedResponse StringEnumValues:[] MapType:Response Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of ResponseValue suitable for JSON or YAML export. -func (m *ResponseValue) ToRawInfo() interface{} { +func (m *ResponseValue) ToRawInfo() *yaml.Node { // ONE OF WRAPPER // ResponseValue // {Name:response Type:Response StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -8515,163 +8451,187 @@ func (m *ResponseValue) ToRawInfo() interface{} { if v1 != nil { return v1.ToRawInfo() } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of Responses suitable for JSON or YAML export. -func (m *Responses) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Responses) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.ResponseCode != nil { for _, item := range m.ResponseCode { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:ResponseCode Type:NamedResponseValue StringEnumValues:[] MapType:ResponseValue Repeated:true Pattern:^([0-9]{3})$|^(default)$ Implicit:true Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of Schema suitable for JSON or YAML export. -func (m *Schema) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Schema) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.XRef != "" { - info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("$ref")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.XRef)) } if m.Format != "" { - info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("format")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Format)) } if m.Title != "" { - info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("title")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Title)) } if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.Default != nil { - info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("default")) + info.Content = append(info.Content, m.Default.ToRawInfo()) } - // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("multipleOf")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.MultipleOf)) } if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Maximum)) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMaximum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMaximum)) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForFloat(m.Minimum)) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("exclusiveMinimum")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ExclusiveMinimum)) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxLength)) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minLength")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinLength)) } if m.Pattern != "" { - info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("pattern")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Pattern)) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxItems)) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinItems)) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("uniqueItems")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.UniqueItems)) } if m.MaxProperties != 0 { - info = append(info, yaml.MapItem{Key: "maxProperties", Value: m.MaxProperties}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("maxProperties")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MaxProperties)) } if m.MinProperties != 0 { - info = append(info, yaml.MapItem{Key: "minProperties", Value: m.MinProperties}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("minProperties")) + info.Content = append(info.Content, compiler.NewScalarNodeForInt(m.MinProperties)) } if len(m.Required) != 0 { - info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("required")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Required)) } if len(m.Enum) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Enum { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "enum", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("enum")) + info.Content = append(info.Content, items) } - // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.AdditionalProperties != nil { - info = append(info, yaml.MapItem{Key: "additionalProperties", Value: m.AdditionalProperties.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("additionalProperties")) + info.Content = append(info.Content, m.AdditionalProperties.ToRawInfo()) } - // &{Name:additionalProperties Type:AdditionalPropertiesItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Type != nil { if len(m.Type.Value) == 1 { - info = append(info, yaml.MapItem{Key: "type", Value: m.Type.Value[0]}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Type.Value[0])) } else { - info = append(info, yaml.MapItem{Key: "type", Value: m.Type.Value}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("type")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Type.Value)) } } - // &{Name:type Type:TypeItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Items != nil { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.Items.Schema { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) + } + if len(items.Content) == 1 { + items = items.Content[0] } - info = append(info, yaml.MapItem{Key: "items", Value: items[0]}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("items")) + info.Content = append(info.Content, items) } - // &{Name:items Type:ItemsItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.AllOf) != 0 { - items := make([]interface{}, 0) + items := compiler.NewSequenceNode() for _, item := range m.AllOf { - items = append(items, item.ToRawInfo()) + items.Content = append(items.Content, item.ToRawInfo()) } - info = append(info, yaml.MapItem{Key: "allOf", Value: items}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("allOf")) + info.Content = append(info.Content, items) } - // &{Name:allOf Type:Schema StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.Properties != nil { - info = append(info, yaml.MapItem{Key: "properties", Value: m.Properties.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("properties")) + info.Content = append(info.Content, m.Properties.ToRawInfo()) } - // &{Name:properties Type:Properties StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Discriminator != "" { - info = append(info, yaml.MapItem{Key: "discriminator", Value: m.Discriminator}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("discriminator")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Discriminator)) } if m.ReadOnly != false { - info = append(info, yaml.MapItem{Key: "readOnly", Value: m.ReadOnly}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("readOnly")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.ReadOnly)) } if m.Xml != nil { - info = append(info, yaml.MapItem{Key: "xml", Value: m.Xml.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("xml")) + info.Content = append(info.Content, m.Xml.ToRawInfo()) } - // &{Name:xml Type:Xml StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("externalDocs")) + info.Content = append(info.Content, m.ExternalDocs.ToRawInfo()) } - // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("example")) + info.Content = append(info.Content, m.Example.ToRawInfo()) } - // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of SchemaItem suitable for JSON or YAML export. -func (m *SchemaItem) ToRawInfo() interface{} { +func (m *SchemaItem) ToRawInfo() *yaml.Node { // ONE OF WRAPPER // SchemaItem // {Name:schema Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -8684,26 +8644,26 @@ func (m *SchemaItem) ToRawInfo() interface{} { if v1 != nil { return v1.ToRawInfo() } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of SecurityDefinitions suitable for JSON or YAML export. -func (m *SecurityDefinitions) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *SecurityDefinitions) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedSecurityDefinitionsItem StringEnumValues:[] MapType:SecurityDefinitionsItem Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of SecurityDefinitionsItem suitable for JSON or YAML export. -func (m *SecurityDefinitionsItem) ToRawInfo() interface{} { +func (m *SecurityDefinitionsItem) ToRawInfo() *yaml.Node { // ONE OF WRAPPER // SecurityDefinitionsItem // {Name:basicAuthenticationSecurity Type:BasicAuthenticationSecurity StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -8736,107 +8696,115 @@ func (m *SecurityDefinitionsItem) ToRawInfo() interface{} { if v5 != nil { return v5.ToRawInfo() } - return nil + return compiler.NewNullNode() } // ToRawInfo returns a description of SecurityRequirement suitable for JSON or YAML export. -func (m *SecurityRequirement) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *SecurityRequirement) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedStringArray StringEnumValues:[] MapType:StringArray Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of StringArray suitable for JSON or YAML export. -func (m *StringArray) ToRawInfo() interface{} { - return m.Value +func (m *StringArray) ToRawInfo() *yaml.Node { + return compiler.NewSequenceNodeForStringArray(m.Value) } // ToRawInfo returns a description of Tag suitable for JSON or YAML export. -func (m *Tag) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Tag) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } // always include this required field. - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) if m.Description != "" { - info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("description")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Description)) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("externalDocs")) + info.Content = append(info.Content, m.ExternalDocs.ToRawInfo()) } - // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } // ToRawInfo returns a description of TypeItem suitable for JSON or YAML export. -func (m *TypeItem) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *TypeItem) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if len(m.Value) != 0 { - info = append(info, yaml.MapItem{Key: "value", Value: m.Value}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("value")) + info.Content = append(info.Content, compiler.NewSequenceNodeForStringArray(m.Value)) } return info } // ToRawInfo returns a description of VendorExtension suitable for JSON or YAML export. -func (m *VendorExtension) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *VendorExtension) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:true Description:} return info } // ToRawInfo returns a description of Xml suitable for JSON or YAML export. -func (m *Xml) ToRawInfo() interface{} { - info := yaml.MapSlice{} +func (m *Xml) ToRawInfo() *yaml.Node { + info := compiler.NewMappingNode() if m == nil { return info } if m.Name != "" { - info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("name")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Name)) } if m.Namespace != "" { - info = append(info, yaml.MapItem{Key: "namespace", Value: m.Namespace}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("namespace")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Namespace)) } if m.Prefix != "" { - info = append(info, yaml.MapItem{Key: "prefix", Value: m.Prefix}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("prefix")) + info.Content = append(info.Content, compiler.NewScalarNodeForString(m.Prefix)) } if m.Attribute != false { - info = append(info, yaml.MapItem{Key: "attribute", Value: m.Attribute}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("attribute")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Attribute)) } if m.Wrapped != false { - info = append(info, yaml.MapItem{Key: "wrapped", Value: m.Wrapped}) + info.Content = append(info.Content, compiler.NewScalarNodeForString("wrapped")) + info.Content = append(info.Content, compiler.NewScalarNodeForBool(m.Wrapped)) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) + info.Content = append(info.Content, compiler.NewScalarNodeForString(item.Name)) + info.Content = append(info.Content, item.Value.ToRawInfo()) } } - // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} return info } diff --git a/vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.pb.go b/vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.pb.go new file mode 100644 index 00000000..4320dc37 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.pb.go @@ -0,0 +1,7347 @@ +// Copyright 2020 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// THIS FILE IS AUTOMATICALLY GENERATED. + +// Code generated by protoc-gen-go. DO NOT EDIT. +// versions: +// protoc-gen-go v1.24.0 +// protoc v3.12.0 +// source: openapiv2/OpenAPIv2.proto + +package openapi_v2 + +import ( + proto "github.com/golang/protobuf/proto" + any "github.com/golang/protobuf/ptypes/any" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + reflect "reflect" + sync "sync" +) + +const ( + // Verify that this generated code is sufficiently up-to-date. + _ = protoimpl.EnforceVersion(20 - protoimpl.MinVersion) + // Verify that runtime/protoimpl is sufficiently up-to-date. + _ = protoimpl.EnforceVersion(protoimpl.MaxVersion - 20) +) + +// This is a compile-time assertion that a sufficiently up-to-date version +// of the legacy proto package is being used. +const _ = proto.ProtoPackageIsVersion4 + +type AdditionalPropertiesItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Types that are assignable to Oneof: + // *AdditionalPropertiesItem_Schema + // *AdditionalPropertiesItem_Boolean + Oneof isAdditionalPropertiesItem_Oneof `protobuf_oneof:"oneof"` +} + +func (x *AdditionalPropertiesItem) Reset() { + *x = AdditionalPropertiesItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[0] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *AdditionalPropertiesItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*AdditionalPropertiesItem) ProtoMessage() {} + +func (x *AdditionalPropertiesItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[0] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use AdditionalPropertiesItem.ProtoReflect.Descriptor instead. +func (*AdditionalPropertiesItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{0} +} + +func (m *AdditionalPropertiesItem) GetOneof() isAdditionalPropertiesItem_Oneof { + if m != nil { + return m.Oneof + } + return nil +} + +func (x *AdditionalPropertiesItem) GetSchema() *Schema { + if x, ok := x.GetOneof().(*AdditionalPropertiesItem_Schema); ok { + return x.Schema + } + return nil +} + +func (x *AdditionalPropertiesItem) GetBoolean() bool { + if x, ok := x.GetOneof().(*AdditionalPropertiesItem_Boolean); ok { + return x.Boolean + } + return false +} + +type isAdditionalPropertiesItem_Oneof interface { + isAdditionalPropertiesItem_Oneof() +} + +type AdditionalPropertiesItem_Schema struct { + Schema *Schema `protobuf:"bytes,1,opt,name=schema,proto3,oneof"` +} + +type AdditionalPropertiesItem_Boolean struct { + Boolean bool `protobuf:"varint,2,opt,name=boolean,proto3,oneof"` +} + +func (*AdditionalPropertiesItem_Schema) isAdditionalPropertiesItem_Oneof() {} + +func (*AdditionalPropertiesItem_Boolean) isAdditionalPropertiesItem_Oneof() {} + +type Any struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Value *any.Any `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` + Yaml string `protobuf:"bytes,2,opt,name=yaml,proto3" json:"yaml,omitempty"` +} + +func (x *Any) Reset() { + *x = Any{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[1] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Any) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Any) ProtoMessage() {} + +func (x *Any) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[1] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Any.ProtoReflect.Descriptor instead. +func (*Any) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{1} +} + +func (x *Any) GetValue() *any.Any { + if x != nil { + return x.Value + } + return nil +} + +func (x *Any) GetYaml() string { + if x != nil { + return x.Yaml + } + return "" +} + +type ApiKeySecurity struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` + In string `protobuf:"bytes,3,opt,name=in,proto3" json:"in,omitempty"` + Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,5,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *ApiKeySecurity) Reset() { + *x = ApiKeySecurity{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[2] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *ApiKeySecurity) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*ApiKeySecurity) ProtoMessage() {} + +func (x *ApiKeySecurity) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[2] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use ApiKeySecurity.ProtoReflect.Descriptor instead. +func (*ApiKeySecurity) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{2} +} + +func (x *ApiKeySecurity) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *ApiKeySecurity) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *ApiKeySecurity) GetIn() string { + if x != nil { + return x.In + } + return "" +} + +func (x *ApiKeySecurity) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *ApiKeySecurity) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type BasicAuthenticationSecurity struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,3,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *BasicAuthenticationSecurity) Reset() { + *x = BasicAuthenticationSecurity{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[3] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *BasicAuthenticationSecurity) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*BasicAuthenticationSecurity) ProtoMessage() {} + +func (x *BasicAuthenticationSecurity) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[3] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use BasicAuthenticationSecurity.ProtoReflect.Descriptor instead. +func (*BasicAuthenticationSecurity) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{3} +} + +func (x *BasicAuthenticationSecurity) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *BasicAuthenticationSecurity) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *BasicAuthenticationSecurity) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type BodyParameter struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` + // The name of the parameter. + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` + // Determines the location of the parameter. + In string `protobuf:"bytes,3,opt,name=in,proto3" json:"in,omitempty"` + // Determines whether or not this parameter is required or optional. + Required bool `protobuf:"varint,4,opt,name=required,proto3" json:"required,omitempty"` + Schema *Schema `protobuf:"bytes,5,opt,name=schema,proto3" json:"schema,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *BodyParameter) Reset() { + *x = BodyParameter{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[4] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *BodyParameter) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*BodyParameter) ProtoMessage() {} + +func (x *BodyParameter) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[4] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use BodyParameter.ProtoReflect.Descriptor instead. +func (*BodyParameter) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{4} +} + +func (x *BodyParameter) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *BodyParameter) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *BodyParameter) GetIn() string { + if x != nil { + return x.In + } + return "" +} + +func (x *BodyParameter) GetRequired() bool { + if x != nil { + return x.Required + } + return false +} + +func (x *BodyParameter) GetSchema() *Schema { + if x != nil { + return x.Schema + } + return nil +} + +func (x *BodyParameter) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +// Contact information for the owners of the API. +type Contact struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // The identifying name of the contact person/organization. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The URL pointing to the contact information. + Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` + // The email address of the contact person/organization. + Email string `protobuf:"bytes,3,opt,name=email,proto3" json:"email,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,4,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Contact) Reset() { + *x = Contact{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[5] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Contact) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Contact) ProtoMessage() {} + +func (x *Contact) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[5] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Contact.ProtoReflect.Descriptor instead. +func (*Contact) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{5} +} + +func (x *Contact) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *Contact) GetUrl() string { + if x != nil { + return x.Url + } + return "" +} + +func (x *Contact) GetEmail() string { + if x != nil { + return x.Email + } + return "" +} + +func (x *Contact) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Default struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedAny `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *Default) Reset() { + *x = Default{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[6] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Default) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Default) ProtoMessage() {} + +func (x *Default) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[6] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Default.ProtoReflect.Descriptor instead. +func (*Default) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{6} +} + +func (x *Default) GetAdditionalProperties() []*NamedAny { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +// One or more JSON objects describing the schemas being consumed and produced by the API. +type Definitions struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedSchema `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *Definitions) Reset() { + *x = Definitions{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[7] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Definitions) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Definitions) ProtoMessage() {} + +func (x *Definitions) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[7] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Definitions.ProtoReflect.Descriptor instead. +func (*Definitions) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{7} +} + +func (x *Definitions) GetAdditionalProperties() []*NamedSchema { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type Document struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // The Swagger version of this document. + Swagger string `protobuf:"bytes,1,opt,name=swagger,proto3" json:"swagger,omitempty"` + Info *Info `protobuf:"bytes,2,opt,name=info,proto3" json:"info,omitempty"` + // The host (name or ip) of the API. Example: 'swagger.io' + Host string `protobuf:"bytes,3,opt,name=host,proto3" json:"host,omitempty"` + // The base path to the API. Example: '/api'. + BasePath string `protobuf:"bytes,4,opt,name=base_path,json=basePath,proto3" json:"base_path,omitempty"` + // The transfer protocol of the API. + Schemes []string `protobuf:"bytes,5,rep,name=schemes,proto3" json:"schemes,omitempty"` + // A list of MIME types accepted by the API. + Consumes []string `protobuf:"bytes,6,rep,name=consumes,proto3" json:"consumes,omitempty"` + // A list of MIME types the API can produce. + Produces []string `protobuf:"bytes,7,rep,name=produces,proto3" json:"produces,omitempty"` + Paths *Paths `protobuf:"bytes,8,opt,name=paths,proto3" json:"paths,omitempty"` + Definitions *Definitions `protobuf:"bytes,9,opt,name=definitions,proto3" json:"definitions,omitempty"` + Parameters *ParameterDefinitions `protobuf:"bytes,10,opt,name=parameters,proto3" json:"parameters,omitempty"` + Responses *ResponseDefinitions `protobuf:"bytes,11,opt,name=responses,proto3" json:"responses,omitempty"` + Security []*SecurityRequirement `protobuf:"bytes,12,rep,name=security,proto3" json:"security,omitempty"` + SecurityDefinitions *SecurityDefinitions `protobuf:"bytes,13,opt,name=security_definitions,json=securityDefinitions,proto3" json:"security_definitions,omitempty"` + Tags []*Tag `protobuf:"bytes,14,rep,name=tags,proto3" json:"tags,omitempty"` + ExternalDocs *ExternalDocs `protobuf:"bytes,15,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,16,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Document) Reset() { + *x = Document{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[8] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Document) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Document) ProtoMessage() {} + +func (x *Document) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[8] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Document.ProtoReflect.Descriptor instead. +func (*Document) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{8} +} + +func (x *Document) GetSwagger() string { + if x != nil { + return x.Swagger + } + return "" +} + +func (x *Document) GetInfo() *Info { + if x != nil { + return x.Info + } + return nil +} + +func (x *Document) GetHost() string { + if x != nil { + return x.Host + } + return "" +} + +func (x *Document) GetBasePath() string { + if x != nil { + return x.BasePath + } + return "" +} + +func (x *Document) GetSchemes() []string { + if x != nil { + return x.Schemes + } + return nil +} + +func (x *Document) GetConsumes() []string { + if x != nil { + return x.Consumes + } + return nil +} + +func (x *Document) GetProduces() []string { + if x != nil { + return x.Produces + } + return nil +} + +func (x *Document) GetPaths() *Paths { + if x != nil { + return x.Paths + } + return nil +} + +func (x *Document) GetDefinitions() *Definitions { + if x != nil { + return x.Definitions + } + return nil +} + +func (x *Document) GetParameters() *ParameterDefinitions { + if x != nil { + return x.Parameters + } + return nil +} + +func (x *Document) GetResponses() *ResponseDefinitions { + if x != nil { + return x.Responses + } + return nil +} + +func (x *Document) GetSecurity() []*SecurityRequirement { + if x != nil { + return x.Security + } + return nil +} + +func (x *Document) GetSecurityDefinitions() *SecurityDefinitions { + if x != nil { + return x.SecurityDefinitions + } + return nil +} + +func (x *Document) GetTags() []*Tag { + if x != nil { + return x.Tags + } + return nil +} + +func (x *Document) GetExternalDocs() *ExternalDocs { + if x != nil { + return x.ExternalDocs + } + return nil +} + +func (x *Document) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Examples struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedAny `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *Examples) Reset() { + *x = Examples{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[9] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Examples) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Examples) ProtoMessage() {} + +func (x *Examples) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[9] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Examples.ProtoReflect.Descriptor instead. +func (*Examples) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{9} +} + +func (x *Examples) GetAdditionalProperties() []*NamedAny { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +// information about external documentation +type ExternalDocs struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` + Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,3,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *ExternalDocs) Reset() { + *x = ExternalDocs{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[10] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *ExternalDocs) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*ExternalDocs) ProtoMessage() {} + +func (x *ExternalDocs) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[10] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use ExternalDocs.ProtoReflect.Descriptor instead. +func (*ExternalDocs) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{10} +} + +func (x *ExternalDocs) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *ExternalDocs) GetUrl() string { + if x != nil { + return x.Url + } + return "" +} + +func (x *ExternalDocs) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +// A deterministic version of a JSON Schema object. +type FileSchema struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Format string `protobuf:"bytes,1,opt,name=format,proto3" json:"format,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` + Default *Any `protobuf:"bytes,4,opt,name=default,proto3" json:"default,omitempty"` + Required []string `protobuf:"bytes,5,rep,name=required,proto3" json:"required,omitempty"` + Type string `protobuf:"bytes,6,opt,name=type,proto3" json:"type,omitempty"` + ReadOnly bool `protobuf:"varint,7,opt,name=read_only,json=readOnly,proto3" json:"read_only,omitempty"` + ExternalDocs *ExternalDocs `protobuf:"bytes,8,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` + Example *Any `protobuf:"bytes,9,opt,name=example,proto3" json:"example,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,10,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *FileSchema) Reset() { + *x = FileSchema{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[11] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *FileSchema) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*FileSchema) ProtoMessage() {} + +func (x *FileSchema) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[11] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use FileSchema.ProtoReflect.Descriptor instead. +func (*FileSchema) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{11} +} + +func (x *FileSchema) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *FileSchema) GetTitle() string { + if x != nil { + return x.Title + } + return "" +} + +func (x *FileSchema) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *FileSchema) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *FileSchema) GetRequired() []string { + if x != nil { + return x.Required + } + return nil +} + +func (x *FileSchema) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *FileSchema) GetReadOnly() bool { + if x != nil { + return x.ReadOnly + } + return false +} + +func (x *FileSchema) GetExternalDocs() *ExternalDocs { + if x != nil { + return x.ExternalDocs + } + return nil +} + +func (x *FileSchema) GetExample() *Any { + if x != nil { + return x.Example + } + return nil +} + +func (x *FileSchema) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type FormDataParameterSubSchema struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Determines whether or not this parameter is required or optional. + Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` + // Determines the location of the parameter. + In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` + // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` + // The name of the parameter. + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` + // allows sending a parameter by name only or with an empty value. + AllowEmptyValue bool `protobuf:"varint,5,opt,name=allow_empty_value,json=allowEmptyValue,proto3" json:"allow_empty_value,omitempty"` + Type string `protobuf:"bytes,6,opt,name=type,proto3" json:"type,omitempty"` + Format string `protobuf:"bytes,7,opt,name=format,proto3" json:"format,omitempty"` + Items *PrimitivesItems `protobuf:"bytes,8,opt,name=items,proto3" json:"items,omitempty"` + CollectionFormat string `protobuf:"bytes,9,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` + Default *Any `protobuf:"bytes,10,opt,name=default,proto3" json:"default,omitempty"` + Maximum float64 `protobuf:"fixed64,11,opt,name=maximum,proto3" json:"maximum,omitempty"` + ExclusiveMaximum bool `protobuf:"varint,12,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` + Minimum float64 `protobuf:"fixed64,13,opt,name=minimum,proto3" json:"minimum,omitempty"` + ExclusiveMinimum bool `protobuf:"varint,14,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` + MaxLength int64 `protobuf:"varint,15,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` + MinLength int64 `protobuf:"varint,16,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` + Pattern string `protobuf:"bytes,17,opt,name=pattern,proto3" json:"pattern,omitempty"` + MaxItems int64 `protobuf:"varint,18,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` + MinItems int64 `protobuf:"varint,19,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` + UniqueItems bool `protobuf:"varint,20,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` + Enum []*Any `protobuf:"bytes,21,rep,name=enum,proto3" json:"enum,omitempty"` + MultipleOf float64 `protobuf:"fixed64,22,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,23,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *FormDataParameterSubSchema) Reset() { + *x = FormDataParameterSubSchema{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[12] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *FormDataParameterSubSchema) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*FormDataParameterSubSchema) ProtoMessage() {} + +func (x *FormDataParameterSubSchema) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[12] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use FormDataParameterSubSchema.ProtoReflect.Descriptor instead. +func (*FormDataParameterSubSchema) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{12} +} + +func (x *FormDataParameterSubSchema) GetRequired() bool { + if x != nil { + return x.Required + } + return false +} + +func (x *FormDataParameterSubSchema) GetIn() string { + if x != nil { + return x.In + } + return "" +} + +func (x *FormDataParameterSubSchema) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *FormDataParameterSubSchema) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *FormDataParameterSubSchema) GetAllowEmptyValue() bool { + if x != nil { + return x.AllowEmptyValue + } + return false +} + +func (x *FormDataParameterSubSchema) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *FormDataParameterSubSchema) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *FormDataParameterSubSchema) GetItems() *PrimitivesItems { + if x != nil { + return x.Items + } + return nil +} + +func (x *FormDataParameterSubSchema) GetCollectionFormat() string { + if x != nil { + return x.CollectionFormat + } + return "" +} + +func (x *FormDataParameterSubSchema) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *FormDataParameterSubSchema) GetMaximum() float64 { + if x != nil { + return x.Maximum + } + return 0 +} + +func (x *FormDataParameterSubSchema) GetExclusiveMaximum() bool { + if x != nil { + return x.ExclusiveMaximum + } + return false +} + +func (x *FormDataParameterSubSchema) GetMinimum() float64 { + if x != nil { + return x.Minimum + } + return 0 +} + +func (x *FormDataParameterSubSchema) GetExclusiveMinimum() bool { + if x != nil { + return x.ExclusiveMinimum + } + return false +} + +func (x *FormDataParameterSubSchema) GetMaxLength() int64 { + if x != nil { + return x.MaxLength + } + return 0 +} + +func (x *FormDataParameterSubSchema) GetMinLength() int64 { + if x != nil { + return x.MinLength + } + return 0 +} + +func (x *FormDataParameterSubSchema) GetPattern() string { + if x != nil { + return x.Pattern + } + return "" +} + +func (x *FormDataParameterSubSchema) GetMaxItems() int64 { + if x != nil { + return x.MaxItems + } + return 0 +} + +func (x *FormDataParameterSubSchema) GetMinItems() int64 { + if x != nil { + return x.MinItems + } + return 0 +} + +func (x *FormDataParameterSubSchema) GetUniqueItems() bool { + if x != nil { + return x.UniqueItems + } + return false +} + +func (x *FormDataParameterSubSchema) GetEnum() []*Any { + if x != nil { + return x.Enum + } + return nil +} + +func (x *FormDataParameterSubSchema) GetMultipleOf() float64 { + if x != nil { + return x.MultipleOf + } + return 0 +} + +func (x *FormDataParameterSubSchema) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Header struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Format string `protobuf:"bytes,2,opt,name=format,proto3" json:"format,omitempty"` + Items *PrimitivesItems `protobuf:"bytes,3,opt,name=items,proto3" json:"items,omitempty"` + CollectionFormat string `protobuf:"bytes,4,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` + Default *Any `protobuf:"bytes,5,opt,name=default,proto3" json:"default,omitempty"` + Maximum float64 `protobuf:"fixed64,6,opt,name=maximum,proto3" json:"maximum,omitempty"` + ExclusiveMaximum bool `protobuf:"varint,7,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` + Minimum float64 `protobuf:"fixed64,8,opt,name=minimum,proto3" json:"minimum,omitempty"` + ExclusiveMinimum bool `protobuf:"varint,9,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` + MaxLength int64 `protobuf:"varint,10,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` + MinLength int64 `protobuf:"varint,11,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` + Pattern string `protobuf:"bytes,12,opt,name=pattern,proto3" json:"pattern,omitempty"` + MaxItems int64 `protobuf:"varint,13,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` + MinItems int64 `protobuf:"varint,14,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` + UniqueItems bool `protobuf:"varint,15,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` + Enum []*Any `protobuf:"bytes,16,rep,name=enum,proto3" json:"enum,omitempty"` + MultipleOf float64 `protobuf:"fixed64,17,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` + Description string `protobuf:"bytes,18,opt,name=description,proto3" json:"description,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,19,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Header) Reset() { + *x = Header{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[13] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Header) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Header) ProtoMessage() {} + +func (x *Header) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[13] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Header.ProtoReflect.Descriptor instead. +func (*Header) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{13} +} + +func (x *Header) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *Header) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *Header) GetItems() *PrimitivesItems { + if x != nil { + return x.Items + } + return nil +} + +func (x *Header) GetCollectionFormat() string { + if x != nil { + return x.CollectionFormat + } + return "" +} + +func (x *Header) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *Header) GetMaximum() float64 { + if x != nil { + return x.Maximum + } + return 0 +} + +func (x *Header) GetExclusiveMaximum() bool { + if x != nil { + return x.ExclusiveMaximum + } + return false +} + +func (x *Header) GetMinimum() float64 { + if x != nil { + return x.Minimum + } + return 0 +} + +func (x *Header) GetExclusiveMinimum() bool { + if x != nil { + return x.ExclusiveMinimum + } + return false +} + +func (x *Header) GetMaxLength() int64 { + if x != nil { + return x.MaxLength + } + return 0 +} + +func (x *Header) GetMinLength() int64 { + if x != nil { + return x.MinLength + } + return 0 +} + +func (x *Header) GetPattern() string { + if x != nil { + return x.Pattern + } + return "" +} + +func (x *Header) GetMaxItems() int64 { + if x != nil { + return x.MaxItems + } + return 0 +} + +func (x *Header) GetMinItems() int64 { + if x != nil { + return x.MinItems + } + return 0 +} + +func (x *Header) GetUniqueItems() bool { + if x != nil { + return x.UniqueItems + } + return false +} + +func (x *Header) GetEnum() []*Any { + if x != nil { + return x.Enum + } + return nil +} + +func (x *Header) GetMultipleOf() float64 { + if x != nil { + return x.MultipleOf + } + return 0 +} + +func (x *Header) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Header) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type HeaderParameterSubSchema struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Determines whether or not this parameter is required or optional. + Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` + // Determines the location of the parameter. + In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` + // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` + // The name of the parameter. + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` + Type string `protobuf:"bytes,5,opt,name=type,proto3" json:"type,omitempty"` + Format string `protobuf:"bytes,6,opt,name=format,proto3" json:"format,omitempty"` + Items *PrimitivesItems `protobuf:"bytes,7,opt,name=items,proto3" json:"items,omitempty"` + CollectionFormat string `protobuf:"bytes,8,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` + Default *Any `protobuf:"bytes,9,opt,name=default,proto3" json:"default,omitempty"` + Maximum float64 `protobuf:"fixed64,10,opt,name=maximum,proto3" json:"maximum,omitempty"` + ExclusiveMaximum bool `protobuf:"varint,11,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` + Minimum float64 `protobuf:"fixed64,12,opt,name=minimum,proto3" json:"minimum,omitempty"` + ExclusiveMinimum bool `protobuf:"varint,13,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` + MaxLength int64 `protobuf:"varint,14,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` + MinLength int64 `protobuf:"varint,15,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` + Pattern string `protobuf:"bytes,16,opt,name=pattern,proto3" json:"pattern,omitempty"` + MaxItems int64 `protobuf:"varint,17,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` + MinItems int64 `protobuf:"varint,18,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` + UniqueItems bool `protobuf:"varint,19,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` + Enum []*Any `protobuf:"bytes,20,rep,name=enum,proto3" json:"enum,omitempty"` + MultipleOf float64 `protobuf:"fixed64,21,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,22,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *HeaderParameterSubSchema) Reset() { + *x = HeaderParameterSubSchema{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[14] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *HeaderParameterSubSchema) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*HeaderParameterSubSchema) ProtoMessage() {} + +func (x *HeaderParameterSubSchema) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[14] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use HeaderParameterSubSchema.ProtoReflect.Descriptor instead. +func (*HeaderParameterSubSchema) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{14} +} + +func (x *HeaderParameterSubSchema) GetRequired() bool { + if x != nil { + return x.Required + } + return false +} + +func (x *HeaderParameterSubSchema) GetIn() string { + if x != nil { + return x.In + } + return "" +} + +func (x *HeaderParameterSubSchema) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *HeaderParameterSubSchema) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *HeaderParameterSubSchema) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *HeaderParameterSubSchema) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *HeaderParameterSubSchema) GetItems() *PrimitivesItems { + if x != nil { + return x.Items + } + return nil +} + +func (x *HeaderParameterSubSchema) GetCollectionFormat() string { + if x != nil { + return x.CollectionFormat + } + return "" +} + +func (x *HeaderParameterSubSchema) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *HeaderParameterSubSchema) GetMaximum() float64 { + if x != nil { + return x.Maximum + } + return 0 +} + +func (x *HeaderParameterSubSchema) GetExclusiveMaximum() bool { + if x != nil { + return x.ExclusiveMaximum + } + return false +} + +func (x *HeaderParameterSubSchema) GetMinimum() float64 { + if x != nil { + return x.Minimum + } + return 0 +} + +func (x *HeaderParameterSubSchema) GetExclusiveMinimum() bool { + if x != nil { + return x.ExclusiveMinimum + } + return false +} + +func (x *HeaderParameterSubSchema) GetMaxLength() int64 { + if x != nil { + return x.MaxLength + } + return 0 +} + +func (x *HeaderParameterSubSchema) GetMinLength() int64 { + if x != nil { + return x.MinLength + } + return 0 +} + +func (x *HeaderParameterSubSchema) GetPattern() string { + if x != nil { + return x.Pattern + } + return "" +} + +func (x *HeaderParameterSubSchema) GetMaxItems() int64 { + if x != nil { + return x.MaxItems + } + return 0 +} + +func (x *HeaderParameterSubSchema) GetMinItems() int64 { + if x != nil { + return x.MinItems + } + return 0 +} + +func (x *HeaderParameterSubSchema) GetUniqueItems() bool { + if x != nil { + return x.UniqueItems + } + return false +} + +func (x *HeaderParameterSubSchema) GetEnum() []*Any { + if x != nil { + return x.Enum + } + return nil +} + +func (x *HeaderParameterSubSchema) GetMultipleOf() float64 { + if x != nil { + return x.MultipleOf + } + return 0 +} + +func (x *HeaderParameterSubSchema) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Headers struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedHeader `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *Headers) Reset() { + *x = Headers{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[15] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Headers) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Headers) ProtoMessage() {} + +func (x *Headers) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[15] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Headers.ProtoReflect.Descriptor instead. +func (*Headers) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{15} +} + +func (x *Headers) GetAdditionalProperties() []*NamedHeader { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +// General information about the API. +type Info struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // A unique and precise title of the API. + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` + // A semantic version number of the API. + Version string `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` + // A longer description of the API. Should be different from the title. GitHub Flavored Markdown is allowed. + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` + // The terms of service for the API. + TermsOfService string `protobuf:"bytes,4,opt,name=terms_of_service,json=termsOfService,proto3" json:"terms_of_service,omitempty"` + Contact *Contact `protobuf:"bytes,5,opt,name=contact,proto3" json:"contact,omitempty"` + License *License `protobuf:"bytes,6,opt,name=license,proto3" json:"license,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,7,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Info) Reset() { + *x = Info{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[16] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Info) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Info) ProtoMessage() {} + +func (x *Info) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[16] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Info.ProtoReflect.Descriptor instead. +func (*Info) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{16} +} + +func (x *Info) GetTitle() string { + if x != nil { + return x.Title + } + return "" +} + +func (x *Info) GetVersion() string { + if x != nil { + return x.Version + } + return "" +} + +func (x *Info) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Info) GetTermsOfService() string { + if x != nil { + return x.TermsOfService + } + return "" +} + +func (x *Info) GetContact() *Contact { + if x != nil { + return x.Contact + } + return nil +} + +func (x *Info) GetLicense() *License { + if x != nil { + return x.License + } + return nil +} + +func (x *Info) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type ItemsItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Schema []*Schema `protobuf:"bytes,1,rep,name=schema,proto3" json:"schema,omitempty"` +} + +func (x *ItemsItem) Reset() { + *x = ItemsItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[17] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *ItemsItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*ItemsItem) ProtoMessage() {} + +func (x *ItemsItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[17] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use ItemsItem.ProtoReflect.Descriptor instead. +func (*ItemsItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{17} +} + +func (x *ItemsItem) GetSchema() []*Schema { + if x != nil { + return x.Schema + } + return nil +} + +type JsonReference struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + XRef string `protobuf:"bytes,1,opt,name=_ref,json=Ref,proto3" json:"_ref,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` +} + +func (x *JsonReference) Reset() { + *x = JsonReference{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[18] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *JsonReference) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*JsonReference) ProtoMessage() {} + +func (x *JsonReference) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[18] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use JsonReference.ProtoReflect.Descriptor instead. +func (*JsonReference) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{18} +} + +func (x *JsonReference) GetXRef() string { + if x != nil { + return x.XRef + } + return "" +} + +func (x *JsonReference) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +type License struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // The name of the license type. It's encouraged to use an OSI compatible license. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The URL pointing to the license. + Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,3,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *License) Reset() { + *x = License{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[19] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *License) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*License) ProtoMessage() {} + +func (x *License) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[19] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use License.ProtoReflect.Descriptor instead. +func (*License) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{19} +} + +func (x *License) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *License) GetUrl() string { + if x != nil { + return x.Url + } + return "" +} + +func (x *License) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +// Automatically-generated message used to represent maps of Any as ordered (name,value) pairs. +type NamedAny struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *Any `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedAny) Reset() { + *x = NamedAny{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[20] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedAny) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedAny) ProtoMessage() {} + +func (x *NamedAny) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[20] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedAny.ProtoReflect.Descriptor instead. +func (*NamedAny) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{20} +} + +func (x *NamedAny) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedAny) GetValue() *Any { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of Header as ordered (name,value) pairs. +type NamedHeader struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *Header `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedHeader) Reset() { + *x = NamedHeader{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[21] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedHeader) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedHeader) ProtoMessage() {} + +func (x *NamedHeader) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[21] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedHeader.ProtoReflect.Descriptor instead. +func (*NamedHeader) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{21} +} + +func (x *NamedHeader) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedHeader) GetValue() *Header { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of Parameter as ordered (name,value) pairs. +type NamedParameter struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *Parameter `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedParameter) Reset() { + *x = NamedParameter{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[22] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedParameter) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedParameter) ProtoMessage() {} + +func (x *NamedParameter) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[22] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedParameter.ProtoReflect.Descriptor instead. +func (*NamedParameter) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{22} +} + +func (x *NamedParameter) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedParameter) GetValue() *Parameter { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of PathItem as ordered (name,value) pairs. +type NamedPathItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *PathItem `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedPathItem) Reset() { + *x = NamedPathItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[23] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedPathItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedPathItem) ProtoMessage() {} + +func (x *NamedPathItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[23] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedPathItem.ProtoReflect.Descriptor instead. +func (*NamedPathItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{23} +} + +func (x *NamedPathItem) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedPathItem) GetValue() *PathItem { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of Response as ordered (name,value) pairs. +type NamedResponse struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *Response `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedResponse) Reset() { + *x = NamedResponse{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[24] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedResponse) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedResponse) ProtoMessage() {} + +func (x *NamedResponse) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[24] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedResponse.ProtoReflect.Descriptor instead. +func (*NamedResponse) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{24} +} + +func (x *NamedResponse) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedResponse) GetValue() *Response { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of ResponseValue as ordered (name,value) pairs. +type NamedResponseValue struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *ResponseValue `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedResponseValue) Reset() { + *x = NamedResponseValue{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[25] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedResponseValue) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedResponseValue) ProtoMessage() {} + +func (x *NamedResponseValue) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[25] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedResponseValue.ProtoReflect.Descriptor instead. +func (*NamedResponseValue) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{25} +} + +func (x *NamedResponseValue) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedResponseValue) GetValue() *ResponseValue { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of Schema as ordered (name,value) pairs. +type NamedSchema struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *Schema `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedSchema) Reset() { + *x = NamedSchema{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[26] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedSchema) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedSchema) ProtoMessage() {} + +func (x *NamedSchema) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[26] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedSchema.ProtoReflect.Descriptor instead. +func (*NamedSchema) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{26} +} + +func (x *NamedSchema) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedSchema) GetValue() *Schema { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of SecurityDefinitionsItem as ordered (name,value) pairs. +type NamedSecurityDefinitionsItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *SecurityDefinitionsItem `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedSecurityDefinitionsItem) Reset() { + *x = NamedSecurityDefinitionsItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[27] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedSecurityDefinitionsItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedSecurityDefinitionsItem) ProtoMessage() {} + +func (x *NamedSecurityDefinitionsItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[27] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedSecurityDefinitionsItem.ProtoReflect.Descriptor instead. +func (*NamedSecurityDefinitionsItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{27} +} + +func (x *NamedSecurityDefinitionsItem) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedSecurityDefinitionsItem) GetValue() *SecurityDefinitionsItem { + if x != nil { + return x.Value + } + return nil +} + +// Automatically-generated message used to represent maps of string as ordered (name,value) pairs. +type NamedString struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedString) Reset() { + *x = NamedString{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[28] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedString) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedString) ProtoMessage() {} + +func (x *NamedString) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[28] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedString.ProtoReflect.Descriptor instead. +func (*NamedString) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{28} +} + +func (x *NamedString) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedString) GetValue() string { + if x != nil { + return x.Value + } + return "" +} + +// Automatically-generated message used to represent maps of StringArray as ordered (name,value) pairs. +type NamedStringArray struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Map key + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Mapped value + Value *StringArray `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *NamedStringArray) Reset() { + *x = NamedStringArray{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[29] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NamedStringArray) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NamedStringArray) ProtoMessage() {} + +func (x *NamedStringArray) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[29] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NamedStringArray.ProtoReflect.Descriptor instead. +func (*NamedStringArray) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{29} +} + +func (x *NamedStringArray) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *NamedStringArray) GetValue() *StringArray { + if x != nil { + return x.Value + } + return nil +} + +type NonBodyParameter struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Types that are assignable to Oneof: + // *NonBodyParameter_HeaderParameterSubSchema + // *NonBodyParameter_FormDataParameterSubSchema + // *NonBodyParameter_QueryParameterSubSchema + // *NonBodyParameter_PathParameterSubSchema + Oneof isNonBodyParameter_Oneof `protobuf_oneof:"oneof"` +} + +func (x *NonBodyParameter) Reset() { + *x = NonBodyParameter{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[30] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *NonBodyParameter) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*NonBodyParameter) ProtoMessage() {} + +func (x *NonBodyParameter) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[30] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use NonBodyParameter.ProtoReflect.Descriptor instead. +func (*NonBodyParameter) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{30} +} + +func (m *NonBodyParameter) GetOneof() isNonBodyParameter_Oneof { + if m != nil { + return m.Oneof + } + return nil +} + +func (x *NonBodyParameter) GetHeaderParameterSubSchema() *HeaderParameterSubSchema { + if x, ok := x.GetOneof().(*NonBodyParameter_HeaderParameterSubSchema); ok { + return x.HeaderParameterSubSchema + } + return nil +} + +func (x *NonBodyParameter) GetFormDataParameterSubSchema() *FormDataParameterSubSchema { + if x, ok := x.GetOneof().(*NonBodyParameter_FormDataParameterSubSchema); ok { + return x.FormDataParameterSubSchema + } + return nil +} + +func (x *NonBodyParameter) GetQueryParameterSubSchema() *QueryParameterSubSchema { + if x, ok := x.GetOneof().(*NonBodyParameter_QueryParameterSubSchema); ok { + return x.QueryParameterSubSchema + } + return nil +} + +func (x *NonBodyParameter) GetPathParameterSubSchema() *PathParameterSubSchema { + if x, ok := x.GetOneof().(*NonBodyParameter_PathParameterSubSchema); ok { + return x.PathParameterSubSchema + } + return nil +} + +type isNonBodyParameter_Oneof interface { + isNonBodyParameter_Oneof() +} + +type NonBodyParameter_HeaderParameterSubSchema struct { + HeaderParameterSubSchema *HeaderParameterSubSchema `protobuf:"bytes,1,opt,name=header_parameter_sub_schema,json=headerParameterSubSchema,proto3,oneof"` +} + +type NonBodyParameter_FormDataParameterSubSchema struct { + FormDataParameterSubSchema *FormDataParameterSubSchema `protobuf:"bytes,2,opt,name=form_data_parameter_sub_schema,json=formDataParameterSubSchema,proto3,oneof"` +} + +type NonBodyParameter_QueryParameterSubSchema struct { + QueryParameterSubSchema *QueryParameterSubSchema `protobuf:"bytes,3,opt,name=query_parameter_sub_schema,json=queryParameterSubSchema,proto3,oneof"` +} + +type NonBodyParameter_PathParameterSubSchema struct { + PathParameterSubSchema *PathParameterSubSchema `protobuf:"bytes,4,opt,name=path_parameter_sub_schema,json=pathParameterSubSchema,proto3,oneof"` +} + +func (*NonBodyParameter_HeaderParameterSubSchema) isNonBodyParameter_Oneof() {} + +func (*NonBodyParameter_FormDataParameterSubSchema) isNonBodyParameter_Oneof() {} + +func (*NonBodyParameter_QueryParameterSubSchema) isNonBodyParameter_Oneof() {} + +func (*NonBodyParameter_PathParameterSubSchema) isNonBodyParameter_Oneof() {} + +type Oauth2AccessCodeSecurity struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` + Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` + AuthorizationUrl string `protobuf:"bytes,4,opt,name=authorization_url,json=authorizationUrl,proto3" json:"authorization_url,omitempty"` + TokenUrl string `protobuf:"bytes,5,opt,name=token_url,json=tokenUrl,proto3" json:"token_url,omitempty"` + Description string `protobuf:"bytes,6,opt,name=description,proto3" json:"description,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,7,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Oauth2AccessCodeSecurity) Reset() { + *x = Oauth2AccessCodeSecurity{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[31] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Oauth2AccessCodeSecurity) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Oauth2AccessCodeSecurity) ProtoMessage() {} + +func (x *Oauth2AccessCodeSecurity) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[31] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Oauth2AccessCodeSecurity.ProtoReflect.Descriptor instead. +func (*Oauth2AccessCodeSecurity) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{31} +} + +func (x *Oauth2AccessCodeSecurity) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *Oauth2AccessCodeSecurity) GetFlow() string { + if x != nil { + return x.Flow + } + return "" +} + +func (x *Oauth2AccessCodeSecurity) GetScopes() *Oauth2Scopes { + if x != nil { + return x.Scopes + } + return nil +} + +func (x *Oauth2AccessCodeSecurity) GetAuthorizationUrl() string { + if x != nil { + return x.AuthorizationUrl + } + return "" +} + +func (x *Oauth2AccessCodeSecurity) GetTokenUrl() string { + if x != nil { + return x.TokenUrl + } + return "" +} + +func (x *Oauth2AccessCodeSecurity) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Oauth2AccessCodeSecurity) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Oauth2ApplicationSecurity struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` + Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` + TokenUrl string `protobuf:"bytes,4,opt,name=token_url,json=tokenUrl,proto3" json:"token_url,omitempty"` + Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Oauth2ApplicationSecurity) Reset() { + *x = Oauth2ApplicationSecurity{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[32] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Oauth2ApplicationSecurity) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Oauth2ApplicationSecurity) ProtoMessage() {} + +func (x *Oauth2ApplicationSecurity) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[32] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Oauth2ApplicationSecurity.ProtoReflect.Descriptor instead. +func (*Oauth2ApplicationSecurity) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{32} +} + +func (x *Oauth2ApplicationSecurity) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *Oauth2ApplicationSecurity) GetFlow() string { + if x != nil { + return x.Flow + } + return "" +} + +func (x *Oauth2ApplicationSecurity) GetScopes() *Oauth2Scopes { + if x != nil { + return x.Scopes + } + return nil +} + +func (x *Oauth2ApplicationSecurity) GetTokenUrl() string { + if x != nil { + return x.TokenUrl + } + return "" +} + +func (x *Oauth2ApplicationSecurity) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Oauth2ApplicationSecurity) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Oauth2ImplicitSecurity struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` + Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` + AuthorizationUrl string `protobuf:"bytes,4,opt,name=authorization_url,json=authorizationUrl,proto3" json:"authorization_url,omitempty"` + Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Oauth2ImplicitSecurity) Reset() { + *x = Oauth2ImplicitSecurity{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[33] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Oauth2ImplicitSecurity) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Oauth2ImplicitSecurity) ProtoMessage() {} + +func (x *Oauth2ImplicitSecurity) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[33] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Oauth2ImplicitSecurity.ProtoReflect.Descriptor instead. +func (*Oauth2ImplicitSecurity) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{33} +} + +func (x *Oauth2ImplicitSecurity) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *Oauth2ImplicitSecurity) GetFlow() string { + if x != nil { + return x.Flow + } + return "" +} + +func (x *Oauth2ImplicitSecurity) GetScopes() *Oauth2Scopes { + if x != nil { + return x.Scopes + } + return nil +} + +func (x *Oauth2ImplicitSecurity) GetAuthorizationUrl() string { + if x != nil { + return x.AuthorizationUrl + } + return "" +} + +func (x *Oauth2ImplicitSecurity) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Oauth2ImplicitSecurity) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Oauth2PasswordSecurity struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Flow string `protobuf:"bytes,2,opt,name=flow,proto3" json:"flow,omitempty"` + Scopes *Oauth2Scopes `protobuf:"bytes,3,opt,name=scopes,proto3" json:"scopes,omitempty"` + TokenUrl string `protobuf:"bytes,4,opt,name=token_url,json=tokenUrl,proto3" json:"token_url,omitempty"` + Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Oauth2PasswordSecurity) Reset() { + *x = Oauth2PasswordSecurity{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[34] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Oauth2PasswordSecurity) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Oauth2PasswordSecurity) ProtoMessage() {} + +func (x *Oauth2PasswordSecurity) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[34] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Oauth2PasswordSecurity.ProtoReflect.Descriptor instead. +func (*Oauth2PasswordSecurity) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{34} +} + +func (x *Oauth2PasswordSecurity) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *Oauth2PasswordSecurity) GetFlow() string { + if x != nil { + return x.Flow + } + return "" +} + +func (x *Oauth2PasswordSecurity) GetScopes() *Oauth2Scopes { + if x != nil { + return x.Scopes + } + return nil +} + +func (x *Oauth2PasswordSecurity) GetTokenUrl() string { + if x != nil { + return x.TokenUrl + } + return "" +} + +func (x *Oauth2PasswordSecurity) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Oauth2PasswordSecurity) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Oauth2Scopes struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedString `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *Oauth2Scopes) Reset() { + *x = Oauth2Scopes{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[35] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Oauth2Scopes) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Oauth2Scopes) ProtoMessage() {} + +func (x *Oauth2Scopes) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[35] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Oauth2Scopes.ProtoReflect.Descriptor instead. +func (*Oauth2Scopes) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{35} +} + +func (x *Oauth2Scopes) GetAdditionalProperties() []*NamedString { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type Operation struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Tags []string `protobuf:"bytes,1,rep,name=tags,proto3" json:"tags,omitempty"` + // A brief summary of the operation. + Summary string `protobuf:"bytes,2,opt,name=summary,proto3" json:"summary,omitempty"` + // A longer description of the operation, GitHub Flavored Markdown is allowed. + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` + ExternalDocs *ExternalDocs `protobuf:"bytes,4,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` + // A unique identifier of the operation. + OperationId string `protobuf:"bytes,5,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` + // A list of MIME types the API can produce. + Produces []string `protobuf:"bytes,6,rep,name=produces,proto3" json:"produces,omitempty"` + // A list of MIME types the API can consume. + Consumes []string `protobuf:"bytes,7,rep,name=consumes,proto3" json:"consumes,omitempty"` + // The parameters needed to send a valid API call. + Parameters []*ParametersItem `protobuf:"bytes,8,rep,name=parameters,proto3" json:"parameters,omitempty"` + Responses *Responses `protobuf:"bytes,9,opt,name=responses,proto3" json:"responses,omitempty"` + // The transfer protocol of the API. + Schemes []string `protobuf:"bytes,10,rep,name=schemes,proto3" json:"schemes,omitempty"` + Deprecated bool `protobuf:"varint,11,opt,name=deprecated,proto3" json:"deprecated,omitempty"` + Security []*SecurityRequirement `protobuf:"bytes,12,rep,name=security,proto3" json:"security,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,13,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Operation) Reset() { + *x = Operation{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[36] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Operation) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Operation) ProtoMessage() {} + +func (x *Operation) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[36] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Operation.ProtoReflect.Descriptor instead. +func (*Operation) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{36} +} + +func (x *Operation) GetTags() []string { + if x != nil { + return x.Tags + } + return nil +} + +func (x *Operation) GetSummary() string { + if x != nil { + return x.Summary + } + return "" +} + +func (x *Operation) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Operation) GetExternalDocs() *ExternalDocs { + if x != nil { + return x.ExternalDocs + } + return nil +} + +func (x *Operation) GetOperationId() string { + if x != nil { + return x.OperationId + } + return "" +} + +func (x *Operation) GetProduces() []string { + if x != nil { + return x.Produces + } + return nil +} + +func (x *Operation) GetConsumes() []string { + if x != nil { + return x.Consumes + } + return nil +} + +func (x *Operation) GetParameters() []*ParametersItem { + if x != nil { + return x.Parameters + } + return nil +} + +func (x *Operation) GetResponses() *Responses { + if x != nil { + return x.Responses + } + return nil +} + +func (x *Operation) GetSchemes() []string { + if x != nil { + return x.Schemes + } + return nil +} + +func (x *Operation) GetDeprecated() bool { + if x != nil { + return x.Deprecated + } + return false +} + +func (x *Operation) GetSecurity() []*SecurityRequirement { + if x != nil { + return x.Security + } + return nil +} + +func (x *Operation) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Parameter struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Types that are assignable to Oneof: + // *Parameter_BodyParameter + // *Parameter_NonBodyParameter + Oneof isParameter_Oneof `protobuf_oneof:"oneof"` +} + +func (x *Parameter) Reset() { + *x = Parameter{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[37] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Parameter) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Parameter) ProtoMessage() {} + +func (x *Parameter) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[37] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Parameter.ProtoReflect.Descriptor instead. +func (*Parameter) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{37} +} + +func (m *Parameter) GetOneof() isParameter_Oneof { + if m != nil { + return m.Oneof + } + return nil +} + +func (x *Parameter) GetBodyParameter() *BodyParameter { + if x, ok := x.GetOneof().(*Parameter_BodyParameter); ok { + return x.BodyParameter + } + return nil +} + +func (x *Parameter) GetNonBodyParameter() *NonBodyParameter { + if x, ok := x.GetOneof().(*Parameter_NonBodyParameter); ok { + return x.NonBodyParameter + } + return nil +} + +type isParameter_Oneof interface { + isParameter_Oneof() +} + +type Parameter_BodyParameter struct { + BodyParameter *BodyParameter `protobuf:"bytes,1,opt,name=body_parameter,json=bodyParameter,proto3,oneof"` +} + +type Parameter_NonBodyParameter struct { + NonBodyParameter *NonBodyParameter `protobuf:"bytes,2,opt,name=non_body_parameter,json=nonBodyParameter,proto3,oneof"` +} + +func (*Parameter_BodyParameter) isParameter_Oneof() {} + +func (*Parameter_NonBodyParameter) isParameter_Oneof() {} + +// One or more JSON representations for parameters +type ParameterDefinitions struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedParameter `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *ParameterDefinitions) Reset() { + *x = ParameterDefinitions{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[38] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *ParameterDefinitions) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*ParameterDefinitions) ProtoMessage() {} + +func (x *ParameterDefinitions) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[38] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use ParameterDefinitions.ProtoReflect.Descriptor instead. +func (*ParameterDefinitions) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{38} +} + +func (x *ParameterDefinitions) GetAdditionalProperties() []*NamedParameter { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type ParametersItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Types that are assignable to Oneof: + // *ParametersItem_Parameter + // *ParametersItem_JsonReference + Oneof isParametersItem_Oneof `protobuf_oneof:"oneof"` +} + +func (x *ParametersItem) Reset() { + *x = ParametersItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[39] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *ParametersItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*ParametersItem) ProtoMessage() {} + +func (x *ParametersItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[39] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use ParametersItem.ProtoReflect.Descriptor instead. +func (*ParametersItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{39} +} + +func (m *ParametersItem) GetOneof() isParametersItem_Oneof { + if m != nil { + return m.Oneof + } + return nil +} + +func (x *ParametersItem) GetParameter() *Parameter { + if x, ok := x.GetOneof().(*ParametersItem_Parameter); ok { + return x.Parameter + } + return nil +} + +func (x *ParametersItem) GetJsonReference() *JsonReference { + if x, ok := x.GetOneof().(*ParametersItem_JsonReference); ok { + return x.JsonReference + } + return nil +} + +type isParametersItem_Oneof interface { + isParametersItem_Oneof() +} + +type ParametersItem_Parameter struct { + Parameter *Parameter `protobuf:"bytes,1,opt,name=parameter,proto3,oneof"` +} + +type ParametersItem_JsonReference struct { + JsonReference *JsonReference `protobuf:"bytes,2,opt,name=json_reference,json=jsonReference,proto3,oneof"` +} + +func (*ParametersItem_Parameter) isParametersItem_Oneof() {} + +func (*ParametersItem_JsonReference) isParametersItem_Oneof() {} + +type PathItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + XRef string `protobuf:"bytes,1,opt,name=_ref,json=Ref,proto3" json:"_ref,omitempty"` + Get *Operation `protobuf:"bytes,2,opt,name=get,proto3" json:"get,omitempty"` + Put *Operation `protobuf:"bytes,3,opt,name=put,proto3" json:"put,omitempty"` + Post *Operation `protobuf:"bytes,4,opt,name=post,proto3" json:"post,omitempty"` + Delete *Operation `protobuf:"bytes,5,opt,name=delete,proto3" json:"delete,omitempty"` + Options *Operation `protobuf:"bytes,6,opt,name=options,proto3" json:"options,omitempty"` + Head *Operation `protobuf:"bytes,7,opt,name=head,proto3" json:"head,omitempty"` + Patch *Operation `protobuf:"bytes,8,opt,name=patch,proto3" json:"patch,omitempty"` + // The parameters needed to send a valid API call. + Parameters []*ParametersItem `protobuf:"bytes,9,rep,name=parameters,proto3" json:"parameters,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,10,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *PathItem) Reset() { + *x = PathItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[40] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *PathItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*PathItem) ProtoMessage() {} + +func (x *PathItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[40] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use PathItem.ProtoReflect.Descriptor instead. +func (*PathItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{40} +} + +func (x *PathItem) GetXRef() string { + if x != nil { + return x.XRef + } + return "" +} + +func (x *PathItem) GetGet() *Operation { + if x != nil { + return x.Get + } + return nil +} + +func (x *PathItem) GetPut() *Operation { + if x != nil { + return x.Put + } + return nil +} + +func (x *PathItem) GetPost() *Operation { + if x != nil { + return x.Post + } + return nil +} + +func (x *PathItem) GetDelete() *Operation { + if x != nil { + return x.Delete + } + return nil +} + +func (x *PathItem) GetOptions() *Operation { + if x != nil { + return x.Options + } + return nil +} + +func (x *PathItem) GetHead() *Operation { + if x != nil { + return x.Head + } + return nil +} + +func (x *PathItem) GetPatch() *Operation { + if x != nil { + return x.Patch + } + return nil +} + +func (x *PathItem) GetParameters() []*ParametersItem { + if x != nil { + return x.Parameters + } + return nil +} + +func (x *PathItem) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type PathParameterSubSchema struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Determines whether or not this parameter is required or optional. + Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` + // Determines the location of the parameter. + In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` + // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` + // The name of the parameter. + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` + Type string `protobuf:"bytes,5,opt,name=type,proto3" json:"type,omitempty"` + Format string `protobuf:"bytes,6,opt,name=format,proto3" json:"format,omitempty"` + Items *PrimitivesItems `protobuf:"bytes,7,opt,name=items,proto3" json:"items,omitempty"` + CollectionFormat string `protobuf:"bytes,8,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` + Default *Any `protobuf:"bytes,9,opt,name=default,proto3" json:"default,omitempty"` + Maximum float64 `protobuf:"fixed64,10,opt,name=maximum,proto3" json:"maximum,omitempty"` + ExclusiveMaximum bool `protobuf:"varint,11,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` + Minimum float64 `protobuf:"fixed64,12,opt,name=minimum,proto3" json:"minimum,omitempty"` + ExclusiveMinimum bool `protobuf:"varint,13,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` + MaxLength int64 `protobuf:"varint,14,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` + MinLength int64 `protobuf:"varint,15,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` + Pattern string `protobuf:"bytes,16,opt,name=pattern,proto3" json:"pattern,omitempty"` + MaxItems int64 `protobuf:"varint,17,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` + MinItems int64 `protobuf:"varint,18,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` + UniqueItems bool `protobuf:"varint,19,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` + Enum []*Any `protobuf:"bytes,20,rep,name=enum,proto3" json:"enum,omitempty"` + MultipleOf float64 `protobuf:"fixed64,21,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,22,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *PathParameterSubSchema) Reset() { + *x = PathParameterSubSchema{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[41] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *PathParameterSubSchema) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*PathParameterSubSchema) ProtoMessage() {} + +func (x *PathParameterSubSchema) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[41] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use PathParameterSubSchema.ProtoReflect.Descriptor instead. +func (*PathParameterSubSchema) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{41} +} + +func (x *PathParameterSubSchema) GetRequired() bool { + if x != nil { + return x.Required + } + return false +} + +func (x *PathParameterSubSchema) GetIn() string { + if x != nil { + return x.In + } + return "" +} + +func (x *PathParameterSubSchema) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *PathParameterSubSchema) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *PathParameterSubSchema) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *PathParameterSubSchema) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *PathParameterSubSchema) GetItems() *PrimitivesItems { + if x != nil { + return x.Items + } + return nil +} + +func (x *PathParameterSubSchema) GetCollectionFormat() string { + if x != nil { + return x.CollectionFormat + } + return "" +} + +func (x *PathParameterSubSchema) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *PathParameterSubSchema) GetMaximum() float64 { + if x != nil { + return x.Maximum + } + return 0 +} + +func (x *PathParameterSubSchema) GetExclusiveMaximum() bool { + if x != nil { + return x.ExclusiveMaximum + } + return false +} + +func (x *PathParameterSubSchema) GetMinimum() float64 { + if x != nil { + return x.Minimum + } + return 0 +} + +func (x *PathParameterSubSchema) GetExclusiveMinimum() bool { + if x != nil { + return x.ExclusiveMinimum + } + return false +} + +func (x *PathParameterSubSchema) GetMaxLength() int64 { + if x != nil { + return x.MaxLength + } + return 0 +} + +func (x *PathParameterSubSchema) GetMinLength() int64 { + if x != nil { + return x.MinLength + } + return 0 +} + +func (x *PathParameterSubSchema) GetPattern() string { + if x != nil { + return x.Pattern + } + return "" +} + +func (x *PathParameterSubSchema) GetMaxItems() int64 { + if x != nil { + return x.MaxItems + } + return 0 +} + +func (x *PathParameterSubSchema) GetMinItems() int64 { + if x != nil { + return x.MinItems + } + return 0 +} + +func (x *PathParameterSubSchema) GetUniqueItems() bool { + if x != nil { + return x.UniqueItems + } + return false +} + +func (x *PathParameterSubSchema) GetEnum() []*Any { + if x != nil { + return x.Enum + } + return nil +} + +func (x *PathParameterSubSchema) GetMultipleOf() float64 { + if x != nil { + return x.MultipleOf + } + return 0 +} + +func (x *PathParameterSubSchema) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +// Relative paths to the individual endpoints. They must be relative to the 'basePath'. +type Paths struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + VendorExtension []*NamedAny `protobuf:"bytes,1,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` + Path []*NamedPathItem `protobuf:"bytes,2,rep,name=path,proto3" json:"path,omitempty"` +} + +func (x *Paths) Reset() { + *x = Paths{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[42] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Paths) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Paths) ProtoMessage() {} + +func (x *Paths) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[42] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Paths.ProtoReflect.Descriptor instead. +func (*Paths) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{42} +} + +func (x *Paths) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +func (x *Paths) GetPath() []*NamedPathItem { + if x != nil { + return x.Path + } + return nil +} + +type PrimitivesItems struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + Format string `protobuf:"bytes,2,opt,name=format,proto3" json:"format,omitempty"` + Items *PrimitivesItems `protobuf:"bytes,3,opt,name=items,proto3" json:"items,omitempty"` + CollectionFormat string `protobuf:"bytes,4,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` + Default *Any `protobuf:"bytes,5,opt,name=default,proto3" json:"default,omitempty"` + Maximum float64 `protobuf:"fixed64,6,opt,name=maximum,proto3" json:"maximum,omitempty"` + ExclusiveMaximum bool `protobuf:"varint,7,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` + Minimum float64 `protobuf:"fixed64,8,opt,name=minimum,proto3" json:"minimum,omitempty"` + ExclusiveMinimum bool `protobuf:"varint,9,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` + MaxLength int64 `protobuf:"varint,10,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` + MinLength int64 `protobuf:"varint,11,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` + Pattern string `protobuf:"bytes,12,opt,name=pattern,proto3" json:"pattern,omitempty"` + MaxItems int64 `protobuf:"varint,13,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` + MinItems int64 `protobuf:"varint,14,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` + UniqueItems bool `protobuf:"varint,15,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` + Enum []*Any `protobuf:"bytes,16,rep,name=enum,proto3" json:"enum,omitempty"` + MultipleOf float64 `protobuf:"fixed64,17,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,18,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *PrimitivesItems) Reset() { + *x = PrimitivesItems{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[43] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *PrimitivesItems) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*PrimitivesItems) ProtoMessage() {} + +func (x *PrimitivesItems) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[43] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use PrimitivesItems.ProtoReflect.Descriptor instead. +func (*PrimitivesItems) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{43} +} + +func (x *PrimitivesItems) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *PrimitivesItems) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *PrimitivesItems) GetItems() *PrimitivesItems { + if x != nil { + return x.Items + } + return nil +} + +func (x *PrimitivesItems) GetCollectionFormat() string { + if x != nil { + return x.CollectionFormat + } + return "" +} + +func (x *PrimitivesItems) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *PrimitivesItems) GetMaximum() float64 { + if x != nil { + return x.Maximum + } + return 0 +} + +func (x *PrimitivesItems) GetExclusiveMaximum() bool { + if x != nil { + return x.ExclusiveMaximum + } + return false +} + +func (x *PrimitivesItems) GetMinimum() float64 { + if x != nil { + return x.Minimum + } + return 0 +} + +func (x *PrimitivesItems) GetExclusiveMinimum() bool { + if x != nil { + return x.ExclusiveMinimum + } + return false +} + +func (x *PrimitivesItems) GetMaxLength() int64 { + if x != nil { + return x.MaxLength + } + return 0 +} + +func (x *PrimitivesItems) GetMinLength() int64 { + if x != nil { + return x.MinLength + } + return 0 +} + +func (x *PrimitivesItems) GetPattern() string { + if x != nil { + return x.Pattern + } + return "" +} + +func (x *PrimitivesItems) GetMaxItems() int64 { + if x != nil { + return x.MaxItems + } + return 0 +} + +func (x *PrimitivesItems) GetMinItems() int64 { + if x != nil { + return x.MinItems + } + return 0 +} + +func (x *PrimitivesItems) GetUniqueItems() bool { + if x != nil { + return x.UniqueItems + } + return false +} + +func (x *PrimitivesItems) GetEnum() []*Any { + if x != nil { + return x.Enum + } + return nil +} + +func (x *PrimitivesItems) GetMultipleOf() float64 { + if x != nil { + return x.MultipleOf + } + return 0 +} + +func (x *PrimitivesItems) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Properties struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedSchema `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *Properties) Reset() { + *x = Properties{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[44] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Properties) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Properties) ProtoMessage() {} + +func (x *Properties) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[44] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Properties.ProtoReflect.Descriptor instead. +func (*Properties) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{44} +} + +func (x *Properties) GetAdditionalProperties() []*NamedSchema { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type QueryParameterSubSchema struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Determines whether or not this parameter is required or optional. + Required bool `protobuf:"varint,1,opt,name=required,proto3" json:"required,omitempty"` + // Determines the location of the parameter. + In string `protobuf:"bytes,2,opt,name=in,proto3" json:"in,omitempty"` + // A brief description of the parameter. This could contain examples of use. GitHub Flavored Markdown is allowed. + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` + // The name of the parameter. + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` + // allows sending a parameter by name only or with an empty value. + AllowEmptyValue bool `protobuf:"varint,5,opt,name=allow_empty_value,json=allowEmptyValue,proto3" json:"allow_empty_value,omitempty"` + Type string `protobuf:"bytes,6,opt,name=type,proto3" json:"type,omitempty"` + Format string `protobuf:"bytes,7,opt,name=format,proto3" json:"format,omitempty"` + Items *PrimitivesItems `protobuf:"bytes,8,opt,name=items,proto3" json:"items,omitempty"` + CollectionFormat string `protobuf:"bytes,9,opt,name=collection_format,json=collectionFormat,proto3" json:"collection_format,omitempty"` + Default *Any `protobuf:"bytes,10,opt,name=default,proto3" json:"default,omitempty"` + Maximum float64 `protobuf:"fixed64,11,opt,name=maximum,proto3" json:"maximum,omitempty"` + ExclusiveMaximum bool `protobuf:"varint,12,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` + Minimum float64 `protobuf:"fixed64,13,opt,name=minimum,proto3" json:"minimum,omitempty"` + ExclusiveMinimum bool `protobuf:"varint,14,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` + MaxLength int64 `protobuf:"varint,15,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` + MinLength int64 `protobuf:"varint,16,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` + Pattern string `protobuf:"bytes,17,opt,name=pattern,proto3" json:"pattern,omitempty"` + MaxItems int64 `protobuf:"varint,18,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` + MinItems int64 `protobuf:"varint,19,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` + UniqueItems bool `protobuf:"varint,20,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` + Enum []*Any `protobuf:"bytes,21,rep,name=enum,proto3" json:"enum,omitempty"` + MultipleOf float64 `protobuf:"fixed64,22,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,23,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *QueryParameterSubSchema) Reset() { + *x = QueryParameterSubSchema{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[45] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *QueryParameterSubSchema) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*QueryParameterSubSchema) ProtoMessage() {} + +func (x *QueryParameterSubSchema) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[45] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use QueryParameterSubSchema.ProtoReflect.Descriptor instead. +func (*QueryParameterSubSchema) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{45} +} + +func (x *QueryParameterSubSchema) GetRequired() bool { + if x != nil { + return x.Required + } + return false +} + +func (x *QueryParameterSubSchema) GetIn() string { + if x != nil { + return x.In + } + return "" +} + +func (x *QueryParameterSubSchema) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *QueryParameterSubSchema) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *QueryParameterSubSchema) GetAllowEmptyValue() bool { + if x != nil { + return x.AllowEmptyValue + } + return false +} + +func (x *QueryParameterSubSchema) GetType() string { + if x != nil { + return x.Type + } + return "" +} + +func (x *QueryParameterSubSchema) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *QueryParameterSubSchema) GetItems() *PrimitivesItems { + if x != nil { + return x.Items + } + return nil +} + +func (x *QueryParameterSubSchema) GetCollectionFormat() string { + if x != nil { + return x.CollectionFormat + } + return "" +} + +func (x *QueryParameterSubSchema) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *QueryParameterSubSchema) GetMaximum() float64 { + if x != nil { + return x.Maximum + } + return 0 +} + +func (x *QueryParameterSubSchema) GetExclusiveMaximum() bool { + if x != nil { + return x.ExclusiveMaximum + } + return false +} + +func (x *QueryParameterSubSchema) GetMinimum() float64 { + if x != nil { + return x.Minimum + } + return 0 +} + +func (x *QueryParameterSubSchema) GetExclusiveMinimum() bool { + if x != nil { + return x.ExclusiveMinimum + } + return false +} + +func (x *QueryParameterSubSchema) GetMaxLength() int64 { + if x != nil { + return x.MaxLength + } + return 0 +} + +func (x *QueryParameterSubSchema) GetMinLength() int64 { + if x != nil { + return x.MinLength + } + return 0 +} + +func (x *QueryParameterSubSchema) GetPattern() string { + if x != nil { + return x.Pattern + } + return "" +} + +func (x *QueryParameterSubSchema) GetMaxItems() int64 { + if x != nil { + return x.MaxItems + } + return 0 +} + +func (x *QueryParameterSubSchema) GetMinItems() int64 { + if x != nil { + return x.MinItems + } + return 0 +} + +func (x *QueryParameterSubSchema) GetUniqueItems() bool { + if x != nil { + return x.UniqueItems + } + return false +} + +func (x *QueryParameterSubSchema) GetEnum() []*Any { + if x != nil { + return x.Enum + } + return nil +} + +func (x *QueryParameterSubSchema) GetMultipleOf() float64 { + if x != nil { + return x.MultipleOf + } + return 0 +} + +func (x *QueryParameterSubSchema) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type Response struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` + Schema *SchemaItem `protobuf:"bytes,2,opt,name=schema,proto3" json:"schema,omitempty"` + Headers *Headers `protobuf:"bytes,3,opt,name=headers,proto3" json:"headers,omitempty"` + Examples *Examples `protobuf:"bytes,4,opt,name=examples,proto3" json:"examples,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,5,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Response) Reset() { + *x = Response{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[46] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Response) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Response) ProtoMessage() {} + +func (x *Response) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[46] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Response.ProtoReflect.Descriptor instead. +func (*Response) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{46} +} + +func (x *Response) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Response) GetSchema() *SchemaItem { + if x != nil { + return x.Schema + } + return nil +} + +func (x *Response) GetHeaders() *Headers { + if x != nil { + return x.Headers + } + return nil +} + +func (x *Response) GetExamples() *Examples { + if x != nil { + return x.Examples + } + return nil +} + +func (x *Response) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +// One or more JSON representations for responses +type ResponseDefinitions struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedResponse `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *ResponseDefinitions) Reset() { + *x = ResponseDefinitions{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[47] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *ResponseDefinitions) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*ResponseDefinitions) ProtoMessage() {} + +func (x *ResponseDefinitions) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[47] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use ResponseDefinitions.ProtoReflect.Descriptor instead. +func (*ResponseDefinitions) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{47} +} + +func (x *ResponseDefinitions) GetAdditionalProperties() []*NamedResponse { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type ResponseValue struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Types that are assignable to Oneof: + // *ResponseValue_Response + // *ResponseValue_JsonReference + Oneof isResponseValue_Oneof `protobuf_oneof:"oneof"` +} + +func (x *ResponseValue) Reset() { + *x = ResponseValue{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[48] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *ResponseValue) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*ResponseValue) ProtoMessage() {} + +func (x *ResponseValue) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[48] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use ResponseValue.ProtoReflect.Descriptor instead. +func (*ResponseValue) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{48} +} + +func (m *ResponseValue) GetOneof() isResponseValue_Oneof { + if m != nil { + return m.Oneof + } + return nil +} + +func (x *ResponseValue) GetResponse() *Response { + if x, ok := x.GetOneof().(*ResponseValue_Response); ok { + return x.Response + } + return nil +} + +func (x *ResponseValue) GetJsonReference() *JsonReference { + if x, ok := x.GetOneof().(*ResponseValue_JsonReference); ok { + return x.JsonReference + } + return nil +} + +type isResponseValue_Oneof interface { + isResponseValue_Oneof() +} + +type ResponseValue_Response struct { + Response *Response `protobuf:"bytes,1,opt,name=response,proto3,oneof"` +} + +type ResponseValue_JsonReference struct { + JsonReference *JsonReference `protobuf:"bytes,2,opt,name=json_reference,json=jsonReference,proto3,oneof"` +} + +func (*ResponseValue_Response) isResponseValue_Oneof() {} + +func (*ResponseValue_JsonReference) isResponseValue_Oneof() {} + +// Response objects names can either be any valid HTTP status code or 'default'. +type Responses struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + ResponseCode []*NamedResponseValue `protobuf:"bytes,1,rep,name=response_code,json=responseCode,proto3" json:"response_code,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,2,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Responses) Reset() { + *x = Responses{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[49] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Responses) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Responses) ProtoMessage() {} + +func (x *Responses) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[49] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Responses.ProtoReflect.Descriptor instead. +func (*Responses) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{49} +} + +func (x *Responses) GetResponseCode() []*NamedResponseValue { + if x != nil { + return x.ResponseCode + } + return nil +} + +func (x *Responses) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +// A deterministic version of a JSON Schema object. +type Schema struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + XRef string `protobuf:"bytes,1,opt,name=_ref,json=Ref,proto3" json:"_ref,omitempty"` + Format string `protobuf:"bytes,2,opt,name=format,proto3" json:"format,omitempty"` + Title string `protobuf:"bytes,3,opt,name=title,proto3" json:"title,omitempty"` + Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` + Default *Any `protobuf:"bytes,5,opt,name=default,proto3" json:"default,omitempty"` + MultipleOf float64 `protobuf:"fixed64,6,opt,name=multiple_of,json=multipleOf,proto3" json:"multiple_of,omitempty"` + Maximum float64 `protobuf:"fixed64,7,opt,name=maximum,proto3" json:"maximum,omitempty"` + ExclusiveMaximum bool `protobuf:"varint,8,opt,name=exclusive_maximum,json=exclusiveMaximum,proto3" json:"exclusive_maximum,omitempty"` + Minimum float64 `protobuf:"fixed64,9,opt,name=minimum,proto3" json:"minimum,omitempty"` + ExclusiveMinimum bool `protobuf:"varint,10,opt,name=exclusive_minimum,json=exclusiveMinimum,proto3" json:"exclusive_minimum,omitempty"` + MaxLength int64 `protobuf:"varint,11,opt,name=max_length,json=maxLength,proto3" json:"max_length,omitempty"` + MinLength int64 `protobuf:"varint,12,opt,name=min_length,json=minLength,proto3" json:"min_length,omitempty"` + Pattern string `protobuf:"bytes,13,opt,name=pattern,proto3" json:"pattern,omitempty"` + MaxItems int64 `protobuf:"varint,14,opt,name=max_items,json=maxItems,proto3" json:"max_items,omitempty"` + MinItems int64 `protobuf:"varint,15,opt,name=min_items,json=minItems,proto3" json:"min_items,omitempty"` + UniqueItems bool `protobuf:"varint,16,opt,name=unique_items,json=uniqueItems,proto3" json:"unique_items,omitempty"` + MaxProperties int64 `protobuf:"varint,17,opt,name=max_properties,json=maxProperties,proto3" json:"max_properties,omitempty"` + MinProperties int64 `protobuf:"varint,18,opt,name=min_properties,json=minProperties,proto3" json:"min_properties,omitempty"` + Required []string `protobuf:"bytes,19,rep,name=required,proto3" json:"required,omitempty"` + Enum []*Any `protobuf:"bytes,20,rep,name=enum,proto3" json:"enum,omitempty"` + AdditionalProperties *AdditionalPropertiesItem `protobuf:"bytes,21,opt,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` + Type *TypeItem `protobuf:"bytes,22,opt,name=type,proto3" json:"type,omitempty"` + Items *ItemsItem `protobuf:"bytes,23,opt,name=items,proto3" json:"items,omitempty"` + AllOf []*Schema `protobuf:"bytes,24,rep,name=all_of,json=allOf,proto3" json:"all_of,omitempty"` + Properties *Properties `protobuf:"bytes,25,opt,name=properties,proto3" json:"properties,omitempty"` + Discriminator string `protobuf:"bytes,26,opt,name=discriminator,proto3" json:"discriminator,omitempty"` + ReadOnly bool `protobuf:"varint,27,opt,name=read_only,json=readOnly,proto3" json:"read_only,omitempty"` + Xml *Xml `protobuf:"bytes,28,opt,name=xml,proto3" json:"xml,omitempty"` + ExternalDocs *ExternalDocs `protobuf:"bytes,29,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` + Example *Any `protobuf:"bytes,30,opt,name=example,proto3" json:"example,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,31,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Schema) Reset() { + *x = Schema{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[50] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Schema) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Schema) ProtoMessage() {} + +func (x *Schema) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[50] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Schema.ProtoReflect.Descriptor instead. +func (*Schema) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{50} +} + +func (x *Schema) GetXRef() string { + if x != nil { + return x.XRef + } + return "" +} + +func (x *Schema) GetFormat() string { + if x != nil { + return x.Format + } + return "" +} + +func (x *Schema) GetTitle() string { + if x != nil { + return x.Title + } + return "" +} + +func (x *Schema) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Schema) GetDefault() *Any { + if x != nil { + return x.Default + } + return nil +} + +func (x *Schema) GetMultipleOf() float64 { + if x != nil { + return x.MultipleOf + } + return 0 +} + +func (x *Schema) GetMaximum() float64 { + if x != nil { + return x.Maximum + } + return 0 +} + +func (x *Schema) GetExclusiveMaximum() bool { + if x != nil { + return x.ExclusiveMaximum + } + return false +} + +func (x *Schema) GetMinimum() float64 { + if x != nil { + return x.Minimum + } + return 0 +} + +func (x *Schema) GetExclusiveMinimum() bool { + if x != nil { + return x.ExclusiveMinimum + } + return false +} + +func (x *Schema) GetMaxLength() int64 { + if x != nil { + return x.MaxLength + } + return 0 +} + +func (x *Schema) GetMinLength() int64 { + if x != nil { + return x.MinLength + } + return 0 +} + +func (x *Schema) GetPattern() string { + if x != nil { + return x.Pattern + } + return "" +} + +func (x *Schema) GetMaxItems() int64 { + if x != nil { + return x.MaxItems + } + return 0 +} + +func (x *Schema) GetMinItems() int64 { + if x != nil { + return x.MinItems + } + return 0 +} + +func (x *Schema) GetUniqueItems() bool { + if x != nil { + return x.UniqueItems + } + return false +} + +func (x *Schema) GetMaxProperties() int64 { + if x != nil { + return x.MaxProperties + } + return 0 +} + +func (x *Schema) GetMinProperties() int64 { + if x != nil { + return x.MinProperties + } + return 0 +} + +func (x *Schema) GetRequired() []string { + if x != nil { + return x.Required + } + return nil +} + +func (x *Schema) GetEnum() []*Any { + if x != nil { + return x.Enum + } + return nil +} + +func (x *Schema) GetAdditionalProperties() *AdditionalPropertiesItem { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +func (x *Schema) GetType() *TypeItem { + if x != nil { + return x.Type + } + return nil +} + +func (x *Schema) GetItems() *ItemsItem { + if x != nil { + return x.Items + } + return nil +} + +func (x *Schema) GetAllOf() []*Schema { + if x != nil { + return x.AllOf + } + return nil +} + +func (x *Schema) GetProperties() *Properties { + if x != nil { + return x.Properties + } + return nil +} + +func (x *Schema) GetDiscriminator() string { + if x != nil { + return x.Discriminator + } + return "" +} + +func (x *Schema) GetReadOnly() bool { + if x != nil { + return x.ReadOnly + } + return false +} + +func (x *Schema) GetXml() *Xml { + if x != nil { + return x.Xml + } + return nil +} + +func (x *Schema) GetExternalDocs() *ExternalDocs { + if x != nil { + return x.ExternalDocs + } + return nil +} + +func (x *Schema) GetExample() *Any { + if x != nil { + return x.Example + } + return nil +} + +func (x *Schema) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type SchemaItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Types that are assignable to Oneof: + // *SchemaItem_Schema + // *SchemaItem_FileSchema + Oneof isSchemaItem_Oneof `protobuf_oneof:"oneof"` +} + +func (x *SchemaItem) Reset() { + *x = SchemaItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[51] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *SchemaItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*SchemaItem) ProtoMessage() {} + +func (x *SchemaItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[51] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use SchemaItem.ProtoReflect.Descriptor instead. +func (*SchemaItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{51} +} + +func (m *SchemaItem) GetOneof() isSchemaItem_Oneof { + if m != nil { + return m.Oneof + } + return nil +} + +func (x *SchemaItem) GetSchema() *Schema { + if x, ok := x.GetOneof().(*SchemaItem_Schema); ok { + return x.Schema + } + return nil +} + +func (x *SchemaItem) GetFileSchema() *FileSchema { + if x, ok := x.GetOneof().(*SchemaItem_FileSchema); ok { + return x.FileSchema + } + return nil +} + +type isSchemaItem_Oneof interface { + isSchemaItem_Oneof() +} + +type SchemaItem_Schema struct { + Schema *Schema `protobuf:"bytes,1,opt,name=schema,proto3,oneof"` +} + +type SchemaItem_FileSchema struct { + FileSchema *FileSchema `protobuf:"bytes,2,opt,name=file_schema,json=fileSchema,proto3,oneof"` +} + +func (*SchemaItem_Schema) isSchemaItem_Oneof() {} + +func (*SchemaItem_FileSchema) isSchemaItem_Oneof() {} + +type SecurityDefinitions struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedSecurityDefinitionsItem `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *SecurityDefinitions) Reset() { + *x = SecurityDefinitions{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[52] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *SecurityDefinitions) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*SecurityDefinitions) ProtoMessage() {} + +func (x *SecurityDefinitions) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[52] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use SecurityDefinitions.ProtoReflect.Descriptor instead. +func (*SecurityDefinitions) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{52} +} + +func (x *SecurityDefinitions) GetAdditionalProperties() []*NamedSecurityDefinitionsItem { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type SecurityDefinitionsItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Types that are assignable to Oneof: + // *SecurityDefinitionsItem_BasicAuthenticationSecurity + // *SecurityDefinitionsItem_ApiKeySecurity + // *SecurityDefinitionsItem_Oauth2ImplicitSecurity + // *SecurityDefinitionsItem_Oauth2PasswordSecurity + // *SecurityDefinitionsItem_Oauth2ApplicationSecurity + // *SecurityDefinitionsItem_Oauth2AccessCodeSecurity + Oneof isSecurityDefinitionsItem_Oneof `protobuf_oneof:"oneof"` +} + +func (x *SecurityDefinitionsItem) Reset() { + *x = SecurityDefinitionsItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[53] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *SecurityDefinitionsItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*SecurityDefinitionsItem) ProtoMessage() {} + +func (x *SecurityDefinitionsItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[53] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use SecurityDefinitionsItem.ProtoReflect.Descriptor instead. +func (*SecurityDefinitionsItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{53} +} + +func (m *SecurityDefinitionsItem) GetOneof() isSecurityDefinitionsItem_Oneof { + if m != nil { + return m.Oneof + } + return nil +} + +func (x *SecurityDefinitionsItem) GetBasicAuthenticationSecurity() *BasicAuthenticationSecurity { + if x, ok := x.GetOneof().(*SecurityDefinitionsItem_BasicAuthenticationSecurity); ok { + return x.BasicAuthenticationSecurity + } + return nil +} + +func (x *SecurityDefinitionsItem) GetApiKeySecurity() *ApiKeySecurity { + if x, ok := x.GetOneof().(*SecurityDefinitionsItem_ApiKeySecurity); ok { + return x.ApiKeySecurity + } + return nil +} + +func (x *SecurityDefinitionsItem) GetOauth2ImplicitSecurity() *Oauth2ImplicitSecurity { + if x, ok := x.GetOneof().(*SecurityDefinitionsItem_Oauth2ImplicitSecurity); ok { + return x.Oauth2ImplicitSecurity + } + return nil +} + +func (x *SecurityDefinitionsItem) GetOauth2PasswordSecurity() *Oauth2PasswordSecurity { + if x, ok := x.GetOneof().(*SecurityDefinitionsItem_Oauth2PasswordSecurity); ok { + return x.Oauth2PasswordSecurity + } + return nil +} + +func (x *SecurityDefinitionsItem) GetOauth2ApplicationSecurity() *Oauth2ApplicationSecurity { + if x, ok := x.GetOneof().(*SecurityDefinitionsItem_Oauth2ApplicationSecurity); ok { + return x.Oauth2ApplicationSecurity + } + return nil +} + +func (x *SecurityDefinitionsItem) GetOauth2AccessCodeSecurity() *Oauth2AccessCodeSecurity { + if x, ok := x.GetOneof().(*SecurityDefinitionsItem_Oauth2AccessCodeSecurity); ok { + return x.Oauth2AccessCodeSecurity + } + return nil +} + +type isSecurityDefinitionsItem_Oneof interface { + isSecurityDefinitionsItem_Oneof() +} + +type SecurityDefinitionsItem_BasicAuthenticationSecurity struct { + BasicAuthenticationSecurity *BasicAuthenticationSecurity `protobuf:"bytes,1,opt,name=basic_authentication_security,json=basicAuthenticationSecurity,proto3,oneof"` +} + +type SecurityDefinitionsItem_ApiKeySecurity struct { + ApiKeySecurity *ApiKeySecurity `protobuf:"bytes,2,opt,name=api_key_security,json=apiKeySecurity,proto3,oneof"` +} + +type SecurityDefinitionsItem_Oauth2ImplicitSecurity struct { + Oauth2ImplicitSecurity *Oauth2ImplicitSecurity `protobuf:"bytes,3,opt,name=oauth2_implicit_security,json=oauth2ImplicitSecurity,proto3,oneof"` +} + +type SecurityDefinitionsItem_Oauth2PasswordSecurity struct { + Oauth2PasswordSecurity *Oauth2PasswordSecurity `protobuf:"bytes,4,opt,name=oauth2_password_security,json=oauth2PasswordSecurity,proto3,oneof"` +} + +type SecurityDefinitionsItem_Oauth2ApplicationSecurity struct { + Oauth2ApplicationSecurity *Oauth2ApplicationSecurity `protobuf:"bytes,5,opt,name=oauth2_application_security,json=oauth2ApplicationSecurity,proto3,oneof"` +} + +type SecurityDefinitionsItem_Oauth2AccessCodeSecurity struct { + Oauth2AccessCodeSecurity *Oauth2AccessCodeSecurity `protobuf:"bytes,6,opt,name=oauth2_access_code_security,json=oauth2AccessCodeSecurity,proto3,oneof"` +} + +func (*SecurityDefinitionsItem_BasicAuthenticationSecurity) isSecurityDefinitionsItem_Oneof() {} + +func (*SecurityDefinitionsItem_ApiKeySecurity) isSecurityDefinitionsItem_Oneof() {} + +func (*SecurityDefinitionsItem_Oauth2ImplicitSecurity) isSecurityDefinitionsItem_Oneof() {} + +func (*SecurityDefinitionsItem_Oauth2PasswordSecurity) isSecurityDefinitionsItem_Oneof() {} + +func (*SecurityDefinitionsItem_Oauth2ApplicationSecurity) isSecurityDefinitionsItem_Oneof() {} + +func (*SecurityDefinitionsItem_Oauth2AccessCodeSecurity) isSecurityDefinitionsItem_Oneof() {} + +type SecurityRequirement struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedStringArray `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *SecurityRequirement) Reset() { + *x = SecurityRequirement{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[54] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *SecurityRequirement) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*SecurityRequirement) ProtoMessage() {} + +func (x *SecurityRequirement) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[54] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use SecurityRequirement.ProtoReflect.Descriptor instead. +func (*SecurityRequirement) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{54} +} + +func (x *SecurityRequirement) GetAdditionalProperties() []*NamedStringArray { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type StringArray struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Value []string `protobuf:"bytes,1,rep,name=value,proto3" json:"value,omitempty"` +} + +func (x *StringArray) Reset() { + *x = StringArray{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[55] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *StringArray) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*StringArray) ProtoMessage() {} + +func (x *StringArray) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[55] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use StringArray.ProtoReflect.Descriptor instead. +func (*StringArray) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{55} +} + +func (x *StringArray) GetValue() []string { + if x != nil { + return x.Value + } + return nil +} + +type Tag struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` + ExternalDocs *ExternalDocs `protobuf:"bytes,3,opt,name=external_docs,json=externalDocs,proto3" json:"external_docs,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,4,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Tag) Reset() { + *x = Tag{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[56] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Tag) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Tag) ProtoMessage() {} + +func (x *Tag) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[56] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Tag.ProtoReflect.Descriptor instead. +func (*Tag) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{56} +} + +func (x *Tag) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *Tag) GetDescription() string { + if x != nil { + return x.Description + } + return "" +} + +func (x *Tag) GetExternalDocs() *ExternalDocs { + if x != nil { + return x.ExternalDocs + } + return nil +} + +func (x *Tag) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +type TypeItem struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Value []string `protobuf:"bytes,1,rep,name=value,proto3" json:"value,omitempty"` +} + +func (x *TypeItem) Reset() { + *x = TypeItem{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[57] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *TypeItem) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*TypeItem) ProtoMessage() {} + +func (x *TypeItem) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[57] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use TypeItem.ProtoReflect.Descriptor instead. +func (*TypeItem) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{57} +} + +func (x *TypeItem) GetValue() []string { + if x != nil { + return x.Value + } + return nil +} + +// Any property starting with x- is valid. +type VendorExtension struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + AdditionalProperties []*NamedAny `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties,proto3" json:"additional_properties,omitempty"` +} + +func (x *VendorExtension) Reset() { + *x = VendorExtension{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[58] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *VendorExtension) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*VendorExtension) ProtoMessage() {} + +func (x *VendorExtension) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[58] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use VendorExtension.ProtoReflect.Descriptor instead. +func (*VendorExtension) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{58} +} + +func (x *VendorExtension) GetAdditionalProperties() []*NamedAny { + if x != nil { + return x.AdditionalProperties + } + return nil +} + +type Xml struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + Namespace string `protobuf:"bytes,2,opt,name=namespace,proto3" json:"namespace,omitempty"` + Prefix string `protobuf:"bytes,3,opt,name=prefix,proto3" json:"prefix,omitempty"` + Attribute bool `protobuf:"varint,4,opt,name=attribute,proto3" json:"attribute,omitempty"` + Wrapped bool `protobuf:"varint,5,opt,name=wrapped,proto3" json:"wrapped,omitempty"` + VendorExtension []*NamedAny `protobuf:"bytes,6,rep,name=vendor_extension,json=vendorExtension,proto3" json:"vendor_extension,omitempty"` +} + +func (x *Xml) Reset() { + *x = Xml{} + if protoimpl.UnsafeEnabled { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[59] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Xml) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Xml) ProtoMessage() {} + +func (x *Xml) ProtoReflect() protoreflect.Message { + mi := &file_openapiv2_OpenAPIv2_proto_msgTypes[59] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Xml.ProtoReflect.Descriptor instead. +func (*Xml) Descriptor() ([]byte, []int) { + return file_openapiv2_OpenAPIv2_proto_rawDescGZIP(), []int{59} +} + +func (x *Xml) GetName() string { + if x != nil { + return x.Name + } + return "" +} + +func (x *Xml) GetNamespace() string { + if x != nil { + return x.Namespace + } + return "" +} + +func (x *Xml) GetPrefix() string { + if x != nil { + return x.Prefix + } + return "" +} + +func (x *Xml) GetAttribute() bool { + if x != nil { + return x.Attribute + } + return false +} + +func (x *Xml) GetWrapped() bool { + if x != nil { + return x.Wrapped + } + return false +} + +func (x *Xml) GetVendorExtension() []*NamedAny { + if x != nil { + return x.VendorExtension + } + return nil +} + +var File_openapiv2_OpenAPIv2_proto protoreflect.FileDescriptor + +var file_openapiv2_OpenAPIv2_proto_rawDesc = []byte{ + 0x0a, 0x19, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x76, 0x32, 0x2f, 0x4f, 0x70, 0x65, 0x6e, + 0x41, 0x50, 0x49, 0x76, 0x32, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x0a, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x1a, 0x19, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, + 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x61, 0x6e, 0x79, 0x2e, 0x70, 0x72, 0x6f, + 0x74, 0x6f, 0x22, 0x6d, 0x0a, 0x18, 0x41, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, + 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x12, 0x2c, + 0x0a, 0x06, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x12, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x63, 0x68, 0x65, + 0x6d, 0x61, 0x48, 0x00, 0x52, 0x06, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x1a, 0x0a, 0x07, + 0x62, 0x6f, 0x6f, 0x6c, 0x65, 0x61, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x08, 0x48, 0x00, 0x52, + 0x07, 0x62, 0x6f, 0x6f, 0x6c, 0x65, 0x61, 0x6e, 0x42, 0x07, 0x0a, 0x05, 0x6f, 0x6e, 0x65, 0x6f, + 0x66, 0x22, 0x45, 0x0a, 0x03, 0x41, 0x6e, 0x79, 0x12, 0x2a, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, + 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, + 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x05, 0x76, + 0x61, 0x6c, 0x75, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x79, 0x61, 0x6d, 0x6c, 0x18, 0x02, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x04, 0x79, 0x61, 0x6d, 0x6c, 0x22, 0xab, 0x01, 0x0a, 0x0e, 0x41, 0x70, 0x69, + 0x4b, 0x65, 0x79, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x12, 0x0a, 0x04, 0x74, + 0x79, 0x70, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, + 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, + 0x61, 0x6d, 0x65, 0x12, 0x0e, 0x0a, 0x02, 0x69, 0x6e, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x02, 0x69, 0x6e, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, + 0x6f, 0x6e, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, + 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, + 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x05, 0x20, 0x03, 0x28, 0x0b, 0x32, + 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, + 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, + 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x94, 0x01, 0x0a, 0x1b, 0x42, 0x61, 0x73, 0x69, 0x63, + 0x41, 0x75, 0x74, 0x68, 0x65, 0x6e, 0x74, 0x69, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x53, 0x65, + 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x01, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, + 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3f, 0x0a, 0x10, + 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, + 0x18, 0x03, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, + 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, + 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xde, 0x01, + 0x0a, 0x0d, 0x42, 0x6f, 0x64, 0x79, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x12, + 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x01, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, + 0x6e, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x0e, 0x0a, 0x02, 0x69, 0x6e, 0x18, 0x03, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x02, 0x69, 0x6e, 0x12, 0x1a, 0x0a, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, + 0x64, 0x18, 0x04, 0x20, 0x01, 0x28, 0x08, 0x52, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, + 0x64, 0x12, 0x2a, 0x0a, 0x06, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x18, 0x05, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x12, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, + 0x63, 0x68, 0x65, 0x6d, 0x61, 0x52, 0x06, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x3f, 0x0a, + 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, + 0x6e, 0x18, 0x06, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, + 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, + 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x86, + 0x01, 0x0a, 0x07, 0x43, 0x6f, 0x6e, 0x74, 0x61, 0x63, 0x74, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, + 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x10, + 0x0a, 0x03, 0x75, 0x72, 0x6c, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x75, 0x72, 0x6c, + 0x12, 0x14, 0x0a, 0x05, 0x65, 0x6d, 0x61, 0x69, 0x6c, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x05, 0x65, 0x6d, 0x61, 0x69, 0x6c, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, + 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x04, 0x20, 0x03, 0x28, 0x0b, + 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, + 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, + 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x54, 0x0a, 0x07, 0x44, 0x65, 0x66, 0x61, 0x75, + 0x6c, 0x74, 0x12, 0x49, 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, + 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, + 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, + 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, + 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0x5b, 0x0a, + 0x0b, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x12, 0x4c, 0x0a, 0x15, + 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, + 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x17, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x63, + 0x68, 0x65, 0x6d, 0x61, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, + 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0xe8, 0x05, 0x0a, 0x08, 0x44, + 0x6f, 0x63, 0x75, 0x6d, 0x65, 0x6e, 0x74, 0x12, 0x18, 0x0a, 0x07, 0x73, 0x77, 0x61, 0x67, 0x67, + 0x65, 0x72, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x73, 0x77, 0x61, 0x67, 0x67, 0x65, + 0x72, 0x12, 0x24, 0x0a, 0x04, 0x69, 0x6e, 0x66, 0x6f, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x10, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x49, 0x6e, 0x66, + 0x6f, 0x52, 0x04, 0x69, 0x6e, 0x66, 0x6f, 0x12, 0x12, 0x0a, 0x04, 0x68, 0x6f, 0x73, 0x74, 0x18, + 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x68, 0x6f, 0x73, 0x74, 0x12, 0x1b, 0x0a, 0x09, 0x62, + 0x61, 0x73, 0x65, 0x5f, 0x70, 0x61, 0x74, 0x68, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, + 0x62, 0x61, 0x73, 0x65, 0x50, 0x61, 0x74, 0x68, 0x12, 0x18, 0x0a, 0x07, 0x73, 0x63, 0x68, 0x65, + 0x6d, 0x65, 0x73, 0x18, 0x05, 0x20, 0x03, 0x28, 0x09, 0x52, 0x07, 0x73, 0x63, 0x68, 0x65, 0x6d, + 0x65, 0x73, 0x12, 0x1a, 0x0a, 0x08, 0x63, 0x6f, 0x6e, 0x73, 0x75, 0x6d, 0x65, 0x73, 0x18, 0x06, + 0x20, 0x03, 0x28, 0x09, 0x52, 0x08, 0x63, 0x6f, 0x6e, 0x73, 0x75, 0x6d, 0x65, 0x73, 0x12, 0x1a, + 0x0a, 0x08, 0x70, 0x72, 0x6f, 0x64, 0x75, 0x63, 0x65, 0x73, 0x18, 0x07, 0x20, 0x03, 0x28, 0x09, + 0x52, 0x08, 0x70, 0x72, 0x6f, 0x64, 0x75, 0x63, 0x65, 0x73, 0x12, 0x27, 0x0a, 0x05, 0x70, 0x61, + 0x74, 0x68, 0x73, 0x18, 0x08, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x11, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x61, 0x74, 0x68, 0x73, 0x52, 0x05, 0x70, 0x61, + 0x74, 0x68, 0x73, 0x12, 0x39, 0x0a, 0x0b, 0x64, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, + 0x6e, 0x73, 0x18, 0x09, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x17, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, + 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, + 0x73, 0x52, 0x0b, 0x64, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x12, 0x40, + 0x0a, 0x0a, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x73, 0x18, 0x0a, 0x20, 0x01, + 0x28, 0x0b, 0x32, 0x20, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, + 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, + 0x69, 0x6f, 0x6e, 0x73, 0x52, 0x0a, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x73, + 0x12, 0x3d, 0x0a, 0x09, 0x72, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x73, 0x18, 0x0b, 0x20, + 0x01, 0x28, 0x0b, 0x32, 0x1f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, + 0x69, 0x6f, 0x6e, 0x73, 0x52, 0x09, 0x72, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x73, 0x12, + 0x3b, 0x0a, 0x08, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x18, 0x0c, 0x20, 0x03, 0x28, + 0x0b, 0x32, 0x1f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, + 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x52, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x6d, 0x65, + 0x6e, 0x74, 0x52, 0x08, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x52, 0x0a, 0x14, + 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x5f, 0x64, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, + 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x0d, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1f, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, + 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x52, 0x13, 0x73, 0x65, 0x63, + 0x75, 0x72, 0x69, 0x74, 0x79, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, + 0x12, 0x23, 0x0a, 0x04, 0x74, 0x61, 0x67, 0x73, 0x18, 0x0e, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x0f, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x54, 0x61, 0x67, 0x52, + 0x04, 0x74, 0x61, 0x67, 0x73, 0x12, 0x3d, 0x0a, 0x0d, 0x65, 0x78, 0x74, 0x65, 0x72, 0x6e, 0x61, + 0x6c, 0x5f, 0x64, 0x6f, 0x63, 0x73, 0x18, 0x0f, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x18, 0x2e, 0x6f, + 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x45, 0x78, 0x74, 0x65, 0x72, 0x6e, + 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x52, 0x0c, 0x65, 0x78, 0x74, 0x65, 0x72, 0x6e, 0x61, 0x6c, + 0x44, 0x6f, 0x63, 0x73, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, + 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x10, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, + 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, + 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x55, 0x0a, 0x08, 0x45, 0x78, 0x61, 0x6d, 0x70, 0x6c, 0x65, + 0x73, 0x12, 0x49, 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, + 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, + 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, + 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, + 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0x83, 0x01, 0x0a, + 0x0c, 0x45, 0x78, 0x74, 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x12, 0x20, 0x0a, + 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, + 0x10, 0x0a, 0x03, 0x75, 0x72, 0x6c, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x75, 0x72, + 0x6c, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, + 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x03, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, + 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, + 0x6f, 0x6e, 0x22, 0xff, 0x02, 0x0a, 0x0a, 0x46, 0x69, 0x6c, 0x65, 0x53, 0x63, 0x68, 0x65, 0x6d, + 0x61, 0x12, 0x16, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x01, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x14, 0x0a, 0x05, 0x74, 0x69, 0x74, + 0x6c, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x05, 0x74, 0x69, 0x74, 0x6c, 0x65, 0x12, + 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x03, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, + 0x6e, 0x12, 0x29, 0x0a, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x18, 0x04, 0x20, 0x01, + 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, + 0x41, 0x6e, 0x79, 0x52, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x12, 0x1a, 0x0a, 0x08, + 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x18, 0x05, 0x20, 0x03, 0x28, 0x09, 0x52, 0x08, + 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, + 0x18, 0x06, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x1b, 0x0a, 0x09, + 0x72, 0x65, 0x61, 0x64, 0x5f, 0x6f, 0x6e, 0x6c, 0x79, 0x18, 0x07, 0x20, 0x01, 0x28, 0x08, 0x52, + 0x08, 0x72, 0x65, 0x61, 0x64, 0x4f, 0x6e, 0x6c, 0x79, 0x12, 0x3d, 0x0a, 0x0d, 0x65, 0x78, 0x74, + 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x5f, 0x64, 0x6f, 0x63, 0x73, 0x18, 0x08, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x18, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x45, 0x78, + 0x74, 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x52, 0x0c, 0x65, 0x78, 0x74, 0x65, + 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x12, 0x29, 0x0a, 0x07, 0x65, 0x78, 0x61, 0x6d, + 0x70, 0x6c, 0x65, 0x18, 0x09, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x07, 0x65, 0x78, 0x61, 0x6d, + 0x70, 0x6c, 0x65, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, + 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x0a, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, + 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, + 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xab, 0x06, 0x0a, 0x1a, 0x46, 0x6f, 0x72, 0x6d, 0x44, 0x61, 0x74, + 0x61, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, + 0x65, 0x6d, 0x61, 0x12, 0x1a, 0x0a, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x08, 0x52, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x12, + 0x0e, 0x0a, 0x02, 0x69, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x02, 0x69, 0x6e, 0x12, + 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x03, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, + 0x6e, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x2a, 0x0a, 0x11, 0x61, 0x6c, 0x6c, 0x6f, 0x77, 0x5f, 0x65, + 0x6d, 0x70, 0x74, 0x79, 0x5f, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x05, 0x20, 0x01, 0x28, 0x08, + 0x52, 0x0f, 0x61, 0x6c, 0x6c, 0x6f, 0x77, 0x45, 0x6d, 0x70, 0x74, 0x79, 0x56, 0x61, 0x6c, 0x75, + 0x65, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x06, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, + 0x07, 0x20, 0x01, 0x28, 0x09, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x31, 0x0a, + 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x08, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1b, 0x2e, 0x6f, + 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x72, 0x69, 0x6d, 0x69, 0x74, + 0x69, 0x76, 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x52, 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, + 0x12, 0x2b, 0x0a, 0x11, 0x63, 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x66, + 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x09, 0x20, 0x01, 0x28, 0x09, 0x52, 0x10, 0x63, 0x6f, 0x6c, + 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x29, 0x0a, + 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x18, 0x0a, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, + 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x61, 0x78, 0x69, + 0x6d, 0x75, 0x6d, 0x18, 0x0b, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, + 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, + 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0c, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, + 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, + 0x18, 0x0a, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0d, 0x20, 0x01, 0x28, 0x01, + 0x52, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, + 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0e, + 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, + 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x61, 0x78, 0x5f, 0x6c, 0x65, + 0x6e, 0x67, 0x74, 0x68, 0x18, 0x0f, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x61, 0x78, 0x4c, + 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x69, 0x6e, 0x5f, 0x6c, 0x65, 0x6e, + 0x67, 0x74, 0x68, 0x18, 0x10, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x69, 0x6e, 0x4c, 0x65, + 0x6e, 0x67, 0x74, 0x68, 0x12, 0x18, 0x0a, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x18, + 0x11, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x12, 0x1b, + 0x0a, 0x09, 0x6d, 0x61, 0x78, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x12, 0x20, 0x01, 0x28, + 0x03, 0x52, 0x08, 0x6d, 0x61, 0x78, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x1b, 0x0a, 0x09, 0x6d, + 0x69, 0x6e, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x13, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, + 0x6d, 0x69, 0x6e, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x21, 0x0a, 0x0c, 0x75, 0x6e, 0x69, 0x71, + 0x75, 0x65, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x14, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0b, + 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x23, 0x0a, 0x04, 0x65, + 0x6e, 0x75, 0x6d, 0x18, 0x15, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x04, 0x65, 0x6e, 0x75, 0x6d, + 0x12, 0x1f, 0x0a, 0x0b, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x5f, 0x6f, 0x66, 0x18, + 0x16, 0x20, 0x01, 0x28, 0x01, 0x52, 0x0a, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x4f, + 0x66, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, + 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x17, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, + 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, + 0x6f, 0x6e, 0x22, 0xab, 0x05, 0x0a, 0x06, 0x48, 0x65, 0x61, 0x64, 0x65, 0x72, 0x12, 0x12, 0x0a, + 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, + 0x65, 0x12, 0x16, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x02, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x31, 0x0a, 0x05, 0x69, 0x74, 0x65, + 0x6d, 0x73, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1b, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, + 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x72, 0x69, 0x6d, 0x69, 0x74, 0x69, 0x76, 0x65, 0x73, + 0x49, 0x74, 0x65, 0x6d, 0x73, 0x52, 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x2b, 0x0a, 0x11, + 0x63, 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x66, 0x6f, 0x72, 0x6d, 0x61, + 0x74, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x10, 0x63, 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, + 0x69, 0x6f, 0x6e, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x29, 0x0a, 0x07, 0x64, 0x65, 0x66, + 0x61, 0x75, 0x6c, 0x74, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x07, 0x64, 0x65, 0x66, + 0x61, 0x75, 0x6c, 0x74, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, + 0x06, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, + 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x61, 0x78, 0x69, + 0x6d, 0x75, 0x6d, 0x18, 0x07, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, + 0x73, 0x69, 0x76, 0x65, 0x4d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x18, 0x0a, 0x07, 0x6d, + 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x08, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x69, + 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, + 0x76, 0x65, 0x5f, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x09, 0x20, 0x01, 0x28, 0x08, + 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x69, 0x6e, 0x69, 0x6d, + 0x75, 0x6d, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x61, 0x78, 0x5f, 0x6c, 0x65, 0x6e, 0x67, 0x74, 0x68, + 0x18, 0x0a, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x61, 0x78, 0x4c, 0x65, 0x6e, 0x67, 0x74, + 0x68, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x69, 0x6e, 0x5f, 0x6c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x18, + 0x0b, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x69, 0x6e, 0x4c, 0x65, 0x6e, 0x67, 0x74, 0x68, + 0x12, 0x18, 0x0a, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x18, 0x0c, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x61, + 0x78, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x0d, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, + 0x61, 0x78, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x69, 0x6e, 0x5f, 0x69, + 0x74, 0x65, 0x6d, 0x73, 0x18, 0x0e, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, 0x69, 0x6e, 0x49, + 0x74, 0x65, 0x6d, 0x73, 0x12, 0x21, 0x0a, 0x0c, 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, 0x5f, 0x69, + 0x74, 0x65, 0x6d, 0x73, 0x18, 0x0f, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0b, 0x75, 0x6e, 0x69, 0x71, + 0x75, 0x65, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x23, 0x0a, 0x04, 0x65, 0x6e, 0x75, 0x6d, 0x18, + 0x10, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x04, 0x65, 0x6e, 0x75, 0x6d, 0x12, 0x1f, 0x0a, 0x0b, + 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x5f, 0x6f, 0x66, 0x18, 0x11, 0x20, 0x01, 0x28, + 0x01, 0x52, 0x0a, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x4f, 0x66, 0x12, 0x20, 0x0a, + 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x12, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, + 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, + 0x69, 0x6f, 0x6e, 0x18, 0x13, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, + 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, + 0x22, 0xfd, 0x05, 0x0a, 0x18, 0x48, 0x65, 0x61, 0x64, 0x65, 0x72, 0x50, 0x61, 0x72, 0x61, 0x6d, + 0x65, 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x1a, 0x0a, + 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x18, 0x01, 0x20, 0x01, 0x28, 0x08, 0x52, + 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x12, 0x0e, 0x0a, 0x02, 0x69, 0x6e, 0x18, + 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x02, 0x69, 0x6e, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, + 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, + 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x12, 0x0a, 0x04, 0x6e, + 0x61, 0x6d, 0x65, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, + 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x05, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, + 0x79, 0x70, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x06, 0x20, + 0x01, 0x28, 0x09, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x31, 0x0a, 0x05, 0x69, + 0x74, 0x65, 0x6d, 0x73, 0x18, 0x07, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1b, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x72, 0x69, 0x6d, 0x69, 0x74, 0x69, 0x76, + 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x52, 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x2b, + 0x0a, 0x11, 0x63, 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x66, 0x6f, 0x72, + 0x6d, 0x61, 0x74, 0x18, 0x08, 0x20, 0x01, 0x28, 0x09, 0x52, 0x10, 0x63, 0x6f, 0x6c, 0x6c, 0x65, + 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x29, 0x0a, 0x07, 0x64, + 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x18, 0x09, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, + 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x07, 0x64, + 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, + 0x6d, 0x18, 0x0a, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, + 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x61, + 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0b, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, + 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x18, 0x0a, + 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0c, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, + 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, + 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0d, 0x20, 0x01, + 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x69, 0x6e, + 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x61, 0x78, 0x5f, 0x6c, 0x65, 0x6e, 0x67, + 0x74, 0x68, 0x18, 0x0e, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x61, 0x78, 0x4c, 0x65, 0x6e, + 0x67, 0x74, 0x68, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x69, 0x6e, 0x5f, 0x6c, 0x65, 0x6e, 0x67, 0x74, + 0x68, 0x18, 0x0f, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x69, 0x6e, 0x4c, 0x65, 0x6e, 0x67, + 0x74, 0x68, 0x12, 0x18, 0x0a, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x18, 0x10, 0x20, + 0x01, 0x28, 0x09, 0x52, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x12, 0x1b, 0x0a, 0x09, + 0x6d, 0x61, 0x78, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x11, 0x20, 0x01, 0x28, 0x03, 0x52, + 0x08, 0x6d, 0x61, 0x78, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x69, 0x6e, + 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x12, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, 0x69, + 0x6e, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x21, 0x0a, 0x0c, 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, + 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x13, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0b, 0x75, 0x6e, + 0x69, 0x71, 0x75, 0x65, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x23, 0x0a, 0x04, 0x65, 0x6e, 0x75, + 0x6d, 0x18, 0x14, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, + 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x04, 0x65, 0x6e, 0x75, 0x6d, 0x12, 0x1f, + 0x0a, 0x0b, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x5f, 0x6f, 0x66, 0x18, 0x15, 0x20, + 0x01, 0x28, 0x01, 0x52, 0x0a, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x4f, 0x66, 0x12, + 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, + 0x69, 0x6f, 0x6e, 0x18, 0x16, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, + 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, + 0x22, 0x57, 0x0a, 0x07, 0x48, 0x65, 0x61, 0x64, 0x65, 0x72, 0x73, 0x12, 0x4c, 0x0a, 0x15, 0x61, + 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, + 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x17, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x48, 0x65, 0x61, + 0x64, 0x65, 0x72, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x50, + 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0xa1, 0x02, 0x0a, 0x04, 0x49, 0x6e, + 0x66, 0x6f, 0x12, 0x14, 0x0a, 0x05, 0x74, 0x69, 0x74, 0x6c, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x05, 0x74, 0x69, 0x74, 0x6c, 0x65, 0x12, 0x18, 0x0a, 0x07, 0x76, 0x65, 0x72, 0x73, + 0x69, 0x6f, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x76, 0x65, 0x72, 0x73, 0x69, + 0x6f, 0x6e, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, + 0x6e, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, + 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x28, 0x0a, 0x10, 0x74, 0x65, 0x72, 0x6d, 0x73, 0x5f, 0x6f, 0x66, + 0x5f, 0x73, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0e, + 0x74, 0x65, 0x72, 0x6d, 0x73, 0x4f, 0x66, 0x53, 0x65, 0x72, 0x76, 0x69, 0x63, 0x65, 0x12, 0x2d, + 0x0a, 0x07, 0x63, 0x6f, 0x6e, 0x74, 0x61, 0x63, 0x74, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x13, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x43, 0x6f, 0x6e, + 0x74, 0x61, 0x63, 0x74, 0x52, 0x07, 0x63, 0x6f, 0x6e, 0x74, 0x61, 0x63, 0x74, 0x12, 0x2d, 0x0a, + 0x07, 0x6c, 0x69, 0x63, 0x65, 0x6e, 0x73, 0x65, 0x18, 0x06, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x13, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4c, 0x69, 0x63, 0x65, + 0x6e, 0x73, 0x65, 0x52, 0x07, 0x6c, 0x69, 0x63, 0x65, 0x6e, 0x73, 0x65, 0x12, 0x3f, 0x0a, 0x10, + 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, + 0x18, 0x07, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, + 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, + 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x37, 0x0a, + 0x09, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x12, 0x2a, 0x0a, 0x06, 0x73, 0x63, + 0x68, 0x65, 0x6d, 0x61, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x12, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x52, 0x06, + 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x22, 0x44, 0x0a, 0x0d, 0x4a, 0x73, 0x6f, 0x6e, 0x52, 0x65, + 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x12, 0x11, 0x0a, 0x04, 0x5f, 0x72, 0x65, 0x66, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x52, 0x65, 0x66, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, + 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x22, 0x70, 0x0a, 0x07, + 0x4c, 0x69, 0x63, 0x65, 0x6e, 0x73, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x10, 0x0a, 0x03, 0x75, + 0x72, 0x6c, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x03, 0x75, 0x72, 0x6c, 0x12, 0x3f, 0x0a, + 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, + 0x6e, 0x18, 0x03, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, + 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, + 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x45, + 0x0a, 0x08, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, + 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x25, + 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x05, + 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x4b, 0x0a, 0x0b, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x48, 0x65, + 0x61, 0x64, 0x65, 0x72, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x28, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, + 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x12, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, + 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x48, 0x65, 0x61, 0x64, 0x65, 0x72, 0x52, 0x05, 0x76, 0x61, 0x6c, + 0x75, 0x65, 0x22, 0x51, 0x0a, 0x0e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x50, 0x61, 0x72, 0x61, 0x6d, + 0x65, 0x74, 0x65, 0x72, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x2b, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, + 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, + 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x52, 0x05, + 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x4f, 0x0a, 0x0d, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x50, 0x61, + 0x74, 0x68, 0x49, 0x74, 0x65, 0x6d, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x2a, 0x0a, 0x05, 0x76, 0x61, + 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x61, 0x74, 0x68, 0x49, 0x74, 0x65, 0x6d, 0x52, + 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x4f, 0x0a, 0x0d, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x52, + 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x2a, 0x0a, 0x05, 0x76, + 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, + 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x59, 0x0a, 0x12, 0x4e, 0x61, 0x6d, 0x65, 0x64, + 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x12, 0x0a, + 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, + 0x65, 0x12, 0x2f, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x19, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x52, 0x65, + 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x52, 0x05, 0x76, 0x61, 0x6c, + 0x75, 0x65, 0x22, 0x4b, 0x0a, 0x0b, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x63, 0x68, 0x65, 0x6d, + 0x61, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x28, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, + 0x20, 0x01, 0x28, 0x0b, 0x32, 0x12, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, + 0x32, 0x2e, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, + 0x6d, 0x0a, 0x1c, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, + 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x12, + 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, + 0x61, 0x6d, 0x65, 0x12, 0x39, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, + 0x28, 0x0b, 0x32, 0x23, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, + 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, + 0x6f, 0x6e, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x37, + 0x0a, 0x0b, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x74, 0x72, 0x69, 0x6e, 0x67, 0x12, 0x12, 0x0a, + 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, + 0x65, 0x12, 0x14, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, + 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x55, 0x0a, 0x10, 0x4e, 0x61, 0x6d, 0x65, 0x64, + 0x53, 0x74, 0x72, 0x69, 0x6e, 0x67, 0x41, 0x72, 0x72, 0x61, 0x79, 0x12, 0x12, 0x0a, 0x04, 0x6e, + 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, + 0x2d, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x17, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x74, 0x72, 0x69, + 0x6e, 0x67, 0x41, 0x72, 0x72, 0x61, 0x79, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0xb5, + 0x03, 0x0a, 0x10, 0x4e, 0x6f, 0x6e, 0x42, 0x6f, 0x64, 0x79, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, + 0x74, 0x65, 0x72, 0x12, 0x65, 0x0a, 0x1b, 0x68, 0x65, 0x61, 0x64, 0x65, 0x72, 0x5f, 0x70, 0x61, + 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x5f, 0x73, 0x75, 0x62, 0x5f, 0x73, 0x63, 0x68, 0x65, + 0x6d, 0x61, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x24, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, + 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x48, 0x65, 0x61, 0x64, 0x65, 0x72, 0x50, 0x61, 0x72, 0x61, + 0x6d, 0x65, 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x48, 0x00, + 0x52, 0x18, 0x68, 0x65, 0x61, 0x64, 0x65, 0x72, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, + 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x6c, 0x0a, 0x1e, 0x66, 0x6f, + 0x72, 0x6d, 0x5f, 0x64, 0x61, 0x74, 0x61, 0x5f, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, + 0x72, 0x5f, 0x73, 0x75, 0x62, 0x5f, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x18, 0x02, 0x20, 0x01, + 0x28, 0x0b, 0x32, 0x26, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, + 0x46, 0x6f, 0x72, 0x6d, 0x44, 0x61, 0x74, 0x61, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, + 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x48, 0x00, 0x52, 0x1a, 0x66, 0x6f, + 0x72, 0x6d, 0x44, 0x61, 0x74, 0x61, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x53, + 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x62, 0x0a, 0x1a, 0x71, 0x75, 0x65, 0x72, + 0x79, 0x5f, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x5f, 0x73, 0x75, 0x62, 0x5f, + 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x23, 0x2e, 0x6f, + 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x51, 0x75, 0x65, 0x72, 0x79, 0x50, + 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, + 0x61, 0x48, 0x00, 0x52, 0x17, 0x71, 0x75, 0x65, 0x72, 0x79, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, + 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x5f, 0x0a, 0x19, + 0x70, 0x61, 0x74, 0x68, 0x5f, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x5f, 0x73, + 0x75, 0x62, 0x5f, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x18, 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x22, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x61, 0x74, + 0x68, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, + 0x65, 0x6d, 0x61, 0x48, 0x00, 0x52, 0x16, 0x70, 0x61, 0x74, 0x68, 0x50, 0x61, 0x72, 0x61, 0x6d, + 0x65, 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x42, 0x07, 0x0a, + 0x05, 0x6f, 0x6e, 0x65, 0x6f, 0x66, 0x22, 0xa1, 0x02, 0x0a, 0x18, 0x4f, 0x61, 0x75, 0x74, 0x68, + 0x32, 0x41, 0x63, 0x63, 0x65, 0x73, 0x73, 0x43, 0x6f, 0x64, 0x65, 0x53, 0x65, 0x63, 0x75, 0x72, + 0x69, 0x74, 0x79, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x66, 0x6c, 0x6f, 0x77, 0x18, + 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x66, 0x6c, 0x6f, 0x77, 0x12, 0x30, 0x0a, 0x06, 0x73, + 0x63, 0x6f, 0x70, 0x65, 0x73, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x18, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x53, + 0x63, 0x6f, 0x70, 0x65, 0x73, 0x52, 0x06, 0x73, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x12, 0x2b, 0x0a, + 0x11, 0x61, 0x75, 0x74, 0x68, 0x6f, 0x72, 0x69, 0x7a, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x75, + 0x72, 0x6c, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x10, 0x61, 0x75, 0x74, 0x68, 0x6f, 0x72, + 0x69, 0x7a, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x55, 0x72, 0x6c, 0x12, 0x1b, 0x0a, 0x09, 0x74, 0x6f, + 0x6b, 0x65, 0x6e, 0x5f, 0x75, 0x72, 0x6c, 0x18, 0x05, 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, 0x74, + 0x6f, 0x6b, 0x65, 0x6e, 0x55, 0x72, 0x6c, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, + 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x06, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, + 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, + 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x07, 0x20, + 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, + 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xf5, 0x01, 0x0a, 0x19, 0x4f, + 0x61, 0x75, 0x74, 0x68, 0x32, 0x41, 0x70, 0x70, 0x6c, 0x69, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, + 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, + 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x12, 0x0a, 0x04, + 0x66, 0x6c, 0x6f, 0x77, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x66, 0x6c, 0x6f, 0x77, + 0x12, 0x30, 0x0a, 0x06, 0x73, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x18, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, 0x61, + 0x75, 0x74, 0x68, 0x32, 0x53, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x52, 0x06, 0x73, 0x63, 0x6f, 0x70, + 0x65, 0x73, 0x12, 0x1b, 0x0a, 0x09, 0x74, 0x6f, 0x6b, 0x65, 0x6e, 0x5f, 0x75, 0x72, 0x6c, 0x18, + 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x08, 0x74, 0x6f, 0x6b, 0x65, 0x6e, 0x55, 0x72, 0x6c, 0x12, + 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x05, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, + 0x6e, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, + 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x06, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, + 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, + 0x6f, 0x6e, 0x22, 0x82, 0x02, 0x0a, 0x16, 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x49, 0x6d, 0x70, + 0x6c, 0x69, 0x63, 0x69, 0x74, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x12, 0x0a, + 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, + 0x65, 0x12, 0x12, 0x0a, 0x04, 0x66, 0x6c, 0x6f, 0x77, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x04, 0x66, 0x6c, 0x6f, 0x77, 0x12, 0x30, 0x0a, 0x06, 0x73, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x18, + 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x18, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x53, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x52, + 0x06, 0x73, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x12, 0x2b, 0x0a, 0x11, 0x61, 0x75, 0x74, 0x68, 0x6f, + 0x72, 0x69, 0x7a, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x75, 0x72, 0x6c, 0x18, 0x04, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x10, 0x61, 0x75, 0x74, 0x68, 0x6f, 0x72, 0x69, 0x7a, 0x61, 0x74, 0x69, 0x6f, + 0x6e, 0x55, 0x72, 0x6c, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, + 0x69, 0x6f, 0x6e, 0x18, 0x05, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, + 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, + 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x06, 0x20, 0x03, 0x28, 0x0b, + 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, + 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, + 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xf2, 0x01, 0x0a, 0x16, 0x4f, 0x61, 0x75, 0x74, + 0x68, 0x32, 0x50, 0x61, 0x73, 0x73, 0x77, 0x6f, 0x72, 0x64, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, + 0x74, 0x79, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, + 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x66, 0x6c, 0x6f, 0x77, 0x18, 0x02, + 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x66, 0x6c, 0x6f, 0x77, 0x12, 0x30, 0x0a, 0x06, 0x73, 0x63, + 0x6f, 0x70, 0x65, 0x73, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x18, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x53, 0x63, + 0x6f, 0x70, 0x65, 0x73, 0x52, 0x06, 0x73, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x12, 0x1b, 0x0a, 0x09, + 0x74, 0x6f, 0x6b, 0x65, 0x6e, 0x5f, 0x75, 0x72, 0x6c, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x08, 0x74, 0x6f, 0x6b, 0x65, 0x6e, 0x55, 0x72, 0x6c, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, + 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x05, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, + 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3f, 0x0a, 0x10, 0x76, + 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, + 0x06, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, + 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x5c, 0x0a, 0x0c, + 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x53, 0x63, 0x6f, 0x70, 0x65, 0x73, 0x12, 0x4c, 0x0a, 0x15, + 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, + 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x17, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x74, + 0x72, 0x69, 0x6e, 0x67, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, + 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0x9e, 0x04, 0x0a, 0x09, 0x4f, + 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x61, 0x67, 0x73, + 0x18, 0x01, 0x20, 0x03, 0x28, 0x09, 0x52, 0x04, 0x74, 0x61, 0x67, 0x73, 0x12, 0x18, 0x0a, 0x07, + 0x73, 0x75, 0x6d, 0x6d, 0x61, 0x72, 0x79, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x73, + 0x75, 0x6d, 0x6d, 0x61, 0x72, 0x79, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, + 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, + 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3d, 0x0a, 0x0d, 0x65, 0x78, 0x74, 0x65, + 0x72, 0x6e, 0x61, 0x6c, 0x5f, 0x64, 0x6f, 0x63, 0x73, 0x18, 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x18, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x45, 0x78, 0x74, + 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x52, 0x0c, 0x65, 0x78, 0x74, 0x65, 0x72, + 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x12, 0x21, 0x0a, 0x0c, 0x6f, 0x70, 0x65, 0x72, 0x61, + 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x69, 0x64, 0x18, 0x05, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x6f, + 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x49, 0x64, 0x12, 0x1a, 0x0a, 0x08, 0x70, 0x72, + 0x6f, 0x64, 0x75, 0x63, 0x65, 0x73, 0x18, 0x06, 0x20, 0x03, 0x28, 0x09, 0x52, 0x08, 0x70, 0x72, + 0x6f, 0x64, 0x75, 0x63, 0x65, 0x73, 0x12, 0x1a, 0x0a, 0x08, 0x63, 0x6f, 0x6e, 0x73, 0x75, 0x6d, + 0x65, 0x73, 0x18, 0x07, 0x20, 0x03, 0x28, 0x09, 0x52, 0x08, 0x63, 0x6f, 0x6e, 0x73, 0x75, 0x6d, + 0x65, 0x73, 0x12, 0x3a, 0x0a, 0x0a, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x73, + 0x18, 0x08, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x1a, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, + 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x73, 0x49, 0x74, + 0x65, 0x6d, 0x52, 0x0a, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x73, 0x12, 0x33, + 0x0a, 0x09, 0x72, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x73, 0x18, 0x09, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x52, + 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x73, 0x52, 0x09, 0x72, 0x65, 0x73, 0x70, 0x6f, 0x6e, + 0x73, 0x65, 0x73, 0x12, 0x18, 0x0a, 0x07, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x65, 0x73, 0x18, 0x0a, + 0x20, 0x03, 0x28, 0x09, 0x52, 0x07, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x65, 0x73, 0x12, 0x1e, 0x0a, + 0x0a, 0x64, 0x65, 0x70, 0x72, 0x65, 0x63, 0x61, 0x74, 0x65, 0x64, 0x18, 0x0b, 0x20, 0x01, 0x28, + 0x08, 0x52, 0x0a, 0x64, 0x65, 0x70, 0x72, 0x65, 0x63, 0x61, 0x74, 0x65, 0x64, 0x12, 0x3b, 0x0a, + 0x08, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x18, 0x0c, 0x20, 0x03, 0x28, 0x0b, 0x32, + 0x1f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x65, 0x63, + 0x75, 0x72, 0x69, 0x74, 0x79, 0x52, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x6d, 0x65, 0x6e, 0x74, + 0x52, 0x08, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, + 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x0d, + 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, + 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, + 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xa6, 0x01, 0x0a, 0x09, + 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x12, 0x42, 0x0a, 0x0e, 0x62, 0x6f, 0x64, + 0x79, 0x5f, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x18, 0x01, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x19, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x42, + 0x6f, 0x64, 0x79, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x48, 0x00, 0x52, 0x0d, + 0x62, 0x6f, 0x64, 0x79, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x12, 0x4c, 0x0a, + 0x12, 0x6e, 0x6f, 0x6e, 0x5f, 0x62, 0x6f, 0x64, 0x79, 0x5f, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, + 0x74, 0x65, 0x72, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1c, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x6f, 0x6e, 0x42, 0x6f, 0x64, 0x79, 0x50, 0x61, + 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x48, 0x00, 0x52, 0x10, 0x6e, 0x6f, 0x6e, 0x42, 0x6f, + 0x64, 0x79, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x42, 0x07, 0x0a, 0x05, 0x6f, + 0x6e, 0x65, 0x6f, 0x66, 0x22, 0x67, 0x0a, 0x14, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, + 0x72, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x12, 0x4f, 0x0a, 0x15, + 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, + 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x1a, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x50, 0x61, + 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, + 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0x94, 0x01, + 0x0a, 0x0e, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x73, 0x49, 0x74, 0x65, 0x6d, + 0x12, 0x35, 0x0a, 0x09, 0x70, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x18, 0x01, 0x20, + 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x48, 0x00, 0x52, 0x09, 0x70, 0x61, + 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x12, 0x42, 0x0a, 0x0e, 0x6a, 0x73, 0x6f, 0x6e, 0x5f, + 0x72, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x19, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4a, 0x73, 0x6f, + 0x6e, 0x52, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x48, 0x00, 0x52, 0x0d, 0x6a, 0x73, + 0x6f, 0x6e, 0x52, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x42, 0x07, 0x0a, 0x05, 0x6f, + 0x6e, 0x65, 0x6f, 0x66, 0x22, 0xcf, 0x03, 0x0a, 0x08, 0x50, 0x61, 0x74, 0x68, 0x49, 0x74, 0x65, + 0x6d, 0x12, 0x11, 0x0a, 0x04, 0x5f, 0x72, 0x65, 0x66, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x03, 0x52, 0x65, 0x66, 0x12, 0x27, 0x0a, 0x03, 0x67, 0x65, 0x74, 0x18, 0x02, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, + 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x03, 0x67, 0x65, 0x74, 0x12, 0x27, 0x0a, + 0x03, 0x70, 0x75, 0x74, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, + 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, + 0x6e, 0x52, 0x03, 0x70, 0x75, 0x74, 0x12, 0x29, 0x0a, 0x04, 0x70, 0x6f, 0x73, 0x74, 0x18, 0x04, + 0x20, 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, + 0x32, 0x2e, 0x4f, 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x04, 0x70, 0x6f, 0x73, + 0x74, 0x12, 0x2d, 0x0a, 0x06, 0x64, 0x65, 0x6c, 0x65, 0x74, 0x65, 0x18, 0x05, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, + 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x06, 0x64, 0x65, 0x6c, 0x65, 0x74, 0x65, + 0x12, 0x2f, 0x0a, 0x07, 0x6f, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x18, 0x06, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, + 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x07, 0x6f, 0x70, 0x74, 0x69, 0x6f, 0x6e, + 0x73, 0x12, 0x29, 0x0a, 0x04, 0x68, 0x65, 0x61, 0x64, 0x18, 0x07, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, 0x70, 0x65, + 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x04, 0x68, 0x65, 0x61, 0x64, 0x12, 0x2b, 0x0a, 0x05, + 0x70, 0x61, 0x74, 0x63, 0x68, 0x18, 0x08, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, + 0x6f, 0x6e, 0x52, 0x05, 0x70, 0x61, 0x74, 0x63, 0x68, 0x12, 0x3a, 0x0a, 0x0a, 0x70, 0x61, 0x72, + 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x73, 0x18, 0x09, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x1a, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x61, 0x72, 0x61, 0x6d, + 0x65, 0x74, 0x65, 0x72, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x52, 0x0a, 0x70, 0x61, 0x72, 0x61, 0x6d, + 0x65, 0x74, 0x65, 0x72, 0x73, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, + 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x0a, 0x20, 0x03, 0x28, 0x0b, 0x32, + 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, + 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, + 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xfb, 0x05, 0x0a, 0x16, 0x50, 0x61, 0x74, 0x68, 0x50, + 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, + 0x61, 0x12, 0x1a, 0x0a, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x18, 0x01, 0x20, + 0x01, 0x28, 0x08, 0x52, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x12, 0x0e, 0x0a, + 0x02, 0x69, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x02, 0x69, 0x6e, 0x12, 0x20, 0x0a, + 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x03, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, + 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, + 0x61, 0x6d, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x05, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, + 0x74, 0x18, 0x06, 0x20, 0x01, 0x28, 0x09, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, + 0x31, 0x0a, 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x07, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1b, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x72, 0x69, 0x6d, + 0x69, 0x74, 0x69, 0x76, 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x52, 0x05, 0x69, 0x74, 0x65, + 0x6d, 0x73, 0x12, 0x2b, 0x0a, 0x11, 0x63, 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, + 0x5f, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x08, 0x20, 0x01, 0x28, 0x09, 0x52, 0x10, 0x63, + 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, + 0x29, 0x0a, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x18, 0x09, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, + 0x79, 0x52, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x61, + 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0a, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x61, 0x78, + 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, + 0x65, 0x5f, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0b, 0x20, 0x01, 0x28, 0x08, 0x52, + 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x61, 0x78, 0x69, 0x6d, 0x75, + 0x6d, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0c, 0x20, 0x01, + 0x28, 0x01, 0x52, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, + 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, + 0x18, 0x0d, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, + 0x65, 0x4d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x61, 0x78, 0x5f, + 0x6c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x18, 0x0e, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x61, + 0x78, 0x4c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x69, 0x6e, 0x5f, 0x6c, + 0x65, 0x6e, 0x67, 0x74, 0x68, 0x18, 0x0f, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x69, 0x6e, + 0x4c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, 0x18, 0x0a, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, + 0x6e, 0x18, 0x10, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, + 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x61, 0x78, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x11, 0x20, + 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, 0x61, 0x78, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x1b, 0x0a, + 0x09, 0x6d, 0x69, 0x6e, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x12, 0x20, 0x01, 0x28, 0x03, + 0x52, 0x08, 0x6d, 0x69, 0x6e, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x21, 0x0a, 0x0c, 0x75, 0x6e, + 0x69, 0x71, 0x75, 0x65, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x13, 0x20, 0x01, 0x28, 0x08, + 0x52, 0x0b, 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x23, 0x0a, + 0x04, 0x65, 0x6e, 0x75, 0x6d, 0x18, 0x14, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x04, 0x65, 0x6e, + 0x75, 0x6d, 0x12, 0x1f, 0x0a, 0x0b, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x5f, 0x6f, + 0x66, 0x18, 0x15, 0x20, 0x01, 0x28, 0x01, 0x52, 0x0a, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, + 0x65, 0x4f, 0x66, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, + 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x16, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, + 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, + 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x77, 0x0a, 0x05, 0x50, 0x61, 0x74, 0x68, 0x73, 0x12, 0x3f, 0x0a, + 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, + 0x6e, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, + 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, + 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x2d, + 0x0a, 0x04, 0x70, 0x61, 0x74, 0x68, 0x18, 0x02, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x19, 0x2e, 0x6f, + 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x50, + 0x61, 0x74, 0x68, 0x49, 0x74, 0x65, 0x6d, 0x52, 0x04, 0x70, 0x61, 0x74, 0x68, 0x22, 0x92, 0x05, + 0x0a, 0x0f, 0x50, 0x72, 0x69, 0x6d, 0x69, 0x74, 0x69, 0x76, 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, + 0x73, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, + 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x31, 0x0a, + 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1b, 0x2e, 0x6f, + 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x50, 0x72, 0x69, 0x6d, 0x69, 0x74, + 0x69, 0x76, 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x52, 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, + 0x12, 0x2b, 0x0a, 0x11, 0x63, 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x66, + 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x10, 0x63, 0x6f, 0x6c, + 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x46, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x29, 0x0a, + 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, + 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, + 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x61, 0x78, 0x69, + 0x6d, 0x75, 0x6d, 0x18, 0x06, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, + 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, + 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x07, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, + 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, + 0x18, 0x0a, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x08, 0x20, 0x01, 0x28, 0x01, + 0x52, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, + 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x09, + 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, + 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x61, 0x78, 0x5f, 0x6c, 0x65, + 0x6e, 0x67, 0x74, 0x68, 0x18, 0x0a, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x61, 0x78, 0x4c, + 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x69, 0x6e, 0x5f, 0x6c, 0x65, 0x6e, + 0x67, 0x74, 0x68, 0x18, 0x0b, 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x69, 0x6e, 0x4c, 0x65, + 0x6e, 0x67, 0x74, 0x68, 0x12, 0x18, 0x0a, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x18, + 0x0c, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x12, 0x1b, + 0x0a, 0x09, 0x6d, 0x61, 0x78, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x0d, 0x20, 0x01, 0x28, + 0x03, 0x52, 0x08, 0x6d, 0x61, 0x78, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x1b, 0x0a, 0x09, 0x6d, + 0x69, 0x6e, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x0e, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, + 0x6d, 0x69, 0x6e, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x21, 0x0a, 0x0c, 0x75, 0x6e, 0x69, 0x71, + 0x75, 0x65, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x0f, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0b, + 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x23, 0x0a, 0x04, 0x65, + 0x6e, 0x75, 0x6d, 0x18, 0x10, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x04, 0x65, 0x6e, 0x75, 0x6d, + 0x12, 0x1f, 0x0a, 0x0b, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x5f, 0x6f, 0x66, 0x18, + 0x11, 0x20, 0x01, 0x28, 0x01, 0x52, 0x0a, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x4f, + 0x66, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, + 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x12, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, + 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, + 0x6f, 0x6e, 0x22, 0x5a, 0x0a, 0x0a, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, + 0x12, 0x4c, 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, 0x70, + 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, + 0x17, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, + 0x65, 0x64, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, + 0x6f, 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0xa8, + 0x06, 0x0a, 0x17, 0x51, 0x75, 0x65, 0x72, 0x79, 0x50, 0x61, 0x72, 0x61, 0x6d, 0x65, 0x74, 0x65, + 0x72, 0x53, 0x75, 0x62, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x1a, 0x0a, 0x08, 0x72, 0x65, + 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x18, 0x01, 0x20, 0x01, 0x28, 0x08, 0x52, 0x08, 0x72, 0x65, + 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x12, 0x0e, 0x0a, 0x02, 0x69, 0x6e, 0x18, 0x02, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x02, 0x69, 0x6e, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, + 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, + 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, + 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x2a, 0x0a, 0x11, + 0x61, 0x6c, 0x6c, 0x6f, 0x77, 0x5f, 0x65, 0x6d, 0x70, 0x74, 0x79, 0x5f, 0x76, 0x61, 0x6c, 0x75, + 0x65, 0x18, 0x05, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0f, 0x61, 0x6c, 0x6c, 0x6f, 0x77, 0x45, 0x6d, + 0x70, 0x74, 0x79, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x12, 0x12, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, + 0x18, 0x06, 0x20, 0x01, 0x28, 0x09, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x16, 0x0a, 0x06, + 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x07, 0x20, 0x01, 0x28, 0x09, 0x52, 0x06, 0x66, 0x6f, + 0x72, 0x6d, 0x61, 0x74, 0x12, 0x31, 0x0a, 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x08, 0x20, + 0x01, 0x28, 0x0b, 0x32, 0x1b, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x50, 0x72, 0x69, 0x6d, 0x69, 0x74, 0x69, 0x76, 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x73, + 0x52, 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x2b, 0x0a, 0x11, 0x63, 0x6f, 0x6c, 0x6c, 0x65, + 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x18, 0x09, 0x20, 0x01, + 0x28, 0x09, 0x52, 0x10, 0x63, 0x6f, 0x6c, 0x6c, 0x65, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x46, 0x6f, + 0x72, 0x6d, 0x61, 0x74, 0x12, 0x29, 0x0a, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x18, + 0x0a, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, 0x74, 0x12, + 0x18, 0x0a, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0b, 0x20, 0x01, 0x28, 0x01, + 0x52, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, + 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0c, + 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, + 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, + 0x6d, 0x18, 0x0d, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, + 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x69, + 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0e, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, + 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x1d, 0x0a, + 0x0a, 0x6d, 0x61, 0x78, 0x5f, 0x6c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x18, 0x0f, 0x20, 0x01, 0x28, + 0x03, 0x52, 0x09, 0x6d, 0x61, 0x78, 0x4c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, 0x1d, 0x0a, 0x0a, + 0x6d, 0x69, 0x6e, 0x5f, 0x6c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x18, 0x10, 0x20, 0x01, 0x28, 0x03, + 0x52, 0x09, 0x6d, 0x69, 0x6e, 0x4c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, 0x18, 0x0a, 0x07, 0x70, + 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x18, 0x11, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x70, 0x61, + 0x74, 0x74, 0x65, 0x72, 0x6e, 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x61, 0x78, 0x5f, 0x69, 0x74, 0x65, + 0x6d, 0x73, 0x18, 0x12, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, 0x61, 0x78, 0x49, 0x74, 0x65, + 0x6d, 0x73, 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x69, 0x6e, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, + 0x13, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, 0x69, 0x6e, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, + 0x21, 0x0a, 0x0c, 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, 0x5f, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, + 0x14, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0b, 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, 0x49, 0x74, 0x65, + 0x6d, 0x73, 0x12, 0x23, 0x0a, 0x04, 0x65, 0x6e, 0x75, 0x6d, 0x18, 0x15, 0x20, 0x03, 0x28, 0x0b, + 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, + 0x79, 0x52, 0x04, 0x65, 0x6e, 0x75, 0x6d, 0x12, 0x1f, 0x0a, 0x0b, 0x6d, 0x75, 0x6c, 0x74, 0x69, + 0x70, 0x6c, 0x65, 0x5f, 0x6f, 0x66, 0x18, 0x16, 0x20, 0x01, 0x28, 0x01, 0x52, 0x0a, 0x6d, 0x75, + 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x4f, 0x66, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, + 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x17, 0x20, 0x03, + 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, + 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, + 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xfe, 0x01, 0x0a, 0x08, 0x52, 0x65, + 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, + 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, + 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x2e, 0x0a, 0x06, 0x73, 0x63, 0x68, 0x65, + 0x6d, 0x61, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x16, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, + 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x49, 0x74, 0x65, 0x6d, + 0x52, 0x06, 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x2d, 0x0a, 0x07, 0x68, 0x65, 0x61, 0x64, + 0x65, 0x72, 0x73, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x13, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x48, 0x65, 0x61, 0x64, 0x65, 0x72, 0x73, 0x52, 0x07, + 0x68, 0x65, 0x61, 0x64, 0x65, 0x72, 0x73, 0x12, 0x30, 0x0a, 0x08, 0x65, 0x78, 0x61, 0x6d, 0x70, + 0x6c, 0x65, 0x73, 0x18, 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x45, 0x78, 0x61, 0x6d, 0x70, 0x6c, 0x65, 0x73, 0x52, + 0x08, 0x65, 0x78, 0x61, 0x6d, 0x70, 0x6c, 0x65, 0x73, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, + 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x05, 0x20, + 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, + 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x65, 0x0a, 0x13, 0x52, 0x65, + 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, + 0x73, 0x12, 0x4e, 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, + 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, + 0x32, 0x19, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, + 0x6d, 0x65, 0x64, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x52, 0x14, 0x61, 0x64, 0x64, + 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, + 0x73, 0x22, 0x90, 0x01, 0x0a, 0x0d, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x56, 0x61, + 0x6c, 0x75, 0x65, 0x12, 0x32, 0x0a, 0x08, 0x72, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x48, 0x00, 0x52, 0x08, 0x72, + 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x12, 0x42, 0x0a, 0x0e, 0x6a, 0x73, 0x6f, 0x6e, 0x5f, + 0x72, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x19, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4a, 0x73, 0x6f, + 0x6e, 0x52, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x48, 0x00, 0x52, 0x0d, 0x6a, 0x73, + 0x6f, 0x6e, 0x52, 0x65, 0x66, 0x65, 0x72, 0x65, 0x6e, 0x63, 0x65, 0x42, 0x07, 0x0a, 0x05, 0x6f, + 0x6e, 0x65, 0x6f, 0x66, 0x22, 0x91, 0x01, 0x0a, 0x09, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, + 0x65, 0x73, 0x12, 0x43, 0x0a, 0x0d, 0x72, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x5f, 0x63, + 0x6f, 0x64, 0x65, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x1e, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x52, 0x65, 0x73, 0x70, + 0x6f, 0x6e, 0x73, 0x65, 0x56, 0x61, 0x6c, 0x75, 0x65, 0x52, 0x0c, 0x72, 0x65, 0x73, 0x70, 0x6f, + 0x6e, 0x73, 0x65, 0x43, 0x6f, 0x64, 0x65, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, + 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x02, 0x20, 0x03, 0x28, + 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, + 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, + 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0xaf, 0x09, 0x0a, 0x06, 0x53, 0x63, 0x68, + 0x65, 0x6d, 0x61, 0x12, 0x11, 0x0a, 0x04, 0x5f, 0x72, 0x65, 0x66, 0x18, 0x01, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x03, 0x52, 0x65, 0x66, 0x12, 0x16, 0x0a, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, + 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x06, 0x66, 0x6f, 0x72, 0x6d, 0x61, 0x74, 0x12, 0x14, + 0x0a, 0x05, 0x74, 0x69, 0x74, 0x6c, 0x65, 0x18, 0x03, 0x20, 0x01, 0x28, 0x09, 0x52, 0x05, 0x74, + 0x69, 0x74, 0x6c, 0x65, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, + 0x69, 0x6f, 0x6e, 0x18, 0x04, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, + 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x29, 0x0a, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, + 0x74, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, + 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x07, 0x64, 0x65, 0x66, 0x61, 0x75, 0x6c, + 0x74, 0x12, 0x1f, 0x0a, 0x0b, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, 0x5f, 0x6f, 0x66, + 0x18, 0x06, 0x20, 0x01, 0x28, 0x01, 0x52, 0x0a, 0x6d, 0x75, 0x6c, 0x74, 0x69, 0x70, 0x6c, 0x65, + 0x4f, 0x66, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x07, 0x20, + 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, + 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x5f, 0x6d, 0x61, 0x78, 0x69, 0x6d, 0x75, + 0x6d, 0x18, 0x08, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, + 0x76, 0x65, 0x4d, 0x61, 0x78, 0x69, 0x6d, 0x75, 0x6d, 0x12, 0x18, 0x0a, 0x07, 0x6d, 0x69, 0x6e, + 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x09, 0x20, 0x01, 0x28, 0x01, 0x52, 0x07, 0x6d, 0x69, 0x6e, 0x69, + 0x6d, 0x75, 0x6d, 0x12, 0x2b, 0x0a, 0x11, 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, + 0x5f, 0x6d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, 0x18, 0x0a, 0x20, 0x01, 0x28, 0x08, 0x52, 0x10, + 0x65, 0x78, 0x63, 0x6c, 0x75, 0x73, 0x69, 0x76, 0x65, 0x4d, 0x69, 0x6e, 0x69, 0x6d, 0x75, 0x6d, + 0x12, 0x1d, 0x0a, 0x0a, 0x6d, 0x61, 0x78, 0x5f, 0x6c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x18, 0x0b, + 0x20, 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x61, 0x78, 0x4c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, + 0x1d, 0x0a, 0x0a, 0x6d, 0x69, 0x6e, 0x5f, 0x6c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x18, 0x0c, 0x20, + 0x01, 0x28, 0x03, 0x52, 0x09, 0x6d, 0x69, 0x6e, 0x4c, 0x65, 0x6e, 0x67, 0x74, 0x68, 0x12, 0x18, + 0x0a, 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x18, 0x0d, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x07, 0x70, 0x61, 0x74, 0x74, 0x65, 0x72, 0x6e, 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x61, 0x78, 0x5f, + 0x69, 0x74, 0x65, 0x6d, 0x73, 0x18, 0x0e, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, 0x61, 0x78, + 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x1b, 0x0a, 0x09, 0x6d, 0x69, 0x6e, 0x5f, 0x69, 0x74, 0x65, + 0x6d, 0x73, 0x18, 0x0f, 0x20, 0x01, 0x28, 0x03, 0x52, 0x08, 0x6d, 0x69, 0x6e, 0x49, 0x74, 0x65, + 0x6d, 0x73, 0x12, 0x21, 0x0a, 0x0c, 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, 0x5f, 0x69, 0x74, 0x65, + 0x6d, 0x73, 0x18, 0x10, 0x20, 0x01, 0x28, 0x08, 0x52, 0x0b, 0x75, 0x6e, 0x69, 0x71, 0x75, 0x65, + 0x49, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x25, 0x0a, 0x0e, 0x6d, 0x61, 0x78, 0x5f, 0x70, 0x72, 0x6f, + 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x11, 0x20, 0x01, 0x28, 0x03, 0x52, 0x0d, 0x6d, + 0x61, 0x78, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x12, 0x25, 0x0a, 0x0e, + 0x6d, 0x69, 0x6e, 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x12, + 0x20, 0x01, 0x28, 0x03, 0x52, 0x0d, 0x6d, 0x69, 0x6e, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, + 0x69, 0x65, 0x73, 0x12, 0x1a, 0x0a, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x18, + 0x13, 0x20, 0x03, 0x28, 0x09, 0x52, 0x08, 0x72, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x64, 0x12, + 0x23, 0x0a, 0x04, 0x65, 0x6e, 0x75, 0x6d, 0x18, 0x14, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x0f, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, 0x52, 0x04, + 0x65, 0x6e, 0x75, 0x6d, 0x12, 0x59, 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, + 0x61, 0x6c, 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x15, 0x20, + 0x01, 0x28, 0x0b, 0x32, 0x24, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x41, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, + 0x72, 0x74, 0x69, 0x65, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, + 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x12, + 0x28, 0x0a, 0x04, 0x74, 0x79, 0x70, 0x65, 0x18, 0x16, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x14, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x54, 0x79, 0x70, 0x65, 0x49, + 0x74, 0x65, 0x6d, 0x52, 0x04, 0x74, 0x79, 0x70, 0x65, 0x12, 0x2b, 0x0a, 0x05, 0x69, 0x74, 0x65, + 0x6d, 0x73, 0x18, 0x17, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x15, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, + 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x49, 0x74, 0x65, 0x6d, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x52, + 0x05, 0x69, 0x74, 0x65, 0x6d, 0x73, 0x12, 0x29, 0x0a, 0x06, 0x61, 0x6c, 0x6c, 0x5f, 0x6f, 0x66, + 0x18, 0x18, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x12, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, + 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x52, 0x05, 0x61, 0x6c, 0x6c, 0x4f, + 0x66, 0x12, 0x36, 0x0a, 0x0a, 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, + 0x19, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x16, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x52, 0x0a, 0x70, + 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x12, 0x24, 0x0a, 0x0d, 0x64, 0x69, 0x73, + 0x63, 0x72, 0x69, 0x6d, 0x69, 0x6e, 0x61, 0x74, 0x6f, 0x72, 0x18, 0x1a, 0x20, 0x01, 0x28, 0x09, + 0x52, 0x0d, 0x64, 0x69, 0x73, 0x63, 0x72, 0x69, 0x6d, 0x69, 0x6e, 0x61, 0x74, 0x6f, 0x72, 0x12, + 0x1b, 0x0a, 0x09, 0x72, 0x65, 0x61, 0x64, 0x5f, 0x6f, 0x6e, 0x6c, 0x79, 0x18, 0x1b, 0x20, 0x01, + 0x28, 0x08, 0x52, 0x08, 0x72, 0x65, 0x61, 0x64, 0x4f, 0x6e, 0x6c, 0x79, 0x12, 0x21, 0x0a, 0x03, + 0x78, 0x6d, 0x6c, 0x18, 0x1c, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x58, 0x6d, 0x6c, 0x52, 0x03, 0x78, 0x6d, 0x6c, 0x12, + 0x3d, 0x0a, 0x0d, 0x65, 0x78, 0x74, 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x5f, 0x64, 0x6f, 0x63, 0x73, + 0x18, 0x1d, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x18, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, + 0x2e, 0x76, 0x32, 0x2e, 0x45, 0x78, 0x74, 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, + 0x52, 0x0c, 0x65, 0x78, 0x74, 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x12, 0x29, + 0x0a, 0x07, 0x65, 0x78, 0x61, 0x6d, 0x70, 0x6c, 0x65, 0x18, 0x1e, 0x20, 0x01, 0x28, 0x0b, 0x32, + 0x0f, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x41, 0x6e, 0x79, + 0x52, 0x07, 0x65, 0x78, 0x61, 0x6d, 0x70, 0x6c, 0x65, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, + 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x1f, 0x20, + 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, + 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x7e, 0x0a, 0x0a, 0x53, 0x63, + 0x68, 0x65, 0x6d, 0x61, 0x49, 0x74, 0x65, 0x6d, 0x12, 0x2c, 0x0a, 0x06, 0x73, 0x63, 0x68, 0x65, + 0x6d, 0x61, 0x18, 0x01, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x12, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, + 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x53, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x48, 0x00, 0x52, 0x06, + 0x73, 0x63, 0x68, 0x65, 0x6d, 0x61, 0x12, 0x39, 0x0a, 0x0b, 0x66, 0x69, 0x6c, 0x65, 0x5f, 0x73, + 0x63, 0x68, 0x65, 0x6d, 0x61, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x16, 0x2e, 0x6f, 0x70, + 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x46, 0x69, 0x6c, 0x65, 0x53, 0x63, 0x68, + 0x65, 0x6d, 0x61, 0x48, 0x00, 0x52, 0x0a, 0x66, 0x69, 0x6c, 0x65, 0x53, 0x63, 0x68, 0x65, 0x6d, + 0x61, 0x42, 0x07, 0x0a, 0x05, 0x6f, 0x6e, 0x65, 0x6f, 0x66, 0x22, 0x74, 0x0a, 0x13, 0x53, 0x65, + 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x44, 0x65, 0x66, 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, + 0x73, 0x12, 0x5d, 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, + 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, + 0x32, 0x28, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, + 0x6d, 0x65, 0x64, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x44, 0x65, 0x66, 0x69, 0x6e, + 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, + 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, + 0x22, 0xe9, 0x04, 0x0a, 0x17, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x44, 0x65, 0x66, + 0x69, 0x6e, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x49, 0x74, 0x65, 0x6d, 0x12, 0x6d, 0x0a, 0x1d, + 0x62, 0x61, 0x73, 0x69, 0x63, 0x5f, 0x61, 0x75, 0x74, 0x68, 0x65, 0x6e, 0x74, 0x69, 0x63, 0x61, + 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x18, 0x01, 0x20, + 0x01, 0x28, 0x0b, 0x32, 0x27, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, + 0x2e, 0x42, 0x61, 0x73, 0x69, 0x63, 0x41, 0x75, 0x74, 0x68, 0x65, 0x6e, 0x74, 0x69, 0x63, 0x61, + 0x74, 0x69, 0x6f, 0x6e, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x48, 0x00, 0x52, 0x1b, + 0x62, 0x61, 0x73, 0x69, 0x63, 0x41, 0x75, 0x74, 0x68, 0x65, 0x6e, 0x74, 0x69, 0x63, 0x61, 0x74, + 0x69, 0x6f, 0x6e, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x46, 0x0a, 0x10, 0x61, + 0x70, 0x69, 0x5f, 0x6b, 0x65, 0x79, 0x5f, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x18, + 0x02, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x1a, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x41, 0x70, 0x69, 0x4b, 0x65, 0x79, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, + 0x79, 0x48, 0x00, 0x52, 0x0e, 0x61, 0x70, 0x69, 0x4b, 0x65, 0x79, 0x53, 0x65, 0x63, 0x75, 0x72, + 0x69, 0x74, 0x79, 0x12, 0x5e, 0x0a, 0x18, 0x6f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x5f, 0x69, 0x6d, + 0x70, 0x6c, 0x69, 0x63, 0x69, 0x74, 0x5f, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x18, + 0x03, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x49, 0x6d, 0x70, 0x6c, 0x69, 0x63, 0x69, + 0x74, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x48, 0x00, 0x52, 0x16, 0x6f, 0x61, 0x75, + 0x74, 0x68, 0x32, 0x49, 0x6d, 0x70, 0x6c, 0x69, 0x63, 0x69, 0x74, 0x53, 0x65, 0x63, 0x75, 0x72, + 0x69, 0x74, 0x79, 0x12, 0x5e, 0x0a, 0x18, 0x6f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x5f, 0x70, 0x61, + 0x73, 0x73, 0x77, 0x6f, 0x72, 0x64, 0x5f, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x18, + 0x04, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x22, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, + 0x76, 0x32, 0x2e, 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x50, 0x61, 0x73, 0x73, 0x77, 0x6f, 0x72, + 0x64, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x48, 0x00, 0x52, 0x16, 0x6f, 0x61, 0x75, + 0x74, 0x68, 0x32, 0x50, 0x61, 0x73, 0x73, 0x77, 0x6f, 0x72, 0x64, 0x53, 0x65, 0x63, 0x75, 0x72, + 0x69, 0x74, 0x79, 0x12, 0x67, 0x0a, 0x1b, 0x6f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x5f, 0x61, 0x70, + 0x70, 0x6c, 0x69, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, + 0x74, 0x79, 0x18, 0x05, 0x20, 0x01, 0x28, 0x0b, 0x32, 0x25, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, + 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x41, 0x70, 0x70, 0x6c, + 0x69, 0x63, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x48, + 0x00, 0x52, 0x19, 0x6f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x41, 0x70, 0x70, 0x6c, 0x69, 0x63, 0x61, + 0x74, 0x69, 0x6f, 0x6e, 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x12, 0x65, 0x0a, 0x1b, + 0x6f, 0x61, 0x75, 0x74, 0x68, 0x32, 0x5f, 0x61, 0x63, 0x63, 0x65, 0x73, 0x73, 0x5f, 0x63, 0x6f, + 0x64, 0x65, 0x5f, 0x73, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x18, 0x06, 0x20, 0x01, 0x28, + 0x0b, 0x32, 0x24, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4f, + 0x61, 0x75, 0x74, 0x68, 0x32, 0x41, 0x63, 0x63, 0x65, 0x73, 0x73, 0x43, 0x6f, 0x64, 0x65, 0x53, + 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x48, 0x00, 0x52, 0x18, 0x6f, 0x61, 0x75, 0x74, 0x68, + 0x32, 0x41, 0x63, 0x63, 0x65, 0x73, 0x73, 0x43, 0x6f, 0x64, 0x65, 0x53, 0x65, 0x63, 0x75, 0x72, + 0x69, 0x74, 0x79, 0x42, 0x07, 0x0a, 0x05, 0x6f, 0x6e, 0x65, 0x6f, 0x66, 0x22, 0x68, 0x0a, 0x13, + 0x53, 0x65, 0x63, 0x75, 0x72, 0x69, 0x74, 0x79, 0x52, 0x65, 0x71, 0x75, 0x69, 0x72, 0x65, 0x6d, + 0x65, 0x6e, 0x74, 0x12, 0x51, 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, + 0x6c, 0x5f, 0x70, 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, + 0x28, 0x0b, 0x32, 0x1c, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, + 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x53, 0x74, 0x72, 0x69, 0x6e, 0x67, 0x41, 0x72, 0x72, 0x61, 0x79, + 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x50, 0x72, 0x6f, 0x70, + 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0x23, 0x0a, 0x0b, 0x53, 0x74, 0x72, 0x69, 0x6e, 0x67, + 0x41, 0x72, 0x72, 0x61, 0x79, 0x12, 0x14, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x01, + 0x20, 0x03, 0x28, 0x09, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0xbb, 0x01, 0x0a, 0x03, + 0x54, 0x61, 0x67, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, + 0x09, 0x52, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x20, 0x0a, 0x0b, 0x64, 0x65, 0x73, 0x63, 0x72, + 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x0b, 0x64, 0x65, + 0x73, 0x63, 0x72, 0x69, 0x70, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x3d, 0x0a, 0x0d, 0x65, 0x78, 0x74, + 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x5f, 0x64, 0x6f, 0x63, 0x73, 0x18, 0x03, 0x20, 0x01, 0x28, 0x0b, + 0x32, 0x18, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x45, 0x78, + 0x74, 0x65, 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x52, 0x0c, 0x65, 0x78, 0x74, 0x65, + 0x72, 0x6e, 0x61, 0x6c, 0x44, 0x6f, 0x63, 0x73, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, + 0x6f, 0x72, 0x5f, 0x65, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x04, 0x20, 0x03, + 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, + 0x4e, 0x61, 0x6d, 0x65, 0x64, 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, + 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x22, 0x20, 0x0a, 0x08, 0x54, 0x79, 0x70, + 0x65, 0x49, 0x74, 0x65, 0x6d, 0x12, 0x14, 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x01, + 0x20, 0x03, 0x28, 0x09, 0x52, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x22, 0x5c, 0x0a, 0x0f, 0x56, + 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x12, 0x49, + 0x0a, 0x15, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x5f, 0x70, 0x72, 0x6f, + 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x18, 0x01, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, + 0x41, 0x6e, 0x79, 0x52, 0x14, 0x61, 0x64, 0x64, 0x69, 0x74, 0x69, 0x6f, 0x6e, 0x61, 0x6c, 0x50, + 0x72, 0x6f, 0x70, 0x65, 0x72, 0x74, 0x69, 0x65, 0x73, 0x22, 0xc8, 0x01, 0x0a, 0x03, 0x58, 0x6d, + 0x6c, 0x12, 0x12, 0x0a, 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x04, 0x6e, 0x61, 0x6d, 0x65, 0x12, 0x1c, 0x0a, 0x09, 0x6e, 0x61, 0x6d, 0x65, 0x73, 0x70, 0x61, + 0x63, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, 0x09, 0x6e, 0x61, 0x6d, 0x65, 0x73, 0x70, + 0x61, 0x63, 0x65, 0x12, 0x16, 0x0a, 0x06, 0x70, 0x72, 0x65, 0x66, 0x69, 0x78, 0x18, 0x03, 0x20, + 0x01, 0x28, 0x09, 0x52, 0x06, 0x70, 0x72, 0x65, 0x66, 0x69, 0x78, 0x12, 0x1c, 0x0a, 0x09, 0x61, + 0x74, 0x74, 0x72, 0x69, 0x62, 0x75, 0x74, 0x65, 0x18, 0x04, 0x20, 0x01, 0x28, 0x08, 0x52, 0x09, + 0x61, 0x74, 0x74, 0x72, 0x69, 0x62, 0x75, 0x74, 0x65, 0x12, 0x18, 0x0a, 0x07, 0x77, 0x72, 0x61, + 0x70, 0x70, 0x65, 0x64, 0x18, 0x05, 0x20, 0x01, 0x28, 0x08, 0x52, 0x07, 0x77, 0x72, 0x61, 0x70, + 0x70, 0x65, 0x64, 0x12, 0x3f, 0x0a, 0x10, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x5f, 0x65, 0x78, + 0x74, 0x65, 0x6e, 0x73, 0x69, 0x6f, 0x6e, 0x18, 0x06, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, + 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x2e, 0x76, 0x32, 0x2e, 0x4e, 0x61, 0x6d, 0x65, 0x64, + 0x41, 0x6e, 0x79, 0x52, 0x0f, 0x76, 0x65, 0x6e, 0x64, 0x6f, 0x72, 0x45, 0x78, 0x74, 0x65, 0x6e, + 0x73, 0x69, 0x6f, 0x6e, 0x42, 0x3c, 0x0a, 0x0e, 0x6f, 0x72, 0x67, 0x2e, 0x6f, 0x70, 0x65, 0x6e, + 0x61, 0x70, 0x69, 0x5f, 0x76, 0x32, 0x42, 0x0c, 0x4f, 0x70, 0x65, 0x6e, 0x41, 0x50, 0x49, 0x50, + 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, 0x14, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x76, + 0x32, 0x3b, 0x6f, 0x70, 0x65, 0x6e, 0x61, 0x70, 0x69, 0x5f, 0x76, 0x32, 0xa2, 0x02, 0x03, 0x4f, + 0x41, 0x53, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33, +} + +var ( + file_openapiv2_OpenAPIv2_proto_rawDescOnce sync.Once + file_openapiv2_OpenAPIv2_proto_rawDescData = file_openapiv2_OpenAPIv2_proto_rawDesc +) + +func file_openapiv2_OpenAPIv2_proto_rawDescGZIP() []byte { + file_openapiv2_OpenAPIv2_proto_rawDescOnce.Do(func() { + file_openapiv2_OpenAPIv2_proto_rawDescData = protoimpl.X.CompressGZIP(file_openapiv2_OpenAPIv2_proto_rawDescData) + }) + return file_openapiv2_OpenAPIv2_proto_rawDescData +} + +var file_openapiv2_OpenAPIv2_proto_msgTypes = make([]protoimpl.MessageInfo, 60) +var file_openapiv2_OpenAPIv2_proto_goTypes = []interface{}{ + (*AdditionalPropertiesItem)(nil), // 0: openapi.v2.AdditionalPropertiesItem + (*Any)(nil), // 1: openapi.v2.Any + (*ApiKeySecurity)(nil), // 2: openapi.v2.ApiKeySecurity + (*BasicAuthenticationSecurity)(nil), // 3: openapi.v2.BasicAuthenticationSecurity + (*BodyParameter)(nil), // 4: openapi.v2.BodyParameter + (*Contact)(nil), // 5: openapi.v2.Contact + (*Default)(nil), // 6: openapi.v2.Default + (*Definitions)(nil), // 7: openapi.v2.Definitions + (*Document)(nil), // 8: openapi.v2.Document + (*Examples)(nil), // 9: openapi.v2.Examples + (*ExternalDocs)(nil), // 10: openapi.v2.ExternalDocs + (*FileSchema)(nil), // 11: openapi.v2.FileSchema + (*FormDataParameterSubSchema)(nil), // 12: openapi.v2.FormDataParameterSubSchema + (*Header)(nil), // 13: openapi.v2.Header + (*HeaderParameterSubSchema)(nil), // 14: openapi.v2.HeaderParameterSubSchema + (*Headers)(nil), // 15: openapi.v2.Headers + (*Info)(nil), // 16: openapi.v2.Info + (*ItemsItem)(nil), // 17: openapi.v2.ItemsItem + (*JsonReference)(nil), // 18: openapi.v2.JsonReference + (*License)(nil), // 19: openapi.v2.License + (*NamedAny)(nil), // 20: openapi.v2.NamedAny + (*NamedHeader)(nil), // 21: openapi.v2.NamedHeader + (*NamedParameter)(nil), // 22: openapi.v2.NamedParameter + (*NamedPathItem)(nil), // 23: openapi.v2.NamedPathItem + (*NamedResponse)(nil), // 24: openapi.v2.NamedResponse + (*NamedResponseValue)(nil), // 25: openapi.v2.NamedResponseValue + (*NamedSchema)(nil), // 26: openapi.v2.NamedSchema + (*NamedSecurityDefinitionsItem)(nil), // 27: openapi.v2.NamedSecurityDefinitionsItem + (*NamedString)(nil), // 28: openapi.v2.NamedString + (*NamedStringArray)(nil), // 29: openapi.v2.NamedStringArray + (*NonBodyParameter)(nil), // 30: openapi.v2.NonBodyParameter + (*Oauth2AccessCodeSecurity)(nil), // 31: openapi.v2.Oauth2AccessCodeSecurity + (*Oauth2ApplicationSecurity)(nil), // 32: openapi.v2.Oauth2ApplicationSecurity + (*Oauth2ImplicitSecurity)(nil), // 33: openapi.v2.Oauth2ImplicitSecurity + (*Oauth2PasswordSecurity)(nil), // 34: openapi.v2.Oauth2PasswordSecurity + (*Oauth2Scopes)(nil), // 35: openapi.v2.Oauth2Scopes + (*Operation)(nil), // 36: openapi.v2.Operation + (*Parameter)(nil), // 37: openapi.v2.Parameter + (*ParameterDefinitions)(nil), // 38: openapi.v2.ParameterDefinitions + (*ParametersItem)(nil), // 39: openapi.v2.ParametersItem + (*PathItem)(nil), // 40: openapi.v2.PathItem + (*PathParameterSubSchema)(nil), // 41: openapi.v2.PathParameterSubSchema + (*Paths)(nil), // 42: openapi.v2.Paths + (*PrimitivesItems)(nil), // 43: openapi.v2.PrimitivesItems + (*Properties)(nil), // 44: openapi.v2.Properties + (*QueryParameterSubSchema)(nil), // 45: openapi.v2.QueryParameterSubSchema + (*Response)(nil), // 46: openapi.v2.Response + (*ResponseDefinitions)(nil), // 47: openapi.v2.ResponseDefinitions + (*ResponseValue)(nil), // 48: openapi.v2.ResponseValue + (*Responses)(nil), // 49: openapi.v2.Responses + (*Schema)(nil), // 50: openapi.v2.Schema + (*SchemaItem)(nil), // 51: openapi.v2.SchemaItem + (*SecurityDefinitions)(nil), // 52: openapi.v2.SecurityDefinitions + (*SecurityDefinitionsItem)(nil), // 53: openapi.v2.SecurityDefinitionsItem + (*SecurityRequirement)(nil), // 54: openapi.v2.SecurityRequirement + (*StringArray)(nil), // 55: openapi.v2.StringArray + (*Tag)(nil), // 56: openapi.v2.Tag + (*TypeItem)(nil), // 57: openapi.v2.TypeItem + (*VendorExtension)(nil), // 58: openapi.v2.VendorExtension + (*Xml)(nil), // 59: openapi.v2.Xml + (*any.Any)(nil), // 60: google.protobuf.Any +} +var file_openapiv2_OpenAPIv2_proto_depIdxs = []int32{ + 50, // 0: openapi.v2.AdditionalPropertiesItem.schema:type_name -> openapi.v2.Schema + 60, // 1: openapi.v2.Any.value:type_name -> google.protobuf.Any + 20, // 2: openapi.v2.ApiKeySecurity.vendor_extension:type_name -> openapi.v2.NamedAny + 20, // 3: openapi.v2.BasicAuthenticationSecurity.vendor_extension:type_name -> openapi.v2.NamedAny + 50, // 4: openapi.v2.BodyParameter.schema:type_name -> openapi.v2.Schema + 20, // 5: openapi.v2.BodyParameter.vendor_extension:type_name -> openapi.v2.NamedAny + 20, // 6: openapi.v2.Contact.vendor_extension:type_name -> openapi.v2.NamedAny + 20, // 7: openapi.v2.Default.additional_properties:type_name -> openapi.v2.NamedAny + 26, // 8: openapi.v2.Definitions.additional_properties:type_name -> openapi.v2.NamedSchema + 16, // 9: openapi.v2.Document.info:type_name -> openapi.v2.Info + 42, // 10: openapi.v2.Document.paths:type_name -> openapi.v2.Paths + 7, // 11: openapi.v2.Document.definitions:type_name -> openapi.v2.Definitions + 38, // 12: openapi.v2.Document.parameters:type_name -> openapi.v2.ParameterDefinitions + 47, // 13: openapi.v2.Document.responses:type_name -> openapi.v2.ResponseDefinitions + 54, // 14: openapi.v2.Document.security:type_name -> openapi.v2.SecurityRequirement + 52, // 15: openapi.v2.Document.security_definitions:type_name -> openapi.v2.SecurityDefinitions + 56, // 16: openapi.v2.Document.tags:type_name -> openapi.v2.Tag + 10, // 17: openapi.v2.Document.external_docs:type_name -> openapi.v2.ExternalDocs + 20, // 18: openapi.v2.Document.vendor_extension:type_name -> openapi.v2.NamedAny + 20, // 19: openapi.v2.Examples.additional_properties:type_name -> openapi.v2.NamedAny + 20, // 20: openapi.v2.ExternalDocs.vendor_extension:type_name -> openapi.v2.NamedAny + 1, // 21: openapi.v2.FileSchema.default:type_name -> openapi.v2.Any + 10, // 22: openapi.v2.FileSchema.external_docs:type_name -> openapi.v2.ExternalDocs + 1, // 23: openapi.v2.FileSchema.example:type_name -> openapi.v2.Any + 20, // 24: openapi.v2.FileSchema.vendor_extension:type_name -> openapi.v2.NamedAny + 43, // 25: openapi.v2.FormDataParameterSubSchema.items:type_name -> openapi.v2.PrimitivesItems + 1, // 26: openapi.v2.FormDataParameterSubSchema.default:type_name -> openapi.v2.Any + 1, // 27: openapi.v2.FormDataParameterSubSchema.enum:type_name -> openapi.v2.Any + 20, // 28: openapi.v2.FormDataParameterSubSchema.vendor_extension:type_name -> openapi.v2.NamedAny + 43, // 29: openapi.v2.Header.items:type_name -> openapi.v2.PrimitivesItems + 1, // 30: openapi.v2.Header.default:type_name -> openapi.v2.Any + 1, // 31: openapi.v2.Header.enum:type_name -> openapi.v2.Any + 20, // 32: openapi.v2.Header.vendor_extension:type_name -> openapi.v2.NamedAny + 43, // 33: openapi.v2.HeaderParameterSubSchema.items:type_name -> openapi.v2.PrimitivesItems + 1, // 34: openapi.v2.HeaderParameterSubSchema.default:type_name -> openapi.v2.Any + 1, // 35: openapi.v2.HeaderParameterSubSchema.enum:type_name -> openapi.v2.Any + 20, // 36: openapi.v2.HeaderParameterSubSchema.vendor_extension:type_name -> openapi.v2.NamedAny + 21, // 37: openapi.v2.Headers.additional_properties:type_name -> openapi.v2.NamedHeader + 5, // 38: openapi.v2.Info.contact:type_name -> openapi.v2.Contact + 19, // 39: openapi.v2.Info.license:type_name -> openapi.v2.License + 20, // 40: openapi.v2.Info.vendor_extension:type_name -> openapi.v2.NamedAny + 50, // 41: openapi.v2.ItemsItem.schema:type_name -> openapi.v2.Schema + 20, // 42: openapi.v2.License.vendor_extension:type_name -> openapi.v2.NamedAny + 1, // 43: openapi.v2.NamedAny.value:type_name -> openapi.v2.Any + 13, // 44: openapi.v2.NamedHeader.value:type_name -> openapi.v2.Header + 37, // 45: openapi.v2.NamedParameter.value:type_name -> openapi.v2.Parameter + 40, // 46: openapi.v2.NamedPathItem.value:type_name -> openapi.v2.PathItem + 46, // 47: openapi.v2.NamedResponse.value:type_name -> openapi.v2.Response + 48, // 48: openapi.v2.NamedResponseValue.value:type_name -> openapi.v2.ResponseValue + 50, // 49: openapi.v2.NamedSchema.value:type_name -> openapi.v2.Schema + 53, // 50: openapi.v2.NamedSecurityDefinitionsItem.value:type_name -> openapi.v2.SecurityDefinitionsItem + 55, // 51: openapi.v2.NamedStringArray.value:type_name -> openapi.v2.StringArray + 14, // 52: openapi.v2.NonBodyParameter.header_parameter_sub_schema:type_name -> openapi.v2.HeaderParameterSubSchema + 12, // 53: openapi.v2.NonBodyParameter.form_data_parameter_sub_schema:type_name -> openapi.v2.FormDataParameterSubSchema + 45, // 54: openapi.v2.NonBodyParameter.query_parameter_sub_schema:type_name -> openapi.v2.QueryParameterSubSchema + 41, // 55: openapi.v2.NonBodyParameter.path_parameter_sub_schema:type_name -> openapi.v2.PathParameterSubSchema + 35, // 56: openapi.v2.Oauth2AccessCodeSecurity.scopes:type_name -> openapi.v2.Oauth2Scopes + 20, // 57: openapi.v2.Oauth2AccessCodeSecurity.vendor_extension:type_name -> openapi.v2.NamedAny + 35, // 58: openapi.v2.Oauth2ApplicationSecurity.scopes:type_name -> openapi.v2.Oauth2Scopes + 20, // 59: openapi.v2.Oauth2ApplicationSecurity.vendor_extension:type_name -> openapi.v2.NamedAny + 35, // 60: openapi.v2.Oauth2ImplicitSecurity.scopes:type_name -> openapi.v2.Oauth2Scopes + 20, // 61: openapi.v2.Oauth2ImplicitSecurity.vendor_extension:type_name -> openapi.v2.NamedAny + 35, // 62: openapi.v2.Oauth2PasswordSecurity.scopes:type_name -> openapi.v2.Oauth2Scopes + 20, // 63: openapi.v2.Oauth2PasswordSecurity.vendor_extension:type_name -> openapi.v2.NamedAny + 28, // 64: openapi.v2.Oauth2Scopes.additional_properties:type_name -> openapi.v2.NamedString + 10, // 65: openapi.v2.Operation.external_docs:type_name -> openapi.v2.ExternalDocs + 39, // 66: openapi.v2.Operation.parameters:type_name -> openapi.v2.ParametersItem + 49, // 67: openapi.v2.Operation.responses:type_name -> openapi.v2.Responses + 54, // 68: openapi.v2.Operation.security:type_name -> openapi.v2.SecurityRequirement + 20, // 69: openapi.v2.Operation.vendor_extension:type_name -> openapi.v2.NamedAny + 4, // 70: openapi.v2.Parameter.body_parameter:type_name -> openapi.v2.BodyParameter + 30, // 71: openapi.v2.Parameter.non_body_parameter:type_name -> openapi.v2.NonBodyParameter + 22, // 72: openapi.v2.ParameterDefinitions.additional_properties:type_name -> openapi.v2.NamedParameter + 37, // 73: openapi.v2.ParametersItem.parameter:type_name -> openapi.v2.Parameter + 18, // 74: openapi.v2.ParametersItem.json_reference:type_name -> openapi.v2.JsonReference + 36, // 75: openapi.v2.PathItem.get:type_name -> openapi.v2.Operation + 36, // 76: openapi.v2.PathItem.put:type_name -> openapi.v2.Operation + 36, // 77: openapi.v2.PathItem.post:type_name -> openapi.v2.Operation + 36, // 78: openapi.v2.PathItem.delete:type_name -> openapi.v2.Operation + 36, // 79: openapi.v2.PathItem.options:type_name -> openapi.v2.Operation + 36, // 80: openapi.v2.PathItem.head:type_name -> openapi.v2.Operation + 36, // 81: openapi.v2.PathItem.patch:type_name -> openapi.v2.Operation + 39, // 82: openapi.v2.PathItem.parameters:type_name -> openapi.v2.ParametersItem + 20, // 83: openapi.v2.PathItem.vendor_extension:type_name -> openapi.v2.NamedAny + 43, // 84: openapi.v2.PathParameterSubSchema.items:type_name -> openapi.v2.PrimitivesItems + 1, // 85: openapi.v2.PathParameterSubSchema.default:type_name -> openapi.v2.Any + 1, // 86: openapi.v2.PathParameterSubSchema.enum:type_name -> openapi.v2.Any + 20, // 87: openapi.v2.PathParameterSubSchema.vendor_extension:type_name -> openapi.v2.NamedAny + 20, // 88: openapi.v2.Paths.vendor_extension:type_name -> openapi.v2.NamedAny + 23, // 89: openapi.v2.Paths.path:type_name -> openapi.v2.NamedPathItem + 43, // 90: openapi.v2.PrimitivesItems.items:type_name -> openapi.v2.PrimitivesItems + 1, // 91: openapi.v2.PrimitivesItems.default:type_name -> openapi.v2.Any + 1, // 92: openapi.v2.PrimitivesItems.enum:type_name -> openapi.v2.Any + 20, // 93: openapi.v2.PrimitivesItems.vendor_extension:type_name -> openapi.v2.NamedAny + 26, // 94: openapi.v2.Properties.additional_properties:type_name -> openapi.v2.NamedSchema + 43, // 95: openapi.v2.QueryParameterSubSchema.items:type_name -> openapi.v2.PrimitivesItems + 1, // 96: openapi.v2.QueryParameterSubSchema.default:type_name -> openapi.v2.Any + 1, // 97: openapi.v2.QueryParameterSubSchema.enum:type_name -> openapi.v2.Any + 20, // 98: openapi.v2.QueryParameterSubSchema.vendor_extension:type_name -> openapi.v2.NamedAny + 51, // 99: openapi.v2.Response.schema:type_name -> openapi.v2.SchemaItem + 15, // 100: openapi.v2.Response.headers:type_name -> openapi.v2.Headers + 9, // 101: openapi.v2.Response.examples:type_name -> openapi.v2.Examples + 20, // 102: openapi.v2.Response.vendor_extension:type_name -> openapi.v2.NamedAny + 24, // 103: openapi.v2.ResponseDefinitions.additional_properties:type_name -> openapi.v2.NamedResponse + 46, // 104: openapi.v2.ResponseValue.response:type_name -> openapi.v2.Response + 18, // 105: openapi.v2.ResponseValue.json_reference:type_name -> openapi.v2.JsonReference + 25, // 106: openapi.v2.Responses.response_code:type_name -> openapi.v2.NamedResponseValue + 20, // 107: openapi.v2.Responses.vendor_extension:type_name -> openapi.v2.NamedAny + 1, // 108: openapi.v2.Schema.default:type_name -> openapi.v2.Any + 1, // 109: openapi.v2.Schema.enum:type_name -> openapi.v2.Any + 0, // 110: openapi.v2.Schema.additional_properties:type_name -> openapi.v2.AdditionalPropertiesItem + 57, // 111: openapi.v2.Schema.type:type_name -> openapi.v2.TypeItem + 17, // 112: openapi.v2.Schema.items:type_name -> openapi.v2.ItemsItem + 50, // 113: openapi.v2.Schema.all_of:type_name -> openapi.v2.Schema + 44, // 114: openapi.v2.Schema.properties:type_name -> openapi.v2.Properties + 59, // 115: openapi.v2.Schema.xml:type_name -> openapi.v2.Xml + 10, // 116: openapi.v2.Schema.external_docs:type_name -> openapi.v2.ExternalDocs + 1, // 117: openapi.v2.Schema.example:type_name -> openapi.v2.Any + 20, // 118: openapi.v2.Schema.vendor_extension:type_name -> openapi.v2.NamedAny + 50, // 119: openapi.v2.SchemaItem.schema:type_name -> openapi.v2.Schema + 11, // 120: openapi.v2.SchemaItem.file_schema:type_name -> openapi.v2.FileSchema + 27, // 121: openapi.v2.SecurityDefinitions.additional_properties:type_name -> openapi.v2.NamedSecurityDefinitionsItem + 3, // 122: openapi.v2.SecurityDefinitionsItem.basic_authentication_security:type_name -> openapi.v2.BasicAuthenticationSecurity + 2, // 123: openapi.v2.SecurityDefinitionsItem.api_key_security:type_name -> openapi.v2.ApiKeySecurity + 33, // 124: openapi.v2.SecurityDefinitionsItem.oauth2_implicit_security:type_name -> openapi.v2.Oauth2ImplicitSecurity + 34, // 125: openapi.v2.SecurityDefinitionsItem.oauth2_password_security:type_name -> openapi.v2.Oauth2PasswordSecurity + 32, // 126: openapi.v2.SecurityDefinitionsItem.oauth2_application_security:type_name -> openapi.v2.Oauth2ApplicationSecurity + 31, // 127: openapi.v2.SecurityDefinitionsItem.oauth2_access_code_security:type_name -> openapi.v2.Oauth2AccessCodeSecurity + 29, // 128: openapi.v2.SecurityRequirement.additional_properties:type_name -> openapi.v2.NamedStringArray + 10, // 129: openapi.v2.Tag.external_docs:type_name -> openapi.v2.ExternalDocs + 20, // 130: openapi.v2.Tag.vendor_extension:type_name -> openapi.v2.NamedAny + 20, // 131: openapi.v2.VendorExtension.additional_properties:type_name -> openapi.v2.NamedAny + 20, // 132: openapi.v2.Xml.vendor_extension:type_name -> openapi.v2.NamedAny + 133, // [133:133] is the sub-list for method output_type + 133, // [133:133] is the sub-list for method input_type + 133, // [133:133] is the sub-list for extension type_name + 133, // [133:133] is the sub-list for extension extendee + 0, // [0:133] is the sub-list for field type_name +} + +func init() { file_openapiv2_OpenAPIv2_proto_init() } +func file_openapiv2_OpenAPIv2_proto_init() { + if File_openapiv2_OpenAPIv2_proto != nil { + return + } + if !protoimpl.UnsafeEnabled { + file_openapiv2_OpenAPIv2_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*AdditionalPropertiesItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[1].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Any); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[2].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ApiKeySecurity); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[3].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*BasicAuthenticationSecurity); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[4].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*BodyParameter); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[5].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Contact); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[6].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Default); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[7].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Definitions); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[8].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Document); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[9].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Examples); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[10].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ExternalDocs); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[11].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*FileSchema); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[12].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*FormDataParameterSubSchema); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[13].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Header); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[14].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*HeaderParameterSubSchema); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[15].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Headers); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[16].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Info); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[17].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ItemsItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[18].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*JsonReference); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[19].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*License); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[20].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedAny); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[21].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedHeader); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[22].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedParameter); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[23].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedPathItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[24].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedResponse); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[25].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedResponseValue); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[26].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedSchema); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[27].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedSecurityDefinitionsItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[28].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedString); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[29].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NamedStringArray); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[30].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*NonBodyParameter); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[31].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Oauth2AccessCodeSecurity); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[32].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Oauth2ApplicationSecurity); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[33].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Oauth2ImplicitSecurity); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[34].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Oauth2PasswordSecurity); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[35].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Oauth2Scopes); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[36].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Operation); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[37].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Parameter); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[38].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ParameterDefinitions); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[39].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ParametersItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[40].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*PathItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[41].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*PathParameterSubSchema); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[42].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Paths); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[43].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*PrimitivesItems); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[44].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Properties); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[45].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*QueryParameterSubSchema); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[46].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Response); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[47].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ResponseDefinitions); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[48].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*ResponseValue); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[49].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Responses); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[50].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Schema); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[51].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*SchemaItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[52].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*SecurityDefinitions); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[53].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*SecurityDefinitionsItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[54].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*SecurityRequirement); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[55].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*StringArray); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[56].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Tag); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[57].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*TypeItem); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[58].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*VendorExtension); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[59].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Xml); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + } + file_openapiv2_OpenAPIv2_proto_msgTypes[0].OneofWrappers = []interface{}{ + (*AdditionalPropertiesItem_Schema)(nil), + (*AdditionalPropertiesItem_Boolean)(nil), + } + file_openapiv2_OpenAPIv2_proto_msgTypes[30].OneofWrappers = []interface{}{ + (*NonBodyParameter_HeaderParameterSubSchema)(nil), + (*NonBodyParameter_FormDataParameterSubSchema)(nil), + (*NonBodyParameter_QueryParameterSubSchema)(nil), + (*NonBodyParameter_PathParameterSubSchema)(nil), + } + file_openapiv2_OpenAPIv2_proto_msgTypes[37].OneofWrappers = []interface{}{ + (*Parameter_BodyParameter)(nil), + (*Parameter_NonBodyParameter)(nil), + } + file_openapiv2_OpenAPIv2_proto_msgTypes[39].OneofWrappers = []interface{}{ + (*ParametersItem_Parameter)(nil), + (*ParametersItem_JsonReference)(nil), + } + file_openapiv2_OpenAPIv2_proto_msgTypes[48].OneofWrappers = []interface{}{ + (*ResponseValue_Response)(nil), + (*ResponseValue_JsonReference)(nil), + } + file_openapiv2_OpenAPIv2_proto_msgTypes[51].OneofWrappers = []interface{}{ + (*SchemaItem_Schema)(nil), + (*SchemaItem_FileSchema)(nil), + } + file_openapiv2_OpenAPIv2_proto_msgTypes[53].OneofWrappers = []interface{}{ + (*SecurityDefinitionsItem_BasicAuthenticationSecurity)(nil), + (*SecurityDefinitionsItem_ApiKeySecurity)(nil), + (*SecurityDefinitionsItem_Oauth2ImplicitSecurity)(nil), + (*SecurityDefinitionsItem_Oauth2PasswordSecurity)(nil), + (*SecurityDefinitionsItem_Oauth2ApplicationSecurity)(nil), + (*SecurityDefinitionsItem_Oauth2AccessCodeSecurity)(nil), + } + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_openapiv2_OpenAPIv2_proto_rawDesc, + NumEnums: 0, + NumMessages: 60, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_openapiv2_OpenAPIv2_proto_goTypes, + DependencyIndexes: file_openapiv2_OpenAPIv2_proto_depIdxs, + MessageInfos: file_openapiv2_OpenAPIv2_proto_msgTypes, + }.Build() + File_openapiv2_OpenAPIv2_proto = out.File + file_openapiv2_OpenAPIv2_proto_rawDesc = nil + file_openapiv2_OpenAPIv2_proto_goTypes = nil + file_openapiv2_OpenAPIv2_proto_depIdxs = nil +} diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.proto b/vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.proto similarity index 99% rename from vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.proto rename to vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.proto index 557c8807..00ac1b0a 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.proto +++ b/vendor/github.com/googleapis/gnostic/openapiv2/OpenAPIv2.proto @@ -1,4 +1,4 @@ -// Copyright 2017 Google Inc. All Rights Reserved. +// Copyright 2020 Google LLC. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -41,6 +41,9 @@ option java_package = "org.openapi_v2"; // the future. 'GPB' is reserved for the protocol buffer implementation itself. option objc_class_prefix = "OAS"; +// The Go package name. +option go_package = "openapiv2;openapi_v2"; + message AdditionalPropertiesItem { oneof oneof { Schema schema = 1; @@ -553,7 +556,7 @@ message Response { repeated NamedAny vendor_extension = 5; } -// One or more JSON representations for parameters +// One or more JSON representations for responses message ResponseDefinitions { repeated NamedResponse additional_properties = 1; } diff --git a/vendor/github.com/googleapis/gnostic/openapiv2/README.md b/vendor/github.com/googleapis/gnostic/openapiv2/README.md new file mode 100644 index 00000000..5276128d --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/openapiv2/README.md @@ -0,0 +1,14 @@ +# OpenAPI v2 Protocol Buffer Models + +This directory contains a Protocol Buffer-language model and related code for +supporting OpenAPI v2. + +Gnostic applications and plugins can use OpenAPIv2.proto to generate Protocol +Buffer support code for their preferred languages. + +OpenAPIv2.go is used by Gnostic to read JSON and YAML OpenAPI descriptions into +the Protocol Buffer-based datastructures generated from OpenAPIv2.proto. + +OpenAPIv2.proto and OpenAPIv2.go are generated by the Gnostic compiler +generator, and OpenAPIv2.pb.go is generated by protoc, the Protocol Buffer +compiler, and protoc-gen-go, the Protocol Buffer Go code generation plugin. diff --git a/vendor/github.com/googleapis/gnostic/openapiv2/document.go b/vendor/github.com/googleapis/gnostic/openapiv2/document.go new file mode 100644 index 00000000..56e5966b --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/openapiv2/document.go @@ -0,0 +1,41 @@ +// Copyright 2020 Google LLC. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package openapi_v2 + +import ( + "github.com/googleapis/gnostic/compiler" + "gopkg.in/yaml.v3" +) + +// ParseDocument reads an OpenAPI v2 description from a YAML/JSON representation. +func ParseDocument(b []byte) (*Document, error) { + info, err := compiler.ReadInfoFromBytes("", b) + if err != nil { + return nil, err + } + root := info.Content[0] + return NewDocument(root, compiler.NewContextWithExtensions("$root", root, nil, nil)) +} + +// YAMLValue produces a serialized YAML representation of the document. +func (d *Document) YAMLValue(comment string) ([]byte, error) { + rawInfo := d.ToRawInfo() + rawInfo = &yaml.Node{ + Kind: yaml.DocumentNode, + Content: []*yaml.Node{rawInfo}, + HeadComment: comment, + } + return yaml.Marshal(rawInfo) +} diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv2/openapi-2.0.json b/vendor/github.com/googleapis/gnostic/openapiv2/openapi-2.0.json similarity index 99% rename from vendor/github.com/googleapis/gnostic/OpenAPIv2/openapi-2.0.json rename to vendor/github.com/googleapis/gnostic/openapiv2/openapi-2.0.json index 2815a26e..afa12b79 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv2/openapi-2.0.json +++ b/vendor/github.com/googleapis/gnostic/openapiv2/openapi-2.0.json @@ -203,7 +203,7 @@ "additionalProperties": { "$ref": "#/definitions/response" }, - "description": "One or more JSON representations for parameters" + "description": "One or more JSON representations for responses" }, "externalDocs": { "type": "object", @@ -1607,4 +1607,4 @@ } } } -} +} \ No newline at end of file diff --git a/vendor/github.com/gophercloud/gophercloud/.gitignore b/vendor/github.com/gophercloud/gophercloud/.gitignore deleted file mode 100644 index dd91ed20..00000000 --- a/vendor/github.com/gophercloud/gophercloud/.gitignore +++ /dev/null @@ -1,3 +0,0 @@ -**/*.swp -.idea -.vscode diff --git a/vendor/github.com/gophercloud/gophercloud/.travis.yml b/vendor/github.com/gophercloud/gophercloud/.travis.yml deleted file mode 100644 index 9153a00f..00000000 --- a/vendor/github.com/gophercloud/gophercloud/.travis.yml +++ /dev/null @@ -1,25 +0,0 @@ -language: go -sudo: false -install: -- GO111MODULE=off go get golang.org/x/crypto/ssh -- GO111MODULE=off go get -v -tags 'fixtures acceptance' ./... -- GO111MODULE=off go get github.com/wadey/gocovmerge -- GO111MODULE=off go get github.com/mattn/goveralls -- GO111MODULE=off go get golang.org/x/tools/cmd/goimports -go: -- "1.10" -- "1.11" -- "1.12" -- "tip" -env: - global: - - secure: "xSQsAG5wlL9emjbCdxzz/hYQsSpJ/bABO1kkbwMSISVcJ3Nk0u4ywF+LS4bgeOnwPfmFvNTOqVDu3RwEvMeWXSI76t1piCPcObutb2faKLVD/hLoAS76gYX+Z8yGWGHrSB7Do5vTPj1ERe2UljdrnsSeOXzoDwFxYRaZLX4bBOB4AyoGvRniil5QXPATiA1tsWX1VMicj8a4F8X+xeESzjt1Q5Iy31e7vkptu71bhvXCaoo5QhYwT+pLR9dN0S1b7Ro0KVvkRefmr1lUOSYd2e74h6Lc34tC1h3uYZCS4h47t7v5cOXvMNxinEj2C51RvbjvZI1RLVdkuAEJD1Iz4+Ote46nXbZ//6XRZMZz/YxQ13l7ux1PFjgEB6HAapmF5Xd8PRsgeTU9LRJxpiTJ3P5QJ3leS1va8qnziM5kYipj/Rn+V8g2ad/rgkRox9LSiR9VYZD2Pe45YCb1mTKSl2aIJnV7nkOqsShY5LNB4JZSg7xIffA+9YVDktw8dJlATjZqt7WvJJ49g6A61mIUV4C15q2JPGKTkZzDiG81NtmS7hFa7k0yaE2ELgYocbcuyUcAahhxntYTC0i23nJmEHVNiZmBO3u7EgpWe4KGVfumU+lt12tIn5b3dZRBBUk3QakKKozSK1QPHGpk/AZGrhu7H6l8to6IICKWtDcyMPQ=" - - GO111MODULE=on -before_script: -- go vet ./... -script: -- ./script/coverage -- ./script/unittest -- ./script/format -after_success: -- $HOME/gopath/bin/goveralls -service=travis-ci -coverprofile=cover.out diff --git a/vendor/github.com/gophercloud/gophercloud/.zuul.yaml b/vendor/github.com/gophercloud/gophercloud/.zuul.yaml deleted file mode 100644 index 135e3b20..00000000 --- a/vendor/github.com/gophercloud/gophercloud/.zuul.yaml +++ /dev/null @@ -1,114 +0,0 @@ -- job: - name: gophercloud-unittest - parent: golang-test - description: | - Run gophercloud unit test - run: .zuul/playbooks/gophercloud-unittest/run.yaml - nodeset: ubuntu-xenial-ut - -- job: - name: gophercloud-acceptance-test - parent: golang-test - description: | - Run gophercloud acceptance test on master branch - run: .zuul/playbooks/gophercloud-acceptance-test/run.yaml - -- job: - name: gophercloud-acceptance-test-ironic - parent: golang-test - description: | - Run gophercloud ironic acceptance test on master branch - run: .zuul/playbooks/gophercloud-acceptance-test-ironic/run.yaml - -- job: - name: gophercloud-acceptance-test-stein - parent: gophercloud-acceptance-test - description: | - Run gophercloud acceptance test on stein branch - vars: - global_env: - OS_BRANCH: stable/stein - -- job: - name: gophercloud-acceptance-test-rocky - parent: gophercloud-acceptance-test - description: | - Run gophercloud acceptance test on rocky branch - vars: - global_env: - OS_BRANCH: stable/rocky - -- job: - name: gophercloud-acceptance-test-queens - parent: gophercloud-acceptance-test - description: | - Run gophercloud acceptance test on queens branch - vars: - global_env: - OS_BRANCH: stable/queens - -- job: - name: gophercloud-acceptance-test-pike - parent: gophercloud-acceptance-test - description: | - Run gophercloud acceptance test on pike branch - vars: - global_env: - OS_BRANCH: stable/pike - -- job: - name: gophercloud-acceptance-test-ocata - parent: gophercloud-acceptance-test - description: | - Run gophercloud acceptance test on ocata branch - vars: - global_env: - OS_BRANCH: stable/ocata - -- job: - name: gophercloud-acceptance-test-newton - parent: gophercloud-acceptance-test - description: | - Run gophercloud acceptance test on newton branch - vars: - global_env: - OS_BRANCH: stable/newton - -- job: - name: gophercloud-acceptance-test-mitaka - parent: gophercloud-acceptance-test - description: | - Run gophercloud acceptance test on mitaka branch - vars: - global_env: - OS_BRANCH: stable/mitaka - nodeset: ubuntu-trusty - -- project: - name: gophercloud/gophercloud - check: - jobs: - - gophercloud-unittest - - gophercloud-acceptance-test - - gophercloud-acceptance-test-ironic - recheck-mitaka: - jobs: - - gophercloud-acceptance-test-mitaka - recheck-newton: - jobs: - - gophercloud-acceptance-test-newton - recheck-ocata: - jobs: - - gophercloud-acceptance-test-ocata - recheck-pike: - jobs: - - gophercloud-acceptance-test-pike - recheck-queens: - jobs: - - gophercloud-acceptance-test-queens - recheck-rocky: - jobs: - - gophercloud-acceptance-test-rocky - recheck-stein: - jobs: - - gophercloud-acceptance-test-stein diff --git a/vendor/github.com/gophercloud/gophercloud/CHANGELOG.md b/vendor/github.com/gophercloud/gophercloud/CHANGELOG.md deleted file mode 100644 index e69de29b..00000000 diff --git a/vendor/github.com/gophercloud/gophercloud/README.md b/vendor/github.com/gophercloud/gophercloud/README.md deleted file mode 100644 index ad29041d..00000000 --- a/vendor/github.com/gophercloud/gophercloud/README.md +++ /dev/null @@ -1,159 +0,0 @@ -# Gophercloud: an OpenStack SDK for Go -[![Build Status](https://travis-ci.org/gophercloud/gophercloud.svg?branch=master)](https://travis-ci.org/gophercloud/gophercloud) -[![Coverage Status](https://coveralls.io/repos/github/gophercloud/gophercloud/badge.svg?branch=master)](https://coveralls.io/github/gophercloud/gophercloud?branch=master) - -Gophercloud is an OpenStack Go SDK. - -## Useful links - -* [Reference documentation](http://godoc.org/github.com/gophercloud/gophercloud) -* [Effective Go](https://golang.org/doc/effective_go.html) - -## How to install - -Before installing, you need to ensure that your [GOPATH environment variable](https://golang.org/doc/code.html#GOPATH) -is pointing to an appropriate directory where you want to install Gophercloud: - -```bash -mkdir $HOME/go -export GOPATH=$HOME/go -``` - -To protect yourself against changes in your dependencies, we highly recommend choosing a -[dependency management solution](https://github.com/golang/go/wiki/PackageManagementTools) for -your projects, such as [godep](https://github.com/tools/godep). Once this is set up, you can install -Gophercloud as a dependency like so: - -```bash -go get github.com/gophercloud/gophercloud - -# Edit your code to import relevant packages from "github.com/gophercloud/gophercloud" - -godep save ./... -``` - -This will install all the source files you need into a `Godeps/_workspace` directory, which is -referenceable from your own source files when you use the `godep go` command. - -## Getting started - -### Credentials - -Because you'll be hitting an API, you will need to retrieve your OpenStack -credentials and either store them as environment variables or in your local Go -files. The first method is recommended because it decouples credential -information from source code, allowing you to push the latter to your version -control system without any security risk. - -You will need to retrieve the following: - -* username -* password -* a valid Keystone identity URL - -For users that have the OpenStack dashboard installed, there's a shortcut. If -you visit the `project/access_and_security` path in Horizon and click on the -"Download OpenStack RC File" button at the top right hand corner, you will -download a bash file that exports all of your access details to environment -variables. To execute the file, run `source admin-openrc.sh` and you will be -prompted for your password. - -### Authentication - -Once you have access to your credentials, you can begin plugging them into -Gophercloud. The next step is authentication, and this is handled by a base -"Provider" struct. To get one, you can either pass in your credentials -explicitly, or tell Gophercloud to use environment variables: - -```go -import ( - "github.com/gophercloud/gophercloud" - "github.com/gophercloud/gophercloud/openstack" - "github.com/gophercloud/gophercloud/openstack/utils" -) - -// Option 1: Pass in the values yourself -opts := gophercloud.AuthOptions{ - IdentityEndpoint: "https://openstack.example.com:5000/v2.0", - Username: "{username}", - Password: "{password}", -} - -// Option 2: Use a utility function to retrieve all your environment variables -opts, err := openstack.AuthOptionsFromEnv() -``` - -Once you have the `opts` variable, you can pass it in and get back a -`ProviderClient` struct: - -```go -provider, err := openstack.AuthenticatedClient(opts) -``` - -The `ProviderClient` is the top-level client that all of your OpenStack services -derive from. The provider contains all of the authentication details that allow -your Go code to access the API - such as the base URL and token ID. - -### Provision a server - -Once we have a base Provider, we inject it as a dependency into each OpenStack -service. In order to work with the Compute API, we need a Compute service -client; which can be created like so: - -```go -client, err := openstack.NewComputeV2(provider, gophercloud.EndpointOpts{ - Region: os.Getenv("OS_REGION_NAME"), -}) -``` - -We then use this `client` for any Compute API operation we want. In our case, -we want to provision a new server - so we invoke the `Create` method and pass -in the flavor ID (hardware specification) and image ID (operating system) we're -interested in: - -```go -import "github.com/gophercloud/gophercloud/openstack/compute/v2/servers" - -server, err := servers.Create(client, servers.CreateOpts{ - Name: "My new server!", - FlavorRef: "flavor_id", - ImageRef: "image_id", -}).Extract() -``` - -The above code sample creates a new server with the parameters, and embodies the -new resource in the `server` variable (a -[`servers.Server`](http://godoc.org/github.com/gophercloud/gophercloud) struct). - -## Advanced Usage - -Have a look at the [FAQ](./docs/FAQ.md) for some tips on customizing the way Gophercloud works. - -## Backwards-Compatibility Guarantees - -None. Vendor it and write tests covering the parts you use. - -## Contributing - -See the [contributing guide](./.github/CONTRIBUTING.md). - -## Help and feedback - -If you're struggling with something or have spotted a potential bug, feel free -to submit an issue to our [bug tracker](https://github.com/gophercloud/gophercloud/issues). - -## Thank You - -We'd like to extend special thanks and appreciation to the following: - -### OpenLab - - - -OpenLab is providing a full CI environment to test each PR and merge for a variety of OpenStack releases. - -### VEXXHOST - - - -VEXXHOST is providing their services to assist with the development and testing of Gophercloud. diff --git a/vendor/github.com/gophercloud/gophercloud/auth_options.go b/vendor/github.com/gophercloud/gophercloud/auth_options.go deleted file mode 100644 index 5ffa8d1e..00000000 --- a/vendor/github.com/gophercloud/gophercloud/auth_options.go +++ /dev/null @@ -1,437 +0,0 @@ -package gophercloud - -/* -AuthOptions stores information needed to authenticate to an OpenStack Cloud. -You can populate one manually, or use a provider's AuthOptionsFromEnv() function -to read relevant information from the standard environment variables. Pass one -to a provider's AuthenticatedClient function to authenticate and obtain a -ProviderClient representing an active session on that provider. - -Its fields are the union of those recognized by each identity implementation and -provider. - -An example of manually providing authentication information: - - opts := gophercloud.AuthOptions{ - IdentityEndpoint: "https://openstack.example.com:5000/v2.0", - Username: "{username}", - Password: "{password}", - TenantID: "{tenant_id}", - } - - provider, err := openstack.AuthenticatedClient(opts) - -An example of using AuthOptionsFromEnv(), where the environment variables can -be read from a file, such as a standard openrc file: - - opts, err := openstack.AuthOptionsFromEnv() - provider, err := openstack.AuthenticatedClient(opts) -*/ -type AuthOptions struct { - // IdentityEndpoint specifies the HTTP endpoint that is required to work with - // the Identity API of the appropriate version. While it's ultimately needed by - // all of the identity services, it will often be populated by a provider-level - // function. - // - // The IdentityEndpoint is typically referred to as the "auth_url" or - // "OS_AUTH_URL" in the information provided by the cloud operator. - IdentityEndpoint string `json:"-"` - - // Username is required if using Identity V2 API. Consult with your provider's - // control panel to discover your account's username. In Identity V3, either - // UserID or a combination of Username and DomainID or DomainName are needed. - Username string `json:"username,omitempty"` - UserID string `json:"-"` - - Password string `json:"password,omitempty"` - - // At most one of DomainID and DomainName must be provided if using Username - // with Identity V3. Otherwise, either are optional. - DomainID string `json:"-"` - DomainName string `json:"name,omitempty"` - - // The TenantID and TenantName fields are optional for the Identity V2 API. - // The same fields are known as project_id and project_name in the Identity - // V3 API, but are collected as TenantID and TenantName here in both cases. - // Some providers allow you to specify a TenantName instead of the TenantId. - // Some require both. Your provider's authentication policies will determine - // how these fields influence authentication. - // If DomainID or DomainName are provided, they will also apply to TenantName. - // It is not currently possible to authenticate with Username and a Domain - // and scope to a Project in a different Domain by using TenantName. To - // accomplish that, the ProjectID will need to be provided as the TenantID - // option. - TenantID string `json:"tenantId,omitempty"` - TenantName string `json:"tenantName,omitempty"` - - // AllowReauth should be set to true if you grant permission for Gophercloud to - // cache your credentials in memory, and to allow Gophercloud to attempt to - // re-authenticate automatically if/when your token expires. If you set it to - // false, it will not cache these settings, but re-authentication will not be - // possible. This setting defaults to false. - // - // NOTE: The reauth function will try to re-authenticate endlessly if left - // unchecked. The way to limit the number of attempts is to provide a custom - // HTTP client to the provider client and provide a transport that implements - // the RoundTripper interface and stores the number of failed retries. For an - // example of this, see here: - // https://github.com/rackspace/rack/blob/1.0.0/auth/clients.go#L311 - AllowReauth bool `json:"-"` - - // TokenID allows users to authenticate (possibly as another user) with an - // authentication token ID. - TokenID string `json:"-"` - - // Scope determines the scoping of the authentication request. - Scope *AuthScope `json:"-"` - - // Authentication through Application Credentials requires supplying name, project and secret - // For project we can use TenantID - ApplicationCredentialID string `json:"-"` - ApplicationCredentialName string `json:"-"` - ApplicationCredentialSecret string `json:"-"` -} - -// AuthScope allows a created token to be limited to a specific domain or project. -type AuthScope struct { - ProjectID string - ProjectName string - DomainID string - DomainName string -} - -// ToTokenV2CreateMap allows AuthOptions to satisfy the AuthOptionsBuilder -// interface in the v2 tokens package -func (opts AuthOptions) ToTokenV2CreateMap() (map[string]interface{}, error) { - // Populate the request map. - authMap := make(map[string]interface{}) - - if opts.Username != "" { - if opts.Password != "" { - authMap["passwordCredentials"] = map[string]interface{}{ - "username": opts.Username, - "password": opts.Password, - } - } else { - return nil, ErrMissingInput{Argument: "Password"} - } - } else if opts.TokenID != "" { - authMap["token"] = map[string]interface{}{ - "id": opts.TokenID, - } - } else { - return nil, ErrMissingInput{Argument: "Username"} - } - - if opts.TenantID != "" { - authMap["tenantId"] = opts.TenantID - } - if opts.TenantName != "" { - authMap["tenantName"] = opts.TenantName - } - - return map[string]interface{}{"auth": authMap}, nil -} - -func (opts *AuthOptions) ToTokenV3CreateMap(scope map[string]interface{}) (map[string]interface{}, error) { - type domainReq struct { - ID *string `json:"id,omitempty"` - Name *string `json:"name,omitempty"` - } - - type projectReq struct { - Domain *domainReq `json:"domain,omitempty"` - Name *string `json:"name,omitempty"` - ID *string `json:"id,omitempty"` - } - - type userReq struct { - ID *string `json:"id,omitempty"` - Name *string `json:"name,omitempty"` - Password string `json:"password,omitempty"` - Domain *domainReq `json:"domain,omitempty"` - } - - type passwordReq struct { - User userReq `json:"user"` - } - - type tokenReq struct { - ID string `json:"id"` - } - - type applicationCredentialReq struct { - ID *string `json:"id,omitempty"` - Name *string `json:"name,omitempty"` - User *userReq `json:"user,omitempty"` - Secret *string `json:"secret,omitempty"` - } - - type identityReq struct { - Methods []string `json:"methods"` - Password *passwordReq `json:"password,omitempty"` - Token *tokenReq `json:"token,omitempty"` - ApplicationCredential *applicationCredentialReq `json:"application_credential,omitempty"` - } - - type authReq struct { - Identity identityReq `json:"identity"` - } - - type request struct { - Auth authReq `json:"auth"` - } - - // Populate the request structure based on the provided arguments. Create and return an error - // if insufficient or incompatible information is present. - var req request - - if opts.Password == "" { - if opts.TokenID != "" { - // Because we aren't using password authentication, it's an error to also provide any of the user-based authentication - // parameters. - if opts.Username != "" { - return nil, ErrUsernameWithToken{} - } - if opts.UserID != "" { - return nil, ErrUserIDWithToken{} - } - if opts.DomainID != "" { - return nil, ErrDomainIDWithToken{} - } - if opts.DomainName != "" { - return nil, ErrDomainNameWithToken{} - } - - // Configure the request for Token authentication. - req.Auth.Identity.Methods = []string{"token"} - req.Auth.Identity.Token = &tokenReq{ - ID: opts.TokenID, - } - - } else if opts.ApplicationCredentialID != "" { - // Configure the request for ApplicationCredentialID authentication. - // https://github.com/openstack/keystoneauth/blob/stable/rocky/keystoneauth1/identity/v3/application_credential.py#L48-L67 - // There are three kinds of possible application_credential requests - // 1. application_credential id + secret - // 2. application_credential name + secret + user_id - // 3. application_credential name + secret + username + domain_id / domain_name - if opts.ApplicationCredentialSecret == "" { - return nil, ErrAppCredMissingSecret{} - } - req.Auth.Identity.Methods = []string{"application_credential"} - req.Auth.Identity.ApplicationCredential = &applicationCredentialReq{ - ID: &opts.ApplicationCredentialID, - Secret: &opts.ApplicationCredentialSecret, - } - } else if opts.ApplicationCredentialName != "" { - if opts.ApplicationCredentialSecret == "" { - return nil, ErrAppCredMissingSecret{} - } - - var userRequest *userReq - - if opts.UserID != "" { - // UserID could be used without the domain information - userRequest = &userReq{ - ID: &opts.UserID, - } - } - - if userRequest == nil && opts.Username == "" { - // Make sure that Username or UserID are provided - return nil, ErrUsernameOrUserID{} - } - - if userRequest == nil && opts.DomainID != "" { - userRequest = &userReq{ - Name: &opts.Username, - Domain: &domainReq{ID: &opts.DomainID}, - } - } - - if userRequest == nil && opts.DomainName != "" { - userRequest = &userReq{ - Name: &opts.Username, - Domain: &domainReq{Name: &opts.DomainName}, - } - } - - // Make sure that DomainID or DomainName are provided among Username - if userRequest == nil { - return nil, ErrDomainIDOrDomainName{} - } - - req.Auth.Identity.Methods = []string{"application_credential"} - req.Auth.Identity.ApplicationCredential = &applicationCredentialReq{ - Name: &opts.ApplicationCredentialName, - User: userRequest, - Secret: &opts.ApplicationCredentialSecret, - } - } else { - // If no password or token ID or ApplicationCredential are available, authentication can't continue. - return nil, ErrMissingPassword{} - } - } else { - // Password authentication. - req.Auth.Identity.Methods = []string{"password"} - - // At least one of Username and UserID must be specified. - if opts.Username == "" && opts.UserID == "" { - return nil, ErrUsernameOrUserID{} - } - - if opts.Username != "" { - // If Username is provided, UserID may not be provided. - if opts.UserID != "" { - return nil, ErrUsernameOrUserID{} - } - - // Either DomainID or DomainName must also be specified. - if opts.DomainID == "" && opts.DomainName == "" { - return nil, ErrDomainIDOrDomainName{} - } - - if opts.DomainID != "" { - if opts.DomainName != "" { - return nil, ErrDomainIDOrDomainName{} - } - - // Configure the request for Username and Password authentication with a DomainID. - req.Auth.Identity.Password = &passwordReq{ - User: userReq{ - Name: &opts.Username, - Password: opts.Password, - Domain: &domainReq{ID: &opts.DomainID}, - }, - } - } - - if opts.DomainName != "" { - // Configure the request for Username and Password authentication with a DomainName. - req.Auth.Identity.Password = &passwordReq{ - User: userReq{ - Name: &opts.Username, - Password: opts.Password, - Domain: &domainReq{Name: &opts.DomainName}, - }, - } - } - } - - if opts.UserID != "" { - // If UserID is specified, neither DomainID nor DomainName may be. - if opts.DomainID != "" { - return nil, ErrDomainIDWithUserID{} - } - if opts.DomainName != "" { - return nil, ErrDomainNameWithUserID{} - } - - // Configure the request for UserID and Password authentication. - req.Auth.Identity.Password = &passwordReq{ - User: userReq{ID: &opts.UserID, Password: opts.Password}, - } - } - } - - b, err := BuildRequestBody(req, "") - if err != nil { - return nil, err - } - - if len(scope) != 0 { - b["auth"].(map[string]interface{})["scope"] = scope - } - - return b, nil -} - -func (opts *AuthOptions) ToTokenV3ScopeMap() (map[string]interface{}, error) { - // For backwards compatibility. - // If AuthOptions.Scope was not set, try to determine it. - // This works well for common scenarios. - if opts.Scope == nil { - opts.Scope = new(AuthScope) - if opts.TenantID != "" { - opts.Scope.ProjectID = opts.TenantID - } else { - if opts.TenantName != "" { - opts.Scope.ProjectName = opts.TenantName - opts.Scope.DomainID = opts.DomainID - opts.Scope.DomainName = opts.DomainName - } - } - } - - if opts.Scope.ProjectName != "" { - // ProjectName provided: either DomainID or DomainName must also be supplied. - // ProjectID may not be supplied. - if opts.Scope.DomainID == "" && opts.Scope.DomainName == "" { - return nil, ErrScopeDomainIDOrDomainName{} - } - if opts.Scope.ProjectID != "" { - return nil, ErrScopeProjectIDOrProjectName{} - } - - if opts.Scope.DomainID != "" { - // ProjectName + DomainID - return map[string]interface{}{ - "project": map[string]interface{}{ - "name": &opts.Scope.ProjectName, - "domain": map[string]interface{}{"id": &opts.Scope.DomainID}, - }, - }, nil - } - - if opts.Scope.DomainName != "" { - // ProjectName + DomainName - return map[string]interface{}{ - "project": map[string]interface{}{ - "name": &opts.Scope.ProjectName, - "domain": map[string]interface{}{"name": &opts.Scope.DomainName}, - }, - }, nil - } - } else if opts.Scope.ProjectID != "" { - // ProjectID provided. ProjectName, DomainID, and DomainName may not be provided. - if opts.Scope.DomainID != "" { - return nil, ErrScopeProjectIDAlone{} - } - if opts.Scope.DomainName != "" { - return nil, ErrScopeProjectIDAlone{} - } - - // ProjectID - return map[string]interface{}{ - "project": map[string]interface{}{ - "id": &opts.Scope.ProjectID, - }, - }, nil - } else if opts.Scope.DomainID != "" { - // DomainID provided. ProjectID, ProjectName, and DomainName may not be provided. - if opts.Scope.DomainName != "" { - return nil, ErrScopeDomainIDOrDomainName{} - } - - // DomainID - return map[string]interface{}{ - "domain": map[string]interface{}{ - "id": &opts.Scope.DomainID, - }, - }, nil - } else if opts.Scope.DomainName != "" { - // DomainName - return map[string]interface{}{ - "domain": map[string]interface{}{ - "name": &opts.Scope.DomainName, - }, - }, nil - } - - return nil, nil -} - -func (opts AuthOptions) CanReauth() bool { - return opts.AllowReauth -} diff --git a/vendor/github.com/gophercloud/gophercloud/auth_result.go b/vendor/github.com/gophercloud/gophercloud/auth_result.go deleted file mode 100644 index 2e4699b9..00000000 --- a/vendor/github.com/gophercloud/gophercloud/auth_result.go +++ /dev/null @@ -1,52 +0,0 @@ -package gophercloud - -/* -AuthResult is the result from the request that was used to obtain a provider -client's Keystone token. It is returned from ProviderClient.GetAuthResult(). - -The following types satisfy this interface: - - github.com/gophercloud/gophercloud/openstack/identity/v2/tokens.CreateResult - github.com/gophercloud/gophercloud/openstack/identity/v3/tokens.CreateResult - -Usage example: - - import ( - "github.com/gophercloud/gophercloud" - tokens2 "github.com/gophercloud/gophercloud/openstack/identity/v2/tokens" - tokens3 "github.com/gophercloud/gophercloud/openstack/identity/v3/tokens" - ) - - func GetAuthenticatedUserID(providerClient *gophercloud.ProviderClient) (string, error) { - r := providerClient.GetAuthResult() - if r == nil { - //ProviderClient did not use openstack.Authenticate(), e.g. because token - //was set manually with ProviderClient.SetToken() - return "", errors.New("no AuthResult available") - } - switch r := r.(type) { - case tokens2.CreateResult: - u, err := r.ExtractUser() - if err != nil { - return "", err - } - return u.ID, nil - case tokens3.CreateResult: - u, err := r.ExtractUser() - if err != nil { - return "", err - } - return u.ID, nil - default: - panic(fmt.Sprintf("got unexpected AuthResult type %t", r)) - } - } - -Both implementing types share a lot of methods by name, like ExtractUser() in -this example. But those methods cannot be part of the AuthResult interface -because the return types are different (in this case, type tokens2.User vs. -type tokens3.User). -*/ -type AuthResult interface { - ExtractTokenID() (string, error) -} diff --git a/vendor/github.com/gophercloud/gophercloud/doc.go b/vendor/github.com/gophercloud/gophercloud/doc.go deleted file mode 100644 index 953ca822..00000000 --- a/vendor/github.com/gophercloud/gophercloud/doc.go +++ /dev/null @@ -1,110 +0,0 @@ -/* -Package gophercloud provides a multi-vendor interface to OpenStack-compatible -clouds. The library has a three-level hierarchy: providers, services, and -resources. - -Authenticating with Providers - -Provider structs represent the cloud providers that offer and manage a -collection of services. You will generally want to create one Provider -client per OpenStack cloud. - - It is now recommended to use the `clientconfig` package found at - https://github.com/gophercloud/utils/tree/master/openstack/clientconfig - for all authentication purposes. - - The below documentation is still relevant. clientconfig simply implements - the below and presents it in an easier and more flexible way. - -Use your OpenStack credentials to create a Provider client. The -IdentityEndpoint is typically refered to as "auth_url" or "OS_AUTH_URL" in -information provided by the cloud operator. Additionally, the cloud may refer to -TenantID or TenantName as project_id and project_name. Credentials are -specified like so: - - opts := gophercloud.AuthOptions{ - IdentityEndpoint: "https://openstack.example.com:5000/v2.0", - Username: "{username}", - Password: "{password}", - TenantID: "{tenant_id}", - } - - provider, err := openstack.AuthenticatedClient(opts) - -You can authenticate with a token by doing: - - opts := gophercloud.AuthOptions{ - IdentityEndpoint: "https://openstack.example.com:5000/v2.0", - TokenID: "{token_id}", - TenantID: "{tenant_id}", - } - - provider, err := openstack.AuthenticatedClient(opts) - -You may also use the openstack.AuthOptionsFromEnv() helper function. This -function reads in standard environment variables frequently found in an -OpenStack `openrc` file. Again note that Gophercloud currently uses "tenant" -instead of "project". - - opts, err := openstack.AuthOptionsFromEnv() - provider, err := openstack.AuthenticatedClient(opts) - -Service Clients - -Service structs are specific to a provider and handle all of the logic and -operations for a particular OpenStack service. Examples of services include: -Compute, Object Storage, Block Storage. In order to define one, you need to -pass in the parent provider, like so: - - opts := gophercloud.EndpointOpts{Region: "RegionOne"} - - client, err := openstack.NewComputeV2(provider, opts) - -Resources - -Resource structs are the domain models that services make use of in order -to work with and represent the state of API resources: - - server, err := servers.Get(client, "{serverId}").Extract() - -Intermediate Result structs are returned for API operations, which allow -generic access to the HTTP headers, response body, and any errors associated -with the network transaction. To turn a result into a usable resource struct, -you must call the Extract method which is chained to the response, or an -Extract function from an applicable extension: - - result := servers.Get(client, "{serverId}") - - // Attempt to extract the disk configuration from the OS-DCF disk config - // extension: - config, err := diskconfig.ExtractGet(result) - -All requests that enumerate a collection return a Pager struct that is used to -iterate through the results one page at a time. Use the EachPage method on that -Pager to handle each successive Page in a closure, then use the appropriate -extraction method from that request's package to interpret that Page as a slice -of results: - - err := servers.List(client, nil).EachPage(func (page pagination.Page) (bool, error) { - s, err := servers.ExtractServers(page) - if err != nil { - return false, err - } - - // Handle the []servers.Server slice. - - // Return "false" or an error to prematurely stop fetching new pages. - return true, nil - }) - -If you want to obtain the entire collection of pages without doing any -intermediary processing on each page, you can use the AllPages method: - - allPages, err := servers.List(client, nil).AllPages() - allServers, err := servers.ExtractServers(allPages) - -This top-level package contains utility functions and data types that are used -throughout the provider and service packages. Of particular note for end users -are the AuthOptions and EndpointOpts structs. -*/ -package gophercloud diff --git a/vendor/github.com/gophercloud/gophercloud/endpoint_search.go b/vendor/github.com/gophercloud/gophercloud/endpoint_search.go deleted file mode 100644 index 2fbc3c97..00000000 --- a/vendor/github.com/gophercloud/gophercloud/endpoint_search.go +++ /dev/null @@ -1,76 +0,0 @@ -package gophercloud - -// Availability indicates to whom a specific service endpoint is accessible: -// the internet at large, internal networks only, or only to administrators. -// Different identity services use different terminology for these. Identity v2 -// lists them as different kinds of URLs within the service catalog ("adminURL", -// "internalURL", and "publicURL"), while v3 lists them as "Interfaces" in an -// endpoint's response. -type Availability string - -const ( - // AvailabilityAdmin indicates that an endpoint is only available to - // administrators. - AvailabilityAdmin Availability = "admin" - - // AvailabilityPublic indicates that an endpoint is available to everyone on - // the internet. - AvailabilityPublic Availability = "public" - - // AvailabilityInternal indicates that an endpoint is only available within - // the cluster's internal network. - AvailabilityInternal Availability = "internal" -) - -// EndpointOpts specifies search criteria used by queries against an -// OpenStack service catalog. The options must contain enough information to -// unambiguously identify one, and only one, endpoint within the catalog. -// -// Usually, these are passed to service client factory functions in a provider -// package, like "openstack.NewComputeV2()". -type EndpointOpts struct { - // Type [required] is the service type for the client (e.g., "compute", - // "object-store"). Generally, this will be supplied by the service client - // function, but a user-given value will be honored if provided. - Type string - - // Name [optional] is the service name for the client (e.g., "nova") as it - // appears in the service catalog. Services can have the same Type but a - // different Name, which is why both Type and Name are sometimes needed. - Name string - - // Region [required] is the geographic region in which the endpoint resides, - // generally specifying which datacenter should house your resources. - // Required only for services that span multiple regions. - Region string - - // Availability [optional] is the visibility of the endpoint to be returned. - // Valid types include the constants AvailabilityPublic, AvailabilityInternal, - // or AvailabilityAdmin from this package. - // - // Availability is not required, and defaults to AvailabilityPublic. Not all - // providers or services offer all Availability options. - Availability Availability -} - -/* -EndpointLocator is an internal function to be used by provider implementations. - -It provides an implementation that locates a single endpoint from a service -catalog for a specific ProviderClient based on user-provided EndpointOpts. The -provider then uses it to discover related ServiceClients. -*/ -type EndpointLocator func(EndpointOpts) (string, error) - -// ApplyDefaults is an internal method to be used by provider implementations. -// -// It sets EndpointOpts fields if not already set, including a default type. -// Currently, EndpointOpts.Availability defaults to the public endpoint. -func (eo *EndpointOpts) ApplyDefaults(t string) { - if eo.Type == "" { - eo.Type = t - } - if eo.Availability == "" { - eo.Availability = AvailabilityPublic - } -} diff --git a/vendor/github.com/gophercloud/gophercloud/errors.go b/vendor/github.com/gophercloud/gophercloud/errors.go deleted file mode 100644 index 0bcb3af7..00000000 --- a/vendor/github.com/gophercloud/gophercloud/errors.go +++ /dev/null @@ -1,471 +0,0 @@ -package gophercloud - -import ( - "fmt" - "strings" -) - -// BaseError is an error type that all other error types embed. -type BaseError struct { - DefaultErrString string - Info string -} - -func (e BaseError) Error() string { - e.DefaultErrString = "An error occurred while executing a Gophercloud request." - return e.choseErrString() -} - -func (e BaseError) choseErrString() string { - if e.Info != "" { - return e.Info - } - return e.DefaultErrString -} - -// ErrMissingInput is the error when input is required in a particular -// situation but not provided by the user -type ErrMissingInput struct { - BaseError - Argument string -} - -func (e ErrMissingInput) Error() string { - e.DefaultErrString = fmt.Sprintf("Missing input for argument [%s]", e.Argument) - return e.choseErrString() -} - -// ErrInvalidInput is an error type used for most non-HTTP Gophercloud errors. -type ErrInvalidInput struct { - ErrMissingInput - Value interface{} -} - -func (e ErrInvalidInput) Error() string { - e.DefaultErrString = fmt.Sprintf("Invalid input provided for argument [%s]: [%+v]", e.Argument, e.Value) - return e.choseErrString() -} - -// ErrMissingEnvironmentVariable is the error when environment variable is required -// in a particular situation but not provided by the user -type ErrMissingEnvironmentVariable struct { - BaseError - EnvironmentVariable string -} - -func (e ErrMissingEnvironmentVariable) Error() string { - e.DefaultErrString = fmt.Sprintf("Missing environment variable [%s]", e.EnvironmentVariable) - return e.choseErrString() -} - -// ErrMissingAnyoneOfEnvironmentVariables is the error when anyone of the environment variables -// is required in a particular situation but not provided by the user -type ErrMissingAnyoneOfEnvironmentVariables struct { - BaseError - EnvironmentVariables []string -} - -func (e ErrMissingAnyoneOfEnvironmentVariables) Error() string { - e.DefaultErrString = fmt.Sprintf( - "Missing one of the following environment variables [%s]", - strings.Join(e.EnvironmentVariables, ", "), - ) - return e.choseErrString() -} - -// ErrUnexpectedResponseCode is returned by the Request method when a response code other than -// those listed in OkCodes is encountered. -type ErrUnexpectedResponseCode struct { - BaseError - URL string - Method string - Expected []int - Actual int - Body []byte -} - -func (e ErrUnexpectedResponseCode) Error() string { - e.DefaultErrString = fmt.Sprintf( - "Expected HTTP response code %v when accessing [%s %s], but got %d instead\n%s", - e.Expected, e.Method, e.URL, e.Actual, e.Body, - ) - return e.choseErrString() -} - -// ErrDefault400 is the default error type returned on a 400 HTTP response code. -type ErrDefault400 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault401 is the default error type returned on a 401 HTTP response code. -type ErrDefault401 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault403 is the default error type returned on a 403 HTTP response code. -type ErrDefault403 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault404 is the default error type returned on a 404 HTTP response code. -type ErrDefault404 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault405 is the default error type returned on a 405 HTTP response code. -type ErrDefault405 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault408 is the default error type returned on a 408 HTTP response code. -type ErrDefault408 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault409 is the default error type returned on a 409 HTTP response code. -type ErrDefault409 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault429 is the default error type returned on a 429 HTTP response code. -type ErrDefault429 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault500 is the default error type returned on a 500 HTTP response code. -type ErrDefault500 struct { - ErrUnexpectedResponseCode -} - -// ErrDefault503 is the default error type returned on a 503 HTTP response code. -type ErrDefault503 struct { - ErrUnexpectedResponseCode -} - -func (e ErrDefault400) Error() string { - e.DefaultErrString = fmt.Sprintf( - "Bad request with: [%s %s], error message: %s", - e.Method, e.URL, e.Body, - ) - return e.choseErrString() -} -func (e ErrDefault401) Error() string { - return "Authentication failed" -} -func (e ErrDefault403) Error() string { - e.DefaultErrString = fmt.Sprintf( - "Request forbidden: [%s %s], error message: %s", - e.Method, e.URL, e.Body, - ) - return e.choseErrString() -} -func (e ErrDefault404) Error() string { - return "Resource not found" -} -func (e ErrDefault405) Error() string { - return "Method not allowed" -} -func (e ErrDefault408) Error() string { - return "The server timed out waiting for the request" -} -func (e ErrDefault429) Error() string { - return "Too many requests have been sent in a given amount of time. Pause" + - " requests, wait up to one minute, and try again." -} -func (e ErrDefault500) Error() string { - return "Internal Server Error" -} -func (e ErrDefault503) Error() string { - return "The service is currently unable to handle the request due to a temporary" + - " overloading or maintenance. This is a temporary condition. Try again later." -} - -// Err400er is the interface resource error types implement to override the error message -// from a 400 error. -type Err400er interface { - Error400(ErrUnexpectedResponseCode) error -} - -// Err401er is the interface resource error types implement to override the error message -// from a 401 error. -type Err401er interface { - Error401(ErrUnexpectedResponseCode) error -} - -// Err403er is the interface resource error types implement to override the error message -// from a 403 error. -type Err403er interface { - Error403(ErrUnexpectedResponseCode) error -} - -// Err404er is the interface resource error types implement to override the error message -// from a 404 error. -type Err404er interface { - Error404(ErrUnexpectedResponseCode) error -} - -// Err405er is the interface resource error types implement to override the error message -// from a 405 error. -type Err405er interface { - Error405(ErrUnexpectedResponseCode) error -} - -// Err408er is the interface resource error types implement to override the error message -// from a 408 error. -type Err408er interface { - Error408(ErrUnexpectedResponseCode) error -} - -// Err409er is the interface resource error types implement to override the error message -// from a 409 error. -type Err409er interface { - Error409(ErrUnexpectedResponseCode) error -} - -// Err429er is the interface resource error types implement to override the error message -// from a 429 error. -type Err429er interface { - Error429(ErrUnexpectedResponseCode) error -} - -// Err500er is the interface resource error types implement to override the error message -// from a 500 error. -type Err500er interface { - Error500(ErrUnexpectedResponseCode) error -} - -// Err503er is the interface resource error types implement to override the error message -// from a 503 error. -type Err503er interface { - Error503(ErrUnexpectedResponseCode) error -} - -// ErrTimeOut is the error type returned when an operations times out. -type ErrTimeOut struct { - BaseError -} - -func (e ErrTimeOut) Error() string { - e.DefaultErrString = "A time out occurred" - return e.choseErrString() -} - -// ErrUnableToReauthenticate is the error type returned when reauthentication fails. -type ErrUnableToReauthenticate struct { - BaseError - ErrOriginal error -} - -func (e ErrUnableToReauthenticate) Error() string { - e.DefaultErrString = fmt.Sprintf("Unable to re-authenticate: %s", e.ErrOriginal) - return e.choseErrString() -} - -// ErrErrorAfterReauthentication is the error type returned when reauthentication -// succeeds, but an error occurs afterword (usually an HTTP error). -type ErrErrorAfterReauthentication struct { - BaseError - ErrOriginal error -} - -func (e ErrErrorAfterReauthentication) Error() string { - e.DefaultErrString = fmt.Sprintf("Successfully re-authenticated, but got error executing request: %s", e.ErrOriginal) - return e.choseErrString() -} - -// ErrServiceNotFound is returned when no service in a service catalog matches -// the provided EndpointOpts. This is generally returned by provider service -// factory methods like "NewComputeV2()" and can mean that a service is not -// enabled for your account. -type ErrServiceNotFound struct { - BaseError -} - -func (e ErrServiceNotFound) Error() string { - e.DefaultErrString = "No suitable service could be found in the service catalog." - return e.choseErrString() -} - -// ErrEndpointNotFound is returned when no available endpoints match the -// provided EndpointOpts. This is also generally returned by provider service -// factory methods, and usually indicates that a region was specified -// incorrectly. -type ErrEndpointNotFound struct { - BaseError -} - -func (e ErrEndpointNotFound) Error() string { - e.DefaultErrString = "No suitable endpoint could be found in the service catalog." - return e.choseErrString() -} - -// ErrResourceNotFound is the error when trying to retrieve a resource's -// ID by name and the resource doesn't exist. -type ErrResourceNotFound struct { - BaseError - Name string - ResourceType string -} - -func (e ErrResourceNotFound) Error() string { - e.DefaultErrString = fmt.Sprintf("Unable to find %s with name %s", e.ResourceType, e.Name) - return e.choseErrString() -} - -// ErrMultipleResourcesFound is the error when trying to retrieve a resource's -// ID by name and multiple resources have the user-provided name. -type ErrMultipleResourcesFound struct { - BaseError - Name string - Count int - ResourceType string -} - -func (e ErrMultipleResourcesFound) Error() string { - e.DefaultErrString = fmt.Sprintf("Found %d %ss matching %s", e.Count, e.ResourceType, e.Name) - return e.choseErrString() -} - -// ErrUnexpectedType is the error when an unexpected type is encountered -type ErrUnexpectedType struct { - BaseError - Expected string - Actual string -} - -func (e ErrUnexpectedType) Error() string { - e.DefaultErrString = fmt.Sprintf("Expected %s but got %s", e.Expected, e.Actual) - return e.choseErrString() -} - -func unacceptedAttributeErr(attribute string) string { - return fmt.Sprintf("The base Identity V3 API does not accept authentication by %s", attribute) -} - -func redundantWithTokenErr(attribute string) string { - return fmt.Sprintf("%s may not be provided when authenticating with a TokenID", attribute) -} - -func redundantWithUserID(attribute string) string { - return fmt.Sprintf("%s may not be provided when authenticating with a UserID", attribute) -} - -// ErrAPIKeyProvided indicates that an APIKey was provided but can't be used. -type ErrAPIKeyProvided struct{ BaseError } - -func (e ErrAPIKeyProvided) Error() string { - return unacceptedAttributeErr("APIKey") -} - -// ErrTenantIDProvided indicates that a TenantID was provided but can't be used. -type ErrTenantIDProvided struct{ BaseError } - -func (e ErrTenantIDProvided) Error() string { - return unacceptedAttributeErr("TenantID") -} - -// ErrTenantNameProvided indicates that a TenantName was provided but can't be used. -type ErrTenantNameProvided struct{ BaseError } - -func (e ErrTenantNameProvided) Error() string { - return unacceptedAttributeErr("TenantName") -} - -// ErrUsernameWithToken indicates that a Username was provided, but token authentication is being used instead. -type ErrUsernameWithToken struct{ BaseError } - -func (e ErrUsernameWithToken) Error() string { - return redundantWithTokenErr("Username") -} - -// ErrUserIDWithToken indicates that a UserID was provided, but token authentication is being used instead. -type ErrUserIDWithToken struct{ BaseError } - -func (e ErrUserIDWithToken) Error() string { - return redundantWithTokenErr("UserID") -} - -// ErrDomainIDWithToken indicates that a DomainID was provided, but token authentication is being used instead. -type ErrDomainIDWithToken struct{ BaseError } - -func (e ErrDomainIDWithToken) Error() string { - return redundantWithTokenErr("DomainID") -} - -// ErrDomainNameWithToken indicates that a DomainName was provided, but token authentication is being used instead.s -type ErrDomainNameWithToken struct{ BaseError } - -func (e ErrDomainNameWithToken) Error() string { - return redundantWithTokenErr("DomainName") -} - -// ErrUsernameOrUserID indicates that neither username nor userID are specified, or both are at once. -type ErrUsernameOrUserID struct{ BaseError } - -func (e ErrUsernameOrUserID) Error() string { - return "Exactly one of Username and UserID must be provided for password authentication" -} - -// ErrDomainIDWithUserID indicates that a DomainID was provided, but unnecessary because a UserID is being used. -type ErrDomainIDWithUserID struct{ BaseError } - -func (e ErrDomainIDWithUserID) Error() string { - return redundantWithUserID("DomainID") -} - -// ErrDomainNameWithUserID indicates that a DomainName was provided, but unnecessary because a UserID is being used. -type ErrDomainNameWithUserID struct{ BaseError } - -func (e ErrDomainNameWithUserID) Error() string { - return redundantWithUserID("DomainName") -} - -// ErrDomainIDOrDomainName indicates that a username was provided, but no domain to scope it. -// It may also indicate that both a DomainID and a DomainName were provided at once. -type ErrDomainIDOrDomainName struct{ BaseError } - -func (e ErrDomainIDOrDomainName) Error() string { - return "You must provide exactly one of DomainID or DomainName to authenticate by Username" -} - -// ErrMissingPassword indicates that no password was provided and no token is available. -type ErrMissingPassword struct{ BaseError } - -func (e ErrMissingPassword) Error() string { - return "You must provide a password to authenticate" -} - -// ErrScopeDomainIDOrDomainName indicates that a domain ID or Name was required in a Scope, but not present. -type ErrScopeDomainIDOrDomainName struct{ BaseError } - -func (e ErrScopeDomainIDOrDomainName) Error() string { - return "You must provide exactly one of DomainID or DomainName in a Scope with ProjectName" -} - -// ErrScopeProjectIDOrProjectName indicates that both a ProjectID and a ProjectName were provided in a Scope. -type ErrScopeProjectIDOrProjectName struct{ BaseError } - -func (e ErrScopeProjectIDOrProjectName) Error() string { - return "You must provide at most one of ProjectID or ProjectName in a Scope" -} - -// ErrScopeProjectIDAlone indicates that a ProjectID was provided with other constraints in a Scope. -type ErrScopeProjectIDAlone struct{ BaseError } - -func (e ErrScopeProjectIDAlone) Error() string { - return "ProjectID must be supplied alone in a Scope" -} - -// ErrScopeEmpty indicates that no credentials were provided in a Scope. -type ErrScopeEmpty struct{ BaseError } - -func (e ErrScopeEmpty) Error() string { - return "You must provide either a Project or Domain in a Scope" -} - -// ErrAppCredMissingSecret indicates that no Application Credential Secret was provided with Application Credential ID or Name -type ErrAppCredMissingSecret struct{ BaseError } - -func (e ErrAppCredMissingSecret) Error() string { - return "You must provide an Application Credential Secret" -} diff --git a/vendor/github.com/gophercloud/gophercloud/go.mod b/vendor/github.com/gophercloud/gophercloud/go.mod deleted file mode 100644 index d1ee3b47..00000000 --- a/vendor/github.com/gophercloud/gophercloud/go.mod +++ /dev/null @@ -1,7 +0,0 @@ -module github.com/gophercloud/gophercloud - -require ( - golang.org/x/crypto v0.0.0-20190211182817-74369b46fc67 - golang.org/x/sys v0.0.0-20190209173611-3b5209105503 // indirect - gopkg.in/yaml.v2 v2.2.2 -) diff --git a/vendor/github.com/gophercloud/gophercloud/go.sum b/vendor/github.com/gophercloud/gophercloud/go.sum deleted file mode 100644 index 33cb0be8..00000000 --- a/vendor/github.com/gophercloud/gophercloud/go.sum +++ /dev/null @@ -1,8 +0,0 @@ -golang.org/x/crypto v0.0.0-20190211182817-74369b46fc67 h1:ng3VDlRp5/DHpSWl02R4rM9I+8M2rhmsuLwAMmkLQWE= -golang.org/x/crypto v0.0.0-20190211182817-74369b46fc67/go.mod h1:6SG95UA2DQfeDnfUPMdvaQW0Q7yPrPDi9nlGo2tz2b4= -golang.org/x/sys v0.0.0-20190209173611-3b5209105503 h1:5SvYFrOM3W8Mexn9/oA44Ji7vhXAZQ9hiP+1Q/DMrWg= -golang.org/x/sys v0.0.0-20190209173611-3b5209105503/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= -gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM= -gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= -gopkg.in/yaml.v2 v2.2.2 h1:ZCJp+EgiOT7lHqUV2J862kp8Qj64Jo6az82+3Td9dZw= -gopkg.in/yaml.v2 v2.2.2/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/auth_env.go b/vendor/github.com/gophercloud/gophercloud/openstack/auth_env.go deleted file mode 100644 index 0e8d90ff..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/auth_env.go +++ /dev/null @@ -1,125 +0,0 @@ -package openstack - -import ( - "os" - - "github.com/gophercloud/gophercloud" -) - -var nilOptions = gophercloud.AuthOptions{} - -/* -AuthOptionsFromEnv fills out an identity.AuthOptions structure with the -settings found on the various OpenStack OS_* environment variables. - -The following variables provide sources of truth: OS_AUTH_URL, OS_USERNAME, -OS_PASSWORD and OS_PROJECT_ID. - -Of these, OS_USERNAME, OS_PASSWORD, and OS_AUTH_URL must have settings, -or an error will result. OS_PROJECT_ID, is optional. - -OS_TENANT_ID and OS_TENANT_NAME are deprecated forms of OS_PROJECT_ID and -OS_PROJECT_NAME and the latter are expected against a v3 auth api. - -If OS_PROJECT_ID and OS_PROJECT_NAME are set, they will still be referred -as "tenant" in Gophercloud. - -If OS_PROJECT_NAME is set, it requires OS_PROJECT_ID to be set as well to -handle projects not on the default domain. - -To use this function, first set the OS_* environment variables (for example, -by sourcing an `openrc` file), then: - - opts, err := openstack.AuthOptionsFromEnv() - provider, err := openstack.AuthenticatedClient(opts) -*/ -func AuthOptionsFromEnv() (gophercloud.AuthOptions, error) { - authURL := os.Getenv("OS_AUTH_URL") - username := os.Getenv("OS_USERNAME") - userID := os.Getenv("OS_USERID") - password := os.Getenv("OS_PASSWORD") - tenantID := os.Getenv("OS_TENANT_ID") - tenantName := os.Getenv("OS_TENANT_NAME") - domainID := os.Getenv("OS_DOMAIN_ID") - domainName := os.Getenv("OS_DOMAIN_NAME") - applicationCredentialID := os.Getenv("OS_APPLICATION_CREDENTIAL_ID") - applicationCredentialName := os.Getenv("OS_APPLICATION_CREDENTIAL_NAME") - applicationCredentialSecret := os.Getenv("OS_APPLICATION_CREDENTIAL_SECRET") - - // If OS_PROJECT_ID is set, overwrite tenantID with the value. - if v := os.Getenv("OS_PROJECT_ID"); v != "" { - tenantID = v - } - - // If OS_PROJECT_NAME is set, overwrite tenantName with the value. - if v := os.Getenv("OS_PROJECT_NAME"); v != "" { - tenantName = v - } - - if authURL == "" { - err := gophercloud.ErrMissingEnvironmentVariable{ - EnvironmentVariable: "OS_AUTH_URL", - } - return nilOptions, err - } - - if userID == "" && username == "" { - // Empty username and userID could be ignored, when applicationCredentialID and applicationCredentialSecret are set - if applicationCredentialID == "" && applicationCredentialSecret == "" { - err := gophercloud.ErrMissingAnyoneOfEnvironmentVariables{ - EnvironmentVariables: []string{"OS_USERID", "OS_USERNAME"}, - } - return nilOptions, err - } - } - - if password == "" && applicationCredentialID == "" && applicationCredentialName == "" { - err := gophercloud.ErrMissingEnvironmentVariable{ - EnvironmentVariable: "OS_PASSWORD", - } - return nilOptions, err - } - - if (applicationCredentialID != "" || applicationCredentialName != "") && applicationCredentialSecret == "" { - err := gophercloud.ErrMissingEnvironmentVariable{ - EnvironmentVariable: "OS_APPLICATION_CREDENTIAL_SECRET", - } - return nilOptions, err - } - - if domainID == "" && domainName == "" && tenantID == "" && tenantName != "" { - err := gophercloud.ErrMissingEnvironmentVariable{ - EnvironmentVariable: "OS_PROJECT_ID", - } - return nilOptions, err - } - - if applicationCredentialID == "" && applicationCredentialName != "" && applicationCredentialSecret != "" { - if userID == "" && username == "" { - return nilOptions, gophercloud.ErrMissingAnyoneOfEnvironmentVariables{ - EnvironmentVariables: []string{"OS_USERID", "OS_USERNAME"}, - } - } - if username != "" && domainID == "" && domainName == "" { - return nilOptions, gophercloud.ErrMissingAnyoneOfEnvironmentVariables{ - EnvironmentVariables: []string{"OS_DOMAIN_ID", "OS_DOMAIN_NAME"}, - } - } - } - - ao := gophercloud.AuthOptions{ - IdentityEndpoint: authURL, - UserID: userID, - Username: username, - Password: password, - TenantID: tenantID, - TenantName: tenantName, - DomainID: domainID, - DomainName: domainName, - ApplicationCredentialID: applicationCredentialID, - ApplicationCredentialName: applicationCredentialName, - ApplicationCredentialSecret: applicationCredentialSecret, - } - - return ao, nil -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/client.go b/vendor/github.com/gophercloud/gophercloud/openstack/client.go deleted file mode 100644 index 50f23971..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/client.go +++ /dev/null @@ -1,438 +0,0 @@ -package openstack - -import ( - "fmt" - "reflect" - - "github.com/gophercloud/gophercloud" - tokens2 "github.com/gophercloud/gophercloud/openstack/identity/v2/tokens" - tokens3 "github.com/gophercloud/gophercloud/openstack/identity/v3/tokens" - "github.com/gophercloud/gophercloud/openstack/utils" -) - -const ( - // v2 represents Keystone v2. - // It should never increase beyond 2.0. - v2 = "v2.0" - - // v3 represents Keystone v3. - // The version can be anything from v3 to v3.x. - v3 = "v3" -) - -/* -NewClient prepares an unauthenticated ProviderClient instance. -Most users will probably prefer using the AuthenticatedClient function -instead. - -This is useful if you wish to explicitly control the version of the identity -service that's used for authentication explicitly, for example. - -A basic example of using this would be: - - ao, err := openstack.AuthOptionsFromEnv() - provider, err := openstack.NewClient(ao.IdentityEndpoint) - client, err := openstack.NewIdentityV3(provider, gophercloud.EndpointOpts{}) -*/ -func NewClient(endpoint string) (*gophercloud.ProviderClient, error) { - base, err := utils.BaseEndpoint(endpoint) - if err != nil { - return nil, err - } - - endpoint = gophercloud.NormalizeURL(endpoint) - base = gophercloud.NormalizeURL(base) - - p := new(gophercloud.ProviderClient) - p.IdentityBase = base - p.IdentityEndpoint = endpoint - p.UseTokenLock() - - return p, nil -} - -/* -AuthenticatedClient logs in to an OpenStack cloud found at the identity endpoint -specified by the options, acquires a token, and returns a Provider Client -instance that's ready to operate. - -If the full path to a versioned identity endpoint was specified (example: -http://example.com:5000/v3), that path will be used as the endpoint to query. - -If a versionless endpoint was specified (example: http://example.com:5000/), -the endpoint will be queried to determine which versions of the identity service -are available, then chooses the most recent or most supported version. - -Example: - - ao, err := openstack.AuthOptionsFromEnv() - provider, err := openstack.AuthenticatedClient(ao) - client, err := openstack.NewNetworkV2(client, gophercloud.EndpointOpts{ - Region: os.Getenv("OS_REGION_NAME"), - }) -*/ -func AuthenticatedClient(options gophercloud.AuthOptions) (*gophercloud.ProviderClient, error) { - client, err := NewClient(options.IdentityEndpoint) - if err != nil { - return nil, err - } - - err = Authenticate(client, options) - if err != nil { - return nil, err - } - return client, nil -} - -// Authenticate or re-authenticate against the most recent identity service -// supported at the provided endpoint. -func Authenticate(client *gophercloud.ProviderClient, options gophercloud.AuthOptions) error { - versions := []*utils.Version{ - {ID: v2, Priority: 20, Suffix: "/v2.0/"}, - {ID: v3, Priority: 30, Suffix: "/v3/"}, - } - - chosen, endpoint, err := utils.ChooseVersion(client, versions) - if err != nil { - return err - } - - switch chosen.ID { - case v2: - return v2auth(client, endpoint, options, gophercloud.EndpointOpts{}) - case v3: - return v3auth(client, endpoint, &options, gophercloud.EndpointOpts{}) - default: - // The switch statement must be out of date from the versions list. - return fmt.Errorf("Unrecognized identity version: %s", chosen.ID) - } -} - -// AuthenticateV2 explicitly authenticates against the identity v2 endpoint. -func AuthenticateV2(client *gophercloud.ProviderClient, options gophercloud.AuthOptions, eo gophercloud.EndpointOpts) error { - return v2auth(client, "", options, eo) -} - -func v2auth(client *gophercloud.ProviderClient, endpoint string, options gophercloud.AuthOptions, eo gophercloud.EndpointOpts) error { - v2Client, err := NewIdentityV2(client, eo) - if err != nil { - return err - } - - if endpoint != "" { - v2Client.Endpoint = endpoint - } - - v2Opts := tokens2.AuthOptions{ - IdentityEndpoint: options.IdentityEndpoint, - Username: options.Username, - Password: options.Password, - TenantID: options.TenantID, - TenantName: options.TenantName, - AllowReauth: options.AllowReauth, - TokenID: options.TokenID, - } - - result := tokens2.Create(v2Client, v2Opts) - - err = client.SetTokenAndAuthResult(result) - if err != nil { - return err - } - - catalog, err := result.ExtractServiceCatalog() - if err != nil { - return err - } - - if options.AllowReauth { - // here we're creating a throw-away client (tac). it's a copy of the user's provider client, but - // with the token and reauth func zeroed out. combined with setting `AllowReauth` to `false`, - // this should retry authentication only once - tac := *client - tac.SetThrowaway(true) - tac.ReauthFunc = nil - tac.SetTokenAndAuthResult(nil) - tao := options - tao.AllowReauth = false - client.ReauthFunc = func() error { - err := v2auth(&tac, endpoint, tao, eo) - if err != nil { - return err - } - client.CopyTokenFrom(&tac) - return nil - } - } - client.EndpointLocator = func(opts gophercloud.EndpointOpts) (string, error) { - return V2EndpointURL(catalog, opts) - } - - return nil -} - -// AuthenticateV3 explicitly authenticates against the identity v3 service. -func AuthenticateV3(client *gophercloud.ProviderClient, options tokens3.AuthOptionsBuilder, eo gophercloud.EndpointOpts) error { - return v3auth(client, "", options, eo) -} - -func v3auth(client *gophercloud.ProviderClient, endpoint string, opts tokens3.AuthOptionsBuilder, eo gophercloud.EndpointOpts) error { - // Override the generated service endpoint with the one returned by the version endpoint. - v3Client, err := NewIdentityV3(client, eo) - if err != nil { - return err - } - - if endpoint != "" { - v3Client.Endpoint = endpoint - } - - result := tokens3.Create(v3Client, opts) - - err = client.SetTokenAndAuthResult(result) - if err != nil { - return err - } - - catalog, err := result.ExtractServiceCatalog() - if err != nil { - return err - } - - if opts.CanReauth() { - // here we're creating a throw-away client (tac). it's a copy of the user's provider client, but - // with the token and reauth func zeroed out. combined with setting `AllowReauth` to `false`, - // this should retry authentication only once - tac := *client - tac.SetThrowaway(true) - tac.ReauthFunc = nil - tac.SetTokenAndAuthResult(nil) - var tao tokens3.AuthOptionsBuilder - switch ot := opts.(type) { - case *gophercloud.AuthOptions: - o := *ot - o.AllowReauth = false - tao = &o - case *tokens3.AuthOptions: - o := *ot - o.AllowReauth = false - tao = &o - default: - tao = opts - } - client.ReauthFunc = func() error { - err := v3auth(&tac, endpoint, tao, eo) - if err != nil { - return err - } - client.CopyTokenFrom(&tac) - return nil - } - } - client.EndpointLocator = func(opts gophercloud.EndpointOpts) (string, error) { - return V3EndpointURL(catalog, opts) - } - - return nil -} - -// NewIdentityV2 creates a ServiceClient that may be used to interact with the -// v2 identity service. -func NewIdentityV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - endpoint := client.IdentityBase + "v2.0/" - clientType := "identity" - var err error - if !reflect.DeepEqual(eo, gophercloud.EndpointOpts{}) { - eo.ApplyDefaults(clientType) - endpoint, err = client.EndpointLocator(eo) - if err != nil { - return nil, err - } - } - - return &gophercloud.ServiceClient{ - ProviderClient: client, - Endpoint: endpoint, - Type: clientType, - }, nil -} - -// NewIdentityV3 creates a ServiceClient that may be used to access the v3 -// identity service. -func NewIdentityV3(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - endpoint := client.IdentityBase + "v3/" - clientType := "identity" - var err error - if !reflect.DeepEqual(eo, gophercloud.EndpointOpts{}) { - eo.ApplyDefaults(clientType) - endpoint, err = client.EndpointLocator(eo) - if err != nil { - return nil, err - } - } - - // Ensure endpoint still has a suffix of v3. - // This is because EndpointLocator might have found a versionless - // endpoint or the published endpoint is still /v2.0. In both - // cases, we need to fix the endpoint to point to /v3. - base, err := utils.BaseEndpoint(endpoint) - if err != nil { - return nil, err - } - - base = gophercloud.NormalizeURL(base) - - endpoint = base + "v3/" - - return &gophercloud.ServiceClient{ - ProviderClient: client, - Endpoint: endpoint, - Type: clientType, - }, nil -} - -func initClientOpts(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts, clientType string) (*gophercloud.ServiceClient, error) { - sc := new(gophercloud.ServiceClient) - eo.ApplyDefaults(clientType) - url, err := client.EndpointLocator(eo) - if err != nil { - return sc, err - } - sc.ProviderClient = client - sc.Endpoint = url - sc.Type = clientType - return sc, nil -} - -// NewBareMetalV1 creates a ServiceClient that may be used with the v1 -// bare metal package. -func NewBareMetalV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "baremetal") -} - -// NewBareMetalIntrospectionV1 creates a ServiceClient that may be used with the v1 -// bare metal introspection package. -func NewBareMetalIntrospectionV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "baremetal-inspector") -} - -// NewObjectStorageV1 creates a ServiceClient that may be used with the v1 -// object storage package. -func NewObjectStorageV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "object-store") -} - -// NewComputeV2 creates a ServiceClient that may be used with the v2 compute -// package. -func NewComputeV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "compute") -} - -// NewNetworkV2 creates a ServiceClient that may be used with the v2 network -// package. -func NewNetworkV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - sc, err := initClientOpts(client, eo, "network") - sc.ResourceBase = sc.Endpoint + "v2.0/" - return sc, err -} - -// NewBlockStorageV1 creates a ServiceClient that may be used to access the v1 -// block storage service. -func NewBlockStorageV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "volume") -} - -// NewBlockStorageV2 creates a ServiceClient that may be used to access the v2 -// block storage service. -func NewBlockStorageV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "volumev2") -} - -// NewBlockStorageV3 creates a ServiceClient that may be used to access the v3 block storage service. -func NewBlockStorageV3(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "volumev3") -} - -// NewSharedFileSystemV2 creates a ServiceClient that may be used to access the v2 shared file system service. -func NewSharedFileSystemV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "sharev2") -} - -// NewCDNV1 creates a ServiceClient that may be used to access the OpenStack v1 -// CDN service. -func NewCDNV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "cdn") -} - -// NewOrchestrationV1 creates a ServiceClient that may be used to access the v1 -// orchestration service. -func NewOrchestrationV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "orchestration") -} - -// NewDBV1 creates a ServiceClient that may be used to access the v1 DB service. -func NewDBV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "database") -} - -// NewDNSV2 creates a ServiceClient that may be used to access the v2 DNS -// service. -func NewDNSV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - sc, err := initClientOpts(client, eo, "dns") - sc.ResourceBase = sc.Endpoint + "v2/" - return sc, err -} - -// NewImageServiceV2 creates a ServiceClient that may be used to access the v2 -// image service. -func NewImageServiceV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - sc, err := initClientOpts(client, eo, "image") - sc.ResourceBase = sc.Endpoint + "v2/" - return sc, err -} - -// NewLoadBalancerV2 creates a ServiceClient that may be used to access the v2 -// load balancer service. -func NewLoadBalancerV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - sc, err := initClientOpts(client, eo, "load-balancer") - sc.ResourceBase = sc.Endpoint + "v2.0/" - return sc, err -} - -// NewClusteringV1 creates a ServiceClient that may be used with the v1 clustering -// package. -func NewClusteringV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "clustering") -} - -// NewMessagingV2 creates a ServiceClient that may be used with the v2 messaging -// service. -func NewMessagingV2(client *gophercloud.ProviderClient, clientID string, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - sc, err := initClientOpts(client, eo, "messaging") - sc.MoreHeaders = map[string]string{"Client-ID": clientID} - return sc, err -} - -// NewContainerV1 creates a ServiceClient that may be used with v1 container package -func NewContainerV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "container") -} - -// NewKeyManagerV1 creates a ServiceClient that may be used with the v1 key -// manager service. -func NewKeyManagerV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - sc, err := initClientOpts(client, eo, "key-manager") - sc.ResourceBase = sc.Endpoint + "v1/" - return sc, err -} - -// NewContainerInfraV1 creates a ServiceClient that may be used with the v1 container infra management -// package. -func NewContainerInfraV1(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "container-infra") -} - -// NewWorkflowV2 creates a ServiceClient that may be used with the v2 workflow management package. -func NewWorkflowV2(client *gophercloud.ProviderClient, eo gophercloud.EndpointOpts) (*gophercloud.ServiceClient, error) { - return initClientOpts(client, eo, "workflowv2") -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/doc.go b/vendor/github.com/gophercloud/gophercloud/openstack/doc.go deleted file mode 100644 index cedf1f4d..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/doc.go +++ /dev/null @@ -1,14 +0,0 @@ -/* -Package openstack contains resources for the individual OpenStack projects -supported in Gophercloud. It also includes functions to authenticate to an -OpenStack cloud and for provisioning various service-level clients. - -Example of Creating a Service Client - - ao, err := openstack.AuthOptionsFromEnv() - provider, err := openstack.AuthenticatedClient(ao) - client, err := openstack.NewNetworkV2(client, gophercloud.EndpointOpts{ - Region: os.Getenv("OS_REGION_NAME"), - }) -*/ -package openstack diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/endpoint_location.go b/vendor/github.com/gophercloud/gophercloud/openstack/endpoint_location.go deleted file mode 100644 index 12c8aebc..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/endpoint_location.go +++ /dev/null @@ -1,107 +0,0 @@ -package openstack - -import ( - "github.com/gophercloud/gophercloud" - tokens2 "github.com/gophercloud/gophercloud/openstack/identity/v2/tokens" - tokens3 "github.com/gophercloud/gophercloud/openstack/identity/v3/tokens" -) - -/* -V2EndpointURL discovers the endpoint URL for a specific service from a -ServiceCatalog acquired during the v2 identity service. - -The specified EndpointOpts are used to identify a unique, unambiguous endpoint -to return. It's an error both when multiple endpoints match the provided -criteria and when none do. The minimum that can be specified is a Type, but you -will also often need to specify a Name and/or a Region depending on what's -available on your OpenStack deployment. -*/ -func V2EndpointURL(catalog *tokens2.ServiceCatalog, opts gophercloud.EndpointOpts) (string, error) { - // Extract Endpoints from the catalog entries that match the requested Type, Name if provided, and Region if provided. - var endpoints = make([]tokens2.Endpoint, 0, 1) - for _, entry := range catalog.Entries { - if (entry.Type == opts.Type) && (opts.Name == "" || entry.Name == opts.Name) { - for _, endpoint := range entry.Endpoints { - if opts.Region == "" || endpoint.Region == opts.Region { - endpoints = append(endpoints, endpoint) - } - } - } - } - - // Report an error if the options were ambiguous. - if len(endpoints) > 1 { - err := &ErrMultipleMatchingEndpointsV2{} - err.Endpoints = endpoints - return "", err - } - - // Extract the appropriate URL from the matching Endpoint. - for _, endpoint := range endpoints { - switch opts.Availability { - case gophercloud.AvailabilityPublic: - return gophercloud.NormalizeURL(endpoint.PublicURL), nil - case gophercloud.AvailabilityInternal: - return gophercloud.NormalizeURL(endpoint.InternalURL), nil - case gophercloud.AvailabilityAdmin: - return gophercloud.NormalizeURL(endpoint.AdminURL), nil - default: - err := &ErrInvalidAvailabilityProvided{} - err.Argument = "Availability" - err.Value = opts.Availability - return "", err - } - } - - // Report an error if there were no matching endpoints. - err := &gophercloud.ErrEndpointNotFound{} - return "", err -} - -/* -V3EndpointURL discovers the endpoint URL for a specific service from a Catalog -acquired during the v3 identity service. - -The specified EndpointOpts are used to identify a unique, unambiguous endpoint -to return. It's an error both when multiple endpoints match the provided -criteria and when none do. The minimum that can be specified is a Type, but you -will also often need to specify a Name and/or a Region depending on what's -available on your OpenStack deployment. -*/ -func V3EndpointURL(catalog *tokens3.ServiceCatalog, opts gophercloud.EndpointOpts) (string, error) { - // Extract Endpoints from the catalog entries that match the requested Type, Interface, - // Name if provided, and Region if provided. - var endpoints = make([]tokens3.Endpoint, 0, 1) - for _, entry := range catalog.Entries { - if (entry.Type == opts.Type) && (opts.Name == "" || entry.Name == opts.Name) { - for _, endpoint := range entry.Endpoints { - if opts.Availability != gophercloud.AvailabilityAdmin && - opts.Availability != gophercloud.AvailabilityPublic && - opts.Availability != gophercloud.AvailabilityInternal { - err := &ErrInvalidAvailabilityProvided{} - err.Argument = "Availability" - err.Value = opts.Availability - return "", err - } - if (opts.Availability == gophercloud.Availability(endpoint.Interface)) && - (opts.Region == "" || endpoint.Region == opts.Region || endpoint.RegionID == opts.Region) { - endpoints = append(endpoints, endpoint) - } - } - } - } - - // Report an error if the options were ambiguous. - if len(endpoints) > 1 { - return "", ErrMultipleMatchingEndpointsV3{Endpoints: endpoints} - } - - // Extract the URL from the matching Endpoint. - for _, endpoint := range endpoints { - return gophercloud.NormalizeURL(endpoint.URL), nil - } - - // Report an error if there were no matching endpoints. - err := &gophercloud.ErrEndpointNotFound{} - return "", err -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/errors.go b/vendor/github.com/gophercloud/gophercloud/openstack/errors.go deleted file mode 100644 index df410b1c..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/errors.go +++ /dev/null @@ -1,71 +0,0 @@ -package openstack - -import ( - "fmt" - - "github.com/gophercloud/gophercloud" - tokens2 "github.com/gophercloud/gophercloud/openstack/identity/v2/tokens" - tokens3 "github.com/gophercloud/gophercloud/openstack/identity/v3/tokens" -) - -// ErrEndpointNotFound is the error when no suitable endpoint can be found -// in the user's catalog -type ErrEndpointNotFound struct{ gophercloud.BaseError } - -func (e ErrEndpointNotFound) Error() string { - return "No suitable endpoint could be found in the service catalog." -} - -// ErrInvalidAvailabilityProvided is the error when an invalid endpoint -// availability is provided -type ErrInvalidAvailabilityProvided struct{ gophercloud.ErrInvalidInput } - -func (e ErrInvalidAvailabilityProvided) Error() string { - return fmt.Sprintf("Unexpected availability in endpoint query: %s", e.Value) -} - -// ErrMultipleMatchingEndpointsV2 is the error when more than one endpoint -// for the given options is found in the v2 catalog -type ErrMultipleMatchingEndpointsV2 struct { - gophercloud.BaseError - Endpoints []tokens2.Endpoint -} - -func (e ErrMultipleMatchingEndpointsV2) Error() string { - return fmt.Sprintf("Discovered %d matching endpoints: %#v", len(e.Endpoints), e.Endpoints) -} - -// ErrMultipleMatchingEndpointsV3 is the error when more than one endpoint -// for the given options is found in the v3 catalog -type ErrMultipleMatchingEndpointsV3 struct { - gophercloud.BaseError - Endpoints []tokens3.Endpoint -} - -func (e ErrMultipleMatchingEndpointsV3) Error() string { - return fmt.Sprintf("Discovered %d matching endpoints: %#v", len(e.Endpoints), e.Endpoints) -} - -// ErrNoAuthURL is the error when the OS_AUTH_URL environment variable is not -// found -type ErrNoAuthURL struct{ gophercloud.ErrInvalidInput } - -func (e ErrNoAuthURL) Error() string { - return "Environment variable OS_AUTH_URL needs to be set." -} - -// ErrNoUsername is the error when the OS_USERNAME environment variable is not -// found -type ErrNoUsername struct{ gophercloud.ErrInvalidInput } - -func (e ErrNoUsername) Error() string { - return "Environment variable OS_USERNAME needs to be set." -} - -// ErrNoPassword is the error when the OS_PASSWORD environment variable is not -// found -type ErrNoPassword struct{ gophercloud.ErrInvalidInput } - -func (e ErrNoPassword) Error() string { - return "Environment variable OS_PASSWORD needs to be set." -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/doc.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/doc.go deleted file mode 100644 index 45623369..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/doc.go +++ /dev/null @@ -1,65 +0,0 @@ -/* -Package tenants provides information and interaction with the -tenants API resource for the OpenStack Identity service. - -See http://developer.openstack.org/api-ref-identity-v2.html#identity-auth-v2 -and http://developer.openstack.org/api-ref-identity-v2.html#admin-tenants -for more information. - -Example to List Tenants - - listOpts := tenants.ListOpts{ - Limit: 2, - } - - allPages, err := tenants.List(identityClient, listOpts).AllPages() - if err != nil { - panic(err) - } - - allTenants, err := tenants.ExtractTenants(allPages) - if err != nil { - panic(err) - } - - for _, tenant := range allTenants { - fmt.Printf("%+v\n", tenant) - } - -Example to Create a Tenant - - createOpts := tenants.CreateOpts{ - Name: "tenant_name", - Description: "this is a tenant", - Enabled: gophercloud.Enabled, - } - - tenant, err := tenants.Create(identityClient, createOpts).Extract() - if err != nil { - panic(err) - } - -Example to Update a Tenant - - tenantID := "e6db6ed6277c461a853458589063b295" - - updateOpts := tenants.UpdateOpts{ - Description: "this is a new description", - Enabled: gophercloud.Disabled, - } - - tenant, err := tenants.Update(identityClient, tenantID, updateOpts).Extract() - if err != nil { - panic(err) - } - -Example to Delete a Tenant - - tenantID := "e6db6ed6277c461a853458589063b295" - - err := tenants.Delete(identitYClient, tenantID).ExtractErr() - if err != nil { - panic(err) - } -*/ -package tenants diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/requests.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/requests.go deleted file mode 100644 index f21a58f1..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/requests.go +++ /dev/null @@ -1,116 +0,0 @@ -package tenants - -import ( - "github.com/gophercloud/gophercloud" - "github.com/gophercloud/gophercloud/pagination" -) - -// ListOpts filters the Tenants that are returned by the List call. -type ListOpts struct { - // Marker is the ID of the last Tenant on the previous page. - Marker string `q:"marker"` - - // Limit specifies the page size. - Limit int `q:"limit"` -} - -// List enumerates the Tenants to which the current token has access. -func List(client *gophercloud.ServiceClient, opts *ListOpts) pagination.Pager { - url := listURL(client) - if opts != nil { - q, err := gophercloud.BuildQueryString(opts) - if err != nil { - return pagination.Pager{Err: err} - } - url += q.String() - } - return pagination.NewPager(client, url, func(r pagination.PageResult) pagination.Page { - return TenantPage{pagination.LinkedPageBase{PageResult: r}} - }) -} - -// CreateOpts represents the options needed when creating new tenant. -type CreateOpts struct { - // Name is the name of the tenant. - Name string `json:"name" required:"true"` - - // Description is the description of the tenant. - Description string `json:"description,omitempty"` - - // Enabled sets the tenant status to enabled or disabled. - Enabled *bool `json:"enabled,omitempty"` -} - -// CreateOptsBuilder enables extensions to add additional parameters to the -// Create request. -type CreateOptsBuilder interface { - ToTenantCreateMap() (map[string]interface{}, error) -} - -// ToTenantCreateMap assembles a request body based on the contents of -// a CreateOpts. -func (opts CreateOpts) ToTenantCreateMap() (map[string]interface{}, error) { - return gophercloud.BuildRequestBody(opts, "tenant") -} - -// Create is the operation responsible for creating new tenant. -func Create(client *gophercloud.ServiceClient, opts CreateOptsBuilder) (r CreateResult) { - b, err := opts.ToTenantCreateMap() - if err != nil { - r.Err = err - return - } - _, r.Err = client.Post(createURL(client), b, &r.Body, &gophercloud.RequestOpts{ - OkCodes: []int{200, 201}, - }) - return -} - -// Get requests details on a single tenant by ID. -func Get(client *gophercloud.ServiceClient, id string) (r GetResult) { - _, r.Err = client.Get(getURL(client, id), &r.Body, nil) - return -} - -// UpdateOptsBuilder allows extensions to add additional parameters to the -// Update request. -type UpdateOptsBuilder interface { - ToTenantUpdateMap() (map[string]interface{}, error) -} - -// UpdateOpts specifies the base attributes that may be updated on an existing -// tenant. -type UpdateOpts struct { - // Name is the name of the tenant. - Name string `json:"name,omitempty"` - - // Description is the description of the tenant. - Description *string `json:"description,omitempty"` - - // Enabled sets the tenant status to enabled or disabled. - Enabled *bool `json:"enabled,omitempty"` -} - -// ToTenantUpdateMap formats an UpdateOpts structure into a request body. -func (opts UpdateOpts) ToTenantUpdateMap() (map[string]interface{}, error) { - return gophercloud.BuildRequestBody(opts, "tenant") -} - -// Update is the operation responsible for updating exist tenants by their TenantID. -func Update(client *gophercloud.ServiceClient, id string, opts UpdateOptsBuilder) (r UpdateResult) { - b, err := opts.ToTenantUpdateMap() - if err != nil { - r.Err = err - return - } - _, r.Err = client.Put(updateURL(client, id), &b, &r.Body, &gophercloud.RequestOpts{ - OkCodes: []int{200}, - }) - return -} - -// Delete is the operation responsible for permanently deleting a tenant. -func Delete(client *gophercloud.ServiceClient, id string) (r DeleteResult) { - _, r.Err = client.Delete(deleteURL(client, id), nil) - return -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/results.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/results.go deleted file mode 100644 index bb6c2c6b..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/results.go +++ /dev/null @@ -1,91 +0,0 @@ -package tenants - -import ( - "github.com/gophercloud/gophercloud" - "github.com/gophercloud/gophercloud/pagination" -) - -// Tenant is a grouping of users in the identity service. -type Tenant struct { - // ID is a unique identifier for this tenant. - ID string `json:"id"` - - // Name is a friendlier user-facing name for this tenant. - Name string `json:"name"` - - // Description is a human-readable explanation of this Tenant's purpose. - Description string `json:"description"` - - // Enabled indicates whether or not a tenant is active. - Enabled bool `json:"enabled"` -} - -// TenantPage is a single page of Tenant results. -type TenantPage struct { - pagination.LinkedPageBase -} - -// IsEmpty determines whether or not a page of Tenants contains any results. -func (r TenantPage) IsEmpty() (bool, error) { - tenants, err := ExtractTenants(r) - return len(tenants) == 0, err -} - -// NextPageURL extracts the "next" link from the tenants_links section of the result. -func (r TenantPage) NextPageURL() (string, error) { - var s struct { - Links []gophercloud.Link `json:"tenants_links"` - } - err := r.ExtractInto(&s) - if err != nil { - return "", err - } - return gophercloud.ExtractNextURL(s.Links) -} - -// ExtractTenants returns a slice of Tenants contained in a single page of -// results. -func ExtractTenants(r pagination.Page) ([]Tenant, error) { - var s struct { - Tenants []Tenant `json:"tenants"` - } - err := (r.(TenantPage)).ExtractInto(&s) - return s.Tenants, err -} - -type tenantResult struct { - gophercloud.Result -} - -// Extract interprets any tenantResults as a Tenant. -func (r tenantResult) Extract() (*Tenant, error) { - var s struct { - Tenant *Tenant `json:"tenant"` - } - err := r.ExtractInto(&s) - return s.Tenant, err -} - -// GetResult is the response from a Get request. Call its Extract method to -// interpret it as a Tenant. -type GetResult struct { - tenantResult -} - -// CreateResult is the response from a Create request. Call its Extract method -// to interpret it as a Tenant. -type CreateResult struct { - tenantResult -} - -// DeleteResult is the response from a Get request. Call its ExtractErr method -// to determine if the call succeeded or failed. -type DeleteResult struct { - gophercloud.ErrResult -} - -// UpdateResult is the response from a Update request. Call its Extract method -// to interpret it as a Tenant. -type UpdateResult struct { - tenantResult -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/urls.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/urls.go deleted file mode 100644 index 0f026690..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tenants/urls.go +++ /dev/null @@ -1,23 +0,0 @@ -package tenants - -import "github.com/gophercloud/gophercloud" - -func listURL(client *gophercloud.ServiceClient) string { - return client.ServiceURL("tenants") -} - -func getURL(client *gophercloud.ServiceClient, tenantID string) string { - return client.ServiceURL("tenants", tenantID) -} - -func createURL(client *gophercloud.ServiceClient) string { - return client.ServiceURL("tenants") -} - -func deleteURL(client *gophercloud.ServiceClient, tenantID string) string { - return client.ServiceURL("tenants", tenantID) -} - -func updateURL(client *gophercloud.ServiceClient, tenantID string) string { - return client.ServiceURL("tenants", tenantID) -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/doc.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/doc.go deleted file mode 100644 index 5375eea8..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/doc.go +++ /dev/null @@ -1,46 +0,0 @@ -/* -Package tokens provides information and interaction with the token API -resource for the OpenStack Identity service. - -For more information, see: -http://developer.openstack.org/api-ref-identity-v2.html#identity-auth-v2 - -Example to Create an Unscoped Token from a Password - - authOpts := gophercloud.AuthOptions{ - Username: "user", - Password: "pass" - } - - token, err := tokens.Create(identityClient, authOpts).ExtractToken() - if err != nil { - panic(err) - } - -Example to Create a Token from a Tenant ID and Password - - authOpts := gophercloud.AuthOptions{ - Username: "user", - Password: "password", - TenantID: "fc394f2ab2df4114bde39905f800dc57" - } - - token, err := tokens.Create(identityClient, authOpts).ExtractToken() - if err != nil { - panic(err) - } - -Example to Create a Token from a Tenant Name and Password - - authOpts := gophercloud.AuthOptions{ - Username: "user", - Password: "password", - TenantName: "tenantname" - } - - token, err := tokens.Create(identityClient, authOpts).ExtractToken() - if err != nil { - panic(err) - } -*/ -package tokens diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/requests.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/requests.go deleted file mode 100644 index ab32368c..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/requests.go +++ /dev/null @@ -1,103 +0,0 @@ -package tokens - -import "github.com/gophercloud/gophercloud" - -// PasswordCredentialsV2 represents the required options to authenticate -// with a username and password. -type PasswordCredentialsV2 struct { - Username string `json:"username" required:"true"` - Password string `json:"password" required:"true"` -} - -// TokenCredentialsV2 represents the required options to authenticate -// with a token. -type TokenCredentialsV2 struct { - ID string `json:"id,omitempty" required:"true"` -} - -// AuthOptionsV2 wraps a gophercloud AuthOptions in order to adhere to the -// AuthOptionsBuilder interface. -type AuthOptionsV2 struct { - PasswordCredentials *PasswordCredentialsV2 `json:"passwordCredentials,omitempty" xor:"TokenCredentials"` - - // The TenantID and TenantName fields are optional for the Identity V2 API. - // Some providers allow you to specify a TenantName instead of the TenantId. - // Some require both. Your provider's authentication policies will determine - // how these fields influence authentication. - TenantID string `json:"tenantId,omitempty"` - TenantName string `json:"tenantName,omitempty"` - - // TokenCredentials allows users to authenticate (possibly as another user) - // with an authentication token ID. - TokenCredentials *TokenCredentialsV2 `json:"token,omitempty" xor:"PasswordCredentials"` -} - -// AuthOptionsBuilder allows extensions to add additional parameters to the -// token create request. -type AuthOptionsBuilder interface { - // ToTokenCreateMap assembles the Create request body, returning an error - // if parameters are missing or inconsistent. - ToTokenV2CreateMap() (map[string]interface{}, error) -} - -// AuthOptions are the valid options for Openstack Identity v2 authentication. -// For field descriptions, see gophercloud.AuthOptions. -type AuthOptions struct { - IdentityEndpoint string `json:"-"` - Username string `json:"username,omitempty"` - Password string `json:"password,omitempty"` - TenantID string `json:"tenantId,omitempty"` - TenantName string `json:"tenantName,omitempty"` - AllowReauth bool `json:"-"` - TokenID string -} - -// ToTokenV2CreateMap builds a token request body from the given AuthOptions. -func (opts AuthOptions) ToTokenV2CreateMap() (map[string]interface{}, error) { - v2Opts := AuthOptionsV2{ - TenantID: opts.TenantID, - TenantName: opts.TenantName, - } - - if opts.Password != "" { - v2Opts.PasswordCredentials = &PasswordCredentialsV2{ - Username: opts.Username, - Password: opts.Password, - } - } else { - v2Opts.TokenCredentials = &TokenCredentialsV2{ - ID: opts.TokenID, - } - } - - b, err := gophercloud.BuildRequestBody(v2Opts, "auth") - if err != nil { - return nil, err - } - return b, nil -} - -// Create authenticates to the identity service and attempts to acquire a Token. -// Generally, rather than interact with this call directly, end users should -// call openstack.AuthenticatedClient(), which abstracts all of the gory details -// about navigating service catalogs and such. -func Create(client *gophercloud.ServiceClient, auth AuthOptionsBuilder) (r CreateResult) { - b, err := auth.ToTokenV2CreateMap() - if err != nil { - r.Err = err - return - } - _, r.Err = client.Post(CreateURL(client), b, &r.Body, &gophercloud.RequestOpts{ - OkCodes: []int{200, 203}, - MoreHeaders: map[string]string{"X-Auth-Token": ""}, - }) - return -} - -// Get validates and retrieves information for user's token. -func Get(client *gophercloud.ServiceClient, token string) (r GetResult) { - _, r.Err = client.Get(GetURL(client, token), &r.Body, &gophercloud.RequestOpts{ - OkCodes: []int{200, 203}, - }) - return -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/results.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/results.go deleted file mode 100644 index ee5da37f..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/results.go +++ /dev/null @@ -1,174 +0,0 @@ -package tokens - -import ( - "time" - - "github.com/gophercloud/gophercloud" - "github.com/gophercloud/gophercloud/openstack/identity/v2/tenants" -) - -// Token provides only the most basic information related to an authentication -// token. -type Token struct { - // ID provides the primary means of identifying a user to the OpenStack API. - // OpenStack defines this field as an opaque value, so do not depend on its - // content. It is safe, however, to compare for equality. - ID string - - // ExpiresAt provides a timestamp in ISO 8601 format, indicating when the - // authentication token becomes invalid. After this point in time, future - // API requests made using this authentication token will respond with - // errors. Either the caller will need to reauthenticate manually, or more - // preferably, the caller should exploit automatic re-authentication. - // See the AuthOptions structure for more details. - ExpiresAt time.Time - - // Tenant provides information about the tenant to which this token grants - // access. - Tenant tenants.Tenant -} - -// Role is a role for a user. -type Role struct { - Name string `json:"name"` -} - -// User is an OpenStack user. -type User struct { - ID string `json:"id"` - Name string `json:"name"` - UserName string `json:"username"` - Roles []Role `json:"roles"` -} - -// Endpoint represents a single API endpoint offered by a service. -// It provides the public and internal URLs, if supported, along with a region -// specifier, again if provided. -// -// The significance of the Region field will depend upon your provider. -// -// In addition, the interface offered by the service will have version -// information associated with it through the VersionId, VersionInfo, and -// VersionList fields, if provided or supported. -// -// In all cases, fields which aren't supported by the provider and service -// combined will assume a zero-value (""). -type Endpoint struct { - TenantID string `json:"tenantId"` - PublicURL string `json:"publicURL"` - InternalURL string `json:"internalURL"` - AdminURL string `json:"adminURL"` - Region string `json:"region"` - VersionID string `json:"versionId"` - VersionInfo string `json:"versionInfo"` - VersionList string `json:"versionList"` -} - -// CatalogEntry provides a type-safe interface to an Identity API V2 service -// catalog listing. -// -// Each class of service, such as cloud DNS or block storage services, will have -// a single CatalogEntry representing it. -// -// Note: when looking for the desired service, try, whenever possible, to key -// off the type field. Otherwise, you'll tie the representation of the service -// to a specific provider. -type CatalogEntry struct { - // Name will contain the provider-specified name for the service. - Name string `json:"name"` - - // Type will contain a type string if OpenStack defines a type for the - // service. Otherwise, for provider-specific services, the provider may assign - // their own type strings. - Type string `json:"type"` - - // Endpoints will let the caller iterate over all the different endpoints that - // may exist for the service. - Endpoints []Endpoint `json:"endpoints"` -} - -// ServiceCatalog provides a view into the service catalog from a previous, -// successful authentication. -type ServiceCatalog struct { - Entries []CatalogEntry -} - -// CreateResult is the response from a Create request. Use ExtractToken() to -// interpret it as a Token, or ExtractServiceCatalog() to interpret it as a -// service catalog. -type CreateResult struct { - gophercloud.Result -} - -// GetResult is the deferred response from a Get call, which is the same with a -// Created token. Use ExtractUser() to interpret it as a User. -type GetResult struct { - CreateResult -} - -// ExtractToken returns the just-created Token from a CreateResult. -func (r CreateResult) ExtractToken() (*Token, error) { - var s struct { - Access struct { - Token struct { - Expires string `json:"expires"` - ID string `json:"id"` - Tenant tenants.Tenant `json:"tenant"` - } `json:"token"` - } `json:"access"` - } - - err := r.ExtractInto(&s) - if err != nil { - return nil, err - } - - expiresTs, err := time.Parse(gophercloud.RFC3339Milli, s.Access.Token.Expires) - if err != nil { - return nil, err - } - - return &Token{ - ID: s.Access.Token.ID, - ExpiresAt: expiresTs, - Tenant: s.Access.Token.Tenant, - }, nil -} - -// ExtractTokenID implements the gophercloud.AuthResult interface. The returned -// string is the same as the ID field of the Token struct returned from -// ExtractToken(). -func (r CreateResult) ExtractTokenID() (string, error) { - var s struct { - Access struct { - Token struct { - ID string `json:"id"` - } `json:"token"` - } `json:"access"` - } - err := r.ExtractInto(&s) - return s.Access.Token.ID, err -} - -// ExtractServiceCatalog returns the ServiceCatalog that was generated along -// with the user's Token. -func (r CreateResult) ExtractServiceCatalog() (*ServiceCatalog, error) { - var s struct { - Access struct { - Entries []CatalogEntry `json:"serviceCatalog"` - } `json:"access"` - } - err := r.ExtractInto(&s) - return &ServiceCatalog{Entries: s.Access.Entries}, err -} - -// ExtractUser returns the User from a GetResult. -func (r GetResult) ExtractUser() (*User, error) { - var s struct { - Access struct { - User User `json:"user"` - } `json:"access"` - } - err := r.ExtractInto(&s) - return &s.Access.User, err -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/urls.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/urls.go deleted file mode 100644 index ee0a28f2..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v2/tokens/urls.go +++ /dev/null @@ -1,13 +0,0 @@ -package tokens - -import "github.com/gophercloud/gophercloud" - -// CreateURL generates the URL used to create new Tokens. -func CreateURL(client *gophercloud.ServiceClient) string { - return client.ServiceURL("tokens") -} - -// GetURL generates the URL used to Validate Tokens. -func GetURL(client *gophercloud.ServiceClient, token string) string { - return client.ServiceURL("tokens", token) -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/doc.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/doc.go deleted file mode 100644 index 966e128f..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/doc.go +++ /dev/null @@ -1,108 +0,0 @@ -/* -Package tokens provides information and interaction with the token API -resource for the OpenStack Identity service. - -For more information, see: -http://developer.openstack.org/api-ref-identity-v3.html#tokens-v3 - -Example to Create a Token From a Username and Password - - authOptions := tokens.AuthOptions{ - UserID: "username", - Password: "password", - } - - token, err := tokens.Create(identityClient, authOptions).ExtractToken() - if err != nil { - panic(err) - } - -Example to Create a Token From a Username, Password, and Domain - - authOptions := tokens.AuthOptions{ - UserID: "username", - Password: "password", - DomainID: "default", - } - - token, err := tokens.Create(identityClient, authOptions).ExtractToken() - if err != nil { - panic(err) - } - - authOptions = tokens.AuthOptions{ - UserID: "username", - Password: "password", - DomainName: "default", - } - - token, err = tokens.Create(identityClient, authOptions).ExtractToken() - if err != nil { - panic(err) - } - -Example to Create a Token From a Token - - authOptions := tokens.AuthOptions{ - TokenID: "token_id", - } - - token, err := tokens.Create(identityClient, authOptions).ExtractToken() - if err != nil { - panic(err) - } - -Example to Create a Token from a Username and Password with Project ID Scope - - scope := tokens.Scope{ - ProjectID: "0fe36e73809d46aeae6705c39077b1b3", - } - - authOptions := tokens.AuthOptions{ - Scope: &scope, - UserID: "username", - Password: "password", - } - - token, err = tokens.Create(identityClient, authOptions).ExtractToken() - if err != nil { - panic(err) - } - -Example to Create a Token from a Username and Password with Domain ID Scope - - scope := tokens.Scope{ - DomainID: "default", - } - - authOptions := tokens.AuthOptions{ - Scope: &scope, - UserID: "username", - Password: "password", - } - - token, err = tokens.Create(identityClient, authOptions).ExtractToken() - if err != nil { - panic(err) - } - -Example to Create a Token from a Username and Password with Project Name Scope - - scope := tokens.Scope{ - ProjectName: "project_name", - DomainID: "default", - } - - authOptions := tokens.AuthOptions{ - Scope: &scope, - UserID: "username", - Password: "password", - } - - token, err = tokens.Create(identityClient, authOptions).ExtractToken() - if err != nil { - panic(err) - } - -*/ -package tokens diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/requests.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/requests.go deleted file mode 100644 index e4d766b2..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/requests.go +++ /dev/null @@ -1,162 +0,0 @@ -package tokens - -import "github.com/gophercloud/gophercloud" - -// Scope allows a created token to be limited to a specific domain or project. -type Scope struct { - ProjectID string - ProjectName string - DomainID string - DomainName string -} - -// AuthOptionsBuilder provides the ability for extensions to add additional -// parameters to AuthOptions. Extensions must satisfy all required methods. -type AuthOptionsBuilder interface { - // ToTokenV3CreateMap assembles the Create request body, returning an error - // if parameters are missing or inconsistent. - ToTokenV3CreateMap(map[string]interface{}) (map[string]interface{}, error) - ToTokenV3ScopeMap() (map[string]interface{}, error) - CanReauth() bool -} - -// AuthOptions represents options for authenticating a user. -type AuthOptions struct { - // IdentityEndpoint specifies the HTTP endpoint that is required to work with - // the Identity API of the appropriate version. While it's ultimately needed - // by all of the identity services, it will often be populated by a - // provider-level function. - IdentityEndpoint string `json:"-"` - - // Username is required if using Identity V2 API. Consult with your provider's - // control panel to discover your account's username. In Identity V3, either - // UserID or a combination of Username and DomainID or DomainName are needed. - Username string `json:"username,omitempty"` - UserID string `json:"id,omitempty"` - - Password string `json:"password,omitempty"` - - // At most one of DomainID and DomainName must be provided if using Username - // with Identity V3. Otherwise, either are optional. - DomainID string `json:"-"` - DomainName string `json:"name,omitempty"` - - // AllowReauth should be set to true if you grant permission for Gophercloud - // to cache your credentials in memory, and to allow Gophercloud to attempt - // to re-authenticate automatically if/when your token expires. If you set - // it to false, it will not cache these settings, but re-authentication will - // not be possible. This setting defaults to false. - AllowReauth bool `json:"-"` - - // TokenID allows users to authenticate (possibly as another user) with an - // authentication token ID. - TokenID string `json:"-"` - - // Authentication through Application Credentials requires supplying name, project and secret - // For project we can use TenantID - ApplicationCredentialID string `json:"-"` - ApplicationCredentialName string `json:"-"` - ApplicationCredentialSecret string `json:"-"` - - Scope Scope `json:"-"` -} - -// ToTokenV3CreateMap builds a request body from AuthOptions. -func (opts *AuthOptions) ToTokenV3CreateMap(scope map[string]interface{}) (map[string]interface{}, error) { - gophercloudAuthOpts := gophercloud.AuthOptions{ - Username: opts.Username, - UserID: opts.UserID, - Password: opts.Password, - DomainID: opts.DomainID, - DomainName: opts.DomainName, - AllowReauth: opts.AllowReauth, - TokenID: opts.TokenID, - ApplicationCredentialID: opts.ApplicationCredentialID, - ApplicationCredentialName: opts.ApplicationCredentialName, - ApplicationCredentialSecret: opts.ApplicationCredentialSecret, - } - - return gophercloudAuthOpts.ToTokenV3CreateMap(scope) -} - -// ToTokenV3CreateMap builds a scope request body from AuthOptions. -func (opts *AuthOptions) ToTokenV3ScopeMap() (map[string]interface{}, error) { - scope := gophercloud.AuthScope(opts.Scope) - - gophercloudAuthOpts := gophercloud.AuthOptions{ - Scope: &scope, - DomainID: opts.DomainID, - DomainName: opts.DomainName, - } - - return gophercloudAuthOpts.ToTokenV3ScopeMap() -} - -func (opts *AuthOptions) CanReauth() bool { - return opts.AllowReauth -} - -func subjectTokenHeaders(c *gophercloud.ServiceClient, subjectToken string) map[string]string { - return map[string]string{ - "X-Subject-Token": subjectToken, - } -} - -// Create authenticates and either generates a new token, or changes the Scope -// of an existing token. -func Create(c *gophercloud.ServiceClient, opts AuthOptionsBuilder) (r CreateResult) { - scope, err := opts.ToTokenV3ScopeMap() - if err != nil { - r.Err = err - return - } - - b, err := opts.ToTokenV3CreateMap(scope) - if err != nil { - r.Err = err - return - } - - resp, err := c.Post(tokenURL(c), b, &r.Body, &gophercloud.RequestOpts{ - MoreHeaders: map[string]string{"X-Auth-Token": ""}, - }) - r.Err = err - if resp != nil { - r.Header = resp.Header - } - return -} - -// Get validates and retrieves information about another token. -func Get(c *gophercloud.ServiceClient, token string) (r GetResult) { - resp, err := c.Get(tokenURL(c), &r.Body, &gophercloud.RequestOpts{ - MoreHeaders: subjectTokenHeaders(c, token), - OkCodes: []int{200, 203}, - }) - if resp != nil { - r.Header = resp.Header - } - r.Err = err - return -} - -// Validate determines if a specified token is valid or not. -func Validate(c *gophercloud.ServiceClient, token string) (bool, error) { - resp, err := c.Head(tokenURL(c), &gophercloud.RequestOpts{ - MoreHeaders: subjectTokenHeaders(c, token), - OkCodes: []int{200, 204, 404}, - }) - if err != nil { - return false, err - } - - return resp.StatusCode == 200 || resp.StatusCode == 204, nil -} - -// Revoke immediately makes specified token invalid. -func Revoke(c *gophercloud.ServiceClient, token string) (r RevokeResult) { - _, r.Err = c.Delete(tokenURL(c), &gophercloud.RequestOpts{ - MoreHeaders: subjectTokenHeaders(c, token), - }) - return -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/results.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/results.go deleted file mode 100644 index 6f26c96b..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/results.go +++ /dev/null @@ -1,178 +0,0 @@ -package tokens - -import ( - "time" - - "github.com/gophercloud/gophercloud" -) - -// Endpoint represents a single API endpoint offered by a service. -// It matches either a public, internal or admin URL. -// If supported, it contains a region specifier, again if provided. -// The significance of the Region field will depend upon your provider. -type Endpoint struct { - ID string `json:"id"` - Region string `json:"region"` - RegionID string `json:"region_id"` - Interface string `json:"interface"` - URL string `json:"url"` -} - -// CatalogEntry provides a type-safe interface to an Identity API V3 service -// catalog listing. Each class of service, such as cloud DNS or block storage -// services, could have multiple CatalogEntry representing it (one by interface -// type, e.g public, admin or internal). -// -// Note: when looking for the desired service, try, whenever possible, to key -// off the type field. Otherwise, you'll tie the representation of the service -// to a specific provider. -type CatalogEntry struct { - // Service ID - ID string `json:"id"` - - // Name will contain the provider-specified name for the service. - Name string `json:"name"` - - // Type will contain a type string if OpenStack defines a type for the - // service. Otherwise, for provider-specific services, the provider may - // assign their own type strings. - Type string `json:"type"` - - // Endpoints will let the caller iterate over all the different endpoints that - // may exist for the service. - Endpoints []Endpoint `json:"endpoints"` -} - -// ServiceCatalog provides a view into the service catalog from a previous, -// successful authentication. -type ServiceCatalog struct { - Entries []CatalogEntry `json:"catalog"` -} - -// Domain provides information about the domain to which this token grants -// access. -type Domain struct { - ID string `json:"id"` - Name string `json:"name"` -} - -// User represents a user resource that exists in the Identity Service. -type User struct { - Domain Domain `json:"domain"` - ID string `json:"id"` - Name string `json:"name"` -} - -// Role provides information about roles to which User is authorized. -type Role struct { - ID string `json:"id"` - Name string `json:"name"` -} - -// Project provides information about project to which User is authorized. -type Project struct { - Domain Domain `json:"domain"` - ID string `json:"id"` - Name string `json:"name"` -} - -// commonResult is the response from a request. A commonResult has various -// methods which can be used to extract different details about the result. -type commonResult struct { - gophercloud.Result -} - -// Extract is a shortcut for ExtractToken. -// This function is deprecated and still present for backward compatibility. -func (r commonResult) Extract() (*Token, error) { - return r.ExtractToken() -} - -// ExtractToken interprets a commonResult as a Token. -func (r commonResult) ExtractToken() (*Token, error) { - var s Token - err := r.ExtractInto(&s) - if err != nil { - return nil, err - } - - // Parse the token itself from the stored headers. - s.ID = r.Header.Get("X-Subject-Token") - - return &s, err -} - -// ExtractTokenID implements the gophercloud.AuthResult interface. The returned -// string is the same as the ID field of the Token struct returned from -// ExtractToken(). -func (r CreateResult) ExtractTokenID() (string, error) { - return r.Header.Get("X-Subject-Token"), r.Err -} - -// ExtractServiceCatalog returns the ServiceCatalog that was generated along -// with the user's Token. -func (r commonResult) ExtractServiceCatalog() (*ServiceCatalog, error) { - var s ServiceCatalog - err := r.ExtractInto(&s) - return &s, err -} - -// ExtractUser returns the User that is the owner of the Token. -func (r commonResult) ExtractUser() (*User, error) { - var s struct { - User *User `json:"user"` - } - err := r.ExtractInto(&s) - return s.User, err -} - -// ExtractRoles returns Roles to which User is authorized. -func (r commonResult) ExtractRoles() ([]Role, error) { - var s struct { - Roles []Role `json:"roles"` - } - err := r.ExtractInto(&s) - return s.Roles, err -} - -// ExtractProject returns Project to which User is authorized. -func (r commonResult) ExtractProject() (*Project, error) { - var s struct { - Project *Project `json:"project"` - } - err := r.ExtractInto(&s) - return s.Project, err -} - -// CreateResult is the response from a Create request. Use ExtractToken() -// to interpret it as a Token, or ExtractServiceCatalog() to interpret it -// as a service catalog. -type CreateResult struct { - commonResult -} - -// GetResult is the response from a Get request. Use ExtractToken() -// to interpret it as a Token, or ExtractServiceCatalog() to interpret it -// as a service catalog. -type GetResult struct { - commonResult -} - -// RevokeResult is response from a Revoke request. -type RevokeResult struct { - commonResult -} - -// Token is a string that grants a user access to a controlled set of services -// in an OpenStack provider. Each Token is valid for a set length of time. -type Token struct { - // ID is the issued token. - ID string `json:"id"` - - // ExpiresAt is the timestamp at which this token will no longer be accepted. - ExpiresAt time.Time `json:"expires_at"` -} - -func (r commonResult) ExtractInto(v interface{}) error { - return r.ExtractIntoStructPtr(v, "token") -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/urls.go b/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/urls.go deleted file mode 100644 index 2f864a31..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/identity/v3/tokens/urls.go +++ /dev/null @@ -1,7 +0,0 @@ -package tokens - -import "github.com/gophercloud/gophercloud" - -func tokenURL(c *gophercloud.ServiceClient) string { - return c.ServiceURL("auth", "tokens") -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/utils/base_endpoint.go b/vendor/github.com/gophercloud/gophercloud/openstack/utils/base_endpoint.go deleted file mode 100644 index 40080f7a..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/utils/base_endpoint.go +++ /dev/null @@ -1,28 +0,0 @@ -package utils - -import ( - "net/url" - "regexp" - "strings" -) - -// BaseEndpoint will return a URL without the /vX.Y -// portion of the URL. -func BaseEndpoint(endpoint string) (string, error) { - u, err := url.Parse(endpoint) - if err != nil { - return "", err - } - - u.RawQuery, u.Fragment = "", "" - - path := u.Path - versionRe := regexp.MustCompile("v[0-9.]+/?") - - if version := versionRe.FindString(path); version != "" { - versionIndex := strings.Index(path, version) - u.Path = path[:versionIndex] - } - - return u.String(), nil -} diff --git a/vendor/github.com/gophercloud/gophercloud/openstack/utils/choose_version.go b/vendor/github.com/gophercloud/gophercloud/openstack/utils/choose_version.go deleted file mode 100644 index 27da19f9..00000000 --- a/vendor/github.com/gophercloud/gophercloud/openstack/utils/choose_version.go +++ /dev/null @@ -1,111 +0,0 @@ -package utils - -import ( - "fmt" - "strings" - - "github.com/gophercloud/gophercloud" -) - -// Version is a supported API version, corresponding to a vN package within the appropriate service. -type Version struct { - ID string - Suffix string - Priority int -} - -var goodStatus = map[string]bool{ - "current": true, - "supported": true, - "stable": true, -} - -// ChooseVersion queries the base endpoint of an API to choose the most recent non-experimental alternative from a service's -// published versions. -// It returns the highest-Priority Version among the alternatives that are provided, as well as its corresponding endpoint. -func ChooseVersion(client *gophercloud.ProviderClient, recognized []*Version) (*Version, string, error) { - type linkResp struct { - Href string `json:"href"` - Rel string `json:"rel"` - } - - type valueResp struct { - ID string `json:"id"` - Status string `json:"status"` - Links []linkResp `json:"links"` - } - - type versionsResp struct { - Values []valueResp `json:"values"` - } - - type response struct { - Versions versionsResp `json:"versions"` - } - - normalize := func(endpoint string) string { - if !strings.HasSuffix(endpoint, "/") { - return endpoint + "/" - } - return endpoint - } - identityEndpoint := normalize(client.IdentityEndpoint) - - // If a full endpoint is specified, check version suffixes for a match first. - for _, v := range recognized { - if strings.HasSuffix(identityEndpoint, v.Suffix) { - return v, identityEndpoint, nil - } - } - - var resp response - _, err := client.Request("GET", client.IdentityBase, &gophercloud.RequestOpts{ - JSONResponse: &resp, - OkCodes: []int{200, 300}, - }) - - if err != nil { - return nil, "", err - } - - var highest *Version - var endpoint string - - for _, value := range resp.Versions.Values { - href := "" - for _, link := range value.Links { - if link.Rel == "self" { - href = normalize(link.Href) - } - } - - for _, version := range recognized { - if strings.Contains(value.ID, version.ID) { - // Prefer a version that exactly matches the provided endpoint. - if href == identityEndpoint { - if href == "" { - return nil, "", fmt.Errorf("Endpoint missing in version %s response from %s", value.ID, client.IdentityBase) - } - return version, href, nil - } - - // Otherwise, find the highest-priority version with a whitelisted status. - if goodStatus[strings.ToLower(value.Status)] { - if highest == nil || version.Priority > highest.Priority { - highest = version - endpoint = href - } - } - } - } - } - - if highest == nil { - return nil, "", fmt.Errorf("No supported version available from endpoint %s", client.IdentityBase) - } - if endpoint == "" { - return nil, "", fmt.Errorf("Endpoint missing in version %s response from %s", highest.ID, client.IdentityBase) - } - - return highest, endpoint, nil -} diff --git a/vendor/github.com/gophercloud/gophercloud/pagination/http.go b/vendor/github.com/gophercloud/gophercloud/pagination/http.go deleted file mode 100644 index 757295c4..00000000 --- a/vendor/github.com/gophercloud/gophercloud/pagination/http.go +++ /dev/null @@ -1,60 +0,0 @@ -package pagination - -import ( - "encoding/json" - "io/ioutil" - "net/http" - "net/url" - "strings" - - "github.com/gophercloud/gophercloud" -) - -// PageResult stores the HTTP response that returned the current page of results. -type PageResult struct { - gophercloud.Result - url.URL -} - -// PageResultFrom parses an HTTP response as JSON and returns a PageResult containing the -// results, interpreting it as JSON if the content type indicates. -func PageResultFrom(resp *http.Response) (PageResult, error) { - var parsedBody interface{} - - defer resp.Body.Close() - rawBody, err := ioutil.ReadAll(resp.Body) - if err != nil { - return PageResult{}, err - } - - if strings.HasPrefix(resp.Header.Get("Content-Type"), "application/json") { - err = json.Unmarshal(rawBody, &parsedBody) - if err != nil { - return PageResult{}, err - } - } else { - parsedBody = rawBody - } - - return PageResultFromParsed(resp, parsedBody), err -} - -// PageResultFromParsed constructs a PageResult from an HTTP response that has already had its -// body parsed as JSON (and closed). -func PageResultFromParsed(resp *http.Response, body interface{}) PageResult { - return PageResult{ - Result: gophercloud.Result{ - Body: body, - Header: resp.Header, - }, - URL: *resp.Request.URL, - } -} - -// Request performs an HTTP request and extracts the http.Response from the result. -func Request(client *gophercloud.ServiceClient, headers map[string]string, url string) (*http.Response, error) { - return client.Get(url, nil, &gophercloud.RequestOpts{ - MoreHeaders: headers, - OkCodes: []int{200, 204, 300}, - }) -} diff --git a/vendor/github.com/gophercloud/gophercloud/pagination/linked.go b/vendor/github.com/gophercloud/gophercloud/pagination/linked.go deleted file mode 100644 index 3656fb7f..00000000 --- a/vendor/github.com/gophercloud/gophercloud/pagination/linked.go +++ /dev/null @@ -1,92 +0,0 @@ -package pagination - -import ( - "fmt" - "reflect" - - "github.com/gophercloud/gophercloud" -) - -// LinkedPageBase may be embedded to implement a page that provides navigational "Next" and "Previous" links within its result. -type LinkedPageBase struct { - PageResult - - // LinkPath lists the keys that should be traversed within a response to arrive at the "next" pointer. - // If any link along the path is missing, an empty URL will be returned. - // If any link results in an unexpected value type, an error will be returned. - // When left as "nil", []string{"links", "next"} will be used as a default. - LinkPath []string -} - -// NextPageURL extracts the pagination structure from a JSON response and returns the "next" link, if one is present. -// It assumes that the links are available in a "links" element of the top-level response object. -// If this is not the case, override NextPageURL on your result type. -func (current LinkedPageBase) NextPageURL() (string, error) { - var path []string - var key string - - if current.LinkPath == nil { - path = []string{"links", "next"} - } else { - path = current.LinkPath - } - - submap, ok := current.Body.(map[string]interface{}) - if !ok { - err := gophercloud.ErrUnexpectedType{} - err.Expected = "map[string]interface{}" - err.Actual = fmt.Sprintf("%v", reflect.TypeOf(current.Body)) - return "", err - } - - for { - key, path = path[0], path[1:len(path)] - - value, ok := submap[key] - if !ok { - return "", nil - } - - if len(path) > 0 { - submap, ok = value.(map[string]interface{}) - if !ok { - err := gophercloud.ErrUnexpectedType{} - err.Expected = "map[string]interface{}" - err.Actual = fmt.Sprintf("%v", reflect.TypeOf(value)) - return "", err - } - } else { - if value == nil { - // Actual null element. - return "", nil - } - - url, ok := value.(string) - if !ok { - err := gophercloud.ErrUnexpectedType{} - err.Expected = "string" - err.Actual = fmt.Sprintf("%v", reflect.TypeOf(value)) - return "", err - } - - return url, nil - } - } -} - -// IsEmpty satisifies the IsEmpty method of the Page interface -func (current LinkedPageBase) IsEmpty() (bool, error) { - if b, ok := current.Body.([]interface{}); ok { - return len(b) == 0, nil - } - err := gophercloud.ErrUnexpectedType{} - err.Expected = "[]interface{}" - err.Actual = fmt.Sprintf("%v", reflect.TypeOf(current.Body)) - return true, err -} - -// GetBody returns the linked page's body. This method is needed to satisfy the -// Page interface. -func (current LinkedPageBase) GetBody() interface{} { - return current.Body -} diff --git a/vendor/github.com/gophercloud/gophercloud/pagination/marker.go b/vendor/github.com/gophercloud/gophercloud/pagination/marker.go deleted file mode 100644 index 52e53bae..00000000 --- a/vendor/github.com/gophercloud/gophercloud/pagination/marker.go +++ /dev/null @@ -1,58 +0,0 @@ -package pagination - -import ( - "fmt" - "reflect" - - "github.com/gophercloud/gophercloud" -) - -// MarkerPage is a stricter Page interface that describes additional functionality required for use with NewMarkerPager. -// For convenience, embed the MarkedPageBase struct. -type MarkerPage interface { - Page - - // LastMarker returns the last "marker" value on this page. - LastMarker() (string, error) -} - -// MarkerPageBase is a page in a collection that's paginated by "limit" and "marker" query parameters. -type MarkerPageBase struct { - PageResult - - // Owner is a reference to the embedding struct. - Owner MarkerPage -} - -// NextPageURL generates the URL for the page of results after this one. -func (current MarkerPageBase) NextPageURL() (string, error) { - currentURL := current.URL - - mark, err := current.Owner.LastMarker() - if err != nil { - return "", err - } - - q := currentURL.Query() - q.Set("marker", mark) - currentURL.RawQuery = q.Encode() - - return currentURL.String(), nil -} - -// IsEmpty satisifies the IsEmpty method of the Page interface -func (current MarkerPageBase) IsEmpty() (bool, error) { - if b, ok := current.Body.([]interface{}); ok { - return len(b) == 0, nil - } - err := gophercloud.ErrUnexpectedType{} - err.Expected = "[]interface{}" - err.Actual = fmt.Sprintf("%v", reflect.TypeOf(current.Body)) - return true, err -} - -// GetBody returns the linked page's body. This method is needed to satisfy the -// Page interface. -func (current MarkerPageBase) GetBody() interface{} { - return current.Body -} diff --git a/vendor/github.com/gophercloud/gophercloud/pagination/pager.go b/vendor/github.com/gophercloud/gophercloud/pagination/pager.go deleted file mode 100644 index 42c0b2db..00000000 --- a/vendor/github.com/gophercloud/gophercloud/pagination/pager.go +++ /dev/null @@ -1,251 +0,0 @@ -package pagination - -import ( - "errors" - "fmt" - "net/http" - "reflect" - "strings" - - "github.com/gophercloud/gophercloud" -) - -var ( - // ErrPageNotAvailable is returned from a Pager when a next or previous page is requested, but does not exist. - ErrPageNotAvailable = errors.New("The requested page does not exist.") -) - -// Page must be satisfied by the result type of any resource collection. -// It allows clients to interact with the resource uniformly, regardless of whether or not or how it's paginated. -// Generally, rather than implementing this interface directly, implementors should embed one of the concrete PageBase structs, -// instead. -// Depending on the pagination strategy of a particular resource, there may be an additional subinterface that the result type -// will need to implement. -type Page interface { - // NextPageURL generates the URL for the page of data that follows this collection. - // Return "" if no such page exists. - NextPageURL() (string, error) - - // IsEmpty returns true if this Page has no items in it. - IsEmpty() (bool, error) - - // GetBody returns the Page Body. This is used in the `AllPages` method. - GetBody() interface{} -} - -// Pager knows how to advance through a specific resource collection, one page at a time. -type Pager struct { - client *gophercloud.ServiceClient - - initialURL string - - createPage func(r PageResult) Page - - firstPage Page - - Err error - - // Headers supplies additional HTTP headers to populate on each paged request. - Headers map[string]string -} - -// NewPager constructs a manually-configured pager. -// Supply the URL for the first page, a function that requests a specific page given a URL, and a function that counts a page. -func NewPager(client *gophercloud.ServiceClient, initialURL string, createPage func(r PageResult) Page) Pager { - return Pager{ - client: client, - initialURL: initialURL, - createPage: createPage, - } -} - -// WithPageCreator returns a new Pager that substitutes a different page creation function. This is -// useful for overriding List functions in delegation. -func (p Pager) WithPageCreator(createPage func(r PageResult) Page) Pager { - return Pager{ - client: p.client, - initialURL: p.initialURL, - createPage: createPage, - } -} - -func (p Pager) fetchNextPage(url string) (Page, error) { - resp, err := Request(p.client, p.Headers, url) - if err != nil { - return nil, err - } - - remembered, err := PageResultFrom(resp) - if err != nil { - return nil, err - } - - return p.createPage(remembered), nil -} - -// EachPage iterates over each page returned by a Pager, yielding one at a time to a handler function. -// Return "false" from the handler to prematurely stop iterating. -func (p Pager) EachPage(handler func(Page) (bool, error)) error { - if p.Err != nil { - return p.Err - } - currentURL := p.initialURL - for { - var currentPage Page - - // if first page has already been fetched, no need to fetch it again - if p.firstPage != nil { - currentPage = p.firstPage - p.firstPage = nil - } else { - var err error - currentPage, err = p.fetchNextPage(currentURL) - if err != nil { - return err - } - } - - empty, err := currentPage.IsEmpty() - if err != nil { - return err - } - if empty { - return nil - } - - ok, err := handler(currentPage) - if err != nil { - return err - } - if !ok { - return nil - } - - currentURL, err = currentPage.NextPageURL() - if err != nil { - return err - } - if currentURL == "" { - return nil - } - } -} - -// AllPages returns all the pages from a `List` operation in a single page, -// allowing the user to retrieve all the pages at once. -func (p Pager) AllPages() (Page, error) { - // pagesSlice holds all the pages until they get converted into as Page Body. - var pagesSlice []interface{} - // body will contain the final concatenated Page body. - var body reflect.Value - - // Grab a first page to ascertain the page body type. - firstPage, err := p.fetchNextPage(p.initialURL) - if err != nil { - return nil, err - } - // Store the page type so we can use reflection to create a new mega-page of - // that type. - pageType := reflect.TypeOf(firstPage) - - // if it's a single page, just return the firstPage (first page) - if _, found := pageType.FieldByName("SinglePageBase"); found { - return firstPage, nil - } - - // store the first page to avoid getting it twice - p.firstPage = firstPage - - // Switch on the page body type. Recognized types are `map[string]interface{}`, - // `[]byte`, and `[]interface{}`. - switch pb := firstPage.GetBody().(type) { - case map[string]interface{}: - // key is the map key for the page body if the body type is `map[string]interface{}`. - var key string - // Iterate over the pages to concatenate the bodies. - err = p.EachPage(func(page Page) (bool, error) { - b := page.GetBody().(map[string]interface{}) - for k, v := range b { - // If it's a linked page, we don't want the `links`, we want the other one. - if !strings.HasSuffix(k, "links") { - // check the field's type. we only want []interface{} (which is really []map[string]interface{}) - switch vt := v.(type) { - case []interface{}: - key = k - pagesSlice = append(pagesSlice, vt...) - } - } - } - return true, nil - }) - if err != nil { - return nil, err - } - // Set body to value of type `map[string]interface{}` - body = reflect.MakeMap(reflect.MapOf(reflect.TypeOf(key), reflect.TypeOf(pagesSlice))) - body.SetMapIndex(reflect.ValueOf(key), reflect.ValueOf(pagesSlice)) - case []byte: - // Iterate over the pages to concatenate the bodies. - err = p.EachPage(func(page Page) (bool, error) { - b := page.GetBody().([]byte) - pagesSlice = append(pagesSlice, b) - // seperate pages with a comma - pagesSlice = append(pagesSlice, []byte{10}) - return true, nil - }) - if err != nil { - return nil, err - } - if len(pagesSlice) > 0 { - // Remove the trailing comma. - pagesSlice = pagesSlice[:len(pagesSlice)-1] - } - var b []byte - // Combine the slice of slices in to a single slice. - for _, slice := range pagesSlice { - b = append(b, slice.([]byte)...) - } - // Set body to value of type `bytes`. - body = reflect.New(reflect.TypeOf(b)).Elem() - body.SetBytes(b) - case []interface{}: - // Iterate over the pages to concatenate the bodies. - err = p.EachPage(func(page Page) (bool, error) { - b := page.GetBody().([]interface{}) - pagesSlice = append(pagesSlice, b...) - return true, nil - }) - if err != nil { - return nil, err - } - // Set body to value of type `[]interface{}` - body = reflect.MakeSlice(reflect.TypeOf(pagesSlice), len(pagesSlice), len(pagesSlice)) - for i, s := range pagesSlice { - body.Index(i).Set(reflect.ValueOf(s)) - } - default: - err := gophercloud.ErrUnexpectedType{} - err.Expected = "map[string]interface{}/[]byte/[]interface{}" - err.Actual = fmt.Sprintf("%T", pb) - return nil, err - } - - // Each `Extract*` function is expecting a specific type of page coming back, - // otherwise the type assertion in those functions will fail. pageType is needed - // to create a type in this method that has the same type that the `Extract*` - // function is expecting and set the Body of that object to the concatenated - // pages. - page := reflect.New(pageType) - // Set the page body to be the concatenated pages. - page.Elem().FieldByName("Body").Set(body) - // Set any additional headers that were pass along. The `objectstorage` pacakge, - // for example, passes a Content-Type header. - h := make(http.Header) - for k, v := range p.Headers { - h.Add(k, v) - } - page.Elem().FieldByName("Header").Set(reflect.ValueOf(h)) - // Type assert the page to a Page interface so that the type assertion in the - // `Extract*` methods will work. - return page.Elem().Interface().(Page), err -} diff --git a/vendor/github.com/gophercloud/gophercloud/pagination/pkg.go b/vendor/github.com/gophercloud/gophercloud/pagination/pkg.go deleted file mode 100644 index 912daea3..00000000 --- a/vendor/github.com/gophercloud/gophercloud/pagination/pkg.go +++ /dev/null @@ -1,4 +0,0 @@ -/* -Package pagination contains utilities and convenience structs that implement common pagination idioms within OpenStack APIs. -*/ -package pagination diff --git a/vendor/github.com/gophercloud/gophercloud/pagination/single.go b/vendor/github.com/gophercloud/gophercloud/pagination/single.go deleted file mode 100644 index 4251d649..00000000 --- a/vendor/github.com/gophercloud/gophercloud/pagination/single.go +++ /dev/null @@ -1,33 +0,0 @@ -package pagination - -import ( - "fmt" - "reflect" - - "github.com/gophercloud/gophercloud" -) - -// SinglePageBase may be embedded in a Page that contains all of the results from an operation at once. -type SinglePageBase PageResult - -// NextPageURL always returns "" to indicate that there are no more pages to return. -func (current SinglePageBase) NextPageURL() (string, error) { - return "", nil -} - -// IsEmpty satisifies the IsEmpty method of the Page interface -func (current SinglePageBase) IsEmpty() (bool, error) { - if b, ok := current.Body.([]interface{}); ok { - return len(b) == 0, nil - } - err := gophercloud.ErrUnexpectedType{} - err.Expected = "[]interface{}" - err.Actual = fmt.Sprintf("%v", reflect.TypeOf(current.Body)) - return true, err -} - -// GetBody returns the single page's body. This method is needed to satisfy the -// Page interface. -func (current SinglePageBase) GetBody() interface{} { - return current.Body -} diff --git a/vendor/github.com/gophercloud/gophercloud/params.go b/vendor/github.com/gophercloud/gophercloud/params.go deleted file mode 100644 index b9986660..00000000 --- a/vendor/github.com/gophercloud/gophercloud/params.go +++ /dev/null @@ -1,491 +0,0 @@ -package gophercloud - -import ( - "encoding/json" - "fmt" - "net/url" - "reflect" - "strconv" - "strings" - "time" -) - -/* -BuildRequestBody builds a map[string]interface from the given `struct`. If -parent is not an empty string, the final map[string]interface returned will -encapsulate the built one. For example: - - disk := 1 - createOpts := flavors.CreateOpts{ - ID: "1", - Name: "m1.tiny", - Disk: &disk, - RAM: 512, - VCPUs: 1, - RxTxFactor: 1.0, - } - - body, err := gophercloud.BuildRequestBody(createOpts, "flavor") - -The above example can be run as-is, however it is recommended to look at how -BuildRequestBody is used within Gophercloud to more fully understand how it -fits within the request process as a whole rather than use it directly as shown -above. -*/ -func BuildRequestBody(opts interface{}, parent string) (map[string]interface{}, error) { - optsValue := reflect.ValueOf(opts) - if optsValue.Kind() == reflect.Ptr { - optsValue = optsValue.Elem() - } - - optsType := reflect.TypeOf(opts) - if optsType.Kind() == reflect.Ptr { - optsType = optsType.Elem() - } - - optsMap := make(map[string]interface{}) - if optsValue.Kind() == reflect.Struct { - //fmt.Printf("optsValue.Kind() is a reflect.Struct: %+v\n", optsValue.Kind()) - for i := 0; i < optsValue.NumField(); i++ { - v := optsValue.Field(i) - f := optsType.Field(i) - - if f.Name != strings.Title(f.Name) { - //fmt.Printf("Skipping field: %s...\n", f.Name) - continue - } - - //fmt.Printf("Starting on field: %s...\n", f.Name) - - zero := isZero(v) - //fmt.Printf("v is zero?: %v\n", zero) - - // if the field has a required tag that's set to "true" - if requiredTag := f.Tag.Get("required"); requiredTag == "true" { - //fmt.Printf("Checking required field [%s]:\n\tv: %+v\n\tisZero:%v\n", f.Name, v.Interface(), zero) - // if the field's value is zero, return a missing-argument error - if zero { - // if the field has a 'required' tag, it can't have a zero-value - err := ErrMissingInput{} - err.Argument = f.Name - return nil, err - } - } - - if xorTag := f.Tag.Get("xor"); xorTag != "" { - //fmt.Printf("Checking `xor` tag for field [%s] with value %+v:\n\txorTag: %s\n", f.Name, v, xorTag) - xorField := optsValue.FieldByName(xorTag) - var xorFieldIsZero bool - if reflect.ValueOf(xorField.Interface()) == reflect.Zero(xorField.Type()) { - xorFieldIsZero = true - } else { - if xorField.Kind() == reflect.Ptr { - xorField = xorField.Elem() - } - xorFieldIsZero = isZero(xorField) - } - if !(zero != xorFieldIsZero) { - err := ErrMissingInput{} - err.Argument = fmt.Sprintf("%s/%s", f.Name, xorTag) - err.Info = fmt.Sprintf("Exactly one of %s and %s must be provided", f.Name, xorTag) - return nil, err - } - } - - if orTag := f.Tag.Get("or"); orTag != "" { - //fmt.Printf("Checking `or` tag for field with:\n\tname: %+v\n\torTag:%s\n", f.Name, orTag) - //fmt.Printf("field is zero?: %v\n", zero) - if zero { - orField := optsValue.FieldByName(orTag) - var orFieldIsZero bool - if reflect.ValueOf(orField.Interface()) == reflect.Zero(orField.Type()) { - orFieldIsZero = true - } else { - if orField.Kind() == reflect.Ptr { - orField = orField.Elem() - } - orFieldIsZero = isZero(orField) - } - if orFieldIsZero { - err := ErrMissingInput{} - err.Argument = fmt.Sprintf("%s/%s", f.Name, orTag) - err.Info = fmt.Sprintf("At least one of %s and %s must be provided", f.Name, orTag) - return nil, err - } - } - } - - jsonTag := f.Tag.Get("json") - if jsonTag == "-" { - continue - } - - if v.Kind() == reflect.Slice || (v.Kind() == reflect.Ptr && v.Elem().Kind() == reflect.Slice) { - sliceValue := v - if sliceValue.Kind() == reflect.Ptr { - sliceValue = sliceValue.Elem() - } - - for i := 0; i < sliceValue.Len(); i++ { - element := sliceValue.Index(i) - if element.Kind() == reflect.Struct || (element.Kind() == reflect.Ptr && element.Elem().Kind() == reflect.Struct) { - _, err := BuildRequestBody(element.Interface(), "") - if err != nil { - return nil, err - } - } - } - } - if v.Kind() == reflect.Struct || (v.Kind() == reflect.Ptr && v.Elem().Kind() == reflect.Struct) { - if zero { - //fmt.Printf("value before change: %+v\n", optsValue.Field(i)) - if jsonTag != "" { - jsonTagPieces := strings.Split(jsonTag, ",") - if len(jsonTagPieces) > 1 && jsonTagPieces[1] == "omitempty" { - if v.CanSet() { - if !v.IsNil() { - if v.Kind() == reflect.Ptr { - v.Set(reflect.Zero(v.Type())) - } - } - //fmt.Printf("value after change: %+v\n", optsValue.Field(i)) - } - } - } - continue - } - - //fmt.Printf("Calling BuildRequestBody with:\n\tv: %+v\n\tf.Name:%s\n", v.Interface(), f.Name) - _, err := BuildRequestBody(v.Interface(), f.Name) - if err != nil { - return nil, err - } - } - } - - //fmt.Printf("opts: %+v \n", opts) - - b, err := json.Marshal(opts) - if err != nil { - return nil, err - } - - //fmt.Printf("string(b): %s\n", string(b)) - - err = json.Unmarshal(b, &optsMap) - if err != nil { - return nil, err - } - - //fmt.Printf("optsMap: %+v\n", optsMap) - - if parent != "" { - optsMap = map[string]interface{}{parent: optsMap} - } - //fmt.Printf("optsMap after parent added: %+v\n", optsMap) - return optsMap, nil - } - // Return an error if the underlying type of 'opts' isn't a struct. - return nil, fmt.Errorf("Options type is not a struct.") -} - -// EnabledState is a convenience type, mostly used in Create and Update -// operations. Because the zero value of a bool is FALSE, we need to use a -// pointer instead to indicate zero-ness. -type EnabledState *bool - -// Convenience vars for EnabledState values. -var ( - iTrue = true - iFalse = false - - Enabled EnabledState = &iTrue - Disabled EnabledState = &iFalse -) - -// IPVersion is a type for the possible IP address versions. Valid instances -// are IPv4 and IPv6 -type IPVersion int - -const ( - // IPv4 is used for IP version 4 addresses - IPv4 IPVersion = 4 - // IPv6 is used for IP version 6 addresses - IPv6 IPVersion = 6 -) - -// IntToPointer is a function for converting integers into integer pointers. -// This is useful when passing in options to operations. -func IntToPointer(i int) *int { - return &i -} - -/* -MaybeString is an internal function to be used by request methods in individual -resource packages. - -It takes a string that might be a zero value and returns either a pointer to its -address or nil. This is useful for allowing users to conveniently omit values -from an options struct by leaving them zeroed, but still pass nil to the JSON -serializer so they'll be omitted from the request body. -*/ -func MaybeString(original string) *string { - if original != "" { - return &original - } - return nil -} - -/* -MaybeInt is an internal function to be used by request methods in individual -resource packages. - -Like MaybeString, it accepts an int that may or may not be a zero value, and -returns either a pointer to its address or nil. It's intended to hint that the -JSON serializer should omit its field. -*/ -func MaybeInt(original int) *int { - if original != 0 { - return &original - } - return nil -} - -/* -func isUnderlyingStructZero(v reflect.Value) bool { - switch v.Kind() { - case reflect.Ptr: - return isUnderlyingStructZero(v.Elem()) - default: - return isZero(v) - } -} -*/ - -var t time.Time - -func isZero(v reflect.Value) bool { - //fmt.Printf("\n\nchecking isZero for value: %+v\n", v) - switch v.Kind() { - case reflect.Ptr: - if v.IsNil() { - return true - } - return false - case reflect.Func, reflect.Map, reflect.Slice: - return v.IsNil() - case reflect.Array: - z := true - for i := 0; i < v.Len(); i++ { - z = z && isZero(v.Index(i)) - } - return z - case reflect.Struct: - if v.Type() == reflect.TypeOf(t) { - if v.Interface().(time.Time).IsZero() { - return true - } - return false - } - z := true - for i := 0; i < v.NumField(); i++ { - z = z && isZero(v.Field(i)) - } - return z - } - // Compare other types directly: - z := reflect.Zero(v.Type()) - //fmt.Printf("zero type for value: %+v\n\n\n", z) - return v.Interface() == z.Interface() -} - -/* -BuildQueryString is an internal function to be used by request methods in -individual resource packages. - -It accepts a tagged structure and expands it into a URL struct. Field names are -converted into query parameters based on a "q" tag. For example: - - type struct Something { - Bar string `q:"x_bar"` - Baz int `q:"lorem_ipsum"` - } - - instance := Something{ - Bar: "AAA", - Baz: "BBB", - } - -will be converted into "?x_bar=AAA&lorem_ipsum=BBB". - -The struct's fields may be strings, integers, or boolean values. Fields left at -their type's zero value will be omitted from the query. -*/ -func BuildQueryString(opts interface{}) (*url.URL, error) { - optsValue := reflect.ValueOf(opts) - if optsValue.Kind() == reflect.Ptr { - optsValue = optsValue.Elem() - } - - optsType := reflect.TypeOf(opts) - if optsType.Kind() == reflect.Ptr { - optsType = optsType.Elem() - } - - params := url.Values{} - - if optsValue.Kind() == reflect.Struct { - for i := 0; i < optsValue.NumField(); i++ { - v := optsValue.Field(i) - f := optsType.Field(i) - qTag := f.Tag.Get("q") - - // if the field has a 'q' tag, it goes in the query string - if qTag != "" { - tags := strings.Split(qTag, ",") - - // if the field is set, add it to the slice of query pieces - if !isZero(v) { - loop: - switch v.Kind() { - case reflect.Ptr: - v = v.Elem() - goto loop - case reflect.String: - params.Add(tags[0], v.String()) - case reflect.Int: - params.Add(tags[0], strconv.FormatInt(v.Int(), 10)) - case reflect.Bool: - params.Add(tags[0], strconv.FormatBool(v.Bool())) - case reflect.Slice: - switch v.Type().Elem() { - case reflect.TypeOf(0): - for i := 0; i < v.Len(); i++ { - params.Add(tags[0], strconv.FormatInt(v.Index(i).Int(), 10)) - } - default: - for i := 0; i < v.Len(); i++ { - params.Add(tags[0], v.Index(i).String()) - } - } - case reflect.Map: - if v.Type().Key().Kind() == reflect.String && v.Type().Elem().Kind() == reflect.String { - var s []string - for _, k := range v.MapKeys() { - value := v.MapIndex(k).String() - s = append(s, fmt.Sprintf("'%s':'%s'", k.String(), value)) - } - params.Add(tags[0], fmt.Sprintf("{%s}", strings.Join(s, ", "))) - } - } - } else { - // if the field has a 'required' tag, it can't have a zero-value - if requiredTag := f.Tag.Get("required"); requiredTag == "true" { - return &url.URL{}, fmt.Errorf("Required query parameter [%s] not set.", f.Name) - } - } - } - } - - return &url.URL{RawQuery: params.Encode()}, nil - } - // Return an error if the underlying type of 'opts' isn't a struct. - return nil, fmt.Errorf("Options type is not a struct.") -} - -/* -BuildHeaders is an internal function to be used by request methods in -individual resource packages. - -It accepts an arbitrary tagged structure and produces a string map that's -suitable for use as the HTTP headers of an outgoing request. Field names are -mapped to header names based in "h" tags. - - type struct Something { - Bar string `h:"x_bar"` - Baz int `h:"lorem_ipsum"` - } - - instance := Something{ - Bar: "AAA", - Baz: "BBB", - } - -will be converted into: - - map[string]string{ - "x_bar": "AAA", - "lorem_ipsum": "BBB", - } - -Untagged fields and fields left at their zero values are skipped. Integers, -booleans and string values are supported. -*/ -func BuildHeaders(opts interface{}) (map[string]string, error) { - optsValue := reflect.ValueOf(opts) - if optsValue.Kind() == reflect.Ptr { - optsValue = optsValue.Elem() - } - - optsType := reflect.TypeOf(opts) - if optsType.Kind() == reflect.Ptr { - optsType = optsType.Elem() - } - - optsMap := make(map[string]string) - if optsValue.Kind() == reflect.Struct { - for i := 0; i < optsValue.NumField(); i++ { - v := optsValue.Field(i) - f := optsType.Field(i) - hTag := f.Tag.Get("h") - - // if the field has a 'h' tag, it goes in the header - if hTag != "" { - tags := strings.Split(hTag, ",") - - // if the field is set, add it to the slice of query pieces - if !isZero(v) { - switch v.Kind() { - case reflect.String: - optsMap[tags[0]] = v.String() - case reflect.Int: - optsMap[tags[0]] = strconv.FormatInt(v.Int(), 10) - case reflect.Bool: - optsMap[tags[0]] = strconv.FormatBool(v.Bool()) - } - } else { - // if the field has a 'required' tag, it can't have a zero-value - if requiredTag := f.Tag.Get("required"); requiredTag == "true" { - return optsMap, fmt.Errorf("Required header [%s] not set.", f.Name) - } - } - } - - } - return optsMap, nil - } - // Return an error if the underlying type of 'opts' isn't a struct. - return optsMap, fmt.Errorf("Options type is not a struct.") -} - -// IDSliceToQueryString takes a slice of elements and converts them into a query -// string. For example, if name=foo and slice=[]int{20, 40, 60}, then the -// result would be `?name=20&name=40&name=60' -func IDSliceToQueryString(name string, ids []int) string { - str := "" - for k, v := range ids { - if k == 0 { - str += "?" - } else { - str += "&" - } - str += fmt.Sprintf("%s=%s", name, strconv.Itoa(v)) - } - return str -} - -// IntWithinRange returns TRUE if an integer falls within a defined range, and -// FALSE if not. -func IntWithinRange(val, min, max int) bool { - return val > min && val < max -} diff --git a/vendor/github.com/gophercloud/gophercloud/provider_client.go b/vendor/github.com/gophercloud/gophercloud/provider_client.go deleted file mode 100644 index fce00462..00000000 --- a/vendor/github.com/gophercloud/gophercloud/provider_client.go +++ /dev/null @@ -1,501 +0,0 @@ -package gophercloud - -import ( - "bytes" - "context" - "encoding/json" - "errors" - "io" - "io/ioutil" - "net/http" - "strings" - "sync" -) - -// DefaultUserAgent is the default User-Agent string set in the request header. -const DefaultUserAgent = "gophercloud/2.0.0" - -// UserAgent represents a User-Agent header. -type UserAgent struct { - // prepend is the slice of User-Agent strings to prepend to DefaultUserAgent. - // All the strings to prepend are accumulated and prepended in the Join method. - prepend []string -} - -// Prepend prepends a user-defined string to the default User-Agent string. Users -// may pass in one or more strings to prepend. -func (ua *UserAgent) Prepend(s ...string) { - ua.prepend = append(s, ua.prepend...) -} - -// Join concatenates all the user-defined User-Agend strings with the default -// Gophercloud User-Agent string. -func (ua *UserAgent) Join() string { - uaSlice := append(ua.prepend, DefaultUserAgent) - return strings.Join(uaSlice, " ") -} - -// ProviderClient stores details that are required to interact with any -// services within a specific provider's API. -// -// Generally, you acquire a ProviderClient by calling the NewClient method in -// the appropriate provider's child package, providing whatever authentication -// credentials are required. -type ProviderClient struct { - // IdentityBase is the base URL used for a particular provider's identity - // service - it will be used when issuing authenticatation requests. It - // should point to the root resource of the identity service, not a specific - // identity version. - IdentityBase string - - // IdentityEndpoint is the identity endpoint. This may be a specific version - // of the identity service. If this is the case, this endpoint is used rather - // than querying versions first. - IdentityEndpoint string - - // TokenID is the ID of the most recently issued valid token. - // NOTE: Aside from within a custom ReauthFunc, this field shouldn't be set by an application. - // To safely read or write this value, call `Token` or `SetToken`, respectively - TokenID string - - // EndpointLocator describes how this provider discovers the endpoints for - // its constituent services. - EndpointLocator EndpointLocator - - // HTTPClient allows users to interject arbitrary http, https, or other transit behaviors. - HTTPClient http.Client - - // UserAgent represents the User-Agent header in the HTTP request. - UserAgent UserAgent - - // ReauthFunc is the function used to re-authenticate the user if the request - // fails with a 401 HTTP response code. This a needed because there may be multiple - // authentication functions for different Identity service versions. - ReauthFunc func() error - - // Throwaway determines whether if this client is a throw-away client. It's a copy of user's provider client - // with the token and reauth func zeroed. Such client can be used to perform reauthorization. - Throwaway bool - - // Context is the context passed to the HTTP request. - Context context.Context - - // mut is a mutex for the client. It protects read and write access to client attributes such as getting - // and setting the TokenID. - mut *sync.RWMutex - - // reauthmut is a mutex for reauthentication it attempts to ensure that only one reauthentication - // attempt happens at one time. - reauthmut *reauthlock - - authResult AuthResult -} - -// reauthlock represents a set of attributes used to help in the reauthentication process. -type reauthlock struct { - sync.RWMutex - reauthing bool - reauthingErr error - done *sync.Cond -} - -// AuthenticatedHeaders returns a map of HTTP headers that are common for all -// authenticated service requests. Blocks if Reauthenticate is in progress. -func (client *ProviderClient) AuthenticatedHeaders() (m map[string]string) { - if client.IsThrowaway() { - return - } - if client.reauthmut != nil { - client.reauthmut.Lock() - for client.reauthmut.reauthing { - client.reauthmut.done.Wait() - } - client.reauthmut.Unlock() - } - t := client.Token() - if t == "" { - return - } - return map[string]string{"X-Auth-Token": t} -} - -// UseTokenLock creates a mutex that is used to allow safe concurrent access to the auth token. -// If the application's ProviderClient is not used concurrently, this doesn't need to be called. -func (client *ProviderClient) UseTokenLock() { - client.mut = new(sync.RWMutex) - client.reauthmut = new(reauthlock) -} - -// GetAuthResult returns the result from the request that was used to obtain a -// provider client's Keystone token. -// -// The result is nil when authentication has not yet taken place, when the token -// was set manually with SetToken(), or when a ReauthFunc was used that does not -// record the AuthResult. -func (client *ProviderClient) GetAuthResult() AuthResult { - if client.mut != nil { - client.mut.RLock() - defer client.mut.RUnlock() - } - return client.authResult -} - -// Token safely reads the value of the auth token from the ProviderClient. Applications should -// call this method to access the token instead of the TokenID field -func (client *ProviderClient) Token() string { - if client.mut != nil { - client.mut.RLock() - defer client.mut.RUnlock() - } - return client.TokenID -} - -// SetToken safely sets the value of the auth token in the ProviderClient. Applications may -// use this method in a custom ReauthFunc. -// -// WARNING: This function is deprecated. Use SetTokenAndAuthResult() instead. -func (client *ProviderClient) SetToken(t string) { - if client.mut != nil { - client.mut.Lock() - defer client.mut.Unlock() - } - client.TokenID = t - client.authResult = nil -} - -// SetTokenAndAuthResult safely sets the value of the auth token in the -// ProviderClient and also records the AuthResult that was returned from the -// token creation request. Applications may call this in a custom ReauthFunc. -func (client *ProviderClient) SetTokenAndAuthResult(r AuthResult) error { - tokenID := "" - var err error - if r != nil { - tokenID, err = r.ExtractTokenID() - if err != nil { - return err - } - } - - if client.mut != nil { - client.mut.Lock() - defer client.mut.Unlock() - } - client.TokenID = tokenID - client.authResult = r - return nil -} - -// CopyTokenFrom safely copies the token from another ProviderClient into the -// this one. -func (client *ProviderClient) CopyTokenFrom(other *ProviderClient) { - if client.mut != nil { - client.mut.Lock() - defer client.mut.Unlock() - } - if other.mut != nil && other.mut != client.mut { - other.mut.RLock() - defer other.mut.RUnlock() - } - client.TokenID = other.TokenID - client.authResult = other.authResult -} - -// IsThrowaway safely reads the value of the client Throwaway field. -func (client *ProviderClient) IsThrowaway() bool { - if client.reauthmut != nil { - client.reauthmut.RLock() - defer client.reauthmut.RUnlock() - } - return client.Throwaway -} - -// SetThrowaway safely sets the value of the client Throwaway field. -func (client *ProviderClient) SetThrowaway(v bool) { - if client.reauthmut != nil { - client.reauthmut.Lock() - defer client.reauthmut.Unlock() - } - client.Throwaway = v -} - -// Reauthenticate calls client.ReauthFunc in a thread-safe way. If this is -// called because of a 401 response, the caller may pass the previous token. In -// this case, the reauthentication can be skipped if another thread has already -// reauthenticated in the meantime. If no previous token is known, an empty -// string should be passed instead to force unconditional reauthentication. -func (client *ProviderClient) Reauthenticate(previousToken string) (err error) { - if client.ReauthFunc == nil { - return nil - } - - if client.reauthmut == nil { - return client.ReauthFunc() - } - - client.reauthmut.Lock() - if client.reauthmut.reauthing { - for !client.reauthmut.reauthing { - client.reauthmut.done.Wait() - } - err = client.reauthmut.reauthingErr - client.reauthmut.Unlock() - return err - } - client.reauthmut.Unlock() - - client.reauthmut.Lock() - client.reauthmut.reauthing = true - client.reauthmut.done = sync.NewCond(client.reauthmut) - client.reauthmut.reauthingErr = nil - client.reauthmut.Unlock() - - if previousToken == "" || client.TokenID == previousToken { - err = client.ReauthFunc() - } - - client.reauthmut.Lock() - client.reauthmut.reauthing = false - client.reauthmut.reauthingErr = err - client.reauthmut.done.Broadcast() - client.reauthmut.Unlock() - return -} - -// RequestOpts customizes the behavior of the provider.Request() method. -type RequestOpts struct { - // JSONBody, if provided, will be encoded as JSON and used as the body of the HTTP request. The - // content type of the request will default to "application/json" unless overridden by MoreHeaders. - // It's an error to specify both a JSONBody and a RawBody. - JSONBody interface{} - // RawBody contains an io.Reader that will be consumed by the request directly. No content-type - // will be set unless one is provided explicitly by MoreHeaders. - RawBody io.Reader - // JSONResponse, if provided, will be populated with the contents of the response body parsed as - // JSON. - JSONResponse interface{} - // OkCodes contains a list of numeric HTTP status codes that should be interpreted as success. If - // the response has a different code, an error will be returned. - OkCodes []int - // MoreHeaders specifies additional HTTP headers to be provide on the request. If a header is - // provided with a blank value (""), that header will be *omitted* instead: use this to suppress - // the default Accept header or an inferred Content-Type, for example. - MoreHeaders map[string]string - // ErrorContext specifies the resource error type to return if an error is encountered. - // This lets resources override default error messages based on the response status code. - ErrorContext error -} - -var applicationJSON = "application/json" - -// Request performs an HTTP request using the ProviderClient's current HTTPClient. An authentication -// header will automatically be provided. -func (client *ProviderClient) Request(method, url string, options *RequestOpts) (*http.Response, error) { - var body io.Reader - var contentType *string - - // Derive the content body by either encoding an arbitrary object as JSON, or by taking a provided - // io.ReadSeeker as-is. Default the content-type to application/json. - if options.JSONBody != nil { - if options.RawBody != nil { - return nil, errors.New("please provide only one of JSONBody or RawBody to gophercloud.Request()") - } - - rendered, err := json.Marshal(options.JSONBody) - if err != nil { - return nil, err - } - - body = bytes.NewReader(rendered) - contentType = &applicationJSON - } - - if options.RawBody != nil { - body = options.RawBody - } - - // Construct the http.Request. - req, err := http.NewRequest(method, url, body) - if err != nil { - return nil, err - } - if client.Context != nil { - req = req.WithContext(client.Context) - } - - // Populate the request headers. Apply options.MoreHeaders last, to give the caller the chance to - // modify or omit any header. - if contentType != nil { - req.Header.Set("Content-Type", *contentType) - } - req.Header.Set("Accept", applicationJSON) - - // Set the User-Agent header - req.Header.Set("User-Agent", client.UserAgent.Join()) - - if options.MoreHeaders != nil { - for k, v := range options.MoreHeaders { - if v != "" { - req.Header.Set(k, v) - } else { - req.Header.Del(k) - } - } - } - - // get latest token from client - for k, v := range client.AuthenticatedHeaders() { - req.Header.Set(k, v) - } - - // Set connection parameter to close the connection immediately when we've got the response - req.Close = true - - prereqtok := req.Header.Get("X-Auth-Token") - - // Issue the request. - resp, err := client.HTTPClient.Do(req) - if err != nil { - return nil, err - } - - // Allow default OkCodes if none explicitly set - okc := options.OkCodes - if okc == nil { - okc = defaultOkCodes(method) - } - - // Validate the HTTP response status. - var ok bool - for _, code := range okc { - if resp.StatusCode == code { - ok = true - break - } - } - - if !ok { - body, _ := ioutil.ReadAll(resp.Body) - resp.Body.Close() - respErr := ErrUnexpectedResponseCode{ - URL: url, - Method: method, - Expected: options.OkCodes, - Actual: resp.StatusCode, - Body: body, - } - - errType := options.ErrorContext - switch resp.StatusCode { - case http.StatusBadRequest: - err = ErrDefault400{respErr} - if error400er, ok := errType.(Err400er); ok { - err = error400er.Error400(respErr) - } - case http.StatusUnauthorized: - if client.ReauthFunc != nil { - err = client.Reauthenticate(prereqtok) - if err != nil { - e := &ErrUnableToReauthenticate{} - e.ErrOriginal = respErr - return nil, e - } - if options.RawBody != nil { - if seeker, ok := options.RawBody.(io.Seeker); ok { - seeker.Seek(0, 0) - } - } - resp, err = client.Request(method, url, options) - if err != nil { - switch err.(type) { - case *ErrUnexpectedResponseCode: - e := &ErrErrorAfterReauthentication{} - e.ErrOriginal = err.(*ErrUnexpectedResponseCode) - return nil, e - default: - e := &ErrErrorAfterReauthentication{} - e.ErrOriginal = err - return nil, e - } - } - return resp, nil - } - err = ErrDefault401{respErr} - if error401er, ok := errType.(Err401er); ok { - err = error401er.Error401(respErr) - } - case http.StatusForbidden: - err = ErrDefault403{respErr} - if error403er, ok := errType.(Err403er); ok { - err = error403er.Error403(respErr) - } - case http.StatusNotFound: - err = ErrDefault404{respErr} - if error404er, ok := errType.(Err404er); ok { - err = error404er.Error404(respErr) - } - case http.StatusMethodNotAllowed: - err = ErrDefault405{respErr} - if error405er, ok := errType.(Err405er); ok { - err = error405er.Error405(respErr) - } - case http.StatusRequestTimeout: - err = ErrDefault408{respErr} - if error408er, ok := errType.(Err408er); ok { - err = error408er.Error408(respErr) - } - case http.StatusConflict: - err = ErrDefault409{respErr} - if error409er, ok := errType.(Err409er); ok { - err = error409er.Error409(respErr) - } - case 429: - err = ErrDefault429{respErr} - if error429er, ok := errType.(Err429er); ok { - err = error429er.Error429(respErr) - } - case http.StatusInternalServerError: - err = ErrDefault500{respErr} - if error500er, ok := errType.(Err500er); ok { - err = error500er.Error500(respErr) - } - case http.StatusServiceUnavailable: - err = ErrDefault503{respErr} - if error503er, ok := errType.(Err503er); ok { - err = error503er.Error503(respErr) - } - } - - if err == nil { - err = respErr - } - - return resp, err - } - - // Parse the response body as JSON, if requested to do so. - if options.JSONResponse != nil { - defer resp.Body.Close() - if err := json.NewDecoder(resp.Body).Decode(options.JSONResponse); err != nil { - return nil, err - } - } - - return resp, nil -} - -func defaultOkCodes(method string) []int { - switch { - case method == "GET": - return []int{200} - case method == "POST": - return []int{201, 202} - case method == "PUT": - return []int{201, 202} - case method == "PATCH": - return []int{200, 202, 204} - case method == "DELETE": - return []int{202, 204} - } - - return []int{} -} diff --git a/vendor/github.com/gophercloud/gophercloud/results.go b/vendor/github.com/gophercloud/gophercloud/results.go deleted file mode 100644 index 94a16bff..00000000 --- a/vendor/github.com/gophercloud/gophercloud/results.go +++ /dev/null @@ -1,448 +0,0 @@ -package gophercloud - -import ( - "bytes" - "encoding/json" - "fmt" - "io" - "net/http" - "reflect" - "strconv" - "time" -) - -/* -Result is an internal type to be used by individual resource packages, but its -methods will be available on a wide variety of user-facing embedding types. - -It acts as a base struct that other Result types, returned from request -functions, can embed for convenience. All Results capture basic information -from the HTTP transaction that was performed, including the response body, -HTTP headers, and any errors that happened. - -Generally, each Result type will have an Extract method that can be used to -further interpret the result's payload in a specific context. Extensions or -providers can then provide additional extraction functions to pull out -provider- or extension-specific information as well. -*/ -type Result struct { - // Body is the payload of the HTTP response from the server. In most cases, - // this will be the deserialized JSON structure. - Body interface{} - - // Header contains the HTTP header structure from the original response. - Header http.Header - - // Err is an error that occurred during the operation. It's deferred until - // extraction to make it easier to chain the Extract call. - Err error -} - -// ExtractInto allows users to provide an object into which `Extract` will extract -// the `Result.Body`. This would be useful for OpenStack providers that have -// different fields in the response object than OpenStack proper. -func (r Result) ExtractInto(to interface{}) error { - if r.Err != nil { - return r.Err - } - - if reader, ok := r.Body.(io.Reader); ok { - if readCloser, ok := reader.(io.Closer); ok { - defer readCloser.Close() - } - return json.NewDecoder(reader).Decode(to) - } - - b, err := json.Marshal(r.Body) - if err != nil { - return err - } - err = json.Unmarshal(b, to) - - return err -} - -func (r Result) extractIntoPtr(to interface{}, label string) error { - if label == "" { - return r.ExtractInto(&to) - } - - var m map[string]interface{} - err := r.ExtractInto(&m) - if err != nil { - return err - } - - b, err := json.Marshal(m[label]) - if err != nil { - return err - } - - toValue := reflect.ValueOf(to) - if toValue.Kind() == reflect.Ptr { - toValue = toValue.Elem() - } - - switch toValue.Kind() { - case reflect.Slice: - typeOfV := toValue.Type().Elem() - if typeOfV.Kind() == reflect.Struct { - if typeOfV.NumField() > 0 && typeOfV.Field(0).Anonymous { - newSlice := reflect.MakeSlice(reflect.SliceOf(typeOfV), 0, 0) - - if mSlice, ok := m[label].([]interface{}); ok { - for _, v := range mSlice { - // For each iteration of the slice, we create a new struct. - // This is to work around a bug where elements of a slice - // are reused and not overwritten when the same copy of the - // struct is used: - // - // https://github.com/golang/go/issues/21092 - // https://github.com/golang/go/issues/24155 - // https://play.golang.org/p/NHo3ywlPZli - newType := reflect.New(typeOfV).Elem() - - b, err := json.Marshal(v) - if err != nil { - return err - } - - // This is needed for structs with an UnmarshalJSON method. - // Technically this is just unmarshalling the response into - // a struct that is never used, but it's good enough to - // trigger the UnmarshalJSON method. - for i := 0; i < newType.NumField(); i++ { - s := newType.Field(i).Addr().Interface() - - // Unmarshal is used rather than NewDecoder to also work - // around the above-mentioned bug. - err = json.Unmarshal(b, s) - if err != nil { - return err - } - } - - newSlice = reflect.Append(newSlice, newType) - } - } - - // "to" should now be properly modeled to receive the - // JSON response body and unmarshal into all the correct - // fields of the struct or composed extension struct - // at the end of this method. - toValue.Set(newSlice) - } - } - case reflect.Struct: - typeOfV := toValue.Type() - if typeOfV.NumField() > 0 && typeOfV.Field(0).Anonymous { - for i := 0; i < toValue.NumField(); i++ { - toField := toValue.Field(i) - if toField.Kind() == reflect.Struct { - s := toField.Addr().Interface() - err = json.NewDecoder(bytes.NewReader(b)).Decode(s) - if err != nil { - return err - } - } - } - } - } - - err = json.Unmarshal(b, &to) - return err -} - -// ExtractIntoStructPtr will unmarshal the Result (r) into the provided -// interface{} (to). -// -// NOTE: For internal use only -// -// `to` must be a pointer to an underlying struct type -// -// If provided, `label` will be filtered out of the response -// body prior to `r` being unmarshalled into `to`. -func (r Result) ExtractIntoStructPtr(to interface{}, label string) error { - if r.Err != nil { - return r.Err - } - - t := reflect.TypeOf(to) - if k := t.Kind(); k != reflect.Ptr { - return fmt.Errorf("Expected pointer, got %v", k) - } - switch t.Elem().Kind() { - case reflect.Struct: - return r.extractIntoPtr(to, label) - default: - return fmt.Errorf("Expected pointer to struct, got: %v", t) - } -} - -// ExtractIntoSlicePtr will unmarshal the Result (r) into the provided -// interface{} (to). -// -// NOTE: For internal use only -// -// `to` must be a pointer to an underlying slice type -// -// If provided, `label` will be filtered out of the response -// body prior to `r` being unmarshalled into `to`. -func (r Result) ExtractIntoSlicePtr(to interface{}, label string) error { - if r.Err != nil { - return r.Err - } - - t := reflect.TypeOf(to) - if k := t.Kind(); k != reflect.Ptr { - return fmt.Errorf("Expected pointer, got %v", k) - } - switch t.Elem().Kind() { - case reflect.Slice: - return r.extractIntoPtr(to, label) - default: - return fmt.Errorf("Expected pointer to slice, got: %v", t) - } -} - -// PrettyPrintJSON creates a string containing the full response body as -// pretty-printed JSON. It's useful for capturing test fixtures and for -// debugging extraction bugs. If you include its output in an issue related to -// a buggy extraction function, we will all love you forever. -func (r Result) PrettyPrintJSON() string { - pretty, err := json.MarshalIndent(r.Body, "", " ") - if err != nil { - panic(err.Error()) - } - return string(pretty) -} - -// ErrResult is an internal type to be used by individual resource packages, but -// its methods will be available on a wide variety of user-facing embedding -// types. -// -// It represents results that only contain a potential error and -// nothing else. Usually, if the operation executed successfully, the Err field -// will be nil; otherwise it will be stocked with a relevant error. Use the -// ExtractErr method -// to cleanly pull it out. -type ErrResult struct { - Result -} - -// ExtractErr is a function that extracts error information, or nil, from a result. -func (r ErrResult) ExtractErr() error { - return r.Err -} - -/* -HeaderResult is an internal type to be used by individual resource packages, but -its methods will be available on a wide variety of user-facing embedding types. - -It represents a result that only contains an error (possibly nil) and an -http.Header. This is used, for example, by the objectstorage packages in -openstack, because most of the operations don't return response bodies, but do -have relevant information in headers. -*/ -type HeaderResult struct { - Result -} - -// ExtractInto allows users to provide an object into which `Extract` will -// extract the http.Header headers of the result. -func (r HeaderResult) ExtractInto(to interface{}) error { - if r.Err != nil { - return r.Err - } - - tmpHeaderMap := map[string]string{} - for k, v := range r.Header { - if len(v) > 0 { - tmpHeaderMap[k] = v[0] - } - } - - b, err := json.Marshal(tmpHeaderMap) - if err != nil { - return err - } - err = json.Unmarshal(b, to) - - return err -} - -// RFC3339Milli describes a common time format used by some API responses. -const RFC3339Milli = "2006-01-02T15:04:05.999999Z" - -type JSONRFC3339Milli time.Time - -func (jt *JSONRFC3339Milli) UnmarshalJSON(data []byte) error { - b := bytes.NewBuffer(data) - dec := json.NewDecoder(b) - var s string - if err := dec.Decode(&s); err != nil { - return err - } - t, err := time.Parse(RFC3339Milli, s) - if err != nil { - return err - } - *jt = JSONRFC3339Milli(t) - return nil -} - -const RFC3339MilliNoZ = "2006-01-02T15:04:05.999999" - -type JSONRFC3339MilliNoZ time.Time - -func (jt *JSONRFC3339MilliNoZ) UnmarshalJSON(data []byte) error { - var s string - if err := json.Unmarshal(data, &s); err != nil { - return err - } - if s == "" { - return nil - } - t, err := time.Parse(RFC3339MilliNoZ, s) - if err != nil { - return err - } - *jt = JSONRFC3339MilliNoZ(t) - return nil -} - -type JSONRFC1123 time.Time - -func (jt *JSONRFC1123) UnmarshalJSON(data []byte) error { - var s string - if err := json.Unmarshal(data, &s); err != nil { - return err - } - if s == "" { - return nil - } - t, err := time.Parse(time.RFC1123, s) - if err != nil { - return err - } - *jt = JSONRFC1123(t) - return nil -} - -type JSONUnix time.Time - -func (jt *JSONUnix) UnmarshalJSON(data []byte) error { - var s string - if err := json.Unmarshal(data, &s); err != nil { - return err - } - if s == "" { - return nil - } - unix, err := strconv.ParseInt(s, 10, 64) - if err != nil { - return err - } - t = time.Unix(unix, 0) - *jt = JSONUnix(t) - return nil -} - -// RFC3339NoZ is the time format used in Heat (Orchestration). -const RFC3339NoZ = "2006-01-02T15:04:05" - -type JSONRFC3339NoZ time.Time - -func (jt *JSONRFC3339NoZ) UnmarshalJSON(data []byte) error { - var s string - if err := json.Unmarshal(data, &s); err != nil { - return err - } - if s == "" { - return nil - } - t, err := time.Parse(RFC3339NoZ, s) - if err != nil { - return err - } - *jt = JSONRFC3339NoZ(t) - return nil -} - -// RFC3339ZNoT is the time format used in Zun (Containers Service). -const RFC3339ZNoT = "2006-01-02 15:04:05-07:00" - -type JSONRFC3339ZNoT time.Time - -func (jt *JSONRFC3339ZNoT) UnmarshalJSON(data []byte) error { - var s string - if err := json.Unmarshal(data, &s); err != nil { - return err - } - if s == "" { - return nil - } - t, err := time.Parse(RFC3339ZNoT, s) - if err != nil { - return err - } - *jt = JSONRFC3339ZNoT(t) - return nil -} - -// RFC3339ZNoTNoZ is another time format used in Zun (Containers Service). -const RFC3339ZNoTNoZ = "2006-01-02 15:04:05" - -type JSONRFC3339ZNoTNoZ time.Time - -func (jt *JSONRFC3339ZNoTNoZ) UnmarshalJSON(data []byte) error { - var s string - if err := json.Unmarshal(data, &s); err != nil { - return err - } - if s == "" { - return nil - } - t, err := time.Parse(RFC3339ZNoTNoZ, s) - if err != nil { - return err - } - *jt = JSONRFC3339ZNoTNoZ(t) - return nil -} - -/* -Link is an internal type to be used in packages of collection resources that are -paginated in a certain way. - -It's a response substructure common to many paginated collection results that is -used to point to related pages. Usually, the one we care about is the one with -Rel field set to "next". -*/ -type Link struct { - Href string `json:"href"` - Rel string `json:"rel"` -} - -/* -ExtractNextURL is an internal function useful for packages of collection -resources that are paginated in a certain way. - -It attempts to extract the "next" URL from slice of Link structs, or -"" if no such URL is present. -*/ -func ExtractNextURL(links []Link) (string, error) { - var url string - - for _, l := range links { - if l.Rel == "next" { - url = l.Href - } - } - - if url == "" { - return "", nil - } - - return url, nil -} diff --git a/vendor/github.com/gophercloud/gophercloud/service_client.go b/vendor/github.com/gophercloud/gophercloud/service_client.go deleted file mode 100644 index f222f05a..00000000 --- a/vendor/github.com/gophercloud/gophercloud/service_client.go +++ /dev/null @@ -1,154 +0,0 @@ -package gophercloud - -import ( - "io" - "net/http" - "strings" -) - -// ServiceClient stores details required to interact with a specific service API implemented by a provider. -// Generally, you'll acquire these by calling the appropriate `New` method on a ProviderClient. -type ServiceClient struct { - // ProviderClient is a reference to the provider that implements this service. - *ProviderClient - - // Endpoint is the base URL of the service's API, acquired from a service catalog. - // It MUST end with a /. - Endpoint string - - // ResourceBase is the base URL shared by the resources within a service's API. It should include - // the API version and, like Endpoint, MUST end with a / if set. If not set, the Endpoint is used - // as-is, instead. - ResourceBase string - - // This is the service client type (e.g. compute, sharev2). - // NOTE: FOR INTERNAL USE ONLY. DO NOT SET. GOPHERCLOUD WILL SET THIS. - // It is only exported because it gets set in a different package. - Type string - - // The microversion of the service to use. Set this to use a particular microversion. - Microversion string - - // MoreHeaders allows users (or Gophercloud) to set service-wide headers on requests. Put another way, - // values set in this field will be set on all the HTTP requests the service client sends. - MoreHeaders map[string]string -} - -// ResourceBaseURL returns the base URL of any resources used by this service. It MUST end with a /. -func (client *ServiceClient) ResourceBaseURL() string { - if client.ResourceBase != "" { - return client.ResourceBase - } - return client.Endpoint -} - -// ServiceURL constructs a URL for a resource belonging to this provider. -func (client *ServiceClient) ServiceURL(parts ...string) string { - return client.ResourceBaseURL() + strings.Join(parts, "/") -} - -func (client *ServiceClient) initReqOpts(url string, JSONBody interface{}, JSONResponse interface{}, opts *RequestOpts) { - if v, ok := (JSONBody).(io.Reader); ok { - opts.RawBody = v - } else if JSONBody != nil { - opts.JSONBody = JSONBody - } - - if JSONResponse != nil { - opts.JSONResponse = JSONResponse - } - - if opts.MoreHeaders == nil { - opts.MoreHeaders = make(map[string]string) - } - - if client.Microversion != "" { - client.setMicroversionHeader(opts) - } -} - -// Get calls `Request` with the "GET" HTTP verb. -func (client *ServiceClient) Get(url string, JSONResponse interface{}, opts *RequestOpts) (*http.Response, error) { - if opts == nil { - opts = new(RequestOpts) - } - client.initReqOpts(url, nil, JSONResponse, opts) - return client.Request("GET", url, opts) -} - -// Post calls `Request` with the "POST" HTTP verb. -func (client *ServiceClient) Post(url string, JSONBody interface{}, JSONResponse interface{}, opts *RequestOpts) (*http.Response, error) { - if opts == nil { - opts = new(RequestOpts) - } - client.initReqOpts(url, JSONBody, JSONResponse, opts) - return client.Request("POST", url, opts) -} - -// Put calls `Request` with the "PUT" HTTP verb. -func (client *ServiceClient) Put(url string, JSONBody interface{}, JSONResponse interface{}, opts *RequestOpts) (*http.Response, error) { - if opts == nil { - opts = new(RequestOpts) - } - client.initReqOpts(url, JSONBody, JSONResponse, opts) - return client.Request("PUT", url, opts) -} - -// Patch calls `Request` with the "PATCH" HTTP verb. -func (client *ServiceClient) Patch(url string, JSONBody interface{}, JSONResponse interface{}, opts *RequestOpts) (*http.Response, error) { - if opts == nil { - opts = new(RequestOpts) - } - client.initReqOpts(url, JSONBody, JSONResponse, opts) - return client.Request("PATCH", url, opts) -} - -// Delete calls `Request` with the "DELETE" HTTP verb. -func (client *ServiceClient) Delete(url string, opts *RequestOpts) (*http.Response, error) { - if opts == nil { - opts = new(RequestOpts) - } - client.initReqOpts(url, nil, nil, opts) - return client.Request("DELETE", url, opts) -} - -// Head calls `Request` with the "HEAD" HTTP verb. -func (client *ServiceClient) Head(url string, opts *RequestOpts) (*http.Response, error) { - if opts == nil { - opts = new(RequestOpts) - } - client.initReqOpts(url, nil, nil, opts) - return client.Request("HEAD", url, opts) -} - -func (client *ServiceClient) setMicroversionHeader(opts *RequestOpts) { - switch client.Type { - case "compute": - opts.MoreHeaders["X-OpenStack-Nova-API-Version"] = client.Microversion - case "sharev2": - opts.MoreHeaders["X-OpenStack-Manila-API-Version"] = client.Microversion - case "volume": - opts.MoreHeaders["X-OpenStack-Volume-API-Version"] = client.Microversion - case "baremetal": - opts.MoreHeaders["X-OpenStack-Ironic-API-Version"] = client.Microversion - case "baremetal-introspection": - opts.MoreHeaders["X-OpenStack-Ironic-Inspector-API-Version"] = client.Microversion - } - - if client.Type != "" { - opts.MoreHeaders["OpenStack-API-Version"] = client.Type + " " + client.Microversion - } -} - -// Request carries out the HTTP operation for the service client -func (client *ServiceClient) Request(method, url string, options *RequestOpts) (*http.Response, error) { - if len(client.MoreHeaders) > 0 { - if options == nil { - options = new(RequestOpts) - } - for k, v := range client.MoreHeaders { - options.MoreHeaders[k] = v - } - } - return client.ProviderClient.Request(method, url, options) -} diff --git a/vendor/github.com/gophercloud/gophercloud/util.go b/vendor/github.com/gophercloud/gophercloud/util.go deleted file mode 100644 index 68f9a5d3..00000000 --- a/vendor/github.com/gophercloud/gophercloud/util.go +++ /dev/null @@ -1,102 +0,0 @@ -package gophercloud - -import ( - "fmt" - "net/url" - "path/filepath" - "strings" - "time" -) - -// WaitFor polls a predicate function, once per second, up to a timeout limit. -// This is useful to wait for a resource to transition to a certain state. -// To handle situations when the predicate might hang indefinitely, the -// predicate will be prematurely cancelled after the timeout. -// Resource packages will wrap this in a more convenient function that's -// specific to a certain resource, but it can also be useful on its own. -func WaitFor(timeout int, predicate func() (bool, error)) error { - type WaitForResult struct { - Success bool - Error error - } - - start := time.Now().Unix() - - for { - // If a timeout is set, and that's been exceeded, shut it down. - if timeout >= 0 && time.Now().Unix()-start >= int64(timeout) { - return fmt.Errorf("A timeout occurred") - } - - time.Sleep(1 * time.Second) - - var result WaitForResult - ch := make(chan bool, 1) - go func() { - defer close(ch) - satisfied, err := predicate() - result.Success = satisfied - result.Error = err - }() - - select { - case <-ch: - if result.Error != nil { - return result.Error - } - if result.Success { - return nil - } - // If the predicate has not finished by the timeout, cancel it. - case <-time.After(time.Duration(timeout) * time.Second): - return fmt.Errorf("A timeout occurred") - } - } -} - -// NormalizeURL is an internal function to be used by provider clients. -// -// It ensures that each endpoint URL has a closing `/`, as expected by -// ServiceClient's methods. -func NormalizeURL(url string) string { - if !strings.HasSuffix(url, "/") { - return url + "/" - } - return url -} - -// NormalizePathURL is used to convert rawPath to a fqdn, using basePath as -// a reference in the filesystem, if necessary. basePath is assumed to contain -// either '.' when first used, or the file:// type fqdn of the parent resource. -// e.g. myFavScript.yaml => file://opt/lib/myFavScript.yaml -func NormalizePathURL(basePath, rawPath string) (string, error) { - u, err := url.Parse(rawPath) - if err != nil { - return "", err - } - // if a scheme is defined, it must be a fqdn already - if u.Scheme != "" { - return u.String(), nil - } - // if basePath is a url, then child resources are assumed to be relative to it - bu, err := url.Parse(basePath) - if err != nil { - return "", err - } - var basePathSys, absPathSys string - if bu.Scheme != "" { - basePathSys = filepath.FromSlash(bu.Path) - absPathSys = filepath.Join(basePathSys, rawPath) - bu.Path = filepath.ToSlash(absPathSys) - return bu.String(), nil - } - - absPathSys = filepath.Join(basePath, rawPath) - u.Path = filepath.ToSlash(absPathSys) - if err != nil { - return "", err - } - u.Scheme = "file" - return u.String(), nil - -} diff --git a/vendor/github.com/json-iterator/go/README.md b/vendor/github.com/json-iterator/go/README.md index 50d56ffb..52b111d5 100644 --- a/vendor/github.com/json-iterator/go/README.md +++ b/vendor/github.com/json-iterator/go/README.md @@ -1,5 +1,5 @@ [![Sourcegraph](https://sourcegraph.com/github.com/json-iterator/go/-/badge.svg)](https://sourcegraph.com/github.com/json-iterator/go?badge) -[![GoDoc](http://img.shields.io/badge/go-documentation-blue.svg?style=flat-square)](http://godoc.org/github.com/json-iterator/go) +[![GoDoc](http://img.shields.io/badge/go-documentation-blue.svg?style=flat-square)](https://pkg.go.dev/github.com/json-iterator/go) [![Build Status](https://travis-ci.org/json-iterator/go.svg?branch=master)](https://travis-ci.org/json-iterator/go) [![codecov](https://codecov.io/gh/json-iterator/go/branch/master/graph/badge.svg)](https://codecov.io/gh/json-iterator/go) [![rcard](https://goreportcard.com/badge/github.com/json-iterator/go)](https://goreportcard.com/report/github.com/json-iterator/go) @@ -18,16 +18,16 @@ Source code: https://github.com/json-iterator/go-benchmark/blob/master/src/githu Raw Result (easyjson requires static code generation) -| | ns/op | allocation bytes | allocation times | -| --- | --- | --- | --- | -| std decode | 35510 ns/op | 1960 B/op | 99 allocs/op | -| easyjson decode | 8499 ns/op | 160 B/op | 4 allocs/op | -| jsoniter decode | 5623 ns/op | 160 B/op | 3 allocs/op | -| std encode | 2213 ns/op | 712 B/op | 5 allocs/op | -| easyjson encode | 883 ns/op | 576 B/op | 3 allocs/op | -| jsoniter encode | 837 ns/op | 384 B/op | 4 allocs/op | +| | ns/op | allocation bytes | allocation times | +| --------------- | ----------- | ---------------- | ---------------- | +| std decode | 35510 ns/op | 1960 B/op | 99 allocs/op | +| easyjson decode | 8499 ns/op | 160 B/op | 4 allocs/op | +| jsoniter decode | 5623 ns/op | 160 B/op | 3 allocs/op | +| std encode | 2213 ns/op | 712 B/op | 5 allocs/op | +| easyjson encode | 883 ns/op | 576 B/op | 3 allocs/op | +| jsoniter encode | 837 ns/op | 384 B/op | 4 allocs/op | -Always benchmark with your own workload. +Always benchmark with your own workload. The result depends heavily on the data input. # Usage @@ -41,10 +41,10 @@ import "encoding/json" json.Marshal(&data) ``` -with +with ```go -import "github.com/json-iterator/go" +import jsoniter "github.com/json-iterator/go" var json = jsoniter.ConfigCompatibleWithStandardLibrary json.Marshal(&data) @@ -60,7 +60,7 @@ json.Unmarshal(input, &data) with ```go -import "github.com/json-iterator/go" +import jsoniter "github.com/json-iterator/go" var json = jsoniter.ConfigCompatibleWithStandardLibrary json.Unmarshal(input, &data) @@ -78,10 +78,10 @@ go get github.com/json-iterator/go Contributors -* [thockin](https://github.com/thockin) -* [mattn](https://github.com/mattn) -* [cch123](https://github.com/cch123) -* [Oleg Shaldybin](https://github.com/olegshaldybin) -* [Jason Toffaletti](https://github.com/toffaletti) +- [thockin](https://github.com/thockin) +- [mattn](https://github.com/mattn) +- [cch123](https://github.com/cch123) +- [Oleg Shaldybin](https://github.com/olegshaldybin) +- [Jason Toffaletti](https://github.com/toffaletti) Report issue or pull request, or email taowen@gmail.com, or [![Gitter chat](https://badges.gitter.im/gitterHQ/gitter.png)](https://gitter.im/json-iterator/Lobby) diff --git a/vendor/github.com/json-iterator/go/any_str.go b/vendor/github.com/json-iterator/go/any_str.go index a4b93c78..1f12f661 100644 --- a/vendor/github.com/json-iterator/go/any_str.go +++ b/vendor/github.com/json-iterator/go/any_str.go @@ -64,7 +64,6 @@ func (any *stringAny) ToInt64() int64 { flag := 1 startPos := 0 - endPos := 0 if any.val[0] == '+' || any.val[0] == '-' { startPos = 1 } @@ -73,6 +72,7 @@ func (any *stringAny) ToInt64() int64 { flag = -1 } + endPos := startPos for i := startPos; i < len(any.val); i++ { if any.val[i] >= '0' && any.val[i] <= '9' { endPos = i + 1 @@ -98,7 +98,6 @@ func (any *stringAny) ToUint64() uint64 { } startPos := 0 - endPos := 0 if any.val[0] == '-' { return 0 @@ -107,6 +106,7 @@ func (any *stringAny) ToUint64() uint64 { startPos = 1 } + endPos := startPos for i := startPos; i < len(any.val); i++ { if any.val[i] >= '0' && any.val[i] <= '9' { endPos = i + 1 diff --git a/vendor/github.com/json-iterator/go/config.go b/vendor/github.com/json-iterator/go/config.go index 8c58fcba..2adcdc3b 100644 --- a/vendor/github.com/json-iterator/go/config.go +++ b/vendor/github.com/json-iterator/go/config.go @@ -183,11 +183,11 @@ func (cfg *frozenConfig) validateJsonRawMessage(extension EncoderExtension) { encoder := &funcEncoder{func(ptr unsafe.Pointer, stream *Stream) { rawMessage := *(*json.RawMessage)(ptr) iter := cfg.BorrowIterator([]byte(rawMessage)) + defer cfg.ReturnIterator(iter) iter.Read() - if iter.Error != nil { + if iter.Error != nil && iter.Error != io.EOF { stream.WriteRaw("null") } else { - cfg.ReturnIterator(iter) stream.WriteRaw(string(rawMessage)) } }, func(ptr unsafe.Pointer) bool { diff --git a/vendor/github.com/json-iterator/go/iter_object.go b/vendor/github.com/json-iterator/go/iter_object.go index b6513711..58ee89c8 100644 --- a/vendor/github.com/json-iterator/go/iter_object.go +++ b/vendor/github.com/json-iterator/go/iter_object.go @@ -150,7 +150,7 @@ func (iter *Iterator) ReadObjectCB(callback func(*Iterator, string) bool) bool { if c == '}' { return iter.decrementDepth() } - iter.ReportError("ReadObjectCB", `expect " after }, but found `+string([]byte{c})) + iter.ReportError("ReadObjectCB", `expect " after {, but found `+string([]byte{c})) iter.decrementDepth() return false } @@ -206,7 +206,7 @@ func (iter *Iterator) ReadMapCB(callback func(*Iterator, string) bool) bool { if c == '}' { return iter.decrementDepth() } - iter.ReportError("ReadMapCB", `expect " after }, but found `+string([]byte{c})) + iter.ReportError("ReadMapCB", `expect " after {, but found `+string([]byte{c})) iter.decrementDepth() return false } diff --git a/vendor/github.com/json-iterator/go/reflect_extension.go b/vendor/github.com/json-iterator/go/reflect_extension.go index 80320cd6..74a97bfe 100644 --- a/vendor/github.com/json-iterator/go/reflect_extension.go +++ b/vendor/github.com/json-iterator/go/reflect_extension.go @@ -475,7 +475,7 @@ func calcFieldNames(originalFieldName string, tagProvidedFieldName string, whole fieldNames = []string{tagProvidedFieldName} } // private? - isNotExported := unicode.IsLower(rune(originalFieldName[0])) + isNotExported := unicode.IsLower(rune(originalFieldName[0])) || originalFieldName[0] == '_' if isNotExported { fieldNames = []string{} } diff --git a/vendor/github.com/json-iterator/go/reflect_map.go b/vendor/github.com/json-iterator/go/reflect_map.go index 9e2b623f..58296713 100644 --- a/vendor/github.com/json-iterator/go/reflect_map.go +++ b/vendor/github.com/json-iterator/go/reflect_map.go @@ -49,6 +49,33 @@ func decoderOfMapKey(ctx *ctx, typ reflect2.Type) ValDecoder { return decoder } } + + ptrType := reflect2.PtrTo(typ) + if ptrType.Implements(unmarshalerType) { + return &referenceDecoder{ + &unmarshalerDecoder{ + valType: ptrType, + }, + } + } + if typ.Implements(unmarshalerType) { + return &unmarshalerDecoder{ + valType: typ, + } + } + if ptrType.Implements(textUnmarshalerType) { + return &referenceDecoder{ + &textUnmarshalerDecoder{ + valType: ptrType, + }, + } + } + if typ.Implements(textUnmarshalerType) { + return &textUnmarshalerDecoder{ + valType: typ, + } + } + switch typ.Kind() { case reflect.String: return decoderOfType(ctx, reflect2.DefaultTypeOfKind(reflect.String)) @@ -63,31 +90,6 @@ func decoderOfMapKey(ctx *ctx, typ reflect2.Type) ValDecoder { typ = reflect2.DefaultTypeOfKind(typ.Kind()) return &numericMapKeyDecoder{decoderOfType(ctx, typ)} default: - ptrType := reflect2.PtrTo(typ) - if ptrType.Implements(unmarshalerType) { - return &referenceDecoder{ - &unmarshalerDecoder{ - valType: ptrType, - }, - } - } - if typ.Implements(unmarshalerType) { - return &unmarshalerDecoder{ - valType: typ, - } - } - if ptrType.Implements(textUnmarshalerType) { - return &referenceDecoder{ - &textUnmarshalerDecoder{ - valType: ptrType, - }, - } - } - if typ.Implements(textUnmarshalerType) { - return &textUnmarshalerDecoder{ - valType: typ, - } - } return &lazyErrorDecoder{err: fmt.Errorf("unsupported map key type: %v", typ)} } } @@ -103,6 +105,19 @@ func encoderOfMapKey(ctx *ctx, typ reflect2.Type) ValEncoder { return encoder } } + + if typ == textMarshalerType { + return &directTextMarshalerEncoder{ + stringEncoder: ctx.EncoderOf(reflect2.TypeOf("")), + } + } + if typ.Implements(textMarshalerType) { + return &textMarshalerEncoder{ + valType: typ, + stringEncoder: ctx.EncoderOf(reflect2.TypeOf("")), + } + } + switch typ.Kind() { case reflect.String: return encoderOfType(ctx, reflect2.DefaultTypeOfKind(reflect.String)) @@ -117,17 +132,6 @@ func encoderOfMapKey(ctx *ctx, typ reflect2.Type) ValEncoder { typ = reflect2.DefaultTypeOfKind(typ.Kind()) return &numericMapKeyEncoder{encoderOfType(ctx, typ)} default: - if typ == textMarshalerType { - return &directTextMarshalerEncoder{ - stringEncoder: ctx.EncoderOf(reflect2.TypeOf("")), - } - } - if typ.Implements(textMarshalerType) { - return &textMarshalerEncoder{ - valType: typ, - stringEncoder: ctx.EncoderOf(reflect2.TypeOf("")), - } - } if typ.Kind() == reflect.Interface { return &dynamicMapKeyEncoder{ctx, typ} } @@ -163,10 +167,6 @@ func (decoder *mapDecoder) Decode(ptr unsafe.Pointer, iter *Iterator) { if c == '}' { return } - if c != '"' { - iter.ReportError("ReadMapCB", `expect " after }, but found `+string([]byte{c})) - return - } iter.unreadByte() key := decoder.keyType.UnsafeNew() decoder.keyDecoder.Decode(key, iter) diff --git a/vendor/github.com/json-iterator/go/reflect_optional.go b/vendor/github.com/json-iterator/go/reflect_optional.go index 43ec71d6..fa71f474 100644 --- a/vendor/github.com/json-iterator/go/reflect_optional.go +++ b/vendor/github.com/json-iterator/go/reflect_optional.go @@ -2,7 +2,6 @@ package jsoniter import ( "github.com/modern-go/reflect2" - "reflect" "unsafe" ) @@ -10,9 +9,6 @@ func decoderOfOptional(ctx *ctx, typ reflect2.Type) ValDecoder { ptrType := typ.(*reflect2.UnsafePtrType) elemType := ptrType.Elem() decoder := decoderOfType(ctx, elemType) - if ctx.prefix == "" && elemType.Kind() == reflect.Ptr { - return &dereferenceDecoder{elemType, decoder} - } return &OptionalDecoder{elemType, decoder} } diff --git a/vendor/github.com/json-iterator/go/reflect_struct_decoder.go b/vendor/github.com/json-iterator/go/reflect_struct_decoder.go index 5ad5cc56..d7eb0eb5 100644 --- a/vendor/github.com/json-iterator/go/reflect_struct_decoder.go +++ b/vendor/github.com/json-iterator/go/reflect_struct_decoder.go @@ -507,7 +507,7 @@ func (decoder *generalStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterator) for c = ','; c == ','; c = iter.nextToken() { decoder.decodeOneField(ptr, iter) } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } if c != '}' { @@ -588,7 +588,7 @@ func (decoder *oneFieldStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterator) break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -622,7 +622,7 @@ func (decoder *twoFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterator break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -660,7 +660,7 @@ func (decoder *threeFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterat break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -702,7 +702,7 @@ func (decoder *fourFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterato break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -748,7 +748,7 @@ func (decoder *fiveFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterato break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -798,7 +798,7 @@ func (decoder *sixFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterator break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -852,7 +852,7 @@ func (decoder *sevenFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterat break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -910,7 +910,7 @@ func (decoder *eightFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterat break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -972,7 +972,7 @@ func (decoder *nineFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterato break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() @@ -1038,7 +1038,7 @@ func (decoder *tenFieldsStructDecoder) Decode(ptr unsafe.Pointer, iter *Iterator break } } - if iter.Error != nil && iter.Error != io.EOF { + if iter.Error != nil && iter.Error != io.EOF && len(decoder.typ.Type1().Name()) != 0 { iter.Error = fmt.Errorf("%v.%s", decoder.typ, iter.Error.Error()) } iter.decrementDepth() diff --git a/vendor/github.com/json-iterator/go/stream.go b/vendor/github.com/json-iterator/go/stream.go index 17662fde..23d8a3ad 100644 --- a/vendor/github.com/json-iterator/go/stream.go +++ b/vendor/github.com/json-iterator/go/stream.go @@ -103,14 +103,14 @@ func (stream *Stream) Flush() error { if stream.Error != nil { return stream.Error } - n, err := stream.out.Write(stream.buf) + _, err := stream.out.Write(stream.buf) if err != nil { if stream.Error == nil { stream.Error = err } return err } - stream.buf = stream.buf[n:] + stream.buf = stream.buf[:0] return nil } @@ -177,7 +177,6 @@ func (stream *Stream) WriteEmptyObject() { func (stream *Stream) WriteMore() { stream.writeByte(',') stream.writeIndention(0) - stream.Flush() } // WriteArrayStart write [ with possible indention diff --git a/vendor/github.com/konsorten/go-windows-terminal-sequences/README.md b/vendor/github.com/konsorten/go-windows-terminal-sequences/README.md index 195333e5..09a4a35c 100644 --- a/vendor/github.com/konsorten/go-windows-terminal-sequences/README.md +++ b/vendor/github.com/konsorten/go-windows-terminal-sequences/README.md @@ -27,6 +27,7 @@ We thank all the authors who provided code to this library: * Felix Kollmann * Nicolas Perraut +* @dirty49374 ## License diff --git a/vendor/github.com/konsorten/go-windows-terminal-sequences/sequences.go b/vendor/github.com/konsorten/go-windows-terminal-sequences/sequences.go index ef18d8f9..57f530ae 100644 --- a/vendor/github.com/konsorten/go-windows-terminal-sequences/sequences.go +++ b/vendor/github.com/konsorten/go-windows-terminal-sequences/sequences.go @@ -4,7 +4,6 @@ package sequences import ( "syscall" - "unsafe" ) var ( @@ -27,7 +26,7 @@ func EnableVirtualTerminalProcessing(stream syscall.Handle, enable bool) error { mode &^= ENABLE_VIRTUAL_TERMINAL_PROCESSING } - ret, _, err := setConsoleMode.Call(uintptr(unsafe.Pointer(stream)), uintptr(mode)) + ret, _, err := setConsoleMode.Call(uintptr(stream), uintptr(mode)) if ret == 0 { return err } diff --git a/vendor/github.com/sergi/go-diff/diffmatchpatch/diff.go b/vendor/github.com/sergi/go-diff/diffmatchpatch/diff.go index cb25b437..82ad7bc8 100644 --- a/vendor/github.com/sergi/go-diff/diffmatchpatch/diff.go +++ b/vendor/github.com/sergi/go-diff/diffmatchpatch/diff.go @@ -25,8 +25,6 @@ import ( // Operation defines the operation of a diff item. type Operation int8 -//go:generate stringer -type=Operation -trimprefix=Diff - const ( // DiffDelete item represents a delete diff. DiffDelete Operation = -1 @@ -42,41 +40,8 @@ type Diff struct { Text string } -// splice removes amount elements from slice at index index, replacing them with elements. func splice(slice []Diff, index int, amount int, elements ...Diff) []Diff { - if len(elements) == amount { - // Easy case: overwrite the relevant items. - copy(slice[index:], elements) - return slice - } - if len(elements) < amount { - // Fewer new items than old. - // Copy in the new items. - copy(slice[index:], elements) - // Shift the remaining items left. - copy(slice[index+len(elements):], slice[index+amount:]) - // Calculate the new end of the slice. - end := len(slice) - amount + len(elements) - // Zero stranded elements at end so that they can be garbage collected. - tail := slice[end:] - for i := range tail { - tail[i] = Diff{} - } - return slice[:end] - } - // More new items than old. - // Make room in slice for new elements. - // There's probably an even more efficient way to do this, - // but this is simple and clear. - need := len(slice) - amount + len(elements) - for len(slice) < need { - slice = append(slice, Diff{}) - } - // Shift slice elements right to make room for new elements. - copy(slice[index+len(elements):], slice[index+amount:]) - // Copy in new elements. - copy(slice[index:], elements) - return slice + return append(slice[:index], append(elements, slice[index+amount:]...)...) } // DiffMain finds the differences between two texts. @@ -180,10 +145,7 @@ func (dmp *DiffMatchPatch) diffCompute(text1, text2 []rune, checklines bool, dea diffsA := dmp.diffMainRunes(text1A, text2A, checklines, deadline) diffsB := dmp.diffMainRunes(text1B, text2B, checklines, deadline) // Merge the results. - diffs := diffsA - diffs = append(diffs, Diff{DiffEqual, string(midCommon)}) - diffs = append(diffs, diffsB...) - return diffs + return append(diffsA, append([]Diff{Diff{DiffEqual, string(midCommon)}}, diffsB...)...) } else if checklines && len(text1) > 100 && len(text2) > 100 { return dmp.diffLineMode(text1, text2, deadline) } @@ -285,7 +247,7 @@ func (dmp *DiffMatchPatch) diffBisect(runes1, runes2 []rune, deadline time.Time) k2end := 0 for d := 0; d < maxD; d++ { // Bail out if deadline is reached. - if !deadline.IsZero() && d%16 == 0 && time.Now().After(deadline) { + if !deadline.IsZero() && time.Now().After(deadline) { break } @@ -472,29 +434,48 @@ func (dmp *DiffMatchPatch) DiffCommonSuffix(text1, text2 string) int { // commonPrefixLength returns the length of the common prefix of two rune slices. func commonPrefixLength(text1, text2 []rune) int { - // Linear search. See comment in commonSuffixLength. - n := 0 - for ; n < len(text1) && n < len(text2); n++ { - if text1[n] != text2[n] { - return n + short, long := text1, text2 + if len(short) > len(long) { + short, long = long, short + } + for i, r := range short { + if r != long[i] { + return i } } - return n + return len(short) } // commonSuffixLength returns the length of the common suffix of two rune slices. func commonSuffixLength(text1, text2 []rune) int { - // Use linear search rather than the binary search discussed at https://neil.fraser.name/news/2007/10/09/. - // See discussion at https://github.com/sergi/go-diff/issues/54. - i1 := len(text1) - i2 := len(text2) - for n := 0; ; n++ { - i1-- - i2-- - if i1 < 0 || i2 < 0 || text1[i1] != text2[i2] { - return n + n := min(len(text1), len(text2)) + for i := 0; i < n; i++ { + if text1[len(text1)-i-1] != text2[len(text2)-i-1] { + return i } } + return n + + // TODO research and benchmark this, why is it not activated? https://github.com/sergi/go-diff/issues/54 + // Binary search. + // Performance analysis: http://neil.fraser.name/news/2007/10/09/ + /* + pointermin := 0 + pointermax := math.Min(len(text1), len(text2)) + pointermid := pointermax + pointerend := 0 + for pointermin < pointermid { + if text1[len(text1)-pointermid:len(text1)-pointerend] == + text2[len(text2)-pointermid:len(text2)-pointerend] { + pointermin = pointermid + pointerend = pointermin + } else { + pointermax = pointermid + } + pointermid = math.Floor((pointermax-pointermin)/2 + pointermin) + } + return pointermid + */ } // DiffCommonOverlap determines if the suffix of one string is the prefix of another. @@ -647,7 +628,11 @@ func (dmp *DiffMatchPatch) diffHalfMatchI(l, s []rune, i int) [][]rune { func (dmp *DiffMatchPatch) DiffCleanupSemantic(diffs []Diff) []Diff { changes := false // Stack of indices where equalities are found. - equalities := make([]int, 0, len(diffs)) + type equality struct { + data int + next *equality + } + var equalities *equality var lastequality string // Always equal to diffs[equalities[equalitiesLength - 1]][1] @@ -660,7 +645,11 @@ func (dmp *DiffMatchPatch) DiffCleanupSemantic(diffs []Diff) []Diff { for pointer < len(diffs) { if diffs[pointer].Type == DiffEqual { // Equality found. - equalities = append(equalities, pointer) + + equalities = &equality{ + data: pointer, + next: equalities, + } lengthInsertions1 = lengthInsertions2 lengthDeletions1 = lengthDeletions2 lengthInsertions2 = 0 @@ -681,20 +670,23 @@ func (dmp *DiffMatchPatch) DiffCleanupSemantic(diffs []Diff) []Diff { (len(lastequality) <= difference1) && (len(lastequality) <= difference2) { // Duplicate record. - insPoint := equalities[len(equalities)-1] - diffs = splice(diffs, insPoint, 0, Diff{DiffDelete, lastequality}) + insPoint := equalities.data + diffs = append( + diffs[:insPoint], + append([]Diff{Diff{DiffDelete, lastequality}}, diffs[insPoint:]...)...) // Change second copy to insert. diffs[insPoint+1].Type = DiffInsert // Throw away the equality we just deleted. - equalities = equalities[:len(equalities)-1] + equalities = equalities.next - if len(equalities) > 0 { - equalities = equalities[:len(equalities)-1] + if equalities != nil { + equalities = equalities.next } - pointer = -1 - if len(equalities) > 0 { - pointer = equalities[len(equalities)-1] + if equalities != nil { + pointer = equalities.data + } else { + pointer = -1 } lengthInsertions1 = 0 // Reset the counters. @@ -732,7 +724,10 @@ func (dmp *DiffMatchPatch) DiffCleanupSemantic(diffs []Diff) []Diff { float64(overlapLength1) >= float64(len(insertion))/2 { // Overlap found. Insert an equality and trim the surrounding edits. - diffs = splice(diffs, pointer, 0, Diff{DiffEqual, insertion[:overlapLength1]}) + diffs = append( + diffs[:pointer], + append([]Diff{Diff{DiffEqual, insertion[:overlapLength1]}}, diffs[pointer:]...)...) + diffs[pointer-1].Text = deletion[0 : len(deletion)-overlapLength1] diffs[pointer+1].Text = insertion[overlapLength1:] @@ -743,7 +738,10 @@ func (dmp *DiffMatchPatch) DiffCleanupSemantic(diffs []Diff) []Diff { float64(overlapLength2) >= float64(len(insertion))/2 { // Reverse overlap found. Insert an equality and swap and trim the surrounding edits. overlap := Diff{DiffEqual, deletion[:overlapLength2]} - diffs = splice(diffs, pointer, 0, overlap) + diffs = append( + diffs[:pointer], + append([]Diff{overlap}, diffs[pointer:]...)...) + diffs[pointer-1].Type = DiffInsert diffs[pointer-1].Text = insertion[0 : len(insertion)-overlapLength2] diffs[pointer+1].Type = DiffDelete @@ -956,7 +954,8 @@ func (dmp *DiffMatchPatch) DiffCleanupEfficiency(diffs []Diff) []Diff { insPoint := equalities.data // Duplicate record. - diffs = splice(diffs, insPoint, 0, Diff{DiffDelete, lastequality}) + diffs = append(diffs[:insPoint], + append([]Diff{Diff{DiffDelete, lastequality}}, diffs[insPoint:]...)...) // Change second copy to insert. diffs[insPoint+1].Type = DiffInsert @@ -1236,9 +1235,9 @@ func (dmp *DiffMatchPatch) DiffLevenshtein(diffs []Diff) int { for _, aDiff := range diffs { switch aDiff.Type { case DiffInsert: - insertions += utf8.RuneCountInString(aDiff.Text) + insertions += len(aDiff.Text) case DiffDelete: - deletions += utf8.RuneCountInString(aDiff.Text) + deletions += len(aDiff.Text) case DiffEqual: // A deletion and an insertion is one substitution. levenshtein += max(insertions, deletions) diff --git a/vendor/github.com/sergi/go-diff/diffmatchpatch/operation_string.go b/vendor/github.com/sergi/go-diff/diffmatchpatch/operation_string.go deleted file mode 100644 index 533ec0da..00000000 --- a/vendor/github.com/sergi/go-diff/diffmatchpatch/operation_string.go +++ /dev/null @@ -1,17 +0,0 @@ -// Code generated by "stringer -type=Operation -trimprefix=Diff"; DO NOT EDIT. - -package diffmatchpatch - -import "fmt" - -const _Operation_name = "DeleteEqualInsert" - -var _Operation_index = [...]uint8{0, 6, 11, 17} - -func (i Operation) String() string { - i -= -1 - if i < 0 || i >= Operation(len(_Operation_index)-1) { - return fmt.Sprintf("Operation(%d)", i+-1) - } - return _Operation_name[_Operation_index[i]:_Operation_index[i+1]] -} diff --git a/vendor/github.com/sirupsen/logrus/.golangci.yml b/vendor/github.com/sirupsen/logrus/.golangci.yml new file mode 100644 index 00000000..65dc2850 --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/.golangci.yml @@ -0,0 +1,40 @@ +run: + # do not run on test files yet + tests: false + +# all available settings of specific linters +linters-settings: + errcheck: + # report about not checking of errors in type assetions: `a := b.(MyStruct)`; + # default is false: such cases aren't reported by default. + check-type-assertions: false + + # report about assignment of errors to blank identifier: `num, _ := strconv.Atoi(numStr)`; + # default is false: such cases aren't reported by default. + check-blank: false + + lll: + line-length: 100 + tab-width: 4 + + prealloc: + simple: false + range-loops: false + for-loops: false + + whitespace: + multi-if: false # Enforces newlines (or comments) after every multi-line if statement + multi-func: false # Enforces newlines (or comments) after every multi-line function signature + +linters: + enable: + - megacheck + - govet + disable: + - maligned + - prealloc + disable-all: false + presets: + - bugs + - unused + fast: false diff --git a/vendor/github.com/sirupsen/logrus/.travis.yml b/vendor/github.com/sirupsen/logrus/.travis.yml index 848938a6..5e20aa41 100644 --- a/vendor/github.com/sirupsen/logrus/.travis.yml +++ b/vendor/github.com/sirupsen/logrus/.travis.yml @@ -4,21 +4,13 @@ git: depth: 1 env: - GO111MODULE=on - - GO111MODULE=off -go: [ 1.11.x, 1.12.x ] -os: [ linux, osx ] -matrix: - exclude: - - go: 1.12.x - env: GO111MODULE=off - - go: 1.11.x - os: osx +go: [1.13.x, 1.14.x] +os: [linux, osx] install: - ./travis/install.sh - - if [[ "$GO111MODULE" == "on" ]]; then go mod download; fi - - if [[ "$GO111MODULE" == "off" ]]; then go get github.com/stretchr/testify/assert golang.org/x/sys/unix github.com/konsorten/go-windows-terminal-sequences; fi script: - ./travis/cross_build.sh + - ./travis/lint.sh - export GOMAXPROCS=4 - export GORACE=halt_on_error=1 - go test -race -v ./... diff --git a/vendor/github.com/sirupsen/logrus/CHANGELOG.md b/vendor/github.com/sirupsen/logrus/CHANGELOG.md index 51a7ab0c..584026d6 100644 --- a/vendor/github.com/sirupsen/logrus/CHANGELOG.md +++ b/vendor/github.com/sirupsen/logrus/CHANGELOG.md @@ -1,9 +1,32 @@ +# 1.6.0 +Fixes: + * end of line cleanup + * revert the entry concurrency bug fix whic leads to deadlock under some circumstances + * update dependency on go-windows-terminal-sequences to fix a crash with go 1.14 + +Features: + * add an option to the `TextFormatter` to completely disable fields quoting + +# 1.5.0 +Code quality: + * add golangci linter run on travis + +Fixes: + * add mutex for hooks concurrent access on `Entry` data + * caller function field for go1.14 + * fix build issue for gopherjs target + +Feature: + * add an hooks/writer sub-package whose goal is to split output on different stream depending on the trace level + * add a `DisableHTMLEscape` option in the `JSONFormatter` + * add `ForceQuote` and `PadLevelText` options in the `TextFormatter` + # 1.4.2 * Fixes build break for plan9, nacl, solaris # 1.4.1 This new release introduces: * Enhance TextFormatter to not print caller information when they are empty (#944) - * Remove dependency on golang.org/x/crypto (#932, #943) + * Remove dependency on golang.org/x/crypto (#932, #943) Fixes: * Fix Entry.WithContext method to return a copy of the initial entry (#941) @@ -11,7 +34,7 @@ Fixes: # 1.4.0 This new release introduces: * Add `DeferExitHandler`, similar to `RegisterExitHandler` but prepending the handler to the list of handlers (semantically like `defer`) (#848). - * Add `CallerPrettyfier` to `JSONFormatter` and `TextFormatter (#909, #911) + * Add `CallerPrettyfier` to `JSONFormatter` and `TextFormatter` (#909, #911) * Add `Entry.WithContext()` and `Entry.Context`, to set a context on entries to be used e.g. in hooks (#919). Fixes: diff --git a/vendor/github.com/sirupsen/logrus/README.md b/vendor/github.com/sirupsen/logrus/README.md index a4796eb0..5796706d 100644 --- a/vendor/github.com/sirupsen/logrus/README.md +++ b/vendor/github.com/sirupsen/logrus/README.md @@ -1,8 +1,28 @@ -# Logrus :walrus: [![Build Status](https://travis-ci.org/sirupsen/logrus.svg?branch=master)](https://travis-ci.org/sirupsen/logrus) [![GoDoc](https://godoc.org/github.com/sirupsen/logrus?status.svg)](https://godoc.org/github.com/sirupsen/logrus) +# Logrus :walrus: [![Build Status](https://travis-ci.org/sirupsen/logrus.svg?branch=master)](https://travis-ci.org/sirupsen/logrus) [![GoDoc](https://godoc.org/github.com/sirupsen/logrus?status.svg)](https://godoc.org/github.com/sirupsen/logrus) Logrus is a structured logger for Go (golang), completely API compatible with the standard library logger. +**Logrus is in maintenance-mode.** We will not be introducing new features. It's +simply too hard to do in a way that won't break many people's projects, which is +the last thing you want from your Logging library (again...). + +This does not mean Logrus is dead. Logrus will continue to be maintained for +security, (backwards compatible) bug fixes, and performance (where we are +limited by the interface). + +I believe Logrus' biggest contribution is to have played a part in today's +widespread use of structured logging in Golang. There doesn't seem to be a +reason to do a major, breaking iteration into Logrus V2, since the fantastic Go +community has built those independently. Many fantastic alternatives have sprung +up. Logrus would look like those, had it been re-designed with what we know +about structured logging in Go today. Check out, for example, +[Zerolog][zerolog], [Zap][zap], and [Apex][apex]. + +[zerolog]: https://github.com/rs/zerolog +[zap]: https://github.com/uber-go/zap +[apex]: https://github.com/apex/log + **Seeing weird case-sensitive problems?** It's in the past been possible to import Logrus as both upper- and lower-case. Due to the Go package environment, this caused issues in the community and we needed a standard. Some environments @@ -15,11 +35,6 @@ comments](https://github.com/sirupsen/logrus/issues/553#issuecomment-306591437). For an in-depth explanation of the casing issue, see [this comment](https://github.com/sirupsen/logrus/issues/570#issuecomment-313933276). -**Are you interested in assisting in maintaining Logrus?** Currently I have a -lot of obligations, and I am unable to provide Logrus with the maintainership it -needs. If you'd like to help, please reach out to me at `simon at author's -username dot com`. - Nicely color-coded in development (when a TTY is attached, otherwise just plain text): @@ -187,7 +202,7 @@ func main() { log.Out = os.Stdout // You could set this to any `io.Writer` such as a file - // file, err := os.OpenFile("logrus.log", os.O_CREATE|os.O_WRONLY, 0666) + // file, err := os.OpenFile("logrus.log", os.O_CREATE|os.O_WRONLY|os.O_APPEND, 0666) // if err == nil { // log.Out = file // } else { @@ -272,7 +287,7 @@ func init() { ``` Note: Syslog hook also support connecting to local syslog (Ex. "/dev/log" or "/var/run/syslog" or "/var/run/log"). For the detail, please check the [syslog hook README](hooks/syslog/README.md). -A list of currently known of service hook can be found in this wiki [page](https://github.com/sirupsen/logrus/wiki/Hooks) +A list of currently known service hooks can be found in this wiki [page](https://github.com/sirupsen/logrus/wiki/Hooks) #### Level logging @@ -354,6 +369,7 @@ The built-in logging formatters are: [github.com/mattn/go-colorable](https://github.com/mattn/go-colorable). * When colors are enabled, levels are truncated to 4 characters by default. To disable truncation set the `DisableLevelTruncation` field to `true`. + * When outputting to a TTY, it's often helpful to visually scan down a column where all the levels are the same width. Setting the `PadLevelText` field to `true` enables this behavior, by adding padding to the level text. * All options are listed in the [generated docs](https://godoc.org/github.com/sirupsen/logrus#TextFormatter). * `logrus.JSONFormatter`. Logs fields as JSON. * All options are listed in the [generated docs](https://godoc.org/github.com/sirupsen/logrus#JSONFormatter). @@ -364,8 +380,10 @@ Third party logging formatters: * [`GELF`](https://github.com/fabienm/go-logrus-formatters). Formats entries so they comply to Graylog's [GELF 1.1 specification](http://docs.graylog.org/en/2.4/pages/gelf.html). * [`logstash`](https://github.com/bshuster-repo/logrus-logstash-hook). Logs fields as [Logstash](http://logstash.net) Events. * [`prefixed`](https://github.com/x-cray/logrus-prefixed-formatter). Displays log entry source along with alternative layout. -* [`zalgo`](https://github.com/aybabtme/logzalgo). Invoking the P͉̫o̳̼̊w̖͈̰͎e̬͔̭͂r͚̼̹̲ ̫͓͉̳͈ō̠͕͖̚f̝͍̠ ͕̲̞͖͑Z̖̫̤̫ͪa͉̬͈̗l͖͎g̳̥o̰̥̅!̣͔̲̻͊̄ ̙̘̦̹̦. +* [`zalgo`](https://github.com/aybabtme/logzalgo). Invoking the Power of Zalgo. * [`nested-logrus-formatter`](https://github.com/antonfisher/nested-logrus-formatter). Converts logrus fields to a nested structure. +* [`powerful-logrus-formatter`](https://github.com/zput/zxcTool). get fileName, log's line number and the latest function's name when print log; Sava log to files. +* [`caption-json-formatter`](https://github.com/nolleh/caption_json_formatter). logrus's message json formatter with human-readable caption added. You can define your formatter by implementing the `Formatter` interface, requiring a `Format` method. `Format` takes an `*Entry`. `entry.Data` is a @@ -430,14 +448,14 @@ entries. It should not be a feature of the application-level logger. | Tool | Description | | ---- | ----------- | -|[Logrus Mate](https://github.com/gogap/logrus_mate)|Logrus mate is a tool for Logrus to manage loggers, you can initial logger's level, hook and formatter by config file, the logger will generated with different config at different environment.| +|[Logrus Mate](https://github.com/gogap/logrus_mate)|Logrus mate is a tool for Logrus to manage loggers, you can initial logger's level, hook and formatter by config file, the logger will be generated with different configs in different environments.| |[Logrus Viper Helper](https://github.com/heirko/go-contrib/tree/master/logrusHelper)|An Helper around Logrus to wrap with spf13/Viper to load configuration with fangs! And to simplify Logrus configuration use some behavior of [Logrus Mate](https://github.com/gogap/logrus_mate). [sample](https://github.com/heirko/iris-contrib/blob/master/middleware/logrus-logger/example) | #### Testing Logrus has a built in facility for asserting the presence of log messages. This is implemented through the `test` hook and provides: -* decorators for existing logger (`test.NewLocal` and `test.NewGlobal`) which basically just add the `test` hook +* decorators for existing logger (`test.NewLocal` and `test.NewGlobal`) which basically just adds the `test` hook * a test logger (`test.NewNullLogger`) that just records log messages (and does not output any): ```go @@ -465,7 +483,7 @@ func TestSomething(t*testing.T){ Logrus can register one or more functions that will be called when any `fatal` level message is logged. The registered handlers will be executed before -logrus performs a `os.Exit(1)`. This behavior may be helpful if callers need +logrus performs an `os.Exit(1)`. This behavior may be helpful if callers need to gracefully shutdown. Unlike a `panic("Something went wrong...")` call which can be intercepted with a deferred `recover` a call to `os.Exit(1)` can not be intercepted. ``` @@ -490,6 +508,6 @@ Situation when locking is not needed includes: 1) logger.Out is protected by locks. - 2) logger.Out is a os.File handler opened with `O_APPEND` flag, and every write is smaller than 4k. (This allow multi-thread/multi-process writing) + 2) logger.Out is an os.File handler opened with `O_APPEND` flag, and every write is smaller than 4k. (This allows multi-thread/multi-process writing) (Refer to http://www.notthewizard.com/2014/06/17/are-files-appends-really-atomic/) diff --git a/vendor/github.com/sirupsen/logrus/appveyor.yml b/vendor/github.com/sirupsen/logrus/appveyor.yml index 96c2ce15..df9d65c3 100644 --- a/vendor/github.com/sirupsen/logrus/appveyor.yml +++ b/vendor/github.com/sirupsen/logrus/appveyor.yml @@ -1,14 +1,14 @@ -version: "{build}" -platform: x64 -clone_folder: c:\gopath\src\github.com\sirupsen\logrus -environment: - GOPATH: c:\gopath -branches: - only: - - master -install: - - set PATH=%GOPATH%\bin;c:\go\bin;%PATH% - - go version -build_script: - - go get -t - - go test +version: "{build}" +platform: x64 +clone_folder: c:\gopath\src\github.com\sirupsen\logrus +environment: + GOPATH: c:\gopath +branches: + only: + - master +install: + - set PATH=%GOPATH%\bin;c:\go\bin;%PATH% + - go version +build_script: + - go get -t + - go test diff --git a/vendor/github.com/sirupsen/logrus/entry.go b/vendor/github.com/sirupsen/logrus/entry.go index 63e25583..f6e062a3 100644 --- a/vendor/github.com/sirupsen/logrus/entry.go +++ b/vendor/github.com/sirupsen/logrus/entry.go @@ -85,10 +85,15 @@ func NewEntry(logger *Logger) *Entry { } } +// Returns the bytes representation of this entry from the formatter. +func (entry *Entry) Bytes() ([]byte, error) { + return entry.Logger.Formatter.Format(entry) +} + // Returns the string representation from the reader and ultimately the // formatter. func (entry *Entry) String() (string, error) { - serialized, err := entry.Logger.Formatter.Format(entry) + serialized, err := entry.Bytes() if err != nil { return "", err } @@ -103,7 +108,11 @@ func (entry *Entry) WithError(err error) *Entry { // Add a context to the Entry. func (entry *Entry) WithContext(ctx context.Context) *Entry { - return &Entry{Logger: entry.Logger, Data: entry.Data, Time: entry.Time, err: entry.err, Context: ctx} + dataCopy := make(Fields, len(entry.Data)) + for k, v := range entry.Data { + dataCopy[k] = v + } + return &Entry{Logger: entry.Logger, Data: dataCopy, Time: entry.Time, err: entry.err, Context: ctx} } // Add a single field to the Entry. @@ -144,7 +153,11 @@ func (entry *Entry) WithFields(fields Fields) *Entry { // Overrides the time of the Entry. func (entry *Entry) WithTime(t time.Time) *Entry { - return &Entry{Logger: entry.Logger, Data: entry.Data, Time: t, err: entry.err, Context: entry.Context} + dataCopy := make(Fields, len(entry.Data)) + for k, v := range entry.Data { + dataCopy[k] = v + } + return &Entry{Logger: entry.Logger, Data: dataCopy, Time: t, err: entry.err, Context: entry.Context} } // getPackageName reduces a fully qualified function name to the package name @@ -165,15 +178,20 @@ func getPackageName(f string) string { // getCaller retrieves the name of the first non-logrus calling function func getCaller() *runtime.Frame { - // cache this package's fully-qualified name callerInitOnce.Do(func() { - pcs := make([]uintptr, 2) + pcs := make([]uintptr, maximumCallerDepth) _ = runtime.Callers(0, pcs) - logrusPackage = getPackageName(runtime.FuncForPC(pcs[1]).Name()) - // now that we have the cache, we can skip a minimum count of known-logrus functions - // XXX this is dubious, the number of frames may vary + // dynamic get the package name and the minimum caller depth + for i := 0; i < maximumCallerDepth; i++ { + funcName := runtime.FuncForPC(pcs[i]).Name() + if strings.Contains(funcName, "getCaller") { + logrusPackage = getPackageName(funcName) + break + } + } + minimumCallerDepth = knownLogrusFrames }) @@ -187,7 +205,7 @@ func getCaller() *runtime.Frame { // If the caller isn't part of this package, we're done if pkg != logrusPackage { - return &f + return &f //nolint:scopelint } } @@ -217,9 +235,11 @@ func (entry Entry) log(level Level, msg string) { entry.Level = level entry.Message = msg + entry.Logger.mu.Lock() if entry.Logger.ReportCaller { entry.Caller = getCaller() } + entry.Logger.mu.Unlock() entry.fireHooks() @@ -255,11 +275,10 @@ func (entry *Entry) write() { serialized, err := entry.Logger.Formatter.Format(entry) if err != nil { fmt.Fprintf(os.Stderr, "Failed to obtain reader, %v\n", err) - } else { - _, err = entry.Logger.Out.Write(serialized) - if err != nil { - fmt.Fprintf(os.Stderr, "Failed to write to log, %v\n", err) - } + return + } + if _, err = entry.Logger.Out.Write(serialized); err != nil { + fmt.Fprintf(os.Stderr, "Failed to write to log, %v\n", err) } } diff --git a/vendor/github.com/sirupsen/logrus/exported.go b/vendor/github.com/sirupsen/logrus/exported.go index 62fc2f21..42b04f6c 100644 --- a/vendor/github.com/sirupsen/logrus/exported.go +++ b/vendor/github.com/sirupsen/logrus/exported.go @@ -80,7 +80,7 @@ func WithFields(fields Fields) *Entry { return std.WithFields(fields) } -// WithTime creats an entry from the standard logger and overrides the time of +// WithTime creates an entry from the standard logger and overrides the time of // logs generated with it. // // Note that it doesn't log until you call Debug, Print, Info, Warn, Fatal diff --git a/vendor/github.com/sirupsen/logrus/go.mod b/vendor/github.com/sirupsen/logrus/go.mod index 12fdf989..d4132967 100644 --- a/vendor/github.com/sirupsen/logrus/go.mod +++ b/vendor/github.com/sirupsen/logrus/go.mod @@ -2,9 +2,10 @@ module github.com/sirupsen/logrus require ( github.com/davecgh/go-spew v1.1.1 // indirect - github.com/konsorten/go-windows-terminal-sequences v1.0.1 + github.com/konsorten/go-windows-terminal-sequences v1.0.3 github.com/pmezard/go-difflib v1.0.0 // indirect - github.com/stretchr/objx v0.1.1 // indirect github.com/stretchr/testify v1.2.2 golang.org/x/sys v0.0.0-20190422165155-953cdadca894 ) + +go 1.13 diff --git a/vendor/github.com/sirupsen/logrus/go.sum b/vendor/github.com/sirupsen/logrus/go.sum index 596c318b..49c690f2 100644 --- a/vendor/github.com/sirupsen/logrus/go.sum +++ b/vendor/github.com/sirupsen/logrus/go.sum @@ -1,16 +1,12 @@ github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c= github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= -github.com/konsorten/go-windows-terminal-sequences v0.0.0-20180402223658-b729f2633dfe h1:CHRGQ8V7OlCYtwaKPJi3iA7J+YdNKdo8j7nG5IgDhjs= -github.com/konsorten/go-windows-terminal-sequences v0.0.0-20180402223658-b729f2633dfe/go.mod h1:T0+1ngSBFLxvqU3pZ+m/2kptfBszLMUkC4ZK/EgS/cQ= github.com/konsorten/go-windows-terminal-sequences v1.0.1 h1:mweAR1A6xJ3oS2pRaGiHgQ4OO8tzTaLawm8vnODuwDk= github.com/konsorten/go-windows-terminal-sequences v1.0.1/go.mod h1:T0+1ngSBFLxvqU3pZ+m/2kptfBszLMUkC4ZK/EgS/cQ= +github.com/konsorten/go-windows-terminal-sequences v1.0.3 h1:CE8S1cTafDpPvMhIxNJKvHsGVBgn1xWYf1NbHQhywc8= +github.com/konsorten/go-windows-terminal-sequences v1.0.3/go.mod h1:T0+1ngSBFLxvqU3pZ+m/2kptfBszLMUkC4ZK/EgS/cQ= github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM= github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= -github.com/stretchr/objx v0.1.1 h1:2vfRuCMp5sSVIDSqO8oNnWJq7mPa6KVP3iPIwFBuy8A= -github.com/stretchr/objx v0.1.1/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME= github.com/stretchr/testify v1.2.2 h1:bSDNvY7ZPG5RlJ8otE/7V6gMiyenm9RtJ7IUVIAoJ1w= github.com/stretchr/testify v1.2.2/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXfy6kDkUVs= -golang.org/x/sys v0.0.0-20180905080454-ebe1bf3edb33 h1:I6FyU15t786LL7oL/hn43zqTuEGr4PN7F4XJ1p4E3Y8= -golang.org/x/sys v0.0.0-20180905080454-ebe1bf3edb33/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20190422165155-953cdadca894 h1:Cz4ceDQGXuKRnVBDTS23GTn/pU5OE2C0WrNTOYK1Uuc= golang.org/x/sys v0.0.0-20190422165155-953cdadca894/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= diff --git a/vendor/github.com/sirupsen/logrus/json_formatter.go b/vendor/github.com/sirupsen/logrus/json_formatter.go index 098a21a0..ba7f2371 100644 --- a/vendor/github.com/sirupsen/logrus/json_formatter.go +++ b/vendor/github.com/sirupsen/logrus/json_formatter.go @@ -28,6 +28,9 @@ type JSONFormatter struct { // DisableTimestamp allows disabling automatic timestamps in output DisableTimestamp bool + // DisableHTMLEscape allows disabling html escaping in output + DisableHTMLEscape bool + // DataKey allows users to put all the log entry parameters into a nested dictionary at a given key. DataKey string @@ -110,6 +113,7 @@ func (f *JSONFormatter) Format(entry *Entry) ([]byte, error) { } encoder := json.NewEncoder(b) + encoder.SetEscapeHTML(!f.DisableHTMLEscape) if f.PrettyPrint { encoder.SetIndent("", " ") } diff --git a/vendor/github.com/sirupsen/logrus/logger.go b/vendor/github.com/sirupsen/logrus/logger.go index c0c0b1e5..6fdda748 100644 --- a/vendor/github.com/sirupsen/logrus/logger.go +++ b/vendor/github.com/sirupsen/logrus/logger.go @@ -68,10 +68,10 @@ func (mw *MutexWrap) Disable() { // `Out` and `Hooks` directly on the default logger instance. You can also just // instantiate your own: // -// var log = &Logger{ +// var log = &logrus.Logger{ // Out: os.Stderr, -// Formatter: new(JSONFormatter), -// Hooks: make(LevelHooks), +// Formatter: new(logrus.JSONFormatter), +// Hooks: make(logrus.LevelHooks), // Level: logrus.DebugLevel, // } // @@ -100,8 +100,9 @@ func (logger *Logger) releaseEntry(entry *Entry) { logger.entryPool.Put(entry) } -// Adds a field to the log entry, note that it doesn't log until you call -// Debug, Print, Info, Warn, Error, Fatal or Panic. It only creates a log entry. +// WithField allocates a new entry and adds a field to it. +// Debug, Print, Info, Warn, Error, Fatal or Panic must be then applied to +// this new returned entry. // If you want multiple fields, use `WithFields`. func (logger *Logger) WithField(key string, value interface{}) *Entry { entry := logger.newEntry() diff --git a/vendor/github.com/sirupsen/logrus/logrus.go b/vendor/github.com/sirupsen/logrus/logrus.go index 8644761f..2f16224c 100644 --- a/vendor/github.com/sirupsen/logrus/logrus.go +++ b/vendor/github.com/sirupsen/logrus/logrus.go @@ -51,7 +51,7 @@ func (level *Level) UnmarshalText(text []byte) error { return err } - *level = Level(l) + *level = l return nil } diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_bsd.go b/vendor/github.com/sirupsen/logrus/terminal_check_bsd.go index 3c4f43f9..49978998 100644 --- a/vendor/github.com/sirupsen/logrus/terminal_check_bsd.go +++ b/vendor/github.com/sirupsen/logrus/terminal_check_bsd.go @@ -1,4 +1,5 @@ // +build darwin dragonfly freebsd netbsd openbsd +// +build !js package logrus @@ -10,4 +11,3 @@ func isTerminal(fd int) bool { _, err := unix.IoctlGetTermios(fd, ioctlReadTermios) return err == nil } - diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_js.go b/vendor/github.com/sirupsen/logrus/terminal_check_js.go new file mode 100644 index 00000000..ebdae3ec --- /dev/null +++ b/vendor/github.com/sirupsen/logrus/terminal_check_js.go @@ -0,0 +1,7 @@ +// +build js + +package logrus + +func isTerminal(fd int) bool { + return false +} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_unix.go b/vendor/github.com/sirupsen/logrus/terminal_check_unix.go index 355dc966..cc4fe6e3 100644 --- a/vendor/github.com/sirupsen/logrus/terminal_check_unix.go +++ b/vendor/github.com/sirupsen/logrus/terminal_check_unix.go @@ -1,4 +1,5 @@ // +build linux aix +// +build !js package logrus @@ -10,4 +11,3 @@ func isTerminal(fd int) bool { _, err := unix.IoctlGetTermios(fd, ioctlReadTermios) return err == nil } - diff --git a/vendor/github.com/sirupsen/logrus/text_formatter.go b/vendor/github.com/sirupsen/logrus/text_formatter.go index e01587c4..3c28b54c 100644 --- a/vendor/github.com/sirupsen/logrus/text_formatter.go +++ b/vendor/github.com/sirupsen/logrus/text_formatter.go @@ -6,9 +6,11 @@ import ( "os" "runtime" "sort" + "strconv" "strings" "sync" "time" + "unicode/utf8" ) const ( @@ -32,6 +34,14 @@ type TextFormatter struct { // Force disabling colors. DisableColors bool + // Force quoting of all values + ForceQuote bool + + // DisableQuote disables quoting for all values. + // DisableQuote will have a lower priority than ForceQuote. + // If both of them are set to true, quote will be forced on all values. + DisableQuote bool + // Override coloring based on CLICOLOR and CLICOLOR_FORCE. - https://bixense.com/clicolors/ EnvironmentOverrideColors bool @@ -57,6 +67,10 @@ type TextFormatter struct { // Disables the truncation of the level text to 4 characters. DisableLevelTruncation bool + // PadLevelText Adds padding the level text so that all the levels output at the same length + // PadLevelText is a superset of the DisableLevelTruncation option + PadLevelText bool + // QuoteEmptyFields will wrap empty fields in quotes if true QuoteEmptyFields bool @@ -79,23 +93,32 @@ type TextFormatter struct { CallerPrettyfier func(*runtime.Frame) (function string, file string) terminalInitOnce sync.Once + + // The max length of the level text, generated dynamically on init + levelTextMaxLength int } func (f *TextFormatter) init(entry *Entry) { if entry.Logger != nil { f.isTerminal = checkIfTerminal(entry.Logger.Out) } + // Get the max length of the level text + for _, level := range AllLevels { + levelTextLength := utf8.RuneCount([]byte(level.String())) + if levelTextLength > f.levelTextMaxLength { + f.levelTextMaxLength = levelTextLength + } + } } func (f *TextFormatter) isColored() bool { isColored := f.ForceColors || (f.isTerminal && (runtime.GOOS != "windows")) if f.EnvironmentOverrideColors { - if force, ok := os.LookupEnv("CLICOLOR_FORCE"); ok && force != "0" { + switch force, ok := os.LookupEnv("CLICOLOR_FORCE"); { + case ok && force != "0": isColored = true - } else if ok && force == "0" { - isColored = false - } else if os.Getenv("CLICOLOR") == "0" { + case ok && force == "0", os.Getenv("CLICOLOR") == "0": isColored = false } } @@ -217,9 +240,18 @@ func (f *TextFormatter) printColored(b *bytes.Buffer, entry *Entry, keys []strin } levelText := strings.ToUpper(entry.Level.String()) - if !f.DisableLevelTruncation { + if !f.DisableLevelTruncation && !f.PadLevelText { levelText = levelText[0:4] } + if f.PadLevelText { + // Generates the format string used in the next line, for example "%-6s" or "%-7s". + // Based on the max level text length. + formatString := "%-" + strconv.Itoa(f.levelTextMaxLength) + "s" + // Formats the level text by appending spaces up to the max length, for example: + // - "INFO " + // - "WARNING" + levelText = fmt.Sprintf(formatString, levelText) + } // Remove a single newline if it already exists in the message to keep // the behavior of logrus text_formatter the same as the stdlib log package @@ -243,11 +275,12 @@ func (f *TextFormatter) printColored(b *bytes.Buffer, entry *Entry, keys []strin } } - if f.DisableTimestamp { + switch { + case f.DisableTimestamp: fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m%s %-44s ", levelColor, levelText, caller, entry.Message) - } else if !f.FullTimestamp { + case !f.FullTimestamp: fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%04d]%s %-44s ", levelColor, levelText, int(entry.Time.Sub(baseTimestamp)/time.Second), caller, entry.Message) - } else { + default: fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%s]%s %-44s ", levelColor, levelText, entry.Time.Format(timestampFormat), caller, entry.Message) } for _, k := range keys { @@ -258,9 +291,15 @@ func (f *TextFormatter) printColored(b *bytes.Buffer, entry *Entry, keys []strin } func (f *TextFormatter) needsQuoting(text string) bool { + if f.ForceQuote { + return true + } if f.QuoteEmptyFields && len(text) == 0 { return true } + if f.DisableQuote { + return false + } for _, ch := range text { if !((ch >= 'a' && ch <= 'z') || (ch >= 'A' && ch <= 'Z') || diff --git a/vendor/github.com/sirupsen/logrus/writer.go b/vendor/github.com/sirupsen/logrus/writer.go index 9e1f7513..72e8e3a1 100644 --- a/vendor/github.com/sirupsen/logrus/writer.go +++ b/vendor/github.com/sirupsen/logrus/writer.go @@ -6,10 +6,16 @@ import ( "runtime" ) +// Writer at INFO level. See WriterLevel for details. func (logger *Logger) Writer() *io.PipeWriter { return logger.WriterLevel(InfoLevel) } +// WriterLevel returns an io.Writer that can be used to write arbitrary text to +// the logger at the given log level. Each line written to the writer will be +// printed in the usual way using formatters and hooks. The writer is part of an +// io.Pipe and it is the callers responsibility to close the writer when done. +// This can be used to override the standard library logger easily. func (logger *Logger) WriterLevel(level Level) *io.PipeWriter { return NewEntry(logger).WriterLevel(level) } diff --git a/vendor/github.com/spf13/cobra/README.md b/vendor/github.com/spf13/cobra/README.md index 2f8175bc..9d799342 100644 --- a/vendor/github.com/spf13/cobra/README.md +++ b/vendor/github.com/spf13/cobra/README.md @@ -25,10 +25,10 @@ Many of the most widely used Go projects are built using Cobra, such as: [mattermost-server](https://github.com/mattermost/mattermost-server), [Gardener](https://github.com/gardener/gardenctl), [Linkerd](https://linkerd.io/), +[Github CLI](https://github.com/cli/cli) etc. [![Build Status](https://travis-ci.org/spf13/cobra.svg "Travis CI status")](https://travis-ci.org/spf13/cobra) -[![CircleCI status](https://circleci.com/gh/spf13/cobra.png?circle-token=:circle-token "CircleCI status")](https://circleci.com/gh/spf13/cobra) [![GoDoc](https://godoc.org/github.com/spf13/cobra?status.svg)](https://godoc.org/github.com/spf13/cobra) [![Go Report Card](https://goreportcard.com/badge/github.com/spf13/cobra)](https://goreportcard.com/report/github.com/spf13/cobra) diff --git a/vendor/github.com/spf13/cobra/args.go b/vendor/github.com/spf13/cobra/args.go index c4d820b8..70e9b262 100644 --- a/vendor/github.com/spf13/cobra/args.go +++ b/vendor/github.com/spf13/cobra/args.go @@ -2,6 +2,7 @@ package cobra import ( "fmt" + "strings" ) type PositionalArgs func(cmd *Command, args []string) error @@ -34,8 +35,15 @@ func NoArgs(cmd *Command, args []string) error { // OnlyValidArgs returns an error if any args are not in the list of ValidArgs. func OnlyValidArgs(cmd *Command, args []string) error { if len(cmd.ValidArgs) > 0 { + // Remove any description that may be included in ValidArgs. + // A description is following a tab character. + var validArgs []string + for _, v := range cmd.ValidArgs { + validArgs = append(validArgs, strings.Split(v, "\t")[0]) + } + for _, v := range args { - if !stringInSlice(v, cmd.ValidArgs) { + if !stringInSlice(v, validArgs) { return fmt.Errorf("invalid argument %q for %q%s", v, cmd.CommandPath(), cmd.findSuggestions(args[0])) } } diff --git a/vendor/github.com/spf13/cobra/bash_completions.go b/vendor/github.com/spf13/cobra/bash_completions.go index 1e0e25cf..1e27188c 100644 --- a/vendor/github.com/spf13/cobra/bash_completions.go +++ b/vendor/github.com/spf13/cobra/bash_completions.go @@ -58,6 +58,67 @@ __%[1]s_contains_word() return 1 } +__%[1]s_handle_go_custom_completion() +{ + __%[1]s_debug "${FUNCNAME[0]}: cur is ${cur}, words[*] is ${words[*]}, #words[@] is ${#words[@]}" + + local out requestComp lastParam lastChar comp directive args + + # Prepare the command to request completions for the program. + # Calling ${words[0]} instead of directly %[1]s allows to handle aliases + args=("${words[@]:1}") + requestComp="${words[0]} %[2]s ${args[*]}" + + lastParam=${words[$((${#words[@]}-1))]} + lastChar=${lastParam:$((${#lastParam}-1)):1} + __%[1]s_debug "${FUNCNAME[0]}: lastParam ${lastParam}, lastChar ${lastChar}" + + if [ -z "${cur}" ] && [ "${lastChar}" != "=" ]; then + # If the last parameter is complete (there is a space following it) + # We add an extra empty parameter so we can indicate this to the go method. + __%[1]s_debug "${FUNCNAME[0]}: Adding extra empty parameter" + requestComp="${requestComp} \"\"" + fi + + __%[1]s_debug "${FUNCNAME[0]}: calling ${requestComp}" + # Use eval to handle any environment variables and such + out=$(eval "${requestComp}" 2>/dev/null) + + # Extract the directive integer at the very end of the output following a colon (:) + directive=${out##*:} + # Remove the directive + out=${out%%:*} + if [ "${directive}" = "${out}" ]; then + # There is not directive specified + directive=0 + fi + __%[1]s_debug "${FUNCNAME[0]}: the completion directive is: ${directive}" + __%[1]s_debug "${FUNCNAME[0]}: the completions are: ${out[*]}" + + if [ $((directive & %[3]d)) -ne 0 ]; then + # Error code. No completion. + __%[1]s_debug "${FUNCNAME[0]}: received error from custom completion go code" + return + else + if [ $((directive & %[4]d)) -ne 0 ]; then + if [[ $(type -t compopt) = "builtin" ]]; then + __%[1]s_debug "${FUNCNAME[0]}: activating no space" + compopt -o nospace + fi + fi + if [ $((directive & %[5]d)) -ne 0 ]; then + if [[ $(type -t compopt) = "builtin" ]]; then + __%[1]s_debug "${FUNCNAME[0]}: activating no file completion" + compopt +o default + fi + fi + + while IFS='' read -r comp; do + COMPREPLY+=("$comp") + done < <(compgen -W "${out[*]}" -- "$cur") + fi +} + __%[1]s_handle_reply() { __%[1]s_debug "${FUNCNAME[0]}" @@ -121,6 +182,10 @@ __%[1]s_handle_reply() completions=("${commands[@]}") if [[ ${#must_have_one_noun[@]} -ne 0 ]]; then completions=("${must_have_one_noun[@]}") + elif [[ -n "${has_completion_function}" ]]; then + # if a go completion function is provided, defer to that function + completions=() + __%[1]s_handle_go_custom_completion fi if [[ ${#must_have_one_flag[@]} -ne 0 ]]; then completions+=("${must_have_one_flag[@]}") @@ -279,7 +344,7 @@ __%[1]s_handle_word() __%[1]s_handle_word } -`, name)) +`, name, ShellCompNoDescRequestCmd, ShellCompDirectiveError, ShellCompDirectiveNoSpace, ShellCompDirectiveNoFileComp)) } func writePostscript(buf *bytes.Buffer, name string) { @@ -304,6 +369,7 @@ func writePostscript(buf *bytes.Buffer, name string) { local commands=("%[1]s") local must_have_one_flag=() local must_have_one_noun=() + local has_completion_function local last_command local nouns=() @@ -404,7 +470,22 @@ func writeLocalNonPersistentFlag(buf *bytes.Buffer, flag *pflag.Flag) { buf.WriteString(fmt.Sprintf(format, name)) } +// Setup annotations for go completions for registered flags +func prepareCustomAnnotationsForFlags(cmd *Command) { + for flag := range flagCompletionFunctions { + // Make sure the completion script calls the __*_go_custom_completion function for + // every registered flag. We need to do this here (and not when the flag was registered + // for completion) so that we can know the root command name for the prefix + // of ___go_custom_completion + if flag.Annotations == nil { + flag.Annotations = map[string][]string{} + } + flag.Annotations[BashCompCustom] = []string{fmt.Sprintf("__%[1]s_handle_go_custom_completion", cmd.Root().Name())} + } +} + func writeFlags(buf *bytes.Buffer, cmd *Command) { + prepareCustomAnnotationsForFlags(cmd) buf.WriteString(` flags=() two_word_flags=() local_nonpersistent_flags=() @@ -467,8 +548,14 @@ func writeRequiredNouns(buf *bytes.Buffer, cmd *Command) { buf.WriteString(" must_have_one_noun=()\n") sort.Sort(sort.StringSlice(cmd.ValidArgs)) for _, value := range cmd.ValidArgs { + // Remove any description that may be included following a tab character. + // Descriptions are not supported by bash completion. + value = strings.Split(value, "\t")[0] buf.WriteString(fmt.Sprintf(" must_have_one_noun+=(%q)\n", value)) } + if cmd.ValidArgsFunction != nil { + buf.WriteString(" has_completion_function=1\n") + } } func writeCmdAliases(buf *bytes.Buffer, cmd *Command) { diff --git a/vendor/github.com/spf13/cobra/bash_completions.md b/vendor/github.com/spf13/cobra/bash_completions.md index 4ac61ee1..e61a3a65 100644 --- a/vendor/github.com/spf13/cobra/bash_completions.md +++ b/vendor/github.com/spf13/cobra/bash_completions.md @@ -56,7 +56,149 @@ func main() { `out.sh` will get you completions of subcommands and flags. Copy it to `/etc/bash_completion.d/` as described [here](https://debian-administration.org/article/316/An_introduction_to_bash_completion_part_1) and reset your terminal to use autocompletion. If you make additional annotations to your code, you can get even more intelligent and flexible behavior. -## Creating your own custom functions +## Have the completions code complete your 'nouns' + +### Static completion of nouns + +This method allows you to provide a pre-defined list of completion choices for your nouns using the `validArgs` field. +For example, if you want `kubectl get [tab][tab]` to show a list of valid "nouns" you have to set them. Simplified code from `kubectl get` looks like: + +```go +validArgs []string = { "pod", "node", "service", "replicationcontroller" } + +cmd := &cobra.Command{ + Use: "get [(-o|--output=)json|yaml|template|...] (RESOURCE [NAME] | RESOURCE/NAME ...)", + Short: "Display one or many resources", + Long: get_long, + Example: get_example, + Run: func(cmd *cobra.Command, args []string) { + err := RunGet(f, out, cmd, args) + util.CheckErr(err) + }, + ValidArgs: validArgs, +} +``` + +Notice we put the "ValidArgs" on the "get" subcommand. Doing so will give results like + +```bash +# kubectl get [tab][tab] +node pod replicationcontroller service +``` + +### Plural form and shortcuts for nouns + +If your nouns have a number of aliases, you can define them alongside `ValidArgs` using `ArgAliases`: + +```go +argAliases []string = { "pods", "nodes", "services", "svc", "replicationcontrollers", "rc" } + +cmd := &cobra.Command{ + ... + ValidArgs: validArgs, + ArgAliases: argAliases +} +``` + +The aliases are not shown to the user on tab completion, but they are accepted as valid nouns by +the completion algorithm if entered manually, e.g. in: + +```bash +# kubectl get rc [tab][tab] +backend frontend database +``` + +Note that without declaring `rc` as an alias, the completion algorithm would show the list of nouns +in this example again instead of the replication controllers. + +### Dynamic completion of nouns + +In some cases it is not possible to provide a list of possible completions in advance. Instead, the list of completions must be determined at execution-time. Cobra provides two ways of defining such dynamic completion of nouns. Note that both these methods can be used along-side each other as long as they are not both used for the same command. + +**Note**: *Custom Completions written in Go* will automatically work for other shell-completion scripts (e.g., Fish shell), while *Custom Completions written in Bash* will only work for Bash shell-completion. It is therefore recommended to use *Custom Completions written in Go*. + +#### 1. Custom completions of nouns written in Go + +In a similar fashion as for static completions, you can use the `ValidArgsFunction` field to provide a Go function that Cobra will execute when it needs the list of completion choices for the nouns of a command. Note that either `ValidArgs` or `ValidArgsFunction` can be used for a single cobra command, but not both. +Simplified code from `helm status` looks like: + +```go +cmd := &cobra.Command{ + Use: "status RELEASE_NAME", + Short: "Display the status of the named release", + Long: status_long, + RunE: func(cmd *cobra.Command, args []string) { + RunGet(args[0]) + }, + ValidArgsFunction: func(cmd *cobra.Command, args []string, toComplete string) ([]string, cobra.ShellCompDirective) { + if len(args) != 0 { + return nil, cobra.ShellCompDirectiveNoFileComp + } + return getReleasesFromCluster(toComplete), cobra.ShellCompDirectiveNoFileComp + }, +} +``` +Where `getReleasesFromCluster()` is a Go function that obtains the list of current Helm releases running on the Kubernetes cluster. +Notice we put the `ValidArgsFunction` on the `status` subcommand. Let's assume the Helm releases on the cluster are: `harbor`, `notary`, `rook` and `thanos` then this dynamic completion will give results like + +```bash +# helm status [tab][tab] +harbor notary rook thanos +``` +You may have noticed the use of `cobra.ShellCompDirective`. These directives are bit fields allowing to control some shell completion behaviors for your particular completion. You can combine them with the bit-or operator such as `cobra.ShellCompDirectiveNoSpace | cobra.ShellCompDirectiveNoFileComp` +```go +// Indicates an error occurred and completions should be ignored. +ShellCompDirectiveError +// Indicates that the shell should not add a space after the completion, +// even if there is a single completion provided. +ShellCompDirectiveNoSpace +// Indicates that the shell should not provide file completion even when +// no completion is provided. +// This currently does not work for zsh or bash < 4 +ShellCompDirectiveNoFileComp +// Indicates that the shell will perform its default behavior after completions +// have been provided (this implies !ShellCompDirectiveNoSpace && !ShellCompDirectiveNoFileComp). +ShellCompDirectiveDefault +``` + +When using the `ValidArgsFunction`, Cobra will call your registered function after having parsed all flags and arguments provided in the command-line. You therefore don't need to do this parsing yourself. For example, when a user calls `helm status --namespace my-rook-ns [tab][tab]`, Cobra will call your registered `ValidArgsFunction` after having parsed the `--namespace` flag, as it would have done when calling the `RunE` function. + +##### Debugging + +Cobra achieves dynamic completions written in Go through the use of a hidden command called by the completion script. To debug your Go completion code, you can call this hidden command directly: +```bash +# helm __complete status har +harbor +:4 +Completion ended with directive: ShellCompDirectiveNoFileComp # This is on stderr +``` +***Important:*** If the noun to complete is empty, you must pass an empty parameter to the `__complete` command: +```bash +# helm __complete status "" +harbor +notary +rook +thanos +:4 +Completion ended with directive: ShellCompDirectiveNoFileComp # This is on stderr +``` +Calling the `__complete` command directly allows you to run the Go debugger to troubleshoot your code. You can also add printouts to your code; Cobra provides the following functions to use for printouts in Go completion code: +```go +// Prints to the completion script debug file (if BASH_COMP_DEBUG_FILE +// is set to a file path) and optionally prints to stderr. +cobra.CompDebug(msg string, printToStdErr bool) { +cobra.CompDebugln(msg string, printToStdErr bool) + +// Prints to the completion script debug file (if BASH_COMP_DEBUG_FILE +// is set to a file path) and to stderr. +cobra.CompError(msg string) +cobra.CompErrorln(msg string) +``` +***Important:*** You should **not** leave traces that print to stdout in your completion code as they will be interpreted as completion choices by the completion script. Instead, use the cobra-provided debugging traces functions mentioned above. + +#### 2. Custom completions of nouns written in Bash + +This method allows you to inject bash functions into the completion script. Those bash functions are responsible for providing the completion choices for your own completions. Some more actual code that works in kubernetes: @@ -111,58 +253,6 @@ Find more information at https://github.com/GoogleCloudPlatform/kubernetes.`, The `BashCompletionFunction` option is really only valid/useful on the root command. Doing the above will cause `__kubectl_custom_func()` (`___custom_func()`) to be called when the built in processor was unable to find a solution. In the case of kubernetes a valid command might look something like `kubectl get pod [mypod]`. If you type `kubectl get pod [tab][tab]` the `__kubectl_customc_func()` will run because the cobra.Command only understood "kubectl" and "get." `__kubectl_custom_func()` will see that the cobra.Command is "kubectl_get" and will thus call another helper `__kubectl_get_resource()`. `__kubectl_get_resource` will look at the 'nouns' collected. In our example the only noun will be `pod`. So it will call `__kubectl_parse_get pod`. `__kubectl_parse_get` will actually call out to kubernetes and get any pods. It will then set `COMPREPLY` to valid pods! -## Have the completions code complete your 'nouns' - -In the above example "pod" was assumed to already be typed. But if you want `kubectl get [tab][tab]` to show a list of valid "nouns" you have to set them. Simplified code from `kubectl get` looks like: - -```go -validArgs []string = { "pod", "node", "service", "replicationcontroller" } - -cmd := &cobra.Command{ - Use: "get [(-o|--output=)json|yaml|template|...] (RESOURCE [NAME] | RESOURCE/NAME ...)", - Short: "Display one or many resources", - Long: get_long, - Example: get_example, - Run: func(cmd *cobra.Command, args []string) { - err := RunGet(f, out, cmd, args) - util.CheckErr(err) - }, - ValidArgs: validArgs, -} -``` - -Notice we put the "ValidArgs" on the "get" subcommand. Doing so will give results like - -```bash -# kubectl get [tab][tab] -node pod replicationcontroller service -``` - -## Plural form and shortcuts for nouns - -If your nouns have a number of aliases, you can define them alongside `ValidArgs` using `ArgAliases`: - -```go -argAliases []string = { "pods", "nodes", "services", "svc", "replicationcontrollers", "rc" } - -cmd := &cobra.Command{ - ... - ValidArgs: validArgs, - ArgAliases: argAliases -} -``` - -The aliases are not shown to the user on tab completion, but they are accepted as valid nouns by -the completion algorithm if entered manually, e.g. in: - -```bash -# kubectl get rc [tab][tab] -backend frontend database -``` - -Note that without declaring `rc` as an alias, the completion algorithm would show the list of nouns -in this example again instead of the replication controllers. - ## Mark flags as required Most of the time completions will only show subcommands. But if a flag is required to make a subcommand work, you probably want it to show up when the user types [tab][tab]. Marking a flag as 'Required' is incredibly easy. @@ -211,8 +301,45 @@ So while there are many other files in the CWD it only shows me subdirs and thos # Specify custom flag completion -Similar to the filename completion and filtering using cobra.BashCompFilenameExt, you can specify -a custom flag completion function with cobra.BashCompCustom: +As for nouns, Cobra provides two ways of defining dynamic completion of flags. Note that both these methods can be used along-side each other as long as they are not both used for the same flag. + +**Note**: *Custom Completions written in Go* will automatically work for other shell-completion scripts (e.g., Fish shell), while *Custom Completions written in Bash* will only work for Bash shell-completion. It is therefore recommended to use *Custom Completions written in Go*. + +## 1. Custom completions of flags written in Go + +To provide a Go function that Cobra will execute when it needs the list of completion choices for a flag, you must register the function in the following manner: + +```go +flagName := "output" +cmd.RegisterFlagCompletionFunc(flagName, func(cmd *cobra.Command, args []string, toComplete string) ([]string, cobra.ShellCompDirective) { + return []string{"json", "table", "yaml"}, cobra.ShellCompDirectiveDefault +}) +``` +Notice that calling `RegisterFlagCompletionFunc()` is done through the `command` with which the flag is associated. In our example this dynamic completion will give results like so: + +```bash +# helm status --output [tab][tab] +json table yaml +``` + +### Debugging + +You can also easily debug your Go completion code for flags: +```bash +# helm __complete status --output "" +json +table +yaml +:4 +Completion ended with directive: ShellCompDirectiveNoFileComp # This is on stderr +``` +***Important:*** You should **not** leave traces that print to stdout in your completion code as they will be interpreted as completion choices by the completion script. Instead, use the cobra-provided debugging traces functions mentioned in the above section. + +## 2. Custom completions of flags written in Bash + +Alternatively, you can use bash code for flag custom completion. Similar to the filename +completion and filtering using `cobra.BashCompFilenameExt`, you can specify +a custom flag completion bash function with `cobra.BashCompCustom`: ```go annotation := make(map[string][]string) @@ -226,7 +353,7 @@ a custom flag completion function with cobra.BashCompCustom: cmd.Flags().AddFlag(flag) ``` -In addition add the `__handle_namespace_flag` implementation in the `BashCompletionFunction` +In addition add the `__kubectl_get_namespaces` implementation in the `BashCompletionFunction` value, e.g.: ```bash diff --git a/vendor/github.com/spf13/cobra/command.go b/vendor/github.com/spf13/cobra/command.go index fdac9d27..88e6ed77 100644 --- a/vendor/github.com/spf13/cobra/command.go +++ b/vendor/github.com/spf13/cobra/command.go @@ -57,6 +57,10 @@ type Command struct { // ValidArgs is list of all valid non-flag arguments that are accepted in bash completions ValidArgs []string + // ValidArgsFunction is an optional function that provides valid non-flag arguments for bash completion. + // It is a dynamic version of using ValidArgs. + // Only one of ValidArgs and ValidArgsFunction can be used for a command. + ValidArgsFunction func(cmd *Command, args []string, toComplete string) ([]string, ShellCompDirective) // Expected arguments Args PositionalArgs @@ -911,6 +915,9 @@ func (c *Command) ExecuteC() (cmd *Command, err error) { args = os.Args[1:] } + // initialize the hidden command to be used for bash completion + c.initCompleteCmd(args) + var flags []string if c.TraverseChildren { cmd, flags, err = c.Traverse(args) diff --git a/vendor/github.com/spf13/cobra/custom_completions.go b/vendor/github.com/spf13/cobra/custom_completions.go new file mode 100644 index 00000000..ba57327c --- /dev/null +++ b/vendor/github.com/spf13/cobra/custom_completions.go @@ -0,0 +1,384 @@ +package cobra + +import ( + "errors" + "fmt" + "os" + "strings" + + "github.com/spf13/pflag" +) + +const ( + // ShellCompRequestCmd is the name of the hidden command that is used to request + // completion results from the program. It is used by the shell completion scripts. + ShellCompRequestCmd = "__complete" + // ShellCompNoDescRequestCmd is the name of the hidden command that is used to request + // completion results without their description. It is used by the shell completion scripts. + ShellCompNoDescRequestCmd = "__completeNoDesc" +) + +// Global map of flag completion functions. +var flagCompletionFunctions = map[*pflag.Flag]func(cmd *Command, args []string, toComplete string) ([]string, ShellCompDirective){} + +// ShellCompDirective is a bit map representing the different behaviors the shell +// can be instructed to have once completions have been provided. +type ShellCompDirective int + +const ( + // ShellCompDirectiveError indicates an error occurred and completions should be ignored. + ShellCompDirectiveError ShellCompDirective = 1 << iota + + // ShellCompDirectiveNoSpace indicates that the shell should not add a space + // after the completion even if there is a single completion provided. + ShellCompDirectiveNoSpace + + // ShellCompDirectiveNoFileComp indicates that the shell should not provide + // file completion even when no completion is provided. + // This currently does not work for zsh or bash < 4 + ShellCompDirectiveNoFileComp + + // ShellCompDirectiveDefault indicates to let the shell perform its default + // behavior after completions have been provided. + ShellCompDirectiveDefault ShellCompDirective = 0 +) + +// RegisterFlagCompletionFunc should be called to register a function to provide completion for a flag. +func (c *Command) RegisterFlagCompletionFunc(flagName string, f func(cmd *Command, args []string, toComplete string) ([]string, ShellCompDirective)) error { + flag := c.Flag(flagName) + if flag == nil { + return fmt.Errorf("RegisterFlagCompletionFunc: flag '%s' does not exist", flagName) + } + if _, exists := flagCompletionFunctions[flag]; exists { + return fmt.Errorf("RegisterFlagCompletionFunc: flag '%s' already registered", flagName) + } + flagCompletionFunctions[flag] = f + return nil +} + +// Returns a string listing the different directive enabled in the specified parameter +func (d ShellCompDirective) string() string { + var directives []string + if d&ShellCompDirectiveError != 0 { + directives = append(directives, "ShellCompDirectiveError") + } + if d&ShellCompDirectiveNoSpace != 0 { + directives = append(directives, "ShellCompDirectiveNoSpace") + } + if d&ShellCompDirectiveNoFileComp != 0 { + directives = append(directives, "ShellCompDirectiveNoFileComp") + } + if len(directives) == 0 { + directives = append(directives, "ShellCompDirectiveDefault") + } + + if d > ShellCompDirectiveError+ShellCompDirectiveNoSpace+ShellCompDirectiveNoFileComp { + return fmt.Sprintf("ERROR: unexpected ShellCompDirective value: %d", d) + } + return strings.Join(directives, ", ") +} + +// Adds a special hidden command that can be used to request custom completions. +func (c *Command) initCompleteCmd(args []string) { + completeCmd := &Command{ + Use: fmt.Sprintf("%s [command-line]", ShellCompRequestCmd), + Aliases: []string{ShellCompNoDescRequestCmd}, + DisableFlagsInUseLine: true, + Hidden: true, + DisableFlagParsing: true, + Args: MinimumNArgs(1), + Short: "Request shell completion choices for the specified command-line", + Long: fmt.Sprintf("%[2]s is a special command that is used by the shell completion logic\n%[1]s", + "to request completion choices for the specified command-line.", ShellCompRequestCmd), + Run: func(cmd *Command, args []string) { + finalCmd, completions, directive, err := cmd.getCompletions(args) + if err != nil { + CompErrorln(err.Error()) + // Keep going for multiple reasons: + // 1- There could be some valid completions even though there was an error + // 2- Even without completions, we need to print the directive + } + + noDescriptions := (cmd.CalledAs() == ShellCompNoDescRequestCmd) + for _, comp := range completions { + if noDescriptions { + // Remove any description that may be included following a tab character. + comp = strings.Split(comp, "\t")[0] + } + // Print each possible completion to stdout for the completion script to consume. + fmt.Fprintln(finalCmd.OutOrStdout(), comp) + } + + if directive > ShellCompDirectiveError+ShellCompDirectiveNoSpace+ShellCompDirectiveNoFileComp { + directive = ShellCompDirectiveDefault + } + + // As the last printout, print the completion directive for the completion script to parse. + // The directive integer must be that last character following a single colon (:). + // The completion script expects : + fmt.Fprintf(finalCmd.OutOrStdout(), ":%d\n", directive) + + // Print some helpful info to stderr for the user to understand. + // Output from stderr must be ignored by the completion script. + fmt.Fprintf(finalCmd.ErrOrStderr(), "Completion ended with directive: %s\n", directive.string()) + }, + } + c.AddCommand(completeCmd) + subCmd, _, err := c.Find(args) + if err != nil || subCmd.Name() != ShellCompRequestCmd { + // Only create this special command if it is actually being called. + // This reduces possible side-effects of creating such a command; + // for example, having this command would cause problems to a + // cobra program that only consists of the root command, since this + // command would cause the root command to suddenly have a subcommand. + c.RemoveCommand(completeCmd) + } +} + +func (c *Command) getCompletions(args []string) (*Command, []string, ShellCompDirective, error) { + var completions []string + + // The last argument, which is not completely typed by the user, + // should not be part of the list of arguments + toComplete := args[len(args)-1] + trimmedArgs := args[:len(args)-1] + + // Find the real command for which completion must be performed + finalCmd, finalArgs, err := c.Root().Find(trimmedArgs) + if err != nil { + // Unable to find the real command. E.g., someInvalidCmd + return c, completions, ShellCompDirectiveDefault, fmt.Errorf("Unable to find a command for arguments: %v", trimmedArgs) + } + + // When doing completion of a flag name, as soon as an argument starts with + // a '-' we know it is a flag. We cannot use isFlagArg() here as it requires + // the flag to be complete + if len(toComplete) > 0 && toComplete[0] == '-' && !strings.Contains(toComplete, "=") { + // We are completing a flag name + finalCmd.NonInheritedFlags().VisitAll(func(flag *pflag.Flag) { + completions = append(completions, getFlagNameCompletions(flag, toComplete)...) + }) + finalCmd.InheritedFlags().VisitAll(func(flag *pflag.Flag) { + completions = append(completions, getFlagNameCompletions(flag, toComplete)...) + }) + + directive := ShellCompDirectiveDefault + if len(completions) > 0 { + if strings.HasSuffix(completions[0], "=") { + directive = ShellCompDirectiveNoSpace + } + } + return finalCmd, completions, directive, nil + } + + var flag *pflag.Flag + if !finalCmd.DisableFlagParsing { + // We only do flag completion if we are allowed to parse flags + // This is important for commands which have requested to do their own flag completion. + flag, finalArgs, toComplete, err = checkIfFlagCompletion(finalCmd, finalArgs, toComplete) + if err != nil { + // Error while attempting to parse flags + return finalCmd, completions, ShellCompDirectiveDefault, err + } + } + + if flag == nil { + // Complete subcommand names + for _, subCmd := range finalCmd.Commands() { + if subCmd.IsAvailableCommand() && strings.HasPrefix(subCmd.Name(), toComplete) { + completions = append(completions, fmt.Sprintf("%s\t%s", subCmd.Name(), subCmd.Short)) + } + } + + if len(finalCmd.ValidArgs) > 0 { + // Always complete ValidArgs, even if we are completing a subcommand name. + // This is for commands that have both subcommands and ValidArgs. + for _, validArg := range finalCmd.ValidArgs { + if strings.HasPrefix(validArg, toComplete) { + completions = append(completions, validArg) + } + } + + // If there are ValidArgs specified (even if they don't match), we stop completion. + // Only one of ValidArgs or ValidArgsFunction can be used for a single command. + return finalCmd, completions, ShellCompDirectiveNoFileComp, nil + } + + // Always let the logic continue so as to add any ValidArgsFunction completions, + // even if we already found sub-commands. + // This is for commands that have subcommands but also specify a ValidArgsFunction. + } + + // Parse the flags and extract the arguments to prepare for calling the completion function + if err = finalCmd.ParseFlags(finalArgs); err != nil { + return finalCmd, completions, ShellCompDirectiveDefault, fmt.Errorf("Error while parsing flags from args %v: %s", finalArgs, err.Error()) + } + + // We only remove the flags from the arguments if DisableFlagParsing is not set. + // This is important for commands which have requested to do their own flag completion. + if !finalCmd.DisableFlagParsing { + finalArgs = finalCmd.Flags().Args() + } + + // Find the completion function for the flag or command + var completionFn func(cmd *Command, args []string, toComplete string) ([]string, ShellCompDirective) + if flag != nil { + completionFn = flagCompletionFunctions[flag] + } else { + completionFn = finalCmd.ValidArgsFunction + } + if completionFn == nil { + // Go custom completion not supported/needed for this flag or command + return finalCmd, completions, ShellCompDirectiveDefault, nil + } + + // Call the registered completion function to get the completions + comps, directive := completionFn(finalCmd, finalArgs, toComplete) + completions = append(completions, comps...) + return finalCmd, completions, directive, nil +} + +func getFlagNameCompletions(flag *pflag.Flag, toComplete string) []string { + if nonCompletableFlag(flag) { + return []string{} + } + + var completions []string + flagName := "--" + flag.Name + if strings.HasPrefix(flagName, toComplete) { + // Flag without the = + completions = append(completions, fmt.Sprintf("%s\t%s", flagName, flag.Usage)) + + if len(flag.NoOptDefVal) == 0 { + // Flag requires a value, so it can be suffixed with = + flagName += "=" + completions = append(completions, fmt.Sprintf("%s\t%s", flagName, flag.Usage)) + } + } + + flagName = "-" + flag.Shorthand + if len(flag.Shorthand) > 0 && strings.HasPrefix(flagName, toComplete) { + completions = append(completions, fmt.Sprintf("%s\t%s", flagName, flag.Usage)) + } + + return completions +} + +func checkIfFlagCompletion(finalCmd *Command, args []string, lastArg string) (*pflag.Flag, []string, string, error) { + var flagName string + trimmedArgs := args + flagWithEqual := false + if isFlagArg(lastArg) { + if index := strings.Index(lastArg, "="); index >= 0 { + flagName = strings.TrimLeft(lastArg[:index], "-") + lastArg = lastArg[index+1:] + flagWithEqual = true + } else { + return nil, nil, "", errors.New("Unexpected completion request for flag") + } + } + + if len(flagName) == 0 { + if len(args) > 0 { + prevArg := args[len(args)-1] + if isFlagArg(prevArg) { + // Only consider the case where the flag does not contain an =. + // If the flag contains an = it means it has already been fully processed, + // so we don't need to deal with it here. + if index := strings.Index(prevArg, "="); index < 0 { + flagName = strings.TrimLeft(prevArg, "-") + + // Remove the uncompleted flag or else there could be an error created + // for an invalid value for that flag + trimmedArgs = args[:len(args)-1] + } + } + } + } + + if len(flagName) == 0 { + // Not doing flag completion + return nil, trimmedArgs, lastArg, nil + } + + flag := findFlag(finalCmd, flagName) + if flag == nil { + // Flag not supported by this command, nothing to complete + err := fmt.Errorf("Subcommand '%s' does not support flag '%s'", finalCmd.Name(), flagName) + return nil, nil, "", err + } + + if !flagWithEqual { + if len(flag.NoOptDefVal) != 0 { + // We had assumed dealing with a two-word flag but the flag is a boolean flag. + // In that case, there is no value following it, so we are not really doing flag completion. + // Reset everything to do noun completion. + trimmedArgs = args + flag = nil + } + } + + return flag, trimmedArgs, lastArg, nil +} + +func findFlag(cmd *Command, name string) *pflag.Flag { + flagSet := cmd.Flags() + if len(name) == 1 { + // First convert the short flag into a long flag + // as the cmd.Flag() search only accepts long flags + if short := flagSet.ShorthandLookup(name); short != nil { + name = short.Name + } else { + set := cmd.InheritedFlags() + if short = set.ShorthandLookup(name); short != nil { + name = short.Name + } else { + return nil + } + } + } + return cmd.Flag(name) +} + +// CompDebug prints the specified string to the same file as where the +// completion script prints its logs. +// Note that completion printouts should never be on stdout as they would +// be wrongly interpreted as actual completion choices by the completion script. +func CompDebug(msg string, printToStdErr bool) { + msg = fmt.Sprintf("[Debug] %s", msg) + + // Such logs are only printed when the user has set the environment + // variable BASH_COMP_DEBUG_FILE to the path of some file to be used. + if path := os.Getenv("BASH_COMP_DEBUG_FILE"); path != "" { + f, err := os.OpenFile(path, + os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0644) + if err == nil { + defer f.Close() + f.WriteString(msg) + } + } + + if printToStdErr { + // Must print to stderr for this not to be read by the completion script. + fmt.Fprintf(os.Stderr, msg) + } +} + +// CompDebugln prints the specified string with a newline at the end +// to the same file as where the completion script prints its logs. +// Such logs are only printed when the user has set the environment +// variable BASH_COMP_DEBUG_FILE to the path of some file to be used. +func CompDebugln(msg string, printToStdErr bool) { + CompDebug(fmt.Sprintf("%s\n", msg), printToStdErr) +} + +// CompError prints the specified completion message to stderr. +func CompError(msg string) { + msg = fmt.Sprintf("[Error] %s", msg) + CompDebug(msg, true) +} + +// CompErrorln prints the specified completion message to stderr with a newline at the end. +func CompErrorln(msg string) { + CompError(fmt.Sprintf("%s\n", msg)) +} diff --git a/vendor/github.com/spf13/cobra/fish_completions.go b/vendor/github.com/spf13/cobra/fish_completions.go new file mode 100644 index 00000000..c83609c8 --- /dev/null +++ b/vendor/github.com/spf13/cobra/fish_completions.go @@ -0,0 +1,172 @@ +package cobra + +import ( + "bytes" + "fmt" + "io" + "os" +) + +func genFishComp(buf *bytes.Buffer, name string, includeDesc bool) { + compCmd := ShellCompRequestCmd + if !includeDesc { + compCmd = ShellCompNoDescRequestCmd + } + buf.WriteString(fmt.Sprintf("# fish completion for %-36s -*- shell-script -*-\n", name)) + buf.WriteString(fmt.Sprintf(` +function __%[1]s_debug + set file "$BASH_COMP_DEBUG_FILE" + if test -n "$file" + echo "$argv" >> $file + end +end + +function __%[1]s_perform_completion + __%[1]s_debug "Starting __%[1]s_perform_completion with: $argv" + + set args (string split -- " " "$argv") + set lastArg "$args[-1]" + + __%[1]s_debug "args: $args" + __%[1]s_debug "last arg: $lastArg" + + set emptyArg "" + if test -z "$lastArg" + __%[1]s_debug "Setting emptyArg" + set emptyArg \"\" + end + __%[1]s_debug "emptyArg: $emptyArg" + + set requestComp "$args[1] %[2]s $args[2..-1] $emptyArg" + __%[1]s_debug "Calling $requestComp" + + set results (eval $requestComp 2> /dev/null) + set comps $results[1..-2] + set directiveLine $results[-1] + + # For Fish, when completing a flag with an = (e.g., -n=) + # completions must be prefixed with the flag + set flagPrefix (string match -r -- '-.*=' "$lastArg") + + __%[1]s_debug "Comps: $comps" + __%[1]s_debug "DirectiveLine: $directiveLine" + __%[1]s_debug "flagPrefix: $flagPrefix" + + for comp in $comps + printf "%%s%%s\n" "$flagPrefix" "$comp" + end + + printf "%%s\n" "$directiveLine" +end + +# This function does three things: +# 1- Obtain the completions and store them in the global __%[1]s_comp_results +# 2- Set the __%[1]s_comp_do_file_comp flag if file completion should be performed +# and unset it otherwise +# 3- Return true if the completion results are not empty +function __%[1]s_prepare_completions + # Start fresh + set --erase __%[1]s_comp_do_file_comp + set --erase __%[1]s_comp_results + + # Check if the command-line is already provided. This is useful for testing. + if not set --query __%[1]s_comp_commandLine + set __%[1]s_comp_commandLine (commandline) + end + __%[1]s_debug "commandLine is: $__%[1]s_comp_commandLine" + + set results (__%[1]s_perform_completion "$__%[1]s_comp_commandLine") + set --erase __%[1]s_comp_commandLine + __%[1]s_debug "Completion results: $results" + + if test -z "$results" + __%[1]s_debug "No completion, probably due to a failure" + # Might as well do file completion, in case it helps + set --global __%[1]s_comp_do_file_comp 1 + return 0 + end + + set directive (string sub --start 2 $results[-1]) + set --global __%[1]s_comp_results $results[1..-2] + + __%[1]s_debug "Completions are: $__%[1]s_comp_results" + __%[1]s_debug "Directive is: $directive" + + if test -z "$directive" + set directive 0 + end + + set compErr (math (math --scale 0 $directive / %[3]d) %% 2) + if test $compErr -eq 1 + __%[1]s_debug "Received error directive: aborting." + # Might as well do file completion, in case it helps + set --global __%[1]s_comp_do_file_comp 1 + return 0 + end + + set nospace (math (math --scale 0 $directive / %[4]d) %% 2) + set nofiles (math (math --scale 0 $directive / %[5]d) %% 2) + + __%[1]s_debug "nospace: $nospace, nofiles: $nofiles" + + # Important not to quote the variable for count to work + set numComps (count $__%[1]s_comp_results) + __%[1]s_debug "numComps: $numComps" + + if test $numComps -eq 1; and test $nospace -ne 0 + # To support the "nospace" directive we trick the shell + # by outputting an extra, longer completion. + __%[1]s_debug "Adding second completion to perform nospace directive" + set --append __%[1]s_comp_results $__%[1]s_comp_results[1]. + end + + if test $numComps -eq 0; and test $nofiles -eq 0 + __%[1]s_debug "Requesting file completion" + set --global __%[1]s_comp_do_file_comp 1 + end + + # If we don't want file completion, we must return true even if there + # are no completions found. This is because fish will perform the last + # completion command, even if its condition is false, if no other + # completion command was triggered + return (not set --query __%[1]s_comp_do_file_comp) +end + +# Remove any pre-existing completions for the program since we will be handling all of them +# TODO this cleanup is not sufficient. Fish completions are only loaded once the user triggers +# them, so the below deletion will not work as it is run too early. What else can we do? +complete -c %[1]s -e + +# The order in which the below two lines are defined is very important so that __%[1]s_prepare_completions +# is called first. It is __%[1]s_prepare_completions that sets up the __%[1]s_comp_do_file_comp variable. +# +# This completion will be run second as complete commands are added FILO. +# It triggers file completion choices when __%[1]s_comp_do_file_comp is set. +complete -c %[1]s -n 'set --query __%[1]s_comp_do_file_comp' + +# This completion will be run first as complete commands are added FILO. +# The call to __%[1]s_prepare_completions will setup both __%[1]s_comp_results abd __%[1]s_comp_do_file_comp. +# It provides the program's completion choices. +complete -c %[1]s -n '__%[1]s_prepare_completions' -f -a '$__%[1]s_comp_results' + +`, name, compCmd, ShellCompDirectiveError, ShellCompDirectiveNoSpace, ShellCompDirectiveNoFileComp)) +} + +// GenFishCompletion generates fish completion file and writes to the passed writer. +func (c *Command) GenFishCompletion(w io.Writer, includeDesc bool) error { + buf := new(bytes.Buffer) + genFishComp(buf, c.Name(), includeDesc) + _, err := buf.WriteTo(w) + return err +} + +// GenFishCompletionFile generates fish completion file. +func (c *Command) GenFishCompletionFile(filename string, includeDesc bool) error { + outFile, err := os.Create(filename) + if err != nil { + return err + } + defer outFile.Close() + + return c.GenFishCompletion(outFile, includeDesc) +} diff --git a/vendor/github.com/spf13/cobra/fish_completions.md b/vendor/github.com/spf13/cobra/fish_completions.md new file mode 100644 index 00000000..6bfe5f88 --- /dev/null +++ b/vendor/github.com/spf13/cobra/fish_completions.md @@ -0,0 +1,7 @@ +## Generating Fish Completions for your own cobra.Command + +Cobra supports native Fish completions generated from the root `cobra.Command`. You can use the `command.GenFishCompletion()` or `command.GenFishCompletionFile()` functions. You must provide these functions with a parameter indicating if the completions should be annotated with a description; Cobra will provide the description automatically based on usage information. You can choose to make this option configurable by your users. + +### Limitations + +* Custom completions implemented using the `ValidArgsFunction` and `RegisterFlagCompletionFunc()` are supported automatically but the ones implemented in Bash scripting are not. diff --git a/vendor/github.com/stretchr/testify/assert/assertion_format.go b/vendor/github.com/stretchr/testify/assert/assertion_format.go index e0364e9e..bf89ecd2 100644 --- a/vendor/github.com/stretchr/testify/assert/assertion_format.go +++ b/vendor/github.com/stretchr/testify/assert/assertion_format.go @@ -32,7 +32,8 @@ func Containsf(t TestingT, s interface{}, contains interface{}, msg string, args return Contains(t, s, contains, append([]interface{}{msg}, args...)...) } -// DirExistsf checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// DirExistsf checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func DirExistsf(t TestingT, path string, msg string, args ...interface{}) bool { if h, ok := t.(tHelper); ok { h.Helper() @@ -160,7 +161,8 @@ func Falsef(t TestingT, value bool, msg string, args ...interface{}) bool { return False(t, value, append([]interface{}{msg}, args...)...) } -// FileExistsf checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExistsf checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func FileExistsf(t TestingT, path string, msg string, args ...interface{}) bool { if h, ok := t.(tHelper); ok { h.Helper() @@ -267,7 +269,7 @@ func Implementsf(t TestingT, interfaceObject interface{}, object interface{}, ms // InDeltaf asserts that the two numerals are within delta of each other. // -// assert.InDeltaf(t, math.Pi, (22 / 7.0, "error message %s", "formatted"), 0.01) +// assert.InDeltaf(t, math.Pi, 22/7.0, 0.01, "error message %s", "formatted") func InDeltaf(t TestingT, expected interface{}, actual interface{}, delta float64, msg string, args ...interface{}) bool { if h, ok := t.(tHelper); ok { h.Helper() @@ -325,14 +327,6 @@ func JSONEqf(t TestingT, expected string, actual string, msg string, args ...int return JSONEq(t, expected, actual, append([]interface{}{msg}, args...)...) } -// YAMLEqf asserts that two YAML strings are equivalent. -func YAMLEqf(t TestingT, expected string, actual string, msg string, args ...interface{}) bool { - if h, ok := t.(tHelper); ok { - h.Helper() - } - return YAMLEq(t, expected, actual, append([]interface{}{msg}, args...)...) -} - // Lenf asserts that the specified object has specific length. // Lenf also fails if the object has a type that len() not accept. // @@ -369,6 +363,17 @@ func LessOrEqualf(t TestingT, e1 interface{}, e2 interface{}, msg string, args . return LessOrEqual(t, e1, e2, append([]interface{}{msg}, args...)...) } +// Neverf asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// assert.Neverf(t, func() bool { return false; }, time.Second, 10*time.Millisecond, "error message %s", "formatted") +func Neverf(t TestingT, condition func() bool, waitFor time.Duration, tick time.Duration, msg string, args ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + return Never(t, condition, waitFor, tick, append([]interface{}{msg}, args...)...) +} + // Nilf asserts that the specified object is nil. // // assert.Nilf(t, err, "error message %s", "formatted") @@ -379,6 +384,15 @@ func Nilf(t TestingT, object interface{}, msg string, args ...interface{}) bool return Nil(t, object, append([]interface{}{msg}, args...)...) } +// NoDirExistsf checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func NoDirExistsf(t TestingT, path string, msg string, args ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + return NoDirExists(t, path, append([]interface{}{msg}, args...)...) +} + // NoErrorf asserts that a function returned no error (i.e. `nil`). // // actualObj, err := SomeFunction() @@ -392,6 +406,15 @@ func NoErrorf(t TestingT, err error, msg string, args ...interface{}) bool { return NoError(t, err, append([]interface{}{msg}, args...)...) } +// NoFileExistsf checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func NoFileExistsf(t TestingT, path string, msg string, args ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + return NoFileExists(t, path, append([]interface{}{msg}, args...)...) +} + // NotContainsf asserts that the specified string, list(array, slice...) or map does NOT contain the // specified substring or element. // @@ -462,6 +485,19 @@ func NotRegexpf(t TestingT, rx interface{}, str interface{}, msg string, args .. return NotRegexp(t, rx, str, append([]interface{}{msg}, args...)...) } +// NotSamef asserts that two pointers do not reference the same object. +// +// assert.NotSamef(t, ptr1, ptr2, "error message %s", "formatted") +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func NotSamef(t TestingT, expected interface{}, actual interface{}, msg string, args ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + return NotSame(t, expected, actual, append([]interface{}{msg}, args...)...) +} + // NotSubsetf asserts that the specified list(array, slice...) contains not all // elements given in the specified subset(array, slice...). // @@ -491,6 +527,18 @@ func Panicsf(t TestingT, f PanicTestFunc, msg string, args ...interface{}) bool return Panics(t, f, append([]interface{}{msg}, args...)...) } +// PanicsWithErrorf asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// assert.PanicsWithErrorf(t, "crazy error", func(){ GoCrazy() }, "error message %s", "formatted") +func PanicsWithErrorf(t TestingT, errString string, f PanicTestFunc, msg string, args ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + return PanicsWithError(t, errString, f, append([]interface{}{msg}, args...)...) +} + // PanicsWithValuef asserts that the code inside the specified PanicTestFunc panics, and that // the recovered panic value equals the expected panic value. // @@ -557,6 +605,14 @@ func WithinDurationf(t TestingT, expected time.Time, actual time.Time, delta tim return WithinDuration(t, expected, actual, delta, append([]interface{}{msg}, args...)...) } +// YAMLEqf asserts that two YAML strings are equivalent. +func YAMLEqf(t TestingT, expected string, actual string, msg string, args ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + return YAMLEq(t, expected, actual, append([]interface{}{msg}, args...)...) +} + // Zerof asserts that i is the zero value for its type. func Zerof(t TestingT, i interface{}, msg string, args ...interface{}) bool { if h, ok := t.(tHelper); ok { diff --git a/vendor/github.com/stretchr/testify/assert/assertion_forward.go b/vendor/github.com/stretchr/testify/assert/assertion_forward.go index 26830403..75ecdcaa 100644 --- a/vendor/github.com/stretchr/testify/assert/assertion_forward.go +++ b/vendor/github.com/stretchr/testify/assert/assertion_forward.go @@ -53,7 +53,8 @@ func (a *Assertions) Containsf(s interface{}, contains interface{}, msg string, return Containsf(a.t, s, contains, msg, args...) } -// DirExists checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// DirExists checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func (a *Assertions) DirExists(path string, msgAndArgs ...interface{}) bool { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -61,7 +62,8 @@ func (a *Assertions) DirExists(path string, msgAndArgs ...interface{}) bool { return DirExists(a.t, path, msgAndArgs...) } -// DirExistsf checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// DirExistsf checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func (a *Assertions) DirExistsf(path string, msg string, args ...interface{}) bool { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -309,7 +311,8 @@ func (a *Assertions) Falsef(value bool, msg string, args ...interface{}) bool { return Falsef(a.t, value, msg, args...) } -// FileExists checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExists checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func (a *Assertions) FileExists(path string, msgAndArgs ...interface{}) bool { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -317,7 +320,8 @@ func (a *Assertions) FileExists(path string, msgAndArgs ...interface{}) bool { return FileExists(a.t, path, msgAndArgs...) } -// FileExistsf checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExistsf checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func (a *Assertions) FileExistsf(path string, msg string, args ...interface{}) bool { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -521,7 +525,7 @@ func (a *Assertions) Implementsf(interfaceObject interface{}, object interface{} // InDelta asserts that the two numerals are within delta of each other. // -// a.InDelta(math.Pi, (22 / 7.0), 0.01) +// a.InDelta(math.Pi, 22/7.0, 0.01) func (a *Assertions) InDelta(expected interface{}, actual interface{}, delta float64, msgAndArgs ...interface{}) bool { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -563,7 +567,7 @@ func (a *Assertions) InDeltaSlicef(expected interface{}, actual interface{}, del // InDeltaf asserts that the two numerals are within delta of each other. // -// a.InDeltaf(math.Pi, (22 / 7.0, "error message %s", "formatted"), 0.01) +// a.InDeltaf(math.Pi, 22/7.0, 0.01, "error message %s", "formatted") func (a *Assertions) InDeltaf(expected interface{}, actual interface{}, delta float64, msg string, args ...interface{}) bool { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -639,22 +643,6 @@ func (a *Assertions) JSONEqf(expected string, actual string, msg string, args .. return JSONEqf(a.t, expected, actual, msg, args...) } -// YAMLEq asserts that two YAML strings are equivalent. -func (a *Assertions) YAMLEq(expected string, actual string, msgAndArgs ...interface{}) bool { - if h, ok := a.t.(tHelper); ok { - h.Helper() - } - return YAMLEq(a.t, expected, actual, msgAndArgs...) -} - -// YAMLEqf asserts that two YAML strings are equivalent. -func (a *Assertions) YAMLEqf(expected string, actual string, msg string, args ...interface{}) bool { - if h, ok := a.t.(tHelper); ok { - h.Helper() - } - return YAMLEqf(a.t, expected, actual, msg, args...) -} - // Len asserts that the specified object has specific length. // Len also fails if the object has a type that len() not accept. // @@ -727,6 +715,28 @@ func (a *Assertions) Lessf(e1 interface{}, e2 interface{}, msg string, args ...i return Lessf(a.t, e1, e2, msg, args...) } +// Never asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// a.Never(func() bool { return false; }, time.Second, 10*time.Millisecond) +func (a *Assertions) Never(condition func() bool, waitFor time.Duration, tick time.Duration, msgAndArgs ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return Never(a.t, condition, waitFor, tick, msgAndArgs...) +} + +// Neverf asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// a.Neverf(func() bool { return false; }, time.Second, 10*time.Millisecond, "error message %s", "formatted") +func (a *Assertions) Neverf(condition func() bool, waitFor time.Duration, tick time.Duration, msg string, args ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return Neverf(a.t, condition, waitFor, tick, msg, args...) +} + // Nil asserts that the specified object is nil. // // a.Nil(err) @@ -747,6 +757,24 @@ func (a *Assertions) Nilf(object interface{}, msg string, args ...interface{}) b return Nilf(a.t, object, msg, args...) } +// NoDirExists checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func (a *Assertions) NoDirExists(path string, msgAndArgs ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return NoDirExists(a.t, path, msgAndArgs...) +} + +// NoDirExistsf checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func (a *Assertions) NoDirExistsf(path string, msg string, args ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return NoDirExistsf(a.t, path, msg, args...) +} + // NoError asserts that a function returned no error (i.e. `nil`). // // actualObj, err := SomeFunction() @@ -773,6 +801,24 @@ func (a *Assertions) NoErrorf(err error, msg string, args ...interface{}) bool { return NoErrorf(a.t, err, msg, args...) } +// NoFileExists checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func (a *Assertions) NoFileExists(path string, msgAndArgs ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return NoFileExists(a.t, path, msgAndArgs...) +} + +// NoFileExistsf checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func (a *Assertions) NoFileExistsf(path string, msg string, args ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return NoFileExistsf(a.t, path, msg, args...) +} + // NotContains asserts that the specified string, list(array, slice...) or map does NOT contain the // specified substring or element. // @@ -913,6 +959,32 @@ func (a *Assertions) NotRegexpf(rx interface{}, str interface{}, msg string, arg return NotRegexpf(a.t, rx, str, msg, args...) } +// NotSame asserts that two pointers do not reference the same object. +// +// a.NotSame(ptr1, ptr2) +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func (a *Assertions) NotSame(expected interface{}, actual interface{}, msgAndArgs ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return NotSame(a.t, expected, actual, msgAndArgs...) +} + +// NotSamef asserts that two pointers do not reference the same object. +// +// a.NotSamef(ptr1, ptr2, "error message %s", "formatted") +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func (a *Assertions) NotSamef(expected interface{}, actual interface{}, msg string, args ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return NotSamef(a.t, expected, actual, msg, args...) +} + // NotSubset asserts that the specified list(array, slice...) contains not all // elements given in the specified subset(array, slice...). // @@ -961,6 +1033,30 @@ func (a *Assertions) Panics(f PanicTestFunc, msgAndArgs ...interface{}) bool { return Panics(a.t, f, msgAndArgs...) } +// PanicsWithError asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// a.PanicsWithError("crazy error", func(){ GoCrazy() }) +func (a *Assertions) PanicsWithError(errString string, f PanicTestFunc, msgAndArgs ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return PanicsWithError(a.t, errString, f, msgAndArgs...) +} + +// PanicsWithErrorf asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// a.PanicsWithErrorf("crazy error", func(){ GoCrazy() }, "error message %s", "formatted") +func (a *Assertions) PanicsWithErrorf(errString string, f PanicTestFunc, msg string, args ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return PanicsWithErrorf(a.t, errString, f, msg, args...) +} + // PanicsWithValue asserts that the code inside the specified PanicTestFunc panics, and that // the recovered panic value equals the expected panic value. // @@ -1103,6 +1199,22 @@ func (a *Assertions) WithinDurationf(expected time.Time, actual time.Time, delta return WithinDurationf(a.t, expected, actual, delta, msg, args...) } +// YAMLEq asserts that two YAML strings are equivalent. +func (a *Assertions) YAMLEq(expected string, actual string, msgAndArgs ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return YAMLEq(a.t, expected, actual, msgAndArgs...) +} + +// YAMLEqf asserts that two YAML strings are equivalent. +func (a *Assertions) YAMLEqf(expected string, actual string, msg string, args ...interface{}) bool { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + return YAMLEqf(a.t, expected, actual, msg, args...) +} + // Zero asserts that i is the zero value for its type. func (a *Assertions) Zero(i interface{}, msgAndArgs ...interface{}) bool { if h, ok := a.t.(tHelper); ok { diff --git a/vendor/github.com/stretchr/testify/assert/assertions.go b/vendor/github.com/stretchr/testify/assert/assertions.go index 044da8b0..bdd81389 100644 --- a/vendor/github.com/stretchr/testify/assert/assertions.go +++ b/vendor/github.com/stretchr/testify/assert/assertions.go @@ -11,6 +11,7 @@ import ( "reflect" "regexp" "runtime" + "runtime/debug" "strings" "time" "unicode" @@ -21,7 +22,7 @@ import ( yaml "gopkg.in/yaml.v2" ) -//go:generate go run ../_codegen/main.go -output-package=assert -template=assertion_format.go.tmpl +//go:generate sh -c "cd ../_codegen && go build && cd - && ../_codegen/_codegen -output-package=assert -template=assertion_format.go.tmpl" // TestingT is an interface wrapper around *testing.T type TestingT interface { @@ -351,6 +352,19 @@ func Equal(t TestingT, expected, actual interface{}, msgAndArgs ...interface{}) } +// validateEqualArgs checks whether provided arguments can be safely used in the +// Equal/NotEqual functions. +func validateEqualArgs(expected, actual interface{}) error { + if expected == nil && actual == nil { + return nil + } + + if isFunction(expected) || isFunction(actual) { + return errors.New("cannot take func type as argument") + } + return nil +} + // Same asserts that two pointers reference the same object. // // assert.Same(t, ptr1, ptr2) @@ -362,18 +376,7 @@ func Same(t TestingT, expected, actual interface{}, msgAndArgs ...interface{}) b h.Helper() } - expectedPtr, actualPtr := reflect.ValueOf(expected), reflect.ValueOf(actual) - if expectedPtr.Kind() != reflect.Ptr || actualPtr.Kind() != reflect.Ptr { - return Fail(t, "Invalid operation: both arguments must be pointers", msgAndArgs...) - } - - expectedType, actualType := reflect.TypeOf(expected), reflect.TypeOf(actual) - if expectedType != actualType { - return Fail(t, fmt.Sprintf("Pointer expected to be of type %v, but was %v", - expectedType, actualType), msgAndArgs...) - } - - if expected != actual { + if !samePointers(expected, actual) { return Fail(t, fmt.Sprintf("Not same: \n"+ "expected: %p %#v\n"+ "actual : %p %#v", expected, expected, actual, actual), msgAndArgs...) @@ -382,6 +385,42 @@ func Same(t TestingT, expected, actual interface{}, msgAndArgs ...interface{}) b return true } +// NotSame asserts that two pointers do not reference the same object. +// +// assert.NotSame(t, ptr1, ptr2) +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func NotSame(t TestingT, expected, actual interface{}, msgAndArgs ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + + if samePointers(expected, actual) { + return Fail(t, fmt.Sprintf( + "Expected and actual point to the same object: %p %#v", + expected, expected), msgAndArgs...) + } + return true +} + +// samePointers compares two generic interface objects and returns whether +// they point to the same object +func samePointers(first, second interface{}) bool { + firstPtr, secondPtr := reflect.ValueOf(first), reflect.ValueOf(second) + if firstPtr.Kind() != reflect.Ptr || secondPtr.Kind() != reflect.Ptr { + return false + } + + firstType, secondType := reflect.TypeOf(first), reflect.TypeOf(second) + if firstType != secondType { + return false + } + + // compare pointer addresses + return first == second +} + // formatUnequalValues takes two values of arbitrary types and returns string // representations appropriate to be presented to the user. // @@ -393,9 +432,11 @@ func formatUnequalValues(expected, actual interface{}) (e string, a string) { return fmt.Sprintf("%T(%#v)", expected, expected), fmt.Sprintf("%T(%#v)", actual, actual) } - - return fmt.Sprintf("%#v", expected), - fmt.Sprintf("%#v", actual) + switch expected.(type) { + case time.Duration: + return fmt.Sprintf("%v", expected), fmt.Sprintf("%v", actual) + } + return fmt.Sprintf("%#v", expected), fmt.Sprintf("%#v", actual) } // EqualValues asserts that two objects are equal or convertable to the same types @@ -901,15 +942,17 @@ func Condition(t TestingT, comp Comparison, msgAndArgs ...interface{}) bool { type PanicTestFunc func() // didPanic returns true if the function passed to it panics. Otherwise, it returns false. -func didPanic(f PanicTestFunc) (bool, interface{}) { +func didPanic(f PanicTestFunc) (bool, interface{}, string) { didPanic := false var message interface{} + var stack string func() { defer func() { if message = recover(); message != nil { didPanic = true + stack = string(debug.Stack()) } }() @@ -918,7 +961,7 @@ func didPanic(f PanicTestFunc) (bool, interface{}) { }() - return didPanic, message + return didPanic, message, stack } @@ -930,7 +973,7 @@ func Panics(t TestingT, f PanicTestFunc, msgAndArgs ...interface{}) bool { h.Helper() } - if funcDidPanic, panicValue := didPanic(f); !funcDidPanic { + if funcDidPanic, panicValue, _ := didPanic(f); !funcDidPanic { return Fail(t, fmt.Sprintf("func %#v should panic\n\tPanic value:\t%#v", f, panicValue), msgAndArgs...) } @@ -946,12 +989,34 @@ func PanicsWithValue(t TestingT, expected interface{}, f PanicTestFunc, msgAndAr h.Helper() } - funcDidPanic, panicValue := didPanic(f) + funcDidPanic, panicValue, panickedStack := didPanic(f) if !funcDidPanic { return Fail(t, fmt.Sprintf("func %#v should panic\n\tPanic value:\t%#v", f, panicValue), msgAndArgs...) } if panicValue != expected { - return Fail(t, fmt.Sprintf("func %#v should panic with value:\t%#v\n\tPanic value:\t%#v", f, expected, panicValue), msgAndArgs...) + return Fail(t, fmt.Sprintf("func %#v should panic with value:\t%#v\n\tPanic value:\t%#v\n\tPanic stack:\t%s", f, expected, panicValue, panickedStack), msgAndArgs...) + } + + return true +} + +// PanicsWithError asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// assert.PanicsWithError(t, "crazy error", func(){ GoCrazy() }) +func PanicsWithError(t TestingT, errString string, f PanicTestFunc, msgAndArgs ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + + funcDidPanic, panicValue, panickedStack := didPanic(f) + if !funcDidPanic { + return Fail(t, fmt.Sprintf("func %#v should panic\n\tPanic value:\t%#v", f, panicValue), msgAndArgs...) + } + panicErr, ok := panicValue.(error) + if !ok || panicErr.Error() != errString { + return Fail(t, fmt.Sprintf("func %#v should panic with error message:\t%#v\n\tPanic value:\t%#v\n\tPanic stack:\t%s", f, errString, panicValue, panickedStack), msgAndArgs...) } return true @@ -965,8 +1030,8 @@ func NotPanics(t TestingT, f PanicTestFunc, msgAndArgs ...interface{}) bool { h.Helper() } - if funcDidPanic, panicValue := didPanic(f); funcDidPanic { - return Fail(t, fmt.Sprintf("func %#v should not panic\n\tPanic value:\t%v", f, panicValue), msgAndArgs...) + if funcDidPanic, panicValue, panickedStack := didPanic(f); funcDidPanic { + return Fail(t, fmt.Sprintf("func %#v should not panic\n\tPanic value:\t%v\n\tPanic stack:\t%s", f, panicValue, panickedStack), msgAndArgs...) } return true @@ -1026,7 +1091,7 @@ func toFloat(x interface{}) (float64, bool) { // InDelta asserts that the two numerals are within delta of each other. // -// assert.InDelta(t, math.Pi, (22 / 7.0), 0.01) +// assert.InDelta(t, math.Pi, 22/7.0, 0.01) func InDelta(t TestingT, expected, actual interface{}, delta float64, msgAndArgs ...interface{}) bool { if h, ok := t.(tHelper); ok { h.Helper() @@ -1314,7 +1379,8 @@ func NotZero(t TestingT, i interface{}, msgAndArgs ...interface{}) bool { return true } -// FileExists checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExists checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func FileExists(t TestingT, path string, msgAndArgs ...interface{}) bool { if h, ok := t.(tHelper); ok { h.Helper() @@ -1332,7 +1398,24 @@ func FileExists(t TestingT, path string, msgAndArgs ...interface{}) bool { return true } -// DirExists checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// NoFileExists checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func NoFileExists(t TestingT, path string, msgAndArgs ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + info, err := os.Lstat(path) + if err != nil { + return true + } + if info.IsDir() { + return true + } + return Fail(t, fmt.Sprintf("file %q exists", path), msgAndArgs...) +} + +// DirExists checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func DirExists(t TestingT, path string, msgAndArgs ...interface{}) bool { if h, ok := t.(tHelper); ok { h.Helper() @@ -1350,6 +1433,25 @@ func DirExists(t TestingT, path string, msgAndArgs ...interface{}) bool { return true } +// NoDirExists checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func NoDirExists(t TestingT, path string, msgAndArgs ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + info, err := os.Lstat(path) + if err != nil { + if os.IsNotExist(err) { + return true + } + return true + } + if !info.IsDir() { + return true + } + return Fail(t, fmt.Sprintf("directory %q exists", path), msgAndArgs...) +} + // JSONEq asserts that two JSON strings are equivalent. // // assert.JSONEq(t, `{"hello": "world", "foo": "bar"}`, `{"foo": "bar", "hello": "world"}`) @@ -1439,15 +1541,6 @@ func diff(expected interface{}, actual interface{}) string { return "\n\nDiff:\n" + diff } -// validateEqualArgs checks whether provided arguments can be safely used in the -// Equal/NotEqual functions. -func validateEqualArgs(expected, actual interface{}) error { - if isFunction(expected) || isFunction(actual) { - return errors.New("cannot take func type as argument") - } - return nil -} - func isFunction(arg interface{}) bool { if arg == nil { return false @@ -1475,24 +1568,59 @@ func Eventually(t TestingT, condition func() bool, waitFor time.Duration, tick t h.Helper() } + ch := make(chan bool, 1) + timer := time.NewTimer(waitFor) - ticker := time.NewTicker(tick) - checkPassed := make(chan bool) defer timer.Stop() + + ticker := time.NewTicker(tick) defer ticker.Stop() - defer close(checkPassed) - for { + + for tick := ticker.C; ; { select { case <-timer.C: return Fail(t, "Condition never satisfied", msgAndArgs...) - case result := <-checkPassed: - if result { + case <-tick: + tick = nil + go func() { ch <- condition() }() + case v := <-ch: + if v { return true } - case <-ticker.C: - go func() { - checkPassed <- condition() - }() + tick = ticker.C + } + } +} + +// Never asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// assert.Never(t, func() bool { return false; }, time.Second, 10*time.Millisecond) +func Never(t TestingT, condition func() bool, waitFor time.Duration, tick time.Duration, msgAndArgs ...interface{}) bool { + if h, ok := t.(tHelper); ok { + h.Helper() + } + + ch := make(chan bool, 1) + + timer := time.NewTimer(waitFor) + defer timer.Stop() + + ticker := time.NewTicker(tick) + defer ticker.Stop() + + for tick := ticker.C; ; { + select { + case <-timer.C: + return true + case <-tick: + tick = nil + go func() { ch <- condition() }() + case v := <-ch: + if v { + return Fail(t, "Condition satisfied", msgAndArgs...) + } + tick = ticker.C } } } diff --git a/vendor/github.com/stretchr/testify/assert/forward_assertions.go b/vendor/github.com/stretchr/testify/assert/forward_assertions.go index 9ad56851..df189d23 100644 --- a/vendor/github.com/stretchr/testify/assert/forward_assertions.go +++ b/vendor/github.com/stretchr/testify/assert/forward_assertions.go @@ -13,4 +13,4 @@ func New(t TestingT) *Assertions { } } -//go:generate go run ../_codegen/main.go -output-package=assert -template=assertion_forward.go.tmpl -include-format-funcs +//go:generate sh -c "cd ../_codegen && go build && cd - && ../_codegen/_codegen -output-package=assert -template=assertion_forward.go.tmpl -include-format-funcs" diff --git a/vendor/github.com/stretchr/testify/require/forward_requirements.go b/vendor/github.com/stretchr/testify/require/forward_requirements.go index ac71d405..1dcb2338 100644 --- a/vendor/github.com/stretchr/testify/require/forward_requirements.go +++ b/vendor/github.com/stretchr/testify/require/forward_requirements.go @@ -13,4 +13,4 @@ func New(t TestingT) *Assertions { } } -//go:generate go run ../_codegen/main.go -output-package=require -template=require_forward.go.tmpl -include-format-funcs +//go:generate sh -c "cd ../_codegen && go build && cd - && ../_codegen/_codegen -output-package=require -template=require_forward.go.tmpl -include-format-funcs" diff --git a/vendor/github.com/stretchr/testify/require/require.go b/vendor/github.com/stretchr/testify/require/require.go index c5903f5d..cf6c7b56 100644 --- a/vendor/github.com/stretchr/testify/require/require.go +++ b/vendor/github.com/stretchr/testify/require/require.go @@ -66,7 +66,8 @@ func Containsf(t TestingT, s interface{}, contains interface{}, msg string, args t.FailNow() } -// DirExists checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// DirExists checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func DirExists(t TestingT, path string, msgAndArgs ...interface{}) { if h, ok := t.(tHelper); ok { h.Helper() @@ -77,7 +78,8 @@ func DirExists(t TestingT, path string, msgAndArgs ...interface{}) { t.FailNow() } -// DirExistsf checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// DirExistsf checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func DirExistsf(t TestingT, path string, msg string, args ...interface{}) { if h, ok := t.(tHelper); ok { h.Helper() @@ -275,12 +277,12 @@ func Errorf(t TestingT, err error, msg string, args ...interface{}) { // // assert.Eventually(t, func() bool { return true; }, time.Second, 10*time.Millisecond) func Eventually(t TestingT, condition func() bool, waitFor time.Duration, tick time.Duration, msgAndArgs ...interface{}) { - if assert.Eventually(t, condition, waitFor, tick, msgAndArgs...) { - return - } if h, ok := t.(tHelper); ok { h.Helper() } + if assert.Eventually(t, condition, waitFor, tick, msgAndArgs...) { + return + } t.FailNow() } @@ -289,12 +291,12 @@ func Eventually(t TestingT, condition func() bool, waitFor time.Duration, tick t // // assert.Eventuallyf(t, func() bool { return true; }, time.Second, 10*time.Millisecond, "error message %s", "formatted") func Eventuallyf(t TestingT, condition func() bool, waitFor time.Duration, tick time.Duration, msg string, args ...interface{}) { - if assert.Eventuallyf(t, condition, waitFor, tick, msg, args...) { - return - } if h, ok := t.(tHelper); ok { h.Helper() } + if assert.Eventuallyf(t, condition, waitFor, tick, msg, args...) { + return + } t.FailNow() } @@ -394,7 +396,8 @@ func Falsef(t TestingT, value bool, msg string, args ...interface{}) { t.FailNow() } -// FileExists checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExists checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func FileExists(t TestingT, path string, msgAndArgs ...interface{}) { if h, ok := t.(tHelper); ok { h.Helper() @@ -405,7 +408,8 @@ func FileExists(t TestingT, path string, msgAndArgs ...interface{}) { t.FailNow() } -// FileExistsf checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExistsf checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func FileExistsf(t TestingT, path string, msg string, args ...interface{}) { if h, ok := t.(tHelper); ok { h.Helper() @@ -660,7 +664,7 @@ func Implementsf(t TestingT, interfaceObject interface{}, object interface{}, ms // InDelta asserts that the two numerals are within delta of each other. // -// assert.InDelta(t, math.Pi, (22 / 7.0), 0.01) +// assert.InDelta(t, math.Pi, 22/7.0, 0.01) func InDelta(t TestingT, expected interface{}, actual interface{}, delta float64, msgAndArgs ...interface{}) { if h, ok := t.(tHelper); ok { h.Helper() @@ -717,7 +721,7 @@ func InDeltaSlicef(t TestingT, expected interface{}, actual interface{}, delta f // InDeltaf asserts that the two numerals are within delta of each other. // -// assert.InDeltaf(t, math.Pi, (22 / 7.0, "error message %s", "formatted"), 0.01) +// assert.InDeltaf(t, math.Pi, 22/7.0, 0.01, "error message %s", "formatted") func InDeltaf(t TestingT, expected interface{}, actual interface{}, delta float64, msg string, args ...interface{}) { if h, ok := t.(tHelper); ok { h.Helper() @@ -820,28 +824,6 @@ func JSONEqf(t TestingT, expected string, actual string, msg string, args ...int t.FailNow() } -// YAMLEq asserts that two YAML strings are equivalent. -func YAMLEq(t TestingT, expected string, actual string, msgAndArgs ...interface{}) { - if h, ok := t.(tHelper); ok { - h.Helper() - } - if assert.YAMLEq(t, expected, actual, msgAndArgs...) { - return - } - t.FailNow() -} - -// YAMLEqf asserts that two YAML strings are equivalent. -func YAMLEqf(t TestingT, expected string, actual string, msg string, args ...interface{}) { - if h, ok := t.(tHelper); ok { - h.Helper() - } - if assert.YAMLEqf(t, expected, actual, msg, args...) { - return - } - t.FailNow() -} - // Len asserts that the specified object has specific length. // Len also fails if the object has a type that len() not accept. // @@ -932,6 +914,34 @@ func Lessf(t TestingT, e1 interface{}, e2 interface{}, msg string, args ...inter t.FailNow() } +// Never asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// assert.Never(t, func() bool { return false; }, time.Second, 10*time.Millisecond) +func Never(t TestingT, condition func() bool, waitFor time.Duration, tick time.Duration, msgAndArgs ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.Never(t, condition, waitFor, tick, msgAndArgs...) { + return + } + t.FailNow() +} + +// Neverf asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// assert.Neverf(t, func() bool { return false; }, time.Second, 10*time.Millisecond, "error message %s", "formatted") +func Neverf(t TestingT, condition func() bool, waitFor time.Duration, tick time.Duration, msg string, args ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.Neverf(t, condition, waitFor, tick, msg, args...) { + return + } + t.FailNow() +} + // Nil asserts that the specified object is nil. // // assert.Nil(t, err) @@ -958,6 +968,30 @@ func Nilf(t TestingT, object interface{}, msg string, args ...interface{}) { t.FailNow() } +// NoDirExists checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func NoDirExists(t TestingT, path string, msgAndArgs ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.NoDirExists(t, path, msgAndArgs...) { + return + } + t.FailNow() +} + +// NoDirExistsf checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func NoDirExistsf(t TestingT, path string, msg string, args ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.NoDirExistsf(t, path, msg, args...) { + return + } + t.FailNow() +} + // NoError asserts that a function returned no error (i.e. `nil`). // // actualObj, err := SomeFunction() @@ -990,6 +1024,30 @@ func NoErrorf(t TestingT, err error, msg string, args ...interface{}) { t.FailNow() } +// NoFileExists checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func NoFileExists(t TestingT, path string, msgAndArgs ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.NoFileExists(t, path, msgAndArgs...) { + return + } + t.FailNow() +} + +// NoFileExistsf checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func NoFileExistsf(t TestingT, path string, msg string, args ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.NoFileExistsf(t, path, msg, args...) { + return + } + t.FailNow() +} + // NotContains asserts that the specified string, list(array, slice...) or map does NOT contain the // specified substring or element. // @@ -1166,6 +1224,38 @@ func NotRegexpf(t TestingT, rx interface{}, str interface{}, msg string, args .. t.FailNow() } +// NotSame asserts that two pointers do not reference the same object. +// +// assert.NotSame(t, ptr1, ptr2) +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func NotSame(t TestingT, expected interface{}, actual interface{}, msgAndArgs ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.NotSame(t, expected, actual, msgAndArgs...) { + return + } + t.FailNow() +} + +// NotSamef asserts that two pointers do not reference the same object. +// +// assert.NotSamef(t, ptr1, ptr2, "error message %s", "formatted") +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func NotSamef(t TestingT, expected interface{}, actual interface{}, msg string, args ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.NotSamef(t, expected, actual, msg, args...) { + return + } + t.FailNow() +} + // NotSubset asserts that the specified list(array, slice...) contains not all // elements given in the specified subset(array, slice...). // @@ -1229,6 +1319,36 @@ func Panics(t TestingT, f assert.PanicTestFunc, msgAndArgs ...interface{}) { t.FailNow() } +// PanicsWithError asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// assert.PanicsWithError(t, "crazy error", func(){ GoCrazy() }) +func PanicsWithError(t TestingT, errString string, f assert.PanicTestFunc, msgAndArgs ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.PanicsWithError(t, errString, f, msgAndArgs...) { + return + } + t.FailNow() +} + +// PanicsWithErrorf asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// assert.PanicsWithErrorf(t, "crazy error", func(){ GoCrazy() }, "error message %s", "formatted") +func PanicsWithErrorf(t TestingT, errString string, f assert.PanicTestFunc, msg string, args ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.PanicsWithErrorf(t, errString, f, msg, args...) { + return + } + t.FailNow() +} + // PanicsWithValue asserts that the code inside the specified PanicTestFunc panics, and that // the recovered panic value equals the expected panic value. // @@ -1410,6 +1530,28 @@ func WithinDurationf(t TestingT, expected time.Time, actual time.Time, delta tim t.FailNow() } +// YAMLEq asserts that two YAML strings are equivalent. +func YAMLEq(t TestingT, expected string, actual string, msgAndArgs ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.YAMLEq(t, expected, actual, msgAndArgs...) { + return + } + t.FailNow() +} + +// YAMLEqf asserts that two YAML strings are equivalent. +func YAMLEqf(t TestingT, expected string, actual string, msg string, args ...interface{}) { + if h, ok := t.(tHelper); ok { + h.Helper() + } + if assert.YAMLEqf(t, expected, actual, msg, args...) { + return + } + t.FailNow() +} + // Zero asserts that i is the zero value for its type. func Zero(t TestingT, i interface{}, msgAndArgs ...interface{}) { if h, ok := t.(tHelper); ok { diff --git a/vendor/github.com/stretchr/testify/require/require_forward.go b/vendor/github.com/stretchr/testify/require/require_forward.go index 804fae03..5aac226d 100644 --- a/vendor/github.com/stretchr/testify/require/require_forward.go +++ b/vendor/github.com/stretchr/testify/require/require_forward.go @@ -54,7 +54,8 @@ func (a *Assertions) Containsf(s interface{}, contains interface{}, msg string, Containsf(a.t, s, contains, msg, args...) } -// DirExists checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// DirExists checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func (a *Assertions) DirExists(path string, msgAndArgs ...interface{}) { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -62,7 +63,8 @@ func (a *Assertions) DirExists(path string, msgAndArgs ...interface{}) { DirExists(a.t, path, msgAndArgs...) } -// DirExistsf checks whether a directory exists in the given path. It also fails if the path is a file rather a directory or there is an error checking whether it exists. +// DirExistsf checks whether a directory exists in the given path. It also fails +// if the path is a file rather a directory or there is an error checking whether it exists. func (a *Assertions) DirExistsf(path string, msg string, args ...interface{}) { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -310,7 +312,8 @@ func (a *Assertions) Falsef(value bool, msg string, args ...interface{}) { Falsef(a.t, value, msg, args...) } -// FileExists checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExists checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func (a *Assertions) FileExists(path string, msgAndArgs ...interface{}) { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -318,7 +321,8 @@ func (a *Assertions) FileExists(path string, msgAndArgs ...interface{}) { FileExists(a.t, path, msgAndArgs...) } -// FileExistsf checks whether a file exists in the given path. It also fails if the path points to a directory or there is an error when trying to check the file. +// FileExistsf checks whether a file exists in the given path. It also fails if +// the path points to a directory or there is an error when trying to check the file. func (a *Assertions) FileExistsf(path string, msg string, args ...interface{}) { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -522,7 +526,7 @@ func (a *Assertions) Implementsf(interfaceObject interface{}, object interface{} // InDelta asserts that the two numerals are within delta of each other. // -// a.InDelta(math.Pi, (22 / 7.0), 0.01) +// a.InDelta(math.Pi, 22/7.0, 0.01) func (a *Assertions) InDelta(expected interface{}, actual interface{}, delta float64, msgAndArgs ...interface{}) { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -564,7 +568,7 @@ func (a *Assertions) InDeltaSlicef(expected interface{}, actual interface{}, del // InDeltaf asserts that the two numerals are within delta of each other. // -// a.InDeltaf(math.Pi, (22 / 7.0, "error message %s", "formatted"), 0.01) +// a.InDeltaf(math.Pi, 22/7.0, 0.01, "error message %s", "formatted") func (a *Assertions) InDeltaf(expected interface{}, actual interface{}, delta float64, msg string, args ...interface{}) { if h, ok := a.t.(tHelper); ok { h.Helper() @@ -640,22 +644,6 @@ func (a *Assertions) JSONEqf(expected string, actual string, msg string, args .. JSONEqf(a.t, expected, actual, msg, args...) } -// YAMLEq asserts that two YAML strings are equivalent. -func (a *Assertions) YAMLEq(expected string, actual string, msgAndArgs ...interface{}) { - if h, ok := a.t.(tHelper); ok { - h.Helper() - } - YAMLEq(a.t, expected, actual, msgAndArgs...) -} - -// YAMLEqf asserts that two YAML strings are equivalent. -func (a *Assertions) YAMLEqf(expected string, actual string, msg string, args ...interface{}) { - if h, ok := a.t.(tHelper); ok { - h.Helper() - } - YAMLEqf(a.t, expected, actual, msg, args...) -} - // Len asserts that the specified object has specific length. // Len also fails if the object has a type that len() not accept. // @@ -728,6 +716,28 @@ func (a *Assertions) Lessf(e1 interface{}, e2 interface{}, msg string, args ...i Lessf(a.t, e1, e2, msg, args...) } +// Never asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// a.Never(func() bool { return false; }, time.Second, 10*time.Millisecond) +func (a *Assertions) Never(condition func() bool, waitFor time.Duration, tick time.Duration, msgAndArgs ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + Never(a.t, condition, waitFor, tick, msgAndArgs...) +} + +// Neverf asserts that the given condition doesn't satisfy in waitFor time, +// periodically checking the target function each tick. +// +// a.Neverf(func() bool { return false; }, time.Second, 10*time.Millisecond, "error message %s", "formatted") +func (a *Assertions) Neverf(condition func() bool, waitFor time.Duration, tick time.Duration, msg string, args ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + Neverf(a.t, condition, waitFor, tick, msg, args...) +} + // Nil asserts that the specified object is nil. // // a.Nil(err) @@ -748,6 +758,24 @@ func (a *Assertions) Nilf(object interface{}, msg string, args ...interface{}) { Nilf(a.t, object, msg, args...) } +// NoDirExists checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func (a *Assertions) NoDirExists(path string, msgAndArgs ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + NoDirExists(a.t, path, msgAndArgs...) +} + +// NoDirExistsf checks whether a directory does not exist in the given path. +// It fails if the path points to an existing _directory_ only. +func (a *Assertions) NoDirExistsf(path string, msg string, args ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + NoDirExistsf(a.t, path, msg, args...) +} + // NoError asserts that a function returned no error (i.e. `nil`). // // actualObj, err := SomeFunction() @@ -774,6 +802,24 @@ func (a *Assertions) NoErrorf(err error, msg string, args ...interface{}) { NoErrorf(a.t, err, msg, args...) } +// NoFileExists checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func (a *Assertions) NoFileExists(path string, msgAndArgs ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + NoFileExists(a.t, path, msgAndArgs...) +} + +// NoFileExistsf checks whether a file does not exist in a given path. It fails +// if the path points to an existing _file_ only. +func (a *Assertions) NoFileExistsf(path string, msg string, args ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + NoFileExistsf(a.t, path, msg, args...) +} + // NotContains asserts that the specified string, list(array, slice...) or map does NOT contain the // specified substring or element. // @@ -914,6 +960,32 @@ func (a *Assertions) NotRegexpf(rx interface{}, str interface{}, msg string, arg NotRegexpf(a.t, rx, str, msg, args...) } +// NotSame asserts that two pointers do not reference the same object. +// +// a.NotSame(ptr1, ptr2) +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func (a *Assertions) NotSame(expected interface{}, actual interface{}, msgAndArgs ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + NotSame(a.t, expected, actual, msgAndArgs...) +} + +// NotSamef asserts that two pointers do not reference the same object. +// +// a.NotSamef(ptr1, ptr2, "error message %s", "formatted") +// +// Both arguments must be pointer variables. Pointer variable sameness is +// determined based on the equality of both type and value. +func (a *Assertions) NotSamef(expected interface{}, actual interface{}, msg string, args ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + NotSamef(a.t, expected, actual, msg, args...) +} + // NotSubset asserts that the specified list(array, slice...) contains not all // elements given in the specified subset(array, slice...). // @@ -962,6 +1034,30 @@ func (a *Assertions) Panics(f assert.PanicTestFunc, msgAndArgs ...interface{}) { Panics(a.t, f, msgAndArgs...) } +// PanicsWithError asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// a.PanicsWithError("crazy error", func(){ GoCrazy() }) +func (a *Assertions) PanicsWithError(errString string, f assert.PanicTestFunc, msgAndArgs ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + PanicsWithError(a.t, errString, f, msgAndArgs...) +} + +// PanicsWithErrorf asserts that the code inside the specified PanicTestFunc +// panics, and that the recovered panic value is an error that satisfies the +// EqualError comparison. +// +// a.PanicsWithErrorf("crazy error", func(){ GoCrazy() }, "error message %s", "formatted") +func (a *Assertions) PanicsWithErrorf(errString string, f assert.PanicTestFunc, msg string, args ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + PanicsWithErrorf(a.t, errString, f, msg, args...) +} + // PanicsWithValue asserts that the code inside the specified PanicTestFunc panics, and that // the recovered panic value equals the expected panic value. // @@ -1104,6 +1200,22 @@ func (a *Assertions) WithinDurationf(expected time.Time, actual time.Time, delta WithinDurationf(a.t, expected, actual, delta, msg, args...) } +// YAMLEq asserts that two YAML strings are equivalent. +func (a *Assertions) YAMLEq(expected string, actual string, msgAndArgs ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + YAMLEq(a.t, expected, actual, msgAndArgs...) +} + +// YAMLEqf asserts that two YAML strings are equivalent. +func (a *Assertions) YAMLEqf(expected string, actual string, msg string, args ...interface{}) { + if h, ok := a.t.(tHelper); ok { + h.Helper() + } + YAMLEqf(a.t, expected, actual, msg, args...) +} + // Zero asserts that i is the zero value for its type. func (a *Assertions) Zero(i interface{}, msgAndArgs ...interface{}) { if h, ok := a.t.(tHelper); ok { diff --git a/vendor/github.com/stretchr/testify/require/requirements.go b/vendor/github.com/stretchr/testify/require/requirements.go index 6b85c5ec..91772dfe 100644 --- a/vendor/github.com/stretchr/testify/require/requirements.go +++ b/vendor/github.com/stretchr/testify/require/requirements.go @@ -26,4 +26,4 @@ type BoolAssertionFunc func(TestingT, bool, ...interface{}) // for table driven tests. type ErrorAssertionFunc func(TestingT, error, ...interface{}) -//go:generate go run ../_codegen/main.go -output-package=require -template=require.go.tmpl -include-format-funcs +//go:generate sh -c "cd ../_codegen && go build && cd - && ../_codegen/_codegen -output-package=require -template=require.go.tmpl -include-format-funcs" diff --git a/vendor/golang.org/x/crypto/ssh/terminal/terminal.go b/vendor/golang.org/x/crypto/ssh/terminal/terminal.go index d1b4fca3..2ffb97bf 100644 --- a/vendor/golang.org/x/crypto/ssh/terminal/terminal.go +++ b/vendor/golang.org/x/crypto/ssh/terminal/terminal.go @@ -113,6 +113,7 @@ func NewTerminal(c io.ReadWriter, prompt string) *Terminal { } const ( + keyCtrlC = 3 keyCtrlD = 4 keyCtrlU = 21 keyEnter = '\r' @@ -151,8 +152,12 @@ func bytesToKey(b []byte, pasteActive bool) (rune, []byte) { switch b[0] { case 1: // ^A return keyHome, b[1:] + case 2: // ^B + return keyLeft, b[1:] case 5: // ^E return keyEnd, b[1:] + case 6: // ^F + return keyRight, b[1:] case 8: // ^H return keyBackspace, b[1:] case 11: // ^K @@ -738,6 +743,9 @@ func (t *Terminal) readLine() (line string, err error) { return "", io.EOF } } + if key == keyCtrlC { + return "", io.EOF + } if key == keyPasteStart { t.pasteActive = true if len(t.line) == 0 { diff --git a/vendor/golang.org/x/net/http2/client_conn_pool.go b/vendor/golang.org/x/net/http2/client_conn_pool.go index f4d9b5ec..3a67636f 100644 --- a/vendor/golang.org/x/net/http2/client_conn_pool.go +++ b/vendor/golang.org/x/net/http2/client_conn_pool.go @@ -107,6 +107,7 @@ func (p *clientConnPool) getClientConn(req *http.Request, addr string, dialOnMis // dialCall is an in-flight Transport dial call to a host. type dialCall struct { + _ incomparable p *clientConnPool done chan struct{} // closed when done res *ClientConn // valid after done is closed @@ -180,6 +181,7 @@ func (p *clientConnPool) addConnIfNeeded(key string, t *Transport, c *tls.Conn) } type addConnCall struct { + _ incomparable p *clientConnPool done chan struct{} // closed when done err error @@ -200,12 +202,6 @@ func (c *addConnCall) run(t *Transport, key string, tc *tls.Conn) { close(c.done) } -func (p *clientConnPool) addConn(key string, cc *ClientConn) { - p.mu.Lock() - p.addConnLocked(key, cc) - p.mu.Unlock() -} - // p.mu must be held func (p *clientConnPool) addConnLocked(key string, cc *ClientConn) { for _, v := range p.conns[key] { diff --git a/vendor/golang.org/x/net/http2/flow.go b/vendor/golang.org/x/net/http2/flow.go index cea601fc..b51f0e0c 100644 --- a/vendor/golang.org/x/net/http2/flow.go +++ b/vendor/golang.org/x/net/http2/flow.go @@ -8,6 +8,8 @@ package http2 // flow is the flow control window's size. type flow struct { + _ incomparable + // n is the number of DATA bytes we're allowed to send. // A flow is kept both on a conn and a per-stream. n int32 diff --git a/vendor/golang.org/x/net/http2/hpack/huffman.go b/vendor/golang.org/x/net/http2/hpack/huffman.go index b412a96c..a1ab2f05 100644 --- a/vendor/golang.org/x/net/http2/hpack/huffman.go +++ b/vendor/golang.org/x/net/http2/hpack/huffman.go @@ -105,7 +105,14 @@ func huffmanDecode(buf *bytes.Buffer, maxLen int, v []byte) error { return nil } +// incomparable is a zero-width, non-comparable type. Adding it to a struct +// makes that struct also non-comparable, and generally doesn't add +// any size (as long as it's first). +type incomparable [0]func() + type node struct { + _ incomparable + // children is non-nil for internal nodes children *[256]*node diff --git a/vendor/golang.org/x/net/http2/http2.go b/vendor/golang.org/x/net/http2/http2.go index 27cc893c..5571ccfd 100644 --- a/vendor/golang.org/x/net/http2/http2.go +++ b/vendor/golang.org/x/net/http2/http2.go @@ -241,6 +241,7 @@ func (cw closeWaiter) Wait() { // Its buffered writer is lazily allocated as needed, to minimize // idle memory usage with many connections. type bufferedWriter struct { + _ incomparable w io.Writer // immutable bw *bufio.Writer // non-nil when data is buffered } @@ -313,6 +314,7 @@ func bodyAllowedForStatus(status int) bool { } type httpError struct { + _ incomparable msg string timeout bool } @@ -376,3 +378,8 @@ func (s *sorter) SortStrings(ss []string) { func validPseudoPath(v string) bool { return (len(v) > 0 && v[0] == '/') || v == "*" } + +// incomparable is a zero-width, non-comparable type. Adding it to a struct +// makes that struct also non-comparable, and generally doesn't add +// any size (as long as it's first). +type incomparable [0]func() diff --git a/vendor/golang.org/x/net/http2/server.go b/vendor/golang.org/x/net/http2/server.go index bc9e41a1..345b7cd8 100644 --- a/vendor/golang.org/x/net/http2/server.go +++ b/vendor/golang.org/x/net/http2/server.go @@ -761,6 +761,7 @@ func (sc *serverConn) readFrames() { // frameWriteResult is the message passed from writeFrameAsync to the serve goroutine. type frameWriteResult struct { + _ incomparable wr FrameWriteRequest // what was written (or attempted) err error // result of the writeFrame call } @@ -771,7 +772,7 @@ type frameWriteResult struct { // serverConn. func (sc *serverConn) writeFrameAsync(wr FrameWriteRequest) { err := wr.write.writeFrame(sc) - sc.wroteFrameCh <- frameWriteResult{wr, err} + sc.wroteFrameCh <- frameWriteResult{wr: wr, err: err} } func (sc *serverConn) closeAllStreamsOnConnClose() { @@ -1161,7 +1162,7 @@ func (sc *serverConn) startFrameWrite(wr FrameWriteRequest) { if wr.write.staysWithinBuffer(sc.bw.Available()) { sc.writingFrameAsync = false err := wr.write.writeFrame(sc) - sc.wroteFrame(frameWriteResult{wr, err}) + sc.wroteFrame(frameWriteResult{wr: wr, err: err}) } else { sc.writingFrameAsync = true go sc.writeFrameAsync(wr) @@ -2057,7 +2058,7 @@ func (sc *serverConn) newWriterAndRequestNoBody(st *stream, rp requestParam) (*r var trailer http.Header for _, v := range rp.header["Trailer"] { for _, key := range strings.Split(v, ",") { - key = http.CanonicalHeaderKey(strings.TrimSpace(key)) + key = http.CanonicalHeaderKey(textproto.TrimString(key)) switch key { case "Transfer-Encoding", "Trailer", "Content-Length": // Bogus. (copy of http1 rules) @@ -2275,6 +2276,7 @@ func (sc *serverConn) sendWindowUpdate32(st *stream, n int32) { // requestBody is the Handler's Request.Body type. // Read and Close may be called concurrently. type requestBody struct { + _ incomparable stream *stream conn *serverConn closed bool // for use by Close only diff --git a/vendor/golang.org/x/net/http2/transport.go b/vendor/golang.org/x/net/http2/transport.go index e4fb0253..76a92e0c 100644 --- a/vendor/golang.org/x/net/http2/transport.go +++ b/vendor/golang.org/x/net/http2/transport.go @@ -108,6 +108,19 @@ type Transport struct { // waiting for their turn. StrictMaxConcurrentStreams bool + // ReadIdleTimeout is the timeout after which a health check using ping + // frame will be carried out if no frame is received on the connection. + // Note that a ping response will is considered a received frame, so if + // there is no other traffic on the connection, the health check will + // be performed every ReadIdleTimeout interval. + // If zero, no health check is performed. + ReadIdleTimeout time.Duration + + // PingTimeout is the timeout after which the connection will be closed + // if a response to Ping is not received. + // Defaults to 15s. + PingTimeout time.Duration + // t1, if non-nil, is the standard library Transport using // this transport. Its settings are used (but not its // RoundTrip method, etc). @@ -131,6 +144,14 @@ func (t *Transport) disableCompression() bool { return t.DisableCompression || (t.t1 != nil && t.t1.DisableCompression) } +func (t *Transport) pingTimeout() time.Duration { + if t.PingTimeout == 0 { + return 15 * time.Second + } + return t.PingTimeout + +} + // ConfigureTransport configures a net/http HTTP/1 Transport to use HTTP/2. // It returns an error if t1 has already been HTTP/2-enabled. func ConfigureTransport(t1 *http.Transport) error { @@ -675,6 +696,20 @@ func (t *Transport) newClientConn(c net.Conn, singleUse bool) (*ClientConn, erro return cc, nil } +func (cc *ClientConn) healthCheck() { + pingTimeout := cc.t.pingTimeout() + // We don't need to periodically ping in the health check, because the readLoop of ClientConn will + // trigger the healthCheck again if there is no frame received. + ctx, cancel := context.WithTimeout(context.Background(), pingTimeout) + defer cancel() + err := cc.Ping(ctx) + if err != nil { + cc.closeForLostPing() + cc.t.connPool().MarkDead(cc) + return + } +} + func (cc *ClientConn) setGoAway(f *GoAwayFrame) { cc.mu.Lock() defer cc.mu.Unlock() @@ -846,14 +881,12 @@ func (cc *ClientConn) sendGoAway() error { return nil } -// Close closes the client connection immediately. -// -// In-flight requests are interrupted. For a graceful shutdown, use Shutdown instead. -func (cc *ClientConn) Close() error { +// closes the client connection immediately. In-flight requests are interrupted. +// err is sent to streams. +func (cc *ClientConn) closeForError(err error) error { cc.mu.Lock() defer cc.cond.Broadcast() defer cc.mu.Unlock() - err := errors.New("http2: client connection force closed via ClientConn.Close") for id, cs := range cc.streams { select { case cs.resc <- resAndError{err: err}: @@ -866,6 +899,20 @@ func (cc *ClientConn) Close() error { return cc.tconn.Close() } +// Close closes the client connection immediately. +// +// In-flight requests are interrupted. For a graceful shutdown, use Shutdown instead. +func (cc *ClientConn) Close() error { + err := errors.New("http2: client connection force closed via ClientConn.Close") + return cc.closeForError(err) +} + +// closes the client connection immediately. In-flight requests are interrupted. +func (cc *ClientConn) closeForLostPing() error { + err := errors.New("http2: client connection lost") + return cc.closeForError(err) +} + const maxAllocFrameSize = 512 << 10 // frameBuffer returns a scratch buffer suitable for writing DATA frames. @@ -916,7 +963,7 @@ func commaSeparatedTrailers(req *http.Request) (string, error) { k = http.CanonicalHeaderKey(k) switch k { case "Transfer-Encoding", "Trailer", "Content-Length": - return "", &badStringError{"invalid Trailer key", k} + return "", fmt.Errorf("invalid Trailer key %q", k) } keys = append(keys, k) } @@ -1394,13 +1441,6 @@ func (cs *clientStream) awaitFlowControl(maxBytes int) (taken int32, err error) } } -type badStringError struct { - what string - str string -} - -func (e *badStringError) Error() string { return fmt.Sprintf("%s %q", e.what, e.str) } - // requires cc.mu be held. func (cc *ClientConn) encodeHeaders(req *http.Request, addGzipHeader bool, trailers string, contentLength int64) ([]byte, error) { cc.hbuf.Reset() @@ -1616,6 +1656,7 @@ func (cc *ClientConn) writeHeader(name, value string) { } type resAndError struct { + _ incomparable res *http.Response err error } @@ -1663,6 +1704,7 @@ func (cc *ClientConn) streamByID(id uint32, andRemove bool) *clientStream { // clientConnReadLoop is the state owned by the clientConn's frame-reading readLoop. type clientConnReadLoop struct { + _ incomparable cc *ClientConn closeWhenIdle bool } @@ -1742,8 +1784,17 @@ func (rl *clientConnReadLoop) run() error { rl.closeWhenIdle = cc.t.disableKeepAlives() || cc.singleUse gotReply := false // ever saw a HEADERS reply gotSettings := false + readIdleTimeout := cc.t.ReadIdleTimeout + var t *time.Timer + if readIdleTimeout != 0 { + t = time.AfterFunc(readIdleTimeout, cc.healthCheck) + defer t.Stop() + } for { f, err := cc.fr.ReadFrame() + if t != nil { + t.Reset(readIdleTimeout) + } if err != nil { cc.vlogf("http2: Transport readFrame error on conn %p: (%T) %v", cc, err, err) } @@ -2479,6 +2530,7 @@ func (rt erringRoundTripper) RoundTrip(*http.Request) (*http.Response, error) { // gzipReader wraps a response body so it can lazily // call gzip.NewReader on the first call to Read type gzipReader struct { + _ incomparable body io.ReadCloser // underlying Response.Body zr *gzip.Reader // lazily-initialized gzip reader zerr error // sticky error diff --git a/vendor/golang.org/x/sys/internal/unsafeheader/unsafeheader.go b/vendor/golang.org/x/sys/internal/unsafeheader/unsafeheader.go new file mode 100644 index 00000000..e07899b9 --- /dev/null +++ b/vendor/golang.org/x/sys/internal/unsafeheader/unsafeheader.go @@ -0,0 +1,30 @@ +// Copyright 2020 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package unsafeheader contains header declarations for the Go runtime's +// slice and string implementations. +// +// This package allows x/sys to use types equivalent to +// reflect.SliceHeader and reflect.StringHeader without introducing +// a dependency on the (relatively heavy) "reflect" package. +package unsafeheader + +import ( + "unsafe" +) + +// Slice is the runtime representation of a slice. +// It cannot be used safely or portably and its representation may change in a later release. +type Slice struct { + Data unsafe.Pointer + Len int + Cap int +} + +// String is the runtime representation of a string. +// It cannot be used safely or portably and its representation may change in a later release. +type String struct { + Data unsafe.Pointer + Len int +} diff --git a/vendor/golang.org/x/sys/unix/mkerrors.sh b/vendor/golang.org/x/sys/unix/mkerrors.sh index ab09aafc..780e387e 100644 --- a/vendor/golang.org/x/sys/unix/mkerrors.sh +++ b/vendor/golang.org/x/sys/unix/mkerrors.sh @@ -187,6 +187,7 @@ struct ltchars { #include #include #include +#include #include #include #include @@ -480,7 +481,7 @@ ccflags="$@" $2 ~ /^(MS|MNT|UMOUNT)_/ || $2 ~ /^NS_GET_/ || $2 ~ /^TUN(SET|GET|ATTACH|DETACH)/ || - $2 ~ /^(O|F|[ES]?FD|NAME|S|PTRACE|PT)_/ || + $2 ~ /^(O|F|[ES]?FD|NAME|S|PTRACE|PT|TFD)_/ || $2 ~ /^KEXEC_/ || $2 ~ /^LINUX_REBOOT_CMD_/ || $2 ~ /^LINUX_REBOOT_MAGIC[12]$/ || diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin.1_13.go b/vendor/golang.org/x/sys/unix/syscall_darwin.1_13.go index f911617b..dc0befee 100644 --- a/vendor/golang.org/x/sys/unix/syscall_darwin.1_13.go +++ b/vendor/golang.org/x/sys/unix/syscall_darwin.1_13.go @@ -6,7 +6,11 @@ package unix -import "unsafe" +import ( + "unsafe" + + "golang.org/x/sys/internal/unsafeheader" +) //sys closedir(dir uintptr) (err error) //sys readdir_r(dir uintptr, entry *Dirent, result **Dirent) (res Errno) @@ -71,6 +75,7 @@ func Getdirentries(fd int, buf []byte, basep *uintptr) (n int, err error) { cnt++ continue } + reclen := int(entry.Reclen) if reclen > len(buf) { // Not enough room. Return for now. @@ -79,13 +84,15 @@ func Getdirentries(fd int, buf []byte, basep *uintptr) (n int, err error) { // restarting is O(n^2) in the length of the directory. Oh well. break } + // Copy entry into return buffer. - s := struct { - ptr unsafe.Pointer - siz int - cap int - }{ptr: unsafe.Pointer(&entry), siz: reclen, cap: reclen} - copy(buf, *(*[]byte)(unsafe.Pointer(&s))) + var s []byte + hdr := (*unsafeheader.Slice)(unsafe.Pointer(&s)) + hdr.Data = unsafe.Pointer(&entry) + hdr.Cap = reclen + hdr.Len = reclen + copy(buf, s) + buf = buf[reclen:] n += reclen cnt++ diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin.go b/vendor/golang.org/x/sys/unix/syscall_darwin.go index 9a5a6ee5..0cf31acf 100644 --- a/vendor/golang.org/x/sys/unix/syscall_darwin.go +++ b/vendor/golang.org/x/sys/unix/syscall_darwin.go @@ -423,6 +423,7 @@ func Sendfile(outfd int, infd int, offset *int64, count int) (written int, err e //sysnb Getrlimit(which int, lim *Rlimit) (err error) //sysnb Getrusage(who int, rusage *Rusage) (err error) //sysnb Getsid(pid int) (sid int, err error) +//sysnb Gettimeofday(tp *Timeval) (err error) //sysnb Getuid() (uid int) //sysnb Issetugid() (tainted bool) //sys Kqueue() (fd int, err error) diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin_386.go b/vendor/golang.org/x/sys/unix/syscall_darwin_386.go index 707ba4f5..2724e3a5 100644 --- a/vendor/golang.org/x/sys/unix/syscall_darwin_386.go +++ b/vendor/golang.org/x/sys/unix/syscall_darwin_386.go @@ -20,17 +20,6 @@ func setTimeval(sec, usec int64) Timeval { return Timeval{Sec: int32(sec), Usec: int32(usec)} } -//sysnb gettimeofday(tp *Timeval) (sec int32, usec int32, err error) -func Gettimeofday(tv *Timeval) (err error) { - // The tv passed to gettimeofday must be non-nil - // but is otherwise unused. The answers come back - // in the two registers. - sec, usec, err := gettimeofday(tv) - tv.Sec = int32(sec) - tv.Usec = int32(usec) - return err -} - func SetKevent(k *Kevent_t, fd, mode, flags int) { k.Ident = uint32(fd) k.Filter = int16(mode) diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin_amd64.go b/vendor/golang.org/x/sys/unix/syscall_darwin_amd64.go index fdbfb591..ce2e0d24 100644 --- a/vendor/golang.org/x/sys/unix/syscall_darwin_amd64.go +++ b/vendor/golang.org/x/sys/unix/syscall_darwin_amd64.go @@ -20,17 +20,6 @@ func setTimeval(sec, usec int64) Timeval { return Timeval{Sec: sec, Usec: int32(usec)} } -//sysnb gettimeofday(tp *Timeval) (sec int64, usec int32, err error) -func Gettimeofday(tv *Timeval) (err error) { - // The tv passed to gettimeofday must be non-nil - // but is otherwise unused. The answers come back - // in the two registers. - sec, usec, err := gettimeofday(tv) - tv.Sec = sec - tv.Usec = usec - return err -} - func SetKevent(k *Kevent_t, fd, mode, flags int) { k.Ident = uint64(fd) k.Filter = int16(mode) diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin_arm.go b/vendor/golang.org/x/sys/unix/syscall_darwin_arm.go index f8bc4cfb..fc17a3f2 100644 --- a/vendor/golang.org/x/sys/unix/syscall_darwin_arm.go +++ b/vendor/golang.org/x/sys/unix/syscall_darwin_arm.go @@ -20,17 +20,6 @@ func setTimeval(sec, usec int64) Timeval { return Timeval{Sec: int32(sec), Usec: int32(usec)} } -//sysnb gettimeofday(tp *Timeval) (sec int32, usec int32, err error) -func Gettimeofday(tv *Timeval) (err error) { - // The tv passed to gettimeofday must be non-nil - // but is otherwise unused. The answers come back - // in the two registers. - sec, usec, err := gettimeofday(tv) - tv.Sec = int32(sec) - tv.Usec = int32(usec) - return err -} - func SetKevent(k *Kevent_t, fd, mode, flags int) { k.Ident = uint32(fd) k.Filter = int16(mode) diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin_arm64.go b/vendor/golang.org/x/sys/unix/syscall_darwin_arm64.go index 5ede3ac3..1e91ddf3 100644 --- a/vendor/golang.org/x/sys/unix/syscall_darwin_arm64.go +++ b/vendor/golang.org/x/sys/unix/syscall_darwin_arm64.go @@ -22,17 +22,6 @@ func setTimeval(sec, usec int64) Timeval { return Timeval{Sec: sec, Usec: int32(usec)} } -//sysnb gettimeofday(tp *Timeval) (sec int64, usec int32, err error) -func Gettimeofday(tv *Timeval) (err error) { - // The tv passed to gettimeofday must be non-nil - // but is otherwise unused. The answers come back - // in the two registers. - sec, usec, err := gettimeofday(tv) - tv.Sec = sec - tv.Usec = usec - return err -} - func SetKevent(k *Kevent_t, fd, mode, flags int) { k.Ident = uint64(fd) k.Filter = int16(mode) diff --git a/vendor/golang.org/x/sys/unix/syscall_linux.go b/vendor/golang.org/x/sys/unix/syscall_linux.go index 2b151eba..e50e4cb2 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux.go @@ -97,6 +97,12 @@ func IoctlSetRTCTime(fd int, value *RTCTime) error { return err } +func IoctlSetRTCWkAlrm(fd int, value *RTCWkAlrm) error { + err := ioctl(fd, RTC_WKALM_SET, uintptr(unsafe.Pointer(value))) + runtime.KeepAlive(value) + return err +} + func IoctlGetUint32(fd int, req uint) (uint32, error) { var value uint32 err := ioctl(fd, req, uintptr(unsafe.Pointer(&value))) @@ -109,6 +115,12 @@ func IoctlGetRTCTime(fd int) (*RTCTime, error) { return &value, err } +func IoctlGetRTCWkAlrm(fd int) (*RTCWkAlrm, error) { + var value RTCWkAlrm + err := ioctl(fd, RTC_WKALM_RD, uintptr(unsafe.Pointer(&value))) + return &value, err +} + //sys Linkat(olddirfd int, oldpath string, newdirfd int, newpath string, flags int) (err error) func Link(oldpath string, newpath string) (err error) { @@ -133,12 +145,6 @@ func Openat(dirfd int, path string, flags int, mode uint32) (fd int, err error) return openat(dirfd, path, flags|O_LARGEFILE, mode) } -//sys openat2(dirfd int, path string, open_how *OpenHow, size int) (fd int, err error) - -func Openat2(dirfd int, path string, how *OpenHow) (fd int, err error) { - return openat2(dirfd, path, how, SizeofOpenHow) -} - //sys ppoll(fds *PollFd, nfds int, timeout *Timespec, sigmask *Sigset_t) (n int, err error) func Ppoll(fds []PollFd, timeout *Timespec, sigmask *Sigset_t) (n int, err error) { @@ -1639,6 +1645,15 @@ func Sendfile(outfd int, infd int, offset *int64, count int) (written int, err e //sys CopyFileRange(rfd int, roff *int64, wfd int, woff *int64, len int, flags int) (n int, err error) //sys DeleteModule(name string, flags int) (err error) //sys Dup(oldfd int) (fd int, err error) + +func Dup2(oldfd, newfd int) error { + // Android O and newer blocks dup2; riscv and arm64 don't implement dup2. + if runtime.GOOS == "android" || runtime.GOARCH == "riscv64" || runtime.GOARCH == "arm64" { + return Dup3(oldfd, newfd, 0) + } + return dup2(oldfd, newfd) +} + //sys Dup3(oldfd int, newfd int, flags int) (err error) //sysnb EpollCreate1(flag int) (fd int, err error) //sysnb EpollCtl(epfd int, op int, fd int, event *EpollEvent) (err error) @@ -1763,6 +1778,9 @@ func Signalfd(fd int, sigmask *Sigset_t, flags int) (newfd int, err error) { //sys Syncfs(fd int) (err error) //sysnb Sysinfo(info *Sysinfo_t) (err error) //sys Tee(rfd int, wfd int, len int, flags int) (n int64, err error) +//sysnb TimerfdCreate(clockid int, flags int) (fd int, err error) +//sysnb TimerfdGettime(fd int, currValue *ItimerSpec) (err error) +//sysnb TimerfdSettime(fd int, flags int, newValue *ItimerSpec, oldValue *ItimerSpec) (err error) //sysnb Tgkill(tgid int, tid int, sig syscall.Signal) (err error) //sysnb Times(tms *Tms) (ticks uintptr, err error) //sysnb Umask(mask int) (oldmask int) @@ -1932,6 +1950,20 @@ func Vmsplice(fd int, iovs []Iovec, flags int) (int, error) { return int(n), nil } +func isGroupMember(gid int) bool { + groups, err := Getgroups() + if err != nil { + return false + } + + for _, g := range groups { + if g == gid { + return true + } + } + return false +} + //sys faccessat(dirfd int, path string, mode uint32) (err error) func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { @@ -1989,7 +2021,7 @@ func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { gid = Getgid() } - if uint32(gid) == st.Gid { + if uint32(gid) == st.Gid || isGroupMember(gid) { fmode = (st.Mode >> 3) & 7 } else { fmode = st.Mode & 7 @@ -2184,7 +2216,6 @@ func Klogset(typ int, arg int) (err error) { // TimerGetoverrun // TimerGettime // TimerSettime -// Timerfd // Tkill (obsolete) // Tuxcall // Umount2 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_386.go b/vendor/golang.org/x/sys/unix/syscall_linux_386.go index a8374b67..048d18e3 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_386.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_386.go @@ -49,7 +49,7 @@ func Pipe2(p []int, flags int) (err error) { // 64-bit file system and 32-bit uid calls // (386 default is 32-bit file system and 16-bit uid). -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64_64 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go b/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go index 8ed1d546..72efe86e 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go @@ -6,7 +6,7 @@ package unix -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_arm.go b/vendor/golang.org/x/sys/unix/syscall_linux_arm.go index 99ae6137..e1913e2c 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_arm.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_arm.go @@ -80,7 +80,7 @@ func Seek(fd int, offset int64, whence int) (newoffset int64, err error) { // 64-bit file system and 32-bit uid calls // (16-bit uid calls are not always supported in newer kernels) -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fchown(fd int, uid int, gid int) (err error) = SYS_FCHOWN32 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go b/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go index 807a0b20..c6de6b91 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go @@ -25,7 +25,7 @@ func EpollCreate(size int) (fd int, err error) { //sysnb Getegid() (egid int) //sysnb Geteuid() (euid int) //sysnb Getgid() (gid int) -//sysnb Getrlimit(resource int, rlim *Rlimit) (err error) +//sysnb getrlimit(resource int, rlim *Rlimit) (err error) //sysnb Getuid() (uid int) //sys Listen(s int, n int) (err error) //sys Pread(fd int, p []byte, offset int64) (n int, err error) = SYS_PREAD64 @@ -47,7 +47,7 @@ func Select(nfd int, r *FdSet, w *FdSet, e *FdSet, timeout *Timeval) (n int, err //sysnb Setregid(rgid int, egid int) (err error) //sysnb Setresgid(rgid int, egid int, sgid int) (err error) //sysnb Setresuid(ruid int, euid int, suid int) (err error) -//sysnb Setrlimit(resource int, rlim *Rlimit) (err error) +//sysnb setrlimit(resource int, rlim *Rlimit) (err error) //sysnb Setreuid(ruid int, euid int) (err error) //sys Shutdown(fd int, how int) (err error) //sys Splice(rfd int, roff *int64, wfd int, woff *int64, len int, flags int) (n int64, err error) @@ -168,6 +168,24 @@ func Pipe2(p []int, flags int) (err error) { return } +// Getrlimit prefers the prlimit64 system call. See issue 38604. +func Getrlimit(resource int, rlim *Rlimit) error { + err := prlimit(0, resource, nil, rlim) + if err != ENOSYS { + return err + } + return getrlimit(resource, rlim) +} + +// Setrlimit prefers the prlimit64 system call. See issue 38604. +func Setrlimit(resource int, rlim *Rlimit) error { + err := prlimit(0, resource, rlim, nil) + if err != ENOSYS { + return err + } + return setrlimit(resource, rlim) +} + func (r *PtraceRegs) PC() uint64 { return r.Pc } func (r *PtraceRegs) SetPC(pc uint64) { r.Pc = pc } @@ -192,9 +210,9 @@ func InotifyInit() (fd int, err error) { return InotifyInit1(0) } -func Dup2(oldfd int, newfd int) (err error) { - return Dup3(oldfd, newfd, 0) -} +// dup2 exists because func Dup3 in syscall_linux.go references +// it in an unreachable path. dup2 isn't available on arm64. +func dup2(oldfd int, newfd int) error func Pause() error { _, err := ppoll(nil, 0, nil, nil) diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go b/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go index af77e6e2..f0287476 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go @@ -7,7 +7,7 @@ package unix -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go b/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go index e286c6ba..c1132811 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go @@ -14,7 +14,7 @@ import ( func Syscall9(trap, a1, a2, a3, a4, a5, a6, a7, a8, a9 uintptr) (r1, r2 uintptr, err syscall.Errno) -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go b/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go index ca0345aa..34937440 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go @@ -7,7 +7,7 @@ package unix -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_riscv64.go b/vendor/golang.org/x/sys/unix/syscall_linux_riscv64.go index abdabbac..b0b15055 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_riscv64.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_riscv64.go @@ -191,10 +191,6 @@ func InotifyInit() (fd int, err error) { return InotifyInit1(0) } -func Dup2(oldfd int, newfd int) (err error) { - return Dup3(oldfd, newfd, 0) -} - func Pause() error { _, err := ppoll(nil, 0, nil, nil) return err @@ -228,3 +224,7 @@ func KexecFileLoad(kernelFd int, initrdFd int, cmdline string, flags int) error } return kexecFileLoad(kernelFd, initrdFd, cmdlineLen, cmdline, flags) } + +// dup2 exists because func Dup3 in syscall_linux.go references +// it in an unreachable path. dup2 isn't available on arm64. +func dup2(oldfd int, newfd int) error diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go b/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go index 533e9305..2363f749 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go @@ -10,7 +10,7 @@ import ( "unsafe" ) -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go b/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go index d890a227..d389f151 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go @@ -8,7 +8,7 @@ package unix //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 -//sys Dup2(oldfd int, newfd int) (err error) +//sys dup2(oldfd int, newfd int) (err error) //sys Fchown(fd int, uid int, gid int) (err error) //sys Fstat(fd int, stat *Stat_t) (err error) //sys Fstatat(dirfd int, path string, stat *Stat_t, flags int) (err error) = SYS_FSTATAT64 diff --git a/vendor/golang.org/x/sys/unix/syscall_unix.go b/vendor/golang.org/x/sys/unix/syscall_unix.go index 8f710d01..400ba9fb 100644 --- a/vendor/golang.org/x/sys/unix/syscall_unix.go +++ b/vendor/golang.org/x/sys/unix/syscall_unix.go @@ -12,6 +12,8 @@ import ( "sync" "syscall" "unsafe" + + "golang.org/x/sys/internal/unsafeheader" ) var ( @@ -113,15 +115,12 @@ func (m *mmapper) Mmap(fd int, offset int64, length int, prot int, flags int) (d return nil, errno } - // Slice memory layout - var sl = struct { - addr uintptr - len int - cap int - }{addr, length, length} - - // Use unsafe to turn sl into a []byte. - b := *(*[]byte)(unsafe.Pointer(&sl)) + // Use unsafe to convert addr into a []byte. + var b []byte + hdr := (*unsafeheader.Slice)(unsafe.Pointer(&b)) + hdr.Data = unsafe.Pointer(addr) + hdr.Cap = length + hdr.Len = length // Register mapping in m and return it. p := &b[cap(b)-1] diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux.go b/vendor/golang.org/x/sys/unix/zerrors_linux.go index 21973940..f8bd50c1 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux.go @@ -160,78 +160,28 @@ const ( BPF_A = 0x10 BPF_ABS = 0x20 BPF_ADD = 0x0 - BPF_ADJ_ROOM_ENCAP_L2_MASK = 0xff - BPF_ADJ_ROOM_ENCAP_L2_SHIFT = 0x38 BPF_ALU = 0x4 BPF_ALU64 = 0x7 BPF_AND = 0x50 - BPF_ANY = 0x0 BPF_ARSH = 0xc0 BPF_B = 0x10 BPF_BUILD_ID_SIZE = 0x14 BPF_CALL = 0x80 - BPF_DEVCG_ACC_MKNOD = 0x1 - BPF_DEVCG_ACC_READ = 0x2 - BPF_DEVCG_ACC_WRITE = 0x4 - BPF_DEVCG_DEV_BLOCK = 0x1 - BPF_DEVCG_DEV_CHAR = 0x2 BPF_DIV = 0x30 BPF_DW = 0x18 BPF_END = 0xd0 - BPF_EXIST = 0x2 BPF_EXIT = 0x90 - BPF_FLOW_DISSECTOR_F_PARSE_1ST_FRAG = 0x1 - BPF_FLOW_DISSECTOR_F_STOP_AT_ENCAP = 0x4 - BPF_FLOW_DISSECTOR_F_STOP_AT_FLOW_LABEL = 0x2 BPF_FROM_BE = 0x8 BPF_FROM_LE = 0x0 BPF_FS_MAGIC = 0xcafe4a11 - BPF_F_ADJ_ROOM_ENCAP_L3_IPV4 = 0x2 - BPF_F_ADJ_ROOM_ENCAP_L3_IPV6 = 0x4 - BPF_F_ADJ_ROOM_ENCAP_L4_GRE = 0x8 - BPF_F_ADJ_ROOM_ENCAP_L4_UDP = 0x10 - BPF_F_ADJ_ROOM_FIXED_GSO = 0x1 BPF_F_ALLOW_MULTI = 0x2 BPF_F_ALLOW_OVERRIDE = 0x1 BPF_F_ANY_ALIGNMENT = 0x2 - BPF_F_CLONE = 0x200 - BPF_F_CTXLEN_MASK = 0xfffff00000000 - BPF_F_CURRENT_CPU = 0xffffffff - BPF_F_CURRENT_NETNS = -0x1 - BPF_F_DONT_FRAGMENT = 0x4 - BPF_F_FAST_STACK_CMP = 0x200 - BPF_F_HDR_FIELD_MASK = 0xf - BPF_F_INDEX_MASK = 0xffffffff - BPF_F_INGRESS = 0x1 - BPF_F_INVALIDATE_HASH = 0x2 - BPF_F_LOCK = 0x4 - BPF_F_MARK_ENFORCE = 0x40 - BPF_F_MARK_MANGLED_0 = 0x20 - BPF_F_MMAPABLE = 0x400 - BPF_F_NO_COMMON_LRU = 0x2 - BPF_F_NO_PREALLOC = 0x1 - BPF_F_NUMA_NODE = 0x4 - BPF_F_PSEUDO_HDR = 0x10 BPF_F_QUERY_EFFECTIVE = 0x1 - BPF_F_RDONLY = 0x8 - BPF_F_RDONLY_PROG = 0x80 - BPF_F_RECOMPUTE_CSUM = 0x1 BPF_F_REPLACE = 0x4 - BPF_F_REUSE_STACKID = 0x400 - BPF_F_SEQ_NUMBER = 0x8 - BPF_F_SKIP_FIELD_MASK = 0xff - BPF_F_STACK_BUILD_ID = 0x20 BPF_F_STRICT_ALIGNMENT = 0x1 - BPF_F_SYSCTL_BASE_NAME = 0x1 BPF_F_TEST_RND_HI32 = 0x4 BPF_F_TEST_STATE_FREQ = 0x8 - BPF_F_TUNINFO_IPV6 = 0x1 - BPF_F_USER_BUILD_ID = 0x800 - BPF_F_USER_STACK = 0x100 - BPF_F_WRONLY = 0x10 - BPF_F_WRONLY_PROG = 0x100 - BPF_F_ZERO_CSUM_TX = 0x2 - BPF_F_ZERO_SEED = 0x40 BPF_H = 0x8 BPF_IMM = 0x0 BPF_IND = 0x40 @@ -267,7 +217,6 @@ const ( BPF_MUL = 0x20 BPF_NEG = 0x80 BPF_NET_OFF = -0x100000 - BPF_NOEXIST = 0x1 BPF_OBJ_NAME_LEN = 0x10 BPF_OR = 0x40 BPF_PSEUDO_CALL = 0x1 @@ -275,12 +224,6 @@ const ( BPF_PSEUDO_MAP_VALUE = 0x2 BPF_RET = 0x6 BPF_RSH = 0x70 - BPF_SK_STORAGE_GET_F_CREATE = 0x1 - BPF_SOCK_OPS_ALL_CB_FLAGS = 0xf - BPF_SOCK_OPS_RETRANS_CB_FLAG = 0x2 - BPF_SOCK_OPS_RTO_CB_FLAG = 0x1 - BPF_SOCK_OPS_RTT_CB_FLAG = 0x8 - BPF_SOCK_OPS_STATE_CB_FLAG = 0x4 BPF_ST = 0x2 BPF_STX = 0x3 BPF_SUB = 0x10 @@ -378,12 +321,14 @@ const ( CLOCK_TXINT = 0x3 CLONE_ARGS_SIZE_VER0 = 0x40 CLONE_ARGS_SIZE_VER1 = 0x50 + CLONE_ARGS_SIZE_VER2 = 0x58 CLONE_CHILD_CLEARTID = 0x200000 CLONE_CHILD_SETTID = 0x1000000 CLONE_CLEAR_SIGHAND = 0x100000000 CLONE_DETACHED = 0x400000 CLONE_FILES = 0x400 CLONE_FS = 0x200 + CLONE_INTO_CGROUP = 0x200000000 CLONE_IO = 0x80000000 CLONE_NEWCGROUP = 0x2000000 CLONE_NEWIPC = 0x8000000 @@ -598,7 +543,9 @@ const ( FAN_DELETE = 0x200 FAN_DELETE_SELF = 0x400 FAN_DENY = 0x2 + FAN_DIR_MODIFY = 0x80000 FAN_ENABLE_AUDIT = 0x40 + FAN_EVENT_INFO_TYPE_DFID_NAME = 0x2 FAN_EVENT_INFO_TYPE_FID = 0x1 FAN_EVENT_METADATA_LEN = 0x18 FAN_EVENT_ON_CHILD = 0x8000000 @@ -2108,8 +2055,6 @@ const ( TCOFLUSH = 0x1 TCOOFF = 0x0 TCOON = 0x1 - TCP_BPF_IW = 0x3e9 - TCP_BPF_SNDCWND_CLAMP = 0x3ea TCP_CC_INFO = 0x1a TCP_CM_INQ = 0x24 TCP_CONGESTION = 0xd @@ -2165,6 +2110,8 @@ const ( TCP_USER_TIMEOUT = 0x12 TCP_WINDOW_CLAMP = 0xa TCP_ZEROCOPY_RECEIVE = 0x23 + TFD_TIMER_ABSTIME = 0x1 + TFD_TIMER_CANCEL_ON_SET = 0x2 TIMER_ABSTIME = 0x1 TIOCM_DTR = 0x2 TIOCM_LE = 0x1 @@ -2382,8 +2329,9 @@ const ( XDP_COPY = 0x2 XDP_FLAGS_DRV_MODE = 0x4 XDP_FLAGS_HW_MODE = 0x8 - XDP_FLAGS_MASK = 0xf + XDP_FLAGS_MASK = 0x1f XDP_FLAGS_MODES = 0xe + XDP_FLAGS_REPLACE = 0x10 XDP_FLAGS_SKB_MODE = 0x2 XDP_FLAGS_UPDATE_IF_NOEXIST = 0x1 XDP_MMAP_OFFSETS = 0x1 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_386.go b/vendor/golang.org/x/sys/unix/zerrors_linux_386.go index 028c9d87..8d207b04 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_386.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_386.go @@ -75,6 +75,7 @@ const ( FP_XSTATE_MAGIC2 = 0x46505845 FS_IOC_ENABLE_VERITY = 0x40806685 FS_IOC_GETFLAGS = 0x80046601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x8010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x400c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x40106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x800c6613 @@ -342,6 +343,8 @@ const ( TCSETXF = 0x5434 TCSETXW = 0x5435 TCXONC = 0x540a + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go index 005970f7..c4bf9cb8 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go @@ -75,6 +75,7 @@ const ( FP_XSTATE_MAGIC2 = 0x46505845 FS_IOC_ENABLE_VERITY = 0x40806685 FS_IOC_GETFLAGS = 0x80086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x8010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x400c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x40106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x800c6613 @@ -343,6 +344,8 @@ const ( TCSETXF = 0x5434 TCSETXW = 0x5435 TCXONC = 0x540a + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go b/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go index 0541f36e..0cab0522 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x1000 FS_IOC_ENABLE_VERITY = 0x40806685 FS_IOC_GETFLAGS = 0x80046601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x8010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x400c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x40106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x800c6613 @@ -349,6 +350,8 @@ const ( TCSETXF = 0x5434 TCSETXW = 0x5435 TCXONC = 0x540a + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go index 9ee8d1bc..370d0a7f 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go @@ -77,6 +77,7 @@ const ( FPSIMD_MAGIC = 0x46508001 FS_IOC_ENABLE_VERITY = 0x40806685 FS_IOC_GETFLAGS = 0x80086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x8010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x400c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x40106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x800c6613 @@ -336,6 +337,8 @@ const ( TCSETXF = 0x5434 TCSETXW = 0x5435 TCXONC = 0x540a + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go index 4826bd70..fbf2f317 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x2000 FS_IOC_ENABLE_VERITY = 0x80806685 FS_IOC_GETFLAGS = 0x40046601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x4010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x800c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x80106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x400c6613 @@ -339,6 +340,8 @@ const ( TCSETSW = 0x540f TCSETSW2 = 0x8030542c TCXONC = 0x5406 + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x80 TIOCCBRK = 0x5428 TIOCCONS = 0x80047478 TIOCEXCL = 0x740d diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go index 2346dc55..25e74b30 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x2000 FS_IOC_ENABLE_VERITY = 0x80806685 FS_IOC_GETFLAGS = 0x40086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x4010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x800c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x80106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x400c6613 @@ -339,6 +340,8 @@ const ( TCSETSW = 0x540f TCSETSW2 = 0x8030542c TCXONC = 0x5406 + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x80 TIOCCBRK = 0x5428 TIOCCONS = 0x80047478 TIOCEXCL = 0x740d diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go index e758b61e..4ecc0bca 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x2000 FS_IOC_ENABLE_VERITY = 0x80806685 FS_IOC_GETFLAGS = 0x40086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x4010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x800c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x80106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x400c6613 @@ -339,6 +340,8 @@ const ( TCSETSW = 0x540f TCSETSW2 = 0x8030542c TCXONC = 0x5406 + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x80 TIOCCBRK = 0x5428 TIOCCONS = 0x80047478 TIOCEXCL = 0x740d diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go index 2dfe6bba..dfb8f88a 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x2000 FS_IOC_ENABLE_VERITY = 0x80806685 FS_IOC_GETFLAGS = 0x40046601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x4010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x800c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x80106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x400c6613 @@ -339,6 +340,8 @@ const ( TCSETSW = 0x540f TCSETSW2 = 0x8030542c TCXONC = 0x5406 + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x80 TIOCCBRK = 0x5428 TIOCCONS = 0x80047478 TIOCEXCL = 0x740d diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go index 51858667..72d8dad5 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x800000 FS_IOC_ENABLE_VERITY = 0x80806685 FS_IOC_GETFLAGS = 0x40086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x4010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x800c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x80106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x400c6613 @@ -393,6 +394,8 @@ const ( TCSETSF = 0x802c7416 TCSETSW = 0x802c7415 TCXONC = 0x2000741e + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go index 4231b20b..ca0e7b52 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x800000 FS_IOC_ENABLE_VERITY = 0x80806685 FS_IOC_GETFLAGS = 0x40086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x4010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x800c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x80106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x400c6613 @@ -393,6 +394,8 @@ const ( TCSETSF = 0x802c7416 TCSETSW = 0x802c7415 TCXONC = 0x2000741e + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_riscv64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_riscv64.go index 6a0b2d29..147511a9 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_riscv64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_riscv64.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x1000 FS_IOC_ENABLE_VERITY = 0x40806685 FS_IOC_GETFLAGS = 0x80086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x8010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x400c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x40106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x800c6613 @@ -330,6 +331,8 @@ const ( TCSETXF = 0x5434 TCSETXW = 0x5435 TCXONC = 0x540a + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go b/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go index 95e950fc..517349da 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go @@ -74,6 +74,7 @@ const ( FLUSHO = 0x1000 FS_IOC_ENABLE_VERITY = 0x40806685 FS_IOC_GETFLAGS = 0x80086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x8010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x400c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x40106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x800c6613 @@ -403,6 +404,8 @@ const ( TCSETXF = 0x5434 TCSETXW = 0x5435 TCXONC = 0x540a + TFD_CLOEXEC = 0x80000 + TFD_NONBLOCK = 0x800 TIOCCBRK = 0x5428 TIOCCONS = 0x541d TIOCEXCL = 0x540c diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go index 079762fa..09482246 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go @@ -78,6 +78,7 @@ const ( FLUSHO = 0x1000 FS_IOC_ENABLE_VERITY = 0x80806685 FS_IOC_GETFLAGS = 0x40086601 + FS_IOC_GET_ENCRYPTION_NONCE = 0x4010661b FS_IOC_GET_ENCRYPTION_POLICY = 0x800c6615 FS_IOC_GET_ENCRYPTION_PWSALT = 0x80106614 FS_IOC_SET_ENCRYPTION_POLICY = 0x400c6613 @@ -392,6 +393,8 @@ const ( TCSETSW = 0x8024540a TCSETSW2 = 0x802c540e TCXONC = 0x20005406 + TFD_CLOEXEC = 0x400000 + TFD_NONBLOCK = 0x4000 TIOCCBRK = 0x2000747a TIOCCONS = 0x20007424 TIOCEXCL = 0x2000740d diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.1_11.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.1_11.go index c1cc0a41..23e94d36 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.1_11.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.1_11.go @@ -966,6 +966,16 @@ func Getsid(pid int) (sid int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := RawSyscall(SYS_GETUID, 0, 0, 0) uid = int(r0) @@ -1709,18 +1719,6 @@ func ptrace(request int, pid int, addr uintptr, data uintptr) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int32, usec int32, err error) { - r0, r1, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int32(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := Syscall(SYS_FSTAT64, uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go index a3fc4900..e2ffb3be 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go @@ -1376,6 +1376,21 @@ func libc_getsid_trampoline() // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +func libc_gettimeofday_trampoline() + +//go:linkname libc_gettimeofday libc_gettimeofday +//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := syscall_rawSyscall(funcPC(libc_getuid_trampoline), 0, 0, 0) uid = int(r0) @@ -2357,23 +2372,6 @@ func libc_ptrace_trampoline() // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int32, usec int32, err error) { - r0, r1, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int32(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -func libc_gettimeofday_trampoline() - -//go:linkname libc_gettimeofday libc_gettimeofday -//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := syscall_syscall(funcPC(libc_fstat64_trampoline), uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.1_11.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.1_11.go index f8e5c37c..10256173 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.1_11.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.1_11.go @@ -966,6 +966,16 @@ func Getsid(pid int) (sid int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := RawSyscall(SYS_GETUID, 0, 0, 0) uid = int(r0) @@ -1709,18 +1719,6 @@ func ptrace(request int, pid int, addr uintptr, data uintptr) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int64, usec int32, err error) { - r0, r1, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int64(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := Syscall(SYS_FSTAT64, uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go index 50d6437e..c67e336e 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go @@ -1376,6 +1376,21 @@ func libc_getsid_trampoline() // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +func libc_gettimeofday_trampoline() + +//go:linkname libc_gettimeofday libc_gettimeofday +//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := syscall_rawSyscall(funcPC(libc_getuid_trampoline), 0, 0, 0) uid = int(r0) @@ -2357,23 +2372,6 @@ func libc_ptrace_trampoline() // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int64, usec int32, err error) { - r0, r1, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int64(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -func libc_gettimeofday_trampoline() - -//go:linkname libc_gettimeofday libc_gettimeofday -//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := syscall_syscall(funcPC(libc_fstat64_trampoline), uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.1_11.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.1_11.go index cea04e04..d34e6df2 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.1_11.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.1_11.go @@ -966,6 +966,16 @@ func Getsid(pid int) (sid int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := RawSyscall(SYS_GETUID, 0, 0, 0) uid = int(r0) @@ -1682,18 +1692,6 @@ func writelen(fd int, buf *byte, nbuf int) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int32, usec int32, err error) { - r0, r1, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int32(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := Syscall(SYS_FSTAT, uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go index 63103950..b759757a 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go @@ -1376,6 +1376,21 @@ func libc_getsid_trampoline() // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +func libc_gettimeofday_trampoline() + +//go:linkname libc_gettimeofday libc_gettimeofday +//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := syscall_rawSyscall(funcPC(libc_getuid_trampoline), 0, 0, 0) uid = int(r0) @@ -2342,23 +2357,6 @@ func writelen(fd int, buf *byte, nbuf int) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int32, usec int32, err error) { - r0, r1, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int32(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -func libc_gettimeofday_trampoline() - -//go:linkname libc_gettimeofday libc_gettimeofday -//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := syscall_syscall(funcPC(libc_fstat_trampoline), uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.1_11.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.1_11.go index 8c3bb3a2..8d39a09f 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.1_11.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.1_11.go @@ -966,6 +966,16 @@ func Getsid(pid int) (sid int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := RawSyscall(SYS_GETUID, 0, 0, 0) uid = int(r0) @@ -1682,18 +1692,6 @@ func writelen(fd int, buf *byte, nbuf int) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int64, usec int32, err error) { - r0, r1, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int64(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := Syscall(SYS_FSTAT, uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go index a8709f72..b2886126 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go @@ -1376,6 +1376,21 @@ func libc_getsid_trampoline() // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Gettimeofday(tp *Timeval) (err error) { + _, _, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +func libc_gettimeofday_trampoline() + +//go:linkname libc_gettimeofday libc_gettimeofday +//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Getuid() (uid int) { r0, _, _ := syscall_rawSyscall(funcPC(libc_getuid_trampoline), 0, 0, 0) uid = int(r0) @@ -2342,23 +2357,6 @@ func writelen(fd int, buf *byte, nbuf int) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func gettimeofday(tp *Timeval) (sec int64, usec int32, err error) { - r0, r1, e1 := syscall_rawSyscall(funcPC(libc_gettimeofday_trampoline), uintptr(unsafe.Pointer(tp)), 0, 0) - sec = int64(r0) - usec = int32(r1) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -func libc_gettimeofday_trampoline() - -//go:linkname libc_gettimeofday libc_gettimeofday -//go:cgo_import_dynamic libc_gettimeofday gettimeofday "/usr/lib/libSystem.B.dylib" - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fstat(fd int, stat *Stat_t) (err error) { _, _, e1 := syscall_syscall(funcPC(libc_fstat_trampoline), uintptr(fd), uintptr(unsafe.Pointer(stat)), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux.go b/vendor/golang.org/x/sys/unix/zsyscall_linux.go index 5c5b7422..df217825 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux.go @@ -83,22 +83,6 @@ func openat(dirfd int, path string, flags int, mode uint32) (fd int, err error) // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func openat2(dirfd int, path string, open_how *OpenHow, size int) (fd int, err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - r0, _, e1 := Syscall6(SYS_OPENAT2, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(open_how)), uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func ppoll(fds *PollFd, nfds int, timeout *Timespec, sigmask *Sigset_t) (n int, err error) { r0, _, e1 := Syscall6(SYS_PPOLL, uintptr(unsafe.Pointer(fds)), uintptr(nfds), uintptr(unsafe.Pointer(timeout)), uintptr(unsafe.Pointer(sigmask)), 0, 0) n = int(r0) @@ -1466,6 +1450,37 @@ func Sysinfo(info *Sysinfo_t) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func TimerfdCreate(clockid int, flags int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_TIMERFD_CREATE, uintptr(clockid), uintptr(flags), 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func TimerfdGettime(fd int, currValue *ItimerSpec) (err error) { + _, _, e1 := RawSyscall(SYS_TIMERFD_GETTIME, uintptr(fd), uintptr(unsafe.Pointer(currValue)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func TimerfdSettime(fd int, flags int, newValue *ItimerSpec, oldValue *ItimerSpec) (err error) { + _, _, e1 := RawSyscall6(SYS_TIMERFD_SETTIME, uintptr(fd), uintptr(flags), uintptr(unsafe.Pointer(newValue)), uintptr(unsafe.Pointer(oldValue)), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Tgkill(tgid int, tid int, sig syscall.Signal) (err error) { _, _, e1 := RawSyscall(SYS_TGKILL, uintptr(tgid), uintptr(tid), uintptr(sig)) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go index ba63af7b..19ebd3ff 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go @@ -55,7 +55,7 @@ func pipe(p *[2]_C_int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go index f64adef4..5c562182 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go index ac19523e..dc69d99c 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go @@ -234,7 +234,7 @@ func sendmsg(s int, msg *Msghdr, flags int) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go index f0d2890b..1b897dee 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go @@ -151,7 +151,7 @@ func Getgid() (gid int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Getrlimit(resource int, rlim *Rlimit) (err error) { +func getrlimit(resource int, rlim *Rlimit) (err error) { _, _, e1 := RawSyscall(SYS_GETRLIMIT, uintptr(resource), uintptr(unsafe.Pointer(rlim)), 0) if e1 != 0 { err = errnoErr(e1) @@ -307,7 +307,7 @@ func Setresuid(ruid int, euid int, suid int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Setrlimit(resource int, rlim *Rlimit) (err error) { +func setrlimit(resource int, rlim *Rlimit) (err error) { _, _, e1 := RawSyscall(SYS_SETRLIMIT, uintptr(resource), uintptr(unsafe.Pointer(rlim)), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go index aecbbca7..49186843 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go index 424fb7fb..9171d3bd 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go index 28c7239c..82286f04 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go index 84596b30..15920621 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go index de022639..73a42e2c 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go index 888f21d3..6b855953 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go index 9bc353f0..d7032ab1 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go @@ -45,7 +45,7 @@ func Tee(rfd int, wfd int, len int, flags int) (n int64, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go index 854e816d..bcbbdd90 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go @@ -72,7 +72,7 @@ func Fadvise(fd int, offset int64, length int64, advice int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { +func dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { err = errnoErr(e1) diff --git a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go index 37dcc74c..102f1ab4 100644 --- a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go +++ b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go @@ -1,4 +1,4 @@ -// mksysctl_openbsd.pl +// go run mksysctl_openbsd.go // Code generated by the command above; DO NOT EDIT. // +build 386,openbsd @@ -30,6 +30,7 @@ var sysctlMib = []mibentry{ {"hw.model", []_C_int{6, 2}}, {"hw.ncpu", []_C_int{6, 3}}, {"hw.ncpufound", []_C_int{6, 21}}, + {"hw.ncpuonline", []_C_int{6, 25}}, {"hw.pagesize", []_C_int{6, 7}}, {"hw.physmem", []_C_int{6, 19}}, {"hw.product", []_C_int{6, 15}}, diff --git a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go index fe6caa6e..4866fced 100644 --- a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go @@ -31,6 +31,7 @@ var sysctlMib = []mibentry{ {"hw.model", []_C_int{6, 2}}, {"hw.ncpu", []_C_int{6, 3}}, {"hw.ncpufound", []_C_int{6, 21}}, + {"hw.ncpuonline", []_C_int{6, 25}}, {"hw.pagesize", []_C_int{6, 7}}, {"hw.perfpolicy", []_C_int{6, 23}}, {"hw.physmem", []_C_int{6, 19}}, diff --git a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go index 6eb8c0b0..d3801eb2 100644 --- a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go +++ b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go @@ -30,6 +30,7 @@ var sysctlMib = []mibentry{ {"hw.model", []_C_int{6, 2}}, {"hw.ncpu", []_C_int{6, 3}}, {"hw.ncpufound", []_C_int{6, 21}}, + {"hw.ncpuonline", []_C_int{6, 25}}, {"hw.pagesize", []_C_int{6, 7}}, {"hw.physmem", []_C_int{6, 19}}, {"hw.product", []_C_int{6, 15}}, diff --git a/vendor/golang.org/x/sys/unix/ztypes_freebsd_arm.go b/vendor/golang.org/x/sys/unix/ztypes_freebsd_arm.go index 6f79227d..b91c2ae0 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_freebsd_arm.go +++ b/vendor/golang.org/x/sys/unix/ztypes_freebsd_arm.go @@ -125,9 +125,9 @@ type Statfs_t struct { Owner uint32 Fsid Fsid Charspare [80]int8 - Fstypename [16]int8 - Mntfromname [1024]int8 - Mntonname [1024]int8 + Fstypename [16]byte + Mntfromname [1024]byte + Mntonname [1024]byte } type statfs_freebsd11_t struct { @@ -150,9 +150,9 @@ type statfs_freebsd11_t struct { Owner uint32 Fsid Fsid Charspare [80]int8 - Fstypename [16]int8 - Mntfromname [88]int8 - Mntonname [88]int8 + Fstypename [16]byte + Mntfromname [88]byte + Mntonname [88]byte } type Flock_t struct { diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux.go b/vendor/golang.org/x/sys/unix/ztypes_linux.go index 1620ad06..27d67ac8 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux.go @@ -18,6 +18,11 @@ type ( _C_long_long int64 ) +type ItimerSpec struct { + Interval Timespec + Value Timespec +} + const ( TIME_OK = 0x0 TIME_INS = 0x1 @@ -691,22 +696,6 @@ const ( AT_EACCESS = 0x200 ) -type OpenHow struct { - Flags uint64 - Mode uint64 - Resolve uint64 -} - -const SizeofOpenHow = 0x18 - -const ( - RESOLVE_BENEATH = 0x8 - RESOLVE_IN_ROOT = 0x10 - RESOLVE_NO_MAGICLINKS = 0x2 - RESOLVE_NO_SYMLINKS = 0x4 - RESOLVE_NO_XDEV = 0x1 -) - type PollFd struct { Fd int32 Events int16 @@ -1882,175 +1871,249 @@ const ( ) const ( - BPF_REG_0 = 0x0 - BPF_REG_1 = 0x1 - BPF_REG_2 = 0x2 - BPF_REG_3 = 0x3 - BPF_REG_4 = 0x4 - BPF_REG_5 = 0x5 - BPF_REG_6 = 0x6 - BPF_REG_7 = 0x7 - BPF_REG_8 = 0x8 - BPF_REG_9 = 0x9 - BPF_REG_10 = 0xa - BPF_MAP_CREATE = 0x0 - BPF_MAP_LOOKUP_ELEM = 0x1 - BPF_MAP_UPDATE_ELEM = 0x2 - BPF_MAP_DELETE_ELEM = 0x3 - BPF_MAP_GET_NEXT_KEY = 0x4 - BPF_PROG_LOAD = 0x5 - BPF_OBJ_PIN = 0x6 - BPF_OBJ_GET = 0x7 - BPF_PROG_ATTACH = 0x8 - BPF_PROG_DETACH = 0x9 - BPF_PROG_TEST_RUN = 0xa - BPF_PROG_GET_NEXT_ID = 0xb - BPF_MAP_GET_NEXT_ID = 0xc - BPF_PROG_GET_FD_BY_ID = 0xd - BPF_MAP_GET_FD_BY_ID = 0xe - BPF_OBJ_GET_INFO_BY_FD = 0xf - BPF_PROG_QUERY = 0x10 - BPF_RAW_TRACEPOINT_OPEN = 0x11 - BPF_BTF_LOAD = 0x12 - BPF_BTF_GET_FD_BY_ID = 0x13 - BPF_TASK_FD_QUERY = 0x14 - BPF_MAP_LOOKUP_AND_DELETE_ELEM = 0x15 - BPF_MAP_FREEZE = 0x16 - BPF_BTF_GET_NEXT_ID = 0x17 - BPF_MAP_TYPE_UNSPEC = 0x0 - BPF_MAP_TYPE_HASH = 0x1 - BPF_MAP_TYPE_ARRAY = 0x2 - BPF_MAP_TYPE_PROG_ARRAY = 0x3 - BPF_MAP_TYPE_PERF_EVENT_ARRAY = 0x4 - BPF_MAP_TYPE_PERCPU_HASH = 0x5 - BPF_MAP_TYPE_PERCPU_ARRAY = 0x6 - BPF_MAP_TYPE_STACK_TRACE = 0x7 - BPF_MAP_TYPE_CGROUP_ARRAY = 0x8 - BPF_MAP_TYPE_LRU_HASH = 0x9 - BPF_MAP_TYPE_LRU_PERCPU_HASH = 0xa - BPF_MAP_TYPE_LPM_TRIE = 0xb - BPF_MAP_TYPE_ARRAY_OF_MAPS = 0xc - BPF_MAP_TYPE_HASH_OF_MAPS = 0xd - BPF_MAP_TYPE_DEVMAP = 0xe - BPF_MAP_TYPE_SOCKMAP = 0xf - BPF_MAP_TYPE_CPUMAP = 0x10 - BPF_MAP_TYPE_XSKMAP = 0x11 - BPF_MAP_TYPE_SOCKHASH = 0x12 - BPF_MAP_TYPE_CGROUP_STORAGE = 0x13 - BPF_MAP_TYPE_REUSEPORT_SOCKARRAY = 0x14 - BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE = 0x15 - BPF_MAP_TYPE_QUEUE = 0x16 - BPF_MAP_TYPE_STACK = 0x17 - BPF_MAP_TYPE_SK_STORAGE = 0x18 - BPF_MAP_TYPE_DEVMAP_HASH = 0x19 - BPF_PROG_TYPE_UNSPEC = 0x0 - BPF_PROG_TYPE_SOCKET_FILTER = 0x1 - BPF_PROG_TYPE_KPROBE = 0x2 - BPF_PROG_TYPE_SCHED_CLS = 0x3 - BPF_PROG_TYPE_SCHED_ACT = 0x4 - BPF_PROG_TYPE_TRACEPOINT = 0x5 - BPF_PROG_TYPE_XDP = 0x6 - BPF_PROG_TYPE_PERF_EVENT = 0x7 - BPF_PROG_TYPE_CGROUP_SKB = 0x8 - BPF_PROG_TYPE_CGROUP_SOCK = 0x9 - BPF_PROG_TYPE_LWT_IN = 0xa - BPF_PROG_TYPE_LWT_OUT = 0xb - BPF_PROG_TYPE_LWT_XMIT = 0xc - BPF_PROG_TYPE_SOCK_OPS = 0xd - BPF_PROG_TYPE_SK_SKB = 0xe - BPF_PROG_TYPE_CGROUP_DEVICE = 0xf - BPF_PROG_TYPE_SK_MSG = 0x10 - BPF_PROG_TYPE_RAW_TRACEPOINT = 0x11 - BPF_PROG_TYPE_CGROUP_SOCK_ADDR = 0x12 - BPF_PROG_TYPE_LWT_SEG6LOCAL = 0x13 - BPF_PROG_TYPE_LIRC_MODE2 = 0x14 - BPF_PROG_TYPE_SK_REUSEPORT = 0x15 - BPF_PROG_TYPE_FLOW_DISSECTOR = 0x16 - BPF_PROG_TYPE_CGROUP_SYSCTL = 0x17 - BPF_PROG_TYPE_RAW_TRACEPOINT_WRITABLE = 0x18 - BPF_PROG_TYPE_CGROUP_SOCKOPT = 0x19 - BPF_PROG_TYPE_TRACING = 0x1a - BPF_CGROUP_INET_INGRESS = 0x0 - BPF_CGROUP_INET_EGRESS = 0x1 - BPF_CGROUP_INET_SOCK_CREATE = 0x2 - BPF_CGROUP_SOCK_OPS = 0x3 - BPF_SK_SKB_STREAM_PARSER = 0x4 - BPF_SK_SKB_STREAM_VERDICT = 0x5 - BPF_CGROUP_DEVICE = 0x6 - BPF_SK_MSG_VERDICT = 0x7 - BPF_CGROUP_INET4_BIND = 0x8 - BPF_CGROUP_INET6_BIND = 0x9 - BPF_CGROUP_INET4_CONNECT = 0xa - BPF_CGROUP_INET6_CONNECT = 0xb - BPF_CGROUP_INET4_POST_BIND = 0xc - BPF_CGROUP_INET6_POST_BIND = 0xd - BPF_CGROUP_UDP4_SENDMSG = 0xe - BPF_CGROUP_UDP6_SENDMSG = 0xf - BPF_LIRC_MODE2 = 0x10 - BPF_FLOW_DISSECTOR = 0x11 - BPF_CGROUP_SYSCTL = 0x12 - BPF_CGROUP_UDP4_RECVMSG = 0x13 - BPF_CGROUP_UDP6_RECVMSG = 0x14 - BPF_CGROUP_GETSOCKOPT = 0x15 - BPF_CGROUP_SETSOCKOPT = 0x16 - BPF_TRACE_RAW_TP = 0x17 - BPF_TRACE_FENTRY = 0x18 - BPF_TRACE_FEXIT = 0x19 - BPF_STACK_BUILD_ID_EMPTY = 0x0 - BPF_STACK_BUILD_ID_VALID = 0x1 - BPF_STACK_BUILD_ID_IP = 0x2 - BPF_ADJ_ROOM_NET = 0x0 - BPF_ADJ_ROOM_MAC = 0x1 - BPF_HDR_START_MAC = 0x0 - BPF_HDR_START_NET = 0x1 - BPF_LWT_ENCAP_SEG6 = 0x0 - BPF_LWT_ENCAP_SEG6_INLINE = 0x1 - BPF_LWT_ENCAP_IP = 0x2 - BPF_OK = 0x0 - BPF_DROP = 0x2 - BPF_REDIRECT = 0x7 - BPF_LWT_REROUTE = 0x80 - BPF_SOCK_OPS_VOID = 0x0 - BPF_SOCK_OPS_TIMEOUT_INIT = 0x1 - BPF_SOCK_OPS_RWND_INIT = 0x2 - BPF_SOCK_OPS_TCP_CONNECT_CB = 0x3 - BPF_SOCK_OPS_ACTIVE_ESTABLISHED_CB = 0x4 - BPF_SOCK_OPS_PASSIVE_ESTABLISHED_CB = 0x5 - BPF_SOCK_OPS_NEEDS_ECN = 0x6 - BPF_SOCK_OPS_BASE_RTT = 0x7 - BPF_SOCK_OPS_RTO_CB = 0x8 - BPF_SOCK_OPS_RETRANS_CB = 0x9 - BPF_SOCK_OPS_STATE_CB = 0xa - BPF_SOCK_OPS_TCP_LISTEN_CB = 0xb - BPF_SOCK_OPS_RTT_CB = 0xc - BPF_TCP_ESTABLISHED = 0x1 - BPF_TCP_SYN_SENT = 0x2 - BPF_TCP_SYN_RECV = 0x3 - BPF_TCP_FIN_WAIT1 = 0x4 - BPF_TCP_FIN_WAIT2 = 0x5 - BPF_TCP_TIME_WAIT = 0x6 - BPF_TCP_CLOSE = 0x7 - BPF_TCP_CLOSE_WAIT = 0x8 - BPF_TCP_LAST_ACK = 0x9 - BPF_TCP_LISTEN = 0xa - BPF_TCP_CLOSING = 0xb - BPF_TCP_NEW_SYN_RECV = 0xc - BPF_TCP_MAX_STATES = 0xd - BPF_FIB_LKUP_RET_SUCCESS = 0x0 - BPF_FIB_LKUP_RET_BLACKHOLE = 0x1 - BPF_FIB_LKUP_RET_UNREACHABLE = 0x2 - BPF_FIB_LKUP_RET_PROHIBIT = 0x3 - BPF_FIB_LKUP_RET_NOT_FWDED = 0x4 - BPF_FIB_LKUP_RET_FWD_DISABLED = 0x5 - BPF_FIB_LKUP_RET_UNSUPP_LWT = 0x6 - BPF_FIB_LKUP_RET_NO_NEIGH = 0x7 - BPF_FIB_LKUP_RET_FRAG_NEEDED = 0x8 - BPF_FD_TYPE_RAW_TRACEPOINT = 0x0 - BPF_FD_TYPE_TRACEPOINT = 0x1 - BPF_FD_TYPE_KPROBE = 0x2 - BPF_FD_TYPE_KRETPROBE = 0x3 - BPF_FD_TYPE_UPROBE = 0x4 - BPF_FD_TYPE_URETPROBE = 0x5 + BPF_REG_0 = 0x0 + BPF_REG_1 = 0x1 + BPF_REG_2 = 0x2 + BPF_REG_3 = 0x3 + BPF_REG_4 = 0x4 + BPF_REG_5 = 0x5 + BPF_REG_6 = 0x6 + BPF_REG_7 = 0x7 + BPF_REG_8 = 0x8 + BPF_REG_9 = 0x9 + BPF_REG_10 = 0xa + BPF_MAP_CREATE = 0x0 + BPF_MAP_LOOKUP_ELEM = 0x1 + BPF_MAP_UPDATE_ELEM = 0x2 + BPF_MAP_DELETE_ELEM = 0x3 + BPF_MAP_GET_NEXT_KEY = 0x4 + BPF_PROG_LOAD = 0x5 + BPF_OBJ_PIN = 0x6 + BPF_OBJ_GET = 0x7 + BPF_PROG_ATTACH = 0x8 + BPF_PROG_DETACH = 0x9 + BPF_PROG_TEST_RUN = 0xa + BPF_PROG_GET_NEXT_ID = 0xb + BPF_MAP_GET_NEXT_ID = 0xc + BPF_PROG_GET_FD_BY_ID = 0xd + BPF_MAP_GET_FD_BY_ID = 0xe + BPF_OBJ_GET_INFO_BY_FD = 0xf + BPF_PROG_QUERY = 0x10 + BPF_RAW_TRACEPOINT_OPEN = 0x11 + BPF_BTF_LOAD = 0x12 + BPF_BTF_GET_FD_BY_ID = 0x13 + BPF_TASK_FD_QUERY = 0x14 + BPF_MAP_LOOKUP_AND_DELETE_ELEM = 0x15 + BPF_MAP_FREEZE = 0x16 + BPF_BTF_GET_NEXT_ID = 0x17 + BPF_MAP_LOOKUP_BATCH = 0x18 + BPF_MAP_LOOKUP_AND_DELETE_BATCH = 0x19 + BPF_MAP_UPDATE_BATCH = 0x1a + BPF_MAP_DELETE_BATCH = 0x1b + BPF_LINK_CREATE = 0x1c + BPF_LINK_UPDATE = 0x1d + BPF_MAP_TYPE_UNSPEC = 0x0 + BPF_MAP_TYPE_HASH = 0x1 + BPF_MAP_TYPE_ARRAY = 0x2 + BPF_MAP_TYPE_PROG_ARRAY = 0x3 + BPF_MAP_TYPE_PERF_EVENT_ARRAY = 0x4 + BPF_MAP_TYPE_PERCPU_HASH = 0x5 + BPF_MAP_TYPE_PERCPU_ARRAY = 0x6 + BPF_MAP_TYPE_STACK_TRACE = 0x7 + BPF_MAP_TYPE_CGROUP_ARRAY = 0x8 + BPF_MAP_TYPE_LRU_HASH = 0x9 + BPF_MAP_TYPE_LRU_PERCPU_HASH = 0xa + BPF_MAP_TYPE_LPM_TRIE = 0xb + BPF_MAP_TYPE_ARRAY_OF_MAPS = 0xc + BPF_MAP_TYPE_HASH_OF_MAPS = 0xd + BPF_MAP_TYPE_DEVMAP = 0xe + BPF_MAP_TYPE_SOCKMAP = 0xf + BPF_MAP_TYPE_CPUMAP = 0x10 + BPF_MAP_TYPE_XSKMAP = 0x11 + BPF_MAP_TYPE_SOCKHASH = 0x12 + BPF_MAP_TYPE_CGROUP_STORAGE = 0x13 + BPF_MAP_TYPE_REUSEPORT_SOCKARRAY = 0x14 + BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE = 0x15 + BPF_MAP_TYPE_QUEUE = 0x16 + BPF_MAP_TYPE_STACK = 0x17 + BPF_MAP_TYPE_SK_STORAGE = 0x18 + BPF_MAP_TYPE_DEVMAP_HASH = 0x19 + BPF_MAP_TYPE_STRUCT_OPS = 0x1a + BPF_PROG_TYPE_UNSPEC = 0x0 + BPF_PROG_TYPE_SOCKET_FILTER = 0x1 + BPF_PROG_TYPE_KPROBE = 0x2 + BPF_PROG_TYPE_SCHED_CLS = 0x3 + BPF_PROG_TYPE_SCHED_ACT = 0x4 + BPF_PROG_TYPE_TRACEPOINT = 0x5 + BPF_PROG_TYPE_XDP = 0x6 + BPF_PROG_TYPE_PERF_EVENT = 0x7 + BPF_PROG_TYPE_CGROUP_SKB = 0x8 + BPF_PROG_TYPE_CGROUP_SOCK = 0x9 + BPF_PROG_TYPE_LWT_IN = 0xa + BPF_PROG_TYPE_LWT_OUT = 0xb + BPF_PROG_TYPE_LWT_XMIT = 0xc + BPF_PROG_TYPE_SOCK_OPS = 0xd + BPF_PROG_TYPE_SK_SKB = 0xe + BPF_PROG_TYPE_CGROUP_DEVICE = 0xf + BPF_PROG_TYPE_SK_MSG = 0x10 + BPF_PROG_TYPE_RAW_TRACEPOINT = 0x11 + BPF_PROG_TYPE_CGROUP_SOCK_ADDR = 0x12 + BPF_PROG_TYPE_LWT_SEG6LOCAL = 0x13 + BPF_PROG_TYPE_LIRC_MODE2 = 0x14 + BPF_PROG_TYPE_SK_REUSEPORT = 0x15 + BPF_PROG_TYPE_FLOW_DISSECTOR = 0x16 + BPF_PROG_TYPE_CGROUP_SYSCTL = 0x17 + BPF_PROG_TYPE_RAW_TRACEPOINT_WRITABLE = 0x18 + BPF_PROG_TYPE_CGROUP_SOCKOPT = 0x19 + BPF_PROG_TYPE_TRACING = 0x1a + BPF_PROG_TYPE_STRUCT_OPS = 0x1b + BPF_PROG_TYPE_EXT = 0x1c + BPF_PROG_TYPE_LSM = 0x1d + BPF_CGROUP_INET_INGRESS = 0x0 + BPF_CGROUP_INET_EGRESS = 0x1 + BPF_CGROUP_INET_SOCK_CREATE = 0x2 + BPF_CGROUP_SOCK_OPS = 0x3 + BPF_SK_SKB_STREAM_PARSER = 0x4 + BPF_SK_SKB_STREAM_VERDICT = 0x5 + BPF_CGROUP_DEVICE = 0x6 + BPF_SK_MSG_VERDICT = 0x7 + BPF_CGROUP_INET4_BIND = 0x8 + BPF_CGROUP_INET6_BIND = 0x9 + BPF_CGROUP_INET4_CONNECT = 0xa + BPF_CGROUP_INET6_CONNECT = 0xb + BPF_CGROUP_INET4_POST_BIND = 0xc + BPF_CGROUP_INET6_POST_BIND = 0xd + BPF_CGROUP_UDP4_SENDMSG = 0xe + BPF_CGROUP_UDP6_SENDMSG = 0xf + BPF_LIRC_MODE2 = 0x10 + BPF_FLOW_DISSECTOR = 0x11 + BPF_CGROUP_SYSCTL = 0x12 + BPF_CGROUP_UDP4_RECVMSG = 0x13 + BPF_CGROUP_UDP6_RECVMSG = 0x14 + BPF_CGROUP_GETSOCKOPT = 0x15 + BPF_CGROUP_SETSOCKOPT = 0x16 + BPF_TRACE_RAW_TP = 0x17 + BPF_TRACE_FENTRY = 0x18 + BPF_TRACE_FEXIT = 0x19 + BPF_MODIFY_RETURN = 0x1a + BPF_LSM_MAC = 0x1b + BPF_ANY = 0x0 + BPF_NOEXIST = 0x1 + BPF_EXIST = 0x2 + BPF_F_LOCK = 0x4 + BPF_F_NO_PREALLOC = 0x1 + BPF_F_NO_COMMON_LRU = 0x2 + BPF_F_NUMA_NODE = 0x4 + BPF_F_RDONLY = 0x8 + BPF_F_WRONLY = 0x10 + BPF_F_STACK_BUILD_ID = 0x20 + BPF_F_ZERO_SEED = 0x40 + BPF_F_RDONLY_PROG = 0x80 + BPF_F_WRONLY_PROG = 0x100 + BPF_F_CLONE = 0x200 + BPF_F_MMAPABLE = 0x400 + BPF_STACK_BUILD_ID_EMPTY = 0x0 + BPF_STACK_BUILD_ID_VALID = 0x1 + BPF_STACK_BUILD_ID_IP = 0x2 + BPF_F_RECOMPUTE_CSUM = 0x1 + BPF_F_INVALIDATE_HASH = 0x2 + BPF_F_HDR_FIELD_MASK = 0xf + BPF_F_PSEUDO_HDR = 0x10 + BPF_F_MARK_MANGLED_0 = 0x20 + BPF_F_MARK_ENFORCE = 0x40 + BPF_F_INGRESS = 0x1 + BPF_F_TUNINFO_IPV6 = 0x1 + BPF_F_SKIP_FIELD_MASK = 0xff + BPF_F_USER_STACK = 0x100 + BPF_F_FAST_STACK_CMP = 0x200 + BPF_F_REUSE_STACKID = 0x400 + BPF_F_USER_BUILD_ID = 0x800 + BPF_F_ZERO_CSUM_TX = 0x2 + BPF_F_DONT_FRAGMENT = 0x4 + BPF_F_SEQ_NUMBER = 0x8 + BPF_F_INDEX_MASK = 0xffffffff + BPF_F_CURRENT_CPU = 0xffffffff + BPF_F_CTXLEN_MASK = 0xfffff00000000 + BPF_F_CURRENT_NETNS = -0x1 + BPF_F_ADJ_ROOM_FIXED_GSO = 0x1 + BPF_F_ADJ_ROOM_ENCAP_L3_IPV4 = 0x2 + BPF_F_ADJ_ROOM_ENCAP_L3_IPV6 = 0x4 + BPF_F_ADJ_ROOM_ENCAP_L4_GRE = 0x8 + BPF_F_ADJ_ROOM_ENCAP_L4_UDP = 0x10 + BPF_ADJ_ROOM_ENCAP_L2_MASK = 0xff + BPF_ADJ_ROOM_ENCAP_L2_SHIFT = 0x38 + BPF_F_SYSCTL_BASE_NAME = 0x1 + BPF_SK_STORAGE_GET_F_CREATE = 0x1 + BPF_F_GET_BRANCH_RECORDS_SIZE = 0x1 + BPF_ADJ_ROOM_NET = 0x0 + BPF_ADJ_ROOM_MAC = 0x1 + BPF_HDR_START_MAC = 0x0 + BPF_HDR_START_NET = 0x1 + BPF_LWT_ENCAP_SEG6 = 0x0 + BPF_LWT_ENCAP_SEG6_INLINE = 0x1 + BPF_LWT_ENCAP_IP = 0x2 + BPF_OK = 0x0 + BPF_DROP = 0x2 + BPF_REDIRECT = 0x7 + BPF_LWT_REROUTE = 0x80 + BPF_SOCK_OPS_RTO_CB_FLAG = 0x1 + BPF_SOCK_OPS_RETRANS_CB_FLAG = 0x2 + BPF_SOCK_OPS_STATE_CB_FLAG = 0x4 + BPF_SOCK_OPS_RTT_CB_FLAG = 0x8 + BPF_SOCK_OPS_ALL_CB_FLAGS = 0xf + BPF_SOCK_OPS_VOID = 0x0 + BPF_SOCK_OPS_TIMEOUT_INIT = 0x1 + BPF_SOCK_OPS_RWND_INIT = 0x2 + BPF_SOCK_OPS_TCP_CONNECT_CB = 0x3 + BPF_SOCK_OPS_ACTIVE_ESTABLISHED_CB = 0x4 + BPF_SOCK_OPS_PASSIVE_ESTABLISHED_CB = 0x5 + BPF_SOCK_OPS_NEEDS_ECN = 0x6 + BPF_SOCK_OPS_BASE_RTT = 0x7 + BPF_SOCK_OPS_RTO_CB = 0x8 + BPF_SOCK_OPS_RETRANS_CB = 0x9 + BPF_SOCK_OPS_STATE_CB = 0xa + BPF_SOCK_OPS_TCP_LISTEN_CB = 0xb + BPF_SOCK_OPS_RTT_CB = 0xc + BPF_TCP_ESTABLISHED = 0x1 + BPF_TCP_SYN_SENT = 0x2 + BPF_TCP_SYN_RECV = 0x3 + BPF_TCP_FIN_WAIT1 = 0x4 + BPF_TCP_FIN_WAIT2 = 0x5 + BPF_TCP_TIME_WAIT = 0x6 + BPF_TCP_CLOSE = 0x7 + BPF_TCP_CLOSE_WAIT = 0x8 + BPF_TCP_LAST_ACK = 0x9 + BPF_TCP_LISTEN = 0xa + BPF_TCP_CLOSING = 0xb + BPF_TCP_NEW_SYN_RECV = 0xc + BPF_TCP_MAX_STATES = 0xd + TCP_BPF_IW = 0x3e9 + TCP_BPF_SNDCWND_CLAMP = 0x3ea + BPF_DEVCG_ACC_MKNOD = 0x1 + BPF_DEVCG_ACC_READ = 0x2 + BPF_DEVCG_ACC_WRITE = 0x4 + BPF_DEVCG_DEV_BLOCK = 0x1 + BPF_DEVCG_DEV_CHAR = 0x2 + BPF_FIB_LOOKUP_DIRECT = 0x1 + BPF_FIB_LOOKUP_OUTPUT = 0x2 + BPF_FIB_LKUP_RET_SUCCESS = 0x0 + BPF_FIB_LKUP_RET_BLACKHOLE = 0x1 + BPF_FIB_LKUP_RET_UNREACHABLE = 0x2 + BPF_FIB_LKUP_RET_PROHIBIT = 0x3 + BPF_FIB_LKUP_RET_NOT_FWDED = 0x4 + BPF_FIB_LKUP_RET_FWD_DISABLED = 0x5 + BPF_FIB_LKUP_RET_UNSUPP_LWT = 0x6 + BPF_FIB_LKUP_RET_NO_NEIGH = 0x7 + BPF_FIB_LKUP_RET_FRAG_NEEDED = 0x8 + BPF_FD_TYPE_RAW_TRACEPOINT = 0x0 + BPF_FD_TYPE_TRACEPOINT = 0x1 + BPF_FD_TYPE_KPROBE = 0x2 + BPF_FD_TYPE_KRETPROBE = 0x3 + BPF_FD_TYPE_UPROBE = 0x4 + BPF_FD_TYPE_URETPROBE = 0x5 + BPF_FLOW_DISSECTOR_F_PARSE_1ST_FRAG = 0x1 + BPF_FLOW_DISSECTOR_F_STOP_AT_FLOW_LABEL = 0x2 + BPF_FLOW_DISSECTOR_F_STOP_AT_ENCAP = 0x4 ) const ( @@ -2216,7 +2279,7 @@ const ( DEVLINK_CMD_DPIPE_ENTRIES_GET = 0x20 DEVLINK_CMD_DPIPE_HEADERS_GET = 0x21 DEVLINK_CMD_DPIPE_TABLE_COUNTERS_SET = 0x22 - DEVLINK_CMD_MAX = 0x44 + DEVLINK_CMD_MAX = 0x48 DEVLINK_PORT_TYPE_NOTSET = 0x0 DEVLINK_PORT_TYPE_AUTO = 0x1 DEVLINK_PORT_TYPE_ETH = 0x2 @@ -2296,7 +2359,7 @@ const ( DEVLINK_ATTR_DPIPE_FIELD_MAPPING_TYPE = 0x3c DEVLINK_ATTR_PAD = 0x3d DEVLINK_ATTR_ESWITCH_ENCAP_MODE = 0x3e - DEVLINK_ATTR_MAX = 0x8c + DEVLINK_ATTR_MAX = 0x90 DEVLINK_DPIPE_FIELD_MAPPING_TYPE_NONE = 0x0 DEVLINK_DPIPE_FIELD_MAPPING_TYPE_IFINDEX = 0x1 DEVLINK_DPIPE_MATCH_TYPE_FIELD_EXACT = 0x0 diff --git a/vendor/golang.org/x/sys/windows/dll_windows.go b/vendor/golang.org/x/sys/windows/dll_windows.go index d7771134..82076fb7 100644 --- a/vendor/golang.org/x/sys/windows/dll_windows.go +++ b/vendor/golang.org/x/sys/windows/dll_windows.go @@ -104,6 +104,35 @@ func (d *DLL) MustFindProc(name string) *Proc { return p } +// FindProcByOrdinal searches DLL d for procedure by ordinal and returns *Proc +// if found. It returns an error if search fails. +func (d *DLL) FindProcByOrdinal(ordinal uintptr) (proc *Proc, err error) { + a, e := GetProcAddressByOrdinal(d.Handle, ordinal) + name := "#" + itoa(int(ordinal)) + if e != nil { + return nil, &DLLError{ + Err: e, + ObjName: name, + Msg: "Failed to find " + name + " procedure in " + d.Name + ": " + e.Error(), + } + } + p := &Proc{ + Dll: d, + Name: name, + addr: a, + } + return p, nil +} + +// MustFindProcByOrdinal is like FindProcByOrdinal but panics if search fails. +func (d *DLL) MustFindProcByOrdinal(ordinal uintptr) *Proc { + p, e := d.FindProcByOrdinal(ordinal) + if e != nil { + panic(e) + } + return p +} + // Release unloads DLL d from memory. func (d *DLL) Release() (err error) { return FreeLibrary(d.Handle) diff --git a/vendor/golang.org/x/sys/windows/env_windows.go b/vendor/golang.org/x/sys/windows/env_windows.go index f482a9fa..92ac05ff 100644 --- a/vendor/golang.org/x/sys/windows/env_windows.go +++ b/vendor/golang.org/x/sys/windows/env_windows.go @@ -8,7 +8,6 @@ package windows import ( "syscall" - "unicode/utf16" "unsafe" ) @@ -40,17 +39,11 @@ func (token Token) Environ(inheritExisting bool) (env []string, err error) { defer DestroyEnvironmentBlock(block) blockp := uintptr(unsafe.Pointer(block)) for { - entry := (*[(1 << 30) - 1]uint16)(unsafe.Pointer(blockp))[:] - for i, v := range entry { - if v == 0 { - entry = entry[:i] - break - } - } + entry := UTF16PtrToString((*uint16)(unsafe.Pointer(blockp))) if len(entry) == 0 { break } - env = append(env, string(utf16.Decode(entry))) + env = append(env, entry) blockp += 2 * (uintptr(len(entry)) + 1) } return env, nil diff --git a/vendor/golang.org/x/sys/windows/memory_windows.go b/vendor/golang.org/x/sys/windows/memory_windows.go index f80a4204..e409d76f 100644 --- a/vendor/golang.org/x/sys/windows/memory_windows.go +++ b/vendor/golang.org/x/sys/windows/memory_windows.go @@ -23,4 +23,9 @@ const ( PAGE_EXECUTE_READ = 0x20 PAGE_EXECUTE_READWRITE = 0x40 PAGE_EXECUTE_WRITECOPY = 0x80 + + QUOTA_LIMITS_HARDWS_MIN_DISABLE = 0x00000002 + QUOTA_LIMITS_HARDWS_MIN_ENABLE = 0x00000001 + QUOTA_LIMITS_HARDWS_MAX_DISABLE = 0x00000008 + QUOTA_LIMITS_HARDWS_MAX_ENABLE = 0x00000004 ) diff --git a/vendor/golang.org/x/sys/windows/security_windows.go b/vendor/golang.org/x/sys/windows/security_windows.go index 4b6eff18..9e3c44a8 100644 --- a/vendor/golang.org/x/sys/windows/security_windows.go +++ b/vendor/golang.org/x/sys/windows/security_windows.go @@ -7,6 +7,8 @@ package windows import ( "syscall" "unsafe" + + "golang.org/x/sys/internal/unsafeheader" ) const ( @@ -1229,7 +1231,7 @@ func (sd *SECURITY_DESCRIPTOR) String() string { return "" } defer LocalFree(Handle(unsafe.Pointer(sddl))) - return UTF16ToString((*[(1 << 30) - 1]uint16)(unsafe.Pointer(sddl))[:]) + return UTF16PtrToString(sddl) } // ToAbsolute converts a self-relative security descriptor into an absolute one. @@ -1307,9 +1309,17 @@ func (absoluteSD *SECURITY_DESCRIPTOR) ToSelfRelative() (selfRelativeSD *SECURIT } func (selfRelativeSD *SECURITY_DESCRIPTOR) copySelfRelativeSecurityDescriptor() *SECURITY_DESCRIPTOR { - sdBytes := make([]byte, selfRelativeSD.Length()) - copy(sdBytes, (*[(1 << 31) - 1]byte)(unsafe.Pointer(selfRelativeSD))[:len(sdBytes)]) - return (*SECURITY_DESCRIPTOR)(unsafe.Pointer(&sdBytes[0])) + sdLen := (int)(selfRelativeSD.Length()) + + var src []byte + h := (*unsafeheader.Slice)(unsafe.Pointer(&src)) + h.Data = unsafe.Pointer(selfRelativeSD) + h.Len = sdLen + h.Cap = sdLen + + dst := make([]byte, sdLen) + copy(dst, src) + return (*SECURITY_DESCRIPTOR)(unsafe.Pointer(&dst[0])) } // SecurityDescriptorFromString converts an SDDL string describing a security descriptor into a @@ -1391,6 +1401,6 @@ func ACLFromEntries(explicitEntries []EXPLICIT_ACCESS, mergedACL *ACL) (acl *ACL } defer LocalFree(Handle(unsafe.Pointer(winHeapACL))) aclBytes := make([]byte, winHeapACL.aclSize) - copy(aclBytes, (*[(1 << 31) - 1]byte)(unsafe.Pointer(winHeapACL))[:len(aclBytes)]) + copy(aclBytes, (*[(1 << 31) - 1]byte)(unsafe.Pointer(winHeapACL))[:len(aclBytes):len(aclBytes)]) return (*ACL)(unsafe.Pointer(&aclBytes[0])), nil } diff --git a/vendor/golang.org/x/sys/windows/syscall_windows.go b/vendor/golang.org/x/sys/windows/syscall_windows.go index 053d664d..62cf70e9 100644 --- a/vendor/golang.org/x/sys/windows/syscall_windows.go +++ b/vendor/golang.org/x/sys/windows/syscall_windows.go @@ -13,6 +13,8 @@ import ( "time" "unicode/utf16" "unsafe" + + "golang.org/x/sys/internal/unsafeheader" ) type Handle uintptr @@ -117,6 +119,32 @@ func UTF16PtrFromString(s string) (*uint16, error) { return &a[0], nil } +// UTF16PtrToString takes a pointer to a UTF-16 sequence and returns the corresponding UTF-8 encoded string. +// If the pointer is nil, this returns the empty string. This assumes that the UTF-16 sequence is terminated +// at a zero word; if the zero word is not present, the program may crash. +func UTF16PtrToString(p *uint16) string { + if p == nil { + return "" + } + if *p == 0 { + return "" + } + + // Find NUL terminator. + n := 0 + for ptr := unsafe.Pointer(p); *(*uint16)(ptr) != 0; n++ { + ptr = unsafe.Pointer(uintptr(ptr) + unsafe.Sizeof(*p)) + } + + var s []uint16 + h := (*unsafeheader.Slice)(unsafe.Pointer(&s)) + h.Data = unsafe.Pointer(p) + h.Len = n + h.Cap = n + + return string(utf16.Decode(s)) +} + func Getpagesize() int { return 4096 } // NewCallback converts a Go function to a function pointer conforming to the stdcall calling convention. @@ -280,6 +308,8 @@ func NewCallbackCDecl(fn interface{}) uintptr { //sys GetProcessId(process Handle) (id uint32, err error) //sys OpenThread(desiredAccess uint32, inheritHandle bool, threadId uint32) (handle Handle, err error) //sys SetProcessPriorityBoost(process Handle, disable bool) (err error) = kernel32.SetProcessPriorityBoost +//sys GetProcessWorkingSetSizeEx(hProcess Handle, lpMinimumWorkingSetSize *uintptr, lpMaximumWorkingSetSize *uintptr, flags *uint32) +//sys SetProcessWorkingSetSizeEx(hProcess Handle, dwMinimumWorkingSetSize uintptr, dwMaximumWorkingSetSize uintptr, flags uint32) (err error) // Volume Management Functions //sys DefineDosDevice(flags uint32, deviceName *uint16, targetPath *uint16) (err error) = DefineDosDeviceW @@ -1181,7 +1211,12 @@ type IPv6Mreq struct { Interface uint32 } -func GetsockoptInt(fd Handle, level, opt int) (int, error) { return -1, syscall.EWINDOWS } +func GetsockoptInt(fd Handle, level, opt int) (int, error) { + v := int32(0) + l := int32(unsafe.Sizeof(v)) + err := Getsockopt(fd, int32(level), int32(opt), (*byte)(unsafe.Pointer(&v)), &l) + return int(v), err +} func SetsockoptLinger(fd Handle, level, opt int, l *Linger) (err error) { sys := sysLinger{Onoff: uint16(l.Onoff), Linger: uint16(l.Linger)} @@ -1378,7 +1413,7 @@ func (t Token) KnownFolderPath(folderID *KNOWNFOLDERID, flags uint32) (string, e return "", err } defer CoTaskMemFree(unsafe.Pointer(p)) - return UTF16ToString((*[(1 << 30) - 1]uint16)(unsafe.Pointer(p))[:]), nil + return UTF16PtrToString(p), nil } // RtlGetVersion returns the version of the underlying operating system, ignoring diff --git a/vendor/golang.org/x/sys/windows/zsyscall_windows.go b/vendor/golang.org/x/sys/windows/zsyscall_windows.go index 2aa4fa64..8a562fee 100644 --- a/vendor/golang.org/x/sys/windows/zsyscall_windows.go +++ b/vendor/golang.org/x/sys/windows/zsyscall_windows.go @@ -217,6 +217,8 @@ var ( procGetProcessId = modkernel32.NewProc("GetProcessId") procOpenThread = modkernel32.NewProc("OpenThread") procSetProcessPriorityBoost = modkernel32.NewProc("SetProcessPriorityBoost") + procGetProcessWorkingSetSizeEx = modkernel32.NewProc("GetProcessWorkingSetSizeEx") + procSetProcessWorkingSetSizeEx = modkernel32.NewProc("SetProcessWorkingSetSizeEx") procDefineDosDeviceW = modkernel32.NewProc("DefineDosDeviceW") procDeleteVolumeMountPointW = modkernel32.NewProc("DeleteVolumeMountPointW") procFindFirstVolumeW = modkernel32.NewProc("FindFirstVolumeW") @@ -2414,6 +2416,23 @@ func SetProcessPriorityBoost(process Handle, disable bool) (err error) { return } +func GetProcessWorkingSetSizeEx(hProcess Handle, lpMinimumWorkingSetSize *uintptr, lpMaximumWorkingSetSize *uintptr, flags *uint32) { + syscall.Syscall6(procGetProcessWorkingSetSizeEx.Addr(), 4, uintptr(hProcess), uintptr(unsafe.Pointer(lpMinimumWorkingSetSize)), uintptr(unsafe.Pointer(lpMaximumWorkingSetSize)), uintptr(unsafe.Pointer(flags)), 0, 0) + return +} + +func SetProcessWorkingSetSizeEx(hProcess Handle, dwMinimumWorkingSetSize uintptr, dwMaximumWorkingSetSize uintptr, flags uint32) (err error) { + r1, _, e1 := syscall.Syscall6(procSetProcessWorkingSetSizeEx.Addr(), 4, uintptr(hProcess), uintptr(dwMinimumWorkingSetSize), uintptr(dwMaximumWorkingSetSize), uintptr(flags), 0, 0) + if r1 == 0 { + if e1 != 0 { + err = errnoErr(e1) + } else { + err = syscall.EINVAL + } + } + return +} + func DefineDosDevice(flags uint32, deviceName *uint16, targetPath *uint16) (err error) { r1, _, e1 := syscall.Syscall(procDefineDosDeviceW.Addr(), 3, uintptr(flags), uintptr(unsafe.Pointer(deviceName)), uintptr(unsafe.Pointer(targetPath))) if r1 == 0 { diff --git a/vendor/golang.org/x/text/encoding/unicode/unicode.go b/vendor/golang.org/x/text/encoding/unicode/unicode.go index 4850ff36..dd99ad14 100644 --- a/vendor/golang.org/x/text/encoding/unicode/unicode.go +++ b/vendor/golang.org/x/text/encoding/unicode/unicode.go @@ -6,6 +6,7 @@ package unicode // import "golang.org/x/text/encoding/unicode" import ( + "bytes" "errors" "unicode/utf16" "unicode/utf8" @@ -25,15 +26,95 @@ import ( // the introduction of some kind of error type for conveying the erroneous code // point. -// UTF8 is the UTF-8 encoding. +// UTF8 is the UTF-8 encoding. It neither removes nor adds byte order marks. var UTF8 encoding.Encoding = utf8enc +// UTF8BOM is an UTF-8 encoding where the decoder strips a leading byte order +// mark while the encoder adds one. +// +// Some editors add a byte order mark as a signature to UTF-8 files. Although +// the byte order mark is not useful for detecting byte order in UTF-8, it is +// sometimes used as a convention to mark UTF-8-encoded files. This relies on +// the observation that the UTF-8 byte order mark is either an illegal or at +// least very unlikely sequence in any other character encoding. +var UTF8BOM encoding.Encoding = utf8bomEncoding{} + +type utf8bomEncoding struct{} + +func (utf8bomEncoding) String() string { + return "UTF-8-BOM" +} + +func (utf8bomEncoding) ID() (identifier.MIB, string) { + return identifier.Unofficial, "x-utf8bom" +} + +func (utf8bomEncoding) NewEncoder() *encoding.Encoder { + return &encoding.Encoder{ + Transformer: &utf8bomEncoder{t: runes.ReplaceIllFormed()}, + } +} + +func (utf8bomEncoding) NewDecoder() *encoding.Decoder { + return &encoding.Decoder{Transformer: &utf8bomDecoder{}} +} + var utf8enc = &internal.Encoding{ &internal.SimpleEncoding{utf8Decoder{}, runes.ReplaceIllFormed()}, "UTF-8", identifier.UTF8, } +type utf8bomDecoder struct { + checked bool +} + +func (t *utf8bomDecoder) Reset() { + t.checked = false +} + +func (t *utf8bomDecoder) Transform(dst, src []byte, atEOF bool) (nDst, nSrc int, err error) { + if !t.checked { + if !atEOF && len(src) < len(utf8BOM) { + if len(src) == 0 { + return 0, 0, nil + } + return 0, 0, transform.ErrShortSrc + } + if bytes.HasPrefix(src, []byte(utf8BOM)) { + nSrc += len(utf8BOM) + src = src[len(utf8BOM):] + } + t.checked = true + } + nDst, n, err := utf8Decoder.Transform(utf8Decoder{}, dst[nDst:], src, atEOF) + nSrc += n + return nDst, nSrc, err +} + +type utf8bomEncoder struct { + written bool + t transform.Transformer +} + +func (t *utf8bomEncoder) Reset() { + t.written = false + t.t.Reset() +} + +func (t *utf8bomEncoder) Transform(dst, src []byte, atEOF bool) (nDst, nSrc int, err error) { + if !t.written { + if len(dst) < len(utf8BOM) { + return nDst, 0, transform.ErrShortDst + } + nDst = copy(dst, utf8BOM) + t.written = true + } + n, nSrc, err := utf8Decoder.Transform(utf8Decoder{}, dst[nDst:], src, atEOF) + nDst += n + return nDst, nSrc, err +} + type utf8Decoder struct{ transform.NopResetter } func (utf8Decoder) Transform(dst, src []byte, atEOF bool) (nDst, nSrc int, err error) { @@ -287,16 +368,13 @@ func (u *utf16Decoder) Reset() { } func (u *utf16Decoder) Transform(dst, src []byte, atEOF bool) (nDst, nSrc int, err error) { + if len(src) < 2 && atEOF && u.current.bomPolicy&requireBOM != 0 { + return 0, 0, ErrMissingBOM + } if len(src) == 0 { - if atEOF && u.current.bomPolicy&requireBOM != 0 { - return 0, 0, ErrMissingBOM - } return 0, 0, nil } - if u.current.bomPolicy&acceptBOM != 0 { - if len(src) < 2 { - return 0, 0, transform.ErrShortSrc - } + if len(src) >= 2 && u.current.bomPolicy&acceptBOM != 0 { switch { case src[0] == 0xfe && src[1] == 0xff: u.current.endianness = BigEndian diff --git a/vendor/golang.org/x/text/transform/transform.go b/vendor/golang.org/x/text/transform/transform.go index 520b9ada..48ec64b4 100644 --- a/vendor/golang.org/x/text/transform/transform.go +++ b/vendor/golang.org/x/text/transform/transform.go @@ -648,7 +648,8 @@ func String(t Transformer, s string) (result string, n int, err error) { // Transform the remaining input, growing dst and src buffers as necessary. for { n := copy(src, s[pSrc:]) - nDst, nSrc, err := t.Transform(dst[pDst:], src[:n], pSrc+n == len(s)) + atEOF := pSrc+n == len(s) + nDst, nSrc, err := t.Transform(dst[pDst:], src[:n], atEOF) pDst += nDst pSrc += nSrc @@ -659,6 +660,9 @@ func String(t Transformer, s string) (result string, n int, err error) { dst = grow(dst, pDst) } } else if err == ErrShortSrc { + if atEOF { + return string(dst[:pDst]), pSrc, err + } if nSrc == 0 { src = grow(src, 0) } diff --git a/vendor/golang.org/x/text/unicode/bidi/core.go b/vendor/golang.org/x/text/unicode/bidi/core.go index 48d14400..50deb660 100644 --- a/vendor/golang.org/x/text/unicode/bidi/core.go +++ b/vendor/golang.org/x/text/unicode/bidi/core.go @@ -480,15 +480,15 @@ func (s *isolatingRunSequence) resolveWeakTypes() { // Rule W1. // Changes all NSMs. - preceedingCharacterType := s.sos + precedingCharacterType := s.sos for i, t := range s.types { if t == NSM { - s.types[i] = preceedingCharacterType + s.types[i] = precedingCharacterType } else { if t.in(LRI, RLI, FSI, PDI) { - preceedingCharacterType = ON + precedingCharacterType = ON } - preceedingCharacterType = t + precedingCharacterType = t } } diff --git a/vendor/golang.org/x/text/unicode/bidi/tables11.0.0.go b/vendor/golang.org/x/text/unicode/bidi/tables11.0.0.go index 022e3c69..16b11db5 100644 --- a/vendor/golang.org/x/text/unicode/bidi/tables11.0.0.go +++ b/vendor/golang.org/x/text/unicode/bidi/tables11.0.0.go @@ -1,6 +1,6 @@ // Code generated by running "go generate" in golang.org/x/text. DO NOT EDIT. -// +build go1.13 +// +build go1.13,!go1.14 package bidi diff --git a/vendor/golang.org/x/text/unicode/bidi/tables12.0.0.go b/vendor/golang.org/x/text/unicode/bidi/tables12.0.0.go new file mode 100644 index 00000000..7ffa3651 --- /dev/null +++ b/vendor/golang.org/x/text/unicode/bidi/tables12.0.0.go @@ -0,0 +1,1923 @@ +// Code generated by running "go generate" in golang.org/x/text. DO NOT EDIT. + +// +build go1.14 + +package bidi + +// UnicodeVersion is the Unicode version from which the tables in this package are derived. +const UnicodeVersion = "12.0.0" + +// xorMasks contains masks to be xor-ed with brackets to get the reverse +// version. +var xorMasks = []int32{ // 8 elements + 0, 1, 6, 7, 3, 15, 29, 63, +} // Size: 56 bytes + +// lookup returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *bidiTrie) lookup(s []byte) (v uint8, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return bidiValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := bidiIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := bidiIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = bidiIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := bidiIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = bidiIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = bidiIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *bidiTrie) lookupUnsafe(s []byte) uint8 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return bidiValues[c0] + } + i := bidiIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = bidiIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = bidiIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// lookupString returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *bidiTrie) lookupString(s string) (v uint8, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return bidiValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := bidiIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := bidiIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = bidiIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := bidiIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = bidiIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = bidiIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupStringUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *bidiTrie) lookupStringUnsafe(s string) uint8 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return bidiValues[c0] + } + i := bidiIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = bidiIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = bidiIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// bidiTrie. Total size: 16896 bytes (16.50 KiB). Checksum: 6f0927067913dc6d. +type bidiTrie struct{} + +func newBidiTrie(i int) *bidiTrie { + return &bidiTrie{} +} + +// lookupValue determines the type of block n and looks up the value for b. +func (t *bidiTrie) lookupValue(n uint32, b byte) uint8 { + switch { + default: + return uint8(bidiValues[n<<6+uint32(b)]) + } +} + +// bidiValues: 240 blocks, 15360 entries, 15360 bytes +// The third block is the zero block. +var bidiValues = [15360]uint8{ + // Block 0x0, offset 0x0 + 0x00: 0x000b, 0x01: 0x000b, 0x02: 0x000b, 0x03: 0x000b, 0x04: 0x000b, 0x05: 0x000b, + 0x06: 0x000b, 0x07: 0x000b, 0x08: 0x000b, 0x09: 0x0008, 0x0a: 0x0007, 0x0b: 0x0008, + 0x0c: 0x0009, 0x0d: 0x0007, 0x0e: 0x000b, 0x0f: 0x000b, 0x10: 0x000b, 0x11: 0x000b, + 0x12: 0x000b, 0x13: 0x000b, 0x14: 0x000b, 0x15: 0x000b, 0x16: 0x000b, 0x17: 0x000b, + 0x18: 0x000b, 0x19: 0x000b, 0x1a: 0x000b, 0x1b: 0x000b, 0x1c: 0x0007, 0x1d: 0x0007, + 0x1e: 0x0007, 0x1f: 0x0008, 0x20: 0x0009, 0x21: 0x000a, 0x22: 0x000a, 0x23: 0x0004, + 0x24: 0x0004, 0x25: 0x0004, 0x26: 0x000a, 0x27: 0x000a, 0x28: 0x003a, 0x29: 0x002a, + 0x2a: 0x000a, 0x2b: 0x0003, 0x2c: 0x0006, 0x2d: 0x0003, 0x2e: 0x0006, 0x2f: 0x0006, + 0x30: 0x0002, 0x31: 0x0002, 0x32: 0x0002, 0x33: 0x0002, 0x34: 0x0002, 0x35: 0x0002, + 0x36: 0x0002, 0x37: 0x0002, 0x38: 0x0002, 0x39: 0x0002, 0x3a: 0x0006, 0x3b: 0x000a, + 0x3c: 0x000a, 0x3d: 0x000a, 0x3e: 0x000a, 0x3f: 0x000a, + // Block 0x1, offset 0x40 + 0x40: 0x000a, + 0x5b: 0x005a, 0x5c: 0x000a, 0x5d: 0x004a, + 0x5e: 0x000a, 0x5f: 0x000a, 0x60: 0x000a, + 0x7b: 0x005a, + 0x7c: 0x000a, 0x7d: 0x004a, 0x7e: 0x000a, 0x7f: 0x000b, + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xc0: 0x000b, 0xc1: 0x000b, 0xc2: 0x000b, 0xc3: 0x000b, 0xc4: 0x000b, 0xc5: 0x0007, + 0xc6: 0x000b, 0xc7: 0x000b, 0xc8: 0x000b, 0xc9: 0x000b, 0xca: 0x000b, 0xcb: 0x000b, + 0xcc: 0x000b, 0xcd: 0x000b, 0xce: 0x000b, 0xcf: 0x000b, 0xd0: 0x000b, 0xd1: 0x000b, + 0xd2: 0x000b, 0xd3: 0x000b, 0xd4: 0x000b, 0xd5: 0x000b, 0xd6: 0x000b, 0xd7: 0x000b, + 0xd8: 0x000b, 0xd9: 0x000b, 0xda: 0x000b, 0xdb: 0x000b, 0xdc: 0x000b, 0xdd: 0x000b, + 0xde: 0x000b, 0xdf: 0x000b, 0xe0: 0x0006, 0xe1: 0x000a, 0xe2: 0x0004, 0xe3: 0x0004, + 0xe4: 0x0004, 0xe5: 0x0004, 0xe6: 0x000a, 0xe7: 0x000a, 0xe8: 0x000a, 0xe9: 0x000a, + 0xeb: 0x000a, 0xec: 0x000a, 0xed: 0x000b, 0xee: 0x000a, 0xef: 0x000a, + 0xf0: 0x0004, 0xf1: 0x0004, 0xf2: 0x0002, 0xf3: 0x0002, 0xf4: 0x000a, + 0xf6: 0x000a, 0xf7: 0x000a, 0xf8: 0x000a, 0xf9: 0x0002, 0xfb: 0x000a, + 0xfc: 0x000a, 0xfd: 0x000a, 0xfe: 0x000a, 0xff: 0x000a, + // Block 0x4, offset 0x100 + 0x117: 0x000a, + 0x137: 0x000a, + // Block 0x5, offset 0x140 + 0x179: 0x000a, 0x17a: 0x000a, + // Block 0x6, offset 0x180 + 0x182: 0x000a, 0x183: 0x000a, 0x184: 0x000a, 0x185: 0x000a, + 0x186: 0x000a, 0x187: 0x000a, 0x188: 0x000a, 0x189: 0x000a, 0x18a: 0x000a, 0x18b: 0x000a, + 0x18c: 0x000a, 0x18d: 0x000a, 0x18e: 0x000a, 0x18f: 0x000a, + 0x192: 0x000a, 0x193: 0x000a, 0x194: 0x000a, 0x195: 0x000a, 0x196: 0x000a, 0x197: 0x000a, + 0x198: 0x000a, 0x199: 0x000a, 0x19a: 0x000a, 0x19b: 0x000a, 0x19c: 0x000a, 0x19d: 0x000a, + 0x19e: 0x000a, 0x19f: 0x000a, + 0x1a5: 0x000a, 0x1a6: 0x000a, 0x1a7: 0x000a, 0x1a8: 0x000a, 0x1a9: 0x000a, + 0x1aa: 0x000a, 0x1ab: 0x000a, 0x1ac: 0x000a, 0x1ad: 0x000a, 0x1af: 0x000a, + 0x1b0: 0x000a, 0x1b1: 0x000a, 0x1b2: 0x000a, 0x1b3: 0x000a, 0x1b4: 0x000a, 0x1b5: 0x000a, + 0x1b6: 0x000a, 0x1b7: 0x000a, 0x1b8: 0x000a, 0x1b9: 0x000a, 0x1ba: 0x000a, 0x1bb: 0x000a, + 0x1bc: 0x000a, 0x1bd: 0x000a, 0x1be: 0x000a, 0x1bf: 0x000a, + // Block 0x7, offset 0x1c0 + 0x1c0: 0x000c, 0x1c1: 0x000c, 0x1c2: 0x000c, 0x1c3: 0x000c, 0x1c4: 0x000c, 0x1c5: 0x000c, + 0x1c6: 0x000c, 0x1c7: 0x000c, 0x1c8: 0x000c, 0x1c9: 0x000c, 0x1ca: 0x000c, 0x1cb: 0x000c, + 0x1cc: 0x000c, 0x1cd: 0x000c, 0x1ce: 0x000c, 0x1cf: 0x000c, 0x1d0: 0x000c, 0x1d1: 0x000c, + 0x1d2: 0x000c, 0x1d3: 0x000c, 0x1d4: 0x000c, 0x1d5: 0x000c, 0x1d6: 0x000c, 0x1d7: 0x000c, + 0x1d8: 0x000c, 0x1d9: 0x000c, 0x1da: 0x000c, 0x1db: 0x000c, 0x1dc: 0x000c, 0x1dd: 0x000c, + 0x1de: 0x000c, 0x1df: 0x000c, 0x1e0: 0x000c, 0x1e1: 0x000c, 0x1e2: 0x000c, 0x1e3: 0x000c, + 0x1e4: 0x000c, 0x1e5: 0x000c, 0x1e6: 0x000c, 0x1e7: 0x000c, 0x1e8: 0x000c, 0x1e9: 0x000c, + 0x1ea: 0x000c, 0x1eb: 0x000c, 0x1ec: 0x000c, 0x1ed: 0x000c, 0x1ee: 0x000c, 0x1ef: 0x000c, + 0x1f0: 0x000c, 0x1f1: 0x000c, 0x1f2: 0x000c, 0x1f3: 0x000c, 0x1f4: 0x000c, 0x1f5: 0x000c, + 0x1f6: 0x000c, 0x1f7: 0x000c, 0x1f8: 0x000c, 0x1f9: 0x000c, 0x1fa: 0x000c, 0x1fb: 0x000c, + 0x1fc: 0x000c, 0x1fd: 0x000c, 0x1fe: 0x000c, 0x1ff: 0x000c, + // Block 0x8, offset 0x200 + 0x200: 0x000c, 0x201: 0x000c, 0x202: 0x000c, 0x203: 0x000c, 0x204: 0x000c, 0x205: 0x000c, + 0x206: 0x000c, 0x207: 0x000c, 0x208: 0x000c, 0x209: 0x000c, 0x20a: 0x000c, 0x20b: 0x000c, + 0x20c: 0x000c, 0x20d: 0x000c, 0x20e: 0x000c, 0x20f: 0x000c, 0x210: 0x000c, 0x211: 0x000c, + 0x212: 0x000c, 0x213: 0x000c, 0x214: 0x000c, 0x215: 0x000c, 0x216: 0x000c, 0x217: 0x000c, + 0x218: 0x000c, 0x219: 0x000c, 0x21a: 0x000c, 0x21b: 0x000c, 0x21c: 0x000c, 0x21d: 0x000c, + 0x21e: 0x000c, 0x21f: 0x000c, 0x220: 0x000c, 0x221: 0x000c, 0x222: 0x000c, 0x223: 0x000c, + 0x224: 0x000c, 0x225: 0x000c, 0x226: 0x000c, 0x227: 0x000c, 0x228: 0x000c, 0x229: 0x000c, + 0x22a: 0x000c, 0x22b: 0x000c, 0x22c: 0x000c, 0x22d: 0x000c, 0x22e: 0x000c, 0x22f: 0x000c, + 0x234: 0x000a, 0x235: 0x000a, + 0x23e: 0x000a, + // Block 0x9, offset 0x240 + 0x244: 0x000a, 0x245: 0x000a, + 0x247: 0x000a, + // Block 0xa, offset 0x280 + 0x2b6: 0x000a, + // Block 0xb, offset 0x2c0 + 0x2c3: 0x000c, 0x2c4: 0x000c, 0x2c5: 0x000c, + 0x2c6: 0x000c, 0x2c7: 0x000c, 0x2c8: 0x000c, 0x2c9: 0x000c, + // Block 0xc, offset 0x300 + 0x30a: 0x000a, + 0x30d: 0x000a, 0x30e: 0x000a, 0x30f: 0x0004, 0x310: 0x0001, 0x311: 0x000c, + 0x312: 0x000c, 0x313: 0x000c, 0x314: 0x000c, 0x315: 0x000c, 0x316: 0x000c, 0x317: 0x000c, + 0x318: 0x000c, 0x319: 0x000c, 0x31a: 0x000c, 0x31b: 0x000c, 0x31c: 0x000c, 0x31d: 0x000c, + 0x31e: 0x000c, 0x31f: 0x000c, 0x320: 0x000c, 0x321: 0x000c, 0x322: 0x000c, 0x323: 0x000c, + 0x324: 0x000c, 0x325: 0x000c, 0x326: 0x000c, 0x327: 0x000c, 0x328: 0x000c, 0x329: 0x000c, + 0x32a: 0x000c, 0x32b: 0x000c, 0x32c: 0x000c, 0x32d: 0x000c, 0x32e: 0x000c, 0x32f: 0x000c, + 0x330: 0x000c, 0x331: 0x000c, 0x332: 0x000c, 0x333: 0x000c, 0x334: 0x000c, 0x335: 0x000c, + 0x336: 0x000c, 0x337: 0x000c, 0x338: 0x000c, 0x339: 0x000c, 0x33a: 0x000c, 0x33b: 0x000c, + 0x33c: 0x000c, 0x33d: 0x000c, 0x33e: 0x0001, 0x33f: 0x000c, + // Block 0xd, offset 0x340 + 0x340: 0x0001, 0x341: 0x000c, 0x342: 0x000c, 0x343: 0x0001, 0x344: 0x000c, 0x345: 0x000c, + 0x346: 0x0001, 0x347: 0x000c, 0x348: 0x0001, 0x349: 0x0001, 0x34a: 0x0001, 0x34b: 0x0001, + 0x34c: 0x0001, 0x34d: 0x0001, 0x34e: 0x0001, 0x34f: 0x0001, 0x350: 0x0001, 0x351: 0x0001, + 0x352: 0x0001, 0x353: 0x0001, 0x354: 0x0001, 0x355: 0x0001, 0x356: 0x0001, 0x357: 0x0001, + 0x358: 0x0001, 0x359: 0x0001, 0x35a: 0x0001, 0x35b: 0x0001, 0x35c: 0x0001, 0x35d: 0x0001, + 0x35e: 0x0001, 0x35f: 0x0001, 0x360: 0x0001, 0x361: 0x0001, 0x362: 0x0001, 0x363: 0x0001, + 0x364: 0x0001, 0x365: 0x0001, 0x366: 0x0001, 0x367: 0x0001, 0x368: 0x0001, 0x369: 0x0001, + 0x36a: 0x0001, 0x36b: 0x0001, 0x36c: 0x0001, 0x36d: 0x0001, 0x36e: 0x0001, 0x36f: 0x0001, + 0x370: 0x0001, 0x371: 0x0001, 0x372: 0x0001, 0x373: 0x0001, 0x374: 0x0001, 0x375: 0x0001, + 0x376: 0x0001, 0x377: 0x0001, 0x378: 0x0001, 0x379: 0x0001, 0x37a: 0x0001, 0x37b: 0x0001, + 0x37c: 0x0001, 0x37d: 0x0001, 0x37e: 0x0001, 0x37f: 0x0001, + // Block 0xe, offset 0x380 + 0x380: 0x0005, 0x381: 0x0005, 0x382: 0x0005, 0x383: 0x0005, 0x384: 0x0005, 0x385: 0x0005, + 0x386: 0x000a, 0x387: 0x000a, 0x388: 0x000d, 0x389: 0x0004, 0x38a: 0x0004, 0x38b: 0x000d, + 0x38c: 0x0006, 0x38d: 0x000d, 0x38e: 0x000a, 0x38f: 0x000a, 0x390: 0x000c, 0x391: 0x000c, + 0x392: 0x000c, 0x393: 0x000c, 0x394: 0x000c, 0x395: 0x000c, 0x396: 0x000c, 0x397: 0x000c, + 0x398: 0x000c, 0x399: 0x000c, 0x39a: 0x000c, 0x39b: 0x000d, 0x39c: 0x000d, 0x39d: 0x000d, + 0x39e: 0x000d, 0x39f: 0x000d, 0x3a0: 0x000d, 0x3a1: 0x000d, 0x3a2: 0x000d, 0x3a3: 0x000d, + 0x3a4: 0x000d, 0x3a5: 0x000d, 0x3a6: 0x000d, 0x3a7: 0x000d, 0x3a8: 0x000d, 0x3a9: 0x000d, + 0x3aa: 0x000d, 0x3ab: 0x000d, 0x3ac: 0x000d, 0x3ad: 0x000d, 0x3ae: 0x000d, 0x3af: 0x000d, + 0x3b0: 0x000d, 0x3b1: 0x000d, 0x3b2: 0x000d, 0x3b3: 0x000d, 0x3b4: 0x000d, 0x3b5: 0x000d, + 0x3b6: 0x000d, 0x3b7: 0x000d, 0x3b8: 0x000d, 0x3b9: 0x000d, 0x3ba: 0x000d, 0x3bb: 0x000d, + 0x3bc: 0x000d, 0x3bd: 0x000d, 0x3be: 0x000d, 0x3bf: 0x000d, + // Block 0xf, offset 0x3c0 + 0x3c0: 0x000d, 0x3c1: 0x000d, 0x3c2: 0x000d, 0x3c3: 0x000d, 0x3c4: 0x000d, 0x3c5: 0x000d, + 0x3c6: 0x000d, 0x3c7: 0x000d, 0x3c8: 0x000d, 0x3c9: 0x000d, 0x3ca: 0x000d, 0x3cb: 0x000c, + 0x3cc: 0x000c, 0x3cd: 0x000c, 0x3ce: 0x000c, 0x3cf: 0x000c, 0x3d0: 0x000c, 0x3d1: 0x000c, + 0x3d2: 0x000c, 0x3d3: 0x000c, 0x3d4: 0x000c, 0x3d5: 0x000c, 0x3d6: 0x000c, 0x3d7: 0x000c, + 0x3d8: 0x000c, 0x3d9: 0x000c, 0x3da: 0x000c, 0x3db: 0x000c, 0x3dc: 0x000c, 0x3dd: 0x000c, + 0x3de: 0x000c, 0x3df: 0x000c, 0x3e0: 0x0005, 0x3e1: 0x0005, 0x3e2: 0x0005, 0x3e3: 0x0005, + 0x3e4: 0x0005, 0x3e5: 0x0005, 0x3e6: 0x0005, 0x3e7: 0x0005, 0x3e8: 0x0005, 0x3e9: 0x0005, + 0x3ea: 0x0004, 0x3eb: 0x0005, 0x3ec: 0x0005, 0x3ed: 0x000d, 0x3ee: 0x000d, 0x3ef: 0x000d, + 0x3f0: 0x000c, 0x3f1: 0x000d, 0x3f2: 0x000d, 0x3f3: 0x000d, 0x3f4: 0x000d, 0x3f5: 0x000d, + 0x3f6: 0x000d, 0x3f7: 0x000d, 0x3f8: 0x000d, 0x3f9: 0x000d, 0x3fa: 0x000d, 0x3fb: 0x000d, + 0x3fc: 0x000d, 0x3fd: 0x000d, 0x3fe: 0x000d, 0x3ff: 0x000d, + // Block 0x10, offset 0x400 + 0x400: 0x000d, 0x401: 0x000d, 0x402: 0x000d, 0x403: 0x000d, 0x404: 0x000d, 0x405: 0x000d, + 0x406: 0x000d, 0x407: 0x000d, 0x408: 0x000d, 0x409: 0x000d, 0x40a: 0x000d, 0x40b: 0x000d, + 0x40c: 0x000d, 0x40d: 0x000d, 0x40e: 0x000d, 0x40f: 0x000d, 0x410: 0x000d, 0x411: 0x000d, + 0x412: 0x000d, 0x413: 0x000d, 0x414: 0x000d, 0x415: 0x000d, 0x416: 0x000d, 0x417: 0x000d, + 0x418: 0x000d, 0x419: 0x000d, 0x41a: 0x000d, 0x41b: 0x000d, 0x41c: 0x000d, 0x41d: 0x000d, + 0x41e: 0x000d, 0x41f: 0x000d, 0x420: 0x000d, 0x421: 0x000d, 0x422: 0x000d, 0x423: 0x000d, + 0x424: 0x000d, 0x425: 0x000d, 0x426: 0x000d, 0x427: 0x000d, 0x428: 0x000d, 0x429: 0x000d, + 0x42a: 0x000d, 0x42b: 0x000d, 0x42c: 0x000d, 0x42d: 0x000d, 0x42e: 0x000d, 0x42f: 0x000d, + 0x430: 0x000d, 0x431: 0x000d, 0x432: 0x000d, 0x433: 0x000d, 0x434: 0x000d, 0x435: 0x000d, + 0x436: 0x000d, 0x437: 0x000d, 0x438: 0x000d, 0x439: 0x000d, 0x43a: 0x000d, 0x43b: 0x000d, + 0x43c: 0x000d, 0x43d: 0x000d, 0x43e: 0x000d, 0x43f: 0x000d, + // Block 0x11, offset 0x440 + 0x440: 0x000d, 0x441: 0x000d, 0x442: 0x000d, 0x443: 0x000d, 0x444: 0x000d, 0x445: 0x000d, + 0x446: 0x000d, 0x447: 0x000d, 0x448: 0x000d, 0x449: 0x000d, 0x44a: 0x000d, 0x44b: 0x000d, + 0x44c: 0x000d, 0x44d: 0x000d, 0x44e: 0x000d, 0x44f: 0x000d, 0x450: 0x000d, 0x451: 0x000d, + 0x452: 0x000d, 0x453: 0x000d, 0x454: 0x000d, 0x455: 0x000d, 0x456: 0x000c, 0x457: 0x000c, + 0x458: 0x000c, 0x459: 0x000c, 0x45a: 0x000c, 0x45b: 0x000c, 0x45c: 0x000c, 0x45d: 0x0005, + 0x45e: 0x000a, 0x45f: 0x000c, 0x460: 0x000c, 0x461: 0x000c, 0x462: 0x000c, 0x463: 0x000c, + 0x464: 0x000c, 0x465: 0x000d, 0x466: 0x000d, 0x467: 0x000c, 0x468: 0x000c, 0x469: 0x000a, + 0x46a: 0x000c, 0x46b: 0x000c, 0x46c: 0x000c, 0x46d: 0x000c, 0x46e: 0x000d, 0x46f: 0x000d, + 0x470: 0x0002, 0x471: 0x0002, 0x472: 0x0002, 0x473: 0x0002, 0x474: 0x0002, 0x475: 0x0002, + 0x476: 0x0002, 0x477: 0x0002, 0x478: 0x0002, 0x479: 0x0002, 0x47a: 0x000d, 0x47b: 0x000d, + 0x47c: 0x000d, 0x47d: 0x000d, 0x47e: 0x000d, 0x47f: 0x000d, + // Block 0x12, offset 0x480 + 0x480: 0x000d, 0x481: 0x000d, 0x482: 0x000d, 0x483: 0x000d, 0x484: 0x000d, 0x485: 0x000d, + 0x486: 0x000d, 0x487: 0x000d, 0x488: 0x000d, 0x489: 0x000d, 0x48a: 0x000d, 0x48b: 0x000d, + 0x48c: 0x000d, 0x48d: 0x000d, 0x48e: 0x000d, 0x48f: 0x000d, 0x490: 0x000d, 0x491: 0x000c, + 0x492: 0x000d, 0x493: 0x000d, 0x494: 0x000d, 0x495: 0x000d, 0x496: 0x000d, 0x497: 0x000d, + 0x498: 0x000d, 0x499: 0x000d, 0x49a: 0x000d, 0x49b: 0x000d, 0x49c: 0x000d, 0x49d: 0x000d, + 0x49e: 0x000d, 0x49f: 0x000d, 0x4a0: 0x000d, 0x4a1: 0x000d, 0x4a2: 0x000d, 0x4a3: 0x000d, + 0x4a4: 0x000d, 0x4a5: 0x000d, 0x4a6: 0x000d, 0x4a7: 0x000d, 0x4a8: 0x000d, 0x4a9: 0x000d, + 0x4aa: 0x000d, 0x4ab: 0x000d, 0x4ac: 0x000d, 0x4ad: 0x000d, 0x4ae: 0x000d, 0x4af: 0x000d, + 0x4b0: 0x000c, 0x4b1: 0x000c, 0x4b2: 0x000c, 0x4b3: 0x000c, 0x4b4: 0x000c, 0x4b5: 0x000c, + 0x4b6: 0x000c, 0x4b7: 0x000c, 0x4b8: 0x000c, 0x4b9: 0x000c, 0x4ba: 0x000c, 0x4bb: 0x000c, + 0x4bc: 0x000c, 0x4bd: 0x000c, 0x4be: 0x000c, 0x4bf: 0x000c, + // Block 0x13, offset 0x4c0 + 0x4c0: 0x000c, 0x4c1: 0x000c, 0x4c2: 0x000c, 0x4c3: 0x000c, 0x4c4: 0x000c, 0x4c5: 0x000c, + 0x4c6: 0x000c, 0x4c7: 0x000c, 0x4c8: 0x000c, 0x4c9: 0x000c, 0x4ca: 0x000c, 0x4cb: 0x000d, + 0x4cc: 0x000d, 0x4cd: 0x000d, 0x4ce: 0x000d, 0x4cf: 0x000d, 0x4d0: 0x000d, 0x4d1: 0x000d, + 0x4d2: 0x000d, 0x4d3: 0x000d, 0x4d4: 0x000d, 0x4d5: 0x000d, 0x4d6: 0x000d, 0x4d7: 0x000d, + 0x4d8: 0x000d, 0x4d9: 0x000d, 0x4da: 0x000d, 0x4db: 0x000d, 0x4dc: 0x000d, 0x4dd: 0x000d, + 0x4de: 0x000d, 0x4df: 0x000d, 0x4e0: 0x000d, 0x4e1: 0x000d, 0x4e2: 0x000d, 0x4e3: 0x000d, + 0x4e4: 0x000d, 0x4e5: 0x000d, 0x4e6: 0x000d, 0x4e7: 0x000d, 0x4e8: 0x000d, 0x4e9: 0x000d, + 0x4ea: 0x000d, 0x4eb: 0x000d, 0x4ec: 0x000d, 0x4ed: 0x000d, 0x4ee: 0x000d, 0x4ef: 0x000d, + 0x4f0: 0x000d, 0x4f1: 0x000d, 0x4f2: 0x000d, 0x4f3: 0x000d, 0x4f4: 0x000d, 0x4f5: 0x000d, + 0x4f6: 0x000d, 0x4f7: 0x000d, 0x4f8: 0x000d, 0x4f9: 0x000d, 0x4fa: 0x000d, 0x4fb: 0x000d, + 0x4fc: 0x000d, 0x4fd: 0x000d, 0x4fe: 0x000d, 0x4ff: 0x000d, + // Block 0x14, offset 0x500 + 0x500: 0x000d, 0x501: 0x000d, 0x502: 0x000d, 0x503: 0x000d, 0x504: 0x000d, 0x505: 0x000d, + 0x506: 0x000d, 0x507: 0x000d, 0x508: 0x000d, 0x509: 0x000d, 0x50a: 0x000d, 0x50b: 0x000d, + 0x50c: 0x000d, 0x50d: 0x000d, 0x50e: 0x000d, 0x50f: 0x000d, 0x510: 0x000d, 0x511: 0x000d, + 0x512: 0x000d, 0x513: 0x000d, 0x514: 0x000d, 0x515: 0x000d, 0x516: 0x000d, 0x517: 0x000d, + 0x518: 0x000d, 0x519: 0x000d, 0x51a: 0x000d, 0x51b: 0x000d, 0x51c: 0x000d, 0x51d: 0x000d, + 0x51e: 0x000d, 0x51f: 0x000d, 0x520: 0x000d, 0x521: 0x000d, 0x522: 0x000d, 0x523: 0x000d, + 0x524: 0x000d, 0x525: 0x000d, 0x526: 0x000c, 0x527: 0x000c, 0x528: 0x000c, 0x529: 0x000c, + 0x52a: 0x000c, 0x52b: 0x000c, 0x52c: 0x000c, 0x52d: 0x000c, 0x52e: 0x000c, 0x52f: 0x000c, + 0x530: 0x000c, 0x531: 0x000d, 0x532: 0x000d, 0x533: 0x000d, 0x534: 0x000d, 0x535: 0x000d, + 0x536: 0x000d, 0x537: 0x000d, 0x538: 0x000d, 0x539: 0x000d, 0x53a: 0x000d, 0x53b: 0x000d, + 0x53c: 0x000d, 0x53d: 0x000d, 0x53e: 0x000d, 0x53f: 0x000d, + // Block 0x15, offset 0x540 + 0x540: 0x0001, 0x541: 0x0001, 0x542: 0x0001, 0x543: 0x0001, 0x544: 0x0001, 0x545: 0x0001, + 0x546: 0x0001, 0x547: 0x0001, 0x548: 0x0001, 0x549: 0x0001, 0x54a: 0x0001, 0x54b: 0x0001, + 0x54c: 0x0001, 0x54d: 0x0001, 0x54e: 0x0001, 0x54f: 0x0001, 0x550: 0x0001, 0x551: 0x0001, + 0x552: 0x0001, 0x553: 0x0001, 0x554: 0x0001, 0x555: 0x0001, 0x556: 0x0001, 0x557: 0x0001, + 0x558: 0x0001, 0x559: 0x0001, 0x55a: 0x0001, 0x55b: 0x0001, 0x55c: 0x0001, 0x55d: 0x0001, + 0x55e: 0x0001, 0x55f: 0x0001, 0x560: 0x0001, 0x561: 0x0001, 0x562: 0x0001, 0x563: 0x0001, + 0x564: 0x0001, 0x565: 0x0001, 0x566: 0x0001, 0x567: 0x0001, 0x568: 0x0001, 0x569: 0x0001, + 0x56a: 0x0001, 0x56b: 0x000c, 0x56c: 0x000c, 0x56d: 0x000c, 0x56e: 0x000c, 0x56f: 0x000c, + 0x570: 0x000c, 0x571: 0x000c, 0x572: 0x000c, 0x573: 0x000c, 0x574: 0x0001, 0x575: 0x0001, + 0x576: 0x000a, 0x577: 0x000a, 0x578: 0x000a, 0x579: 0x000a, 0x57a: 0x0001, 0x57b: 0x0001, + 0x57c: 0x0001, 0x57d: 0x000c, 0x57e: 0x0001, 0x57f: 0x0001, + // Block 0x16, offset 0x580 + 0x580: 0x0001, 0x581: 0x0001, 0x582: 0x0001, 0x583: 0x0001, 0x584: 0x0001, 0x585: 0x0001, + 0x586: 0x0001, 0x587: 0x0001, 0x588: 0x0001, 0x589: 0x0001, 0x58a: 0x0001, 0x58b: 0x0001, + 0x58c: 0x0001, 0x58d: 0x0001, 0x58e: 0x0001, 0x58f: 0x0001, 0x590: 0x0001, 0x591: 0x0001, + 0x592: 0x0001, 0x593: 0x0001, 0x594: 0x0001, 0x595: 0x0001, 0x596: 0x000c, 0x597: 0x000c, + 0x598: 0x000c, 0x599: 0x000c, 0x59a: 0x0001, 0x59b: 0x000c, 0x59c: 0x000c, 0x59d: 0x000c, + 0x59e: 0x000c, 0x59f: 0x000c, 0x5a0: 0x000c, 0x5a1: 0x000c, 0x5a2: 0x000c, 0x5a3: 0x000c, + 0x5a4: 0x0001, 0x5a5: 0x000c, 0x5a6: 0x000c, 0x5a7: 0x000c, 0x5a8: 0x0001, 0x5a9: 0x000c, + 0x5aa: 0x000c, 0x5ab: 0x000c, 0x5ac: 0x000c, 0x5ad: 0x000c, 0x5ae: 0x0001, 0x5af: 0x0001, + 0x5b0: 0x0001, 0x5b1: 0x0001, 0x5b2: 0x0001, 0x5b3: 0x0001, 0x5b4: 0x0001, 0x5b5: 0x0001, + 0x5b6: 0x0001, 0x5b7: 0x0001, 0x5b8: 0x0001, 0x5b9: 0x0001, 0x5ba: 0x0001, 0x5bb: 0x0001, + 0x5bc: 0x0001, 0x5bd: 0x0001, 0x5be: 0x0001, 0x5bf: 0x0001, + // Block 0x17, offset 0x5c0 + 0x5c0: 0x0001, 0x5c1: 0x0001, 0x5c2: 0x0001, 0x5c3: 0x0001, 0x5c4: 0x0001, 0x5c5: 0x0001, + 0x5c6: 0x0001, 0x5c7: 0x0001, 0x5c8: 0x0001, 0x5c9: 0x0001, 0x5ca: 0x0001, 0x5cb: 0x0001, + 0x5cc: 0x0001, 0x5cd: 0x0001, 0x5ce: 0x0001, 0x5cf: 0x0001, 0x5d0: 0x0001, 0x5d1: 0x0001, + 0x5d2: 0x0001, 0x5d3: 0x0001, 0x5d4: 0x0001, 0x5d5: 0x0001, 0x5d6: 0x0001, 0x5d7: 0x0001, + 0x5d8: 0x0001, 0x5d9: 0x000c, 0x5da: 0x000c, 0x5db: 0x000c, 0x5dc: 0x0001, 0x5dd: 0x0001, + 0x5de: 0x0001, 0x5df: 0x0001, 0x5e0: 0x000d, 0x5e1: 0x000d, 0x5e2: 0x000d, 0x5e3: 0x000d, + 0x5e4: 0x000d, 0x5e5: 0x000d, 0x5e6: 0x000d, 0x5e7: 0x000d, 0x5e8: 0x000d, 0x5e9: 0x000d, + 0x5ea: 0x000d, 0x5eb: 0x000d, 0x5ec: 0x000d, 0x5ed: 0x000d, 0x5ee: 0x000d, 0x5ef: 0x000d, + 0x5f0: 0x0001, 0x5f1: 0x0001, 0x5f2: 0x0001, 0x5f3: 0x0001, 0x5f4: 0x0001, 0x5f5: 0x0001, + 0x5f6: 0x0001, 0x5f7: 0x0001, 0x5f8: 0x0001, 0x5f9: 0x0001, 0x5fa: 0x0001, 0x5fb: 0x0001, + 0x5fc: 0x0001, 0x5fd: 0x0001, 0x5fe: 0x0001, 0x5ff: 0x0001, + // Block 0x18, offset 0x600 + 0x600: 0x0001, 0x601: 0x0001, 0x602: 0x0001, 0x603: 0x0001, 0x604: 0x0001, 0x605: 0x0001, + 0x606: 0x0001, 0x607: 0x0001, 0x608: 0x0001, 0x609: 0x0001, 0x60a: 0x0001, 0x60b: 0x0001, + 0x60c: 0x0001, 0x60d: 0x0001, 0x60e: 0x0001, 0x60f: 0x0001, 0x610: 0x0001, 0x611: 0x0001, + 0x612: 0x0001, 0x613: 0x0001, 0x614: 0x0001, 0x615: 0x0001, 0x616: 0x0001, 0x617: 0x0001, + 0x618: 0x0001, 0x619: 0x0001, 0x61a: 0x0001, 0x61b: 0x0001, 0x61c: 0x0001, 0x61d: 0x0001, + 0x61e: 0x0001, 0x61f: 0x0001, 0x620: 0x000d, 0x621: 0x000d, 0x622: 0x000d, 0x623: 0x000d, + 0x624: 0x000d, 0x625: 0x000d, 0x626: 0x000d, 0x627: 0x000d, 0x628: 0x000d, 0x629: 0x000d, + 0x62a: 0x000d, 0x62b: 0x000d, 0x62c: 0x000d, 0x62d: 0x000d, 0x62e: 0x000d, 0x62f: 0x000d, + 0x630: 0x000d, 0x631: 0x000d, 0x632: 0x000d, 0x633: 0x000d, 0x634: 0x000d, 0x635: 0x000d, + 0x636: 0x000d, 0x637: 0x000d, 0x638: 0x000d, 0x639: 0x000d, 0x63a: 0x000d, 0x63b: 0x000d, + 0x63c: 0x000d, 0x63d: 0x000d, 0x63e: 0x000d, 0x63f: 0x000d, + // Block 0x19, offset 0x640 + 0x640: 0x000d, 0x641: 0x000d, 0x642: 0x000d, 0x643: 0x000d, 0x644: 0x000d, 0x645: 0x000d, + 0x646: 0x000d, 0x647: 0x000d, 0x648: 0x000d, 0x649: 0x000d, 0x64a: 0x000d, 0x64b: 0x000d, + 0x64c: 0x000d, 0x64d: 0x000d, 0x64e: 0x000d, 0x64f: 0x000d, 0x650: 0x000d, 0x651: 0x000d, + 0x652: 0x000d, 0x653: 0x000c, 0x654: 0x000c, 0x655: 0x000c, 0x656: 0x000c, 0x657: 0x000c, + 0x658: 0x000c, 0x659: 0x000c, 0x65a: 0x000c, 0x65b: 0x000c, 0x65c: 0x000c, 0x65d: 0x000c, + 0x65e: 0x000c, 0x65f: 0x000c, 0x660: 0x000c, 0x661: 0x000c, 0x662: 0x0005, 0x663: 0x000c, + 0x664: 0x000c, 0x665: 0x000c, 0x666: 0x000c, 0x667: 0x000c, 0x668: 0x000c, 0x669: 0x000c, + 0x66a: 0x000c, 0x66b: 0x000c, 0x66c: 0x000c, 0x66d: 0x000c, 0x66e: 0x000c, 0x66f: 0x000c, + 0x670: 0x000c, 0x671: 0x000c, 0x672: 0x000c, 0x673: 0x000c, 0x674: 0x000c, 0x675: 0x000c, + 0x676: 0x000c, 0x677: 0x000c, 0x678: 0x000c, 0x679: 0x000c, 0x67a: 0x000c, 0x67b: 0x000c, + 0x67c: 0x000c, 0x67d: 0x000c, 0x67e: 0x000c, 0x67f: 0x000c, + // Block 0x1a, offset 0x680 + 0x680: 0x000c, 0x681: 0x000c, 0x682: 0x000c, + 0x6ba: 0x000c, + 0x6bc: 0x000c, + // Block 0x1b, offset 0x6c0 + 0x6c1: 0x000c, 0x6c2: 0x000c, 0x6c3: 0x000c, 0x6c4: 0x000c, 0x6c5: 0x000c, + 0x6c6: 0x000c, 0x6c7: 0x000c, 0x6c8: 0x000c, + 0x6cd: 0x000c, 0x6d1: 0x000c, + 0x6d2: 0x000c, 0x6d3: 0x000c, 0x6d4: 0x000c, 0x6d5: 0x000c, 0x6d6: 0x000c, 0x6d7: 0x000c, + 0x6e2: 0x000c, 0x6e3: 0x000c, + // Block 0x1c, offset 0x700 + 0x701: 0x000c, + 0x73c: 0x000c, + // Block 0x1d, offset 0x740 + 0x741: 0x000c, 0x742: 0x000c, 0x743: 0x000c, 0x744: 0x000c, + 0x74d: 0x000c, + 0x762: 0x000c, 0x763: 0x000c, + 0x772: 0x0004, 0x773: 0x0004, + 0x77b: 0x0004, + 0x77e: 0x000c, + // Block 0x1e, offset 0x780 + 0x781: 0x000c, 0x782: 0x000c, + 0x7bc: 0x000c, + // Block 0x1f, offset 0x7c0 + 0x7c1: 0x000c, 0x7c2: 0x000c, + 0x7c7: 0x000c, 0x7c8: 0x000c, 0x7cb: 0x000c, + 0x7cc: 0x000c, 0x7cd: 0x000c, 0x7d1: 0x000c, + 0x7f0: 0x000c, 0x7f1: 0x000c, 0x7f5: 0x000c, + // Block 0x20, offset 0x800 + 0x801: 0x000c, 0x802: 0x000c, 0x803: 0x000c, 0x804: 0x000c, 0x805: 0x000c, + 0x807: 0x000c, 0x808: 0x000c, + 0x80d: 0x000c, + 0x822: 0x000c, 0x823: 0x000c, + 0x831: 0x0004, + 0x83a: 0x000c, 0x83b: 0x000c, + 0x83c: 0x000c, 0x83d: 0x000c, 0x83e: 0x000c, 0x83f: 0x000c, + // Block 0x21, offset 0x840 + 0x841: 0x000c, + 0x87c: 0x000c, 0x87f: 0x000c, + // Block 0x22, offset 0x880 + 0x881: 0x000c, 0x882: 0x000c, 0x883: 0x000c, 0x884: 0x000c, + 0x88d: 0x000c, + 0x896: 0x000c, + 0x8a2: 0x000c, 0x8a3: 0x000c, + // Block 0x23, offset 0x8c0 + 0x8c2: 0x000c, + // Block 0x24, offset 0x900 + 0x900: 0x000c, + 0x90d: 0x000c, + 0x933: 0x000a, 0x934: 0x000a, 0x935: 0x000a, + 0x936: 0x000a, 0x937: 0x000a, 0x938: 0x000a, 0x939: 0x0004, 0x93a: 0x000a, + // Block 0x25, offset 0x940 + 0x940: 0x000c, 0x944: 0x000c, + 0x97e: 0x000c, 0x97f: 0x000c, + // Block 0x26, offset 0x980 + 0x980: 0x000c, + 0x986: 0x000c, 0x987: 0x000c, 0x988: 0x000c, 0x98a: 0x000c, 0x98b: 0x000c, + 0x98c: 0x000c, 0x98d: 0x000c, + 0x995: 0x000c, 0x996: 0x000c, + 0x9a2: 0x000c, 0x9a3: 0x000c, + 0x9b8: 0x000a, 0x9b9: 0x000a, 0x9ba: 0x000a, 0x9bb: 0x000a, + 0x9bc: 0x000a, 0x9bd: 0x000a, 0x9be: 0x000a, + // Block 0x27, offset 0x9c0 + 0x9cc: 0x000c, 0x9cd: 0x000c, + 0x9e2: 0x000c, 0x9e3: 0x000c, + // Block 0x28, offset 0xa00 + 0xa00: 0x000c, 0xa01: 0x000c, + 0xa3b: 0x000c, + 0xa3c: 0x000c, + // Block 0x29, offset 0xa40 + 0xa41: 0x000c, 0xa42: 0x000c, 0xa43: 0x000c, 0xa44: 0x000c, + 0xa4d: 0x000c, + 0xa62: 0x000c, 0xa63: 0x000c, + // Block 0x2a, offset 0xa80 + 0xa8a: 0x000c, + 0xa92: 0x000c, 0xa93: 0x000c, 0xa94: 0x000c, 0xa96: 0x000c, + // Block 0x2b, offset 0xac0 + 0xaf1: 0x000c, 0xaf4: 0x000c, 0xaf5: 0x000c, + 0xaf6: 0x000c, 0xaf7: 0x000c, 0xaf8: 0x000c, 0xaf9: 0x000c, 0xafa: 0x000c, + 0xaff: 0x0004, + // Block 0x2c, offset 0xb00 + 0xb07: 0x000c, 0xb08: 0x000c, 0xb09: 0x000c, 0xb0a: 0x000c, 0xb0b: 0x000c, + 0xb0c: 0x000c, 0xb0d: 0x000c, 0xb0e: 0x000c, + // Block 0x2d, offset 0xb40 + 0xb71: 0x000c, 0xb74: 0x000c, 0xb75: 0x000c, + 0xb76: 0x000c, 0xb77: 0x000c, 0xb78: 0x000c, 0xb79: 0x000c, 0xb7a: 0x000c, 0xb7b: 0x000c, + 0xb7c: 0x000c, + // Block 0x2e, offset 0xb80 + 0xb88: 0x000c, 0xb89: 0x000c, 0xb8a: 0x000c, 0xb8b: 0x000c, + 0xb8c: 0x000c, 0xb8d: 0x000c, + // Block 0x2f, offset 0xbc0 + 0xbd8: 0x000c, 0xbd9: 0x000c, + 0xbf5: 0x000c, + 0xbf7: 0x000c, 0xbf9: 0x000c, 0xbfa: 0x003a, 0xbfb: 0x002a, + 0xbfc: 0x003a, 0xbfd: 0x002a, + // Block 0x30, offset 0xc00 + 0xc31: 0x000c, 0xc32: 0x000c, 0xc33: 0x000c, 0xc34: 0x000c, 0xc35: 0x000c, + 0xc36: 0x000c, 0xc37: 0x000c, 0xc38: 0x000c, 0xc39: 0x000c, 0xc3a: 0x000c, 0xc3b: 0x000c, + 0xc3c: 0x000c, 0xc3d: 0x000c, 0xc3e: 0x000c, + // Block 0x31, offset 0xc40 + 0xc40: 0x000c, 0xc41: 0x000c, 0xc42: 0x000c, 0xc43: 0x000c, 0xc44: 0x000c, + 0xc46: 0x000c, 0xc47: 0x000c, + 0xc4d: 0x000c, 0xc4e: 0x000c, 0xc4f: 0x000c, 0xc50: 0x000c, 0xc51: 0x000c, + 0xc52: 0x000c, 0xc53: 0x000c, 0xc54: 0x000c, 0xc55: 0x000c, 0xc56: 0x000c, 0xc57: 0x000c, + 0xc59: 0x000c, 0xc5a: 0x000c, 0xc5b: 0x000c, 0xc5c: 0x000c, 0xc5d: 0x000c, + 0xc5e: 0x000c, 0xc5f: 0x000c, 0xc60: 0x000c, 0xc61: 0x000c, 0xc62: 0x000c, 0xc63: 0x000c, + 0xc64: 0x000c, 0xc65: 0x000c, 0xc66: 0x000c, 0xc67: 0x000c, 0xc68: 0x000c, 0xc69: 0x000c, + 0xc6a: 0x000c, 0xc6b: 0x000c, 0xc6c: 0x000c, 0xc6d: 0x000c, 0xc6e: 0x000c, 0xc6f: 0x000c, + 0xc70: 0x000c, 0xc71: 0x000c, 0xc72: 0x000c, 0xc73: 0x000c, 0xc74: 0x000c, 0xc75: 0x000c, + 0xc76: 0x000c, 0xc77: 0x000c, 0xc78: 0x000c, 0xc79: 0x000c, 0xc7a: 0x000c, 0xc7b: 0x000c, + 0xc7c: 0x000c, + // Block 0x32, offset 0xc80 + 0xc86: 0x000c, + // Block 0x33, offset 0xcc0 + 0xced: 0x000c, 0xcee: 0x000c, 0xcef: 0x000c, + 0xcf0: 0x000c, 0xcf2: 0x000c, 0xcf3: 0x000c, 0xcf4: 0x000c, 0xcf5: 0x000c, + 0xcf6: 0x000c, 0xcf7: 0x000c, 0xcf9: 0x000c, 0xcfa: 0x000c, + 0xcfd: 0x000c, 0xcfe: 0x000c, + // Block 0x34, offset 0xd00 + 0xd18: 0x000c, 0xd19: 0x000c, + 0xd1e: 0x000c, 0xd1f: 0x000c, 0xd20: 0x000c, + 0xd31: 0x000c, 0xd32: 0x000c, 0xd33: 0x000c, 0xd34: 0x000c, + // Block 0x35, offset 0xd40 + 0xd42: 0x000c, 0xd45: 0x000c, + 0xd46: 0x000c, + 0xd4d: 0x000c, + 0xd5d: 0x000c, + // Block 0x36, offset 0xd80 + 0xd9d: 0x000c, + 0xd9e: 0x000c, 0xd9f: 0x000c, + // Block 0x37, offset 0xdc0 + 0xdd0: 0x000a, 0xdd1: 0x000a, + 0xdd2: 0x000a, 0xdd3: 0x000a, 0xdd4: 0x000a, 0xdd5: 0x000a, 0xdd6: 0x000a, 0xdd7: 0x000a, + 0xdd8: 0x000a, 0xdd9: 0x000a, + // Block 0x38, offset 0xe00 + 0xe00: 0x000a, + // Block 0x39, offset 0xe40 + 0xe40: 0x0009, + 0xe5b: 0x007a, 0xe5c: 0x006a, + // Block 0x3a, offset 0xe80 + 0xe92: 0x000c, 0xe93: 0x000c, 0xe94: 0x000c, + 0xeb2: 0x000c, 0xeb3: 0x000c, 0xeb4: 0x000c, + // Block 0x3b, offset 0xec0 + 0xed2: 0x000c, 0xed3: 0x000c, + 0xef2: 0x000c, 0xef3: 0x000c, + // Block 0x3c, offset 0xf00 + 0xf34: 0x000c, 0xf35: 0x000c, + 0xf37: 0x000c, 0xf38: 0x000c, 0xf39: 0x000c, 0xf3a: 0x000c, 0xf3b: 0x000c, + 0xf3c: 0x000c, 0xf3d: 0x000c, + // Block 0x3d, offset 0xf40 + 0xf46: 0x000c, 0xf49: 0x000c, 0xf4a: 0x000c, 0xf4b: 0x000c, + 0xf4c: 0x000c, 0xf4d: 0x000c, 0xf4e: 0x000c, 0xf4f: 0x000c, 0xf50: 0x000c, 0xf51: 0x000c, + 0xf52: 0x000c, 0xf53: 0x000c, + 0xf5b: 0x0004, 0xf5d: 0x000c, + 0xf70: 0x000a, 0xf71: 0x000a, 0xf72: 0x000a, 0xf73: 0x000a, 0xf74: 0x000a, 0xf75: 0x000a, + 0xf76: 0x000a, 0xf77: 0x000a, 0xf78: 0x000a, 0xf79: 0x000a, + // Block 0x3e, offset 0xf80 + 0xf80: 0x000a, 0xf81: 0x000a, 0xf82: 0x000a, 0xf83: 0x000a, 0xf84: 0x000a, 0xf85: 0x000a, + 0xf86: 0x000a, 0xf87: 0x000a, 0xf88: 0x000a, 0xf89: 0x000a, 0xf8a: 0x000a, 0xf8b: 0x000c, + 0xf8c: 0x000c, 0xf8d: 0x000c, 0xf8e: 0x000b, + // Block 0x3f, offset 0xfc0 + 0xfc5: 0x000c, + 0xfc6: 0x000c, + 0xfe9: 0x000c, + // Block 0x40, offset 0x1000 + 0x1020: 0x000c, 0x1021: 0x000c, 0x1022: 0x000c, + 0x1027: 0x000c, 0x1028: 0x000c, + 0x1032: 0x000c, + 0x1039: 0x000c, 0x103a: 0x000c, 0x103b: 0x000c, + // Block 0x41, offset 0x1040 + 0x1040: 0x000a, 0x1044: 0x000a, 0x1045: 0x000a, + // Block 0x42, offset 0x1080 + 0x109e: 0x000a, 0x109f: 0x000a, 0x10a0: 0x000a, 0x10a1: 0x000a, 0x10a2: 0x000a, 0x10a3: 0x000a, + 0x10a4: 0x000a, 0x10a5: 0x000a, 0x10a6: 0x000a, 0x10a7: 0x000a, 0x10a8: 0x000a, 0x10a9: 0x000a, + 0x10aa: 0x000a, 0x10ab: 0x000a, 0x10ac: 0x000a, 0x10ad: 0x000a, 0x10ae: 0x000a, 0x10af: 0x000a, + 0x10b0: 0x000a, 0x10b1: 0x000a, 0x10b2: 0x000a, 0x10b3: 0x000a, 0x10b4: 0x000a, 0x10b5: 0x000a, + 0x10b6: 0x000a, 0x10b7: 0x000a, 0x10b8: 0x000a, 0x10b9: 0x000a, 0x10ba: 0x000a, 0x10bb: 0x000a, + 0x10bc: 0x000a, 0x10bd: 0x000a, 0x10be: 0x000a, 0x10bf: 0x000a, + // Block 0x43, offset 0x10c0 + 0x10d7: 0x000c, + 0x10d8: 0x000c, 0x10db: 0x000c, + // Block 0x44, offset 0x1100 + 0x1116: 0x000c, + 0x1118: 0x000c, 0x1119: 0x000c, 0x111a: 0x000c, 0x111b: 0x000c, 0x111c: 0x000c, 0x111d: 0x000c, + 0x111e: 0x000c, 0x1120: 0x000c, 0x1122: 0x000c, + 0x1125: 0x000c, 0x1126: 0x000c, 0x1127: 0x000c, 0x1128: 0x000c, 0x1129: 0x000c, + 0x112a: 0x000c, 0x112b: 0x000c, 0x112c: 0x000c, + 0x1133: 0x000c, 0x1134: 0x000c, 0x1135: 0x000c, + 0x1136: 0x000c, 0x1137: 0x000c, 0x1138: 0x000c, 0x1139: 0x000c, 0x113a: 0x000c, 0x113b: 0x000c, + 0x113c: 0x000c, 0x113f: 0x000c, + // Block 0x45, offset 0x1140 + 0x1170: 0x000c, 0x1171: 0x000c, 0x1172: 0x000c, 0x1173: 0x000c, 0x1174: 0x000c, 0x1175: 0x000c, + 0x1176: 0x000c, 0x1177: 0x000c, 0x1178: 0x000c, 0x1179: 0x000c, 0x117a: 0x000c, 0x117b: 0x000c, + 0x117c: 0x000c, 0x117d: 0x000c, 0x117e: 0x000c, + // Block 0x46, offset 0x1180 + 0x1180: 0x000c, 0x1181: 0x000c, 0x1182: 0x000c, 0x1183: 0x000c, + 0x11b4: 0x000c, + 0x11b6: 0x000c, 0x11b7: 0x000c, 0x11b8: 0x000c, 0x11b9: 0x000c, 0x11ba: 0x000c, + 0x11bc: 0x000c, + // Block 0x47, offset 0x11c0 + 0x11c2: 0x000c, + 0x11eb: 0x000c, 0x11ec: 0x000c, 0x11ed: 0x000c, 0x11ee: 0x000c, 0x11ef: 0x000c, + 0x11f0: 0x000c, 0x11f1: 0x000c, 0x11f2: 0x000c, 0x11f3: 0x000c, + // Block 0x48, offset 0x1200 + 0x1200: 0x000c, 0x1201: 0x000c, + 0x1222: 0x000c, 0x1223: 0x000c, + 0x1224: 0x000c, 0x1225: 0x000c, 0x1228: 0x000c, 0x1229: 0x000c, + 0x122b: 0x000c, 0x122c: 0x000c, 0x122d: 0x000c, + // Block 0x49, offset 0x1240 + 0x1266: 0x000c, 0x1268: 0x000c, 0x1269: 0x000c, + 0x126d: 0x000c, 0x126f: 0x000c, + 0x1270: 0x000c, 0x1271: 0x000c, + // Block 0x4a, offset 0x1280 + 0x12ac: 0x000c, 0x12ad: 0x000c, 0x12ae: 0x000c, 0x12af: 0x000c, + 0x12b0: 0x000c, 0x12b1: 0x000c, 0x12b2: 0x000c, 0x12b3: 0x000c, + 0x12b6: 0x000c, 0x12b7: 0x000c, + // Block 0x4b, offset 0x12c0 + 0x12d0: 0x000c, 0x12d1: 0x000c, + 0x12d2: 0x000c, 0x12d4: 0x000c, 0x12d5: 0x000c, 0x12d6: 0x000c, 0x12d7: 0x000c, + 0x12d8: 0x000c, 0x12d9: 0x000c, 0x12da: 0x000c, 0x12db: 0x000c, 0x12dc: 0x000c, 0x12dd: 0x000c, + 0x12de: 0x000c, 0x12df: 0x000c, 0x12e0: 0x000c, 0x12e2: 0x000c, 0x12e3: 0x000c, + 0x12e4: 0x000c, 0x12e5: 0x000c, 0x12e6: 0x000c, 0x12e7: 0x000c, 0x12e8: 0x000c, + 0x12ed: 0x000c, + 0x12f4: 0x000c, + 0x12f8: 0x000c, 0x12f9: 0x000c, + // Block 0x4c, offset 0x1300 + 0x1300: 0x000c, 0x1301: 0x000c, 0x1302: 0x000c, 0x1303: 0x000c, 0x1304: 0x000c, 0x1305: 0x000c, + 0x1306: 0x000c, 0x1307: 0x000c, 0x1308: 0x000c, 0x1309: 0x000c, 0x130a: 0x000c, 0x130b: 0x000c, + 0x130c: 0x000c, 0x130d: 0x000c, 0x130e: 0x000c, 0x130f: 0x000c, 0x1310: 0x000c, 0x1311: 0x000c, + 0x1312: 0x000c, 0x1313: 0x000c, 0x1314: 0x000c, 0x1315: 0x000c, 0x1316: 0x000c, 0x1317: 0x000c, + 0x1318: 0x000c, 0x1319: 0x000c, 0x131a: 0x000c, 0x131b: 0x000c, 0x131c: 0x000c, 0x131d: 0x000c, + 0x131e: 0x000c, 0x131f: 0x000c, 0x1320: 0x000c, 0x1321: 0x000c, 0x1322: 0x000c, 0x1323: 0x000c, + 0x1324: 0x000c, 0x1325: 0x000c, 0x1326: 0x000c, 0x1327: 0x000c, 0x1328: 0x000c, 0x1329: 0x000c, + 0x132a: 0x000c, 0x132b: 0x000c, 0x132c: 0x000c, 0x132d: 0x000c, 0x132e: 0x000c, 0x132f: 0x000c, + 0x1330: 0x000c, 0x1331: 0x000c, 0x1332: 0x000c, 0x1333: 0x000c, 0x1334: 0x000c, 0x1335: 0x000c, + 0x1336: 0x000c, 0x1337: 0x000c, 0x1338: 0x000c, 0x1339: 0x000c, 0x133b: 0x000c, + 0x133c: 0x000c, 0x133d: 0x000c, 0x133e: 0x000c, 0x133f: 0x000c, + // Block 0x4d, offset 0x1340 + 0x137d: 0x000a, 0x137f: 0x000a, + // Block 0x4e, offset 0x1380 + 0x1380: 0x000a, 0x1381: 0x000a, + 0x138d: 0x000a, 0x138e: 0x000a, 0x138f: 0x000a, + 0x139d: 0x000a, + 0x139e: 0x000a, 0x139f: 0x000a, + 0x13ad: 0x000a, 0x13ae: 0x000a, 0x13af: 0x000a, + 0x13bd: 0x000a, 0x13be: 0x000a, + // Block 0x4f, offset 0x13c0 + 0x13c0: 0x0009, 0x13c1: 0x0009, 0x13c2: 0x0009, 0x13c3: 0x0009, 0x13c4: 0x0009, 0x13c5: 0x0009, + 0x13c6: 0x0009, 0x13c7: 0x0009, 0x13c8: 0x0009, 0x13c9: 0x0009, 0x13ca: 0x0009, 0x13cb: 0x000b, + 0x13cc: 0x000b, 0x13cd: 0x000b, 0x13cf: 0x0001, 0x13d0: 0x000a, 0x13d1: 0x000a, + 0x13d2: 0x000a, 0x13d3: 0x000a, 0x13d4: 0x000a, 0x13d5: 0x000a, 0x13d6: 0x000a, 0x13d7: 0x000a, + 0x13d8: 0x000a, 0x13d9: 0x000a, 0x13da: 0x000a, 0x13db: 0x000a, 0x13dc: 0x000a, 0x13dd: 0x000a, + 0x13de: 0x000a, 0x13df: 0x000a, 0x13e0: 0x000a, 0x13e1: 0x000a, 0x13e2: 0x000a, 0x13e3: 0x000a, + 0x13e4: 0x000a, 0x13e5: 0x000a, 0x13e6: 0x000a, 0x13e7: 0x000a, 0x13e8: 0x0009, 0x13e9: 0x0007, + 0x13ea: 0x000e, 0x13eb: 0x000e, 0x13ec: 0x000e, 0x13ed: 0x000e, 0x13ee: 0x000e, 0x13ef: 0x0006, + 0x13f0: 0x0004, 0x13f1: 0x0004, 0x13f2: 0x0004, 0x13f3: 0x0004, 0x13f4: 0x0004, 0x13f5: 0x000a, + 0x13f6: 0x000a, 0x13f7: 0x000a, 0x13f8: 0x000a, 0x13f9: 0x000a, 0x13fa: 0x000a, 0x13fb: 0x000a, + 0x13fc: 0x000a, 0x13fd: 0x000a, 0x13fe: 0x000a, 0x13ff: 0x000a, + // Block 0x50, offset 0x1400 + 0x1400: 0x000a, 0x1401: 0x000a, 0x1402: 0x000a, 0x1403: 0x000a, 0x1404: 0x0006, 0x1405: 0x009a, + 0x1406: 0x008a, 0x1407: 0x000a, 0x1408: 0x000a, 0x1409: 0x000a, 0x140a: 0x000a, 0x140b: 0x000a, + 0x140c: 0x000a, 0x140d: 0x000a, 0x140e: 0x000a, 0x140f: 0x000a, 0x1410: 0x000a, 0x1411: 0x000a, + 0x1412: 0x000a, 0x1413: 0x000a, 0x1414: 0x000a, 0x1415: 0x000a, 0x1416: 0x000a, 0x1417: 0x000a, + 0x1418: 0x000a, 0x1419: 0x000a, 0x141a: 0x000a, 0x141b: 0x000a, 0x141c: 0x000a, 0x141d: 0x000a, + 0x141e: 0x000a, 0x141f: 0x0009, 0x1420: 0x000b, 0x1421: 0x000b, 0x1422: 0x000b, 0x1423: 0x000b, + 0x1424: 0x000b, 0x1425: 0x000b, 0x1426: 0x000e, 0x1427: 0x000e, 0x1428: 0x000e, 0x1429: 0x000e, + 0x142a: 0x000b, 0x142b: 0x000b, 0x142c: 0x000b, 0x142d: 0x000b, 0x142e: 0x000b, 0x142f: 0x000b, + 0x1430: 0x0002, 0x1434: 0x0002, 0x1435: 0x0002, + 0x1436: 0x0002, 0x1437: 0x0002, 0x1438: 0x0002, 0x1439: 0x0002, 0x143a: 0x0003, 0x143b: 0x0003, + 0x143c: 0x000a, 0x143d: 0x009a, 0x143e: 0x008a, + // Block 0x51, offset 0x1440 + 0x1440: 0x0002, 0x1441: 0x0002, 0x1442: 0x0002, 0x1443: 0x0002, 0x1444: 0x0002, 0x1445: 0x0002, + 0x1446: 0x0002, 0x1447: 0x0002, 0x1448: 0x0002, 0x1449: 0x0002, 0x144a: 0x0003, 0x144b: 0x0003, + 0x144c: 0x000a, 0x144d: 0x009a, 0x144e: 0x008a, + 0x1460: 0x0004, 0x1461: 0x0004, 0x1462: 0x0004, 0x1463: 0x0004, + 0x1464: 0x0004, 0x1465: 0x0004, 0x1466: 0x0004, 0x1467: 0x0004, 0x1468: 0x0004, 0x1469: 0x0004, + 0x146a: 0x0004, 0x146b: 0x0004, 0x146c: 0x0004, 0x146d: 0x0004, 0x146e: 0x0004, 0x146f: 0x0004, + 0x1470: 0x0004, 0x1471: 0x0004, 0x1472: 0x0004, 0x1473: 0x0004, 0x1474: 0x0004, 0x1475: 0x0004, + 0x1476: 0x0004, 0x1477: 0x0004, 0x1478: 0x0004, 0x1479: 0x0004, 0x147a: 0x0004, 0x147b: 0x0004, + 0x147c: 0x0004, 0x147d: 0x0004, 0x147e: 0x0004, 0x147f: 0x0004, + // Block 0x52, offset 0x1480 + 0x1480: 0x0004, 0x1481: 0x0004, 0x1482: 0x0004, 0x1483: 0x0004, 0x1484: 0x0004, 0x1485: 0x0004, + 0x1486: 0x0004, 0x1487: 0x0004, 0x1488: 0x0004, 0x1489: 0x0004, 0x148a: 0x0004, 0x148b: 0x0004, + 0x148c: 0x0004, 0x148d: 0x0004, 0x148e: 0x0004, 0x148f: 0x0004, 0x1490: 0x000c, 0x1491: 0x000c, + 0x1492: 0x000c, 0x1493: 0x000c, 0x1494: 0x000c, 0x1495: 0x000c, 0x1496: 0x000c, 0x1497: 0x000c, + 0x1498: 0x000c, 0x1499: 0x000c, 0x149a: 0x000c, 0x149b: 0x000c, 0x149c: 0x000c, 0x149d: 0x000c, + 0x149e: 0x000c, 0x149f: 0x000c, 0x14a0: 0x000c, 0x14a1: 0x000c, 0x14a2: 0x000c, 0x14a3: 0x000c, + 0x14a4: 0x000c, 0x14a5: 0x000c, 0x14a6: 0x000c, 0x14a7: 0x000c, 0x14a8: 0x000c, 0x14a9: 0x000c, + 0x14aa: 0x000c, 0x14ab: 0x000c, 0x14ac: 0x000c, 0x14ad: 0x000c, 0x14ae: 0x000c, 0x14af: 0x000c, + 0x14b0: 0x000c, + // Block 0x53, offset 0x14c0 + 0x14c0: 0x000a, 0x14c1: 0x000a, 0x14c3: 0x000a, 0x14c4: 0x000a, 0x14c5: 0x000a, + 0x14c6: 0x000a, 0x14c8: 0x000a, 0x14c9: 0x000a, + 0x14d4: 0x000a, 0x14d6: 0x000a, 0x14d7: 0x000a, + 0x14d8: 0x000a, + 0x14de: 0x000a, 0x14df: 0x000a, 0x14e0: 0x000a, 0x14e1: 0x000a, 0x14e2: 0x000a, 0x14e3: 0x000a, + 0x14e5: 0x000a, 0x14e7: 0x000a, 0x14e9: 0x000a, + 0x14ee: 0x0004, + 0x14fa: 0x000a, 0x14fb: 0x000a, + // Block 0x54, offset 0x1500 + 0x1500: 0x000a, 0x1501: 0x000a, 0x1502: 0x000a, 0x1503: 0x000a, 0x1504: 0x000a, + 0x150a: 0x000a, 0x150b: 0x000a, + 0x150c: 0x000a, 0x150d: 0x000a, 0x1510: 0x000a, 0x1511: 0x000a, + 0x1512: 0x000a, 0x1513: 0x000a, 0x1514: 0x000a, 0x1515: 0x000a, 0x1516: 0x000a, 0x1517: 0x000a, + 0x1518: 0x000a, 0x1519: 0x000a, 0x151a: 0x000a, 0x151b: 0x000a, 0x151c: 0x000a, 0x151d: 0x000a, + 0x151e: 0x000a, 0x151f: 0x000a, + // Block 0x55, offset 0x1540 + 0x1549: 0x000a, 0x154a: 0x000a, 0x154b: 0x000a, + 0x1550: 0x000a, 0x1551: 0x000a, + 0x1552: 0x000a, 0x1553: 0x000a, 0x1554: 0x000a, 0x1555: 0x000a, 0x1556: 0x000a, 0x1557: 0x000a, + 0x1558: 0x000a, 0x1559: 0x000a, 0x155a: 0x000a, 0x155b: 0x000a, 0x155c: 0x000a, 0x155d: 0x000a, + 0x155e: 0x000a, 0x155f: 0x000a, 0x1560: 0x000a, 0x1561: 0x000a, 0x1562: 0x000a, 0x1563: 0x000a, + 0x1564: 0x000a, 0x1565: 0x000a, 0x1566: 0x000a, 0x1567: 0x000a, 0x1568: 0x000a, 0x1569: 0x000a, + 0x156a: 0x000a, 0x156b: 0x000a, 0x156c: 0x000a, 0x156d: 0x000a, 0x156e: 0x000a, 0x156f: 0x000a, + 0x1570: 0x000a, 0x1571: 0x000a, 0x1572: 0x000a, 0x1573: 0x000a, 0x1574: 0x000a, 0x1575: 0x000a, + 0x1576: 0x000a, 0x1577: 0x000a, 0x1578: 0x000a, 0x1579: 0x000a, 0x157a: 0x000a, 0x157b: 0x000a, + 0x157c: 0x000a, 0x157d: 0x000a, 0x157e: 0x000a, 0x157f: 0x000a, + // Block 0x56, offset 0x1580 + 0x1580: 0x000a, 0x1581: 0x000a, 0x1582: 0x000a, 0x1583: 0x000a, 0x1584: 0x000a, 0x1585: 0x000a, + 0x1586: 0x000a, 0x1587: 0x000a, 0x1588: 0x000a, 0x1589: 0x000a, 0x158a: 0x000a, 0x158b: 0x000a, + 0x158c: 0x000a, 0x158d: 0x000a, 0x158e: 0x000a, 0x158f: 0x000a, 0x1590: 0x000a, 0x1591: 0x000a, + 0x1592: 0x000a, 0x1593: 0x000a, 0x1594: 0x000a, 0x1595: 0x000a, 0x1596: 0x000a, 0x1597: 0x000a, + 0x1598: 0x000a, 0x1599: 0x000a, 0x159a: 0x000a, 0x159b: 0x000a, 0x159c: 0x000a, 0x159d: 0x000a, + 0x159e: 0x000a, 0x159f: 0x000a, 0x15a0: 0x000a, 0x15a1: 0x000a, 0x15a2: 0x000a, 0x15a3: 0x000a, + 0x15a4: 0x000a, 0x15a5: 0x000a, 0x15a6: 0x000a, 0x15a7: 0x000a, 0x15a8: 0x000a, 0x15a9: 0x000a, + 0x15aa: 0x000a, 0x15ab: 0x000a, 0x15ac: 0x000a, 0x15ad: 0x000a, 0x15ae: 0x000a, 0x15af: 0x000a, + 0x15b0: 0x000a, 0x15b1: 0x000a, 0x15b2: 0x000a, 0x15b3: 0x000a, 0x15b4: 0x000a, 0x15b5: 0x000a, + 0x15b6: 0x000a, 0x15b7: 0x000a, 0x15b8: 0x000a, 0x15b9: 0x000a, 0x15ba: 0x000a, 0x15bb: 0x000a, + 0x15bc: 0x000a, 0x15bd: 0x000a, 0x15be: 0x000a, 0x15bf: 0x000a, + // Block 0x57, offset 0x15c0 + 0x15c0: 0x000a, 0x15c1: 0x000a, 0x15c2: 0x000a, 0x15c3: 0x000a, 0x15c4: 0x000a, 0x15c5: 0x000a, + 0x15c6: 0x000a, 0x15c7: 0x000a, 0x15c8: 0x000a, 0x15c9: 0x000a, 0x15ca: 0x000a, 0x15cb: 0x000a, + 0x15cc: 0x000a, 0x15cd: 0x000a, 0x15ce: 0x000a, 0x15cf: 0x000a, 0x15d0: 0x000a, 0x15d1: 0x000a, + 0x15d2: 0x0003, 0x15d3: 0x0004, 0x15d4: 0x000a, 0x15d5: 0x000a, 0x15d6: 0x000a, 0x15d7: 0x000a, + 0x15d8: 0x000a, 0x15d9: 0x000a, 0x15da: 0x000a, 0x15db: 0x000a, 0x15dc: 0x000a, 0x15dd: 0x000a, + 0x15de: 0x000a, 0x15df: 0x000a, 0x15e0: 0x000a, 0x15e1: 0x000a, 0x15e2: 0x000a, 0x15e3: 0x000a, + 0x15e4: 0x000a, 0x15e5: 0x000a, 0x15e6: 0x000a, 0x15e7: 0x000a, 0x15e8: 0x000a, 0x15e9: 0x000a, + 0x15ea: 0x000a, 0x15eb: 0x000a, 0x15ec: 0x000a, 0x15ed: 0x000a, 0x15ee: 0x000a, 0x15ef: 0x000a, + 0x15f0: 0x000a, 0x15f1: 0x000a, 0x15f2: 0x000a, 0x15f3: 0x000a, 0x15f4: 0x000a, 0x15f5: 0x000a, + 0x15f6: 0x000a, 0x15f7: 0x000a, 0x15f8: 0x000a, 0x15f9: 0x000a, 0x15fa: 0x000a, 0x15fb: 0x000a, + 0x15fc: 0x000a, 0x15fd: 0x000a, 0x15fe: 0x000a, 0x15ff: 0x000a, + // Block 0x58, offset 0x1600 + 0x1600: 0x000a, 0x1601: 0x000a, 0x1602: 0x000a, 0x1603: 0x000a, 0x1604: 0x000a, 0x1605: 0x000a, + 0x1606: 0x000a, 0x1607: 0x000a, 0x1608: 0x003a, 0x1609: 0x002a, 0x160a: 0x003a, 0x160b: 0x002a, + 0x160c: 0x000a, 0x160d: 0x000a, 0x160e: 0x000a, 0x160f: 0x000a, 0x1610: 0x000a, 0x1611: 0x000a, + 0x1612: 0x000a, 0x1613: 0x000a, 0x1614: 0x000a, 0x1615: 0x000a, 0x1616: 0x000a, 0x1617: 0x000a, + 0x1618: 0x000a, 0x1619: 0x000a, 0x161a: 0x000a, 0x161b: 0x000a, 0x161c: 0x000a, 0x161d: 0x000a, + 0x161e: 0x000a, 0x161f: 0x000a, 0x1620: 0x000a, 0x1621: 0x000a, 0x1622: 0x000a, 0x1623: 0x000a, + 0x1624: 0x000a, 0x1625: 0x000a, 0x1626: 0x000a, 0x1627: 0x000a, 0x1628: 0x000a, 0x1629: 0x009a, + 0x162a: 0x008a, 0x162b: 0x000a, 0x162c: 0x000a, 0x162d: 0x000a, 0x162e: 0x000a, 0x162f: 0x000a, + 0x1630: 0x000a, 0x1631: 0x000a, 0x1632: 0x000a, 0x1633: 0x000a, 0x1634: 0x000a, 0x1635: 0x000a, + // Block 0x59, offset 0x1640 + 0x167b: 0x000a, + 0x167c: 0x000a, 0x167d: 0x000a, 0x167e: 0x000a, 0x167f: 0x000a, + // Block 0x5a, offset 0x1680 + 0x1680: 0x000a, 0x1681: 0x000a, 0x1682: 0x000a, 0x1683: 0x000a, 0x1684: 0x000a, 0x1685: 0x000a, + 0x1686: 0x000a, 0x1687: 0x000a, 0x1688: 0x000a, 0x1689: 0x000a, 0x168a: 0x000a, 0x168b: 0x000a, + 0x168c: 0x000a, 0x168d: 0x000a, 0x168e: 0x000a, 0x168f: 0x000a, 0x1690: 0x000a, 0x1691: 0x000a, + 0x1692: 0x000a, 0x1693: 0x000a, 0x1694: 0x000a, 0x1696: 0x000a, 0x1697: 0x000a, + 0x1698: 0x000a, 0x1699: 0x000a, 0x169a: 0x000a, 0x169b: 0x000a, 0x169c: 0x000a, 0x169d: 0x000a, + 0x169e: 0x000a, 0x169f: 0x000a, 0x16a0: 0x000a, 0x16a1: 0x000a, 0x16a2: 0x000a, 0x16a3: 0x000a, + 0x16a4: 0x000a, 0x16a5: 0x000a, 0x16a6: 0x000a, 0x16a7: 0x000a, 0x16a8: 0x000a, 0x16a9: 0x000a, + 0x16aa: 0x000a, 0x16ab: 0x000a, 0x16ac: 0x000a, 0x16ad: 0x000a, 0x16ae: 0x000a, 0x16af: 0x000a, + 0x16b0: 0x000a, 0x16b1: 0x000a, 0x16b2: 0x000a, 0x16b3: 0x000a, 0x16b4: 0x000a, 0x16b5: 0x000a, + 0x16b6: 0x000a, 0x16b7: 0x000a, 0x16b8: 0x000a, 0x16b9: 0x000a, 0x16ba: 0x000a, 0x16bb: 0x000a, + 0x16bc: 0x000a, 0x16bd: 0x000a, 0x16be: 0x000a, 0x16bf: 0x000a, + // Block 0x5b, offset 0x16c0 + 0x16c0: 0x000a, 0x16c1: 0x000a, 0x16c2: 0x000a, 0x16c3: 0x000a, 0x16c4: 0x000a, 0x16c5: 0x000a, + 0x16c6: 0x000a, 0x16c7: 0x000a, 0x16c8: 0x000a, 0x16c9: 0x000a, 0x16ca: 0x000a, 0x16cb: 0x000a, + 0x16cc: 0x000a, 0x16cd: 0x000a, 0x16ce: 0x000a, 0x16cf: 0x000a, 0x16d0: 0x000a, 0x16d1: 0x000a, + 0x16d2: 0x000a, 0x16d3: 0x000a, 0x16d4: 0x000a, 0x16d5: 0x000a, 0x16d6: 0x000a, 0x16d7: 0x000a, + 0x16d8: 0x000a, 0x16d9: 0x000a, 0x16da: 0x000a, 0x16db: 0x000a, 0x16dc: 0x000a, 0x16dd: 0x000a, + 0x16de: 0x000a, 0x16df: 0x000a, 0x16e0: 0x000a, 0x16e1: 0x000a, 0x16e2: 0x000a, 0x16e3: 0x000a, + 0x16e4: 0x000a, 0x16e5: 0x000a, 0x16e6: 0x000a, + // Block 0x5c, offset 0x1700 + 0x1700: 0x000a, 0x1701: 0x000a, 0x1702: 0x000a, 0x1703: 0x000a, 0x1704: 0x000a, 0x1705: 0x000a, + 0x1706: 0x000a, 0x1707: 0x000a, 0x1708: 0x000a, 0x1709: 0x000a, 0x170a: 0x000a, + 0x1720: 0x000a, 0x1721: 0x000a, 0x1722: 0x000a, 0x1723: 0x000a, + 0x1724: 0x000a, 0x1725: 0x000a, 0x1726: 0x000a, 0x1727: 0x000a, 0x1728: 0x000a, 0x1729: 0x000a, + 0x172a: 0x000a, 0x172b: 0x000a, 0x172c: 0x000a, 0x172d: 0x000a, 0x172e: 0x000a, 0x172f: 0x000a, + 0x1730: 0x000a, 0x1731: 0x000a, 0x1732: 0x000a, 0x1733: 0x000a, 0x1734: 0x000a, 0x1735: 0x000a, + 0x1736: 0x000a, 0x1737: 0x000a, 0x1738: 0x000a, 0x1739: 0x000a, 0x173a: 0x000a, 0x173b: 0x000a, + 0x173c: 0x000a, 0x173d: 0x000a, 0x173e: 0x000a, 0x173f: 0x000a, + // Block 0x5d, offset 0x1740 + 0x1740: 0x000a, 0x1741: 0x000a, 0x1742: 0x000a, 0x1743: 0x000a, 0x1744: 0x000a, 0x1745: 0x000a, + 0x1746: 0x000a, 0x1747: 0x000a, 0x1748: 0x0002, 0x1749: 0x0002, 0x174a: 0x0002, 0x174b: 0x0002, + 0x174c: 0x0002, 0x174d: 0x0002, 0x174e: 0x0002, 0x174f: 0x0002, 0x1750: 0x0002, 0x1751: 0x0002, + 0x1752: 0x0002, 0x1753: 0x0002, 0x1754: 0x0002, 0x1755: 0x0002, 0x1756: 0x0002, 0x1757: 0x0002, + 0x1758: 0x0002, 0x1759: 0x0002, 0x175a: 0x0002, 0x175b: 0x0002, + // Block 0x5e, offset 0x1780 + 0x17aa: 0x000a, 0x17ab: 0x000a, 0x17ac: 0x000a, 0x17ad: 0x000a, 0x17ae: 0x000a, 0x17af: 0x000a, + 0x17b0: 0x000a, 0x17b1: 0x000a, 0x17b2: 0x000a, 0x17b3: 0x000a, 0x17b4: 0x000a, 0x17b5: 0x000a, + 0x17b6: 0x000a, 0x17b7: 0x000a, 0x17b8: 0x000a, 0x17b9: 0x000a, 0x17ba: 0x000a, 0x17bb: 0x000a, + 0x17bc: 0x000a, 0x17bd: 0x000a, 0x17be: 0x000a, 0x17bf: 0x000a, + // Block 0x5f, offset 0x17c0 + 0x17c0: 0x000a, 0x17c1: 0x000a, 0x17c2: 0x000a, 0x17c3: 0x000a, 0x17c4: 0x000a, 0x17c5: 0x000a, + 0x17c6: 0x000a, 0x17c7: 0x000a, 0x17c8: 0x000a, 0x17c9: 0x000a, 0x17ca: 0x000a, 0x17cb: 0x000a, + 0x17cc: 0x000a, 0x17cd: 0x000a, 0x17ce: 0x000a, 0x17cf: 0x000a, 0x17d0: 0x000a, 0x17d1: 0x000a, + 0x17d2: 0x000a, 0x17d3: 0x000a, 0x17d4: 0x000a, 0x17d5: 0x000a, 0x17d6: 0x000a, 0x17d7: 0x000a, + 0x17d8: 0x000a, 0x17d9: 0x000a, 0x17da: 0x000a, 0x17db: 0x000a, 0x17dc: 0x000a, 0x17dd: 0x000a, + 0x17de: 0x000a, 0x17df: 0x000a, 0x17e0: 0x000a, 0x17e1: 0x000a, 0x17e2: 0x000a, 0x17e3: 0x000a, + 0x17e4: 0x000a, 0x17e5: 0x000a, 0x17e6: 0x000a, 0x17e7: 0x000a, 0x17e8: 0x000a, 0x17e9: 0x000a, + 0x17ea: 0x000a, 0x17eb: 0x000a, 0x17ed: 0x000a, 0x17ee: 0x000a, 0x17ef: 0x000a, + 0x17f0: 0x000a, 0x17f1: 0x000a, 0x17f2: 0x000a, 0x17f3: 0x000a, 0x17f4: 0x000a, 0x17f5: 0x000a, + 0x17f6: 0x000a, 0x17f7: 0x000a, 0x17f8: 0x000a, 0x17f9: 0x000a, 0x17fa: 0x000a, 0x17fb: 0x000a, + 0x17fc: 0x000a, 0x17fd: 0x000a, 0x17fe: 0x000a, 0x17ff: 0x000a, + // Block 0x60, offset 0x1800 + 0x1800: 0x000a, 0x1801: 0x000a, 0x1802: 0x000a, 0x1803: 0x000a, 0x1804: 0x000a, 0x1805: 0x000a, + 0x1806: 0x000a, 0x1807: 0x000a, 0x1808: 0x000a, 0x1809: 0x000a, 0x180a: 0x000a, 0x180b: 0x000a, + 0x180c: 0x000a, 0x180d: 0x000a, 0x180e: 0x000a, 0x180f: 0x000a, 0x1810: 0x000a, 0x1811: 0x000a, + 0x1812: 0x000a, 0x1813: 0x000a, 0x1814: 0x000a, 0x1815: 0x000a, 0x1816: 0x000a, 0x1817: 0x000a, + 0x1818: 0x000a, 0x1819: 0x000a, 0x181a: 0x000a, 0x181b: 0x000a, 0x181c: 0x000a, 0x181d: 0x000a, + 0x181e: 0x000a, 0x181f: 0x000a, 0x1820: 0x000a, 0x1821: 0x000a, 0x1822: 0x000a, 0x1823: 0x000a, + 0x1824: 0x000a, 0x1825: 0x000a, 0x1826: 0x000a, 0x1827: 0x000a, 0x1828: 0x003a, 0x1829: 0x002a, + 0x182a: 0x003a, 0x182b: 0x002a, 0x182c: 0x003a, 0x182d: 0x002a, 0x182e: 0x003a, 0x182f: 0x002a, + 0x1830: 0x003a, 0x1831: 0x002a, 0x1832: 0x003a, 0x1833: 0x002a, 0x1834: 0x003a, 0x1835: 0x002a, + 0x1836: 0x000a, 0x1837: 0x000a, 0x1838: 0x000a, 0x1839: 0x000a, 0x183a: 0x000a, 0x183b: 0x000a, + 0x183c: 0x000a, 0x183d: 0x000a, 0x183e: 0x000a, 0x183f: 0x000a, + // Block 0x61, offset 0x1840 + 0x1840: 0x000a, 0x1841: 0x000a, 0x1842: 0x000a, 0x1843: 0x000a, 0x1844: 0x000a, 0x1845: 0x009a, + 0x1846: 0x008a, 0x1847: 0x000a, 0x1848: 0x000a, 0x1849: 0x000a, 0x184a: 0x000a, 0x184b: 0x000a, + 0x184c: 0x000a, 0x184d: 0x000a, 0x184e: 0x000a, 0x184f: 0x000a, 0x1850: 0x000a, 0x1851: 0x000a, + 0x1852: 0x000a, 0x1853: 0x000a, 0x1854: 0x000a, 0x1855: 0x000a, 0x1856: 0x000a, 0x1857: 0x000a, + 0x1858: 0x000a, 0x1859: 0x000a, 0x185a: 0x000a, 0x185b: 0x000a, 0x185c: 0x000a, 0x185d: 0x000a, + 0x185e: 0x000a, 0x185f: 0x000a, 0x1860: 0x000a, 0x1861: 0x000a, 0x1862: 0x000a, 0x1863: 0x000a, + 0x1864: 0x000a, 0x1865: 0x000a, 0x1866: 0x003a, 0x1867: 0x002a, 0x1868: 0x003a, 0x1869: 0x002a, + 0x186a: 0x003a, 0x186b: 0x002a, 0x186c: 0x003a, 0x186d: 0x002a, 0x186e: 0x003a, 0x186f: 0x002a, + 0x1870: 0x000a, 0x1871: 0x000a, 0x1872: 0x000a, 0x1873: 0x000a, 0x1874: 0x000a, 0x1875: 0x000a, + 0x1876: 0x000a, 0x1877: 0x000a, 0x1878: 0x000a, 0x1879: 0x000a, 0x187a: 0x000a, 0x187b: 0x000a, + 0x187c: 0x000a, 0x187d: 0x000a, 0x187e: 0x000a, 0x187f: 0x000a, + // Block 0x62, offset 0x1880 + 0x1880: 0x000a, 0x1881: 0x000a, 0x1882: 0x000a, 0x1883: 0x007a, 0x1884: 0x006a, 0x1885: 0x009a, + 0x1886: 0x008a, 0x1887: 0x00ba, 0x1888: 0x00aa, 0x1889: 0x009a, 0x188a: 0x008a, 0x188b: 0x007a, + 0x188c: 0x006a, 0x188d: 0x00da, 0x188e: 0x002a, 0x188f: 0x003a, 0x1890: 0x00ca, 0x1891: 0x009a, + 0x1892: 0x008a, 0x1893: 0x007a, 0x1894: 0x006a, 0x1895: 0x009a, 0x1896: 0x008a, 0x1897: 0x00ba, + 0x1898: 0x00aa, 0x1899: 0x000a, 0x189a: 0x000a, 0x189b: 0x000a, 0x189c: 0x000a, 0x189d: 0x000a, + 0x189e: 0x000a, 0x189f: 0x000a, 0x18a0: 0x000a, 0x18a1: 0x000a, 0x18a2: 0x000a, 0x18a3: 0x000a, + 0x18a4: 0x000a, 0x18a5: 0x000a, 0x18a6: 0x000a, 0x18a7: 0x000a, 0x18a8: 0x000a, 0x18a9: 0x000a, + 0x18aa: 0x000a, 0x18ab: 0x000a, 0x18ac: 0x000a, 0x18ad: 0x000a, 0x18ae: 0x000a, 0x18af: 0x000a, + 0x18b0: 0x000a, 0x18b1: 0x000a, 0x18b2: 0x000a, 0x18b3: 0x000a, 0x18b4: 0x000a, 0x18b5: 0x000a, + 0x18b6: 0x000a, 0x18b7: 0x000a, 0x18b8: 0x000a, 0x18b9: 0x000a, 0x18ba: 0x000a, 0x18bb: 0x000a, + 0x18bc: 0x000a, 0x18bd: 0x000a, 0x18be: 0x000a, 0x18bf: 0x000a, + // Block 0x63, offset 0x18c0 + 0x18c0: 0x000a, 0x18c1: 0x000a, 0x18c2: 0x000a, 0x18c3: 0x000a, 0x18c4: 0x000a, 0x18c5: 0x000a, + 0x18c6: 0x000a, 0x18c7: 0x000a, 0x18c8: 0x000a, 0x18c9: 0x000a, 0x18ca: 0x000a, 0x18cb: 0x000a, + 0x18cc: 0x000a, 0x18cd: 0x000a, 0x18ce: 0x000a, 0x18cf: 0x000a, 0x18d0: 0x000a, 0x18d1: 0x000a, + 0x18d2: 0x000a, 0x18d3: 0x000a, 0x18d4: 0x000a, 0x18d5: 0x000a, 0x18d6: 0x000a, 0x18d7: 0x000a, + 0x18d8: 0x003a, 0x18d9: 0x002a, 0x18da: 0x003a, 0x18db: 0x002a, 0x18dc: 0x000a, 0x18dd: 0x000a, + 0x18de: 0x000a, 0x18df: 0x000a, 0x18e0: 0x000a, 0x18e1: 0x000a, 0x18e2: 0x000a, 0x18e3: 0x000a, + 0x18e4: 0x000a, 0x18e5: 0x000a, 0x18e6: 0x000a, 0x18e7: 0x000a, 0x18e8: 0x000a, 0x18e9: 0x000a, + 0x18ea: 0x000a, 0x18eb: 0x000a, 0x18ec: 0x000a, 0x18ed: 0x000a, 0x18ee: 0x000a, 0x18ef: 0x000a, + 0x18f0: 0x000a, 0x18f1: 0x000a, 0x18f2: 0x000a, 0x18f3: 0x000a, 0x18f4: 0x000a, 0x18f5: 0x000a, + 0x18f6: 0x000a, 0x18f7: 0x000a, 0x18f8: 0x000a, 0x18f9: 0x000a, 0x18fa: 0x000a, 0x18fb: 0x000a, + 0x18fc: 0x003a, 0x18fd: 0x002a, 0x18fe: 0x000a, 0x18ff: 0x000a, + // Block 0x64, offset 0x1900 + 0x1900: 0x000a, 0x1901: 0x000a, 0x1902: 0x000a, 0x1903: 0x000a, 0x1904: 0x000a, 0x1905: 0x000a, + 0x1906: 0x000a, 0x1907: 0x000a, 0x1908: 0x000a, 0x1909: 0x000a, 0x190a: 0x000a, 0x190b: 0x000a, + 0x190c: 0x000a, 0x190d: 0x000a, 0x190e: 0x000a, 0x190f: 0x000a, 0x1910: 0x000a, 0x1911: 0x000a, + 0x1912: 0x000a, 0x1913: 0x000a, 0x1914: 0x000a, 0x1915: 0x000a, 0x1916: 0x000a, 0x1917: 0x000a, + 0x1918: 0x000a, 0x1919: 0x000a, 0x191a: 0x000a, 0x191b: 0x000a, 0x191c: 0x000a, 0x191d: 0x000a, + 0x191e: 0x000a, 0x191f: 0x000a, 0x1920: 0x000a, 0x1921: 0x000a, 0x1922: 0x000a, 0x1923: 0x000a, + 0x1924: 0x000a, 0x1925: 0x000a, 0x1926: 0x000a, 0x1927: 0x000a, 0x1928: 0x000a, 0x1929: 0x000a, + 0x192a: 0x000a, 0x192b: 0x000a, 0x192c: 0x000a, 0x192d: 0x000a, 0x192e: 0x000a, 0x192f: 0x000a, + 0x1930: 0x000a, 0x1931: 0x000a, 0x1932: 0x000a, 0x1933: 0x000a, + 0x1936: 0x000a, 0x1937: 0x000a, 0x1938: 0x000a, 0x1939: 0x000a, 0x193a: 0x000a, 0x193b: 0x000a, + 0x193c: 0x000a, 0x193d: 0x000a, 0x193e: 0x000a, 0x193f: 0x000a, + // Block 0x65, offset 0x1940 + 0x1940: 0x000a, 0x1941: 0x000a, 0x1942: 0x000a, 0x1943: 0x000a, 0x1944: 0x000a, 0x1945: 0x000a, + 0x1946: 0x000a, 0x1947: 0x000a, 0x1948: 0x000a, 0x1949: 0x000a, 0x194a: 0x000a, 0x194b: 0x000a, + 0x194c: 0x000a, 0x194d: 0x000a, 0x194e: 0x000a, 0x194f: 0x000a, 0x1950: 0x000a, 0x1951: 0x000a, + 0x1952: 0x000a, 0x1953: 0x000a, 0x1954: 0x000a, 0x1955: 0x000a, + 0x1958: 0x000a, 0x1959: 0x000a, 0x195a: 0x000a, 0x195b: 0x000a, 0x195c: 0x000a, 0x195d: 0x000a, + 0x195e: 0x000a, 0x195f: 0x000a, 0x1960: 0x000a, 0x1961: 0x000a, 0x1962: 0x000a, 0x1963: 0x000a, + 0x1964: 0x000a, 0x1965: 0x000a, 0x1966: 0x000a, 0x1967: 0x000a, 0x1968: 0x000a, 0x1969: 0x000a, + 0x196a: 0x000a, 0x196b: 0x000a, 0x196c: 0x000a, 0x196d: 0x000a, 0x196e: 0x000a, 0x196f: 0x000a, + 0x1970: 0x000a, 0x1971: 0x000a, 0x1972: 0x000a, 0x1973: 0x000a, 0x1974: 0x000a, 0x1975: 0x000a, + 0x1976: 0x000a, 0x1977: 0x000a, 0x1978: 0x000a, 0x1979: 0x000a, 0x197a: 0x000a, 0x197b: 0x000a, + 0x197c: 0x000a, 0x197d: 0x000a, 0x197e: 0x000a, 0x197f: 0x000a, + // Block 0x66, offset 0x1980 + 0x19a5: 0x000a, 0x19a6: 0x000a, 0x19a7: 0x000a, 0x19a8: 0x000a, 0x19a9: 0x000a, + 0x19aa: 0x000a, 0x19af: 0x000c, + 0x19b0: 0x000c, 0x19b1: 0x000c, + 0x19b9: 0x000a, 0x19ba: 0x000a, 0x19bb: 0x000a, + 0x19bc: 0x000a, 0x19bd: 0x000a, 0x19be: 0x000a, 0x19bf: 0x000a, + // Block 0x67, offset 0x19c0 + 0x19ff: 0x000c, + // Block 0x68, offset 0x1a00 + 0x1a20: 0x000c, 0x1a21: 0x000c, 0x1a22: 0x000c, 0x1a23: 0x000c, + 0x1a24: 0x000c, 0x1a25: 0x000c, 0x1a26: 0x000c, 0x1a27: 0x000c, 0x1a28: 0x000c, 0x1a29: 0x000c, + 0x1a2a: 0x000c, 0x1a2b: 0x000c, 0x1a2c: 0x000c, 0x1a2d: 0x000c, 0x1a2e: 0x000c, 0x1a2f: 0x000c, + 0x1a30: 0x000c, 0x1a31: 0x000c, 0x1a32: 0x000c, 0x1a33: 0x000c, 0x1a34: 0x000c, 0x1a35: 0x000c, + 0x1a36: 0x000c, 0x1a37: 0x000c, 0x1a38: 0x000c, 0x1a39: 0x000c, 0x1a3a: 0x000c, 0x1a3b: 0x000c, + 0x1a3c: 0x000c, 0x1a3d: 0x000c, 0x1a3e: 0x000c, 0x1a3f: 0x000c, + // Block 0x69, offset 0x1a40 + 0x1a40: 0x000a, 0x1a41: 0x000a, 0x1a42: 0x000a, 0x1a43: 0x000a, 0x1a44: 0x000a, 0x1a45: 0x000a, + 0x1a46: 0x000a, 0x1a47: 0x000a, 0x1a48: 0x000a, 0x1a49: 0x000a, 0x1a4a: 0x000a, 0x1a4b: 0x000a, + 0x1a4c: 0x000a, 0x1a4d: 0x000a, 0x1a4e: 0x000a, 0x1a4f: 0x000a, 0x1a50: 0x000a, 0x1a51: 0x000a, + 0x1a52: 0x000a, 0x1a53: 0x000a, 0x1a54: 0x000a, 0x1a55: 0x000a, 0x1a56: 0x000a, 0x1a57: 0x000a, + 0x1a58: 0x000a, 0x1a59: 0x000a, 0x1a5a: 0x000a, 0x1a5b: 0x000a, 0x1a5c: 0x000a, 0x1a5d: 0x000a, + 0x1a5e: 0x000a, 0x1a5f: 0x000a, 0x1a60: 0x000a, 0x1a61: 0x000a, 0x1a62: 0x003a, 0x1a63: 0x002a, + 0x1a64: 0x003a, 0x1a65: 0x002a, 0x1a66: 0x003a, 0x1a67: 0x002a, 0x1a68: 0x003a, 0x1a69: 0x002a, + 0x1a6a: 0x000a, 0x1a6b: 0x000a, 0x1a6c: 0x000a, 0x1a6d: 0x000a, 0x1a6e: 0x000a, 0x1a6f: 0x000a, + 0x1a70: 0x000a, 0x1a71: 0x000a, 0x1a72: 0x000a, 0x1a73: 0x000a, 0x1a74: 0x000a, 0x1a75: 0x000a, + 0x1a76: 0x000a, 0x1a77: 0x000a, 0x1a78: 0x000a, 0x1a79: 0x000a, 0x1a7a: 0x000a, 0x1a7b: 0x000a, + 0x1a7c: 0x000a, 0x1a7d: 0x000a, 0x1a7e: 0x000a, 0x1a7f: 0x000a, + // Block 0x6a, offset 0x1a80 + 0x1a80: 0x000a, 0x1a81: 0x000a, 0x1a82: 0x000a, 0x1a83: 0x000a, 0x1a84: 0x000a, 0x1a85: 0x000a, + 0x1a86: 0x000a, 0x1a87: 0x000a, 0x1a88: 0x000a, 0x1a89: 0x000a, 0x1a8a: 0x000a, 0x1a8b: 0x000a, + 0x1a8c: 0x000a, 0x1a8d: 0x000a, 0x1a8e: 0x000a, 0x1a8f: 0x000a, + // Block 0x6b, offset 0x1ac0 + 0x1ac0: 0x000a, 0x1ac1: 0x000a, 0x1ac2: 0x000a, 0x1ac3: 0x000a, 0x1ac4: 0x000a, 0x1ac5: 0x000a, + 0x1ac6: 0x000a, 0x1ac7: 0x000a, 0x1ac8: 0x000a, 0x1ac9: 0x000a, 0x1aca: 0x000a, 0x1acb: 0x000a, + 0x1acc: 0x000a, 0x1acd: 0x000a, 0x1ace: 0x000a, 0x1acf: 0x000a, 0x1ad0: 0x000a, 0x1ad1: 0x000a, + 0x1ad2: 0x000a, 0x1ad3: 0x000a, 0x1ad4: 0x000a, 0x1ad5: 0x000a, 0x1ad6: 0x000a, 0x1ad7: 0x000a, + 0x1ad8: 0x000a, 0x1ad9: 0x000a, 0x1adb: 0x000a, 0x1adc: 0x000a, 0x1add: 0x000a, + 0x1ade: 0x000a, 0x1adf: 0x000a, 0x1ae0: 0x000a, 0x1ae1: 0x000a, 0x1ae2: 0x000a, 0x1ae3: 0x000a, + 0x1ae4: 0x000a, 0x1ae5: 0x000a, 0x1ae6: 0x000a, 0x1ae7: 0x000a, 0x1ae8: 0x000a, 0x1ae9: 0x000a, + 0x1aea: 0x000a, 0x1aeb: 0x000a, 0x1aec: 0x000a, 0x1aed: 0x000a, 0x1aee: 0x000a, 0x1aef: 0x000a, + 0x1af0: 0x000a, 0x1af1: 0x000a, 0x1af2: 0x000a, 0x1af3: 0x000a, 0x1af4: 0x000a, 0x1af5: 0x000a, + 0x1af6: 0x000a, 0x1af7: 0x000a, 0x1af8: 0x000a, 0x1af9: 0x000a, 0x1afa: 0x000a, 0x1afb: 0x000a, + 0x1afc: 0x000a, 0x1afd: 0x000a, 0x1afe: 0x000a, 0x1aff: 0x000a, + // Block 0x6c, offset 0x1b00 + 0x1b00: 0x000a, 0x1b01: 0x000a, 0x1b02: 0x000a, 0x1b03: 0x000a, 0x1b04: 0x000a, 0x1b05: 0x000a, + 0x1b06: 0x000a, 0x1b07: 0x000a, 0x1b08: 0x000a, 0x1b09: 0x000a, 0x1b0a: 0x000a, 0x1b0b: 0x000a, + 0x1b0c: 0x000a, 0x1b0d: 0x000a, 0x1b0e: 0x000a, 0x1b0f: 0x000a, 0x1b10: 0x000a, 0x1b11: 0x000a, + 0x1b12: 0x000a, 0x1b13: 0x000a, 0x1b14: 0x000a, 0x1b15: 0x000a, 0x1b16: 0x000a, 0x1b17: 0x000a, + 0x1b18: 0x000a, 0x1b19: 0x000a, 0x1b1a: 0x000a, 0x1b1b: 0x000a, 0x1b1c: 0x000a, 0x1b1d: 0x000a, + 0x1b1e: 0x000a, 0x1b1f: 0x000a, 0x1b20: 0x000a, 0x1b21: 0x000a, 0x1b22: 0x000a, 0x1b23: 0x000a, + 0x1b24: 0x000a, 0x1b25: 0x000a, 0x1b26: 0x000a, 0x1b27: 0x000a, 0x1b28: 0x000a, 0x1b29: 0x000a, + 0x1b2a: 0x000a, 0x1b2b: 0x000a, 0x1b2c: 0x000a, 0x1b2d: 0x000a, 0x1b2e: 0x000a, 0x1b2f: 0x000a, + 0x1b30: 0x000a, 0x1b31: 0x000a, 0x1b32: 0x000a, 0x1b33: 0x000a, + // Block 0x6d, offset 0x1b40 + 0x1b40: 0x000a, 0x1b41: 0x000a, 0x1b42: 0x000a, 0x1b43: 0x000a, 0x1b44: 0x000a, 0x1b45: 0x000a, + 0x1b46: 0x000a, 0x1b47: 0x000a, 0x1b48: 0x000a, 0x1b49: 0x000a, 0x1b4a: 0x000a, 0x1b4b: 0x000a, + 0x1b4c: 0x000a, 0x1b4d: 0x000a, 0x1b4e: 0x000a, 0x1b4f: 0x000a, 0x1b50: 0x000a, 0x1b51: 0x000a, + 0x1b52: 0x000a, 0x1b53: 0x000a, 0x1b54: 0x000a, 0x1b55: 0x000a, + 0x1b70: 0x000a, 0x1b71: 0x000a, 0x1b72: 0x000a, 0x1b73: 0x000a, 0x1b74: 0x000a, 0x1b75: 0x000a, + 0x1b76: 0x000a, 0x1b77: 0x000a, 0x1b78: 0x000a, 0x1b79: 0x000a, 0x1b7a: 0x000a, 0x1b7b: 0x000a, + // Block 0x6e, offset 0x1b80 + 0x1b80: 0x0009, 0x1b81: 0x000a, 0x1b82: 0x000a, 0x1b83: 0x000a, 0x1b84: 0x000a, + 0x1b88: 0x003a, 0x1b89: 0x002a, 0x1b8a: 0x003a, 0x1b8b: 0x002a, + 0x1b8c: 0x003a, 0x1b8d: 0x002a, 0x1b8e: 0x003a, 0x1b8f: 0x002a, 0x1b90: 0x003a, 0x1b91: 0x002a, + 0x1b92: 0x000a, 0x1b93: 0x000a, 0x1b94: 0x003a, 0x1b95: 0x002a, 0x1b96: 0x003a, 0x1b97: 0x002a, + 0x1b98: 0x003a, 0x1b99: 0x002a, 0x1b9a: 0x003a, 0x1b9b: 0x002a, 0x1b9c: 0x000a, 0x1b9d: 0x000a, + 0x1b9e: 0x000a, 0x1b9f: 0x000a, 0x1ba0: 0x000a, + 0x1baa: 0x000c, 0x1bab: 0x000c, 0x1bac: 0x000c, 0x1bad: 0x000c, + 0x1bb0: 0x000a, + 0x1bb6: 0x000a, 0x1bb7: 0x000a, + 0x1bbd: 0x000a, 0x1bbe: 0x000a, 0x1bbf: 0x000a, + // Block 0x6f, offset 0x1bc0 + 0x1bd9: 0x000c, 0x1bda: 0x000c, 0x1bdb: 0x000a, 0x1bdc: 0x000a, + 0x1be0: 0x000a, + // Block 0x70, offset 0x1c00 + 0x1c3b: 0x000a, + // Block 0x71, offset 0x1c40 + 0x1c40: 0x000a, 0x1c41: 0x000a, 0x1c42: 0x000a, 0x1c43: 0x000a, 0x1c44: 0x000a, 0x1c45: 0x000a, + 0x1c46: 0x000a, 0x1c47: 0x000a, 0x1c48: 0x000a, 0x1c49: 0x000a, 0x1c4a: 0x000a, 0x1c4b: 0x000a, + 0x1c4c: 0x000a, 0x1c4d: 0x000a, 0x1c4e: 0x000a, 0x1c4f: 0x000a, 0x1c50: 0x000a, 0x1c51: 0x000a, + 0x1c52: 0x000a, 0x1c53: 0x000a, 0x1c54: 0x000a, 0x1c55: 0x000a, 0x1c56: 0x000a, 0x1c57: 0x000a, + 0x1c58: 0x000a, 0x1c59: 0x000a, 0x1c5a: 0x000a, 0x1c5b: 0x000a, 0x1c5c: 0x000a, 0x1c5d: 0x000a, + 0x1c5e: 0x000a, 0x1c5f: 0x000a, 0x1c60: 0x000a, 0x1c61: 0x000a, 0x1c62: 0x000a, 0x1c63: 0x000a, + // Block 0x72, offset 0x1c80 + 0x1c9d: 0x000a, + 0x1c9e: 0x000a, + // Block 0x73, offset 0x1cc0 + 0x1cd0: 0x000a, 0x1cd1: 0x000a, + 0x1cd2: 0x000a, 0x1cd3: 0x000a, 0x1cd4: 0x000a, 0x1cd5: 0x000a, 0x1cd6: 0x000a, 0x1cd7: 0x000a, + 0x1cd8: 0x000a, 0x1cd9: 0x000a, 0x1cda: 0x000a, 0x1cdb: 0x000a, 0x1cdc: 0x000a, 0x1cdd: 0x000a, + 0x1cde: 0x000a, 0x1cdf: 0x000a, + 0x1cfc: 0x000a, 0x1cfd: 0x000a, 0x1cfe: 0x000a, + // Block 0x74, offset 0x1d00 + 0x1d31: 0x000a, 0x1d32: 0x000a, 0x1d33: 0x000a, 0x1d34: 0x000a, 0x1d35: 0x000a, + 0x1d36: 0x000a, 0x1d37: 0x000a, 0x1d38: 0x000a, 0x1d39: 0x000a, 0x1d3a: 0x000a, 0x1d3b: 0x000a, + 0x1d3c: 0x000a, 0x1d3d: 0x000a, 0x1d3e: 0x000a, 0x1d3f: 0x000a, + // Block 0x75, offset 0x1d40 + 0x1d4c: 0x000a, 0x1d4d: 0x000a, 0x1d4e: 0x000a, 0x1d4f: 0x000a, + // Block 0x76, offset 0x1d80 + 0x1db7: 0x000a, 0x1db8: 0x000a, 0x1db9: 0x000a, 0x1dba: 0x000a, + // Block 0x77, offset 0x1dc0 + 0x1dde: 0x000a, 0x1ddf: 0x000a, + 0x1dff: 0x000a, + // Block 0x78, offset 0x1e00 + 0x1e10: 0x000a, 0x1e11: 0x000a, + 0x1e12: 0x000a, 0x1e13: 0x000a, 0x1e14: 0x000a, 0x1e15: 0x000a, 0x1e16: 0x000a, 0x1e17: 0x000a, + 0x1e18: 0x000a, 0x1e19: 0x000a, 0x1e1a: 0x000a, 0x1e1b: 0x000a, 0x1e1c: 0x000a, 0x1e1d: 0x000a, + 0x1e1e: 0x000a, 0x1e1f: 0x000a, 0x1e20: 0x000a, 0x1e21: 0x000a, 0x1e22: 0x000a, 0x1e23: 0x000a, + 0x1e24: 0x000a, 0x1e25: 0x000a, 0x1e26: 0x000a, 0x1e27: 0x000a, 0x1e28: 0x000a, 0x1e29: 0x000a, + 0x1e2a: 0x000a, 0x1e2b: 0x000a, 0x1e2c: 0x000a, 0x1e2d: 0x000a, 0x1e2e: 0x000a, 0x1e2f: 0x000a, + 0x1e30: 0x000a, 0x1e31: 0x000a, 0x1e32: 0x000a, 0x1e33: 0x000a, 0x1e34: 0x000a, 0x1e35: 0x000a, + 0x1e36: 0x000a, 0x1e37: 0x000a, 0x1e38: 0x000a, 0x1e39: 0x000a, 0x1e3a: 0x000a, 0x1e3b: 0x000a, + 0x1e3c: 0x000a, 0x1e3d: 0x000a, 0x1e3e: 0x000a, 0x1e3f: 0x000a, + // Block 0x79, offset 0x1e40 + 0x1e40: 0x000a, 0x1e41: 0x000a, 0x1e42: 0x000a, 0x1e43: 0x000a, 0x1e44: 0x000a, 0x1e45: 0x000a, + 0x1e46: 0x000a, + // Block 0x7a, offset 0x1e80 + 0x1e8d: 0x000a, 0x1e8e: 0x000a, 0x1e8f: 0x000a, + // Block 0x7b, offset 0x1ec0 + 0x1eef: 0x000c, + 0x1ef0: 0x000c, 0x1ef1: 0x000c, 0x1ef2: 0x000c, 0x1ef3: 0x000a, 0x1ef4: 0x000c, 0x1ef5: 0x000c, + 0x1ef6: 0x000c, 0x1ef7: 0x000c, 0x1ef8: 0x000c, 0x1ef9: 0x000c, 0x1efa: 0x000c, 0x1efb: 0x000c, + 0x1efc: 0x000c, 0x1efd: 0x000c, 0x1efe: 0x000a, 0x1eff: 0x000a, + // Block 0x7c, offset 0x1f00 + 0x1f1e: 0x000c, 0x1f1f: 0x000c, + // Block 0x7d, offset 0x1f40 + 0x1f70: 0x000c, 0x1f71: 0x000c, + // Block 0x7e, offset 0x1f80 + 0x1f80: 0x000a, 0x1f81: 0x000a, 0x1f82: 0x000a, 0x1f83: 0x000a, 0x1f84: 0x000a, 0x1f85: 0x000a, + 0x1f86: 0x000a, 0x1f87: 0x000a, 0x1f88: 0x000a, 0x1f89: 0x000a, 0x1f8a: 0x000a, 0x1f8b: 0x000a, + 0x1f8c: 0x000a, 0x1f8d: 0x000a, 0x1f8e: 0x000a, 0x1f8f: 0x000a, 0x1f90: 0x000a, 0x1f91: 0x000a, + 0x1f92: 0x000a, 0x1f93: 0x000a, 0x1f94: 0x000a, 0x1f95: 0x000a, 0x1f96: 0x000a, 0x1f97: 0x000a, + 0x1f98: 0x000a, 0x1f99: 0x000a, 0x1f9a: 0x000a, 0x1f9b: 0x000a, 0x1f9c: 0x000a, 0x1f9d: 0x000a, + 0x1f9e: 0x000a, 0x1f9f: 0x000a, 0x1fa0: 0x000a, 0x1fa1: 0x000a, + // Block 0x7f, offset 0x1fc0 + 0x1fc8: 0x000a, + // Block 0x80, offset 0x2000 + 0x2002: 0x000c, + 0x2006: 0x000c, 0x200b: 0x000c, + 0x2025: 0x000c, 0x2026: 0x000c, 0x2028: 0x000a, 0x2029: 0x000a, + 0x202a: 0x000a, 0x202b: 0x000a, + 0x2038: 0x0004, 0x2039: 0x0004, + // Block 0x81, offset 0x2040 + 0x2074: 0x000a, 0x2075: 0x000a, + 0x2076: 0x000a, 0x2077: 0x000a, + // Block 0x82, offset 0x2080 + 0x2084: 0x000c, 0x2085: 0x000c, + 0x20a0: 0x000c, 0x20a1: 0x000c, 0x20a2: 0x000c, 0x20a3: 0x000c, + 0x20a4: 0x000c, 0x20a5: 0x000c, 0x20a6: 0x000c, 0x20a7: 0x000c, 0x20a8: 0x000c, 0x20a9: 0x000c, + 0x20aa: 0x000c, 0x20ab: 0x000c, 0x20ac: 0x000c, 0x20ad: 0x000c, 0x20ae: 0x000c, 0x20af: 0x000c, + 0x20b0: 0x000c, 0x20b1: 0x000c, + 0x20bf: 0x000c, + // Block 0x83, offset 0x20c0 + 0x20e6: 0x000c, 0x20e7: 0x000c, 0x20e8: 0x000c, 0x20e9: 0x000c, + 0x20ea: 0x000c, 0x20eb: 0x000c, 0x20ec: 0x000c, 0x20ed: 0x000c, + // Block 0x84, offset 0x2100 + 0x2107: 0x000c, 0x2108: 0x000c, 0x2109: 0x000c, 0x210a: 0x000c, 0x210b: 0x000c, + 0x210c: 0x000c, 0x210d: 0x000c, 0x210e: 0x000c, 0x210f: 0x000c, 0x2110: 0x000c, 0x2111: 0x000c, + // Block 0x85, offset 0x2140 + 0x2140: 0x000c, 0x2141: 0x000c, 0x2142: 0x000c, + 0x2173: 0x000c, + 0x2176: 0x000c, 0x2177: 0x000c, 0x2178: 0x000c, 0x2179: 0x000c, + 0x217c: 0x000c, 0x217d: 0x000c, + // Block 0x86, offset 0x2180 + 0x21a5: 0x000c, + // Block 0x87, offset 0x21c0 + 0x21e9: 0x000c, + 0x21ea: 0x000c, 0x21eb: 0x000c, 0x21ec: 0x000c, 0x21ed: 0x000c, 0x21ee: 0x000c, + 0x21f1: 0x000c, 0x21f2: 0x000c, 0x21f5: 0x000c, + 0x21f6: 0x000c, + // Block 0x88, offset 0x2200 + 0x2203: 0x000c, + 0x220c: 0x000c, + 0x223c: 0x000c, + // Block 0x89, offset 0x2240 + 0x2270: 0x000c, 0x2272: 0x000c, 0x2273: 0x000c, 0x2274: 0x000c, + 0x2277: 0x000c, 0x2278: 0x000c, + 0x227e: 0x000c, 0x227f: 0x000c, + // Block 0x8a, offset 0x2280 + 0x2281: 0x000c, + 0x22ac: 0x000c, 0x22ad: 0x000c, + 0x22b6: 0x000c, + // Block 0x8b, offset 0x22c0 + 0x22e5: 0x000c, 0x22e8: 0x000c, + 0x22ed: 0x000c, + // Block 0x8c, offset 0x2300 + 0x231d: 0x0001, + 0x231e: 0x000c, 0x231f: 0x0001, 0x2320: 0x0001, 0x2321: 0x0001, 0x2322: 0x0001, 0x2323: 0x0001, + 0x2324: 0x0001, 0x2325: 0x0001, 0x2326: 0x0001, 0x2327: 0x0001, 0x2328: 0x0001, 0x2329: 0x0003, + 0x232a: 0x0001, 0x232b: 0x0001, 0x232c: 0x0001, 0x232d: 0x0001, 0x232e: 0x0001, 0x232f: 0x0001, + 0x2330: 0x0001, 0x2331: 0x0001, 0x2332: 0x0001, 0x2333: 0x0001, 0x2334: 0x0001, 0x2335: 0x0001, + 0x2336: 0x0001, 0x2337: 0x0001, 0x2338: 0x0001, 0x2339: 0x0001, 0x233a: 0x0001, 0x233b: 0x0001, + 0x233c: 0x0001, 0x233d: 0x0001, 0x233e: 0x0001, 0x233f: 0x0001, + // Block 0x8d, offset 0x2340 + 0x2340: 0x0001, 0x2341: 0x0001, 0x2342: 0x0001, 0x2343: 0x0001, 0x2344: 0x0001, 0x2345: 0x0001, + 0x2346: 0x0001, 0x2347: 0x0001, 0x2348: 0x0001, 0x2349: 0x0001, 0x234a: 0x0001, 0x234b: 0x0001, + 0x234c: 0x0001, 0x234d: 0x0001, 0x234e: 0x0001, 0x234f: 0x0001, 0x2350: 0x000d, 0x2351: 0x000d, + 0x2352: 0x000d, 0x2353: 0x000d, 0x2354: 0x000d, 0x2355: 0x000d, 0x2356: 0x000d, 0x2357: 0x000d, + 0x2358: 0x000d, 0x2359: 0x000d, 0x235a: 0x000d, 0x235b: 0x000d, 0x235c: 0x000d, 0x235d: 0x000d, + 0x235e: 0x000d, 0x235f: 0x000d, 0x2360: 0x000d, 0x2361: 0x000d, 0x2362: 0x000d, 0x2363: 0x000d, + 0x2364: 0x000d, 0x2365: 0x000d, 0x2366: 0x000d, 0x2367: 0x000d, 0x2368: 0x000d, 0x2369: 0x000d, + 0x236a: 0x000d, 0x236b: 0x000d, 0x236c: 0x000d, 0x236d: 0x000d, 0x236e: 0x000d, 0x236f: 0x000d, + 0x2370: 0x000d, 0x2371: 0x000d, 0x2372: 0x000d, 0x2373: 0x000d, 0x2374: 0x000d, 0x2375: 0x000d, + 0x2376: 0x000d, 0x2377: 0x000d, 0x2378: 0x000d, 0x2379: 0x000d, 0x237a: 0x000d, 0x237b: 0x000d, + 0x237c: 0x000d, 0x237d: 0x000d, 0x237e: 0x000d, 0x237f: 0x000d, + // Block 0x8e, offset 0x2380 + 0x2380: 0x000d, 0x2381: 0x000d, 0x2382: 0x000d, 0x2383: 0x000d, 0x2384: 0x000d, 0x2385: 0x000d, + 0x2386: 0x000d, 0x2387: 0x000d, 0x2388: 0x000d, 0x2389: 0x000d, 0x238a: 0x000d, 0x238b: 0x000d, + 0x238c: 0x000d, 0x238d: 0x000d, 0x238e: 0x000d, 0x238f: 0x000d, 0x2390: 0x000d, 0x2391: 0x000d, + 0x2392: 0x000d, 0x2393: 0x000d, 0x2394: 0x000d, 0x2395: 0x000d, 0x2396: 0x000d, 0x2397: 0x000d, + 0x2398: 0x000d, 0x2399: 0x000d, 0x239a: 0x000d, 0x239b: 0x000d, 0x239c: 0x000d, 0x239d: 0x000d, + 0x239e: 0x000d, 0x239f: 0x000d, 0x23a0: 0x000d, 0x23a1: 0x000d, 0x23a2: 0x000d, 0x23a3: 0x000d, + 0x23a4: 0x000d, 0x23a5: 0x000d, 0x23a6: 0x000d, 0x23a7: 0x000d, 0x23a8: 0x000d, 0x23a9: 0x000d, + 0x23aa: 0x000d, 0x23ab: 0x000d, 0x23ac: 0x000d, 0x23ad: 0x000d, 0x23ae: 0x000d, 0x23af: 0x000d, + 0x23b0: 0x000d, 0x23b1: 0x000d, 0x23b2: 0x000d, 0x23b3: 0x000d, 0x23b4: 0x000d, 0x23b5: 0x000d, + 0x23b6: 0x000d, 0x23b7: 0x000d, 0x23b8: 0x000d, 0x23b9: 0x000d, 0x23ba: 0x000d, 0x23bb: 0x000d, + 0x23bc: 0x000d, 0x23bd: 0x000d, 0x23be: 0x000a, 0x23bf: 0x000a, + // Block 0x8f, offset 0x23c0 + 0x23c0: 0x000d, 0x23c1: 0x000d, 0x23c2: 0x000d, 0x23c3: 0x000d, 0x23c4: 0x000d, 0x23c5: 0x000d, + 0x23c6: 0x000d, 0x23c7: 0x000d, 0x23c8: 0x000d, 0x23c9: 0x000d, 0x23ca: 0x000d, 0x23cb: 0x000d, + 0x23cc: 0x000d, 0x23cd: 0x000d, 0x23ce: 0x000d, 0x23cf: 0x000d, 0x23d0: 0x000b, 0x23d1: 0x000b, + 0x23d2: 0x000b, 0x23d3: 0x000b, 0x23d4: 0x000b, 0x23d5: 0x000b, 0x23d6: 0x000b, 0x23d7: 0x000b, + 0x23d8: 0x000b, 0x23d9: 0x000b, 0x23da: 0x000b, 0x23db: 0x000b, 0x23dc: 0x000b, 0x23dd: 0x000b, + 0x23de: 0x000b, 0x23df: 0x000b, 0x23e0: 0x000b, 0x23e1: 0x000b, 0x23e2: 0x000b, 0x23e3: 0x000b, + 0x23e4: 0x000b, 0x23e5: 0x000b, 0x23e6: 0x000b, 0x23e7: 0x000b, 0x23e8: 0x000b, 0x23e9: 0x000b, + 0x23ea: 0x000b, 0x23eb: 0x000b, 0x23ec: 0x000b, 0x23ed: 0x000b, 0x23ee: 0x000b, 0x23ef: 0x000b, + 0x23f0: 0x000d, 0x23f1: 0x000d, 0x23f2: 0x000d, 0x23f3: 0x000d, 0x23f4: 0x000d, 0x23f5: 0x000d, + 0x23f6: 0x000d, 0x23f7: 0x000d, 0x23f8: 0x000d, 0x23f9: 0x000d, 0x23fa: 0x000d, 0x23fb: 0x000d, + 0x23fc: 0x000d, 0x23fd: 0x000a, 0x23fe: 0x000d, 0x23ff: 0x000d, + // Block 0x90, offset 0x2400 + 0x2400: 0x000c, 0x2401: 0x000c, 0x2402: 0x000c, 0x2403: 0x000c, 0x2404: 0x000c, 0x2405: 0x000c, + 0x2406: 0x000c, 0x2407: 0x000c, 0x2408: 0x000c, 0x2409: 0x000c, 0x240a: 0x000c, 0x240b: 0x000c, + 0x240c: 0x000c, 0x240d: 0x000c, 0x240e: 0x000c, 0x240f: 0x000c, 0x2410: 0x000a, 0x2411: 0x000a, + 0x2412: 0x000a, 0x2413: 0x000a, 0x2414: 0x000a, 0x2415: 0x000a, 0x2416: 0x000a, 0x2417: 0x000a, + 0x2418: 0x000a, 0x2419: 0x000a, + 0x2420: 0x000c, 0x2421: 0x000c, 0x2422: 0x000c, 0x2423: 0x000c, + 0x2424: 0x000c, 0x2425: 0x000c, 0x2426: 0x000c, 0x2427: 0x000c, 0x2428: 0x000c, 0x2429: 0x000c, + 0x242a: 0x000c, 0x242b: 0x000c, 0x242c: 0x000c, 0x242d: 0x000c, 0x242e: 0x000c, 0x242f: 0x000c, + 0x2430: 0x000a, 0x2431: 0x000a, 0x2432: 0x000a, 0x2433: 0x000a, 0x2434: 0x000a, 0x2435: 0x000a, + 0x2436: 0x000a, 0x2437: 0x000a, 0x2438: 0x000a, 0x2439: 0x000a, 0x243a: 0x000a, 0x243b: 0x000a, + 0x243c: 0x000a, 0x243d: 0x000a, 0x243e: 0x000a, 0x243f: 0x000a, + // Block 0x91, offset 0x2440 + 0x2440: 0x000a, 0x2441: 0x000a, 0x2442: 0x000a, 0x2443: 0x000a, 0x2444: 0x000a, 0x2445: 0x000a, + 0x2446: 0x000a, 0x2447: 0x000a, 0x2448: 0x000a, 0x2449: 0x000a, 0x244a: 0x000a, 0x244b: 0x000a, + 0x244c: 0x000a, 0x244d: 0x000a, 0x244e: 0x000a, 0x244f: 0x000a, 0x2450: 0x0006, 0x2451: 0x000a, + 0x2452: 0x0006, 0x2454: 0x000a, 0x2455: 0x0006, 0x2456: 0x000a, 0x2457: 0x000a, + 0x2458: 0x000a, 0x2459: 0x009a, 0x245a: 0x008a, 0x245b: 0x007a, 0x245c: 0x006a, 0x245d: 0x009a, + 0x245e: 0x008a, 0x245f: 0x0004, 0x2460: 0x000a, 0x2461: 0x000a, 0x2462: 0x0003, 0x2463: 0x0003, + 0x2464: 0x000a, 0x2465: 0x000a, 0x2466: 0x000a, 0x2468: 0x000a, 0x2469: 0x0004, + 0x246a: 0x0004, 0x246b: 0x000a, + 0x2470: 0x000d, 0x2471: 0x000d, 0x2472: 0x000d, 0x2473: 0x000d, 0x2474: 0x000d, 0x2475: 0x000d, + 0x2476: 0x000d, 0x2477: 0x000d, 0x2478: 0x000d, 0x2479: 0x000d, 0x247a: 0x000d, 0x247b: 0x000d, + 0x247c: 0x000d, 0x247d: 0x000d, 0x247e: 0x000d, 0x247f: 0x000d, + // Block 0x92, offset 0x2480 + 0x2480: 0x000d, 0x2481: 0x000d, 0x2482: 0x000d, 0x2483: 0x000d, 0x2484: 0x000d, 0x2485: 0x000d, + 0x2486: 0x000d, 0x2487: 0x000d, 0x2488: 0x000d, 0x2489: 0x000d, 0x248a: 0x000d, 0x248b: 0x000d, + 0x248c: 0x000d, 0x248d: 0x000d, 0x248e: 0x000d, 0x248f: 0x000d, 0x2490: 0x000d, 0x2491: 0x000d, + 0x2492: 0x000d, 0x2493: 0x000d, 0x2494: 0x000d, 0x2495: 0x000d, 0x2496: 0x000d, 0x2497: 0x000d, + 0x2498: 0x000d, 0x2499: 0x000d, 0x249a: 0x000d, 0x249b: 0x000d, 0x249c: 0x000d, 0x249d: 0x000d, + 0x249e: 0x000d, 0x249f: 0x000d, 0x24a0: 0x000d, 0x24a1: 0x000d, 0x24a2: 0x000d, 0x24a3: 0x000d, + 0x24a4: 0x000d, 0x24a5: 0x000d, 0x24a6: 0x000d, 0x24a7: 0x000d, 0x24a8: 0x000d, 0x24a9: 0x000d, + 0x24aa: 0x000d, 0x24ab: 0x000d, 0x24ac: 0x000d, 0x24ad: 0x000d, 0x24ae: 0x000d, 0x24af: 0x000d, + 0x24b0: 0x000d, 0x24b1: 0x000d, 0x24b2: 0x000d, 0x24b3: 0x000d, 0x24b4: 0x000d, 0x24b5: 0x000d, + 0x24b6: 0x000d, 0x24b7: 0x000d, 0x24b8: 0x000d, 0x24b9: 0x000d, 0x24ba: 0x000d, 0x24bb: 0x000d, + 0x24bc: 0x000d, 0x24bd: 0x000d, 0x24be: 0x000d, 0x24bf: 0x000b, + // Block 0x93, offset 0x24c0 + 0x24c1: 0x000a, 0x24c2: 0x000a, 0x24c3: 0x0004, 0x24c4: 0x0004, 0x24c5: 0x0004, + 0x24c6: 0x000a, 0x24c7: 0x000a, 0x24c8: 0x003a, 0x24c9: 0x002a, 0x24ca: 0x000a, 0x24cb: 0x0003, + 0x24cc: 0x0006, 0x24cd: 0x0003, 0x24ce: 0x0006, 0x24cf: 0x0006, 0x24d0: 0x0002, 0x24d1: 0x0002, + 0x24d2: 0x0002, 0x24d3: 0x0002, 0x24d4: 0x0002, 0x24d5: 0x0002, 0x24d6: 0x0002, 0x24d7: 0x0002, + 0x24d8: 0x0002, 0x24d9: 0x0002, 0x24da: 0x0006, 0x24db: 0x000a, 0x24dc: 0x000a, 0x24dd: 0x000a, + 0x24de: 0x000a, 0x24df: 0x000a, 0x24e0: 0x000a, + 0x24fb: 0x005a, + 0x24fc: 0x000a, 0x24fd: 0x004a, 0x24fe: 0x000a, 0x24ff: 0x000a, + // Block 0x94, offset 0x2500 + 0x2500: 0x000a, + 0x251b: 0x005a, 0x251c: 0x000a, 0x251d: 0x004a, + 0x251e: 0x000a, 0x251f: 0x00fa, 0x2520: 0x00ea, 0x2521: 0x000a, 0x2522: 0x003a, 0x2523: 0x002a, + 0x2524: 0x000a, 0x2525: 0x000a, + // Block 0x95, offset 0x2540 + 0x2560: 0x0004, 0x2561: 0x0004, 0x2562: 0x000a, 0x2563: 0x000a, + 0x2564: 0x000a, 0x2565: 0x0004, 0x2566: 0x0004, 0x2568: 0x000a, 0x2569: 0x000a, + 0x256a: 0x000a, 0x256b: 0x000a, 0x256c: 0x000a, 0x256d: 0x000a, 0x256e: 0x000a, + 0x2570: 0x000b, 0x2571: 0x000b, 0x2572: 0x000b, 0x2573: 0x000b, 0x2574: 0x000b, 0x2575: 0x000b, + 0x2576: 0x000b, 0x2577: 0x000b, 0x2578: 0x000b, 0x2579: 0x000a, 0x257a: 0x000a, 0x257b: 0x000a, + 0x257c: 0x000a, 0x257d: 0x000a, 0x257e: 0x000b, 0x257f: 0x000b, + // Block 0x96, offset 0x2580 + 0x2581: 0x000a, + // Block 0x97, offset 0x25c0 + 0x25c0: 0x000a, 0x25c1: 0x000a, 0x25c2: 0x000a, 0x25c3: 0x000a, 0x25c4: 0x000a, 0x25c5: 0x000a, + 0x25c6: 0x000a, 0x25c7: 0x000a, 0x25c8: 0x000a, 0x25c9: 0x000a, 0x25ca: 0x000a, 0x25cb: 0x000a, + 0x25cc: 0x000a, 0x25d0: 0x000a, 0x25d1: 0x000a, + 0x25d2: 0x000a, 0x25d3: 0x000a, 0x25d4: 0x000a, 0x25d5: 0x000a, 0x25d6: 0x000a, 0x25d7: 0x000a, + 0x25d8: 0x000a, 0x25d9: 0x000a, 0x25da: 0x000a, 0x25db: 0x000a, + 0x25e0: 0x000a, + // Block 0x98, offset 0x2600 + 0x263d: 0x000c, + // Block 0x99, offset 0x2640 + 0x2660: 0x000c, 0x2661: 0x0002, 0x2662: 0x0002, 0x2663: 0x0002, + 0x2664: 0x0002, 0x2665: 0x0002, 0x2666: 0x0002, 0x2667: 0x0002, 0x2668: 0x0002, 0x2669: 0x0002, + 0x266a: 0x0002, 0x266b: 0x0002, 0x266c: 0x0002, 0x266d: 0x0002, 0x266e: 0x0002, 0x266f: 0x0002, + 0x2670: 0x0002, 0x2671: 0x0002, 0x2672: 0x0002, 0x2673: 0x0002, 0x2674: 0x0002, 0x2675: 0x0002, + 0x2676: 0x0002, 0x2677: 0x0002, 0x2678: 0x0002, 0x2679: 0x0002, 0x267a: 0x0002, 0x267b: 0x0002, + // Block 0x9a, offset 0x2680 + 0x26b6: 0x000c, 0x26b7: 0x000c, 0x26b8: 0x000c, 0x26b9: 0x000c, 0x26ba: 0x000c, + // Block 0x9b, offset 0x26c0 + 0x26c0: 0x0001, 0x26c1: 0x0001, 0x26c2: 0x0001, 0x26c3: 0x0001, 0x26c4: 0x0001, 0x26c5: 0x0001, + 0x26c6: 0x0001, 0x26c7: 0x0001, 0x26c8: 0x0001, 0x26c9: 0x0001, 0x26ca: 0x0001, 0x26cb: 0x0001, + 0x26cc: 0x0001, 0x26cd: 0x0001, 0x26ce: 0x0001, 0x26cf: 0x0001, 0x26d0: 0x0001, 0x26d1: 0x0001, + 0x26d2: 0x0001, 0x26d3: 0x0001, 0x26d4: 0x0001, 0x26d5: 0x0001, 0x26d6: 0x0001, 0x26d7: 0x0001, + 0x26d8: 0x0001, 0x26d9: 0x0001, 0x26da: 0x0001, 0x26db: 0x0001, 0x26dc: 0x0001, 0x26dd: 0x0001, + 0x26de: 0x0001, 0x26df: 0x0001, 0x26e0: 0x0001, 0x26e1: 0x0001, 0x26e2: 0x0001, 0x26e3: 0x0001, + 0x26e4: 0x0001, 0x26e5: 0x0001, 0x26e6: 0x0001, 0x26e7: 0x0001, 0x26e8: 0x0001, 0x26e9: 0x0001, + 0x26ea: 0x0001, 0x26eb: 0x0001, 0x26ec: 0x0001, 0x26ed: 0x0001, 0x26ee: 0x0001, 0x26ef: 0x0001, + 0x26f0: 0x0001, 0x26f1: 0x0001, 0x26f2: 0x0001, 0x26f3: 0x0001, 0x26f4: 0x0001, 0x26f5: 0x0001, + 0x26f6: 0x0001, 0x26f7: 0x0001, 0x26f8: 0x0001, 0x26f9: 0x0001, 0x26fa: 0x0001, 0x26fb: 0x0001, + 0x26fc: 0x0001, 0x26fd: 0x0001, 0x26fe: 0x0001, 0x26ff: 0x0001, + // Block 0x9c, offset 0x2700 + 0x2700: 0x0001, 0x2701: 0x0001, 0x2702: 0x0001, 0x2703: 0x0001, 0x2704: 0x0001, 0x2705: 0x0001, + 0x2706: 0x0001, 0x2707: 0x0001, 0x2708: 0x0001, 0x2709: 0x0001, 0x270a: 0x0001, 0x270b: 0x0001, + 0x270c: 0x0001, 0x270d: 0x0001, 0x270e: 0x0001, 0x270f: 0x0001, 0x2710: 0x0001, 0x2711: 0x0001, + 0x2712: 0x0001, 0x2713: 0x0001, 0x2714: 0x0001, 0x2715: 0x0001, 0x2716: 0x0001, 0x2717: 0x0001, + 0x2718: 0x0001, 0x2719: 0x0001, 0x271a: 0x0001, 0x271b: 0x0001, 0x271c: 0x0001, 0x271d: 0x0001, + 0x271e: 0x0001, 0x271f: 0x000a, 0x2720: 0x0001, 0x2721: 0x0001, 0x2722: 0x0001, 0x2723: 0x0001, + 0x2724: 0x0001, 0x2725: 0x0001, 0x2726: 0x0001, 0x2727: 0x0001, 0x2728: 0x0001, 0x2729: 0x0001, + 0x272a: 0x0001, 0x272b: 0x0001, 0x272c: 0x0001, 0x272d: 0x0001, 0x272e: 0x0001, 0x272f: 0x0001, + 0x2730: 0x0001, 0x2731: 0x0001, 0x2732: 0x0001, 0x2733: 0x0001, 0x2734: 0x0001, 0x2735: 0x0001, + 0x2736: 0x0001, 0x2737: 0x0001, 0x2738: 0x0001, 0x2739: 0x0001, 0x273a: 0x0001, 0x273b: 0x0001, + 0x273c: 0x0001, 0x273d: 0x0001, 0x273e: 0x0001, 0x273f: 0x0001, + // Block 0x9d, offset 0x2740 + 0x2740: 0x0001, 0x2741: 0x000c, 0x2742: 0x000c, 0x2743: 0x000c, 0x2744: 0x0001, 0x2745: 0x000c, + 0x2746: 0x000c, 0x2747: 0x0001, 0x2748: 0x0001, 0x2749: 0x0001, 0x274a: 0x0001, 0x274b: 0x0001, + 0x274c: 0x000c, 0x274d: 0x000c, 0x274e: 0x000c, 0x274f: 0x000c, 0x2750: 0x0001, 0x2751: 0x0001, + 0x2752: 0x0001, 0x2753: 0x0001, 0x2754: 0x0001, 0x2755: 0x0001, 0x2756: 0x0001, 0x2757: 0x0001, + 0x2758: 0x0001, 0x2759: 0x0001, 0x275a: 0x0001, 0x275b: 0x0001, 0x275c: 0x0001, 0x275d: 0x0001, + 0x275e: 0x0001, 0x275f: 0x0001, 0x2760: 0x0001, 0x2761: 0x0001, 0x2762: 0x0001, 0x2763: 0x0001, + 0x2764: 0x0001, 0x2765: 0x0001, 0x2766: 0x0001, 0x2767: 0x0001, 0x2768: 0x0001, 0x2769: 0x0001, + 0x276a: 0x0001, 0x276b: 0x0001, 0x276c: 0x0001, 0x276d: 0x0001, 0x276e: 0x0001, 0x276f: 0x0001, + 0x2770: 0x0001, 0x2771: 0x0001, 0x2772: 0x0001, 0x2773: 0x0001, 0x2774: 0x0001, 0x2775: 0x0001, + 0x2776: 0x0001, 0x2777: 0x0001, 0x2778: 0x000c, 0x2779: 0x000c, 0x277a: 0x000c, 0x277b: 0x0001, + 0x277c: 0x0001, 0x277d: 0x0001, 0x277e: 0x0001, 0x277f: 0x000c, + // Block 0x9e, offset 0x2780 + 0x2780: 0x0001, 0x2781: 0x0001, 0x2782: 0x0001, 0x2783: 0x0001, 0x2784: 0x0001, 0x2785: 0x0001, + 0x2786: 0x0001, 0x2787: 0x0001, 0x2788: 0x0001, 0x2789: 0x0001, 0x278a: 0x0001, 0x278b: 0x0001, + 0x278c: 0x0001, 0x278d: 0x0001, 0x278e: 0x0001, 0x278f: 0x0001, 0x2790: 0x0001, 0x2791: 0x0001, + 0x2792: 0x0001, 0x2793: 0x0001, 0x2794: 0x0001, 0x2795: 0x0001, 0x2796: 0x0001, 0x2797: 0x0001, + 0x2798: 0x0001, 0x2799: 0x0001, 0x279a: 0x0001, 0x279b: 0x0001, 0x279c: 0x0001, 0x279d: 0x0001, + 0x279e: 0x0001, 0x279f: 0x0001, 0x27a0: 0x0001, 0x27a1: 0x0001, 0x27a2: 0x0001, 0x27a3: 0x0001, + 0x27a4: 0x0001, 0x27a5: 0x000c, 0x27a6: 0x000c, 0x27a7: 0x0001, 0x27a8: 0x0001, 0x27a9: 0x0001, + 0x27aa: 0x0001, 0x27ab: 0x0001, 0x27ac: 0x0001, 0x27ad: 0x0001, 0x27ae: 0x0001, 0x27af: 0x0001, + 0x27b0: 0x0001, 0x27b1: 0x0001, 0x27b2: 0x0001, 0x27b3: 0x0001, 0x27b4: 0x0001, 0x27b5: 0x0001, + 0x27b6: 0x0001, 0x27b7: 0x0001, 0x27b8: 0x0001, 0x27b9: 0x0001, 0x27ba: 0x0001, 0x27bb: 0x0001, + 0x27bc: 0x0001, 0x27bd: 0x0001, 0x27be: 0x0001, 0x27bf: 0x0001, + // Block 0x9f, offset 0x27c0 + 0x27c0: 0x0001, 0x27c1: 0x0001, 0x27c2: 0x0001, 0x27c3: 0x0001, 0x27c4: 0x0001, 0x27c5: 0x0001, + 0x27c6: 0x0001, 0x27c7: 0x0001, 0x27c8: 0x0001, 0x27c9: 0x0001, 0x27ca: 0x0001, 0x27cb: 0x0001, + 0x27cc: 0x0001, 0x27cd: 0x0001, 0x27ce: 0x0001, 0x27cf: 0x0001, 0x27d0: 0x0001, 0x27d1: 0x0001, + 0x27d2: 0x0001, 0x27d3: 0x0001, 0x27d4: 0x0001, 0x27d5: 0x0001, 0x27d6: 0x0001, 0x27d7: 0x0001, + 0x27d8: 0x0001, 0x27d9: 0x0001, 0x27da: 0x0001, 0x27db: 0x0001, 0x27dc: 0x0001, 0x27dd: 0x0001, + 0x27de: 0x0001, 0x27df: 0x0001, 0x27e0: 0x0001, 0x27e1: 0x0001, 0x27e2: 0x0001, 0x27e3: 0x0001, + 0x27e4: 0x0001, 0x27e5: 0x0001, 0x27e6: 0x0001, 0x27e7: 0x0001, 0x27e8: 0x0001, 0x27e9: 0x0001, + 0x27ea: 0x0001, 0x27eb: 0x0001, 0x27ec: 0x0001, 0x27ed: 0x0001, 0x27ee: 0x0001, 0x27ef: 0x0001, + 0x27f0: 0x0001, 0x27f1: 0x0001, 0x27f2: 0x0001, 0x27f3: 0x0001, 0x27f4: 0x0001, 0x27f5: 0x0001, + 0x27f6: 0x0001, 0x27f7: 0x0001, 0x27f8: 0x0001, 0x27f9: 0x000a, 0x27fa: 0x000a, 0x27fb: 0x000a, + 0x27fc: 0x000a, 0x27fd: 0x000a, 0x27fe: 0x000a, 0x27ff: 0x000a, + // Block 0xa0, offset 0x2800 + 0x2800: 0x000d, 0x2801: 0x000d, 0x2802: 0x000d, 0x2803: 0x000d, 0x2804: 0x000d, 0x2805: 0x000d, + 0x2806: 0x000d, 0x2807: 0x000d, 0x2808: 0x000d, 0x2809: 0x000d, 0x280a: 0x000d, 0x280b: 0x000d, + 0x280c: 0x000d, 0x280d: 0x000d, 0x280e: 0x000d, 0x280f: 0x000d, 0x2810: 0x000d, 0x2811: 0x000d, + 0x2812: 0x000d, 0x2813: 0x000d, 0x2814: 0x000d, 0x2815: 0x000d, 0x2816: 0x000d, 0x2817: 0x000d, + 0x2818: 0x000d, 0x2819: 0x000d, 0x281a: 0x000d, 0x281b: 0x000d, 0x281c: 0x000d, 0x281d: 0x000d, + 0x281e: 0x000d, 0x281f: 0x000d, 0x2820: 0x000d, 0x2821: 0x000d, 0x2822: 0x000d, 0x2823: 0x000d, + 0x2824: 0x000c, 0x2825: 0x000c, 0x2826: 0x000c, 0x2827: 0x000c, 0x2828: 0x000d, 0x2829: 0x000d, + 0x282a: 0x000d, 0x282b: 0x000d, 0x282c: 0x000d, 0x282d: 0x000d, 0x282e: 0x000d, 0x282f: 0x000d, + 0x2830: 0x0005, 0x2831: 0x0005, 0x2832: 0x0005, 0x2833: 0x0005, 0x2834: 0x0005, 0x2835: 0x0005, + 0x2836: 0x0005, 0x2837: 0x0005, 0x2838: 0x0005, 0x2839: 0x0005, 0x283a: 0x000d, 0x283b: 0x000d, + 0x283c: 0x000d, 0x283d: 0x000d, 0x283e: 0x000d, 0x283f: 0x000d, + // Block 0xa1, offset 0x2840 + 0x2840: 0x0001, 0x2841: 0x0001, 0x2842: 0x0001, 0x2843: 0x0001, 0x2844: 0x0001, 0x2845: 0x0001, + 0x2846: 0x0001, 0x2847: 0x0001, 0x2848: 0x0001, 0x2849: 0x0001, 0x284a: 0x0001, 0x284b: 0x0001, + 0x284c: 0x0001, 0x284d: 0x0001, 0x284e: 0x0001, 0x284f: 0x0001, 0x2850: 0x0001, 0x2851: 0x0001, + 0x2852: 0x0001, 0x2853: 0x0001, 0x2854: 0x0001, 0x2855: 0x0001, 0x2856: 0x0001, 0x2857: 0x0001, + 0x2858: 0x0001, 0x2859: 0x0001, 0x285a: 0x0001, 0x285b: 0x0001, 0x285c: 0x0001, 0x285d: 0x0001, + 0x285e: 0x0001, 0x285f: 0x0001, 0x2860: 0x0005, 0x2861: 0x0005, 0x2862: 0x0005, 0x2863: 0x0005, + 0x2864: 0x0005, 0x2865: 0x0005, 0x2866: 0x0005, 0x2867: 0x0005, 0x2868: 0x0005, 0x2869: 0x0005, + 0x286a: 0x0005, 0x286b: 0x0005, 0x286c: 0x0005, 0x286d: 0x0005, 0x286e: 0x0005, 0x286f: 0x0005, + 0x2870: 0x0005, 0x2871: 0x0005, 0x2872: 0x0005, 0x2873: 0x0005, 0x2874: 0x0005, 0x2875: 0x0005, + 0x2876: 0x0005, 0x2877: 0x0005, 0x2878: 0x0005, 0x2879: 0x0005, 0x287a: 0x0005, 0x287b: 0x0005, + 0x287c: 0x0005, 0x287d: 0x0005, 0x287e: 0x0005, 0x287f: 0x0001, + // Block 0xa2, offset 0x2880 + 0x2880: 0x0001, 0x2881: 0x0001, 0x2882: 0x0001, 0x2883: 0x0001, 0x2884: 0x0001, 0x2885: 0x0001, + 0x2886: 0x0001, 0x2887: 0x0001, 0x2888: 0x0001, 0x2889: 0x0001, 0x288a: 0x0001, 0x288b: 0x0001, + 0x288c: 0x0001, 0x288d: 0x0001, 0x288e: 0x0001, 0x288f: 0x0001, 0x2890: 0x0001, 0x2891: 0x0001, + 0x2892: 0x0001, 0x2893: 0x0001, 0x2894: 0x0001, 0x2895: 0x0001, 0x2896: 0x0001, 0x2897: 0x0001, + 0x2898: 0x0001, 0x2899: 0x0001, 0x289a: 0x0001, 0x289b: 0x0001, 0x289c: 0x0001, 0x289d: 0x0001, + 0x289e: 0x0001, 0x289f: 0x0001, 0x28a0: 0x0001, 0x28a1: 0x0001, 0x28a2: 0x0001, 0x28a3: 0x0001, + 0x28a4: 0x0001, 0x28a5: 0x0001, 0x28a6: 0x0001, 0x28a7: 0x0001, 0x28a8: 0x0001, 0x28a9: 0x0001, + 0x28aa: 0x0001, 0x28ab: 0x0001, 0x28ac: 0x0001, 0x28ad: 0x0001, 0x28ae: 0x0001, 0x28af: 0x0001, + 0x28b0: 0x000d, 0x28b1: 0x000d, 0x28b2: 0x000d, 0x28b3: 0x000d, 0x28b4: 0x000d, 0x28b5: 0x000d, + 0x28b6: 0x000d, 0x28b7: 0x000d, 0x28b8: 0x000d, 0x28b9: 0x000d, 0x28ba: 0x000d, 0x28bb: 0x000d, + 0x28bc: 0x000d, 0x28bd: 0x000d, 0x28be: 0x000d, 0x28bf: 0x000d, + // Block 0xa3, offset 0x28c0 + 0x28c0: 0x000d, 0x28c1: 0x000d, 0x28c2: 0x000d, 0x28c3: 0x000d, 0x28c4: 0x000d, 0x28c5: 0x000d, + 0x28c6: 0x000c, 0x28c7: 0x000c, 0x28c8: 0x000c, 0x28c9: 0x000c, 0x28ca: 0x000c, 0x28cb: 0x000c, + 0x28cc: 0x000c, 0x28cd: 0x000c, 0x28ce: 0x000c, 0x28cf: 0x000c, 0x28d0: 0x000c, 0x28d1: 0x000d, + 0x28d2: 0x000d, 0x28d3: 0x000d, 0x28d4: 0x000d, 0x28d5: 0x000d, 0x28d6: 0x000d, 0x28d7: 0x000d, + 0x28d8: 0x000d, 0x28d9: 0x000d, 0x28da: 0x000d, 0x28db: 0x000d, 0x28dc: 0x000d, 0x28dd: 0x000d, + 0x28de: 0x000d, 0x28df: 0x000d, 0x28e0: 0x000d, 0x28e1: 0x000d, 0x28e2: 0x000d, 0x28e3: 0x000d, + 0x28e4: 0x000d, 0x28e5: 0x000d, 0x28e6: 0x000d, 0x28e7: 0x000d, 0x28e8: 0x000d, 0x28e9: 0x000d, + 0x28ea: 0x000d, 0x28eb: 0x000d, 0x28ec: 0x000d, 0x28ed: 0x000d, 0x28ee: 0x000d, 0x28ef: 0x000d, + 0x28f0: 0x0001, 0x28f1: 0x0001, 0x28f2: 0x0001, 0x28f3: 0x0001, 0x28f4: 0x0001, 0x28f5: 0x0001, + 0x28f6: 0x0001, 0x28f7: 0x0001, 0x28f8: 0x0001, 0x28f9: 0x0001, 0x28fa: 0x0001, 0x28fb: 0x0001, + 0x28fc: 0x0001, 0x28fd: 0x0001, 0x28fe: 0x0001, 0x28ff: 0x0001, + // Block 0xa4, offset 0x2900 + 0x2901: 0x000c, + 0x2938: 0x000c, 0x2939: 0x000c, 0x293a: 0x000c, 0x293b: 0x000c, + 0x293c: 0x000c, 0x293d: 0x000c, 0x293e: 0x000c, 0x293f: 0x000c, + // Block 0xa5, offset 0x2940 + 0x2940: 0x000c, 0x2941: 0x000c, 0x2942: 0x000c, 0x2943: 0x000c, 0x2944: 0x000c, 0x2945: 0x000c, + 0x2946: 0x000c, + 0x2952: 0x000a, 0x2953: 0x000a, 0x2954: 0x000a, 0x2955: 0x000a, 0x2956: 0x000a, 0x2957: 0x000a, + 0x2958: 0x000a, 0x2959: 0x000a, 0x295a: 0x000a, 0x295b: 0x000a, 0x295c: 0x000a, 0x295d: 0x000a, + 0x295e: 0x000a, 0x295f: 0x000a, 0x2960: 0x000a, 0x2961: 0x000a, 0x2962: 0x000a, 0x2963: 0x000a, + 0x2964: 0x000a, 0x2965: 0x000a, + 0x297f: 0x000c, + // Block 0xa6, offset 0x2980 + 0x2980: 0x000c, 0x2981: 0x000c, + 0x29b3: 0x000c, 0x29b4: 0x000c, 0x29b5: 0x000c, + 0x29b6: 0x000c, 0x29b9: 0x000c, 0x29ba: 0x000c, + // Block 0xa7, offset 0x29c0 + 0x29c0: 0x000c, 0x29c1: 0x000c, 0x29c2: 0x000c, + 0x29e7: 0x000c, 0x29e8: 0x000c, 0x29e9: 0x000c, + 0x29ea: 0x000c, 0x29eb: 0x000c, 0x29ed: 0x000c, 0x29ee: 0x000c, 0x29ef: 0x000c, + 0x29f0: 0x000c, 0x29f1: 0x000c, 0x29f2: 0x000c, 0x29f3: 0x000c, 0x29f4: 0x000c, + // Block 0xa8, offset 0x2a00 + 0x2a33: 0x000c, + // Block 0xa9, offset 0x2a40 + 0x2a40: 0x000c, 0x2a41: 0x000c, + 0x2a76: 0x000c, 0x2a77: 0x000c, 0x2a78: 0x000c, 0x2a79: 0x000c, 0x2a7a: 0x000c, 0x2a7b: 0x000c, + 0x2a7c: 0x000c, 0x2a7d: 0x000c, 0x2a7e: 0x000c, + // Block 0xaa, offset 0x2a80 + 0x2a89: 0x000c, 0x2a8a: 0x000c, 0x2a8b: 0x000c, + 0x2a8c: 0x000c, + // Block 0xab, offset 0x2ac0 + 0x2aef: 0x000c, + 0x2af0: 0x000c, 0x2af1: 0x000c, 0x2af4: 0x000c, + 0x2af6: 0x000c, 0x2af7: 0x000c, + 0x2afe: 0x000c, + // Block 0xac, offset 0x2b00 + 0x2b1f: 0x000c, 0x2b23: 0x000c, + 0x2b24: 0x000c, 0x2b25: 0x000c, 0x2b26: 0x000c, 0x2b27: 0x000c, 0x2b28: 0x000c, 0x2b29: 0x000c, + 0x2b2a: 0x000c, + // Block 0xad, offset 0x2b40 + 0x2b40: 0x000c, + 0x2b66: 0x000c, 0x2b67: 0x000c, 0x2b68: 0x000c, 0x2b69: 0x000c, + 0x2b6a: 0x000c, 0x2b6b: 0x000c, 0x2b6c: 0x000c, + 0x2b70: 0x000c, 0x2b71: 0x000c, 0x2b72: 0x000c, 0x2b73: 0x000c, 0x2b74: 0x000c, + // Block 0xae, offset 0x2b80 + 0x2bb8: 0x000c, 0x2bb9: 0x000c, 0x2bba: 0x000c, 0x2bbb: 0x000c, + 0x2bbc: 0x000c, 0x2bbd: 0x000c, 0x2bbe: 0x000c, 0x2bbf: 0x000c, + // Block 0xaf, offset 0x2bc0 + 0x2bc2: 0x000c, 0x2bc3: 0x000c, 0x2bc4: 0x000c, + 0x2bc6: 0x000c, + 0x2bde: 0x000c, + // Block 0xb0, offset 0x2c00 + 0x2c33: 0x000c, 0x2c34: 0x000c, 0x2c35: 0x000c, + 0x2c36: 0x000c, 0x2c37: 0x000c, 0x2c38: 0x000c, 0x2c3a: 0x000c, + 0x2c3f: 0x000c, + // Block 0xb1, offset 0x2c40 + 0x2c40: 0x000c, 0x2c42: 0x000c, 0x2c43: 0x000c, + // Block 0xb2, offset 0x2c80 + 0x2cb2: 0x000c, 0x2cb3: 0x000c, 0x2cb4: 0x000c, 0x2cb5: 0x000c, + 0x2cbc: 0x000c, 0x2cbd: 0x000c, 0x2cbf: 0x000c, + // Block 0xb3, offset 0x2cc0 + 0x2cc0: 0x000c, + 0x2cdc: 0x000c, 0x2cdd: 0x000c, + // Block 0xb4, offset 0x2d00 + 0x2d33: 0x000c, 0x2d34: 0x000c, 0x2d35: 0x000c, + 0x2d36: 0x000c, 0x2d37: 0x000c, 0x2d38: 0x000c, 0x2d39: 0x000c, 0x2d3a: 0x000c, + 0x2d3d: 0x000c, 0x2d3f: 0x000c, + // Block 0xb5, offset 0x2d40 + 0x2d40: 0x000c, + 0x2d60: 0x000a, 0x2d61: 0x000a, 0x2d62: 0x000a, 0x2d63: 0x000a, + 0x2d64: 0x000a, 0x2d65: 0x000a, 0x2d66: 0x000a, 0x2d67: 0x000a, 0x2d68: 0x000a, 0x2d69: 0x000a, + 0x2d6a: 0x000a, 0x2d6b: 0x000a, 0x2d6c: 0x000a, + // Block 0xb6, offset 0x2d80 + 0x2dab: 0x000c, 0x2dad: 0x000c, + 0x2db0: 0x000c, 0x2db1: 0x000c, 0x2db2: 0x000c, 0x2db3: 0x000c, 0x2db4: 0x000c, 0x2db5: 0x000c, + 0x2db7: 0x000c, + // Block 0xb7, offset 0x2dc0 + 0x2ddd: 0x000c, + 0x2dde: 0x000c, 0x2ddf: 0x000c, 0x2de2: 0x000c, 0x2de3: 0x000c, + 0x2de4: 0x000c, 0x2de5: 0x000c, 0x2de7: 0x000c, 0x2de8: 0x000c, 0x2de9: 0x000c, + 0x2dea: 0x000c, 0x2deb: 0x000c, + // Block 0xb8, offset 0x2e00 + 0x2e2f: 0x000c, + 0x2e30: 0x000c, 0x2e31: 0x000c, 0x2e32: 0x000c, 0x2e33: 0x000c, 0x2e34: 0x000c, 0x2e35: 0x000c, + 0x2e36: 0x000c, 0x2e37: 0x000c, 0x2e39: 0x000c, 0x2e3a: 0x000c, + // Block 0xb9, offset 0x2e40 + 0x2e54: 0x000c, 0x2e55: 0x000c, 0x2e56: 0x000c, 0x2e57: 0x000c, + 0x2e5a: 0x000c, 0x2e5b: 0x000c, + 0x2e60: 0x000c, + // Block 0xba, offset 0x2e80 + 0x2e81: 0x000c, 0x2e82: 0x000c, 0x2e83: 0x000c, 0x2e84: 0x000c, 0x2e85: 0x000c, + 0x2e86: 0x000c, 0x2e89: 0x000c, 0x2e8a: 0x000c, + 0x2eb3: 0x000c, 0x2eb4: 0x000c, 0x2eb5: 0x000c, + 0x2eb6: 0x000c, 0x2eb7: 0x000c, 0x2eb8: 0x000c, 0x2ebb: 0x000c, + 0x2ebc: 0x000c, 0x2ebd: 0x000c, 0x2ebe: 0x000c, + // Block 0xbb, offset 0x2ec0 + 0x2ec7: 0x000c, + 0x2ed1: 0x000c, + 0x2ed2: 0x000c, 0x2ed3: 0x000c, 0x2ed4: 0x000c, 0x2ed5: 0x000c, 0x2ed6: 0x000c, + 0x2ed9: 0x000c, 0x2eda: 0x000c, 0x2edb: 0x000c, + // Block 0xbc, offset 0x2f00 + 0x2f0a: 0x000c, 0x2f0b: 0x000c, + 0x2f0c: 0x000c, 0x2f0d: 0x000c, 0x2f0e: 0x000c, 0x2f0f: 0x000c, 0x2f10: 0x000c, 0x2f11: 0x000c, + 0x2f12: 0x000c, 0x2f13: 0x000c, 0x2f14: 0x000c, 0x2f15: 0x000c, 0x2f16: 0x000c, + 0x2f18: 0x000c, 0x2f19: 0x000c, + // Block 0xbd, offset 0x2f40 + 0x2f70: 0x000c, 0x2f71: 0x000c, 0x2f72: 0x000c, 0x2f73: 0x000c, 0x2f74: 0x000c, 0x2f75: 0x000c, + 0x2f76: 0x000c, 0x2f78: 0x000c, 0x2f79: 0x000c, 0x2f7a: 0x000c, 0x2f7b: 0x000c, + 0x2f7c: 0x000c, 0x2f7d: 0x000c, + // Block 0xbe, offset 0x2f80 + 0x2f92: 0x000c, 0x2f93: 0x000c, 0x2f94: 0x000c, 0x2f95: 0x000c, 0x2f96: 0x000c, 0x2f97: 0x000c, + 0x2f98: 0x000c, 0x2f99: 0x000c, 0x2f9a: 0x000c, 0x2f9b: 0x000c, 0x2f9c: 0x000c, 0x2f9d: 0x000c, + 0x2f9e: 0x000c, 0x2f9f: 0x000c, 0x2fa0: 0x000c, 0x2fa1: 0x000c, 0x2fa2: 0x000c, 0x2fa3: 0x000c, + 0x2fa4: 0x000c, 0x2fa5: 0x000c, 0x2fa6: 0x000c, 0x2fa7: 0x000c, + 0x2faa: 0x000c, 0x2fab: 0x000c, 0x2fac: 0x000c, 0x2fad: 0x000c, 0x2fae: 0x000c, 0x2faf: 0x000c, + 0x2fb0: 0x000c, 0x2fb2: 0x000c, 0x2fb3: 0x000c, 0x2fb5: 0x000c, + 0x2fb6: 0x000c, + // Block 0xbf, offset 0x2fc0 + 0x2ff1: 0x000c, 0x2ff2: 0x000c, 0x2ff3: 0x000c, 0x2ff4: 0x000c, 0x2ff5: 0x000c, + 0x2ff6: 0x000c, 0x2ffa: 0x000c, + 0x2ffc: 0x000c, 0x2ffd: 0x000c, 0x2fff: 0x000c, + // Block 0xc0, offset 0x3000 + 0x3000: 0x000c, 0x3001: 0x000c, 0x3002: 0x000c, 0x3003: 0x000c, 0x3004: 0x000c, 0x3005: 0x000c, + 0x3007: 0x000c, + // Block 0xc1, offset 0x3040 + 0x3050: 0x000c, 0x3051: 0x000c, + 0x3055: 0x000c, 0x3057: 0x000c, + // Block 0xc2, offset 0x3080 + 0x30b3: 0x000c, 0x30b4: 0x000c, + // Block 0xc3, offset 0x30c0 + 0x30d5: 0x000a, 0x30d6: 0x000a, 0x30d7: 0x000a, + 0x30d8: 0x000a, 0x30d9: 0x000a, 0x30da: 0x000a, 0x30db: 0x000a, 0x30dc: 0x000a, 0x30dd: 0x0004, + 0x30de: 0x0004, 0x30df: 0x0004, 0x30e0: 0x0004, 0x30e1: 0x000a, 0x30e2: 0x000a, 0x30e3: 0x000a, + 0x30e4: 0x000a, 0x30e5: 0x000a, 0x30e6: 0x000a, 0x30e7: 0x000a, 0x30e8: 0x000a, 0x30e9: 0x000a, + 0x30ea: 0x000a, 0x30eb: 0x000a, 0x30ec: 0x000a, 0x30ed: 0x000a, 0x30ee: 0x000a, 0x30ef: 0x000a, + 0x30f0: 0x000a, 0x30f1: 0x000a, + // Block 0xc4, offset 0x3100 + 0x3130: 0x000c, 0x3131: 0x000c, 0x3132: 0x000c, 0x3133: 0x000c, 0x3134: 0x000c, + // Block 0xc5, offset 0x3140 + 0x3170: 0x000c, 0x3171: 0x000c, 0x3172: 0x000c, 0x3173: 0x000c, 0x3174: 0x000c, 0x3175: 0x000c, + 0x3176: 0x000c, + // Block 0xc6, offset 0x3180 + 0x318f: 0x000c, + // Block 0xc7, offset 0x31c0 + 0x31cf: 0x000c, 0x31d0: 0x000c, 0x31d1: 0x000c, + 0x31d2: 0x000c, + // Block 0xc8, offset 0x3200 + 0x3222: 0x000a, + // Block 0xc9, offset 0x3240 + 0x325d: 0x000c, + 0x325e: 0x000c, 0x3260: 0x000b, 0x3261: 0x000b, 0x3262: 0x000b, 0x3263: 0x000b, + // Block 0xca, offset 0x3280 + 0x32a7: 0x000c, 0x32a8: 0x000c, 0x32a9: 0x000c, + 0x32b3: 0x000b, 0x32b4: 0x000b, 0x32b5: 0x000b, + 0x32b6: 0x000b, 0x32b7: 0x000b, 0x32b8: 0x000b, 0x32b9: 0x000b, 0x32ba: 0x000b, 0x32bb: 0x000c, + 0x32bc: 0x000c, 0x32bd: 0x000c, 0x32be: 0x000c, 0x32bf: 0x000c, + // Block 0xcb, offset 0x32c0 + 0x32c0: 0x000c, 0x32c1: 0x000c, 0x32c2: 0x000c, 0x32c5: 0x000c, + 0x32c6: 0x000c, 0x32c7: 0x000c, 0x32c8: 0x000c, 0x32c9: 0x000c, 0x32ca: 0x000c, 0x32cb: 0x000c, + 0x32ea: 0x000c, 0x32eb: 0x000c, 0x32ec: 0x000c, 0x32ed: 0x000c, + // Block 0xcc, offset 0x3300 + 0x3300: 0x000a, 0x3301: 0x000a, 0x3302: 0x000c, 0x3303: 0x000c, 0x3304: 0x000c, 0x3305: 0x000a, + // Block 0xcd, offset 0x3340 + 0x3340: 0x000a, 0x3341: 0x000a, 0x3342: 0x000a, 0x3343: 0x000a, 0x3344: 0x000a, 0x3345: 0x000a, + 0x3346: 0x000a, 0x3347: 0x000a, 0x3348: 0x000a, 0x3349: 0x000a, 0x334a: 0x000a, 0x334b: 0x000a, + 0x334c: 0x000a, 0x334d: 0x000a, 0x334e: 0x000a, 0x334f: 0x000a, 0x3350: 0x000a, 0x3351: 0x000a, + 0x3352: 0x000a, 0x3353: 0x000a, 0x3354: 0x000a, 0x3355: 0x000a, 0x3356: 0x000a, + // Block 0xce, offset 0x3380 + 0x339b: 0x000a, + // Block 0xcf, offset 0x33c0 + 0x33d5: 0x000a, + // Block 0xd0, offset 0x3400 + 0x340f: 0x000a, + // Block 0xd1, offset 0x3440 + 0x3449: 0x000a, + // Block 0xd2, offset 0x3480 + 0x3483: 0x000a, + 0x348e: 0x0002, 0x348f: 0x0002, 0x3490: 0x0002, 0x3491: 0x0002, + 0x3492: 0x0002, 0x3493: 0x0002, 0x3494: 0x0002, 0x3495: 0x0002, 0x3496: 0x0002, 0x3497: 0x0002, + 0x3498: 0x0002, 0x3499: 0x0002, 0x349a: 0x0002, 0x349b: 0x0002, 0x349c: 0x0002, 0x349d: 0x0002, + 0x349e: 0x0002, 0x349f: 0x0002, 0x34a0: 0x0002, 0x34a1: 0x0002, 0x34a2: 0x0002, 0x34a3: 0x0002, + 0x34a4: 0x0002, 0x34a5: 0x0002, 0x34a6: 0x0002, 0x34a7: 0x0002, 0x34a8: 0x0002, 0x34a9: 0x0002, + 0x34aa: 0x0002, 0x34ab: 0x0002, 0x34ac: 0x0002, 0x34ad: 0x0002, 0x34ae: 0x0002, 0x34af: 0x0002, + 0x34b0: 0x0002, 0x34b1: 0x0002, 0x34b2: 0x0002, 0x34b3: 0x0002, 0x34b4: 0x0002, 0x34b5: 0x0002, + 0x34b6: 0x0002, 0x34b7: 0x0002, 0x34b8: 0x0002, 0x34b9: 0x0002, 0x34ba: 0x0002, 0x34bb: 0x0002, + 0x34bc: 0x0002, 0x34bd: 0x0002, 0x34be: 0x0002, 0x34bf: 0x0002, + // Block 0xd3, offset 0x34c0 + 0x34c0: 0x000c, 0x34c1: 0x000c, 0x34c2: 0x000c, 0x34c3: 0x000c, 0x34c4: 0x000c, 0x34c5: 0x000c, + 0x34c6: 0x000c, 0x34c7: 0x000c, 0x34c8: 0x000c, 0x34c9: 0x000c, 0x34ca: 0x000c, 0x34cb: 0x000c, + 0x34cc: 0x000c, 0x34cd: 0x000c, 0x34ce: 0x000c, 0x34cf: 0x000c, 0x34d0: 0x000c, 0x34d1: 0x000c, + 0x34d2: 0x000c, 0x34d3: 0x000c, 0x34d4: 0x000c, 0x34d5: 0x000c, 0x34d6: 0x000c, 0x34d7: 0x000c, + 0x34d8: 0x000c, 0x34d9: 0x000c, 0x34da: 0x000c, 0x34db: 0x000c, 0x34dc: 0x000c, 0x34dd: 0x000c, + 0x34de: 0x000c, 0x34df: 0x000c, 0x34e0: 0x000c, 0x34e1: 0x000c, 0x34e2: 0x000c, 0x34e3: 0x000c, + 0x34e4: 0x000c, 0x34e5: 0x000c, 0x34e6: 0x000c, 0x34e7: 0x000c, 0x34e8: 0x000c, 0x34e9: 0x000c, + 0x34ea: 0x000c, 0x34eb: 0x000c, 0x34ec: 0x000c, 0x34ed: 0x000c, 0x34ee: 0x000c, 0x34ef: 0x000c, + 0x34f0: 0x000c, 0x34f1: 0x000c, 0x34f2: 0x000c, 0x34f3: 0x000c, 0x34f4: 0x000c, 0x34f5: 0x000c, + 0x34f6: 0x000c, 0x34fb: 0x000c, + 0x34fc: 0x000c, 0x34fd: 0x000c, 0x34fe: 0x000c, 0x34ff: 0x000c, + // Block 0xd4, offset 0x3500 + 0x3500: 0x000c, 0x3501: 0x000c, 0x3502: 0x000c, 0x3503: 0x000c, 0x3504: 0x000c, 0x3505: 0x000c, + 0x3506: 0x000c, 0x3507: 0x000c, 0x3508: 0x000c, 0x3509: 0x000c, 0x350a: 0x000c, 0x350b: 0x000c, + 0x350c: 0x000c, 0x350d: 0x000c, 0x350e: 0x000c, 0x350f: 0x000c, 0x3510: 0x000c, 0x3511: 0x000c, + 0x3512: 0x000c, 0x3513: 0x000c, 0x3514: 0x000c, 0x3515: 0x000c, 0x3516: 0x000c, 0x3517: 0x000c, + 0x3518: 0x000c, 0x3519: 0x000c, 0x351a: 0x000c, 0x351b: 0x000c, 0x351c: 0x000c, 0x351d: 0x000c, + 0x351e: 0x000c, 0x351f: 0x000c, 0x3520: 0x000c, 0x3521: 0x000c, 0x3522: 0x000c, 0x3523: 0x000c, + 0x3524: 0x000c, 0x3525: 0x000c, 0x3526: 0x000c, 0x3527: 0x000c, 0x3528: 0x000c, 0x3529: 0x000c, + 0x352a: 0x000c, 0x352b: 0x000c, 0x352c: 0x000c, + 0x3535: 0x000c, + // Block 0xd5, offset 0x3540 + 0x3544: 0x000c, + 0x355b: 0x000c, 0x355c: 0x000c, 0x355d: 0x000c, + 0x355e: 0x000c, 0x355f: 0x000c, 0x3561: 0x000c, 0x3562: 0x000c, 0x3563: 0x000c, + 0x3564: 0x000c, 0x3565: 0x000c, 0x3566: 0x000c, 0x3567: 0x000c, 0x3568: 0x000c, 0x3569: 0x000c, + 0x356a: 0x000c, 0x356b: 0x000c, 0x356c: 0x000c, 0x356d: 0x000c, 0x356e: 0x000c, 0x356f: 0x000c, + // Block 0xd6, offset 0x3580 + 0x3580: 0x000c, 0x3581: 0x000c, 0x3582: 0x000c, 0x3583: 0x000c, 0x3584: 0x000c, 0x3585: 0x000c, + 0x3586: 0x000c, 0x3588: 0x000c, 0x3589: 0x000c, 0x358a: 0x000c, 0x358b: 0x000c, + 0x358c: 0x000c, 0x358d: 0x000c, 0x358e: 0x000c, 0x358f: 0x000c, 0x3590: 0x000c, 0x3591: 0x000c, + 0x3592: 0x000c, 0x3593: 0x000c, 0x3594: 0x000c, 0x3595: 0x000c, 0x3596: 0x000c, 0x3597: 0x000c, + 0x3598: 0x000c, 0x359b: 0x000c, 0x359c: 0x000c, 0x359d: 0x000c, + 0x359e: 0x000c, 0x359f: 0x000c, 0x35a0: 0x000c, 0x35a1: 0x000c, 0x35a3: 0x000c, + 0x35a4: 0x000c, 0x35a6: 0x000c, 0x35a7: 0x000c, 0x35a8: 0x000c, 0x35a9: 0x000c, + 0x35aa: 0x000c, + // Block 0xd7, offset 0x35c0 + 0x35ec: 0x000c, 0x35ed: 0x000c, 0x35ee: 0x000c, 0x35ef: 0x000c, + 0x35ff: 0x0004, + // Block 0xd8, offset 0x3600 + 0x3600: 0x0001, 0x3601: 0x0001, 0x3602: 0x0001, 0x3603: 0x0001, 0x3604: 0x0001, 0x3605: 0x0001, + 0x3606: 0x0001, 0x3607: 0x0001, 0x3608: 0x0001, 0x3609: 0x0001, 0x360a: 0x0001, 0x360b: 0x0001, + 0x360c: 0x0001, 0x360d: 0x0001, 0x360e: 0x0001, 0x360f: 0x0001, 0x3610: 0x000c, 0x3611: 0x000c, + 0x3612: 0x000c, 0x3613: 0x000c, 0x3614: 0x000c, 0x3615: 0x000c, 0x3616: 0x000c, 0x3617: 0x0001, + 0x3618: 0x0001, 0x3619: 0x0001, 0x361a: 0x0001, 0x361b: 0x0001, 0x361c: 0x0001, 0x361d: 0x0001, + 0x361e: 0x0001, 0x361f: 0x0001, 0x3620: 0x0001, 0x3621: 0x0001, 0x3622: 0x0001, 0x3623: 0x0001, + 0x3624: 0x0001, 0x3625: 0x0001, 0x3626: 0x0001, 0x3627: 0x0001, 0x3628: 0x0001, 0x3629: 0x0001, + 0x362a: 0x0001, 0x362b: 0x0001, 0x362c: 0x0001, 0x362d: 0x0001, 0x362e: 0x0001, 0x362f: 0x0001, + 0x3630: 0x0001, 0x3631: 0x0001, 0x3632: 0x0001, 0x3633: 0x0001, 0x3634: 0x0001, 0x3635: 0x0001, + 0x3636: 0x0001, 0x3637: 0x0001, 0x3638: 0x0001, 0x3639: 0x0001, 0x363a: 0x0001, 0x363b: 0x0001, + 0x363c: 0x0001, 0x363d: 0x0001, 0x363e: 0x0001, 0x363f: 0x0001, + // Block 0xd9, offset 0x3640 + 0x3640: 0x0001, 0x3641: 0x0001, 0x3642: 0x0001, 0x3643: 0x0001, 0x3644: 0x000c, 0x3645: 0x000c, + 0x3646: 0x000c, 0x3647: 0x000c, 0x3648: 0x000c, 0x3649: 0x000c, 0x364a: 0x000c, 0x364b: 0x0001, + 0x364c: 0x0001, 0x364d: 0x0001, 0x364e: 0x0001, 0x364f: 0x0001, 0x3650: 0x0001, 0x3651: 0x0001, + 0x3652: 0x0001, 0x3653: 0x0001, 0x3654: 0x0001, 0x3655: 0x0001, 0x3656: 0x0001, 0x3657: 0x0001, + 0x3658: 0x0001, 0x3659: 0x0001, 0x365a: 0x0001, 0x365b: 0x0001, 0x365c: 0x0001, 0x365d: 0x0001, + 0x365e: 0x0001, 0x365f: 0x0001, 0x3660: 0x0001, 0x3661: 0x0001, 0x3662: 0x0001, 0x3663: 0x0001, + 0x3664: 0x0001, 0x3665: 0x0001, 0x3666: 0x0001, 0x3667: 0x0001, 0x3668: 0x0001, 0x3669: 0x0001, + 0x366a: 0x0001, 0x366b: 0x0001, 0x366c: 0x0001, 0x366d: 0x0001, 0x366e: 0x0001, 0x366f: 0x0001, + 0x3670: 0x0001, 0x3671: 0x0001, 0x3672: 0x0001, 0x3673: 0x0001, 0x3674: 0x0001, 0x3675: 0x0001, + 0x3676: 0x0001, 0x3677: 0x0001, 0x3678: 0x0001, 0x3679: 0x0001, 0x367a: 0x0001, 0x367b: 0x0001, + 0x367c: 0x0001, 0x367d: 0x0001, 0x367e: 0x0001, 0x367f: 0x0001, + // Block 0xda, offset 0x3680 + 0x3680: 0x000d, 0x3681: 0x000d, 0x3682: 0x000d, 0x3683: 0x000d, 0x3684: 0x000d, 0x3685: 0x000d, + 0x3686: 0x000d, 0x3687: 0x000d, 0x3688: 0x000d, 0x3689: 0x000d, 0x368a: 0x000d, 0x368b: 0x000d, + 0x368c: 0x000d, 0x368d: 0x000d, 0x368e: 0x000d, 0x368f: 0x000d, 0x3690: 0x0001, 0x3691: 0x0001, + 0x3692: 0x0001, 0x3693: 0x0001, 0x3694: 0x0001, 0x3695: 0x0001, 0x3696: 0x0001, 0x3697: 0x0001, + 0x3698: 0x0001, 0x3699: 0x0001, 0x369a: 0x0001, 0x369b: 0x0001, 0x369c: 0x0001, 0x369d: 0x0001, + 0x369e: 0x0001, 0x369f: 0x0001, 0x36a0: 0x0001, 0x36a1: 0x0001, 0x36a2: 0x0001, 0x36a3: 0x0001, + 0x36a4: 0x0001, 0x36a5: 0x0001, 0x36a6: 0x0001, 0x36a7: 0x0001, 0x36a8: 0x0001, 0x36a9: 0x0001, + 0x36aa: 0x0001, 0x36ab: 0x0001, 0x36ac: 0x0001, 0x36ad: 0x0001, 0x36ae: 0x0001, 0x36af: 0x0001, + 0x36b0: 0x0001, 0x36b1: 0x0001, 0x36b2: 0x0001, 0x36b3: 0x0001, 0x36b4: 0x0001, 0x36b5: 0x0001, + 0x36b6: 0x0001, 0x36b7: 0x0001, 0x36b8: 0x0001, 0x36b9: 0x0001, 0x36ba: 0x0001, 0x36bb: 0x0001, + 0x36bc: 0x0001, 0x36bd: 0x0001, 0x36be: 0x0001, 0x36bf: 0x0001, + // Block 0xdb, offset 0x36c0 + 0x36c0: 0x000d, 0x36c1: 0x000d, 0x36c2: 0x000d, 0x36c3: 0x000d, 0x36c4: 0x000d, 0x36c5: 0x000d, + 0x36c6: 0x000d, 0x36c7: 0x000d, 0x36c8: 0x000d, 0x36c9: 0x000d, 0x36ca: 0x000d, 0x36cb: 0x000d, + 0x36cc: 0x000d, 0x36cd: 0x000d, 0x36ce: 0x000d, 0x36cf: 0x000d, 0x36d0: 0x000d, 0x36d1: 0x000d, + 0x36d2: 0x000d, 0x36d3: 0x000d, 0x36d4: 0x000d, 0x36d5: 0x000d, 0x36d6: 0x000d, 0x36d7: 0x000d, + 0x36d8: 0x000d, 0x36d9: 0x000d, 0x36da: 0x000d, 0x36db: 0x000d, 0x36dc: 0x000d, 0x36dd: 0x000d, + 0x36de: 0x000d, 0x36df: 0x000d, 0x36e0: 0x000d, 0x36e1: 0x000d, 0x36e2: 0x000d, 0x36e3: 0x000d, + 0x36e4: 0x000d, 0x36e5: 0x000d, 0x36e6: 0x000d, 0x36e7: 0x000d, 0x36e8: 0x000d, 0x36e9: 0x000d, + 0x36ea: 0x000d, 0x36eb: 0x000d, 0x36ec: 0x000d, 0x36ed: 0x000d, 0x36ee: 0x000d, 0x36ef: 0x000d, + 0x36f0: 0x000a, 0x36f1: 0x000a, 0x36f2: 0x000d, 0x36f3: 0x000d, 0x36f4: 0x000d, 0x36f5: 0x000d, + 0x36f6: 0x000d, 0x36f7: 0x000d, 0x36f8: 0x000d, 0x36f9: 0x000d, 0x36fa: 0x000d, 0x36fb: 0x000d, + 0x36fc: 0x000d, 0x36fd: 0x000d, 0x36fe: 0x000d, 0x36ff: 0x000d, + // Block 0xdc, offset 0x3700 + 0x3700: 0x000a, 0x3701: 0x000a, 0x3702: 0x000a, 0x3703: 0x000a, 0x3704: 0x000a, 0x3705: 0x000a, + 0x3706: 0x000a, 0x3707: 0x000a, 0x3708: 0x000a, 0x3709: 0x000a, 0x370a: 0x000a, 0x370b: 0x000a, + 0x370c: 0x000a, 0x370d: 0x000a, 0x370e: 0x000a, 0x370f: 0x000a, 0x3710: 0x000a, 0x3711: 0x000a, + 0x3712: 0x000a, 0x3713: 0x000a, 0x3714: 0x000a, 0x3715: 0x000a, 0x3716: 0x000a, 0x3717: 0x000a, + 0x3718: 0x000a, 0x3719: 0x000a, 0x371a: 0x000a, 0x371b: 0x000a, 0x371c: 0x000a, 0x371d: 0x000a, + 0x371e: 0x000a, 0x371f: 0x000a, 0x3720: 0x000a, 0x3721: 0x000a, 0x3722: 0x000a, 0x3723: 0x000a, + 0x3724: 0x000a, 0x3725: 0x000a, 0x3726: 0x000a, 0x3727: 0x000a, 0x3728: 0x000a, 0x3729: 0x000a, + 0x372a: 0x000a, 0x372b: 0x000a, + 0x3730: 0x000a, 0x3731: 0x000a, 0x3732: 0x000a, 0x3733: 0x000a, 0x3734: 0x000a, 0x3735: 0x000a, + 0x3736: 0x000a, 0x3737: 0x000a, 0x3738: 0x000a, 0x3739: 0x000a, 0x373a: 0x000a, 0x373b: 0x000a, + 0x373c: 0x000a, 0x373d: 0x000a, 0x373e: 0x000a, 0x373f: 0x000a, + // Block 0xdd, offset 0x3740 + 0x3740: 0x000a, 0x3741: 0x000a, 0x3742: 0x000a, 0x3743: 0x000a, 0x3744: 0x000a, 0x3745: 0x000a, + 0x3746: 0x000a, 0x3747: 0x000a, 0x3748: 0x000a, 0x3749: 0x000a, 0x374a: 0x000a, 0x374b: 0x000a, + 0x374c: 0x000a, 0x374d: 0x000a, 0x374e: 0x000a, 0x374f: 0x000a, 0x3750: 0x000a, 0x3751: 0x000a, + 0x3752: 0x000a, 0x3753: 0x000a, + 0x3760: 0x000a, 0x3761: 0x000a, 0x3762: 0x000a, 0x3763: 0x000a, + 0x3764: 0x000a, 0x3765: 0x000a, 0x3766: 0x000a, 0x3767: 0x000a, 0x3768: 0x000a, 0x3769: 0x000a, + 0x376a: 0x000a, 0x376b: 0x000a, 0x376c: 0x000a, 0x376d: 0x000a, 0x376e: 0x000a, + 0x3771: 0x000a, 0x3772: 0x000a, 0x3773: 0x000a, 0x3774: 0x000a, 0x3775: 0x000a, + 0x3776: 0x000a, 0x3777: 0x000a, 0x3778: 0x000a, 0x3779: 0x000a, 0x377a: 0x000a, 0x377b: 0x000a, + 0x377c: 0x000a, 0x377d: 0x000a, 0x377e: 0x000a, 0x377f: 0x000a, + // Block 0xde, offset 0x3780 + 0x3781: 0x000a, 0x3782: 0x000a, 0x3783: 0x000a, 0x3784: 0x000a, 0x3785: 0x000a, + 0x3786: 0x000a, 0x3787: 0x000a, 0x3788: 0x000a, 0x3789: 0x000a, 0x378a: 0x000a, 0x378b: 0x000a, + 0x378c: 0x000a, 0x378d: 0x000a, 0x378e: 0x000a, 0x378f: 0x000a, 0x3791: 0x000a, + 0x3792: 0x000a, 0x3793: 0x000a, 0x3794: 0x000a, 0x3795: 0x000a, 0x3796: 0x000a, 0x3797: 0x000a, + 0x3798: 0x000a, 0x3799: 0x000a, 0x379a: 0x000a, 0x379b: 0x000a, 0x379c: 0x000a, 0x379d: 0x000a, + 0x379e: 0x000a, 0x379f: 0x000a, 0x37a0: 0x000a, 0x37a1: 0x000a, 0x37a2: 0x000a, 0x37a3: 0x000a, + 0x37a4: 0x000a, 0x37a5: 0x000a, 0x37a6: 0x000a, 0x37a7: 0x000a, 0x37a8: 0x000a, 0x37a9: 0x000a, + 0x37aa: 0x000a, 0x37ab: 0x000a, 0x37ac: 0x000a, 0x37ad: 0x000a, 0x37ae: 0x000a, 0x37af: 0x000a, + 0x37b0: 0x000a, 0x37b1: 0x000a, 0x37b2: 0x000a, 0x37b3: 0x000a, 0x37b4: 0x000a, 0x37b5: 0x000a, + // Block 0xdf, offset 0x37c0 + 0x37c0: 0x0002, 0x37c1: 0x0002, 0x37c2: 0x0002, 0x37c3: 0x0002, 0x37c4: 0x0002, 0x37c5: 0x0002, + 0x37c6: 0x0002, 0x37c7: 0x0002, 0x37c8: 0x0002, 0x37c9: 0x0002, 0x37ca: 0x0002, 0x37cb: 0x000a, + 0x37cc: 0x000a, + 0x37ef: 0x000a, + // Block 0xe0, offset 0x3800 + 0x382a: 0x000a, 0x382b: 0x000a, 0x382c: 0x000a, + // Block 0xe1, offset 0x3840 + 0x3860: 0x000a, 0x3861: 0x000a, 0x3862: 0x000a, 0x3863: 0x000a, + 0x3864: 0x000a, 0x3865: 0x000a, + // Block 0xe2, offset 0x3880 + 0x3880: 0x000a, 0x3881: 0x000a, 0x3882: 0x000a, 0x3883: 0x000a, 0x3884: 0x000a, 0x3885: 0x000a, + 0x3886: 0x000a, 0x3887: 0x000a, 0x3888: 0x000a, 0x3889: 0x000a, 0x388a: 0x000a, 0x388b: 0x000a, + 0x388c: 0x000a, 0x388d: 0x000a, 0x388e: 0x000a, 0x388f: 0x000a, 0x3890: 0x000a, 0x3891: 0x000a, + 0x3892: 0x000a, 0x3893: 0x000a, 0x3894: 0x000a, 0x3895: 0x000a, + 0x38a0: 0x000a, 0x38a1: 0x000a, 0x38a2: 0x000a, 0x38a3: 0x000a, + 0x38a4: 0x000a, 0x38a5: 0x000a, 0x38a6: 0x000a, 0x38a7: 0x000a, 0x38a8: 0x000a, 0x38a9: 0x000a, + 0x38aa: 0x000a, 0x38ab: 0x000a, 0x38ac: 0x000a, + 0x38b0: 0x000a, 0x38b1: 0x000a, 0x38b2: 0x000a, 0x38b3: 0x000a, 0x38b4: 0x000a, 0x38b5: 0x000a, + 0x38b6: 0x000a, 0x38b7: 0x000a, 0x38b8: 0x000a, 0x38b9: 0x000a, 0x38ba: 0x000a, + // Block 0xe3, offset 0x38c0 + 0x38c0: 0x000a, 0x38c1: 0x000a, 0x38c2: 0x000a, 0x38c3: 0x000a, 0x38c4: 0x000a, 0x38c5: 0x000a, + 0x38c6: 0x000a, 0x38c7: 0x000a, 0x38c8: 0x000a, 0x38c9: 0x000a, 0x38ca: 0x000a, 0x38cb: 0x000a, + 0x38cc: 0x000a, 0x38cd: 0x000a, 0x38ce: 0x000a, 0x38cf: 0x000a, 0x38d0: 0x000a, 0x38d1: 0x000a, + 0x38d2: 0x000a, 0x38d3: 0x000a, 0x38d4: 0x000a, 0x38d5: 0x000a, 0x38d6: 0x000a, 0x38d7: 0x000a, + 0x38d8: 0x000a, + 0x38e0: 0x000a, 0x38e1: 0x000a, 0x38e2: 0x000a, 0x38e3: 0x000a, + 0x38e4: 0x000a, 0x38e5: 0x000a, 0x38e6: 0x000a, 0x38e7: 0x000a, 0x38e8: 0x000a, 0x38e9: 0x000a, + 0x38ea: 0x000a, 0x38eb: 0x000a, + // Block 0xe4, offset 0x3900 + 0x3900: 0x000a, 0x3901: 0x000a, 0x3902: 0x000a, 0x3903: 0x000a, 0x3904: 0x000a, 0x3905: 0x000a, + 0x3906: 0x000a, 0x3907: 0x000a, 0x3908: 0x000a, 0x3909: 0x000a, 0x390a: 0x000a, 0x390b: 0x000a, + 0x3910: 0x000a, 0x3911: 0x000a, + 0x3912: 0x000a, 0x3913: 0x000a, 0x3914: 0x000a, 0x3915: 0x000a, 0x3916: 0x000a, 0x3917: 0x000a, + 0x3918: 0x000a, 0x3919: 0x000a, 0x391a: 0x000a, 0x391b: 0x000a, 0x391c: 0x000a, 0x391d: 0x000a, + 0x391e: 0x000a, 0x391f: 0x000a, 0x3920: 0x000a, 0x3921: 0x000a, 0x3922: 0x000a, 0x3923: 0x000a, + 0x3924: 0x000a, 0x3925: 0x000a, 0x3926: 0x000a, 0x3927: 0x000a, 0x3928: 0x000a, 0x3929: 0x000a, + 0x392a: 0x000a, 0x392b: 0x000a, 0x392c: 0x000a, 0x392d: 0x000a, 0x392e: 0x000a, 0x392f: 0x000a, + 0x3930: 0x000a, 0x3931: 0x000a, 0x3932: 0x000a, 0x3933: 0x000a, 0x3934: 0x000a, 0x3935: 0x000a, + 0x3936: 0x000a, 0x3937: 0x000a, 0x3938: 0x000a, 0x3939: 0x000a, 0x393a: 0x000a, 0x393b: 0x000a, + 0x393c: 0x000a, 0x393d: 0x000a, 0x393e: 0x000a, 0x393f: 0x000a, + // Block 0xe5, offset 0x3940 + 0x3940: 0x000a, 0x3941: 0x000a, 0x3942: 0x000a, 0x3943: 0x000a, 0x3944: 0x000a, 0x3945: 0x000a, + 0x3946: 0x000a, 0x3947: 0x000a, + 0x3950: 0x000a, 0x3951: 0x000a, + 0x3952: 0x000a, 0x3953: 0x000a, 0x3954: 0x000a, 0x3955: 0x000a, 0x3956: 0x000a, 0x3957: 0x000a, + 0x3958: 0x000a, 0x3959: 0x000a, + 0x3960: 0x000a, 0x3961: 0x000a, 0x3962: 0x000a, 0x3963: 0x000a, + 0x3964: 0x000a, 0x3965: 0x000a, 0x3966: 0x000a, 0x3967: 0x000a, 0x3968: 0x000a, 0x3969: 0x000a, + 0x396a: 0x000a, 0x396b: 0x000a, 0x396c: 0x000a, 0x396d: 0x000a, 0x396e: 0x000a, 0x396f: 0x000a, + 0x3970: 0x000a, 0x3971: 0x000a, 0x3972: 0x000a, 0x3973: 0x000a, 0x3974: 0x000a, 0x3975: 0x000a, + 0x3976: 0x000a, 0x3977: 0x000a, 0x3978: 0x000a, 0x3979: 0x000a, 0x397a: 0x000a, 0x397b: 0x000a, + 0x397c: 0x000a, 0x397d: 0x000a, 0x397e: 0x000a, 0x397f: 0x000a, + // Block 0xe6, offset 0x3980 + 0x3980: 0x000a, 0x3981: 0x000a, 0x3982: 0x000a, 0x3983: 0x000a, 0x3984: 0x000a, 0x3985: 0x000a, + 0x3986: 0x000a, 0x3987: 0x000a, + 0x3990: 0x000a, 0x3991: 0x000a, + 0x3992: 0x000a, 0x3993: 0x000a, 0x3994: 0x000a, 0x3995: 0x000a, 0x3996: 0x000a, 0x3997: 0x000a, + 0x3998: 0x000a, 0x3999: 0x000a, 0x399a: 0x000a, 0x399b: 0x000a, 0x399c: 0x000a, 0x399d: 0x000a, + 0x399e: 0x000a, 0x399f: 0x000a, 0x39a0: 0x000a, 0x39a1: 0x000a, 0x39a2: 0x000a, 0x39a3: 0x000a, + 0x39a4: 0x000a, 0x39a5: 0x000a, 0x39a6: 0x000a, 0x39a7: 0x000a, 0x39a8: 0x000a, 0x39a9: 0x000a, + 0x39aa: 0x000a, 0x39ab: 0x000a, 0x39ac: 0x000a, 0x39ad: 0x000a, + // Block 0xe7, offset 0x39c0 + 0x39c0: 0x000a, 0x39c1: 0x000a, 0x39c2: 0x000a, 0x39c3: 0x000a, 0x39c4: 0x000a, 0x39c5: 0x000a, + 0x39c6: 0x000a, 0x39c7: 0x000a, 0x39c8: 0x000a, 0x39c9: 0x000a, 0x39ca: 0x000a, 0x39cb: 0x000a, + 0x39cd: 0x000a, 0x39ce: 0x000a, 0x39cf: 0x000a, 0x39d0: 0x000a, 0x39d1: 0x000a, + 0x39d2: 0x000a, 0x39d3: 0x000a, 0x39d4: 0x000a, 0x39d5: 0x000a, 0x39d6: 0x000a, 0x39d7: 0x000a, + 0x39d8: 0x000a, 0x39d9: 0x000a, 0x39da: 0x000a, 0x39db: 0x000a, 0x39dc: 0x000a, 0x39dd: 0x000a, + 0x39de: 0x000a, 0x39df: 0x000a, 0x39e0: 0x000a, 0x39e1: 0x000a, 0x39e2: 0x000a, 0x39e3: 0x000a, + 0x39e4: 0x000a, 0x39e5: 0x000a, 0x39e6: 0x000a, 0x39e7: 0x000a, 0x39e8: 0x000a, 0x39e9: 0x000a, + 0x39ea: 0x000a, 0x39eb: 0x000a, 0x39ec: 0x000a, 0x39ed: 0x000a, 0x39ee: 0x000a, 0x39ef: 0x000a, + 0x39f0: 0x000a, 0x39f1: 0x000a, 0x39f2: 0x000a, 0x39f3: 0x000a, 0x39f4: 0x000a, 0x39f5: 0x000a, + 0x39f6: 0x000a, 0x39f7: 0x000a, 0x39f8: 0x000a, 0x39f9: 0x000a, 0x39fa: 0x000a, 0x39fb: 0x000a, + 0x39fc: 0x000a, 0x39fd: 0x000a, 0x39fe: 0x000a, 0x39ff: 0x000a, + // Block 0xe8, offset 0x3a00 + 0x3a00: 0x000a, 0x3a01: 0x000a, 0x3a02: 0x000a, 0x3a03: 0x000a, 0x3a04: 0x000a, 0x3a05: 0x000a, + 0x3a06: 0x000a, 0x3a07: 0x000a, 0x3a08: 0x000a, 0x3a09: 0x000a, 0x3a0a: 0x000a, 0x3a0b: 0x000a, + 0x3a0c: 0x000a, 0x3a0d: 0x000a, 0x3a0e: 0x000a, 0x3a0f: 0x000a, 0x3a10: 0x000a, 0x3a11: 0x000a, + 0x3a12: 0x000a, 0x3a13: 0x000a, 0x3a14: 0x000a, 0x3a15: 0x000a, 0x3a16: 0x000a, 0x3a17: 0x000a, + 0x3a18: 0x000a, 0x3a19: 0x000a, 0x3a1a: 0x000a, 0x3a1b: 0x000a, 0x3a1c: 0x000a, 0x3a1d: 0x000a, + 0x3a1e: 0x000a, 0x3a1f: 0x000a, 0x3a20: 0x000a, 0x3a21: 0x000a, 0x3a22: 0x000a, 0x3a23: 0x000a, + 0x3a24: 0x000a, 0x3a25: 0x000a, 0x3a26: 0x000a, 0x3a27: 0x000a, 0x3a28: 0x000a, 0x3a29: 0x000a, + 0x3a2a: 0x000a, 0x3a2b: 0x000a, 0x3a2c: 0x000a, 0x3a2d: 0x000a, 0x3a2e: 0x000a, 0x3a2f: 0x000a, + 0x3a30: 0x000a, 0x3a31: 0x000a, 0x3a33: 0x000a, 0x3a34: 0x000a, 0x3a35: 0x000a, + 0x3a36: 0x000a, 0x3a3a: 0x000a, 0x3a3b: 0x000a, + 0x3a3c: 0x000a, 0x3a3d: 0x000a, 0x3a3e: 0x000a, 0x3a3f: 0x000a, + // Block 0xe9, offset 0x3a40 + 0x3a40: 0x000a, 0x3a41: 0x000a, 0x3a42: 0x000a, 0x3a43: 0x000a, 0x3a44: 0x000a, 0x3a45: 0x000a, + 0x3a46: 0x000a, 0x3a47: 0x000a, 0x3a48: 0x000a, 0x3a49: 0x000a, 0x3a4a: 0x000a, 0x3a4b: 0x000a, + 0x3a4c: 0x000a, 0x3a4d: 0x000a, 0x3a4e: 0x000a, 0x3a4f: 0x000a, 0x3a50: 0x000a, 0x3a51: 0x000a, + 0x3a52: 0x000a, 0x3a53: 0x000a, 0x3a54: 0x000a, 0x3a55: 0x000a, 0x3a56: 0x000a, 0x3a57: 0x000a, + 0x3a58: 0x000a, 0x3a59: 0x000a, 0x3a5a: 0x000a, 0x3a5b: 0x000a, 0x3a5c: 0x000a, 0x3a5d: 0x000a, + 0x3a5e: 0x000a, 0x3a5f: 0x000a, 0x3a60: 0x000a, 0x3a61: 0x000a, 0x3a62: 0x000a, + 0x3a65: 0x000a, 0x3a66: 0x000a, 0x3a67: 0x000a, 0x3a68: 0x000a, 0x3a69: 0x000a, + 0x3a6a: 0x000a, 0x3a6e: 0x000a, 0x3a6f: 0x000a, + 0x3a70: 0x000a, 0x3a71: 0x000a, 0x3a72: 0x000a, 0x3a73: 0x000a, 0x3a74: 0x000a, 0x3a75: 0x000a, + 0x3a76: 0x000a, 0x3a77: 0x000a, 0x3a78: 0x000a, 0x3a79: 0x000a, 0x3a7a: 0x000a, 0x3a7b: 0x000a, + 0x3a7c: 0x000a, 0x3a7d: 0x000a, 0x3a7e: 0x000a, 0x3a7f: 0x000a, + // Block 0xea, offset 0x3a80 + 0x3a80: 0x000a, 0x3a81: 0x000a, 0x3a82: 0x000a, 0x3a83: 0x000a, 0x3a84: 0x000a, 0x3a85: 0x000a, + 0x3a86: 0x000a, 0x3a87: 0x000a, 0x3a88: 0x000a, 0x3a89: 0x000a, 0x3a8a: 0x000a, + 0x3a8d: 0x000a, 0x3a8e: 0x000a, 0x3a8f: 0x000a, 0x3a90: 0x000a, 0x3a91: 0x000a, + 0x3a92: 0x000a, 0x3a93: 0x000a, 0x3a94: 0x000a, 0x3a95: 0x000a, 0x3a96: 0x000a, 0x3a97: 0x000a, + 0x3a98: 0x000a, 0x3a99: 0x000a, 0x3a9a: 0x000a, 0x3a9b: 0x000a, 0x3a9c: 0x000a, 0x3a9d: 0x000a, + 0x3a9e: 0x000a, 0x3a9f: 0x000a, 0x3aa0: 0x000a, 0x3aa1: 0x000a, 0x3aa2: 0x000a, 0x3aa3: 0x000a, + 0x3aa4: 0x000a, 0x3aa5: 0x000a, 0x3aa6: 0x000a, 0x3aa7: 0x000a, 0x3aa8: 0x000a, 0x3aa9: 0x000a, + 0x3aaa: 0x000a, 0x3aab: 0x000a, 0x3aac: 0x000a, 0x3aad: 0x000a, 0x3aae: 0x000a, 0x3aaf: 0x000a, + 0x3ab0: 0x000a, 0x3ab1: 0x000a, 0x3ab2: 0x000a, 0x3ab3: 0x000a, 0x3ab4: 0x000a, 0x3ab5: 0x000a, + 0x3ab6: 0x000a, 0x3ab7: 0x000a, 0x3ab8: 0x000a, 0x3ab9: 0x000a, 0x3aba: 0x000a, 0x3abb: 0x000a, + 0x3abc: 0x000a, 0x3abd: 0x000a, 0x3abe: 0x000a, 0x3abf: 0x000a, + // Block 0xeb, offset 0x3ac0 + 0x3ac0: 0x000a, 0x3ac1: 0x000a, 0x3ac2: 0x000a, 0x3ac3: 0x000a, 0x3ac4: 0x000a, 0x3ac5: 0x000a, + 0x3ac6: 0x000a, 0x3ac7: 0x000a, 0x3ac8: 0x000a, 0x3ac9: 0x000a, 0x3aca: 0x000a, 0x3acb: 0x000a, + 0x3acc: 0x000a, 0x3acd: 0x000a, 0x3ace: 0x000a, 0x3acf: 0x000a, 0x3ad0: 0x000a, 0x3ad1: 0x000a, + 0x3ad2: 0x000a, 0x3ad3: 0x000a, + 0x3ae0: 0x000a, 0x3ae1: 0x000a, 0x3ae2: 0x000a, 0x3ae3: 0x000a, + 0x3ae4: 0x000a, 0x3ae5: 0x000a, 0x3ae6: 0x000a, 0x3ae7: 0x000a, 0x3ae8: 0x000a, 0x3ae9: 0x000a, + 0x3aea: 0x000a, 0x3aeb: 0x000a, 0x3aec: 0x000a, 0x3aed: 0x000a, + 0x3af0: 0x000a, 0x3af1: 0x000a, 0x3af2: 0x000a, 0x3af3: 0x000a, + 0x3af8: 0x000a, 0x3af9: 0x000a, 0x3afa: 0x000a, + // Block 0xec, offset 0x3b00 + 0x3b00: 0x000a, 0x3b01: 0x000a, 0x3b02: 0x000a, + 0x3b10: 0x000a, 0x3b11: 0x000a, + 0x3b12: 0x000a, 0x3b13: 0x000a, 0x3b14: 0x000a, 0x3b15: 0x000a, + // Block 0xed, offset 0x3b40 + 0x3b7e: 0x000b, 0x3b7f: 0x000b, + // Block 0xee, offset 0x3b80 + 0x3b80: 0x000b, 0x3b81: 0x000b, 0x3b82: 0x000b, 0x3b83: 0x000b, 0x3b84: 0x000b, 0x3b85: 0x000b, + 0x3b86: 0x000b, 0x3b87: 0x000b, 0x3b88: 0x000b, 0x3b89: 0x000b, 0x3b8a: 0x000b, 0x3b8b: 0x000b, + 0x3b8c: 0x000b, 0x3b8d: 0x000b, 0x3b8e: 0x000b, 0x3b8f: 0x000b, 0x3b90: 0x000b, 0x3b91: 0x000b, + 0x3b92: 0x000b, 0x3b93: 0x000b, 0x3b94: 0x000b, 0x3b95: 0x000b, 0x3b96: 0x000b, 0x3b97: 0x000b, + 0x3b98: 0x000b, 0x3b99: 0x000b, 0x3b9a: 0x000b, 0x3b9b: 0x000b, 0x3b9c: 0x000b, 0x3b9d: 0x000b, + 0x3b9e: 0x000b, 0x3b9f: 0x000b, 0x3ba0: 0x000b, 0x3ba1: 0x000b, 0x3ba2: 0x000b, 0x3ba3: 0x000b, + 0x3ba4: 0x000b, 0x3ba5: 0x000b, 0x3ba6: 0x000b, 0x3ba7: 0x000b, 0x3ba8: 0x000b, 0x3ba9: 0x000b, + 0x3baa: 0x000b, 0x3bab: 0x000b, 0x3bac: 0x000b, 0x3bad: 0x000b, 0x3bae: 0x000b, 0x3baf: 0x000b, + 0x3bb0: 0x000b, 0x3bb1: 0x000b, 0x3bb2: 0x000b, 0x3bb3: 0x000b, 0x3bb4: 0x000b, 0x3bb5: 0x000b, + 0x3bb6: 0x000b, 0x3bb7: 0x000b, 0x3bb8: 0x000b, 0x3bb9: 0x000b, 0x3bba: 0x000b, 0x3bbb: 0x000b, + 0x3bbc: 0x000b, 0x3bbd: 0x000b, 0x3bbe: 0x000b, 0x3bbf: 0x000b, + // Block 0xef, offset 0x3bc0 + 0x3bc0: 0x000c, 0x3bc1: 0x000c, 0x3bc2: 0x000c, 0x3bc3: 0x000c, 0x3bc4: 0x000c, 0x3bc5: 0x000c, + 0x3bc6: 0x000c, 0x3bc7: 0x000c, 0x3bc8: 0x000c, 0x3bc9: 0x000c, 0x3bca: 0x000c, 0x3bcb: 0x000c, + 0x3bcc: 0x000c, 0x3bcd: 0x000c, 0x3bce: 0x000c, 0x3bcf: 0x000c, 0x3bd0: 0x000c, 0x3bd1: 0x000c, + 0x3bd2: 0x000c, 0x3bd3: 0x000c, 0x3bd4: 0x000c, 0x3bd5: 0x000c, 0x3bd6: 0x000c, 0x3bd7: 0x000c, + 0x3bd8: 0x000c, 0x3bd9: 0x000c, 0x3bda: 0x000c, 0x3bdb: 0x000c, 0x3bdc: 0x000c, 0x3bdd: 0x000c, + 0x3bde: 0x000c, 0x3bdf: 0x000c, 0x3be0: 0x000c, 0x3be1: 0x000c, 0x3be2: 0x000c, 0x3be3: 0x000c, + 0x3be4: 0x000c, 0x3be5: 0x000c, 0x3be6: 0x000c, 0x3be7: 0x000c, 0x3be8: 0x000c, 0x3be9: 0x000c, + 0x3bea: 0x000c, 0x3beb: 0x000c, 0x3bec: 0x000c, 0x3bed: 0x000c, 0x3bee: 0x000c, 0x3bef: 0x000c, + 0x3bf0: 0x000b, 0x3bf1: 0x000b, 0x3bf2: 0x000b, 0x3bf3: 0x000b, 0x3bf4: 0x000b, 0x3bf5: 0x000b, + 0x3bf6: 0x000b, 0x3bf7: 0x000b, 0x3bf8: 0x000b, 0x3bf9: 0x000b, 0x3bfa: 0x000b, 0x3bfb: 0x000b, + 0x3bfc: 0x000b, 0x3bfd: 0x000b, 0x3bfe: 0x000b, 0x3bff: 0x000b, +} + +// bidiIndex: 24 blocks, 1536 entries, 1536 bytes +// Block 0 is the zero block. +var bidiIndex = [1536]uint8{ + // Block 0x0, offset 0x0 + // Block 0x1, offset 0x40 + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xc2: 0x01, 0xc3: 0x02, + 0xca: 0x03, 0xcb: 0x04, 0xcc: 0x05, 0xcd: 0x06, 0xce: 0x07, 0xcf: 0x08, + 0xd2: 0x09, 0xd6: 0x0a, 0xd7: 0x0b, + 0xd8: 0x0c, 0xd9: 0x0d, 0xda: 0x0e, 0xdb: 0x0f, 0xdc: 0x10, 0xdd: 0x11, 0xde: 0x12, 0xdf: 0x13, + 0xe0: 0x02, 0xe1: 0x03, 0xe2: 0x04, 0xe3: 0x05, 0xe4: 0x06, + 0xea: 0x07, 0xef: 0x08, + 0xf0: 0x11, 0xf1: 0x12, 0xf2: 0x12, 0xf3: 0x14, 0xf4: 0x15, + // Block 0x4, offset 0x100 + 0x120: 0x14, 0x121: 0x15, 0x122: 0x16, 0x123: 0x17, 0x124: 0x18, 0x125: 0x19, 0x126: 0x1a, 0x127: 0x1b, + 0x128: 0x1c, 0x129: 0x1d, 0x12a: 0x1c, 0x12b: 0x1e, 0x12c: 0x1f, 0x12d: 0x20, 0x12e: 0x21, 0x12f: 0x22, + 0x130: 0x23, 0x131: 0x24, 0x132: 0x1a, 0x133: 0x25, 0x134: 0x26, 0x135: 0x27, 0x137: 0x28, + 0x138: 0x29, 0x139: 0x2a, 0x13a: 0x2b, 0x13b: 0x2c, 0x13c: 0x2d, 0x13d: 0x2e, 0x13e: 0x2f, 0x13f: 0x30, + // Block 0x5, offset 0x140 + 0x140: 0x31, 0x141: 0x32, 0x142: 0x33, + 0x14d: 0x34, 0x14e: 0x35, + 0x150: 0x36, + 0x15a: 0x37, 0x15c: 0x38, 0x15d: 0x39, 0x15e: 0x3a, 0x15f: 0x3b, + 0x160: 0x3c, 0x162: 0x3d, 0x164: 0x3e, 0x165: 0x3f, 0x167: 0x40, + 0x168: 0x41, 0x169: 0x42, 0x16a: 0x43, 0x16c: 0x44, 0x16d: 0x45, 0x16e: 0x46, 0x16f: 0x47, + 0x170: 0x48, 0x173: 0x49, 0x177: 0x4a, + 0x17e: 0x4b, 0x17f: 0x4c, + // Block 0x6, offset 0x180 + 0x180: 0x4d, 0x181: 0x4e, 0x182: 0x4f, 0x183: 0x50, 0x184: 0x51, 0x185: 0x52, 0x186: 0x53, 0x187: 0x54, + 0x188: 0x55, 0x189: 0x54, 0x18a: 0x54, 0x18b: 0x54, 0x18c: 0x56, 0x18d: 0x57, 0x18e: 0x58, 0x18f: 0x54, + 0x190: 0x59, 0x191: 0x5a, 0x192: 0x5b, 0x193: 0x5c, 0x194: 0x54, 0x195: 0x54, 0x196: 0x54, 0x197: 0x54, + 0x198: 0x54, 0x199: 0x54, 0x19a: 0x5d, 0x19b: 0x54, 0x19c: 0x54, 0x19d: 0x5e, 0x19e: 0x54, 0x19f: 0x5f, + 0x1a4: 0x54, 0x1a5: 0x54, 0x1a6: 0x60, 0x1a7: 0x61, + 0x1a8: 0x54, 0x1a9: 0x54, 0x1aa: 0x54, 0x1ab: 0x54, 0x1ac: 0x54, 0x1ad: 0x62, 0x1ae: 0x63, 0x1af: 0x54, + 0x1b3: 0x64, 0x1b5: 0x65, 0x1b7: 0x66, + 0x1b8: 0x67, 0x1b9: 0x68, 0x1ba: 0x69, 0x1bb: 0x6a, 0x1bc: 0x54, 0x1bd: 0x54, 0x1be: 0x54, 0x1bf: 0x6b, + // Block 0x7, offset 0x1c0 + 0x1c0: 0x6c, 0x1c2: 0x6d, 0x1c3: 0x6e, 0x1c7: 0x6f, + 0x1c8: 0x70, 0x1c9: 0x71, 0x1ca: 0x72, 0x1cb: 0x73, 0x1cd: 0x74, 0x1cf: 0x75, + // Block 0x8, offset 0x200 + 0x237: 0x54, + // Block 0x9, offset 0x240 + 0x252: 0x76, 0x253: 0x77, + 0x258: 0x78, 0x259: 0x79, 0x25a: 0x7a, 0x25b: 0x7b, 0x25c: 0x7c, 0x25e: 0x7d, + 0x260: 0x7e, 0x261: 0x7f, 0x263: 0x80, 0x264: 0x81, 0x265: 0x82, 0x266: 0x83, 0x267: 0x84, + 0x268: 0x85, 0x269: 0x86, 0x26a: 0x87, 0x26b: 0x88, 0x26f: 0x89, + // Block 0xa, offset 0x280 + 0x2ac: 0x8a, 0x2ad: 0x8b, 0x2ae: 0x0e, 0x2af: 0x0e, + 0x2b0: 0x0e, 0x2b1: 0x0e, 0x2b2: 0x0e, 0x2b3: 0x0e, 0x2b4: 0x8c, 0x2b5: 0x0e, 0x2b6: 0x0e, 0x2b7: 0x8d, + 0x2b8: 0x8e, 0x2b9: 0x8f, 0x2ba: 0x0e, 0x2bb: 0x90, 0x2bc: 0x91, 0x2bd: 0x92, 0x2bf: 0x93, + // Block 0xb, offset 0x2c0 + 0x2c4: 0x94, 0x2c5: 0x54, 0x2c6: 0x95, 0x2c7: 0x96, + 0x2cb: 0x97, 0x2cd: 0x98, + 0x2e0: 0x99, 0x2e1: 0x99, 0x2e2: 0x99, 0x2e3: 0x99, 0x2e4: 0x9a, 0x2e5: 0x99, 0x2e6: 0x99, 0x2e7: 0x99, + 0x2e8: 0x9b, 0x2e9: 0x99, 0x2ea: 0x99, 0x2eb: 0x9c, 0x2ec: 0x9d, 0x2ed: 0x99, 0x2ee: 0x99, 0x2ef: 0x99, + 0x2f0: 0x99, 0x2f1: 0x99, 0x2f2: 0x99, 0x2f3: 0x99, 0x2f4: 0x9e, 0x2f5: 0x99, 0x2f6: 0x99, 0x2f7: 0x99, + 0x2f8: 0x99, 0x2f9: 0x9f, 0x2fa: 0x99, 0x2fb: 0x99, 0x2fc: 0xa0, 0x2fd: 0xa1, 0x2fe: 0x99, 0x2ff: 0x99, + // Block 0xc, offset 0x300 + 0x300: 0xa2, 0x301: 0xa3, 0x302: 0xa4, 0x304: 0xa5, 0x305: 0xa6, 0x306: 0xa7, 0x307: 0xa8, + 0x308: 0xa9, 0x30b: 0xaa, 0x30c: 0x26, 0x30d: 0xab, + 0x310: 0xac, 0x311: 0xad, 0x312: 0xae, 0x313: 0xaf, 0x316: 0xb0, 0x317: 0xb1, + 0x318: 0xb2, 0x319: 0xb3, 0x31a: 0xb4, 0x31c: 0xb5, + 0x320: 0xb6, 0x327: 0xb7, + 0x328: 0xb8, 0x329: 0xb9, 0x32a: 0xba, + 0x330: 0xbb, 0x332: 0xbc, 0x334: 0xbd, 0x335: 0xbe, 0x336: 0xbf, + 0x33b: 0xc0, 0x33f: 0xc1, + // Block 0xd, offset 0x340 + 0x36b: 0xc2, 0x36c: 0xc3, + 0x37d: 0xc4, 0x37e: 0xc5, 0x37f: 0xc6, + // Block 0xe, offset 0x380 + 0x3b2: 0xc7, + // Block 0xf, offset 0x3c0 + 0x3c5: 0xc8, 0x3c6: 0xc9, + 0x3c8: 0x54, 0x3c9: 0xca, 0x3cc: 0x54, 0x3cd: 0xcb, + 0x3db: 0xcc, 0x3dc: 0xcd, 0x3dd: 0xce, 0x3de: 0xcf, 0x3df: 0xd0, + 0x3e8: 0xd1, 0x3e9: 0xd2, 0x3ea: 0xd3, + // Block 0x10, offset 0x400 + 0x400: 0xd4, 0x404: 0xc3, + 0x40b: 0xd5, + 0x420: 0x99, 0x421: 0x99, 0x422: 0x99, 0x423: 0xd6, 0x424: 0x99, 0x425: 0xd7, 0x426: 0x99, 0x427: 0x99, + 0x428: 0x99, 0x429: 0x99, 0x42a: 0x99, 0x42b: 0x99, 0x42c: 0x99, 0x42d: 0x99, 0x42e: 0x99, 0x42f: 0x99, + 0x430: 0x99, 0x431: 0xa0, 0x432: 0x0e, 0x433: 0x99, 0x434: 0x0e, 0x435: 0xd8, 0x436: 0x99, 0x437: 0x99, + 0x438: 0x0e, 0x439: 0x0e, 0x43a: 0x0e, 0x43b: 0xd9, 0x43c: 0x99, 0x43d: 0x99, 0x43e: 0x99, 0x43f: 0x99, + // Block 0x11, offset 0x440 + 0x440: 0xda, 0x441: 0x54, 0x442: 0xdb, 0x443: 0xdc, 0x444: 0xdd, 0x445: 0xde, + 0x449: 0xdf, 0x44c: 0x54, 0x44d: 0x54, 0x44e: 0x54, 0x44f: 0x54, + 0x450: 0x54, 0x451: 0x54, 0x452: 0x54, 0x453: 0x54, 0x454: 0x54, 0x455: 0x54, 0x456: 0x54, 0x457: 0x54, + 0x458: 0x54, 0x459: 0x54, 0x45a: 0x54, 0x45b: 0xe0, 0x45c: 0x54, 0x45d: 0x6a, 0x45e: 0x54, 0x45f: 0xe1, + 0x460: 0xe2, 0x461: 0xe3, 0x462: 0xe4, 0x464: 0xe5, 0x465: 0xe6, 0x466: 0xe7, 0x467: 0xe8, + 0x468: 0x54, 0x469: 0xe9, 0x46a: 0xea, + 0x47f: 0xeb, + // Block 0x12, offset 0x480 + 0x4bf: 0xeb, + // Block 0x13, offset 0x4c0 + 0x4d0: 0x09, 0x4d1: 0x0a, 0x4d6: 0x0b, + 0x4db: 0x0c, 0x4dd: 0x0d, 0x4de: 0x0e, 0x4df: 0x0f, + 0x4ef: 0x10, + 0x4ff: 0x10, + // Block 0x14, offset 0x500 + 0x50f: 0x10, + 0x51f: 0x10, + 0x52f: 0x10, + 0x53f: 0x10, + // Block 0x15, offset 0x540 + 0x540: 0xec, 0x541: 0xec, 0x542: 0xec, 0x543: 0xec, 0x544: 0x05, 0x545: 0x05, 0x546: 0x05, 0x547: 0xed, + 0x548: 0xec, 0x549: 0xec, 0x54a: 0xec, 0x54b: 0xec, 0x54c: 0xec, 0x54d: 0xec, 0x54e: 0xec, 0x54f: 0xec, + 0x550: 0xec, 0x551: 0xec, 0x552: 0xec, 0x553: 0xec, 0x554: 0xec, 0x555: 0xec, 0x556: 0xec, 0x557: 0xec, + 0x558: 0xec, 0x559: 0xec, 0x55a: 0xec, 0x55b: 0xec, 0x55c: 0xec, 0x55d: 0xec, 0x55e: 0xec, 0x55f: 0xec, + 0x560: 0xec, 0x561: 0xec, 0x562: 0xec, 0x563: 0xec, 0x564: 0xec, 0x565: 0xec, 0x566: 0xec, 0x567: 0xec, + 0x568: 0xec, 0x569: 0xec, 0x56a: 0xec, 0x56b: 0xec, 0x56c: 0xec, 0x56d: 0xec, 0x56e: 0xec, 0x56f: 0xec, + 0x570: 0xec, 0x571: 0xec, 0x572: 0xec, 0x573: 0xec, 0x574: 0xec, 0x575: 0xec, 0x576: 0xec, 0x577: 0xec, + 0x578: 0xec, 0x579: 0xec, 0x57a: 0xec, 0x57b: 0xec, 0x57c: 0xec, 0x57d: 0xec, 0x57e: 0xec, 0x57f: 0xec, + // Block 0x16, offset 0x580 + 0x58f: 0x10, + 0x59f: 0x10, + 0x5a0: 0x13, + 0x5af: 0x10, + 0x5bf: 0x10, + // Block 0x17, offset 0x5c0 + 0x5cf: 0x10, +} + +// Total table size 16952 bytes (16KiB); checksum: F50EF68C diff --git a/vendor/golang.org/x/text/unicode/norm/tables11.0.0.go b/vendor/golang.org/x/text/unicode/norm/tables11.0.0.go index 7297cce3..2c58f09b 100644 --- a/vendor/golang.org/x/text/unicode/norm/tables11.0.0.go +++ b/vendor/golang.org/x/text/unicode/norm/tables11.0.0.go @@ -1,6 +1,6 @@ // Code generated by running "go generate" in golang.org/x/text. DO NOT EDIT. -// +build go1.13 +// +build go1.13,!go1.14 package norm diff --git a/vendor/golang.org/x/text/unicode/norm/tables12.0.0.go b/vendor/golang.org/x/text/unicode/norm/tables12.0.0.go new file mode 100644 index 00000000..10f5202c --- /dev/null +++ b/vendor/golang.org/x/text/unicode/norm/tables12.0.0.go @@ -0,0 +1,7710 @@ +// Code generated by running "go generate" in golang.org/x/text. DO NOT EDIT. + +// +build go1.14 + +package norm + +import "sync" + +const ( + // Version is the Unicode edition from which the tables are derived. + Version = "12.0.0" + + // MaxTransformChunkSize indicates the maximum number of bytes that Transform + // may need to write atomically for any Form. Making a destination buffer at + // least this size ensures that Transform can always make progress and that + // the user does not need to grow the buffer on an ErrShortDst. + MaxTransformChunkSize = 35 + maxNonStarters*4 +) + +var ccc = [55]uint8{ + 0, 1, 7, 8, 9, 10, 11, 12, + 13, 14, 15, 16, 17, 18, 19, 20, + 21, 22, 23, 24, 25, 26, 27, 28, + 29, 30, 31, 32, 33, 34, 35, 36, + 84, 91, 103, 107, 118, 122, 129, 130, + 132, 202, 214, 216, 218, 220, 222, 224, + 226, 228, 230, 232, 233, 234, 240, +} + +const ( + firstMulti = 0x186D + firstCCC = 0x2CA1 + endMulti = 0x2F63 + firstLeadingCCC = 0x49B1 + firstCCCZeroExcept = 0x4A7B + firstStarterWithNLead = 0x4AA2 + lastDecomp = 0x4AA4 + maxDecomp = 0x8000 +) + +// decomps: 19108 bytes +var decomps = [...]byte{ + // Bytes 0 - 3f + 0x00, 0x41, 0x20, 0x41, 0x21, 0x41, 0x22, 0x41, + 0x23, 0x41, 0x24, 0x41, 0x25, 0x41, 0x26, 0x41, + 0x27, 0x41, 0x28, 0x41, 0x29, 0x41, 0x2A, 0x41, + 0x2B, 0x41, 0x2C, 0x41, 0x2D, 0x41, 0x2E, 0x41, + 0x2F, 0x41, 0x30, 0x41, 0x31, 0x41, 0x32, 0x41, + 0x33, 0x41, 0x34, 0x41, 0x35, 0x41, 0x36, 0x41, + 0x37, 0x41, 0x38, 0x41, 0x39, 0x41, 0x3A, 0x41, + 0x3B, 0x41, 0x3C, 0x41, 0x3D, 0x41, 0x3E, 0x41, + // Bytes 40 - 7f + 0x3F, 0x41, 0x40, 0x41, 0x41, 0x41, 0x42, 0x41, + 0x43, 0x41, 0x44, 0x41, 0x45, 0x41, 0x46, 0x41, + 0x47, 0x41, 0x48, 0x41, 0x49, 0x41, 0x4A, 0x41, + 0x4B, 0x41, 0x4C, 0x41, 0x4D, 0x41, 0x4E, 0x41, + 0x4F, 0x41, 0x50, 0x41, 0x51, 0x41, 0x52, 0x41, + 0x53, 0x41, 0x54, 0x41, 0x55, 0x41, 0x56, 0x41, + 0x57, 0x41, 0x58, 0x41, 0x59, 0x41, 0x5A, 0x41, + 0x5B, 0x41, 0x5C, 0x41, 0x5D, 0x41, 0x5E, 0x41, + // Bytes 80 - bf + 0x5F, 0x41, 0x60, 0x41, 0x61, 0x41, 0x62, 0x41, + 0x63, 0x41, 0x64, 0x41, 0x65, 0x41, 0x66, 0x41, + 0x67, 0x41, 0x68, 0x41, 0x69, 0x41, 0x6A, 0x41, + 0x6B, 0x41, 0x6C, 0x41, 0x6D, 0x41, 0x6E, 0x41, + 0x6F, 0x41, 0x70, 0x41, 0x71, 0x41, 0x72, 0x41, + 0x73, 0x41, 0x74, 0x41, 0x75, 0x41, 0x76, 0x41, + 0x77, 0x41, 0x78, 0x41, 0x79, 0x41, 0x7A, 0x41, + 0x7B, 0x41, 0x7C, 0x41, 0x7D, 0x41, 0x7E, 0x42, + // Bytes c0 - ff + 0xC2, 0xA2, 0x42, 0xC2, 0xA3, 0x42, 0xC2, 0xA5, + 0x42, 0xC2, 0xA6, 0x42, 0xC2, 0xAC, 0x42, 0xC2, + 0xB7, 0x42, 0xC3, 0x86, 0x42, 0xC3, 0xB0, 0x42, + 0xC4, 0xA6, 0x42, 0xC4, 0xA7, 0x42, 0xC4, 0xB1, + 0x42, 0xC5, 0x8B, 0x42, 0xC5, 0x93, 0x42, 0xC6, + 0x8E, 0x42, 0xC6, 0x90, 0x42, 0xC6, 0xAB, 0x42, + 0xC8, 0xA2, 0x42, 0xC8, 0xB7, 0x42, 0xC9, 0x90, + 0x42, 0xC9, 0x91, 0x42, 0xC9, 0x92, 0x42, 0xC9, + // Bytes 100 - 13f + 0x94, 0x42, 0xC9, 0x95, 0x42, 0xC9, 0x99, 0x42, + 0xC9, 0x9B, 0x42, 0xC9, 0x9C, 0x42, 0xC9, 0x9F, + 0x42, 0xC9, 0xA1, 0x42, 0xC9, 0xA3, 0x42, 0xC9, + 0xA5, 0x42, 0xC9, 0xA6, 0x42, 0xC9, 0xA8, 0x42, + 0xC9, 0xA9, 0x42, 0xC9, 0xAA, 0x42, 0xC9, 0xAB, + 0x42, 0xC9, 0xAD, 0x42, 0xC9, 0xAF, 0x42, 0xC9, + 0xB0, 0x42, 0xC9, 0xB1, 0x42, 0xC9, 0xB2, 0x42, + 0xC9, 0xB3, 0x42, 0xC9, 0xB4, 0x42, 0xC9, 0xB5, + // Bytes 140 - 17f + 0x42, 0xC9, 0xB8, 0x42, 0xC9, 0xB9, 0x42, 0xC9, + 0xBB, 0x42, 0xCA, 0x81, 0x42, 0xCA, 0x82, 0x42, + 0xCA, 0x83, 0x42, 0xCA, 0x89, 0x42, 0xCA, 0x8A, + 0x42, 0xCA, 0x8B, 0x42, 0xCA, 0x8C, 0x42, 0xCA, + 0x90, 0x42, 0xCA, 0x91, 0x42, 0xCA, 0x92, 0x42, + 0xCA, 0x95, 0x42, 0xCA, 0x9D, 0x42, 0xCA, 0x9F, + 0x42, 0xCA, 0xB9, 0x42, 0xCE, 0x91, 0x42, 0xCE, + 0x92, 0x42, 0xCE, 0x93, 0x42, 0xCE, 0x94, 0x42, + // Bytes 180 - 1bf + 0xCE, 0x95, 0x42, 0xCE, 0x96, 0x42, 0xCE, 0x97, + 0x42, 0xCE, 0x98, 0x42, 0xCE, 0x99, 0x42, 0xCE, + 0x9A, 0x42, 0xCE, 0x9B, 0x42, 0xCE, 0x9C, 0x42, + 0xCE, 0x9D, 0x42, 0xCE, 0x9E, 0x42, 0xCE, 0x9F, + 0x42, 0xCE, 0xA0, 0x42, 0xCE, 0xA1, 0x42, 0xCE, + 0xA3, 0x42, 0xCE, 0xA4, 0x42, 0xCE, 0xA5, 0x42, + 0xCE, 0xA6, 0x42, 0xCE, 0xA7, 0x42, 0xCE, 0xA8, + 0x42, 0xCE, 0xA9, 0x42, 0xCE, 0xB1, 0x42, 0xCE, + // Bytes 1c0 - 1ff + 0xB2, 0x42, 0xCE, 0xB3, 0x42, 0xCE, 0xB4, 0x42, + 0xCE, 0xB5, 0x42, 0xCE, 0xB6, 0x42, 0xCE, 0xB7, + 0x42, 0xCE, 0xB8, 0x42, 0xCE, 0xB9, 0x42, 0xCE, + 0xBA, 0x42, 0xCE, 0xBB, 0x42, 0xCE, 0xBC, 0x42, + 0xCE, 0xBD, 0x42, 0xCE, 0xBE, 0x42, 0xCE, 0xBF, + 0x42, 0xCF, 0x80, 0x42, 0xCF, 0x81, 0x42, 0xCF, + 0x82, 0x42, 0xCF, 0x83, 0x42, 0xCF, 0x84, 0x42, + 0xCF, 0x85, 0x42, 0xCF, 0x86, 0x42, 0xCF, 0x87, + // Bytes 200 - 23f + 0x42, 0xCF, 0x88, 0x42, 0xCF, 0x89, 0x42, 0xCF, + 0x9C, 0x42, 0xCF, 0x9D, 0x42, 0xD0, 0xBD, 0x42, + 0xD1, 0x8A, 0x42, 0xD1, 0x8C, 0x42, 0xD7, 0x90, + 0x42, 0xD7, 0x91, 0x42, 0xD7, 0x92, 0x42, 0xD7, + 0x93, 0x42, 0xD7, 0x94, 0x42, 0xD7, 0x9B, 0x42, + 0xD7, 0x9C, 0x42, 0xD7, 0x9D, 0x42, 0xD7, 0xA2, + 0x42, 0xD7, 0xA8, 0x42, 0xD7, 0xAA, 0x42, 0xD8, + 0xA1, 0x42, 0xD8, 0xA7, 0x42, 0xD8, 0xA8, 0x42, + // Bytes 240 - 27f + 0xD8, 0xA9, 0x42, 0xD8, 0xAA, 0x42, 0xD8, 0xAB, + 0x42, 0xD8, 0xAC, 0x42, 0xD8, 0xAD, 0x42, 0xD8, + 0xAE, 0x42, 0xD8, 0xAF, 0x42, 0xD8, 0xB0, 0x42, + 0xD8, 0xB1, 0x42, 0xD8, 0xB2, 0x42, 0xD8, 0xB3, + 0x42, 0xD8, 0xB4, 0x42, 0xD8, 0xB5, 0x42, 0xD8, + 0xB6, 0x42, 0xD8, 0xB7, 0x42, 0xD8, 0xB8, 0x42, + 0xD8, 0xB9, 0x42, 0xD8, 0xBA, 0x42, 0xD9, 0x81, + 0x42, 0xD9, 0x82, 0x42, 0xD9, 0x83, 0x42, 0xD9, + // Bytes 280 - 2bf + 0x84, 0x42, 0xD9, 0x85, 0x42, 0xD9, 0x86, 0x42, + 0xD9, 0x87, 0x42, 0xD9, 0x88, 0x42, 0xD9, 0x89, + 0x42, 0xD9, 0x8A, 0x42, 0xD9, 0xAE, 0x42, 0xD9, + 0xAF, 0x42, 0xD9, 0xB1, 0x42, 0xD9, 0xB9, 0x42, + 0xD9, 0xBA, 0x42, 0xD9, 0xBB, 0x42, 0xD9, 0xBE, + 0x42, 0xD9, 0xBF, 0x42, 0xDA, 0x80, 0x42, 0xDA, + 0x83, 0x42, 0xDA, 0x84, 0x42, 0xDA, 0x86, 0x42, + 0xDA, 0x87, 0x42, 0xDA, 0x88, 0x42, 0xDA, 0x8C, + // Bytes 2c0 - 2ff + 0x42, 0xDA, 0x8D, 0x42, 0xDA, 0x8E, 0x42, 0xDA, + 0x91, 0x42, 0xDA, 0x98, 0x42, 0xDA, 0xA1, 0x42, + 0xDA, 0xA4, 0x42, 0xDA, 0xA6, 0x42, 0xDA, 0xA9, + 0x42, 0xDA, 0xAD, 0x42, 0xDA, 0xAF, 0x42, 0xDA, + 0xB1, 0x42, 0xDA, 0xB3, 0x42, 0xDA, 0xBA, 0x42, + 0xDA, 0xBB, 0x42, 0xDA, 0xBE, 0x42, 0xDB, 0x81, + 0x42, 0xDB, 0x85, 0x42, 0xDB, 0x86, 0x42, 0xDB, + 0x87, 0x42, 0xDB, 0x88, 0x42, 0xDB, 0x89, 0x42, + // Bytes 300 - 33f + 0xDB, 0x8B, 0x42, 0xDB, 0x8C, 0x42, 0xDB, 0x90, + 0x42, 0xDB, 0x92, 0x43, 0xE0, 0xBC, 0x8B, 0x43, + 0xE1, 0x83, 0x9C, 0x43, 0xE1, 0x84, 0x80, 0x43, + 0xE1, 0x84, 0x81, 0x43, 0xE1, 0x84, 0x82, 0x43, + 0xE1, 0x84, 0x83, 0x43, 0xE1, 0x84, 0x84, 0x43, + 0xE1, 0x84, 0x85, 0x43, 0xE1, 0x84, 0x86, 0x43, + 0xE1, 0x84, 0x87, 0x43, 0xE1, 0x84, 0x88, 0x43, + 0xE1, 0x84, 0x89, 0x43, 0xE1, 0x84, 0x8A, 0x43, + // Bytes 340 - 37f + 0xE1, 0x84, 0x8B, 0x43, 0xE1, 0x84, 0x8C, 0x43, + 0xE1, 0x84, 0x8D, 0x43, 0xE1, 0x84, 0x8E, 0x43, + 0xE1, 0x84, 0x8F, 0x43, 0xE1, 0x84, 0x90, 0x43, + 0xE1, 0x84, 0x91, 0x43, 0xE1, 0x84, 0x92, 0x43, + 0xE1, 0x84, 0x94, 0x43, 0xE1, 0x84, 0x95, 0x43, + 0xE1, 0x84, 0x9A, 0x43, 0xE1, 0x84, 0x9C, 0x43, + 0xE1, 0x84, 0x9D, 0x43, 0xE1, 0x84, 0x9E, 0x43, + 0xE1, 0x84, 0xA0, 0x43, 0xE1, 0x84, 0xA1, 0x43, + // Bytes 380 - 3bf + 0xE1, 0x84, 0xA2, 0x43, 0xE1, 0x84, 0xA3, 0x43, + 0xE1, 0x84, 0xA7, 0x43, 0xE1, 0x84, 0xA9, 0x43, + 0xE1, 0x84, 0xAB, 0x43, 0xE1, 0x84, 0xAC, 0x43, + 0xE1, 0x84, 0xAD, 0x43, 0xE1, 0x84, 0xAE, 0x43, + 0xE1, 0x84, 0xAF, 0x43, 0xE1, 0x84, 0xB2, 0x43, + 0xE1, 0x84, 0xB6, 0x43, 0xE1, 0x85, 0x80, 0x43, + 0xE1, 0x85, 0x87, 0x43, 0xE1, 0x85, 0x8C, 0x43, + 0xE1, 0x85, 0x97, 0x43, 0xE1, 0x85, 0x98, 0x43, + // Bytes 3c0 - 3ff + 0xE1, 0x85, 0x99, 0x43, 0xE1, 0x85, 0xA0, 0x43, + 0xE1, 0x86, 0x84, 0x43, 0xE1, 0x86, 0x85, 0x43, + 0xE1, 0x86, 0x88, 0x43, 0xE1, 0x86, 0x91, 0x43, + 0xE1, 0x86, 0x92, 0x43, 0xE1, 0x86, 0x94, 0x43, + 0xE1, 0x86, 0x9E, 0x43, 0xE1, 0x86, 0xA1, 0x43, + 0xE1, 0x87, 0x87, 0x43, 0xE1, 0x87, 0x88, 0x43, + 0xE1, 0x87, 0x8C, 0x43, 0xE1, 0x87, 0x8E, 0x43, + 0xE1, 0x87, 0x93, 0x43, 0xE1, 0x87, 0x97, 0x43, + // Bytes 400 - 43f + 0xE1, 0x87, 0x99, 0x43, 0xE1, 0x87, 0x9D, 0x43, + 0xE1, 0x87, 0x9F, 0x43, 0xE1, 0x87, 0xB1, 0x43, + 0xE1, 0x87, 0xB2, 0x43, 0xE1, 0xB4, 0x82, 0x43, + 0xE1, 0xB4, 0x96, 0x43, 0xE1, 0xB4, 0x97, 0x43, + 0xE1, 0xB4, 0x9C, 0x43, 0xE1, 0xB4, 0x9D, 0x43, + 0xE1, 0xB4, 0xA5, 0x43, 0xE1, 0xB5, 0xBB, 0x43, + 0xE1, 0xB6, 0x85, 0x43, 0xE2, 0x80, 0x82, 0x43, + 0xE2, 0x80, 0x83, 0x43, 0xE2, 0x80, 0x90, 0x43, + // Bytes 440 - 47f + 0xE2, 0x80, 0x93, 0x43, 0xE2, 0x80, 0x94, 0x43, + 0xE2, 0x82, 0xA9, 0x43, 0xE2, 0x86, 0x90, 0x43, + 0xE2, 0x86, 0x91, 0x43, 0xE2, 0x86, 0x92, 0x43, + 0xE2, 0x86, 0x93, 0x43, 0xE2, 0x88, 0x82, 0x43, + 0xE2, 0x88, 0x87, 0x43, 0xE2, 0x88, 0x91, 0x43, + 0xE2, 0x88, 0x92, 0x43, 0xE2, 0x94, 0x82, 0x43, + 0xE2, 0x96, 0xA0, 0x43, 0xE2, 0x97, 0x8B, 0x43, + 0xE2, 0xA6, 0x85, 0x43, 0xE2, 0xA6, 0x86, 0x43, + // Bytes 480 - 4bf + 0xE2, 0xB5, 0xA1, 0x43, 0xE3, 0x80, 0x81, 0x43, + 0xE3, 0x80, 0x82, 0x43, 0xE3, 0x80, 0x88, 0x43, + 0xE3, 0x80, 0x89, 0x43, 0xE3, 0x80, 0x8A, 0x43, + 0xE3, 0x80, 0x8B, 0x43, 0xE3, 0x80, 0x8C, 0x43, + 0xE3, 0x80, 0x8D, 0x43, 0xE3, 0x80, 0x8E, 0x43, + 0xE3, 0x80, 0x8F, 0x43, 0xE3, 0x80, 0x90, 0x43, + 0xE3, 0x80, 0x91, 0x43, 0xE3, 0x80, 0x92, 0x43, + 0xE3, 0x80, 0x94, 0x43, 0xE3, 0x80, 0x95, 0x43, + // Bytes 4c0 - 4ff + 0xE3, 0x80, 0x96, 0x43, 0xE3, 0x80, 0x97, 0x43, + 0xE3, 0x82, 0xA1, 0x43, 0xE3, 0x82, 0xA2, 0x43, + 0xE3, 0x82, 0xA3, 0x43, 0xE3, 0x82, 0xA4, 0x43, + 0xE3, 0x82, 0xA5, 0x43, 0xE3, 0x82, 0xA6, 0x43, + 0xE3, 0x82, 0xA7, 0x43, 0xE3, 0x82, 0xA8, 0x43, + 0xE3, 0x82, 0xA9, 0x43, 0xE3, 0x82, 0xAA, 0x43, + 0xE3, 0x82, 0xAB, 0x43, 0xE3, 0x82, 0xAD, 0x43, + 0xE3, 0x82, 0xAF, 0x43, 0xE3, 0x82, 0xB1, 0x43, + // Bytes 500 - 53f + 0xE3, 0x82, 0xB3, 0x43, 0xE3, 0x82, 0xB5, 0x43, + 0xE3, 0x82, 0xB7, 0x43, 0xE3, 0x82, 0xB9, 0x43, + 0xE3, 0x82, 0xBB, 0x43, 0xE3, 0x82, 0xBD, 0x43, + 0xE3, 0x82, 0xBF, 0x43, 0xE3, 0x83, 0x81, 0x43, + 0xE3, 0x83, 0x83, 0x43, 0xE3, 0x83, 0x84, 0x43, + 0xE3, 0x83, 0x86, 0x43, 0xE3, 0x83, 0x88, 0x43, + 0xE3, 0x83, 0x8A, 0x43, 0xE3, 0x83, 0x8B, 0x43, + 0xE3, 0x83, 0x8C, 0x43, 0xE3, 0x83, 0x8D, 0x43, + // Bytes 540 - 57f + 0xE3, 0x83, 0x8E, 0x43, 0xE3, 0x83, 0x8F, 0x43, + 0xE3, 0x83, 0x92, 0x43, 0xE3, 0x83, 0x95, 0x43, + 0xE3, 0x83, 0x98, 0x43, 0xE3, 0x83, 0x9B, 0x43, + 0xE3, 0x83, 0x9E, 0x43, 0xE3, 0x83, 0x9F, 0x43, + 0xE3, 0x83, 0xA0, 0x43, 0xE3, 0x83, 0xA1, 0x43, + 0xE3, 0x83, 0xA2, 0x43, 0xE3, 0x83, 0xA3, 0x43, + 0xE3, 0x83, 0xA4, 0x43, 0xE3, 0x83, 0xA5, 0x43, + 0xE3, 0x83, 0xA6, 0x43, 0xE3, 0x83, 0xA7, 0x43, + // Bytes 580 - 5bf + 0xE3, 0x83, 0xA8, 0x43, 0xE3, 0x83, 0xA9, 0x43, + 0xE3, 0x83, 0xAA, 0x43, 0xE3, 0x83, 0xAB, 0x43, + 0xE3, 0x83, 0xAC, 0x43, 0xE3, 0x83, 0xAD, 0x43, + 0xE3, 0x83, 0xAF, 0x43, 0xE3, 0x83, 0xB0, 0x43, + 0xE3, 0x83, 0xB1, 0x43, 0xE3, 0x83, 0xB2, 0x43, + 0xE3, 0x83, 0xB3, 0x43, 0xE3, 0x83, 0xBB, 0x43, + 0xE3, 0x83, 0xBC, 0x43, 0xE3, 0x92, 0x9E, 0x43, + 0xE3, 0x92, 0xB9, 0x43, 0xE3, 0x92, 0xBB, 0x43, + // Bytes 5c0 - 5ff + 0xE3, 0x93, 0x9F, 0x43, 0xE3, 0x94, 0x95, 0x43, + 0xE3, 0x9B, 0xAE, 0x43, 0xE3, 0x9B, 0xBC, 0x43, + 0xE3, 0x9E, 0x81, 0x43, 0xE3, 0xA0, 0xAF, 0x43, + 0xE3, 0xA1, 0xA2, 0x43, 0xE3, 0xA1, 0xBC, 0x43, + 0xE3, 0xA3, 0x87, 0x43, 0xE3, 0xA3, 0xA3, 0x43, + 0xE3, 0xA4, 0x9C, 0x43, 0xE3, 0xA4, 0xBA, 0x43, + 0xE3, 0xA8, 0xAE, 0x43, 0xE3, 0xA9, 0xAC, 0x43, + 0xE3, 0xAB, 0xA4, 0x43, 0xE3, 0xAC, 0x88, 0x43, + // Bytes 600 - 63f + 0xE3, 0xAC, 0x99, 0x43, 0xE3, 0xAD, 0x89, 0x43, + 0xE3, 0xAE, 0x9D, 0x43, 0xE3, 0xB0, 0x98, 0x43, + 0xE3, 0xB1, 0x8E, 0x43, 0xE3, 0xB4, 0xB3, 0x43, + 0xE3, 0xB6, 0x96, 0x43, 0xE3, 0xBA, 0xAC, 0x43, + 0xE3, 0xBA, 0xB8, 0x43, 0xE3, 0xBC, 0x9B, 0x43, + 0xE3, 0xBF, 0xBC, 0x43, 0xE4, 0x80, 0x88, 0x43, + 0xE4, 0x80, 0x98, 0x43, 0xE4, 0x80, 0xB9, 0x43, + 0xE4, 0x81, 0x86, 0x43, 0xE4, 0x82, 0x96, 0x43, + // Bytes 640 - 67f + 0xE4, 0x83, 0xA3, 0x43, 0xE4, 0x84, 0xAF, 0x43, + 0xE4, 0x88, 0x82, 0x43, 0xE4, 0x88, 0xA7, 0x43, + 0xE4, 0x8A, 0xA0, 0x43, 0xE4, 0x8C, 0x81, 0x43, + 0xE4, 0x8C, 0xB4, 0x43, 0xE4, 0x8D, 0x99, 0x43, + 0xE4, 0x8F, 0x95, 0x43, 0xE4, 0x8F, 0x99, 0x43, + 0xE4, 0x90, 0x8B, 0x43, 0xE4, 0x91, 0xAB, 0x43, + 0xE4, 0x94, 0xAB, 0x43, 0xE4, 0x95, 0x9D, 0x43, + 0xE4, 0x95, 0xA1, 0x43, 0xE4, 0x95, 0xAB, 0x43, + // Bytes 680 - 6bf + 0xE4, 0x97, 0x97, 0x43, 0xE4, 0x97, 0xB9, 0x43, + 0xE4, 0x98, 0xB5, 0x43, 0xE4, 0x9A, 0xBE, 0x43, + 0xE4, 0x9B, 0x87, 0x43, 0xE4, 0xA6, 0x95, 0x43, + 0xE4, 0xA7, 0xA6, 0x43, 0xE4, 0xA9, 0xAE, 0x43, + 0xE4, 0xA9, 0xB6, 0x43, 0xE4, 0xAA, 0xB2, 0x43, + 0xE4, 0xAC, 0xB3, 0x43, 0xE4, 0xAF, 0x8E, 0x43, + 0xE4, 0xB3, 0x8E, 0x43, 0xE4, 0xB3, 0xAD, 0x43, + 0xE4, 0xB3, 0xB8, 0x43, 0xE4, 0xB5, 0x96, 0x43, + // Bytes 6c0 - 6ff + 0xE4, 0xB8, 0x80, 0x43, 0xE4, 0xB8, 0x81, 0x43, + 0xE4, 0xB8, 0x83, 0x43, 0xE4, 0xB8, 0x89, 0x43, + 0xE4, 0xB8, 0x8A, 0x43, 0xE4, 0xB8, 0x8B, 0x43, + 0xE4, 0xB8, 0x8D, 0x43, 0xE4, 0xB8, 0x99, 0x43, + 0xE4, 0xB8, 0xA6, 0x43, 0xE4, 0xB8, 0xA8, 0x43, + 0xE4, 0xB8, 0xAD, 0x43, 0xE4, 0xB8, 0xB2, 0x43, + 0xE4, 0xB8, 0xB6, 0x43, 0xE4, 0xB8, 0xB8, 0x43, + 0xE4, 0xB8, 0xB9, 0x43, 0xE4, 0xB8, 0xBD, 0x43, + // Bytes 700 - 73f + 0xE4, 0xB8, 0xBF, 0x43, 0xE4, 0xB9, 0x81, 0x43, + 0xE4, 0xB9, 0x99, 0x43, 0xE4, 0xB9, 0x9D, 0x43, + 0xE4, 0xBA, 0x82, 0x43, 0xE4, 0xBA, 0x85, 0x43, + 0xE4, 0xBA, 0x86, 0x43, 0xE4, 0xBA, 0x8C, 0x43, + 0xE4, 0xBA, 0x94, 0x43, 0xE4, 0xBA, 0xA0, 0x43, + 0xE4, 0xBA, 0xA4, 0x43, 0xE4, 0xBA, 0xAE, 0x43, + 0xE4, 0xBA, 0xBA, 0x43, 0xE4, 0xBB, 0x80, 0x43, + 0xE4, 0xBB, 0x8C, 0x43, 0xE4, 0xBB, 0xA4, 0x43, + // Bytes 740 - 77f + 0xE4, 0xBC, 0x81, 0x43, 0xE4, 0xBC, 0x91, 0x43, + 0xE4, 0xBD, 0xA0, 0x43, 0xE4, 0xBE, 0x80, 0x43, + 0xE4, 0xBE, 0x86, 0x43, 0xE4, 0xBE, 0x8B, 0x43, + 0xE4, 0xBE, 0xAE, 0x43, 0xE4, 0xBE, 0xBB, 0x43, + 0xE4, 0xBE, 0xBF, 0x43, 0xE5, 0x80, 0x82, 0x43, + 0xE5, 0x80, 0xAB, 0x43, 0xE5, 0x81, 0xBA, 0x43, + 0xE5, 0x82, 0x99, 0x43, 0xE5, 0x83, 0x8F, 0x43, + 0xE5, 0x83, 0x9A, 0x43, 0xE5, 0x83, 0xA7, 0x43, + // Bytes 780 - 7bf + 0xE5, 0x84, 0xAA, 0x43, 0xE5, 0x84, 0xBF, 0x43, + 0xE5, 0x85, 0x80, 0x43, 0xE5, 0x85, 0x85, 0x43, + 0xE5, 0x85, 0x8D, 0x43, 0xE5, 0x85, 0x94, 0x43, + 0xE5, 0x85, 0xA4, 0x43, 0xE5, 0x85, 0xA5, 0x43, + 0xE5, 0x85, 0xA7, 0x43, 0xE5, 0x85, 0xA8, 0x43, + 0xE5, 0x85, 0xA9, 0x43, 0xE5, 0x85, 0xAB, 0x43, + 0xE5, 0x85, 0xAD, 0x43, 0xE5, 0x85, 0xB7, 0x43, + 0xE5, 0x86, 0x80, 0x43, 0xE5, 0x86, 0x82, 0x43, + // Bytes 7c0 - 7ff + 0xE5, 0x86, 0x8D, 0x43, 0xE5, 0x86, 0x92, 0x43, + 0xE5, 0x86, 0x95, 0x43, 0xE5, 0x86, 0x96, 0x43, + 0xE5, 0x86, 0x97, 0x43, 0xE5, 0x86, 0x99, 0x43, + 0xE5, 0x86, 0xA4, 0x43, 0xE5, 0x86, 0xAB, 0x43, + 0xE5, 0x86, 0xAC, 0x43, 0xE5, 0x86, 0xB5, 0x43, + 0xE5, 0x86, 0xB7, 0x43, 0xE5, 0x87, 0x89, 0x43, + 0xE5, 0x87, 0x8C, 0x43, 0xE5, 0x87, 0x9C, 0x43, + 0xE5, 0x87, 0x9E, 0x43, 0xE5, 0x87, 0xA0, 0x43, + // Bytes 800 - 83f + 0xE5, 0x87, 0xB5, 0x43, 0xE5, 0x88, 0x80, 0x43, + 0xE5, 0x88, 0x83, 0x43, 0xE5, 0x88, 0x87, 0x43, + 0xE5, 0x88, 0x97, 0x43, 0xE5, 0x88, 0x9D, 0x43, + 0xE5, 0x88, 0xA9, 0x43, 0xE5, 0x88, 0xBA, 0x43, + 0xE5, 0x88, 0xBB, 0x43, 0xE5, 0x89, 0x86, 0x43, + 0xE5, 0x89, 0x8D, 0x43, 0xE5, 0x89, 0xB2, 0x43, + 0xE5, 0x89, 0xB7, 0x43, 0xE5, 0x8A, 0x89, 0x43, + 0xE5, 0x8A, 0x9B, 0x43, 0xE5, 0x8A, 0xA3, 0x43, + // Bytes 840 - 87f + 0xE5, 0x8A, 0xB3, 0x43, 0xE5, 0x8A, 0xB4, 0x43, + 0xE5, 0x8B, 0x87, 0x43, 0xE5, 0x8B, 0x89, 0x43, + 0xE5, 0x8B, 0x92, 0x43, 0xE5, 0x8B, 0x9E, 0x43, + 0xE5, 0x8B, 0xA4, 0x43, 0xE5, 0x8B, 0xB5, 0x43, + 0xE5, 0x8B, 0xB9, 0x43, 0xE5, 0x8B, 0xBA, 0x43, + 0xE5, 0x8C, 0x85, 0x43, 0xE5, 0x8C, 0x86, 0x43, + 0xE5, 0x8C, 0x95, 0x43, 0xE5, 0x8C, 0x97, 0x43, + 0xE5, 0x8C, 0x9A, 0x43, 0xE5, 0x8C, 0xB8, 0x43, + // Bytes 880 - 8bf + 0xE5, 0x8C, 0xBB, 0x43, 0xE5, 0x8C, 0xBF, 0x43, + 0xE5, 0x8D, 0x81, 0x43, 0xE5, 0x8D, 0x84, 0x43, + 0xE5, 0x8D, 0x85, 0x43, 0xE5, 0x8D, 0x89, 0x43, + 0xE5, 0x8D, 0x91, 0x43, 0xE5, 0x8D, 0x94, 0x43, + 0xE5, 0x8D, 0x9A, 0x43, 0xE5, 0x8D, 0x9C, 0x43, + 0xE5, 0x8D, 0xA9, 0x43, 0xE5, 0x8D, 0xB0, 0x43, + 0xE5, 0x8D, 0xB3, 0x43, 0xE5, 0x8D, 0xB5, 0x43, + 0xE5, 0x8D, 0xBD, 0x43, 0xE5, 0x8D, 0xBF, 0x43, + // Bytes 8c0 - 8ff + 0xE5, 0x8E, 0x82, 0x43, 0xE5, 0x8E, 0xB6, 0x43, + 0xE5, 0x8F, 0x83, 0x43, 0xE5, 0x8F, 0x88, 0x43, + 0xE5, 0x8F, 0x8A, 0x43, 0xE5, 0x8F, 0x8C, 0x43, + 0xE5, 0x8F, 0x9F, 0x43, 0xE5, 0x8F, 0xA3, 0x43, + 0xE5, 0x8F, 0xA5, 0x43, 0xE5, 0x8F, 0xAB, 0x43, + 0xE5, 0x8F, 0xAF, 0x43, 0xE5, 0x8F, 0xB1, 0x43, + 0xE5, 0x8F, 0xB3, 0x43, 0xE5, 0x90, 0x86, 0x43, + 0xE5, 0x90, 0x88, 0x43, 0xE5, 0x90, 0x8D, 0x43, + // Bytes 900 - 93f + 0xE5, 0x90, 0x8F, 0x43, 0xE5, 0x90, 0x9D, 0x43, + 0xE5, 0x90, 0xB8, 0x43, 0xE5, 0x90, 0xB9, 0x43, + 0xE5, 0x91, 0x82, 0x43, 0xE5, 0x91, 0x88, 0x43, + 0xE5, 0x91, 0xA8, 0x43, 0xE5, 0x92, 0x9E, 0x43, + 0xE5, 0x92, 0xA2, 0x43, 0xE5, 0x92, 0xBD, 0x43, + 0xE5, 0x93, 0xB6, 0x43, 0xE5, 0x94, 0x90, 0x43, + 0xE5, 0x95, 0x8F, 0x43, 0xE5, 0x95, 0x93, 0x43, + 0xE5, 0x95, 0x95, 0x43, 0xE5, 0x95, 0xA3, 0x43, + // Bytes 940 - 97f + 0xE5, 0x96, 0x84, 0x43, 0xE5, 0x96, 0x87, 0x43, + 0xE5, 0x96, 0x99, 0x43, 0xE5, 0x96, 0x9D, 0x43, + 0xE5, 0x96, 0xAB, 0x43, 0xE5, 0x96, 0xB3, 0x43, + 0xE5, 0x96, 0xB6, 0x43, 0xE5, 0x97, 0x80, 0x43, + 0xE5, 0x97, 0x82, 0x43, 0xE5, 0x97, 0xA2, 0x43, + 0xE5, 0x98, 0x86, 0x43, 0xE5, 0x99, 0x91, 0x43, + 0xE5, 0x99, 0xA8, 0x43, 0xE5, 0x99, 0xB4, 0x43, + 0xE5, 0x9B, 0x97, 0x43, 0xE5, 0x9B, 0x9B, 0x43, + // Bytes 980 - 9bf + 0xE5, 0x9B, 0xB9, 0x43, 0xE5, 0x9C, 0x96, 0x43, + 0xE5, 0x9C, 0x97, 0x43, 0xE5, 0x9C, 0x9F, 0x43, + 0xE5, 0x9C, 0xB0, 0x43, 0xE5, 0x9E, 0x8B, 0x43, + 0xE5, 0x9F, 0x8E, 0x43, 0xE5, 0x9F, 0xB4, 0x43, + 0xE5, 0xA0, 0x8D, 0x43, 0xE5, 0xA0, 0xB1, 0x43, + 0xE5, 0xA0, 0xB2, 0x43, 0xE5, 0xA1, 0x80, 0x43, + 0xE5, 0xA1, 0x9A, 0x43, 0xE5, 0xA1, 0x9E, 0x43, + 0xE5, 0xA2, 0xA8, 0x43, 0xE5, 0xA2, 0xAC, 0x43, + // Bytes 9c0 - 9ff + 0xE5, 0xA2, 0xB3, 0x43, 0xE5, 0xA3, 0x98, 0x43, + 0xE5, 0xA3, 0x9F, 0x43, 0xE5, 0xA3, 0xAB, 0x43, + 0xE5, 0xA3, 0xAE, 0x43, 0xE5, 0xA3, 0xB0, 0x43, + 0xE5, 0xA3, 0xB2, 0x43, 0xE5, 0xA3, 0xB7, 0x43, + 0xE5, 0xA4, 0x82, 0x43, 0xE5, 0xA4, 0x86, 0x43, + 0xE5, 0xA4, 0x8A, 0x43, 0xE5, 0xA4, 0x95, 0x43, + 0xE5, 0xA4, 0x9A, 0x43, 0xE5, 0xA4, 0x9C, 0x43, + 0xE5, 0xA4, 0xA2, 0x43, 0xE5, 0xA4, 0xA7, 0x43, + // Bytes a00 - a3f + 0xE5, 0xA4, 0xA9, 0x43, 0xE5, 0xA5, 0x84, 0x43, + 0xE5, 0xA5, 0x88, 0x43, 0xE5, 0xA5, 0x91, 0x43, + 0xE5, 0xA5, 0x94, 0x43, 0xE5, 0xA5, 0xA2, 0x43, + 0xE5, 0xA5, 0xB3, 0x43, 0xE5, 0xA7, 0x98, 0x43, + 0xE5, 0xA7, 0xAC, 0x43, 0xE5, 0xA8, 0x9B, 0x43, + 0xE5, 0xA8, 0xA7, 0x43, 0xE5, 0xA9, 0xA2, 0x43, + 0xE5, 0xA9, 0xA6, 0x43, 0xE5, 0xAA, 0xB5, 0x43, + 0xE5, 0xAC, 0x88, 0x43, 0xE5, 0xAC, 0xA8, 0x43, + // Bytes a40 - a7f + 0xE5, 0xAC, 0xBE, 0x43, 0xE5, 0xAD, 0x90, 0x43, + 0xE5, 0xAD, 0x97, 0x43, 0xE5, 0xAD, 0xA6, 0x43, + 0xE5, 0xAE, 0x80, 0x43, 0xE5, 0xAE, 0x85, 0x43, + 0xE5, 0xAE, 0x97, 0x43, 0xE5, 0xAF, 0x83, 0x43, + 0xE5, 0xAF, 0x98, 0x43, 0xE5, 0xAF, 0xA7, 0x43, + 0xE5, 0xAF, 0xAE, 0x43, 0xE5, 0xAF, 0xB3, 0x43, + 0xE5, 0xAF, 0xB8, 0x43, 0xE5, 0xAF, 0xBF, 0x43, + 0xE5, 0xB0, 0x86, 0x43, 0xE5, 0xB0, 0x8F, 0x43, + // Bytes a80 - abf + 0xE5, 0xB0, 0xA2, 0x43, 0xE5, 0xB0, 0xB8, 0x43, + 0xE5, 0xB0, 0xBF, 0x43, 0xE5, 0xB1, 0xA0, 0x43, + 0xE5, 0xB1, 0xA2, 0x43, 0xE5, 0xB1, 0xA4, 0x43, + 0xE5, 0xB1, 0xA5, 0x43, 0xE5, 0xB1, 0xAE, 0x43, + 0xE5, 0xB1, 0xB1, 0x43, 0xE5, 0xB2, 0x8D, 0x43, + 0xE5, 0xB3, 0x80, 0x43, 0xE5, 0xB4, 0x99, 0x43, + 0xE5, 0xB5, 0x83, 0x43, 0xE5, 0xB5, 0x90, 0x43, + 0xE5, 0xB5, 0xAB, 0x43, 0xE5, 0xB5, 0xAE, 0x43, + // Bytes ac0 - aff + 0xE5, 0xB5, 0xBC, 0x43, 0xE5, 0xB6, 0xB2, 0x43, + 0xE5, 0xB6, 0xBA, 0x43, 0xE5, 0xB7, 0x9B, 0x43, + 0xE5, 0xB7, 0xA1, 0x43, 0xE5, 0xB7, 0xA2, 0x43, + 0xE5, 0xB7, 0xA5, 0x43, 0xE5, 0xB7, 0xA6, 0x43, + 0xE5, 0xB7, 0xB1, 0x43, 0xE5, 0xB7, 0xBD, 0x43, + 0xE5, 0xB7, 0xBE, 0x43, 0xE5, 0xB8, 0xA8, 0x43, + 0xE5, 0xB8, 0xBD, 0x43, 0xE5, 0xB9, 0xA9, 0x43, + 0xE5, 0xB9, 0xB2, 0x43, 0xE5, 0xB9, 0xB4, 0x43, + // Bytes b00 - b3f + 0xE5, 0xB9, 0xBA, 0x43, 0xE5, 0xB9, 0xBC, 0x43, + 0xE5, 0xB9, 0xBF, 0x43, 0xE5, 0xBA, 0xA6, 0x43, + 0xE5, 0xBA, 0xB0, 0x43, 0xE5, 0xBA, 0xB3, 0x43, + 0xE5, 0xBA, 0xB6, 0x43, 0xE5, 0xBB, 0x89, 0x43, + 0xE5, 0xBB, 0x8A, 0x43, 0xE5, 0xBB, 0x92, 0x43, + 0xE5, 0xBB, 0x93, 0x43, 0xE5, 0xBB, 0x99, 0x43, + 0xE5, 0xBB, 0xAC, 0x43, 0xE5, 0xBB, 0xB4, 0x43, + 0xE5, 0xBB, 0xBE, 0x43, 0xE5, 0xBC, 0x84, 0x43, + // Bytes b40 - b7f + 0xE5, 0xBC, 0x8B, 0x43, 0xE5, 0xBC, 0x93, 0x43, + 0xE5, 0xBC, 0xA2, 0x43, 0xE5, 0xBD, 0x90, 0x43, + 0xE5, 0xBD, 0x93, 0x43, 0xE5, 0xBD, 0xA1, 0x43, + 0xE5, 0xBD, 0xA2, 0x43, 0xE5, 0xBD, 0xA9, 0x43, + 0xE5, 0xBD, 0xAB, 0x43, 0xE5, 0xBD, 0xB3, 0x43, + 0xE5, 0xBE, 0x8B, 0x43, 0xE5, 0xBE, 0x8C, 0x43, + 0xE5, 0xBE, 0x97, 0x43, 0xE5, 0xBE, 0x9A, 0x43, + 0xE5, 0xBE, 0xA9, 0x43, 0xE5, 0xBE, 0xAD, 0x43, + // Bytes b80 - bbf + 0xE5, 0xBF, 0x83, 0x43, 0xE5, 0xBF, 0x8D, 0x43, + 0xE5, 0xBF, 0x97, 0x43, 0xE5, 0xBF, 0xB5, 0x43, + 0xE5, 0xBF, 0xB9, 0x43, 0xE6, 0x80, 0x92, 0x43, + 0xE6, 0x80, 0x9C, 0x43, 0xE6, 0x81, 0xB5, 0x43, + 0xE6, 0x82, 0x81, 0x43, 0xE6, 0x82, 0x94, 0x43, + 0xE6, 0x83, 0x87, 0x43, 0xE6, 0x83, 0x98, 0x43, + 0xE6, 0x83, 0xA1, 0x43, 0xE6, 0x84, 0x88, 0x43, + 0xE6, 0x85, 0x84, 0x43, 0xE6, 0x85, 0x88, 0x43, + // Bytes bc0 - bff + 0xE6, 0x85, 0x8C, 0x43, 0xE6, 0x85, 0x8E, 0x43, + 0xE6, 0x85, 0xA0, 0x43, 0xE6, 0x85, 0xA8, 0x43, + 0xE6, 0x85, 0xBA, 0x43, 0xE6, 0x86, 0x8E, 0x43, + 0xE6, 0x86, 0x90, 0x43, 0xE6, 0x86, 0xA4, 0x43, + 0xE6, 0x86, 0xAF, 0x43, 0xE6, 0x86, 0xB2, 0x43, + 0xE6, 0x87, 0x9E, 0x43, 0xE6, 0x87, 0xB2, 0x43, + 0xE6, 0x87, 0xB6, 0x43, 0xE6, 0x88, 0x80, 0x43, + 0xE6, 0x88, 0x88, 0x43, 0xE6, 0x88, 0x90, 0x43, + // Bytes c00 - c3f + 0xE6, 0x88, 0x9B, 0x43, 0xE6, 0x88, 0xAE, 0x43, + 0xE6, 0x88, 0xB4, 0x43, 0xE6, 0x88, 0xB6, 0x43, + 0xE6, 0x89, 0x8B, 0x43, 0xE6, 0x89, 0x93, 0x43, + 0xE6, 0x89, 0x9D, 0x43, 0xE6, 0x8A, 0x95, 0x43, + 0xE6, 0x8A, 0xB1, 0x43, 0xE6, 0x8B, 0x89, 0x43, + 0xE6, 0x8B, 0x8F, 0x43, 0xE6, 0x8B, 0x93, 0x43, + 0xE6, 0x8B, 0x94, 0x43, 0xE6, 0x8B, 0xBC, 0x43, + 0xE6, 0x8B, 0xBE, 0x43, 0xE6, 0x8C, 0x87, 0x43, + // Bytes c40 - c7f + 0xE6, 0x8C, 0xBD, 0x43, 0xE6, 0x8D, 0x90, 0x43, + 0xE6, 0x8D, 0x95, 0x43, 0xE6, 0x8D, 0xA8, 0x43, + 0xE6, 0x8D, 0xBB, 0x43, 0xE6, 0x8E, 0x83, 0x43, + 0xE6, 0x8E, 0xA0, 0x43, 0xE6, 0x8E, 0xA9, 0x43, + 0xE6, 0x8F, 0x84, 0x43, 0xE6, 0x8F, 0x85, 0x43, + 0xE6, 0x8F, 0xA4, 0x43, 0xE6, 0x90, 0x9C, 0x43, + 0xE6, 0x90, 0xA2, 0x43, 0xE6, 0x91, 0x92, 0x43, + 0xE6, 0x91, 0xA9, 0x43, 0xE6, 0x91, 0xB7, 0x43, + // Bytes c80 - cbf + 0xE6, 0x91, 0xBE, 0x43, 0xE6, 0x92, 0x9A, 0x43, + 0xE6, 0x92, 0x9D, 0x43, 0xE6, 0x93, 0x84, 0x43, + 0xE6, 0x94, 0xAF, 0x43, 0xE6, 0x94, 0xB4, 0x43, + 0xE6, 0x95, 0x8F, 0x43, 0xE6, 0x95, 0x96, 0x43, + 0xE6, 0x95, 0xAC, 0x43, 0xE6, 0x95, 0xB8, 0x43, + 0xE6, 0x96, 0x87, 0x43, 0xE6, 0x96, 0x97, 0x43, + 0xE6, 0x96, 0x99, 0x43, 0xE6, 0x96, 0xA4, 0x43, + 0xE6, 0x96, 0xB0, 0x43, 0xE6, 0x96, 0xB9, 0x43, + // Bytes cc0 - cff + 0xE6, 0x97, 0x85, 0x43, 0xE6, 0x97, 0xA0, 0x43, + 0xE6, 0x97, 0xA2, 0x43, 0xE6, 0x97, 0xA3, 0x43, + 0xE6, 0x97, 0xA5, 0x43, 0xE6, 0x98, 0x93, 0x43, + 0xE6, 0x98, 0xA0, 0x43, 0xE6, 0x99, 0x89, 0x43, + 0xE6, 0x99, 0xB4, 0x43, 0xE6, 0x9A, 0x88, 0x43, + 0xE6, 0x9A, 0x91, 0x43, 0xE6, 0x9A, 0x9C, 0x43, + 0xE6, 0x9A, 0xB4, 0x43, 0xE6, 0x9B, 0x86, 0x43, + 0xE6, 0x9B, 0xB0, 0x43, 0xE6, 0x9B, 0xB4, 0x43, + // Bytes d00 - d3f + 0xE6, 0x9B, 0xB8, 0x43, 0xE6, 0x9C, 0x80, 0x43, + 0xE6, 0x9C, 0x88, 0x43, 0xE6, 0x9C, 0x89, 0x43, + 0xE6, 0x9C, 0x97, 0x43, 0xE6, 0x9C, 0x9B, 0x43, + 0xE6, 0x9C, 0xA1, 0x43, 0xE6, 0x9C, 0xA8, 0x43, + 0xE6, 0x9D, 0x8E, 0x43, 0xE6, 0x9D, 0x93, 0x43, + 0xE6, 0x9D, 0x96, 0x43, 0xE6, 0x9D, 0x9E, 0x43, + 0xE6, 0x9D, 0xBB, 0x43, 0xE6, 0x9E, 0x85, 0x43, + 0xE6, 0x9E, 0x97, 0x43, 0xE6, 0x9F, 0xB3, 0x43, + // Bytes d40 - d7f + 0xE6, 0x9F, 0xBA, 0x43, 0xE6, 0xA0, 0x97, 0x43, + 0xE6, 0xA0, 0x9F, 0x43, 0xE6, 0xA0, 0xAA, 0x43, + 0xE6, 0xA1, 0x92, 0x43, 0xE6, 0xA2, 0x81, 0x43, + 0xE6, 0xA2, 0x85, 0x43, 0xE6, 0xA2, 0x8E, 0x43, + 0xE6, 0xA2, 0xA8, 0x43, 0xE6, 0xA4, 0x94, 0x43, + 0xE6, 0xA5, 0x82, 0x43, 0xE6, 0xA6, 0xA3, 0x43, + 0xE6, 0xA7, 0xAA, 0x43, 0xE6, 0xA8, 0x82, 0x43, + 0xE6, 0xA8, 0x93, 0x43, 0xE6, 0xAA, 0xA8, 0x43, + // Bytes d80 - dbf + 0xE6, 0xAB, 0x93, 0x43, 0xE6, 0xAB, 0x9B, 0x43, + 0xE6, 0xAC, 0x84, 0x43, 0xE6, 0xAC, 0xA0, 0x43, + 0xE6, 0xAC, 0xA1, 0x43, 0xE6, 0xAD, 0x94, 0x43, + 0xE6, 0xAD, 0xA2, 0x43, 0xE6, 0xAD, 0xA3, 0x43, + 0xE6, 0xAD, 0xB2, 0x43, 0xE6, 0xAD, 0xB7, 0x43, + 0xE6, 0xAD, 0xB9, 0x43, 0xE6, 0xAE, 0x9F, 0x43, + 0xE6, 0xAE, 0xAE, 0x43, 0xE6, 0xAE, 0xB3, 0x43, + 0xE6, 0xAE, 0xBA, 0x43, 0xE6, 0xAE, 0xBB, 0x43, + // Bytes dc0 - dff + 0xE6, 0xAF, 0x8B, 0x43, 0xE6, 0xAF, 0x8D, 0x43, + 0xE6, 0xAF, 0x94, 0x43, 0xE6, 0xAF, 0x9B, 0x43, + 0xE6, 0xB0, 0x8F, 0x43, 0xE6, 0xB0, 0x94, 0x43, + 0xE6, 0xB0, 0xB4, 0x43, 0xE6, 0xB1, 0x8E, 0x43, + 0xE6, 0xB1, 0xA7, 0x43, 0xE6, 0xB2, 0x88, 0x43, + 0xE6, 0xB2, 0xBF, 0x43, 0xE6, 0xB3, 0x8C, 0x43, + 0xE6, 0xB3, 0x8D, 0x43, 0xE6, 0xB3, 0xA5, 0x43, + 0xE6, 0xB3, 0xA8, 0x43, 0xE6, 0xB4, 0x96, 0x43, + // Bytes e00 - e3f + 0xE6, 0xB4, 0x9B, 0x43, 0xE6, 0xB4, 0x9E, 0x43, + 0xE6, 0xB4, 0xB4, 0x43, 0xE6, 0xB4, 0xBE, 0x43, + 0xE6, 0xB5, 0x81, 0x43, 0xE6, 0xB5, 0xA9, 0x43, + 0xE6, 0xB5, 0xAA, 0x43, 0xE6, 0xB5, 0xB7, 0x43, + 0xE6, 0xB5, 0xB8, 0x43, 0xE6, 0xB6, 0x85, 0x43, + 0xE6, 0xB7, 0x8B, 0x43, 0xE6, 0xB7, 0x9A, 0x43, + 0xE6, 0xB7, 0xAA, 0x43, 0xE6, 0xB7, 0xB9, 0x43, + 0xE6, 0xB8, 0x9A, 0x43, 0xE6, 0xB8, 0xAF, 0x43, + // Bytes e40 - e7f + 0xE6, 0xB9, 0xAE, 0x43, 0xE6, 0xBA, 0x80, 0x43, + 0xE6, 0xBA, 0x9C, 0x43, 0xE6, 0xBA, 0xBA, 0x43, + 0xE6, 0xBB, 0x87, 0x43, 0xE6, 0xBB, 0x8B, 0x43, + 0xE6, 0xBB, 0x91, 0x43, 0xE6, 0xBB, 0x9B, 0x43, + 0xE6, 0xBC, 0x8F, 0x43, 0xE6, 0xBC, 0x94, 0x43, + 0xE6, 0xBC, 0xA2, 0x43, 0xE6, 0xBC, 0xA3, 0x43, + 0xE6, 0xBD, 0xAE, 0x43, 0xE6, 0xBF, 0x86, 0x43, + 0xE6, 0xBF, 0xAB, 0x43, 0xE6, 0xBF, 0xBE, 0x43, + // Bytes e80 - ebf + 0xE7, 0x80, 0x9B, 0x43, 0xE7, 0x80, 0x9E, 0x43, + 0xE7, 0x80, 0xB9, 0x43, 0xE7, 0x81, 0x8A, 0x43, + 0xE7, 0x81, 0xAB, 0x43, 0xE7, 0x81, 0xB0, 0x43, + 0xE7, 0x81, 0xB7, 0x43, 0xE7, 0x81, 0xBD, 0x43, + 0xE7, 0x82, 0x99, 0x43, 0xE7, 0x82, 0xAD, 0x43, + 0xE7, 0x83, 0x88, 0x43, 0xE7, 0x83, 0x99, 0x43, + 0xE7, 0x84, 0xA1, 0x43, 0xE7, 0x85, 0x85, 0x43, + 0xE7, 0x85, 0x89, 0x43, 0xE7, 0x85, 0xAE, 0x43, + // Bytes ec0 - eff + 0xE7, 0x86, 0x9C, 0x43, 0xE7, 0x87, 0x8E, 0x43, + 0xE7, 0x87, 0x90, 0x43, 0xE7, 0x88, 0x90, 0x43, + 0xE7, 0x88, 0x9B, 0x43, 0xE7, 0x88, 0xA8, 0x43, + 0xE7, 0x88, 0xAA, 0x43, 0xE7, 0x88, 0xAB, 0x43, + 0xE7, 0x88, 0xB5, 0x43, 0xE7, 0x88, 0xB6, 0x43, + 0xE7, 0x88, 0xBB, 0x43, 0xE7, 0x88, 0xBF, 0x43, + 0xE7, 0x89, 0x87, 0x43, 0xE7, 0x89, 0x90, 0x43, + 0xE7, 0x89, 0x99, 0x43, 0xE7, 0x89, 0x9B, 0x43, + // Bytes f00 - f3f + 0xE7, 0x89, 0xA2, 0x43, 0xE7, 0x89, 0xB9, 0x43, + 0xE7, 0x8A, 0x80, 0x43, 0xE7, 0x8A, 0x95, 0x43, + 0xE7, 0x8A, 0xAC, 0x43, 0xE7, 0x8A, 0xAF, 0x43, + 0xE7, 0x8B, 0x80, 0x43, 0xE7, 0x8B, 0xBC, 0x43, + 0xE7, 0x8C, 0xAA, 0x43, 0xE7, 0x8D, 0xB5, 0x43, + 0xE7, 0x8D, 0xBA, 0x43, 0xE7, 0x8E, 0x84, 0x43, + 0xE7, 0x8E, 0x87, 0x43, 0xE7, 0x8E, 0x89, 0x43, + 0xE7, 0x8E, 0x8B, 0x43, 0xE7, 0x8E, 0xA5, 0x43, + // Bytes f40 - f7f + 0xE7, 0x8E, 0xB2, 0x43, 0xE7, 0x8F, 0x9E, 0x43, + 0xE7, 0x90, 0x86, 0x43, 0xE7, 0x90, 0x89, 0x43, + 0xE7, 0x90, 0xA2, 0x43, 0xE7, 0x91, 0x87, 0x43, + 0xE7, 0x91, 0x9C, 0x43, 0xE7, 0x91, 0xA9, 0x43, + 0xE7, 0x91, 0xB1, 0x43, 0xE7, 0x92, 0x85, 0x43, + 0xE7, 0x92, 0x89, 0x43, 0xE7, 0x92, 0x98, 0x43, + 0xE7, 0x93, 0x8A, 0x43, 0xE7, 0x93, 0x9C, 0x43, + 0xE7, 0x93, 0xA6, 0x43, 0xE7, 0x94, 0x86, 0x43, + // Bytes f80 - fbf + 0xE7, 0x94, 0x98, 0x43, 0xE7, 0x94, 0x9F, 0x43, + 0xE7, 0x94, 0xA4, 0x43, 0xE7, 0x94, 0xA8, 0x43, + 0xE7, 0x94, 0xB0, 0x43, 0xE7, 0x94, 0xB2, 0x43, + 0xE7, 0x94, 0xB3, 0x43, 0xE7, 0x94, 0xB7, 0x43, + 0xE7, 0x94, 0xBB, 0x43, 0xE7, 0x94, 0xBE, 0x43, + 0xE7, 0x95, 0x99, 0x43, 0xE7, 0x95, 0xA5, 0x43, + 0xE7, 0x95, 0xB0, 0x43, 0xE7, 0x96, 0x8B, 0x43, + 0xE7, 0x96, 0x92, 0x43, 0xE7, 0x97, 0xA2, 0x43, + // Bytes fc0 - fff + 0xE7, 0x98, 0x90, 0x43, 0xE7, 0x98, 0x9D, 0x43, + 0xE7, 0x98, 0x9F, 0x43, 0xE7, 0x99, 0x82, 0x43, + 0xE7, 0x99, 0xA9, 0x43, 0xE7, 0x99, 0xB6, 0x43, + 0xE7, 0x99, 0xBD, 0x43, 0xE7, 0x9A, 0xAE, 0x43, + 0xE7, 0x9A, 0xBF, 0x43, 0xE7, 0x9B, 0x8A, 0x43, + 0xE7, 0x9B, 0x9B, 0x43, 0xE7, 0x9B, 0xA3, 0x43, + 0xE7, 0x9B, 0xA7, 0x43, 0xE7, 0x9B, 0xAE, 0x43, + 0xE7, 0x9B, 0xB4, 0x43, 0xE7, 0x9C, 0x81, 0x43, + // Bytes 1000 - 103f + 0xE7, 0x9C, 0x9E, 0x43, 0xE7, 0x9C, 0x9F, 0x43, + 0xE7, 0x9D, 0x80, 0x43, 0xE7, 0x9D, 0x8A, 0x43, + 0xE7, 0x9E, 0x8B, 0x43, 0xE7, 0x9E, 0xA7, 0x43, + 0xE7, 0x9F, 0x9B, 0x43, 0xE7, 0x9F, 0xA2, 0x43, + 0xE7, 0x9F, 0xB3, 0x43, 0xE7, 0xA1, 0x8E, 0x43, + 0xE7, 0xA1, 0xAB, 0x43, 0xE7, 0xA2, 0x8C, 0x43, + 0xE7, 0xA2, 0x91, 0x43, 0xE7, 0xA3, 0x8A, 0x43, + 0xE7, 0xA3, 0x8C, 0x43, 0xE7, 0xA3, 0xBB, 0x43, + // Bytes 1040 - 107f + 0xE7, 0xA4, 0xAA, 0x43, 0xE7, 0xA4, 0xBA, 0x43, + 0xE7, 0xA4, 0xBC, 0x43, 0xE7, 0xA4, 0xBE, 0x43, + 0xE7, 0xA5, 0x88, 0x43, 0xE7, 0xA5, 0x89, 0x43, + 0xE7, 0xA5, 0x90, 0x43, 0xE7, 0xA5, 0x96, 0x43, + 0xE7, 0xA5, 0x9D, 0x43, 0xE7, 0xA5, 0x9E, 0x43, + 0xE7, 0xA5, 0xA5, 0x43, 0xE7, 0xA5, 0xBF, 0x43, + 0xE7, 0xA6, 0x81, 0x43, 0xE7, 0xA6, 0x8D, 0x43, + 0xE7, 0xA6, 0x8E, 0x43, 0xE7, 0xA6, 0x8F, 0x43, + // Bytes 1080 - 10bf + 0xE7, 0xA6, 0xAE, 0x43, 0xE7, 0xA6, 0xB8, 0x43, + 0xE7, 0xA6, 0xBE, 0x43, 0xE7, 0xA7, 0x8A, 0x43, + 0xE7, 0xA7, 0x98, 0x43, 0xE7, 0xA7, 0xAB, 0x43, + 0xE7, 0xA8, 0x9C, 0x43, 0xE7, 0xA9, 0x80, 0x43, + 0xE7, 0xA9, 0x8A, 0x43, 0xE7, 0xA9, 0x8F, 0x43, + 0xE7, 0xA9, 0xB4, 0x43, 0xE7, 0xA9, 0xBA, 0x43, + 0xE7, 0xAA, 0x81, 0x43, 0xE7, 0xAA, 0xB1, 0x43, + 0xE7, 0xAB, 0x8B, 0x43, 0xE7, 0xAB, 0xAE, 0x43, + // Bytes 10c0 - 10ff + 0xE7, 0xAB, 0xB9, 0x43, 0xE7, 0xAC, 0xA0, 0x43, + 0xE7, 0xAE, 0x8F, 0x43, 0xE7, 0xAF, 0x80, 0x43, + 0xE7, 0xAF, 0x86, 0x43, 0xE7, 0xAF, 0x89, 0x43, + 0xE7, 0xB0, 0xBE, 0x43, 0xE7, 0xB1, 0xA0, 0x43, + 0xE7, 0xB1, 0xB3, 0x43, 0xE7, 0xB1, 0xBB, 0x43, + 0xE7, 0xB2, 0x92, 0x43, 0xE7, 0xB2, 0xBE, 0x43, + 0xE7, 0xB3, 0x92, 0x43, 0xE7, 0xB3, 0x96, 0x43, + 0xE7, 0xB3, 0xA3, 0x43, 0xE7, 0xB3, 0xA7, 0x43, + // Bytes 1100 - 113f + 0xE7, 0xB3, 0xA8, 0x43, 0xE7, 0xB3, 0xB8, 0x43, + 0xE7, 0xB4, 0x80, 0x43, 0xE7, 0xB4, 0x90, 0x43, + 0xE7, 0xB4, 0xA2, 0x43, 0xE7, 0xB4, 0xAF, 0x43, + 0xE7, 0xB5, 0x82, 0x43, 0xE7, 0xB5, 0x9B, 0x43, + 0xE7, 0xB5, 0xA3, 0x43, 0xE7, 0xB6, 0xA0, 0x43, + 0xE7, 0xB6, 0xBE, 0x43, 0xE7, 0xB7, 0x87, 0x43, + 0xE7, 0xB7, 0xB4, 0x43, 0xE7, 0xB8, 0x82, 0x43, + 0xE7, 0xB8, 0x89, 0x43, 0xE7, 0xB8, 0xB7, 0x43, + // Bytes 1140 - 117f + 0xE7, 0xB9, 0x81, 0x43, 0xE7, 0xB9, 0x85, 0x43, + 0xE7, 0xBC, 0xB6, 0x43, 0xE7, 0xBC, 0xBE, 0x43, + 0xE7, 0xBD, 0x91, 0x43, 0xE7, 0xBD, 0xB2, 0x43, + 0xE7, 0xBD, 0xB9, 0x43, 0xE7, 0xBD, 0xBA, 0x43, + 0xE7, 0xBE, 0x85, 0x43, 0xE7, 0xBE, 0x8A, 0x43, + 0xE7, 0xBE, 0x95, 0x43, 0xE7, 0xBE, 0x9A, 0x43, + 0xE7, 0xBE, 0xBD, 0x43, 0xE7, 0xBF, 0xBA, 0x43, + 0xE8, 0x80, 0x81, 0x43, 0xE8, 0x80, 0x85, 0x43, + // Bytes 1180 - 11bf + 0xE8, 0x80, 0x8C, 0x43, 0xE8, 0x80, 0x92, 0x43, + 0xE8, 0x80, 0xB3, 0x43, 0xE8, 0x81, 0x86, 0x43, + 0xE8, 0x81, 0xA0, 0x43, 0xE8, 0x81, 0xAF, 0x43, + 0xE8, 0x81, 0xB0, 0x43, 0xE8, 0x81, 0xBE, 0x43, + 0xE8, 0x81, 0xBF, 0x43, 0xE8, 0x82, 0x89, 0x43, + 0xE8, 0x82, 0x8B, 0x43, 0xE8, 0x82, 0xAD, 0x43, + 0xE8, 0x82, 0xB2, 0x43, 0xE8, 0x84, 0x83, 0x43, + 0xE8, 0x84, 0xBE, 0x43, 0xE8, 0x87, 0x98, 0x43, + // Bytes 11c0 - 11ff + 0xE8, 0x87, 0xA3, 0x43, 0xE8, 0x87, 0xA8, 0x43, + 0xE8, 0x87, 0xAA, 0x43, 0xE8, 0x87, 0xAD, 0x43, + 0xE8, 0x87, 0xB3, 0x43, 0xE8, 0x87, 0xBC, 0x43, + 0xE8, 0x88, 0x81, 0x43, 0xE8, 0x88, 0x84, 0x43, + 0xE8, 0x88, 0x8C, 0x43, 0xE8, 0x88, 0x98, 0x43, + 0xE8, 0x88, 0x9B, 0x43, 0xE8, 0x88, 0x9F, 0x43, + 0xE8, 0x89, 0xAE, 0x43, 0xE8, 0x89, 0xAF, 0x43, + 0xE8, 0x89, 0xB2, 0x43, 0xE8, 0x89, 0xB8, 0x43, + // Bytes 1200 - 123f + 0xE8, 0x89, 0xB9, 0x43, 0xE8, 0x8A, 0x8B, 0x43, + 0xE8, 0x8A, 0x91, 0x43, 0xE8, 0x8A, 0x9D, 0x43, + 0xE8, 0x8A, 0xB1, 0x43, 0xE8, 0x8A, 0xB3, 0x43, + 0xE8, 0x8A, 0xBD, 0x43, 0xE8, 0x8B, 0xA5, 0x43, + 0xE8, 0x8B, 0xA6, 0x43, 0xE8, 0x8C, 0x9D, 0x43, + 0xE8, 0x8C, 0xA3, 0x43, 0xE8, 0x8C, 0xB6, 0x43, + 0xE8, 0x8D, 0x92, 0x43, 0xE8, 0x8D, 0x93, 0x43, + 0xE8, 0x8D, 0xA3, 0x43, 0xE8, 0x8E, 0xAD, 0x43, + // Bytes 1240 - 127f + 0xE8, 0x8E, 0xBD, 0x43, 0xE8, 0x8F, 0x89, 0x43, + 0xE8, 0x8F, 0x8A, 0x43, 0xE8, 0x8F, 0x8C, 0x43, + 0xE8, 0x8F, 0x9C, 0x43, 0xE8, 0x8F, 0xA7, 0x43, + 0xE8, 0x8F, 0xAF, 0x43, 0xE8, 0x8F, 0xB1, 0x43, + 0xE8, 0x90, 0xBD, 0x43, 0xE8, 0x91, 0x89, 0x43, + 0xE8, 0x91, 0x97, 0x43, 0xE8, 0x93, 0xAE, 0x43, + 0xE8, 0x93, 0xB1, 0x43, 0xE8, 0x93, 0xB3, 0x43, + 0xE8, 0x93, 0xBC, 0x43, 0xE8, 0x94, 0x96, 0x43, + // Bytes 1280 - 12bf + 0xE8, 0x95, 0xA4, 0x43, 0xE8, 0x97, 0x8D, 0x43, + 0xE8, 0x97, 0xBA, 0x43, 0xE8, 0x98, 0x86, 0x43, + 0xE8, 0x98, 0x92, 0x43, 0xE8, 0x98, 0xAD, 0x43, + 0xE8, 0x98, 0xBF, 0x43, 0xE8, 0x99, 0x8D, 0x43, + 0xE8, 0x99, 0x90, 0x43, 0xE8, 0x99, 0x9C, 0x43, + 0xE8, 0x99, 0xA7, 0x43, 0xE8, 0x99, 0xA9, 0x43, + 0xE8, 0x99, 0xAB, 0x43, 0xE8, 0x9A, 0x88, 0x43, + 0xE8, 0x9A, 0xA9, 0x43, 0xE8, 0x9B, 0xA2, 0x43, + // Bytes 12c0 - 12ff + 0xE8, 0x9C, 0x8E, 0x43, 0xE8, 0x9C, 0xA8, 0x43, + 0xE8, 0x9D, 0xAB, 0x43, 0xE8, 0x9D, 0xB9, 0x43, + 0xE8, 0x9E, 0x86, 0x43, 0xE8, 0x9E, 0xBA, 0x43, + 0xE8, 0x9F, 0xA1, 0x43, 0xE8, 0xA0, 0x81, 0x43, + 0xE8, 0xA0, 0x9F, 0x43, 0xE8, 0xA1, 0x80, 0x43, + 0xE8, 0xA1, 0x8C, 0x43, 0xE8, 0xA1, 0xA0, 0x43, + 0xE8, 0xA1, 0xA3, 0x43, 0xE8, 0xA3, 0x82, 0x43, + 0xE8, 0xA3, 0x8F, 0x43, 0xE8, 0xA3, 0x97, 0x43, + // Bytes 1300 - 133f + 0xE8, 0xA3, 0x9E, 0x43, 0xE8, 0xA3, 0xA1, 0x43, + 0xE8, 0xA3, 0xB8, 0x43, 0xE8, 0xA3, 0xBA, 0x43, + 0xE8, 0xA4, 0x90, 0x43, 0xE8, 0xA5, 0x81, 0x43, + 0xE8, 0xA5, 0xA4, 0x43, 0xE8, 0xA5, 0xBE, 0x43, + 0xE8, 0xA6, 0x86, 0x43, 0xE8, 0xA6, 0x8B, 0x43, + 0xE8, 0xA6, 0x96, 0x43, 0xE8, 0xA7, 0x92, 0x43, + 0xE8, 0xA7, 0xA3, 0x43, 0xE8, 0xA8, 0x80, 0x43, + 0xE8, 0xAA, 0xA0, 0x43, 0xE8, 0xAA, 0xAA, 0x43, + // Bytes 1340 - 137f + 0xE8, 0xAA, 0xBF, 0x43, 0xE8, 0xAB, 0x8B, 0x43, + 0xE8, 0xAB, 0x92, 0x43, 0xE8, 0xAB, 0x96, 0x43, + 0xE8, 0xAB, 0xAD, 0x43, 0xE8, 0xAB, 0xB8, 0x43, + 0xE8, 0xAB, 0xBE, 0x43, 0xE8, 0xAC, 0x81, 0x43, + 0xE8, 0xAC, 0xB9, 0x43, 0xE8, 0xAD, 0x98, 0x43, + 0xE8, 0xAE, 0x80, 0x43, 0xE8, 0xAE, 0x8A, 0x43, + 0xE8, 0xB0, 0xB7, 0x43, 0xE8, 0xB1, 0x86, 0x43, + 0xE8, 0xB1, 0x88, 0x43, 0xE8, 0xB1, 0x95, 0x43, + // Bytes 1380 - 13bf + 0xE8, 0xB1, 0xB8, 0x43, 0xE8, 0xB2, 0x9D, 0x43, + 0xE8, 0xB2, 0xA1, 0x43, 0xE8, 0xB2, 0xA9, 0x43, + 0xE8, 0xB2, 0xAB, 0x43, 0xE8, 0xB3, 0x81, 0x43, + 0xE8, 0xB3, 0x82, 0x43, 0xE8, 0xB3, 0x87, 0x43, + 0xE8, 0xB3, 0x88, 0x43, 0xE8, 0xB3, 0x93, 0x43, + 0xE8, 0xB4, 0x88, 0x43, 0xE8, 0xB4, 0x9B, 0x43, + 0xE8, 0xB5, 0xA4, 0x43, 0xE8, 0xB5, 0xB0, 0x43, + 0xE8, 0xB5, 0xB7, 0x43, 0xE8, 0xB6, 0xB3, 0x43, + // Bytes 13c0 - 13ff + 0xE8, 0xB6, 0xBC, 0x43, 0xE8, 0xB7, 0x8B, 0x43, + 0xE8, 0xB7, 0xAF, 0x43, 0xE8, 0xB7, 0xB0, 0x43, + 0xE8, 0xBA, 0xAB, 0x43, 0xE8, 0xBB, 0x8A, 0x43, + 0xE8, 0xBB, 0x94, 0x43, 0xE8, 0xBC, 0xA6, 0x43, + 0xE8, 0xBC, 0xAA, 0x43, 0xE8, 0xBC, 0xB8, 0x43, + 0xE8, 0xBC, 0xBB, 0x43, 0xE8, 0xBD, 0xA2, 0x43, + 0xE8, 0xBE, 0x9B, 0x43, 0xE8, 0xBE, 0x9E, 0x43, + 0xE8, 0xBE, 0xB0, 0x43, 0xE8, 0xBE, 0xB5, 0x43, + // Bytes 1400 - 143f + 0xE8, 0xBE, 0xB6, 0x43, 0xE9, 0x80, 0xA3, 0x43, + 0xE9, 0x80, 0xB8, 0x43, 0xE9, 0x81, 0x8A, 0x43, + 0xE9, 0x81, 0xA9, 0x43, 0xE9, 0x81, 0xB2, 0x43, + 0xE9, 0x81, 0xBC, 0x43, 0xE9, 0x82, 0x8F, 0x43, + 0xE9, 0x82, 0x91, 0x43, 0xE9, 0x82, 0x94, 0x43, + 0xE9, 0x83, 0x8E, 0x43, 0xE9, 0x83, 0x9E, 0x43, + 0xE9, 0x83, 0xB1, 0x43, 0xE9, 0x83, 0xBD, 0x43, + 0xE9, 0x84, 0x91, 0x43, 0xE9, 0x84, 0x9B, 0x43, + // Bytes 1440 - 147f + 0xE9, 0x85, 0x89, 0x43, 0xE9, 0x85, 0x8D, 0x43, + 0xE9, 0x85, 0xAA, 0x43, 0xE9, 0x86, 0x99, 0x43, + 0xE9, 0x86, 0xB4, 0x43, 0xE9, 0x87, 0x86, 0x43, + 0xE9, 0x87, 0x8C, 0x43, 0xE9, 0x87, 0x8F, 0x43, + 0xE9, 0x87, 0x91, 0x43, 0xE9, 0x88, 0xB4, 0x43, + 0xE9, 0x88, 0xB8, 0x43, 0xE9, 0x89, 0xB6, 0x43, + 0xE9, 0x89, 0xBC, 0x43, 0xE9, 0x8B, 0x97, 0x43, + 0xE9, 0x8B, 0x98, 0x43, 0xE9, 0x8C, 0x84, 0x43, + // Bytes 1480 - 14bf + 0xE9, 0x8D, 0x8A, 0x43, 0xE9, 0x8F, 0xB9, 0x43, + 0xE9, 0x90, 0x95, 0x43, 0xE9, 0x95, 0xB7, 0x43, + 0xE9, 0x96, 0x80, 0x43, 0xE9, 0x96, 0x8B, 0x43, + 0xE9, 0x96, 0xAD, 0x43, 0xE9, 0x96, 0xB7, 0x43, + 0xE9, 0x98, 0x9C, 0x43, 0xE9, 0x98, 0xAE, 0x43, + 0xE9, 0x99, 0x8B, 0x43, 0xE9, 0x99, 0x8D, 0x43, + 0xE9, 0x99, 0xB5, 0x43, 0xE9, 0x99, 0xB8, 0x43, + 0xE9, 0x99, 0xBC, 0x43, 0xE9, 0x9A, 0x86, 0x43, + // Bytes 14c0 - 14ff + 0xE9, 0x9A, 0xA3, 0x43, 0xE9, 0x9A, 0xB6, 0x43, + 0xE9, 0x9A, 0xB7, 0x43, 0xE9, 0x9A, 0xB8, 0x43, + 0xE9, 0x9A, 0xB9, 0x43, 0xE9, 0x9B, 0x83, 0x43, + 0xE9, 0x9B, 0xA2, 0x43, 0xE9, 0x9B, 0xA3, 0x43, + 0xE9, 0x9B, 0xA8, 0x43, 0xE9, 0x9B, 0xB6, 0x43, + 0xE9, 0x9B, 0xB7, 0x43, 0xE9, 0x9C, 0xA3, 0x43, + 0xE9, 0x9C, 0xB2, 0x43, 0xE9, 0x9D, 0x88, 0x43, + 0xE9, 0x9D, 0x91, 0x43, 0xE9, 0x9D, 0x96, 0x43, + // Bytes 1500 - 153f + 0xE9, 0x9D, 0x9E, 0x43, 0xE9, 0x9D, 0xA2, 0x43, + 0xE9, 0x9D, 0xA9, 0x43, 0xE9, 0x9F, 0x8B, 0x43, + 0xE9, 0x9F, 0x9B, 0x43, 0xE9, 0x9F, 0xA0, 0x43, + 0xE9, 0x9F, 0xAD, 0x43, 0xE9, 0x9F, 0xB3, 0x43, + 0xE9, 0x9F, 0xBF, 0x43, 0xE9, 0xA0, 0x81, 0x43, + 0xE9, 0xA0, 0x85, 0x43, 0xE9, 0xA0, 0x8B, 0x43, + 0xE9, 0xA0, 0x98, 0x43, 0xE9, 0xA0, 0xA9, 0x43, + 0xE9, 0xA0, 0xBB, 0x43, 0xE9, 0xA1, 0x9E, 0x43, + // Bytes 1540 - 157f + 0xE9, 0xA2, 0xA8, 0x43, 0xE9, 0xA3, 0x9B, 0x43, + 0xE9, 0xA3, 0x9F, 0x43, 0xE9, 0xA3, 0xA2, 0x43, + 0xE9, 0xA3, 0xAF, 0x43, 0xE9, 0xA3, 0xBC, 0x43, + 0xE9, 0xA4, 0xA8, 0x43, 0xE9, 0xA4, 0xA9, 0x43, + 0xE9, 0xA6, 0x96, 0x43, 0xE9, 0xA6, 0x99, 0x43, + 0xE9, 0xA6, 0xA7, 0x43, 0xE9, 0xA6, 0xAC, 0x43, + 0xE9, 0xA7, 0x82, 0x43, 0xE9, 0xA7, 0xB1, 0x43, + 0xE9, 0xA7, 0xBE, 0x43, 0xE9, 0xA9, 0xAA, 0x43, + // Bytes 1580 - 15bf + 0xE9, 0xAA, 0xA8, 0x43, 0xE9, 0xAB, 0x98, 0x43, + 0xE9, 0xAB, 0x9F, 0x43, 0xE9, 0xAC, 0x92, 0x43, + 0xE9, 0xAC, 0xA5, 0x43, 0xE9, 0xAC, 0xAF, 0x43, + 0xE9, 0xAC, 0xB2, 0x43, 0xE9, 0xAC, 0xBC, 0x43, + 0xE9, 0xAD, 0x9A, 0x43, 0xE9, 0xAD, 0xAF, 0x43, + 0xE9, 0xB1, 0x80, 0x43, 0xE9, 0xB1, 0x97, 0x43, + 0xE9, 0xB3, 0xA5, 0x43, 0xE9, 0xB3, 0xBD, 0x43, + 0xE9, 0xB5, 0xA7, 0x43, 0xE9, 0xB6, 0xB4, 0x43, + // Bytes 15c0 - 15ff + 0xE9, 0xB7, 0xBA, 0x43, 0xE9, 0xB8, 0x9E, 0x43, + 0xE9, 0xB9, 0xB5, 0x43, 0xE9, 0xB9, 0xBF, 0x43, + 0xE9, 0xBA, 0x97, 0x43, 0xE9, 0xBA, 0x9F, 0x43, + 0xE9, 0xBA, 0xA5, 0x43, 0xE9, 0xBA, 0xBB, 0x43, + 0xE9, 0xBB, 0x83, 0x43, 0xE9, 0xBB, 0x8D, 0x43, + 0xE9, 0xBB, 0x8E, 0x43, 0xE9, 0xBB, 0x91, 0x43, + 0xE9, 0xBB, 0xB9, 0x43, 0xE9, 0xBB, 0xBD, 0x43, + 0xE9, 0xBB, 0xBE, 0x43, 0xE9, 0xBC, 0x85, 0x43, + // Bytes 1600 - 163f + 0xE9, 0xBC, 0x8E, 0x43, 0xE9, 0xBC, 0x8F, 0x43, + 0xE9, 0xBC, 0x93, 0x43, 0xE9, 0xBC, 0x96, 0x43, + 0xE9, 0xBC, 0xA0, 0x43, 0xE9, 0xBC, 0xBB, 0x43, + 0xE9, 0xBD, 0x83, 0x43, 0xE9, 0xBD, 0x8A, 0x43, + 0xE9, 0xBD, 0x92, 0x43, 0xE9, 0xBE, 0x8D, 0x43, + 0xE9, 0xBE, 0x8E, 0x43, 0xE9, 0xBE, 0x9C, 0x43, + 0xE9, 0xBE, 0x9F, 0x43, 0xE9, 0xBE, 0xA0, 0x43, + 0xEA, 0x9C, 0xA7, 0x43, 0xEA, 0x9D, 0xAF, 0x43, + // Bytes 1640 - 167f + 0xEA, 0xAC, 0xB7, 0x43, 0xEA, 0xAD, 0x92, 0x44, + 0xF0, 0xA0, 0x84, 0xA2, 0x44, 0xF0, 0xA0, 0x94, + 0x9C, 0x44, 0xF0, 0xA0, 0x94, 0xA5, 0x44, 0xF0, + 0xA0, 0x95, 0x8B, 0x44, 0xF0, 0xA0, 0x98, 0xBA, + 0x44, 0xF0, 0xA0, 0xA0, 0x84, 0x44, 0xF0, 0xA0, + 0xA3, 0x9E, 0x44, 0xF0, 0xA0, 0xA8, 0xAC, 0x44, + 0xF0, 0xA0, 0xAD, 0xA3, 0x44, 0xF0, 0xA1, 0x93, + 0xA4, 0x44, 0xF0, 0xA1, 0x9A, 0xA8, 0x44, 0xF0, + // Bytes 1680 - 16bf + 0xA1, 0x9B, 0xAA, 0x44, 0xF0, 0xA1, 0xA7, 0x88, + 0x44, 0xF0, 0xA1, 0xAC, 0x98, 0x44, 0xF0, 0xA1, + 0xB4, 0x8B, 0x44, 0xF0, 0xA1, 0xB7, 0xA4, 0x44, + 0xF0, 0xA1, 0xB7, 0xA6, 0x44, 0xF0, 0xA2, 0x86, + 0x83, 0x44, 0xF0, 0xA2, 0x86, 0x9F, 0x44, 0xF0, + 0xA2, 0x8C, 0xB1, 0x44, 0xF0, 0xA2, 0x9B, 0x94, + 0x44, 0xF0, 0xA2, 0xA1, 0x84, 0x44, 0xF0, 0xA2, + 0xA1, 0x8A, 0x44, 0xF0, 0xA2, 0xAC, 0x8C, 0x44, + // Bytes 16c0 - 16ff + 0xF0, 0xA2, 0xAF, 0xB1, 0x44, 0xF0, 0xA3, 0x80, + 0x8A, 0x44, 0xF0, 0xA3, 0x8A, 0xB8, 0x44, 0xF0, + 0xA3, 0x8D, 0x9F, 0x44, 0xF0, 0xA3, 0x8E, 0x93, + 0x44, 0xF0, 0xA3, 0x8E, 0x9C, 0x44, 0xF0, 0xA3, + 0x8F, 0x83, 0x44, 0xF0, 0xA3, 0x8F, 0x95, 0x44, + 0xF0, 0xA3, 0x91, 0xAD, 0x44, 0xF0, 0xA3, 0x9A, + 0xA3, 0x44, 0xF0, 0xA3, 0xA2, 0xA7, 0x44, 0xF0, + 0xA3, 0xAA, 0x8D, 0x44, 0xF0, 0xA3, 0xAB, 0xBA, + // Bytes 1700 - 173f + 0x44, 0xF0, 0xA3, 0xB2, 0xBC, 0x44, 0xF0, 0xA3, + 0xB4, 0x9E, 0x44, 0xF0, 0xA3, 0xBB, 0x91, 0x44, + 0xF0, 0xA3, 0xBD, 0x9E, 0x44, 0xF0, 0xA3, 0xBE, + 0x8E, 0x44, 0xF0, 0xA4, 0x89, 0xA3, 0x44, 0xF0, + 0xA4, 0x8B, 0xAE, 0x44, 0xF0, 0xA4, 0x8E, 0xAB, + 0x44, 0xF0, 0xA4, 0x98, 0x88, 0x44, 0xF0, 0xA4, + 0x9C, 0xB5, 0x44, 0xF0, 0xA4, 0xA0, 0x94, 0x44, + 0xF0, 0xA4, 0xB0, 0xB6, 0x44, 0xF0, 0xA4, 0xB2, + // Bytes 1740 - 177f + 0x92, 0x44, 0xF0, 0xA4, 0xBE, 0xA1, 0x44, 0xF0, + 0xA4, 0xBE, 0xB8, 0x44, 0xF0, 0xA5, 0x81, 0x84, + 0x44, 0xF0, 0xA5, 0x83, 0xB2, 0x44, 0xF0, 0xA5, + 0x83, 0xB3, 0x44, 0xF0, 0xA5, 0x84, 0x99, 0x44, + 0xF0, 0xA5, 0x84, 0xB3, 0x44, 0xF0, 0xA5, 0x89, + 0x89, 0x44, 0xF0, 0xA5, 0x90, 0x9D, 0x44, 0xF0, + 0xA5, 0x98, 0xA6, 0x44, 0xF0, 0xA5, 0x9A, 0x9A, + 0x44, 0xF0, 0xA5, 0x9B, 0x85, 0x44, 0xF0, 0xA5, + // Bytes 1780 - 17bf + 0xA5, 0xBC, 0x44, 0xF0, 0xA5, 0xAA, 0xA7, 0x44, + 0xF0, 0xA5, 0xAE, 0xAB, 0x44, 0xF0, 0xA5, 0xB2, + 0x80, 0x44, 0xF0, 0xA5, 0xB3, 0x90, 0x44, 0xF0, + 0xA5, 0xBE, 0x86, 0x44, 0xF0, 0xA6, 0x87, 0x9A, + 0x44, 0xF0, 0xA6, 0x88, 0xA8, 0x44, 0xF0, 0xA6, + 0x89, 0x87, 0x44, 0xF0, 0xA6, 0x8B, 0x99, 0x44, + 0xF0, 0xA6, 0x8C, 0xBE, 0x44, 0xF0, 0xA6, 0x93, + 0x9A, 0x44, 0xF0, 0xA6, 0x94, 0xA3, 0x44, 0xF0, + // Bytes 17c0 - 17ff + 0xA6, 0x96, 0xA8, 0x44, 0xF0, 0xA6, 0x9E, 0xA7, + 0x44, 0xF0, 0xA6, 0x9E, 0xB5, 0x44, 0xF0, 0xA6, + 0xAC, 0xBC, 0x44, 0xF0, 0xA6, 0xB0, 0xB6, 0x44, + 0xF0, 0xA6, 0xB3, 0x95, 0x44, 0xF0, 0xA6, 0xB5, + 0xAB, 0x44, 0xF0, 0xA6, 0xBC, 0xAC, 0x44, 0xF0, + 0xA6, 0xBE, 0xB1, 0x44, 0xF0, 0xA7, 0x83, 0x92, + 0x44, 0xF0, 0xA7, 0x8F, 0x8A, 0x44, 0xF0, 0xA7, + 0x99, 0xA7, 0x44, 0xF0, 0xA7, 0xA2, 0xAE, 0x44, + // Bytes 1800 - 183f + 0xF0, 0xA7, 0xA5, 0xA6, 0x44, 0xF0, 0xA7, 0xB2, + 0xA8, 0x44, 0xF0, 0xA7, 0xBB, 0x93, 0x44, 0xF0, + 0xA7, 0xBC, 0xAF, 0x44, 0xF0, 0xA8, 0x97, 0x92, + 0x44, 0xF0, 0xA8, 0x97, 0xAD, 0x44, 0xF0, 0xA8, + 0x9C, 0xAE, 0x44, 0xF0, 0xA8, 0xAF, 0xBA, 0x44, + 0xF0, 0xA8, 0xB5, 0xB7, 0x44, 0xF0, 0xA9, 0x85, + 0x85, 0x44, 0xF0, 0xA9, 0x87, 0x9F, 0x44, 0xF0, + 0xA9, 0x88, 0x9A, 0x44, 0xF0, 0xA9, 0x90, 0x8A, + // Bytes 1840 - 187f + 0x44, 0xF0, 0xA9, 0x92, 0x96, 0x44, 0xF0, 0xA9, + 0x96, 0xB6, 0x44, 0xF0, 0xA9, 0xAC, 0xB0, 0x44, + 0xF0, 0xAA, 0x83, 0x8E, 0x44, 0xF0, 0xAA, 0x84, + 0x85, 0x44, 0xF0, 0xAA, 0x88, 0x8E, 0x44, 0xF0, + 0xAA, 0x8A, 0x91, 0x44, 0xF0, 0xAA, 0x8E, 0x92, + 0x44, 0xF0, 0xAA, 0x98, 0x80, 0x42, 0x21, 0x21, + 0x42, 0x21, 0x3F, 0x42, 0x2E, 0x2E, 0x42, 0x30, + 0x2C, 0x42, 0x30, 0x2E, 0x42, 0x31, 0x2C, 0x42, + // Bytes 1880 - 18bf + 0x31, 0x2E, 0x42, 0x31, 0x30, 0x42, 0x31, 0x31, + 0x42, 0x31, 0x32, 0x42, 0x31, 0x33, 0x42, 0x31, + 0x34, 0x42, 0x31, 0x35, 0x42, 0x31, 0x36, 0x42, + 0x31, 0x37, 0x42, 0x31, 0x38, 0x42, 0x31, 0x39, + 0x42, 0x32, 0x2C, 0x42, 0x32, 0x2E, 0x42, 0x32, + 0x30, 0x42, 0x32, 0x31, 0x42, 0x32, 0x32, 0x42, + 0x32, 0x33, 0x42, 0x32, 0x34, 0x42, 0x32, 0x35, + 0x42, 0x32, 0x36, 0x42, 0x32, 0x37, 0x42, 0x32, + // Bytes 18c0 - 18ff + 0x38, 0x42, 0x32, 0x39, 0x42, 0x33, 0x2C, 0x42, + 0x33, 0x2E, 0x42, 0x33, 0x30, 0x42, 0x33, 0x31, + 0x42, 0x33, 0x32, 0x42, 0x33, 0x33, 0x42, 0x33, + 0x34, 0x42, 0x33, 0x35, 0x42, 0x33, 0x36, 0x42, + 0x33, 0x37, 0x42, 0x33, 0x38, 0x42, 0x33, 0x39, + 0x42, 0x34, 0x2C, 0x42, 0x34, 0x2E, 0x42, 0x34, + 0x30, 0x42, 0x34, 0x31, 0x42, 0x34, 0x32, 0x42, + 0x34, 0x33, 0x42, 0x34, 0x34, 0x42, 0x34, 0x35, + // Bytes 1900 - 193f + 0x42, 0x34, 0x36, 0x42, 0x34, 0x37, 0x42, 0x34, + 0x38, 0x42, 0x34, 0x39, 0x42, 0x35, 0x2C, 0x42, + 0x35, 0x2E, 0x42, 0x35, 0x30, 0x42, 0x36, 0x2C, + 0x42, 0x36, 0x2E, 0x42, 0x37, 0x2C, 0x42, 0x37, + 0x2E, 0x42, 0x38, 0x2C, 0x42, 0x38, 0x2E, 0x42, + 0x39, 0x2C, 0x42, 0x39, 0x2E, 0x42, 0x3D, 0x3D, + 0x42, 0x3F, 0x21, 0x42, 0x3F, 0x3F, 0x42, 0x41, + 0x55, 0x42, 0x42, 0x71, 0x42, 0x43, 0x44, 0x42, + // Bytes 1940 - 197f + 0x44, 0x4A, 0x42, 0x44, 0x5A, 0x42, 0x44, 0x7A, + 0x42, 0x47, 0x42, 0x42, 0x47, 0x79, 0x42, 0x48, + 0x50, 0x42, 0x48, 0x56, 0x42, 0x48, 0x67, 0x42, + 0x48, 0x7A, 0x42, 0x49, 0x49, 0x42, 0x49, 0x4A, + 0x42, 0x49, 0x55, 0x42, 0x49, 0x56, 0x42, 0x49, + 0x58, 0x42, 0x4B, 0x42, 0x42, 0x4B, 0x4B, 0x42, + 0x4B, 0x4D, 0x42, 0x4C, 0x4A, 0x42, 0x4C, 0x6A, + 0x42, 0x4D, 0x42, 0x42, 0x4D, 0x43, 0x42, 0x4D, + // Bytes 1980 - 19bf + 0x44, 0x42, 0x4D, 0x52, 0x42, 0x4D, 0x56, 0x42, + 0x4D, 0x57, 0x42, 0x4E, 0x4A, 0x42, 0x4E, 0x6A, + 0x42, 0x4E, 0x6F, 0x42, 0x50, 0x48, 0x42, 0x50, + 0x52, 0x42, 0x50, 0x61, 0x42, 0x52, 0x73, 0x42, + 0x53, 0x44, 0x42, 0x53, 0x4D, 0x42, 0x53, 0x53, + 0x42, 0x53, 0x76, 0x42, 0x54, 0x4D, 0x42, 0x56, + 0x49, 0x42, 0x57, 0x43, 0x42, 0x57, 0x5A, 0x42, + 0x57, 0x62, 0x42, 0x58, 0x49, 0x42, 0x63, 0x63, + // Bytes 19c0 - 19ff + 0x42, 0x63, 0x64, 0x42, 0x63, 0x6D, 0x42, 0x64, + 0x42, 0x42, 0x64, 0x61, 0x42, 0x64, 0x6C, 0x42, + 0x64, 0x6D, 0x42, 0x64, 0x7A, 0x42, 0x65, 0x56, + 0x42, 0x66, 0x66, 0x42, 0x66, 0x69, 0x42, 0x66, + 0x6C, 0x42, 0x66, 0x6D, 0x42, 0x68, 0x61, 0x42, + 0x69, 0x69, 0x42, 0x69, 0x6A, 0x42, 0x69, 0x6E, + 0x42, 0x69, 0x76, 0x42, 0x69, 0x78, 0x42, 0x6B, + 0x41, 0x42, 0x6B, 0x56, 0x42, 0x6B, 0x57, 0x42, + // Bytes 1a00 - 1a3f + 0x6B, 0x67, 0x42, 0x6B, 0x6C, 0x42, 0x6B, 0x6D, + 0x42, 0x6B, 0x74, 0x42, 0x6C, 0x6A, 0x42, 0x6C, + 0x6D, 0x42, 0x6C, 0x6E, 0x42, 0x6C, 0x78, 0x42, + 0x6D, 0x32, 0x42, 0x6D, 0x33, 0x42, 0x6D, 0x41, + 0x42, 0x6D, 0x56, 0x42, 0x6D, 0x57, 0x42, 0x6D, + 0x62, 0x42, 0x6D, 0x67, 0x42, 0x6D, 0x6C, 0x42, + 0x6D, 0x6D, 0x42, 0x6D, 0x73, 0x42, 0x6E, 0x41, + 0x42, 0x6E, 0x46, 0x42, 0x6E, 0x56, 0x42, 0x6E, + // Bytes 1a40 - 1a7f + 0x57, 0x42, 0x6E, 0x6A, 0x42, 0x6E, 0x6D, 0x42, + 0x6E, 0x73, 0x42, 0x6F, 0x56, 0x42, 0x70, 0x41, + 0x42, 0x70, 0x46, 0x42, 0x70, 0x56, 0x42, 0x70, + 0x57, 0x42, 0x70, 0x63, 0x42, 0x70, 0x73, 0x42, + 0x73, 0x72, 0x42, 0x73, 0x74, 0x42, 0x76, 0x69, + 0x42, 0x78, 0x69, 0x43, 0x28, 0x31, 0x29, 0x43, + 0x28, 0x32, 0x29, 0x43, 0x28, 0x33, 0x29, 0x43, + 0x28, 0x34, 0x29, 0x43, 0x28, 0x35, 0x29, 0x43, + // Bytes 1a80 - 1abf + 0x28, 0x36, 0x29, 0x43, 0x28, 0x37, 0x29, 0x43, + 0x28, 0x38, 0x29, 0x43, 0x28, 0x39, 0x29, 0x43, + 0x28, 0x41, 0x29, 0x43, 0x28, 0x42, 0x29, 0x43, + 0x28, 0x43, 0x29, 0x43, 0x28, 0x44, 0x29, 0x43, + 0x28, 0x45, 0x29, 0x43, 0x28, 0x46, 0x29, 0x43, + 0x28, 0x47, 0x29, 0x43, 0x28, 0x48, 0x29, 0x43, + 0x28, 0x49, 0x29, 0x43, 0x28, 0x4A, 0x29, 0x43, + 0x28, 0x4B, 0x29, 0x43, 0x28, 0x4C, 0x29, 0x43, + // Bytes 1ac0 - 1aff + 0x28, 0x4D, 0x29, 0x43, 0x28, 0x4E, 0x29, 0x43, + 0x28, 0x4F, 0x29, 0x43, 0x28, 0x50, 0x29, 0x43, + 0x28, 0x51, 0x29, 0x43, 0x28, 0x52, 0x29, 0x43, + 0x28, 0x53, 0x29, 0x43, 0x28, 0x54, 0x29, 0x43, + 0x28, 0x55, 0x29, 0x43, 0x28, 0x56, 0x29, 0x43, + 0x28, 0x57, 0x29, 0x43, 0x28, 0x58, 0x29, 0x43, + 0x28, 0x59, 0x29, 0x43, 0x28, 0x5A, 0x29, 0x43, + 0x28, 0x61, 0x29, 0x43, 0x28, 0x62, 0x29, 0x43, + // Bytes 1b00 - 1b3f + 0x28, 0x63, 0x29, 0x43, 0x28, 0x64, 0x29, 0x43, + 0x28, 0x65, 0x29, 0x43, 0x28, 0x66, 0x29, 0x43, + 0x28, 0x67, 0x29, 0x43, 0x28, 0x68, 0x29, 0x43, + 0x28, 0x69, 0x29, 0x43, 0x28, 0x6A, 0x29, 0x43, + 0x28, 0x6B, 0x29, 0x43, 0x28, 0x6C, 0x29, 0x43, + 0x28, 0x6D, 0x29, 0x43, 0x28, 0x6E, 0x29, 0x43, + 0x28, 0x6F, 0x29, 0x43, 0x28, 0x70, 0x29, 0x43, + 0x28, 0x71, 0x29, 0x43, 0x28, 0x72, 0x29, 0x43, + // Bytes 1b40 - 1b7f + 0x28, 0x73, 0x29, 0x43, 0x28, 0x74, 0x29, 0x43, + 0x28, 0x75, 0x29, 0x43, 0x28, 0x76, 0x29, 0x43, + 0x28, 0x77, 0x29, 0x43, 0x28, 0x78, 0x29, 0x43, + 0x28, 0x79, 0x29, 0x43, 0x28, 0x7A, 0x29, 0x43, + 0x2E, 0x2E, 0x2E, 0x43, 0x31, 0x30, 0x2E, 0x43, + 0x31, 0x31, 0x2E, 0x43, 0x31, 0x32, 0x2E, 0x43, + 0x31, 0x33, 0x2E, 0x43, 0x31, 0x34, 0x2E, 0x43, + 0x31, 0x35, 0x2E, 0x43, 0x31, 0x36, 0x2E, 0x43, + // Bytes 1b80 - 1bbf + 0x31, 0x37, 0x2E, 0x43, 0x31, 0x38, 0x2E, 0x43, + 0x31, 0x39, 0x2E, 0x43, 0x32, 0x30, 0x2E, 0x43, + 0x3A, 0x3A, 0x3D, 0x43, 0x3D, 0x3D, 0x3D, 0x43, + 0x43, 0x6F, 0x2E, 0x43, 0x46, 0x41, 0x58, 0x43, + 0x47, 0x48, 0x7A, 0x43, 0x47, 0x50, 0x61, 0x43, + 0x49, 0x49, 0x49, 0x43, 0x4C, 0x54, 0x44, 0x43, + 0x4C, 0xC2, 0xB7, 0x43, 0x4D, 0x48, 0x7A, 0x43, + 0x4D, 0x50, 0x61, 0x43, 0x4D, 0xCE, 0xA9, 0x43, + // Bytes 1bc0 - 1bff + 0x50, 0x50, 0x4D, 0x43, 0x50, 0x50, 0x56, 0x43, + 0x50, 0x54, 0x45, 0x43, 0x54, 0x45, 0x4C, 0x43, + 0x54, 0x48, 0x7A, 0x43, 0x56, 0x49, 0x49, 0x43, + 0x58, 0x49, 0x49, 0x43, 0x61, 0x2F, 0x63, 0x43, + 0x61, 0x2F, 0x73, 0x43, 0x61, 0xCA, 0xBE, 0x43, + 0x62, 0x61, 0x72, 0x43, 0x63, 0x2F, 0x6F, 0x43, + 0x63, 0x2F, 0x75, 0x43, 0x63, 0x61, 0x6C, 0x43, + 0x63, 0x6D, 0x32, 0x43, 0x63, 0x6D, 0x33, 0x43, + // Bytes 1c00 - 1c3f + 0x64, 0x6D, 0x32, 0x43, 0x64, 0x6D, 0x33, 0x43, + 0x65, 0x72, 0x67, 0x43, 0x66, 0x66, 0x69, 0x43, + 0x66, 0x66, 0x6C, 0x43, 0x67, 0x61, 0x6C, 0x43, + 0x68, 0x50, 0x61, 0x43, 0x69, 0x69, 0x69, 0x43, + 0x6B, 0x48, 0x7A, 0x43, 0x6B, 0x50, 0x61, 0x43, + 0x6B, 0x6D, 0x32, 0x43, 0x6B, 0x6D, 0x33, 0x43, + 0x6B, 0xCE, 0xA9, 0x43, 0x6C, 0x6F, 0x67, 0x43, + 0x6C, 0xC2, 0xB7, 0x43, 0x6D, 0x69, 0x6C, 0x43, + // Bytes 1c40 - 1c7f + 0x6D, 0x6D, 0x32, 0x43, 0x6D, 0x6D, 0x33, 0x43, + 0x6D, 0x6F, 0x6C, 0x43, 0x72, 0x61, 0x64, 0x43, + 0x76, 0x69, 0x69, 0x43, 0x78, 0x69, 0x69, 0x43, + 0xC2, 0xB0, 0x43, 0x43, 0xC2, 0xB0, 0x46, 0x43, + 0xCA, 0xBC, 0x6E, 0x43, 0xCE, 0xBC, 0x41, 0x43, + 0xCE, 0xBC, 0x46, 0x43, 0xCE, 0xBC, 0x56, 0x43, + 0xCE, 0xBC, 0x57, 0x43, 0xCE, 0xBC, 0x67, 0x43, + 0xCE, 0xBC, 0x6C, 0x43, 0xCE, 0xBC, 0x6D, 0x43, + // Bytes 1c80 - 1cbf + 0xCE, 0xBC, 0x73, 0x44, 0x28, 0x31, 0x30, 0x29, + 0x44, 0x28, 0x31, 0x31, 0x29, 0x44, 0x28, 0x31, + 0x32, 0x29, 0x44, 0x28, 0x31, 0x33, 0x29, 0x44, + 0x28, 0x31, 0x34, 0x29, 0x44, 0x28, 0x31, 0x35, + 0x29, 0x44, 0x28, 0x31, 0x36, 0x29, 0x44, 0x28, + 0x31, 0x37, 0x29, 0x44, 0x28, 0x31, 0x38, 0x29, + 0x44, 0x28, 0x31, 0x39, 0x29, 0x44, 0x28, 0x32, + 0x30, 0x29, 0x44, 0x30, 0xE7, 0x82, 0xB9, 0x44, + // Bytes 1cc0 - 1cff + 0x31, 0xE2, 0x81, 0x84, 0x44, 0x31, 0xE6, 0x97, + 0xA5, 0x44, 0x31, 0xE6, 0x9C, 0x88, 0x44, 0x31, + 0xE7, 0x82, 0xB9, 0x44, 0x32, 0xE6, 0x97, 0xA5, + 0x44, 0x32, 0xE6, 0x9C, 0x88, 0x44, 0x32, 0xE7, + 0x82, 0xB9, 0x44, 0x33, 0xE6, 0x97, 0xA5, 0x44, + 0x33, 0xE6, 0x9C, 0x88, 0x44, 0x33, 0xE7, 0x82, + 0xB9, 0x44, 0x34, 0xE6, 0x97, 0xA5, 0x44, 0x34, + 0xE6, 0x9C, 0x88, 0x44, 0x34, 0xE7, 0x82, 0xB9, + // Bytes 1d00 - 1d3f + 0x44, 0x35, 0xE6, 0x97, 0xA5, 0x44, 0x35, 0xE6, + 0x9C, 0x88, 0x44, 0x35, 0xE7, 0x82, 0xB9, 0x44, + 0x36, 0xE6, 0x97, 0xA5, 0x44, 0x36, 0xE6, 0x9C, + 0x88, 0x44, 0x36, 0xE7, 0x82, 0xB9, 0x44, 0x37, + 0xE6, 0x97, 0xA5, 0x44, 0x37, 0xE6, 0x9C, 0x88, + 0x44, 0x37, 0xE7, 0x82, 0xB9, 0x44, 0x38, 0xE6, + 0x97, 0xA5, 0x44, 0x38, 0xE6, 0x9C, 0x88, 0x44, + 0x38, 0xE7, 0x82, 0xB9, 0x44, 0x39, 0xE6, 0x97, + // Bytes 1d40 - 1d7f + 0xA5, 0x44, 0x39, 0xE6, 0x9C, 0x88, 0x44, 0x39, + 0xE7, 0x82, 0xB9, 0x44, 0x56, 0x49, 0x49, 0x49, + 0x44, 0x61, 0x2E, 0x6D, 0x2E, 0x44, 0x6B, 0x63, + 0x61, 0x6C, 0x44, 0x70, 0x2E, 0x6D, 0x2E, 0x44, + 0x76, 0x69, 0x69, 0x69, 0x44, 0xD5, 0xA5, 0xD6, + 0x82, 0x44, 0xD5, 0xB4, 0xD5, 0xA5, 0x44, 0xD5, + 0xB4, 0xD5, 0xAB, 0x44, 0xD5, 0xB4, 0xD5, 0xAD, + 0x44, 0xD5, 0xB4, 0xD5, 0xB6, 0x44, 0xD5, 0xBE, + // Bytes 1d80 - 1dbf + 0xD5, 0xB6, 0x44, 0xD7, 0x90, 0xD7, 0x9C, 0x44, + 0xD8, 0xA7, 0xD9, 0xB4, 0x44, 0xD8, 0xA8, 0xD8, + 0xAC, 0x44, 0xD8, 0xA8, 0xD8, 0xAD, 0x44, 0xD8, + 0xA8, 0xD8, 0xAE, 0x44, 0xD8, 0xA8, 0xD8, 0xB1, + 0x44, 0xD8, 0xA8, 0xD8, 0xB2, 0x44, 0xD8, 0xA8, + 0xD9, 0x85, 0x44, 0xD8, 0xA8, 0xD9, 0x86, 0x44, + 0xD8, 0xA8, 0xD9, 0x87, 0x44, 0xD8, 0xA8, 0xD9, + 0x89, 0x44, 0xD8, 0xA8, 0xD9, 0x8A, 0x44, 0xD8, + // Bytes 1dc0 - 1dff + 0xAA, 0xD8, 0xAC, 0x44, 0xD8, 0xAA, 0xD8, 0xAD, + 0x44, 0xD8, 0xAA, 0xD8, 0xAE, 0x44, 0xD8, 0xAA, + 0xD8, 0xB1, 0x44, 0xD8, 0xAA, 0xD8, 0xB2, 0x44, + 0xD8, 0xAA, 0xD9, 0x85, 0x44, 0xD8, 0xAA, 0xD9, + 0x86, 0x44, 0xD8, 0xAA, 0xD9, 0x87, 0x44, 0xD8, + 0xAA, 0xD9, 0x89, 0x44, 0xD8, 0xAA, 0xD9, 0x8A, + 0x44, 0xD8, 0xAB, 0xD8, 0xAC, 0x44, 0xD8, 0xAB, + 0xD8, 0xB1, 0x44, 0xD8, 0xAB, 0xD8, 0xB2, 0x44, + // Bytes 1e00 - 1e3f + 0xD8, 0xAB, 0xD9, 0x85, 0x44, 0xD8, 0xAB, 0xD9, + 0x86, 0x44, 0xD8, 0xAB, 0xD9, 0x87, 0x44, 0xD8, + 0xAB, 0xD9, 0x89, 0x44, 0xD8, 0xAB, 0xD9, 0x8A, + 0x44, 0xD8, 0xAC, 0xD8, 0xAD, 0x44, 0xD8, 0xAC, + 0xD9, 0x85, 0x44, 0xD8, 0xAC, 0xD9, 0x89, 0x44, + 0xD8, 0xAC, 0xD9, 0x8A, 0x44, 0xD8, 0xAD, 0xD8, + 0xAC, 0x44, 0xD8, 0xAD, 0xD9, 0x85, 0x44, 0xD8, + 0xAD, 0xD9, 0x89, 0x44, 0xD8, 0xAD, 0xD9, 0x8A, + // Bytes 1e40 - 1e7f + 0x44, 0xD8, 0xAE, 0xD8, 0xAC, 0x44, 0xD8, 0xAE, + 0xD8, 0xAD, 0x44, 0xD8, 0xAE, 0xD9, 0x85, 0x44, + 0xD8, 0xAE, 0xD9, 0x89, 0x44, 0xD8, 0xAE, 0xD9, + 0x8A, 0x44, 0xD8, 0xB3, 0xD8, 0xAC, 0x44, 0xD8, + 0xB3, 0xD8, 0xAD, 0x44, 0xD8, 0xB3, 0xD8, 0xAE, + 0x44, 0xD8, 0xB3, 0xD8, 0xB1, 0x44, 0xD8, 0xB3, + 0xD9, 0x85, 0x44, 0xD8, 0xB3, 0xD9, 0x87, 0x44, + 0xD8, 0xB3, 0xD9, 0x89, 0x44, 0xD8, 0xB3, 0xD9, + // Bytes 1e80 - 1ebf + 0x8A, 0x44, 0xD8, 0xB4, 0xD8, 0xAC, 0x44, 0xD8, + 0xB4, 0xD8, 0xAD, 0x44, 0xD8, 0xB4, 0xD8, 0xAE, + 0x44, 0xD8, 0xB4, 0xD8, 0xB1, 0x44, 0xD8, 0xB4, + 0xD9, 0x85, 0x44, 0xD8, 0xB4, 0xD9, 0x87, 0x44, + 0xD8, 0xB4, 0xD9, 0x89, 0x44, 0xD8, 0xB4, 0xD9, + 0x8A, 0x44, 0xD8, 0xB5, 0xD8, 0xAD, 0x44, 0xD8, + 0xB5, 0xD8, 0xAE, 0x44, 0xD8, 0xB5, 0xD8, 0xB1, + 0x44, 0xD8, 0xB5, 0xD9, 0x85, 0x44, 0xD8, 0xB5, + // Bytes 1ec0 - 1eff + 0xD9, 0x89, 0x44, 0xD8, 0xB5, 0xD9, 0x8A, 0x44, + 0xD8, 0xB6, 0xD8, 0xAC, 0x44, 0xD8, 0xB6, 0xD8, + 0xAD, 0x44, 0xD8, 0xB6, 0xD8, 0xAE, 0x44, 0xD8, + 0xB6, 0xD8, 0xB1, 0x44, 0xD8, 0xB6, 0xD9, 0x85, + 0x44, 0xD8, 0xB6, 0xD9, 0x89, 0x44, 0xD8, 0xB6, + 0xD9, 0x8A, 0x44, 0xD8, 0xB7, 0xD8, 0xAD, 0x44, + 0xD8, 0xB7, 0xD9, 0x85, 0x44, 0xD8, 0xB7, 0xD9, + 0x89, 0x44, 0xD8, 0xB7, 0xD9, 0x8A, 0x44, 0xD8, + // Bytes 1f00 - 1f3f + 0xB8, 0xD9, 0x85, 0x44, 0xD8, 0xB9, 0xD8, 0xAC, + 0x44, 0xD8, 0xB9, 0xD9, 0x85, 0x44, 0xD8, 0xB9, + 0xD9, 0x89, 0x44, 0xD8, 0xB9, 0xD9, 0x8A, 0x44, + 0xD8, 0xBA, 0xD8, 0xAC, 0x44, 0xD8, 0xBA, 0xD9, + 0x85, 0x44, 0xD8, 0xBA, 0xD9, 0x89, 0x44, 0xD8, + 0xBA, 0xD9, 0x8A, 0x44, 0xD9, 0x81, 0xD8, 0xAC, + 0x44, 0xD9, 0x81, 0xD8, 0xAD, 0x44, 0xD9, 0x81, + 0xD8, 0xAE, 0x44, 0xD9, 0x81, 0xD9, 0x85, 0x44, + // Bytes 1f40 - 1f7f + 0xD9, 0x81, 0xD9, 0x89, 0x44, 0xD9, 0x81, 0xD9, + 0x8A, 0x44, 0xD9, 0x82, 0xD8, 0xAD, 0x44, 0xD9, + 0x82, 0xD9, 0x85, 0x44, 0xD9, 0x82, 0xD9, 0x89, + 0x44, 0xD9, 0x82, 0xD9, 0x8A, 0x44, 0xD9, 0x83, + 0xD8, 0xA7, 0x44, 0xD9, 0x83, 0xD8, 0xAC, 0x44, + 0xD9, 0x83, 0xD8, 0xAD, 0x44, 0xD9, 0x83, 0xD8, + 0xAE, 0x44, 0xD9, 0x83, 0xD9, 0x84, 0x44, 0xD9, + 0x83, 0xD9, 0x85, 0x44, 0xD9, 0x83, 0xD9, 0x89, + // Bytes 1f80 - 1fbf + 0x44, 0xD9, 0x83, 0xD9, 0x8A, 0x44, 0xD9, 0x84, + 0xD8, 0xA7, 0x44, 0xD9, 0x84, 0xD8, 0xAC, 0x44, + 0xD9, 0x84, 0xD8, 0xAD, 0x44, 0xD9, 0x84, 0xD8, + 0xAE, 0x44, 0xD9, 0x84, 0xD9, 0x85, 0x44, 0xD9, + 0x84, 0xD9, 0x87, 0x44, 0xD9, 0x84, 0xD9, 0x89, + 0x44, 0xD9, 0x84, 0xD9, 0x8A, 0x44, 0xD9, 0x85, + 0xD8, 0xA7, 0x44, 0xD9, 0x85, 0xD8, 0xAC, 0x44, + 0xD9, 0x85, 0xD8, 0xAD, 0x44, 0xD9, 0x85, 0xD8, + // Bytes 1fc0 - 1fff + 0xAE, 0x44, 0xD9, 0x85, 0xD9, 0x85, 0x44, 0xD9, + 0x85, 0xD9, 0x89, 0x44, 0xD9, 0x85, 0xD9, 0x8A, + 0x44, 0xD9, 0x86, 0xD8, 0xAC, 0x44, 0xD9, 0x86, + 0xD8, 0xAD, 0x44, 0xD9, 0x86, 0xD8, 0xAE, 0x44, + 0xD9, 0x86, 0xD8, 0xB1, 0x44, 0xD9, 0x86, 0xD8, + 0xB2, 0x44, 0xD9, 0x86, 0xD9, 0x85, 0x44, 0xD9, + 0x86, 0xD9, 0x86, 0x44, 0xD9, 0x86, 0xD9, 0x87, + 0x44, 0xD9, 0x86, 0xD9, 0x89, 0x44, 0xD9, 0x86, + // Bytes 2000 - 203f + 0xD9, 0x8A, 0x44, 0xD9, 0x87, 0xD8, 0xAC, 0x44, + 0xD9, 0x87, 0xD9, 0x85, 0x44, 0xD9, 0x87, 0xD9, + 0x89, 0x44, 0xD9, 0x87, 0xD9, 0x8A, 0x44, 0xD9, + 0x88, 0xD9, 0xB4, 0x44, 0xD9, 0x8A, 0xD8, 0xAC, + 0x44, 0xD9, 0x8A, 0xD8, 0xAD, 0x44, 0xD9, 0x8A, + 0xD8, 0xAE, 0x44, 0xD9, 0x8A, 0xD8, 0xB1, 0x44, + 0xD9, 0x8A, 0xD8, 0xB2, 0x44, 0xD9, 0x8A, 0xD9, + 0x85, 0x44, 0xD9, 0x8A, 0xD9, 0x86, 0x44, 0xD9, + // Bytes 2040 - 207f + 0x8A, 0xD9, 0x87, 0x44, 0xD9, 0x8A, 0xD9, 0x89, + 0x44, 0xD9, 0x8A, 0xD9, 0x8A, 0x44, 0xD9, 0x8A, + 0xD9, 0xB4, 0x44, 0xDB, 0x87, 0xD9, 0xB4, 0x45, + 0x28, 0xE1, 0x84, 0x80, 0x29, 0x45, 0x28, 0xE1, + 0x84, 0x82, 0x29, 0x45, 0x28, 0xE1, 0x84, 0x83, + 0x29, 0x45, 0x28, 0xE1, 0x84, 0x85, 0x29, 0x45, + 0x28, 0xE1, 0x84, 0x86, 0x29, 0x45, 0x28, 0xE1, + 0x84, 0x87, 0x29, 0x45, 0x28, 0xE1, 0x84, 0x89, + // Bytes 2080 - 20bf + 0x29, 0x45, 0x28, 0xE1, 0x84, 0x8B, 0x29, 0x45, + 0x28, 0xE1, 0x84, 0x8C, 0x29, 0x45, 0x28, 0xE1, + 0x84, 0x8E, 0x29, 0x45, 0x28, 0xE1, 0x84, 0x8F, + 0x29, 0x45, 0x28, 0xE1, 0x84, 0x90, 0x29, 0x45, + 0x28, 0xE1, 0x84, 0x91, 0x29, 0x45, 0x28, 0xE1, + 0x84, 0x92, 0x29, 0x45, 0x28, 0xE4, 0xB8, 0x80, + 0x29, 0x45, 0x28, 0xE4, 0xB8, 0x83, 0x29, 0x45, + 0x28, 0xE4, 0xB8, 0x89, 0x29, 0x45, 0x28, 0xE4, + // Bytes 20c0 - 20ff + 0xB9, 0x9D, 0x29, 0x45, 0x28, 0xE4, 0xBA, 0x8C, + 0x29, 0x45, 0x28, 0xE4, 0xBA, 0x94, 0x29, 0x45, + 0x28, 0xE4, 0xBB, 0xA3, 0x29, 0x45, 0x28, 0xE4, + 0xBC, 0x81, 0x29, 0x45, 0x28, 0xE4, 0xBC, 0x91, + 0x29, 0x45, 0x28, 0xE5, 0x85, 0xAB, 0x29, 0x45, + 0x28, 0xE5, 0x85, 0xAD, 0x29, 0x45, 0x28, 0xE5, + 0x8A, 0xB4, 0x29, 0x45, 0x28, 0xE5, 0x8D, 0x81, + 0x29, 0x45, 0x28, 0xE5, 0x8D, 0x94, 0x29, 0x45, + // Bytes 2100 - 213f + 0x28, 0xE5, 0x90, 0x8D, 0x29, 0x45, 0x28, 0xE5, + 0x91, 0xBC, 0x29, 0x45, 0x28, 0xE5, 0x9B, 0x9B, + 0x29, 0x45, 0x28, 0xE5, 0x9C, 0x9F, 0x29, 0x45, + 0x28, 0xE5, 0xAD, 0xA6, 0x29, 0x45, 0x28, 0xE6, + 0x97, 0xA5, 0x29, 0x45, 0x28, 0xE6, 0x9C, 0x88, + 0x29, 0x45, 0x28, 0xE6, 0x9C, 0x89, 0x29, 0x45, + 0x28, 0xE6, 0x9C, 0xA8, 0x29, 0x45, 0x28, 0xE6, + 0xA0, 0xAA, 0x29, 0x45, 0x28, 0xE6, 0xB0, 0xB4, + // Bytes 2140 - 217f + 0x29, 0x45, 0x28, 0xE7, 0x81, 0xAB, 0x29, 0x45, + 0x28, 0xE7, 0x89, 0xB9, 0x29, 0x45, 0x28, 0xE7, + 0x9B, 0xA3, 0x29, 0x45, 0x28, 0xE7, 0xA4, 0xBE, + 0x29, 0x45, 0x28, 0xE7, 0xA5, 0x9D, 0x29, 0x45, + 0x28, 0xE7, 0xA5, 0xAD, 0x29, 0x45, 0x28, 0xE8, + 0x87, 0xAA, 0x29, 0x45, 0x28, 0xE8, 0x87, 0xB3, + 0x29, 0x45, 0x28, 0xE8, 0xB2, 0xA1, 0x29, 0x45, + 0x28, 0xE8, 0xB3, 0x87, 0x29, 0x45, 0x28, 0xE9, + // Bytes 2180 - 21bf + 0x87, 0x91, 0x29, 0x45, 0x30, 0xE2, 0x81, 0x84, + 0x33, 0x45, 0x31, 0x30, 0xE6, 0x97, 0xA5, 0x45, + 0x31, 0x30, 0xE6, 0x9C, 0x88, 0x45, 0x31, 0x30, + 0xE7, 0x82, 0xB9, 0x45, 0x31, 0x31, 0xE6, 0x97, + 0xA5, 0x45, 0x31, 0x31, 0xE6, 0x9C, 0x88, 0x45, + 0x31, 0x31, 0xE7, 0x82, 0xB9, 0x45, 0x31, 0x32, + 0xE6, 0x97, 0xA5, 0x45, 0x31, 0x32, 0xE6, 0x9C, + 0x88, 0x45, 0x31, 0x32, 0xE7, 0x82, 0xB9, 0x45, + // Bytes 21c0 - 21ff + 0x31, 0x33, 0xE6, 0x97, 0xA5, 0x45, 0x31, 0x33, + 0xE7, 0x82, 0xB9, 0x45, 0x31, 0x34, 0xE6, 0x97, + 0xA5, 0x45, 0x31, 0x34, 0xE7, 0x82, 0xB9, 0x45, + 0x31, 0x35, 0xE6, 0x97, 0xA5, 0x45, 0x31, 0x35, + 0xE7, 0x82, 0xB9, 0x45, 0x31, 0x36, 0xE6, 0x97, + 0xA5, 0x45, 0x31, 0x36, 0xE7, 0x82, 0xB9, 0x45, + 0x31, 0x37, 0xE6, 0x97, 0xA5, 0x45, 0x31, 0x37, + 0xE7, 0x82, 0xB9, 0x45, 0x31, 0x38, 0xE6, 0x97, + // Bytes 2200 - 223f + 0xA5, 0x45, 0x31, 0x38, 0xE7, 0x82, 0xB9, 0x45, + 0x31, 0x39, 0xE6, 0x97, 0xA5, 0x45, 0x31, 0x39, + 0xE7, 0x82, 0xB9, 0x45, 0x31, 0xE2, 0x81, 0x84, + 0x32, 0x45, 0x31, 0xE2, 0x81, 0x84, 0x33, 0x45, + 0x31, 0xE2, 0x81, 0x84, 0x34, 0x45, 0x31, 0xE2, + 0x81, 0x84, 0x35, 0x45, 0x31, 0xE2, 0x81, 0x84, + 0x36, 0x45, 0x31, 0xE2, 0x81, 0x84, 0x37, 0x45, + 0x31, 0xE2, 0x81, 0x84, 0x38, 0x45, 0x31, 0xE2, + // Bytes 2240 - 227f + 0x81, 0x84, 0x39, 0x45, 0x32, 0x30, 0xE6, 0x97, + 0xA5, 0x45, 0x32, 0x30, 0xE7, 0x82, 0xB9, 0x45, + 0x32, 0x31, 0xE6, 0x97, 0xA5, 0x45, 0x32, 0x31, + 0xE7, 0x82, 0xB9, 0x45, 0x32, 0x32, 0xE6, 0x97, + 0xA5, 0x45, 0x32, 0x32, 0xE7, 0x82, 0xB9, 0x45, + 0x32, 0x33, 0xE6, 0x97, 0xA5, 0x45, 0x32, 0x33, + 0xE7, 0x82, 0xB9, 0x45, 0x32, 0x34, 0xE6, 0x97, + 0xA5, 0x45, 0x32, 0x34, 0xE7, 0x82, 0xB9, 0x45, + // Bytes 2280 - 22bf + 0x32, 0x35, 0xE6, 0x97, 0xA5, 0x45, 0x32, 0x36, + 0xE6, 0x97, 0xA5, 0x45, 0x32, 0x37, 0xE6, 0x97, + 0xA5, 0x45, 0x32, 0x38, 0xE6, 0x97, 0xA5, 0x45, + 0x32, 0x39, 0xE6, 0x97, 0xA5, 0x45, 0x32, 0xE2, + 0x81, 0x84, 0x33, 0x45, 0x32, 0xE2, 0x81, 0x84, + 0x35, 0x45, 0x33, 0x30, 0xE6, 0x97, 0xA5, 0x45, + 0x33, 0x31, 0xE6, 0x97, 0xA5, 0x45, 0x33, 0xE2, + 0x81, 0x84, 0x34, 0x45, 0x33, 0xE2, 0x81, 0x84, + // Bytes 22c0 - 22ff + 0x35, 0x45, 0x33, 0xE2, 0x81, 0x84, 0x38, 0x45, + 0x34, 0xE2, 0x81, 0x84, 0x35, 0x45, 0x35, 0xE2, + 0x81, 0x84, 0x36, 0x45, 0x35, 0xE2, 0x81, 0x84, + 0x38, 0x45, 0x37, 0xE2, 0x81, 0x84, 0x38, 0x45, + 0x41, 0xE2, 0x88, 0x95, 0x6D, 0x45, 0x56, 0xE2, + 0x88, 0x95, 0x6D, 0x45, 0x6D, 0xE2, 0x88, 0x95, + 0x73, 0x46, 0x31, 0xE2, 0x81, 0x84, 0x31, 0x30, + 0x46, 0x43, 0xE2, 0x88, 0x95, 0x6B, 0x67, 0x46, + // Bytes 2300 - 233f + 0x6D, 0xE2, 0x88, 0x95, 0x73, 0x32, 0x46, 0xD8, + 0xA8, 0xD8, 0xAD, 0xD9, 0x8A, 0x46, 0xD8, 0xA8, + 0xD8, 0xAE, 0xD9, 0x8A, 0x46, 0xD8, 0xAA, 0xD8, + 0xAC, 0xD9, 0x85, 0x46, 0xD8, 0xAA, 0xD8, 0xAC, + 0xD9, 0x89, 0x46, 0xD8, 0xAA, 0xD8, 0xAC, 0xD9, + 0x8A, 0x46, 0xD8, 0xAA, 0xD8, 0xAD, 0xD8, 0xAC, + 0x46, 0xD8, 0xAA, 0xD8, 0xAD, 0xD9, 0x85, 0x46, + 0xD8, 0xAA, 0xD8, 0xAE, 0xD9, 0x85, 0x46, 0xD8, + // Bytes 2340 - 237f + 0xAA, 0xD8, 0xAE, 0xD9, 0x89, 0x46, 0xD8, 0xAA, + 0xD8, 0xAE, 0xD9, 0x8A, 0x46, 0xD8, 0xAA, 0xD9, + 0x85, 0xD8, 0xAC, 0x46, 0xD8, 0xAA, 0xD9, 0x85, + 0xD8, 0xAD, 0x46, 0xD8, 0xAA, 0xD9, 0x85, 0xD8, + 0xAE, 0x46, 0xD8, 0xAA, 0xD9, 0x85, 0xD9, 0x89, + 0x46, 0xD8, 0xAA, 0xD9, 0x85, 0xD9, 0x8A, 0x46, + 0xD8, 0xAC, 0xD8, 0xAD, 0xD9, 0x89, 0x46, 0xD8, + 0xAC, 0xD8, 0xAD, 0xD9, 0x8A, 0x46, 0xD8, 0xAC, + // Bytes 2380 - 23bf + 0xD9, 0x85, 0xD8, 0xAD, 0x46, 0xD8, 0xAC, 0xD9, + 0x85, 0xD9, 0x89, 0x46, 0xD8, 0xAC, 0xD9, 0x85, + 0xD9, 0x8A, 0x46, 0xD8, 0xAD, 0xD8, 0xAC, 0xD9, + 0x8A, 0x46, 0xD8, 0xAD, 0xD9, 0x85, 0xD9, 0x89, + 0x46, 0xD8, 0xAD, 0xD9, 0x85, 0xD9, 0x8A, 0x46, + 0xD8, 0xB3, 0xD8, 0xAC, 0xD8, 0xAD, 0x46, 0xD8, + 0xB3, 0xD8, 0xAC, 0xD9, 0x89, 0x46, 0xD8, 0xB3, + 0xD8, 0xAD, 0xD8, 0xAC, 0x46, 0xD8, 0xB3, 0xD8, + // Bytes 23c0 - 23ff + 0xAE, 0xD9, 0x89, 0x46, 0xD8, 0xB3, 0xD8, 0xAE, + 0xD9, 0x8A, 0x46, 0xD8, 0xB3, 0xD9, 0x85, 0xD8, + 0xAC, 0x46, 0xD8, 0xB3, 0xD9, 0x85, 0xD8, 0xAD, + 0x46, 0xD8, 0xB3, 0xD9, 0x85, 0xD9, 0x85, 0x46, + 0xD8, 0xB4, 0xD8, 0xAC, 0xD9, 0x8A, 0x46, 0xD8, + 0xB4, 0xD8, 0xAD, 0xD9, 0x85, 0x46, 0xD8, 0xB4, + 0xD8, 0xAD, 0xD9, 0x8A, 0x46, 0xD8, 0xB4, 0xD9, + 0x85, 0xD8, 0xAE, 0x46, 0xD8, 0xB4, 0xD9, 0x85, + // Bytes 2400 - 243f + 0xD9, 0x85, 0x46, 0xD8, 0xB5, 0xD8, 0xAD, 0xD8, + 0xAD, 0x46, 0xD8, 0xB5, 0xD8, 0xAD, 0xD9, 0x8A, + 0x46, 0xD8, 0xB5, 0xD9, 0x84, 0xD9, 0x89, 0x46, + 0xD8, 0xB5, 0xD9, 0x84, 0xDB, 0x92, 0x46, 0xD8, + 0xB5, 0xD9, 0x85, 0xD9, 0x85, 0x46, 0xD8, 0xB6, + 0xD8, 0xAD, 0xD9, 0x89, 0x46, 0xD8, 0xB6, 0xD8, + 0xAD, 0xD9, 0x8A, 0x46, 0xD8, 0xB6, 0xD8, 0xAE, + 0xD9, 0x85, 0x46, 0xD8, 0xB7, 0xD9, 0x85, 0xD8, + // Bytes 2440 - 247f + 0xAD, 0x46, 0xD8, 0xB7, 0xD9, 0x85, 0xD9, 0x85, + 0x46, 0xD8, 0xB7, 0xD9, 0x85, 0xD9, 0x8A, 0x46, + 0xD8, 0xB9, 0xD8, 0xAC, 0xD9, 0x85, 0x46, 0xD8, + 0xB9, 0xD9, 0x85, 0xD9, 0x85, 0x46, 0xD8, 0xB9, + 0xD9, 0x85, 0xD9, 0x89, 0x46, 0xD8, 0xB9, 0xD9, + 0x85, 0xD9, 0x8A, 0x46, 0xD8, 0xBA, 0xD9, 0x85, + 0xD9, 0x85, 0x46, 0xD8, 0xBA, 0xD9, 0x85, 0xD9, + 0x89, 0x46, 0xD8, 0xBA, 0xD9, 0x85, 0xD9, 0x8A, + // Bytes 2480 - 24bf + 0x46, 0xD9, 0x81, 0xD8, 0xAE, 0xD9, 0x85, 0x46, + 0xD9, 0x81, 0xD9, 0x85, 0xD9, 0x8A, 0x46, 0xD9, + 0x82, 0xD9, 0x84, 0xDB, 0x92, 0x46, 0xD9, 0x82, + 0xD9, 0x85, 0xD8, 0xAD, 0x46, 0xD9, 0x82, 0xD9, + 0x85, 0xD9, 0x85, 0x46, 0xD9, 0x82, 0xD9, 0x85, + 0xD9, 0x8A, 0x46, 0xD9, 0x83, 0xD9, 0x85, 0xD9, + 0x85, 0x46, 0xD9, 0x83, 0xD9, 0x85, 0xD9, 0x8A, + 0x46, 0xD9, 0x84, 0xD8, 0xAC, 0xD8, 0xAC, 0x46, + // Bytes 24c0 - 24ff + 0xD9, 0x84, 0xD8, 0xAC, 0xD9, 0x85, 0x46, 0xD9, + 0x84, 0xD8, 0xAC, 0xD9, 0x8A, 0x46, 0xD9, 0x84, + 0xD8, 0xAD, 0xD9, 0x85, 0x46, 0xD9, 0x84, 0xD8, + 0xAD, 0xD9, 0x89, 0x46, 0xD9, 0x84, 0xD8, 0xAD, + 0xD9, 0x8A, 0x46, 0xD9, 0x84, 0xD8, 0xAE, 0xD9, + 0x85, 0x46, 0xD9, 0x84, 0xD9, 0x85, 0xD8, 0xAD, + 0x46, 0xD9, 0x84, 0xD9, 0x85, 0xD9, 0x8A, 0x46, + 0xD9, 0x85, 0xD8, 0xAC, 0xD8, 0xAD, 0x46, 0xD9, + // Bytes 2500 - 253f + 0x85, 0xD8, 0xAC, 0xD8, 0xAE, 0x46, 0xD9, 0x85, + 0xD8, 0xAC, 0xD9, 0x85, 0x46, 0xD9, 0x85, 0xD8, + 0xAC, 0xD9, 0x8A, 0x46, 0xD9, 0x85, 0xD8, 0xAD, + 0xD8, 0xAC, 0x46, 0xD9, 0x85, 0xD8, 0xAD, 0xD9, + 0x85, 0x46, 0xD9, 0x85, 0xD8, 0xAD, 0xD9, 0x8A, + 0x46, 0xD9, 0x85, 0xD8, 0xAE, 0xD8, 0xAC, 0x46, + 0xD9, 0x85, 0xD8, 0xAE, 0xD9, 0x85, 0x46, 0xD9, + 0x85, 0xD8, 0xAE, 0xD9, 0x8A, 0x46, 0xD9, 0x85, + // Bytes 2540 - 257f + 0xD9, 0x85, 0xD9, 0x8A, 0x46, 0xD9, 0x86, 0xD8, + 0xAC, 0xD8, 0xAD, 0x46, 0xD9, 0x86, 0xD8, 0xAC, + 0xD9, 0x85, 0x46, 0xD9, 0x86, 0xD8, 0xAC, 0xD9, + 0x89, 0x46, 0xD9, 0x86, 0xD8, 0xAC, 0xD9, 0x8A, + 0x46, 0xD9, 0x86, 0xD8, 0xAD, 0xD9, 0x85, 0x46, + 0xD9, 0x86, 0xD8, 0xAD, 0xD9, 0x89, 0x46, 0xD9, + 0x86, 0xD8, 0xAD, 0xD9, 0x8A, 0x46, 0xD9, 0x86, + 0xD9, 0x85, 0xD9, 0x89, 0x46, 0xD9, 0x86, 0xD9, + // Bytes 2580 - 25bf + 0x85, 0xD9, 0x8A, 0x46, 0xD9, 0x87, 0xD9, 0x85, + 0xD8, 0xAC, 0x46, 0xD9, 0x87, 0xD9, 0x85, 0xD9, + 0x85, 0x46, 0xD9, 0x8A, 0xD8, 0xAC, 0xD9, 0x8A, + 0x46, 0xD9, 0x8A, 0xD8, 0xAD, 0xD9, 0x8A, 0x46, + 0xD9, 0x8A, 0xD9, 0x85, 0xD9, 0x85, 0x46, 0xD9, + 0x8A, 0xD9, 0x85, 0xD9, 0x8A, 0x46, 0xD9, 0x8A, + 0xD9, 0x94, 0xD8, 0xA7, 0x46, 0xD9, 0x8A, 0xD9, + 0x94, 0xD8, 0xAC, 0x46, 0xD9, 0x8A, 0xD9, 0x94, + // Bytes 25c0 - 25ff + 0xD8, 0xAD, 0x46, 0xD9, 0x8A, 0xD9, 0x94, 0xD8, + 0xAE, 0x46, 0xD9, 0x8A, 0xD9, 0x94, 0xD8, 0xB1, + 0x46, 0xD9, 0x8A, 0xD9, 0x94, 0xD8, 0xB2, 0x46, + 0xD9, 0x8A, 0xD9, 0x94, 0xD9, 0x85, 0x46, 0xD9, + 0x8A, 0xD9, 0x94, 0xD9, 0x86, 0x46, 0xD9, 0x8A, + 0xD9, 0x94, 0xD9, 0x87, 0x46, 0xD9, 0x8A, 0xD9, + 0x94, 0xD9, 0x88, 0x46, 0xD9, 0x8A, 0xD9, 0x94, + 0xD9, 0x89, 0x46, 0xD9, 0x8A, 0xD9, 0x94, 0xD9, + // Bytes 2600 - 263f + 0x8A, 0x46, 0xD9, 0x8A, 0xD9, 0x94, 0xDB, 0x86, + 0x46, 0xD9, 0x8A, 0xD9, 0x94, 0xDB, 0x87, 0x46, + 0xD9, 0x8A, 0xD9, 0x94, 0xDB, 0x88, 0x46, 0xD9, + 0x8A, 0xD9, 0x94, 0xDB, 0x90, 0x46, 0xD9, 0x8A, + 0xD9, 0x94, 0xDB, 0x95, 0x46, 0xE0, 0xB9, 0x8D, + 0xE0, 0xB8, 0xB2, 0x46, 0xE0, 0xBA, 0xAB, 0xE0, + 0xBA, 0x99, 0x46, 0xE0, 0xBA, 0xAB, 0xE0, 0xBA, + 0xA1, 0x46, 0xE0, 0xBB, 0x8D, 0xE0, 0xBA, 0xB2, + // Bytes 2640 - 267f + 0x46, 0xE0, 0xBD, 0x80, 0xE0, 0xBE, 0xB5, 0x46, + 0xE0, 0xBD, 0x82, 0xE0, 0xBE, 0xB7, 0x46, 0xE0, + 0xBD, 0x8C, 0xE0, 0xBE, 0xB7, 0x46, 0xE0, 0xBD, + 0x91, 0xE0, 0xBE, 0xB7, 0x46, 0xE0, 0xBD, 0x96, + 0xE0, 0xBE, 0xB7, 0x46, 0xE0, 0xBD, 0x9B, 0xE0, + 0xBE, 0xB7, 0x46, 0xE0, 0xBE, 0x90, 0xE0, 0xBE, + 0xB5, 0x46, 0xE0, 0xBE, 0x92, 0xE0, 0xBE, 0xB7, + 0x46, 0xE0, 0xBE, 0x9C, 0xE0, 0xBE, 0xB7, 0x46, + // Bytes 2680 - 26bf + 0xE0, 0xBE, 0xA1, 0xE0, 0xBE, 0xB7, 0x46, 0xE0, + 0xBE, 0xA6, 0xE0, 0xBE, 0xB7, 0x46, 0xE0, 0xBE, + 0xAB, 0xE0, 0xBE, 0xB7, 0x46, 0xE2, 0x80, 0xB2, + 0xE2, 0x80, 0xB2, 0x46, 0xE2, 0x80, 0xB5, 0xE2, + 0x80, 0xB5, 0x46, 0xE2, 0x88, 0xAB, 0xE2, 0x88, + 0xAB, 0x46, 0xE2, 0x88, 0xAE, 0xE2, 0x88, 0xAE, + 0x46, 0xE3, 0x81, 0xBB, 0xE3, 0x81, 0x8B, 0x46, + 0xE3, 0x82, 0x88, 0xE3, 0x82, 0x8A, 0x46, 0xE3, + // Bytes 26c0 - 26ff + 0x82, 0xAD, 0xE3, 0x83, 0xAD, 0x46, 0xE3, 0x82, + 0xB3, 0xE3, 0x82, 0xB3, 0x46, 0xE3, 0x82, 0xB3, + 0xE3, 0x83, 0x88, 0x46, 0xE3, 0x83, 0x88, 0xE3, + 0x83, 0xB3, 0x46, 0xE3, 0x83, 0x8A, 0xE3, 0x83, + 0x8E, 0x46, 0xE3, 0x83, 0x9B, 0xE3, 0x83, 0xB3, + 0x46, 0xE3, 0x83, 0x9F, 0xE3, 0x83, 0xAA, 0x46, + 0xE3, 0x83, 0xAA, 0xE3, 0x83, 0xA9, 0x46, 0xE3, + 0x83, 0xAC, 0xE3, 0x83, 0xA0, 0x46, 0xE5, 0xA4, + // Bytes 2700 - 273f + 0xA7, 0xE6, 0xAD, 0xA3, 0x46, 0xE5, 0xB9, 0xB3, + 0xE6, 0x88, 0x90, 0x46, 0xE6, 0x98, 0x8E, 0xE6, + 0xB2, 0xBB, 0x46, 0xE6, 0x98, 0xAD, 0xE5, 0x92, + 0x8C, 0x47, 0x72, 0x61, 0x64, 0xE2, 0x88, 0x95, + 0x73, 0x47, 0xE3, 0x80, 0x94, 0x53, 0xE3, 0x80, + 0x95, 0x48, 0x28, 0xE1, 0x84, 0x80, 0xE1, 0x85, + 0xA1, 0x29, 0x48, 0x28, 0xE1, 0x84, 0x82, 0xE1, + 0x85, 0xA1, 0x29, 0x48, 0x28, 0xE1, 0x84, 0x83, + // Bytes 2740 - 277f + 0xE1, 0x85, 0xA1, 0x29, 0x48, 0x28, 0xE1, 0x84, + 0x85, 0xE1, 0x85, 0xA1, 0x29, 0x48, 0x28, 0xE1, + 0x84, 0x86, 0xE1, 0x85, 0xA1, 0x29, 0x48, 0x28, + 0xE1, 0x84, 0x87, 0xE1, 0x85, 0xA1, 0x29, 0x48, + 0x28, 0xE1, 0x84, 0x89, 0xE1, 0x85, 0xA1, 0x29, + 0x48, 0x28, 0xE1, 0x84, 0x8B, 0xE1, 0x85, 0xA1, + 0x29, 0x48, 0x28, 0xE1, 0x84, 0x8C, 0xE1, 0x85, + 0xA1, 0x29, 0x48, 0x28, 0xE1, 0x84, 0x8C, 0xE1, + // Bytes 2780 - 27bf + 0x85, 0xAE, 0x29, 0x48, 0x28, 0xE1, 0x84, 0x8E, + 0xE1, 0x85, 0xA1, 0x29, 0x48, 0x28, 0xE1, 0x84, + 0x8F, 0xE1, 0x85, 0xA1, 0x29, 0x48, 0x28, 0xE1, + 0x84, 0x90, 0xE1, 0x85, 0xA1, 0x29, 0x48, 0x28, + 0xE1, 0x84, 0x91, 0xE1, 0x85, 0xA1, 0x29, 0x48, + 0x28, 0xE1, 0x84, 0x92, 0xE1, 0x85, 0xA1, 0x29, + 0x48, 0x72, 0x61, 0x64, 0xE2, 0x88, 0x95, 0x73, + 0x32, 0x48, 0xD8, 0xA7, 0xD9, 0x83, 0xD8, 0xA8, + // Bytes 27c0 - 27ff + 0xD8, 0xB1, 0x48, 0xD8, 0xA7, 0xD9, 0x84, 0xD9, + 0x84, 0xD9, 0x87, 0x48, 0xD8, 0xB1, 0xD8, 0xB3, + 0xD9, 0x88, 0xD9, 0x84, 0x48, 0xD8, 0xB1, 0xDB, + 0x8C, 0xD8, 0xA7, 0xD9, 0x84, 0x48, 0xD8, 0xB5, + 0xD9, 0x84, 0xD8, 0xB9, 0xD9, 0x85, 0x48, 0xD8, + 0xB9, 0xD9, 0x84, 0xD9, 0x8A, 0xD9, 0x87, 0x48, + 0xD9, 0x85, 0xD8, 0xAD, 0xD9, 0x85, 0xD8, 0xAF, + 0x48, 0xD9, 0x88, 0xD8, 0xB3, 0xD9, 0x84, 0xD9, + // Bytes 2800 - 283f + 0x85, 0x49, 0xE2, 0x80, 0xB2, 0xE2, 0x80, 0xB2, + 0xE2, 0x80, 0xB2, 0x49, 0xE2, 0x80, 0xB5, 0xE2, + 0x80, 0xB5, 0xE2, 0x80, 0xB5, 0x49, 0xE2, 0x88, + 0xAB, 0xE2, 0x88, 0xAB, 0xE2, 0x88, 0xAB, 0x49, + 0xE2, 0x88, 0xAE, 0xE2, 0x88, 0xAE, 0xE2, 0x88, + 0xAE, 0x49, 0xE3, 0x80, 0x94, 0xE4, 0xB8, 0x89, + 0xE3, 0x80, 0x95, 0x49, 0xE3, 0x80, 0x94, 0xE4, + 0xBA, 0x8C, 0xE3, 0x80, 0x95, 0x49, 0xE3, 0x80, + // Bytes 2840 - 287f + 0x94, 0xE5, 0x8B, 0x9D, 0xE3, 0x80, 0x95, 0x49, + 0xE3, 0x80, 0x94, 0xE5, 0xAE, 0x89, 0xE3, 0x80, + 0x95, 0x49, 0xE3, 0x80, 0x94, 0xE6, 0x89, 0x93, + 0xE3, 0x80, 0x95, 0x49, 0xE3, 0x80, 0x94, 0xE6, + 0x95, 0x97, 0xE3, 0x80, 0x95, 0x49, 0xE3, 0x80, + 0x94, 0xE6, 0x9C, 0xAC, 0xE3, 0x80, 0x95, 0x49, + 0xE3, 0x80, 0x94, 0xE7, 0x82, 0xB9, 0xE3, 0x80, + 0x95, 0x49, 0xE3, 0x80, 0x94, 0xE7, 0x9B, 0x97, + // Bytes 2880 - 28bf + 0xE3, 0x80, 0x95, 0x49, 0xE3, 0x82, 0xA2, 0xE3, + 0x83, 0xBC, 0xE3, 0x83, 0xAB, 0x49, 0xE3, 0x82, + 0xA4, 0xE3, 0x83, 0xB3, 0xE3, 0x83, 0x81, 0x49, + 0xE3, 0x82, 0xA6, 0xE3, 0x82, 0xA9, 0xE3, 0x83, + 0xB3, 0x49, 0xE3, 0x82, 0xAA, 0xE3, 0x83, 0xB3, + 0xE3, 0x82, 0xB9, 0x49, 0xE3, 0x82, 0xAA, 0xE3, + 0x83, 0xBC, 0xE3, 0x83, 0xA0, 0x49, 0xE3, 0x82, + 0xAB, 0xE3, 0x82, 0xA4, 0xE3, 0x83, 0xAA, 0x49, + // Bytes 28c0 - 28ff + 0xE3, 0x82, 0xB1, 0xE3, 0x83, 0xBC, 0xE3, 0x82, + 0xB9, 0x49, 0xE3, 0x82, 0xB3, 0xE3, 0x83, 0xAB, + 0xE3, 0x83, 0x8A, 0x49, 0xE3, 0x82, 0xBB, 0xE3, + 0x83, 0xB3, 0xE3, 0x83, 0x81, 0x49, 0xE3, 0x82, + 0xBB, 0xE3, 0x83, 0xB3, 0xE3, 0x83, 0x88, 0x49, + 0xE3, 0x83, 0x86, 0xE3, 0x82, 0x99, 0xE3, 0x82, + 0xB7, 0x49, 0xE3, 0x83, 0x88, 0xE3, 0x82, 0x99, + 0xE3, 0x83, 0xAB, 0x49, 0xE3, 0x83, 0x8E, 0xE3, + // Bytes 2900 - 293f + 0x83, 0x83, 0xE3, 0x83, 0x88, 0x49, 0xE3, 0x83, + 0x8F, 0xE3, 0x82, 0xA4, 0xE3, 0x83, 0x84, 0x49, + 0xE3, 0x83, 0x92, 0xE3, 0x82, 0x99, 0xE3, 0x83, + 0xAB, 0x49, 0xE3, 0x83, 0x92, 0xE3, 0x82, 0x9A, + 0xE3, 0x82, 0xB3, 0x49, 0xE3, 0x83, 0x95, 0xE3, + 0x83, 0xA9, 0xE3, 0x83, 0xB3, 0x49, 0xE3, 0x83, + 0x98, 0xE3, 0x82, 0x9A, 0xE3, 0x82, 0xBD, 0x49, + 0xE3, 0x83, 0x98, 0xE3, 0x83, 0xAB, 0xE3, 0x83, + // Bytes 2940 - 297f + 0x84, 0x49, 0xE3, 0x83, 0x9B, 0xE3, 0x83, 0xBC, + 0xE3, 0x83, 0xAB, 0x49, 0xE3, 0x83, 0x9B, 0xE3, + 0x83, 0xBC, 0xE3, 0x83, 0xB3, 0x49, 0xE3, 0x83, + 0x9E, 0xE3, 0x82, 0xA4, 0xE3, 0x83, 0xAB, 0x49, + 0xE3, 0x83, 0x9E, 0xE3, 0x83, 0x83, 0xE3, 0x83, + 0x8F, 0x49, 0xE3, 0x83, 0x9E, 0xE3, 0x83, 0xAB, + 0xE3, 0x82, 0xAF, 0x49, 0xE3, 0x83, 0xA4, 0xE3, + 0x83, 0xBC, 0xE3, 0x83, 0xAB, 0x49, 0xE3, 0x83, + // Bytes 2980 - 29bf + 0xA6, 0xE3, 0x82, 0xA2, 0xE3, 0x83, 0xB3, 0x49, + 0xE3, 0x83, 0xAF, 0xE3, 0x83, 0x83, 0xE3, 0x83, + 0x88, 0x4C, 0xE2, 0x80, 0xB2, 0xE2, 0x80, 0xB2, + 0xE2, 0x80, 0xB2, 0xE2, 0x80, 0xB2, 0x4C, 0xE2, + 0x88, 0xAB, 0xE2, 0x88, 0xAB, 0xE2, 0x88, 0xAB, + 0xE2, 0x88, 0xAB, 0x4C, 0xE3, 0x82, 0xA2, 0xE3, + 0x83, 0xAB, 0xE3, 0x83, 0x95, 0xE3, 0x82, 0xA1, + 0x4C, 0xE3, 0x82, 0xA8, 0xE3, 0x83, 0xBC, 0xE3, + // Bytes 29c0 - 29ff + 0x82, 0xAB, 0xE3, 0x83, 0xBC, 0x4C, 0xE3, 0x82, + 0xAB, 0xE3, 0x82, 0x99, 0xE3, 0x83, 0xAD, 0xE3, + 0x83, 0xB3, 0x4C, 0xE3, 0x82, 0xAB, 0xE3, 0x82, + 0x99, 0xE3, 0x83, 0xB3, 0xE3, 0x83, 0x9E, 0x4C, + 0xE3, 0x82, 0xAB, 0xE3, 0x83, 0xA9, 0xE3, 0x83, + 0x83, 0xE3, 0x83, 0x88, 0x4C, 0xE3, 0x82, 0xAB, + 0xE3, 0x83, 0xAD, 0xE3, 0x83, 0xAA, 0xE3, 0x83, + 0xBC, 0x4C, 0xE3, 0x82, 0xAD, 0xE3, 0x82, 0x99, + // Bytes 2a00 - 2a3f + 0xE3, 0x83, 0x8B, 0xE3, 0x83, 0xBC, 0x4C, 0xE3, + 0x82, 0xAD, 0xE3, 0x83, 0xA5, 0xE3, 0x83, 0xAA, + 0xE3, 0x83, 0xBC, 0x4C, 0xE3, 0x82, 0xAF, 0xE3, + 0x82, 0x99, 0xE3, 0x83, 0xA9, 0xE3, 0x83, 0xA0, + 0x4C, 0xE3, 0x82, 0xAF, 0xE3, 0x83, 0xAD, 0xE3, + 0x83, 0xBC, 0xE3, 0x83, 0x8D, 0x4C, 0xE3, 0x82, + 0xB5, 0xE3, 0x82, 0xA4, 0xE3, 0x82, 0xAF, 0xE3, + 0x83, 0xAB, 0x4C, 0xE3, 0x82, 0xBF, 0xE3, 0x82, + // Bytes 2a40 - 2a7f + 0x99, 0xE3, 0x83, 0xBC, 0xE3, 0x82, 0xB9, 0x4C, + 0xE3, 0x83, 0x8F, 0xE3, 0x82, 0x9A, 0xE3, 0x83, + 0xBC, 0xE3, 0x83, 0x84, 0x4C, 0xE3, 0x83, 0x92, + 0xE3, 0x82, 0x9A, 0xE3, 0x82, 0xAF, 0xE3, 0x83, + 0xAB, 0x4C, 0xE3, 0x83, 0x95, 0xE3, 0x82, 0xA3, + 0xE3, 0x83, 0xBC, 0xE3, 0x83, 0x88, 0x4C, 0xE3, + 0x83, 0x98, 0xE3, 0x82, 0x99, 0xE3, 0x83, 0xBC, + 0xE3, 0x82, 0xBF, 0x4C, 0xE3, 0x83, 0x98, 0xE3, + // Bytes 2a80 - 2abf + 0x82, 0x9A, 0xE3, 0x83, 0x8B, 0xE3, 0x83, 0x92, + 0x4C, 0xE3, 0x83, 0x98, 0xE3, 0x82, 0x9A, 0xE3, + 0x83, 0xB3, 0xE3, 0x82, 0xB9, 0x4C, 0xE3, 0x83, + 0x9B, 0xE3, 0x82, 0x99, 0xE3, 0x83, 0xAB, 0xE3, + 0x83, 0x88, 0x4C, 0xE3, 0x83, 0x9E, 0xE3, 0x82, + 0xA4, 0xE3, 0x82, 0xAF, 0xE3, 0x83, 0xAD, 0x4C, + 0xE3, 0x83, 0x9F, 0xE3, 0x82, 0xAF, 0xE3, 0x83, + 0xAD, 0xE3, 0x83, 0xB3, 0x4C, 0xE3, 0x83, 0xA1, + // Bytes 2ac0 - 2aff + 0xE3, 0x83, 0xBC, 0xE3, 0x83, 0x88, 0xE3, 0x83, + 0xAB, 0x4C, 0xE3, 0x83, 0xAA, 0xE3, 0x83, 0x83, + 0xE3, 0x83, 0x88, 0xE3, 0x83, 0xAB, 0x4C, 0xE3, + 0x83, 0xAB, 0xE3, 0x83, 0x92, 0xE3, 0x82, 0x9A, + 0xE3, 0x83, 0xBC, 0x4C, 0xE6, 0xA0, 0xAA, 0xE5, + 0xBC, 0x8F, 0xE4, 0xBC, 0x9A, 0xE7, 0xA4, 0xBE, + 0x4E, 0x28, 0xE1, 0x84, 0x8B, 0xE1, 0x85, 0xA9, + 0xE1, 0x84, 0x92, 0xE1, 0x85, 0xAE, 0x29, 0x4F, + // Bytes 2b00 - 2b3f + 0xD8, 0xAC, 0xD9, 0x84, 0x20, 0xD8, 0xAC, 0xD9, + 0x84, 0xD8, 0xA7, 0xD9, 0x84, 0xD9, 0x87, 0x4F, + 0xE3, 0x82, 0xA2, 0xE3, 0x83, 0x8F, 0xE3, 0x82, + 0x9A, 0xE3, 0x83, 0xBC, 0xE3, 0x83, 0x88, 0x4F, + 0xE3, 0x82, 0xA2, 0xE3, 0x83, 0xB3, 0xE3, 0x83, + 0x98, 0xE3, 0x82, 0x9A, 0xE3, 0x82, 0xA2, 0x4F, + 0xE3, 0x82, 0xAD, 0xE3, 0x83, 0xAD, 0xE3, 0x83, + 0xAF, 0xE3, 0x83, 0x83, 0xE3, 0x83, 0x88, 0x4F, + // Bytes 2b40 - 2b7f + 0xE3, 0x82, 0xB5, 0xE3, 0x83, 0xB3, 0xE3, 0x83, + 0x81, 0xE3, 0x83, 0xBC, 0xE3, 0x83, 0xA0, 0x4F, + 0xE3, 0x83, 0x8F, 0xE3, 0x82, 0x99, 0xE3, 0x83, + 0xBC, 0xE3, 0x83, 0xAC, 0xE3, 0x83, 0xAB, 0x4F, + 0xE3, 0x83, 0x98, 0xE3, 0x82, 0xAF, 0xE3, 0x82, + 0xBF, 0xE3, 0x83, 0xBC, 0xE3, 0x83, 0xAB, 0x4F, + 0xE3, 0x83, 0x9B, 0xE3, 0x82, 0x9A, 0xE3, 0x82, + 0xA4, 0xE3, 0x83, 0xB3, 0xE3, 0x83, 0x88, 0x4F, + // Bytes 2b80 - 2bbf + 0xE3, 0x83, 0x9E, 0xE3, 0x83, 0xB3, 0xE3, 0x82, + 0xB7, 0xE3, 0x83, 0xA7, 0xE3, 0x83, 0xB3, 0x4F, + 0xE3, 0x83, 0xA1, 0xE3, 0x82, 0xAB, 0xE3, 0x82, + 0x99, 0xE3, 0x83, 0x88, 0xE3, 0x83, 0xB3, 0x4F, + 0xE3, 0x83, 0xAB, 0xE3, 0x83, 0xBC, 0xE3, 0x83, + 0x95, 0xE3, 0x82, 0x99, 0xE3, 0x83, 0xAB, 0x51, + 0x28, 0xE1, 0x84, 0x8B, 0xE1, 0x85, 0xA9, 0xE1, + 0x84, 0x8C, 0xE1, 0x85, 0xA5, 0xE1, 0x86, 0xAB, + // Bytes 2bc0 - 2bff + 0x29, 0x52, 0xE3, 0x82, 0xAD, 0xE3, 0x82, 0x99, + 0xE3, 0x83, 0xAB, 0xE3, 0x82, 0xBF, 0xE3, 0x82, + 0x99, 0xE3, 0x83, 0xBC, 0x52, 0xE3, 0x82, 0xAD, + 0xE3, 0x83, 0xAD, 0xE3, 0x82, 0xAF, 0xE3, 0x82, + 0x99, 0xE3, 0x83, 0xA9, 0xE3, 0x83, 0xA0, 0x52, + 0xE3, 0x82, 0xAD, 0xE3, 0x83, 0xAD, 0xE3, 0x83, + 0xA1, 0xE3, 0x83, 0xBC, 0xE3, 0x83, 0x88, 0xE3, + 0x83, 0xAB, 0x52, 0xE3, 0x82, 0xAF, 0xE3, 0x82, + // Bytes 2c00 - 2c3f + 0x99, 0xE3, 0x83, 0xA9, 0xE3, 0x83, 0xA0, 0xE3, + 0x83, 0x88, 0xE3, 0x83, 0xB3, 0x52, 0xE3, 0x82, + 0xAF, 0xE3, 0x83, 0xAB, 0xE3, 0x82, 0xBB, 0xE3, + 0x82, 0x99, 0xE3, 0x82, 0xA4, 0xE3, 0x83, 0xAD, + 0x52, 0xE3, 0x83, 0x8F, 0xE3, 0x82, 0x9A, 0xE3, + 0x83, 0xBC, 0xE3, 0x82, 0xBB, 0xE3, 0x83, 0xB3, + 0xE3, 0x83, 0x88, 0x52, 0xE3, 0x83, 0x92, 0xE3, + 0x82, 0x9A, 0xE3, 0x82, 0xA2, 0xE3, 0x82, 0xB9, + // Bytes 2c40 - 2c7f + 0xE3, 0x83, 0x88, 0xE3, 0x83, 0xAB, 0x52, 0xE3, + 0x83, 0x95, 0xE3, 0x82, 0x99, 0xE3, 0x83, 0x83, + 0xE3, 0x82, 0xB7, 0xE3, 0x82, 0xA7, 0xE3, 0x83, + 0xAB, 0x52, 0xE3, 0x83, 0x9F, 0xE3, 0x83, 0xAA, + 0xE3, 0x83, 0x8F, 0xE3, 0x82, 0x99, 0xE3, 0x83, + 0xBC, 0xE3, 0x83, 0xAB, 0x52, 0xE3, 0x83, 0xAC, + 0xE3, 0x83, 0xB3, 0xE3, 0x83, 0x88, 0xE3, 0x82, + 0xB1, 0xE3, 0x82, 0x99, 0xE3, 0x83, 0xB3, 0x61, + // Bytes 2c80 - 2cbf + 0xD8, 0xB5, 0xD9, 0x84, 0xD9, 0x89, 0x20, 0xD8, + 0xA7, 0xD9, 0x84, 0xD9, 0x84, 0xD9, 0x87, 0x20, + 0xD8, 0xB9, 0xD9, 0x84, 0xD9, 0x8A, 0xD9, 0x87, + 0x20, 0xD9, 0x88, 0xD8, 0xB3, 0xD9, 0x84, 0xD9, + 0x85, 0x06, 0xE0, 0xA7, 0x87, 0xE0, 0xA6, 0xBE, + 0x01, 0x06, 0xE0, 0xA7, 0x87, 0xE0, 0xA7, 0x97, + 0x01, 0x06, 0xE0, 0xAD, 0x87, 0xE0, 0xAC, 0xBE, + 0x01, 0x06, 0xE0, 0xAD, 0x87, 0xE0, 0xAD, 0x96, + // Bytes 2cc0 - 2cff + 0x01, 0x06, 0xE0, 0xAD, 0x87, 0xE0, 0xAD, 0x97, + 0x01, 0x06, 0xE0, 0xAE, 0x92, 0xE0, 0xAF, 0x97, + 0x01, 0x06, 0xE0, 0xAF, 0x86, 0xE0, 0xAE, 0xBE, + 0x01, 0x06, 0xE0, 0xAF, 0x86, 0xE0, 0xAF, 0x97, + 0x01, 0x06, 0xE0, 0xAF, 0x87, 0xE0, 0xAE, 0xBE, + 0x01, 0x06, 0xE0, 0xB2, 0xBF, 0xE0, 0xB3, 0x95, + 0x01, 0x06, 0xE0, 0xB3, 0x86, 0xE0, 0xB3, 0x95, + 0x01, 0x06, 0xE0, 0xB3, 0x86, 0xE0, 0xB3, 0x96, + // Bytes 2d00 - 2d3f + 0x01, 0x06, 0xE0, 0xB5, 0x86, 0xE0, 0xB4, 0xBE, + 0x01, 0x06, 0xE0, 0xB5, 0x86, 0xE0, 0xB5, 0x97, + 0x01, 0x06, 0xE0, 0xB5, 0x87, 0xE0, 0xB4, 0xBE, + 0x01, 0x06, 0xE0, 0xB7, 0x99, 0xE0, 0xB7, 0x9F, + 0x01, 0x06, 0xE1, 0x80, 0xA5, 0xE1, 0x80, 0xAE, + 0x01, 0x06, 0xE1, 0xAC, 0x85, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0x87, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0x89, 0xE1, 0xAC, 0xB5, + // Bytes 2d40 - 2d7f + 0x01, 0x06, 0xE1, 0xAC, 0x8B, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0x8D, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0x91, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0xBA, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0xBC, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0xBE, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAC, 0xBF, 0xE1, 0xAC, 0xB5, + 0x01, 0x06, 0xE1, 0xAD, 0x82, 0xE1, 0xAC, 0xB5, + // Bytes 2d80 - 2dbf + 0x01, 0x08, 0xF0, 0x91, 0x84, 0xB1, 0xF0, 0x91, + 0x84, 0xA7, 0x01, 0x08, 0xF0, 0x91, 0x84, 0xB2, + 0xF0, 0x91, 0x84, 0xA7, 0x01, 0x08, 0xF0, 0x91, + 0x8D, 0x87, 0xF0, 0x91, 0x8C, 0xBE, 0x01, 0x08, + 0xF0, 0x91, 0x8D, 0x87, 0xF0, 0x91, 0x8D, 0x97, + 0x01, 0x08, 0xF0, 0x91, 0x92, 0xB9, 0xF0, 0x91, + 0x92, 0xB0, 0x01, 0x08, 0xF0, 0x91, 0x92, 0xB9, + 0xF0, 0x91, 0x92, 0xBA, 0x01, 0x08, 0xF0, 0x91, + // Bytes 2dc0 - 2dff + 0x92, 0xB9, 0xF0, 0x91, 0x92, 0xBD, 0x01, 0x08, + 0xF0, 0x91, 0x96, 0xB8, 0xF0, 0x91, 0x96, 0xAF, + 0x01, 0x08, 0xF0, 0x91, 0x96, 0xB9, 0xF0, 0x91, + 0x96, 0xAF, 0x01, 0x09, 0xE0, 0xB3, 0x86, 0xE0, + 0xB3, 0x82, 0xE0, 0xB3, 0x95, 0x02, 0x09, 0xE0, + 0xB7, 0x99, 0xE0, 0xB7, 0x8F, 0xE0, 0xB7, 0x8A, + 0x12, 0x44, 0x44, 0x5A, 0xCC, 0x8C, 0xC9, 0x44, + 0x44, 0x7A, 0xCC, 0x8C, 0xC9, 0x44, 0x64, 0x7A, + // Bytes 2e00 - 2e3f + 0xCC, 0x8C, 0xC9, 0x46, 0xD9, 0x84, 0xD8, 0xA7, + 0xD9, 0x93, 0xC9, 0x46, 0xD9, 0x84, 0xD8, 0xA7, + 0xD9, 0x94, 0xC9, 0x46, 0xD9, 0x84, 0xD8, 0xA7, + 0xD9, 0x95, 0xB5, 0x46, 0xE1, 0x84, 0x80, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x82, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x83, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x85, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x86, 0xE1, + // Bytes 2e40 - 2e7f + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x87, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x89, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x8B, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x8B, 0xE1, + 0x85, 0xAE, 0x01, 0x46, 0xE1, 0x84, 0x8C, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x8E, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x8F, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x90, 0xE1, + // Bytes 2e80 - 2ebf + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x91, 0xE1, + 0x85, 0xA1, 0x01, 0x46, 0xE1, 0x84, 0x92, 0xE1, + 0x85, 0xA1, 0x01, 0x49, 0xE3, 0x83, 0xA1, 0xE3, + 0x82, 0xAB, 0xE3, 0x82, 0x99, 0x0D, 0x4C, 0xE1, + 0x84, 0x8C, 0xE1, 0x85, 0xAE, 0xE1, 0x84, 0x8B, + 0xE1, 0x85, 0xB4, 0x01, 0x4C, 0xE3, 0x82, 0xAD, + 0xE3, 0x82, 0x99, 0xE3, 0x82, 0xAB, 0xE3, 0x82, + 0x99, 0x0D, 0x4C, 0xE3, 0x82, 0xB3, 0xE3, 0x83, + // Bytes 2ec0 - 2eff + 0xBC, 0xE3, 0x83, 0x9B, 0xE3, 0x82, 0x9A, 0x0D, + 0x4C, 0xE3, 0x83, 0xA4, 0xE3, 0x83, 0xBC, 0xE3, + 0x83, 0x88, 0xE3, 0x82, 0x99, 0x0D, 0x4F, 0xE1, + 0x84, 0x8E, 0xE1, 0x85, 0xA1, 0xE1, 0x86, 0xB7, + 0xE1, 0x84, 0x80, 0xE1, 0x85, 0xA9, 0x01, 0x4F, + 0xE3, 0x82, 0xA4, 0xE3, 0x83, 0x8B, 0xE3, 0x83, + 0xB3, 0xE3, 0x82, 0xAF, 0xE3, 0x82, 0x99, 0x0D, + 0x4F, 0xE3, 0x82, 0xB7, 0xE3, 0x83, 0xAA, 0xE3, + // Bytes 2f00 - 2f3f + 0x83, 0xB3, 0xE3, 0x82, 0xAF, 0xE3, 0x82, 0x99, + 0x0D, 0x4F, 0xE3, 0x83, 0x98, 0xE3, 0x82, 0x9A, + 0xE3, 0x83, 0xBC, 0xE3, 0x82, 0xB7, 0xE3, 0x82, + 0x99, 0x0D, 0x4F, 0xE3, 0x83, 0x9B, 0xE3, 0x82, + 0x9A, 0xE3, 0x83, 0xB3, 0xE3, 0x83, 0x88, 0xE3, + 0x82, 0x99, 0x0D, 0x52, 0xE3, 0x82, 0xA8, 0xE3, + 0x82, 0xB9, 0xE3, 0x82, 0xAF, 0xE3, 0x83, 0xBC, + 0xE3, 0x83, 0x88, 0xE3, 0x82, 0x99, 0x0D, 0x52, + // Bytes 2f40 - 2f7f + 0xE3, 0x83, 0x95, 0xE3, 0x82, 0xA1, 0xE3, 0x83, + 0xA9, 0xE3, 0x83, 0x83, 0xE3, 0x83, 0x88, 0xE3, + 0x82, 0x99, 0x0D, 0x86, 0xE0, 0xB3, 0x86, 0xE0, + 0xB3, 0x82, 0x01, 0x86, 0xE0, 0xB7, 0x99, 0xE0, + 0xB7, 0x8F, 0x01, 0x03, 0x3C, 0xCC, 0xB8, 0x05, + 0x03, 0x3D, 0xCC, 0xB8, 0x05, 0x03, 0x3E, 0xCC, + 0xB8, 0x05, 0x03, 0x41, 0xCC, 0x80, 0xC9, 0x03, + 0x41, 0xCC, 0x81, 0xC9, 0x03, 0x41, 0xCC, 0x83, + // Bytes 2f80 - 2fbf + 0xC9, 0x03, 0x41, 0xCC, 0x84, 0xC9, 0x03, 0x41, + 0xCC, 0x89, 0xC9, 0x03, 0x41, 0xCC, 0x8C, 0xC9, + 0x03, 0x41, 0xCC, 0x8F, 0xC9, 0x03, 0x41, 0xCC, + 0x91, 0xC9, 0x03, 0x41, 0xCC, 0xA5, 0xB5, 0x03, + 0x41, 0xCC, 0xA8, 0xA5, 0x03, 0x42, 0xCC, 0x87, + 0xC9, 0x03, 0x42, 0xCC, 0xA3, 0xB5, 0x03, 0x42, + 0xCC, 0xB1, 0xB5, 0x03, 0x43, 0xCC, 0x81, 0xC9, + 0x03, 0x43, 0xCC, 0x82, 0xC9, 0x03, 0x43, 0xCC, + // Bytes 2fc0 - 2fff + 0x87, 0xC9, 0x03, 0x43, 0xCC, 0x8C, 0xC9, 0x03, + 0x44, 0xCC, 0x87, 0xC9, 0x03, 0x44, 0xCC, 0x8C, + 0xC9, 0x03, 0x44, 0xCC, 0xA3, 0xB5, 0x03, 0x44, + 0xCC, 0xA7, 0xA5, 0x03, 0x44, 0xCC, 0xAD, 0xB5, + 0x03, 0x44, 0xCC, 0xB1, 0xB5, 0x03, 0x45, 0xCC, + 0x80, 0xC9, 0x03, 0x45, 0xCC, 0x81, 0xC9, 0x03, + 0x45, 0xCC, 0x83, 0xC9, 0x03, 0x45, 0xCC, 0x86, + 0xC9, 0x03, 0x45, 0xCC, 0x87, 0xC9, 0x03, 0x45, + // Bytes 3000 - 303f + 0xCC, 0x88, 0xC9, 0x03, 0x45, 0xCC, 0x89, 0xC9, + 0x03, 0x45, 0xCC, 0x8C, 0xC9, 0x03, 0x45, 0xCC, + 0x8F, 0xC9, 0x03, 0x45, 0xCC, 0x91, 0xC9, 0x03, + 0x45, 0xCC, 0xA8, 0xA5, 0x03, 0x45, 0xCC, 0xAD, + 0xB5, 0x03, 0x45, 0xCC, 0xB0, 0xB5, 0x03, 0x46, + 0xCC, 0x87, 0xC9, 0x03, 0x47, 0xCC, 0x81, 0xC9, + 0x03, 0x47, 0xCC, 0x82, 0xC9, 0x03, 0x47, 0xCC, + 0x84, 0xC9, 0x03, 0x47, 0xCC, 0x86, 0xC9, 0x03, + // Bytes 3040 - 307f + 0x47, 0xCC, 0x87, 0xC9, 0x03, 0x47, 0xCC, 0x8C, + 0xC9, 0x03, 0x47, 0xCC, 0xA7, 0xA5, 0x03, 0x48, + 0xCC, 0x82, 0xC9, 0x03, 0x48, 0xCC, 0x87, 0xC9, + 0x03, 0x48, 0xCC, 0x88, 0xC9, 0x03, 0x48, 0xCC, + 0x8C, 0xC9, 0x03, 0x48, 0xCC, 0xA3, 0xB5, 0x03, + 0x48, 0xCC, 0xA7, 0xA5, 0x03, 0x48, 0xCC, 0xAE, + 0xB5, 0x03, 0x49, 0xCC, 0x80, 0xC9, 0x03, 0x49, + 0xCC, 0x81, 0xC9, 0x03, 0x49, 0xCC, 0x82, 0xC9, + // Bytes 3080 - 30bf + 0x03, 0x49, 0xCC, 0x83, 0xC9, 0x03, 0x49, 0xCC, + 0x84, 0xC9, 0x03, 0x49, 0xCC, 0x86, 0xC9, 0x03, + 0x49, 0xCC, 0x87, 0xC9, 0x03, 0x49, 0xCC, 0x89, + 0xC9, 0x03, 0x49, 0xCC, 0x8C, 0xC9, 0x03, 0x49, + 0xCC, 0x8F, 0xC9, 0x03, 0x49, 0xCC, 0x91, 0xC9, + 0x03, 0x49, 0xCC, 0xA3, 0xB5, 0x03, 0x49, 0xCC, + 0xA8, 0xA5, 0x03, 0x49, 0xCC, 0xB0, 0xB5, 0x03, + 0x4A, 0xCC, 0x82, 0xC9, 0x03, 0x4B, 0xCC, 0x81, + // Bytes 30c0 - 30ff + 0xC9, 0x03, 0x4B, 0xCC, 0x8C, 0xC9, 0x03, 0x4B, + 0xCC, 0xA3, 0xB5, 0x03, 0x4B, 0xCC, 0xA7, 0xA5, + 0x03, 0x4B, 0xCC, 0xB1, 0xB5, 0x03, 0x4C, 0xCC, + 0x81, 0xC9, 0x03, 0x4C, 0xCC, 0x8C, 0xC9, 0x03, + 0x4C, 0xCC, 0xA7, 0xA5, 0x03, 0x4C, 0xCC, 0xAD, + 0xB5, 0x03, 0x4C, 0xCC, 0xB1, 0xB5, 0x03, 0x4D, + 0xCC, 0x81, 0xC9, 0x03, 0x4D, 0xCC, 0x87, 0xC9, + 0x03, 0x4D, 0xCC, 0xA3, 0xB5, 0x03, 0x4E, 0xCC, + // Bytes 3100 - 313f + 0x80, 0xC9, 0x03, 0x4E, 0xCC, 0x81, 0xC9, 0x03, + 0x4E, 0xCC, 0x83, 0xC9, 0x03, 0x4E, 0xCC, 0x87, + 0xC9, 0x03, 0x4E, 0xCC, 0x8C, 0xC9, 0x03, 0x4E, + 0xCC, 0xA3, 0xB5, 0x03, 0x4E, 0xCC, 0xA7, 0xA5, + 0x03, 0x4E, 0xCC, 0xAD, 0xB5, 0x03, 0x4E, 0xCC, + 0xB1, 0xB5, 0x03, 0x4F, 0xCC, 0x80, 0xC9, 0x03, + 0x4F, 0xCC, 0x81, 0xC9, 0x03, 0x4F, 0xCC, 0x86, + 0xC9, 0x03, 0x4F, 0xCC, 0x89, 0xC9, 0x03, 0x4F, + // Bytes 3140 - 317f + 0xCC, 0x8B, 0xC9, 0x03, 0x4F, 0xCC, 0x8C, 0xC9, + 0x03, 0x4F, 0xCC, 0x8F, 0xC9, 0x03, 0x4F, 0xCC, + 0x91, 0xC9, 0x03, 0x50, 0xCC, 0x81, 0xC9, 0x03, + 0x50, 0xCC, 0x87, 0xC9, 0x03, 0x52, 0xCC, 0x81, + 0xC9, 0x03, 0x52, 0xCC, 0x87, 0xC9, 0x03, 0x52, + 0xCC, 0x8C, 0xC9, 0x03, 0x52, 0xCC, 0x8F, 0xC9, + 0x03, 0x52, 0xCC, 0x91, 0xC9, 0x03, 0x52, 0xCC, + 0xA7, 0xA5, 0x03, 0x52, 0xCC, 0xB1, 0xB5, 0x03, + // Bytes 3180 - 31bf + 0x53, 0xCC, 0x82, 0xC9, 0x03, 0x53, 0xCC, 0x87, + 0xC9, 0x03, 0x53, 0xCC, 0xA6, 0xB5, 0x03, 0x53, + 0xCC, 0xA7, 0xA5, 0x03, 0x54, 0xCC, 0x87, 0xC9, + 0x03, 0x54, 0xCC, 0x8C, 0xC9, 0x03, 0x54, 0xCC, + 0xA3, 0xB5, 0x03, 0x54, 0xCC, 0xA6, 0xB5, 0x03, + 0x54, 0xCC, 0xA7, 0xA5, 0x03, 0x54, 0xCC, 0xAD, + 0xB5, 0x03, 0x54, 0xCC, 0xB1, 0xB5, 0x03, 0x55, + 0xCC, 0x80, 0xC9, 0x03, 0x55, 0xCC, 0x81, 0xC9, + // Bytes 31c0 - 31ff + 0x03, 0x55, 0xCC, 0x82, 0xC9, 0x03, 0x55, 0xCC, + 0x86, 0xC9, 0x03, 0x55, 0xCC, 0x89, 0xC9, 0x03, + 0x55, 0xCC, 0x8A, 0xC9, 0x03, 0x55, 0xCC, 0x8B, + 0xC9, 0x03, 0x55, 0xCC, 0x8C, 0xC9, 0x03, 0x55, + 0xCC, 0x8F, 0xC9, 0x03, 0x55, 0xCC, 0x91, 0xC9, + 0x03, 0x55, 0xCC, 0xA3, 0xB5, 0x03, 0x55, 0xCC, + 0xA4, 0xB5, 0x03, 0x55, 0xCC, 0xA8, 0xA5, 0x03, + 0x55, 0xCC, 0xAD, 0xB5, 0x03, 0x55, 0xCC, 0xB0, + // Bytes 3200 - 323f + 0xB5, 0x03, 0x56, 0xCC, 0x83, 0xC9, 0x03, 0x56, + 0xCC, 0xA3, 0xB5, 0x03, 0x57, 0xCC, 0x80, 0xC9, + 0x03, 0x57, 0xCC, 0x81, 0xC9, 0x03, 0x57, 0xCC, + 0x82, 0xC9, 0x03, 0x57, 0xCC, 0x87, 0xC9, 0x03, + 0x57, 0xCC, 0x88, 0xC9, 0x03, 0x57, 0xCC, 0xA3, + 0xB5, 0x03, 0x58, 0xCC, 0x87, 0xC9, 0x03, 0x58, + 0xCC, 0x88, 0xC9, 0x03, 0x59, 0xCC, 0x80, 0xC9, + 0x03, 0x59, 0xCC, 0x81, 0xC9, 0x03, 0x59, 0xCC, + // Bytes 3240 - 327f + 0x82, 0xC9, 0x03, 0x59, 0xCC, 0x83, 0xC9, 0x03, + 0x59, 0xCC, 0x84, 0xC9, 0x03, 0x59, 0xCC, 0x87, + 0xC9, 0x03, 0x59, 0xCC, 0x88, 0xC9, 0x03, 0x59, + 0xCC, 0x89, 0xC9, 0x03, 0x59, 0xCC, 0xA3, 0xB5, + 0x03, 0x5A, 0xCC, 0x81, 0xC9, 0x03, 0x5A, 0xCC, + 0x82, 0xC9, 0x03, 0x5A, 0xCC, 0x87, 0xC9, 0x03, + 0x5A, 0xCC, 0x8C, 0xC9, 0x03, 0x5A, 0xCC, 0xA3, + 0xB5, 0x03, 0x5A, 0xCC, 0xB1, 0xB5, 0x03, 0x61, + // Bytes 3280 - 32bf + 0xCC, 0x80, 0xC9, 0x03, 0x61, 0xCC, 0x81, 0xC9, + 0x03, 0x61, 0xCC, 0x83, 0xC9, 0x03, 0x61, 0xCC, + 0x84, 0xC9, 0x03, 0x61, 0xCC, 0x89, 0xC9, 0x03, + 0x61, 0xCC, 0x8C, 0xC9, 0x03, 0x61, 0xCC, 0x8F, + 0xC9, 0x03, 0x61, 0xCC, 0x91, 0xC9, 0x03, 0x61, + 0xCC, 0xA5, 0xB5, 0x03, 0x61, 0xCC, 0xA8, 0xA5, + 0x03, 0x62, 0xCC, 0x87, 0xC9, 0x03, 0x62, 0xCC, + 0xA3, 0xB5, 0x03, 0x62, 0xCC, 0xB1, 0xB5, 0x03, + // Bytes 32c0 - 32ff + 0x63, 0xCC, 0x81, 0xC9, 0x03, 0x63, 0xCC, 0x82, + 0xC9, 0x03, 0x63, 0xCC, 0x87, 0xC9, 0x03, 0x63, + 0xCC, 0x8C, 0xC9, 0x03, 0x64, 0xCC, 0x87, 0xC9, + 0x03, 0x64, 0xCC, 0x8C, 0xC9, 0x03, 0x64, 0xCC, + 0xA3, 0xB5, 0x03, 0x64, 0xCC, 0xA7, 0xA5, 0x03, + 0x64, 0xCC, 0xAD, 0xB5, 0x03, 0x64, 0xCC, 0xB1, + 0xB5, 0x03, 0x65, 0xCC, 0x80, 0xC9, 0x03, 0x65, + 0xCC, 0x81, 0xC9, 0x03, 0x65, 0xCC, 0x83, 0xC9, + // Bytes 3300 - 333f + 0x03, 0x65, 0xCC, 0x86, 0xC9, 0x03, 0x65, 0xCC, + 0x87, 0xC9, 0x03, 0x65, 0xCC, 0x88, 0xC9, 0x03, + 0x65, 0xCC, 0x89, 0xC9, 0x03, 0x65, 0xCC, 0x8C, + 0xC9, 0x03, 0x65, 0xCC, 0x8F, 0xC9, 0x03, 0x65, + 0xCC, 0x91, 0xC9, 0x03, 0x65, 0xCC, 0xA8, 0xA5, + 0x03, 0x65, 0xCC, 0xAD, 0xB5, 0x03, 0x65, 0xCC, + 0xB0, 0xB5, 0x03, 0x66, 0xCC, 0x87, 0xC9, 0x03, + 0x67, 0xCC, 0x81, 0xC9, 0x03, 0x67, 0xCC, 0x82, + // Bytes 3340 - 337f + 0xC9, 0x03, 0x67, 0xCC, 0x84, 0xC9, 0x03, 0x67, + 0xCC, 0x86, 0xC9, 0x03, 0x67, 0xCC, 0x87, 0xC9, + 0x03, 0x67, 0xCC, 0x8C, 0xC9, 0x03, 0x67, 0xCC, + 0xA7, 0xA5, 0x03, 0x68, 0xCC, 0x82, 0xC9, 0x03, + 0x68, 0xCC, 0x87, 0xC9, 0x03, 0x68, 0xCC, 0x88, + 0xC9, 0x03, 0x68, 0xCC, 0x8C, 0xC9, 0x03, 0x68, + 0xCC, 0xA3, 0xB5, 0x03, 0x68, 0xCC, 0xA7, 0xA5, + 0x03, 0x68, 0xCC, 0xAE, 0xB5, 0x03, 0x68, 0xCC, + // Bytes 3380 - 33bf + 0xB1, 0xB5, 0x03, 0x69, 0xCC, 0x80, 0xC9, 0x03, + 0x69, 0xCC, 0x81, 0xC9, 0x03, 0x69, 0xCC, 0x82, + 0xC9, 0x03, 0x69, 0xCC, 0x83, 0xC9, 0x03, 0x69, + 0xCC, 0x84, 0xC9, 0x03, 0x69, 0xCC, 0x86, 0xC9, + 0x03, 0x69, 0xCC, 0x89, 0xC9, 0x03, 0x69, 0xCC, + 0x8C, 0xC9, 0x03, 0x69, 0xCC, 0x8F, 0xC9, 0x03, + 0x69, 0xCC, 0x91, 0xC9, 0x03, 0x69, 0xCC, 0xA3, + 0xB5, 0x03, 0x69, 0xCC, 0xA8, 0xA5, 0x03, 0x69, + // Bytes 33c0 - 33ff + 0xCC, 0xB0, 0xB5, 0x03, 0x6A, 0xCC, 0x82, 0xC9, + 0x03, 0x6A, 0xCC, 0x8C, 0xC9, 0x03, 0x6B, 0xCC, + 0x81, 0xC9, 0x03, 0x6B, 0xCC, 0x8C, 0xC9, 0x03, + 0x6B, 0xCC, 0xA3, 0xB5, 0x03, 0x6B, 0xCC, 0xA7, + 0xA5, 0x03, 0x6B, 0xCC, 0xB1, 0xB5, 0x03, 0x6C, + 0xCC, 0x81, 0xC9, 0x03, 0x6C, 0xCC, 0x8C, 0xC9, + 0x03, 0x6C, 0xCC, 0xA7, 0xA5, 0x03, 0x6C, 0xCC, + 0xAD, 0xB5, 0x03, 0x6C, 0xCC, 0xB1, 0xB5, 0x03, + // Bytes 3400 - 343f + 0x6D, 0xCC, 0x81, 0xC9, 0x03, 0x6D, 0xCC, 0x87, + 0xC9, 0x03, 0x6D, 0xCC, 0xA3, 0xB5, 0x03, 0x6E, + 0xCC, 0x80, 0xC9, 0x03, 0x6E, 0xCC, 0x81, 0xC9, + 0x03, 0x6E, 0xCC, 0x83, 0xC9, 0x03, 0x6E, 0xCC, + 0x87, 0xC9, 0x03, 0x6E, 0xCC, 0x8C, 0xC9, 0x03, + 0x6E, 0xCC, 0xA3, 0xB5, 0x03, 0x6E, 0xCC, 0xA7, + 0xA5, 0x03, 0x6E, 0xCC, 0xAD, 0xB5, 0x03, 0x6E, + 0xCC, 0xB1, 0xB5, 0x03, 0x6F, 0xCC, 0x80, 0xC9, + // Bytes 3440 - 347f + 0x03, 0x6F, 0xCC, 0x81, 0xC9, 0x03, 0x6F, 0xCC, + 0x86, 0xC9, 0x03, 0x6F, 0xCC, 0x89, 0xC9, 0x03, + 0x6F, 0xCC, 0x8B, 0xC9, 0x03, 0x6F, 0xCC, 0x8C, + 0xC9, 0x03, 0x6F, 0xCC, 0x8F, 0xC9, 0x03, 0x6F, + 0xCC, 0x91, 0xC9, 0x03, 0x70, 0xCC, 0x81, 0xC9, + 0x03, 0x70, 0xCC, 0x87, 0xC9, 0x03, 0x72, 0xCC, + 0x81, 0xC9, 0x03, 0x72, 0xCC, 0x87, 0xC9, 0x03, + 0x72, 0xCC, 0x8C, 0xC9, 0x03, 0x72, 0xCC, 0x8F, + // Bytes 3480 - 34bf + 0xC9, 0x03, 0x72, 0xCC, 0x91, 0xC9, 0x03, 0x72, + 0xCC, 0xA7, 0xA5, 0x03, 0x72, 0xCC, 0xB1, 0xB5, + 0x03, 0x73, 0xCC, 0x82, 0xC9, 0x03, 0x73, 0xCC, + 0x87, 0xC9, 0x03, 0x73, 0xCC, 0xA6, 0xB5, 0x03, + 0x73, 0xCC, 0xA7, 0xA5, 0x03, 0x74, 0xCC, 0x87, + 0xC9, 0x03, 0x74, 0xCC, 0x88, 0xC9, 0x03, 0x74, + 0xCC, 0x8C, 0xC9, 0x03, 0x74, 0xCC, 0xA3, 0xB5, + 0x03, 0x74, 0xCC, 0xA6, 0xB5, 0x03, 0x74, 0xCC, + // Bytes 34c0 - 34ff + 0xA7, 0xA5, 0x03, 0x74, 0xCC, 0xAD, 0xB5, 0x03, + 0x74, 0xCC, 0xB1, 0xB5, 0x03, 0x75, 0xCC, 0x80, + 0xC9, 0x03, 0x75, 0xCC, 0x81, 0xC9, 0x03, 0x75, + 0xCC, 0x82, 0xC9, 0x03, 0x75, 0xCC, 0x86, 0xC9, + 0x03, 0x75, 0xCC, 0x89, 0xC9, 0x03, 0x75, 0xCC, + 0x8A, 0xC9, 0x03, 0x75, 0xCC, 0x8B, 0xC9, 0x03, + 0x75, 0xCC, 0x8C, 0xC9, 0x03, 0x75, 0xCC, 0x8F, + 0xC9, 0x03, 0x75, 0xCC, 0x91, 0xC9, 0x03, 0x75, + // Bytes 3500 - 353f + 0xCC, 0xA3, 0xB5, 0x03, 0x75, 0xCC, 0xA4, 0xB5, + 0x03, 0x75, 0xCC, 0xA8, 0xA5, 0x03, 0x75, 0xCC, + 0xAD, 0xB5, 0x03, 0x75, 0xCC, 0xB0, 0xB5, 0x03, + 0x76, 0xCC, 0x83, 0xC9, 0x03, 0x76, 0xCC, 0xA3, + 0xB5, 0x03, 0x77, 0xCC, 0x80, 0xC9, 0x03, 0x77, + 0xCC, 0x81, 0xC9, 0x03, 0x77, 0xCC, 0x82, 0xC9, + 0x03, 0x77, 0xCC, 0x87, 0xC9, 0x03, 0x77, 0xCC, + 0x88, 0xC9, 0x03, 0x77, 0xCC, 0x8A, 0xC9, 0x03, + // Bytes 3540 - 357f + 0x77, 0xCC, 0xA3, 0xB5, 0x03, 0x78, 0xCC, 0x87, + 0xC9, 0x03, 0x78, 0xCC, 0x88, 0xC9, 0x03, 0x79, + 0xCC, 0x80, 0xC9, 0x03, 0x79, 0xCC, 0x81, 0xC9, + 0x03, 0x79, 0xCC, 0x82, 0xC9, 0x03, 0x79, 0xCC, + 0x83, 0xC9, 0x03, 0x79, 0xCC, 0x84, 0xC9, 0x03, + 0x79, 0xCC, 0x87, 0xC9, 0x03, 0x79, 0xCC, 0x88, + 0xC9, 0x03, 0x79, 0xCC, 0x89, 0xC9, 0x03, 0x79, + 0xCC, 0x8A, 0xC9, 0x03, 0x79, 0xCC, 0xA3, 0xB5, + // Bytes 3580 - 35bf + 0x03, 0x7A, 0xCC, 0x81, 0xC9, 0x03, 0x7A, 0xCC, + 0x82, 0xC9, 0x03, 0x7A, 0xCC, 0x87, 0xC9, 0x03, + 0x7A, 0xCC, 0x8C, 0xC9, 0x03, 0x7A, 0xCC, 0xA3, + 0xB5, 0x03, 0x7A, 0xCC, 0xB1, 0xB5, 0x04, 0xC2, + 0xA8, 0xCC, 0x80, 0xCA, 0x04, 0xC2, 0xA8, 0xCC, + 0x81, 0xCA, 0x04, 0xC2, 0xA8, 0xCD, 0x82, 0xCA, + 0x04, 0xC3, 0x86, 0xCC, 0x81, 0xC9, 0x04, 0xC3, + 0x86, 0xCC, 0x84, 0xC9, 0x04, 0xC3, 0x98, 0xCC, + // Bytes 35c0 - 35ff + 0x81, 0xC9, 0x04, 0xC3, 0xA6, 0xCC, 0x81, 0xC9, + 0x04, 0xC3, 0xA6, 0xCC, 0x84, 0xC9, 0x04, 0xC3, + 0xB8, 0xCC, 0x81, 0xC9, 0x04, 0xC5, 0xBF, 0xCC, + 0x87, 0xC9, 0x04, 0xC6, 0xB7, 0xCC, 0x8C, 0xC9, + 0x04, 0xCA, 0x92, 0xCC, 0x8C, 0xC9, 0x04, 0xCE, + 0x91, 0xCC, 0x80, 0xC9, 0x04, 0xCE, 0x91, 0xCC, + 0x81, 0xC9, 0x04, 0xCE, 0x91, 0xCC, 0x84, 0xC9, + 0x04, 0xCE, 0x91, 0xCC, 0x86, 0xC9, 0x04, 0xCE, + // Bytes 3600 - 363f + 0x91, 0xCD, 0x85, 0xD9, 0x04, 0xCE, 0x95, 0xCC, + 0x80, 0xC9, 0x04, 0xCE, 0x95, 0xCC, 0x81, 0xC9, + 0x04, 0xCE, 0x97, 0xCC, 0x80, 0xC9, 0x04, 0xCE, + 0x97, 0xCC, 0x81, 0xC9, 0x04, 0xCE, 0x97, 0xCD, + 0x85, 0xD9, 0x04, 0xCE, 0x99, 0xCC, 0x80, 0xC9, + 0x04, 0xCE, 0x99, 0xCC, 0x81, 0xC9, 0x04, 0xCE, + 0x99, 0xCC, 0x84, 0xC9, 0x04, 0xCE, 0x99, 0xCC, + 0x86, 0xC9, 0x04, 0xCE, 0x99, 0xCC, 0x88, 0xC9, + // Bytes 3640 - 367f + 0x04, 0xCE, 0x9F, 0xCC, 0x80, 0xC9, 0x04, 0xCE, + 0x9F, 0xCC, 0x81, 0xC9, 0x04, 0xCE, 0xA1, 0xCC, + 0x94, 0xC9, 0x04, 0xCE, 0xA5, 0xCC, 0x80, 0xC9, + 0x04, 0xCE, 0xA5, 0xCC, 0x81, 0xC9, 0x04, 0xCE, + 0xA5, 0xCC, 0x84, 0xC9, 0x04, 0xCE, 0xA5, 0xCC, + 0x86, 0xC9, 0x04, 0xCE, 0xA5, 0xCC, 0x88, 0xC9, + 0x04, 0xCE, 0xA9, 0xCC, 0x80, 0xC9, 0x04, 0xCE, + 0xA9, 0xCC, 0x81, 0xC9, 0x04, 0xCE, 0xA9, 0xCD, + // Bytes 3680 - 36bf + 0x85, 0xD9, 0x04, 0xCE, 0xB1, 0xCC, 0x84, 0xC9, + 0x04, 0xCE, 0xB1, 0xCC, 0x86, 0xC9, 0x04, 0xCE, + 0xB1, 0xCD, 0x85, 0xD9, 0x04, 0xCE, 0xB5, 0xCC, + 0x80, 0xC9, 0x04, 0xCE, 0xB5, 0xCC, 0x81, 0xC9, + 0x04, 0xCE, 0xB7, 0xCD, 0x85, 0xD9, 0x04, 0xCE, + 0xB9, 0xCC, 0x80, 0xC9, 0x04, 0xCE, 0xB9, 0xCC, + 0x81, 0xC9, 0x04, 0xCE, 0xB9, 0xCC, 0x84, 0xC9, + 0x04, 0xCE, 0xB9, 0xCC, 0x86, 0xC9, 0x04, 0xCE, + // Bytes 36c0 - 36ff + 0xB9, 0xCD, 0x82, 0xC9, 0x04, 0xCE, 0xBF, 0xCC, + 0x80, 0xC9, 0x04, 0xCE, 0xBF, 0xCC, 0x81, 0xC9, + 0x04, 0xCF, 0x81, 0xCC, 0x93, 0xC9, 0x04, 0xCF, + 0x81, 0xCC, 0x94, 0xC9, 0x04, 0xCF, 0x85, 0xCC, + 0x80, 0xC9, 0x04, 0xCF, 0x85, 0xCC, 0x81, 0xC9, + 0x04, 0xCF, 0x85, 0xCC, 0x84, 0xC9, 0x04, 0xCF, + 0x85, 0xCC, 0x86, 0xC9, 0x04, 0xCF, 0x85, 0xCD, + 0x82, 0xC9, 0x04, 0xCF, 0x89, 0xCD, 0x85, 0xD9, + // Bytes 3700 - 373f + 0x04, 0xCF, 0x92, 0xCC, 0x81, 0xC9, 0x04, 0xCF, + 0x92, 0xCC, 0x88, 0xC9, 0x04, 0xD0, 0x86, 0xCC, + 0x88, 0xC9, 0x04, 0xD0, 0x90, 0xCC, 0x86, 0xC9, + 0x04, 0xD0, 0x90, 0xCC, 0x88, 0xC9, 0x04, 0xD0, + 0x93, 0xCC, 0x81, 0xC9, 0x04, 0xD0, 0x95, 0xCC, + 0x80, 0xC9, 0x04, 0xD0, 0x95, 0xCC, 0x86, 0xC9, + 0x04, 0xD0, 0x95, 0xCC, 0x88, 0xC9, 0x04, 0xD0, + 0x96, 0xCC, 0x86, 0xC9, 0x04, 0xD0, 0x96, 0xCC, + // Bytes 3740 - 377f + 0x88, 0xC9, 0x04, 0xD0, 0x97, 0xCC, 0x88, 0xC9, + 0x04, 0xD0, 0x98, 0xCC, 0x80, 0xC9, 0x04, 0xD0, + 0x98, 0xCC, 0x84, 0xC9, 0x04, 0xD0, 0x98, 0xCC, + 0x86, 0xC9, 0x04, 0xD0, 0x98, 0xCC, 0x88, 0xC9, + 0x04, 0xD0, 0x9A, 0xCC, 0x81, 0xC9, 0x04, 0xD0, + 0x9E, 0xCC, 0x88, 0xC9, 0x04, 0xD0, 0xA3, 0xCC, + 0x84, 0xC9, 0x04, 0xD0, 0xA3, 0xCC, 0x86, 0xC9, + 0x04, 0xD0, 0xA3, 0xCC, 0x88, 0xC9, 0x04, 0xD0, + // Bytes 3780 - 37bf + 0xA3, 0xCC, 0x8B, 0xC9, 0x04, 0xD0, 0xA7, 0xCC, + 0x88, 0xC9, 0x04, 0xD0, 0xAB, 0xCC, 0x88, 0xC9, + 0x04, 0xD0, 0xAD, 0xCC, 0x88, 0xC9, 0x04, 0xD0, + 0xB0, 0xCC, 0x86, 0xC9, 0x04, 0xD0, 0xB0, 0xCC, + 0x88, 0xC9, 0x04, 0xD0, 0xB3, 0xCC, 0x81, 0xC9, + 0x04, 0xD0, 0xB5, 0xCC, 0x80, 0xC9, 0x04, 0xD0, + 0xB5, 0xCC, 0x86, 0xC9, 0x04, 0xD0, 0xB5, 0xCC, + 0x88, 0xC9, 0x04, 0xD0, 0xB6, 0xCC, 0x86, 0xC9, + // Bytes 37c0 - 37ff + 0x04, 0xD0, 0xB6, 0xCC, 0x88, 0xC9, 0x04, 0xD0, + 0xB7, 0xCC, 0x88, 0xC9, 0x04, 0xD0, 0xB8, 0xCC, + 0x80, 0xC9, 0x04, 0xD0, 0xB8, 0xCC, 0x84, 0xC9, + 0x04, 0xD0, 0xB8, 0xCC, 0x86, 0xC9, 0x04, 0xD0, + 0xB8, 0xCC, 0x88, 0xC9, 0x04, 0xD0, 0xBA, 0xCC, + 0x81, 0xC9, 0x04, 0xD0, 0xBE, 0xCC, 0x88, 0xC9, + 0x04, 0xD1, 0x83, 0xCC, 0x84, 0xC9, 0x04, 0xD1, + 0x83, 0xCC, 0x86, 0xC9, 0x04, 0xD1, 0x83, 0xCC, + // Bytes 3800 - 383f + 0x88, 0xC9, 0x04, 0xD1, 0x83, 0xCC, 0x8B, 0xC9, + 0x04, 0xD1, 0x87, 0xCC, 0x88, 0xC9, 0x04, 0xD1, + 0x8B, 0xCC, 0x88, 0xC9, 0x04, 0xD1, 0x8D, 0xCC, + 0x88, 0xC9, 0x04, 0xD1, 0x96, 0xCC, 0x88, 0xC9, + 0x04, 0xD1, 0xB4, 0xCC, 0x8F, 0xC9, 0x04, 0xD1, + 0xB5, 0xCC, 0x8F, 0xC9, 0x04, 0xD3, 0x98, 0xCC, + 0x88, 0xC9, 0x04, 0xD3, 0x99, 0xCC, 0x88, 0xC9, + 0x04, 0xD3, 0xA8, 0xCC, 0x88, 0xC9, 0x04, 0xD3, + // Bytes 3840 - 387f + 0xA9, 0xCC, 0x88, 0xC9, 0x04, 0xD8, 0xA7, 0xD9, + 0x93, 0xC9, 0x04, 0xD8, 0xA7, 0xD9, 0x94, 0xC9, + 0x04, 0xD8, 0xA7, 0xD9, 0x95, 0xB5, 0x04, 0xD9, + 0x88, 0xD9, 0x94, 0xC9, 0x04, 0xD9, 0x8A, 0xD9, + 0x94, 0xC9, 0x04, 0xDB, 0x81, 0xD9, 0x94, 0xC9, + 0x04, 0xDB, 0x92, 0xD9, 0x94, 0xC9, 0x04, 0xDB, + 0x95, 0xD9, 0x94, 0xC9, 0x05, 0x41, 0xCC, 0x82, + 0xCC, 0x80, 0xCA, 0x05, 0x41, 0xCC, 0x82, 0xCC, + // Bytes 3880 - 38bf + 0x81, 0xCA, 0x05, 0x41, 0xCC, 0x82, 0xCC, 0x83, + 0xCA, 0x05, 0x41, 0xCC, 0x82, 0xCC, 0x89, 0xCA, + 0x05, 0x41, 0xCC, 0x86, 0xCC, 0x80, 0xCA, 0x05, + 0x41, 0xCC, 0x86, 0xCC, 0x81, 0xCA, 0x05, 0x41, + 0xCC, 0x86, 0xCC, 0x83, 0xCA, 0x05, 0x41, 0xCC, + 0x86, 0xCC, 0x89, 0xCA, 0x05, 0x41, 0xCC, 0x87, + 0xCC, 0x84, 0xCA, 0x05, 0x41, 0xCC, 0x88, 0xCC, + 0x84, 0xCA, 0x05, 0x41, 0xCC, 0x8A, 0xCC, 0x81, + // Bytes 38c0 - 38ff + 0xCA, 0x05, 0x41, 0xCC, 0xA3, 0xCC, 0x82, 0xCA, + 0x05, 0x41, 0xCC, 0xA3, 0xCC, 0x86, 0xCA, 0x05, + 0x43, 0xCC, 0xA7, 0xCC, 0x81, 0xCA, 0x05, 0x45, + 0xCC, 0x82, 0xCC, 0x80, 0xCA, 0x05, 0x45, 0xCC, + 0x82, 0xCC, 0x81, 0xCA, 0x05, 0x45, 0xCC, 0x82, + 0xCC, 0x83, 0xCA, 0x05, 0x45, 0xCC, 0x82, 0xCC, + 0x89, 0xCA, 0x05, 0x45, 0xCC, 0x84, 0xCC, 0x80, + 0xCA, 0x05, 0x45, 0xCC, 0x84, 0xCC, 0x81, 0xCA, + // Bytes 3900 - 393f + 0x05, 0x45, 0xCC, 0xA3, 0xCC, 0x82, 0xCA, 0x05, + 0x45, 0xCC, 0xA7, 0xCC, 0x86, 0xCA, 0x05, 0x49, + 0xCC, 0x88, 0xCC, 0x81, 0xCA, 0x05, 0x4C, 0xCC, + 0xA3, 0xCC, 0x84, 0xCA, 0x05, 0x4F, 0xCC, 0x82, + 0xCC, 0x80, 0xCA, 0x05, 0x4F, 0xCC, 0x82, 0xCC, + 0x81, 0xCA, 0x05, 0x4F, 0xCC, 0x82, 0xCC, 0x83, + 0xCA, 0x05, 0x4F, 0xCC, 0x82, 0xCC, 0x89, 0xCA, + 0x05, 0x4F, 0xCC, 0x83, 0xCC, 0x81, 0xCA, 0x05, + // Bytes 3940 - 397f + 0x4F, 0xCC, 0x83, 0xCC, 0x84, 0xCA, 0x05, 0x4F, + 0xCC, 0x83, 0xCC, 0x88, 0xCA, 0x05, 0x4F, 0xCC, + 0x84, 0xCC, 0x80, 0xCA, 0x05, 0x4F, 0xCC, 0x84, + 0xCC, 0x81, 0xCA, 0x05, 0x4F, 0xCC, 0x87, 0xCC, + 0x84, 0xCA, 0x05, 0x4F, 0xCC, 0x88, 0xCC, 0x84, + 0xCA, 0x05, 0x4F, 0xCC, 0x9B, 0xCC, 0x80, 0xCA, + 0x05, 0x4F, 0xCC, 0x9B, 0xCC, 0x81, 0xCA, 0x05, + 0x4F, 0xCC, 0x9B, 0xCC, 0x83, 0xCA, 0x05, 0x4F, + // Bytes 3980 - 39bf + 0xCC, 0x9B, 0xCC, 0x89, 0xCA, 0x05, 0x4F, 0xCC, + 0x9B, 0xCC, 0xA3, 0xB6, 0x05, 0x4F, 0xCC, 0xA3, + 0xCC, 0x82, 0xCA, 0x05, 0x4F, 0xCC, 0xA8, 0xCC, + 0x84, 0xCA, 0x05, 0x52, 0xCC, 0xA3, 0xCC, 0x84, + 0xCA, 0x05, 0x53, 0xCC, 0x81, 0xCC, 0x87, 0xCA, + 0x05, 0x53, 0xCC, 0x8C, 0xCC, 0x87, 0xCA, 0x05, + 0x53, 0xCC, 0xA3, 0xCC, 0x87, 0xCA, 0x05, 0x55, + 0xCC, 0x83, 0xCC, 0x81, 0xCA, 0x05, 0x55, 0xCC, + // Bytes 39c0 - 39ff + 0x84, 0xCC, 0x88, 0xCA, 0x05, 0x55, 0xCC, 0x88, + 0xCC, 0x80, 0xCA, 0x05, 0x55, 0xCC, 0x88, 0xCC, + 0x81, 0xCA, 0x05, 0x55, 0xCC, 0x88, 0xCC, 0x84, + 0xCA, 0x05, 0x55, 0xCC, 0x88, 0xCC, 0x8C, 0xCA, + 0x05, 0x55, 0xCC, 0x9B, 0xCC, 0x80, 0xCA, 0x05, + 0x55, 0xCC, 0x9B, 0xCC, 0x81, 0xCA, 0x05, 0x55, + 0xCC, 0x9B, 0xCC, 0x83, 0xCA, 0x05, 0x55, 0xCC, + 0x9B, 0xCC, 0x89, 0xCA, 0x05, 0x55, 0xCC, 0x9B, + // Bytes 3a00 - 3a3f + 0xCC, 0xA3, 0xB6, 0x05, 0x61, 0xCC, 0x82, 0xCC, + 0x80, 0xCA, 0x05, 0x61, 0xCC, 0x82, 0xCC, 0x81, + 0xCA, 0x05, 0x61, 0xCC, 0x82, 0xCC, 0x83, 0xCA, + 0x05, 0x61, 0xCC, 0x82, 0xCC, 0x89, 0xCA, 0x05, + 0x61, 0xCC, 0x86, 0xCC, 0x80, 0xCA, 0x05, 0x61, + 0xCC, 0x86, 0xCC, 0x81, 0xCA, 0x05, 0x61, 0xCC, + 0x86, 0xCC, 0x83, 0xCA, 0x05, 0x61, 0xCC, 0x86, + 0xCC, 0x89, 0xCA, 0x05, 0x61, 0xCC, 0x87, 0xCC, + // Bytes 3a40 - 3a7f + 0x84, 0xCA, 0x05, 0x61, 0xCC, 0x88, 0xCC, 0x84, + 0xCA, 0x05, 0x61, 0xCC, 0x8A, 0xCC, 0x81, 0xCA, + 0x05, 0x61, 0xCC, 0xA3, 0xCC, 0x82, 0xCA, 0x05, + 0x61, 0xCC, 0xA3, 0xCC, 0x86, 0xCA, 0x05, 0x63, + 0xCC, 0xA7, 0xCC, 0x81, 0xCA, 0x05, 0x65, 0xCC, + 0x82, 0xCC, 0x80, 0xCA, 0x05, 0x65, 0xCC, 0x82, + 0xCC, 0x81, 0xCA, 0x05, 0x65, 0xCC, 0x82, 0xCC, + 0x83, 0xCA, 0x05, 0x65, 0xCC, 0x82, 0xCC, 0x89, + // Bytes 3a80 - 3abf + 0xCA, 0x05, 0x65, 0xCC, 0x84, 0xCC, 0x80, 0xCA, + 0x05, 0x65, 0xCC, 0x84, 0xCC, 0x81, 0xCA, 0x05, + 0x65, 0xCC, 0xA3, 0xCC, 0x82, 0xCA, 0x05, 0x65, + 0xCC, 0xA7, 0xCC, 0x86, 0xCA, 0x05, 0x69, 0xCC, + 0x88, 0xCC, 0x81, 0xCA, 0x05, 0x6C, 0xCC, 0xA3, + 0xCC, 0x84, 0xCA, 0x05, 0x6F, 0xCC, 0x82, 0xCC, + 0x80, 0xCA, 0x05, 0x6F, 0xCC, 0x82, 0xCC, 0x81, + 0xCA, 0x05, 0x6F, 0xCC, 0x82, 0xCC, 0x83, 0xCA, + // Bytes 3ac0 - 3aff + 0x05, 0x6F, 0xCC, 0x82, 0xCC, 0x89, 0xCA, 0x05, + 0x6F, 0xCC, 0x83, 0xCC, 0x81, 0xCA, 0x05, 0x6F, + 0xCC, 0x83, 0xCC, 0x84, 0xCA, 0x05, 0x6F, 0xCC, + 0x83, 0xCC, 0x88, 0xCA, 0x05, 0x6F, 0xCC, 0x84, + 0xCC, 0x80, 0xCA, 0x05, 0x6F, 0xCC, 0x84, 0xCC, + 0x81, 0xCA, 0x05, 0x6F, 0xCC, 0x87, 0xCC, 0x84, + 0xCA, 0x05, 0x6F, 0xCC, 0x88, 0xCC, 0x84, 0xCA, + 0x05, 0x6F, 0xCC, 0x9B, 0xCC, 0x80, 0xCA, 0x05, + // Bytes 3b00 - 3b3f + 0x6F, 0xCC, 0x9B, 0xCC, 0x81, 0xCA, 0x05, 0x6F, + 0xCC, 0x9B, 0xCC, 0x83, 0xCA, 0x05, 0x6F, 0xCC, + 0x9B, 0xCC, 0x89, 0xCA, 0x05, 0x6F, 0xCC, 0x9B, + 0xCC, 0xA3, 0xB6, 0x05, 0x6F, 0xCC, 0xA3, 0xCC, + 0x82, 0xCA, 0x05, 0x6F, 0xCC, 0xA8, 0xCC, 0x84, + 0xCA, 0x05, 0x72, 0xCC, 0xA3, 0xCC, 0x84, 0xCA, + 0x05, 0x73, 0xCC, 0x81, 0xCC, 0x87, 0xCA, 0x05, + 0x73, 0xCC, 0x8C, 0xCC, 0x87, 0xCA, 0x05, 0x73, + // Bytes 3b40 - 3b7f + 0xCC, 0xA3, 0xCC, 0x87, 0xCA, 0x05, 0x75, 0xCC, + 0x83, 0xCC, 0x81, 0xCA, 0x05, 0x75, 0xCC, 0x84, + 0xCC, 0x88, 0xCA, 0x05, 0x75, 0xCC, 0x88, 0xCC, + 0x80, 0xCA, 0x05, 0x75, 0xCC, 0x88, 0xCC, 0x81, + 0xCA, 0x05, 0x75, 0xCC, 0x88, 0xCC, 0x84, 0xCA, + 0x05, 0x75, 0xCC, 0x88, 0xCC, 0x8C, 0xCA, 0x05, + 0x75, 0xCC, 0x9B, 0xCC, 0x80, 0xCA, 0x05, 0x75, + 0xCC, 0x9B, 0xCC, 0x81, 0xCA, 0x05, 0x75, 0xCC, + // Bytes 3b80 - 3bbf + 0x9B, 0xCC, 0x83, 0xCA, 0x05, 0x75, 0xCC, 0x9B, + 0xCC, 0x89, 0xCA, 0x05, 0x75, 0xCC, 0x9B, 0xCC, + 0xA3, 0xB6, 0x05, 0xE1, 0xBE, 0xBF, 0xCC, 0x80, + 0xCA, 0x05, 0xE1, 0xBE, 0xBF, 0xCC, 0x81, 0xCA, + 0x05, 0xE1, 0xBE, 0xBF, 0xCD, 0x82, 0xCA, 0x05, + 0xE1, 0xBF, 0xBE, 0xCC, 0x80, 0xCA, 0x05, 0xE1, + 0xBF, 0xBE, 0xCC, 0x81, 0xCA, 0x05, 0xE1, 0xBF, + 0xBE, 0xCD, 0x82, 0xCA, 0x05, 0xE2, 0x86, 0x90, + // Bytes 3bc0 - 3bff + 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x86, 0x92, 0xCC, + 0xB8, 0x05, 0x05, 0xE2, 0x86, 0x94, 0xCC, 0xB8, + 0x05, 0x05, 0xE2, 0x87, 0x90, 0xCC, 0xB8, 0x05, + 0x05, 0xE2, 0x87, 0x92, 0xCC, 0xB8, 0x05, 0x05, + 0xE2, 0x87, 0x94, 0xCC, 0xB8, 0x05, 0x05, 0xE2, + 0x88, 0x83, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x88, + 0x88, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x88, 0x8B, + 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x88, 0xA3, 0xCC, + // Bytes 3c00 - 3c3f + 0xB8, 0x05, 0x05, 0xE2, 0x88, 0xA5, 0xCC, 0xB8, + 0x05, 0x05, 0xE2, 0x88, 0xBC, 0xCC, 0xB8, 0x05, + 0x05, 0xE2, 0x89, 0x83, 0xCC, 0xB8, 0x05, 0x05, + 0xE2, 0x89, 0x85, 0xCC, 0xB8, 0x05, 0x05, 0xE2, + 0x89, 0x88, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x89, + 0x8D, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x89, 0xA1, + 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x89, 0xA4, 0xCC, + 0xB8, 0x05, 0x05, 0xE2, 0x89, 0xA5, 0xCC, 0xB8, + // Bytes 3c40 - 3c7f + 0x05, 0x05, 0xE2, 0x89, 0xB2, 0xCC, 0xB8, 0x05, + 0x05, 0xE2, 0x89, 0xB3, 0xCC, 0xB8, 0x05, 0x05, + 0xE2, 0x89, 0xB6, 0xCC, 0xB8, 0x05, 0x05, 0xE2, + 0x89, 0xB7, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x89, + 0xBA, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x89, 0xBB, + 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x89, 0xBC, 0xCC, + 0xB8, 0x05, 0x05, 0xE2, 0x89, 0xBD, 0xCC, 0xB8, + 0x05, 0x05, 0xE2, 0x8A, 0x82, 0xCC, 0xB8, 0x05, + // Bytes 3c80 - 3cbf + 0x05, 0xE2, 0x8A, 0x83, 0xCC, 0xB8, 0x05, 0x05, + 0xE2, 0x8A, 0x86, 0xCC, 0xB8, 0x05, 0x05, 0xE2, + 0x8A, 0x87, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x8A, + 0x91, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x8A, 0x92, + 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x8A, 0xA2, 0xCC, + 0xB8, 0x05, 0x05, 0xE2, 0x8A, 0xA8, 0xCC, 0xB8, + 0x05, 0x05, 0xE2, 0x8A, 0xA9, 0xCC, 0xB8, 0x05, + 0x05, 0xE2, 0x8A, 0xAB, 0xCC, 0xB8, 0x05, 0x05, + // Bytes 3cc0 - 3cff + 0xE2, 0x8A, 0xB2, 0xCC, 0xB8, 0x05, 0x05, 0xE2, + 0x8A, 0xB3, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x8A, + 0xB4, 0xCC, 0xB8, 0x05, 0x05, 0xE2, 0x8A, 0xB5, + 0xCC, 0xB8, 0x05, 0x06, 0xCE, 0x91, 0xCC, 0x93, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0x91, 0xCC, 0x94, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0x95, 0xCC, 0x93, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0x95, 0xCC, 0x93, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0x95, 0xCC, 0x94, + // Bytes 3d00 - 3d3f + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0x95, 0xCC, 0x94, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0x97, 0xCC, 0x93, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0x97, 0xCC, 0x94, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0x99, 0xCC, 0x93, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0x99, 0xCC, 0x93, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0x99, 0xCC, 0x93, + 0xCD, 0x82, 0xCA, 0x06, 0xCE, 0x99, 0xCC, 0x94, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0x99, 0xCC, 0x94, + // Bytes 3d40 - 3d7f + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0x99, 0xCC, 0x94, + 0xCD, 0x82, 0xCA, 0x06, 0xCE, 0x9F, 0xCC, 0x93, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0x9F, 0xCC, 0x93, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0x9F, 0xCC, 0x94, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0x9F, 0xCC, 0x94, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xA5, 0xCC, 0x94, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xA5, 0xCC, 0x94, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xA5, 0xCC, 0x94, + // Bytes 3d80 - 3dbf + 0xCD, 0x82, 0xCA, 0x06, 0xCE, 0xA9, 0xCC, 0x93, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xA9, 0xCC, 0x94, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB1, 0xCC, 0x80, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB1, 0xCC, 0x81, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB1, 0xCC, 0x93, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB1, 0xCC, 0x94, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB1, 0xCD, 0x82, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB5, 0xCC, 0x93, + // Bytes 3dc0 - 3dff + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xB5, 0xCC, 0x93, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xB5, 0xCC, 0x94, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xB5, 0xCC, 0x94, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xB7, 0xCC, 0x80, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB7, 0xCC, 0x81, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB7, 0xCC, 0x93, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB7, 0xCC, 0x94, + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB7, 0xCD, 0x82, + // Bytes 3e00 - 3e3f + 0xCD, 0x85, 0xDA, 0x06, 0xCE, 0xB9, 0xCC, 0x88, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x88, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x88, + 0xCD, 0x82, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x93, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x93, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x93, + 0xCD, 0x82, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x94, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x94, + // Bytes 3e40 - 3e7f + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xB9, 0xCC, 0x94, + 0xCD, 0x82, 0xCA, 0x06, 0xCE, 0xBF, 0xCC, 0x93, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xBF, 0xCC, 0x93, + 0xCC, 0x81, 0xCA, 0x06, 0xCE, 0xBF, 0xCC, 0x94, + 0xCC, 0x80, 0xCA, 0x06, 0xCE, 0xBF, 0xCC, 0x94, + 0xCC, 0x81, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x88, + 0xCC, 0x80, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x88, + 0xCC, 0x81, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x88, + // Bytes 3e80 - 3ebf + 0xCD, 0x82, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x93, + 0xCC, 0x80, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x93, + 0xCC, 0x81, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x93, + 0xCD, 0x82, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x94, + 0xCC, 0x80, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x94, + 0xCC, 0x81, 0xCA, 0x06, 0xCF, 0x85, 0xCC, 0x94, + 0xCD, 0x82, 0xCA, 0x06, 0xCF, 0x89, 0xCC, 0x80, + 0xCD, 0x85, 0xDA, 0x06, 0xCF, 0x89, 0xCC, 0x81, + // Bytes 3ec0 - 3eff + 0xCD, 0x85, 0xDA, 0x06, 0xCF, 0x89, 0xCC, 0x93, + 0xCD, 0x85, 0xDA, 0x06, 0xCF, 0x89, 0xCC, 0x94, + 0xCD, 0x85, 0xDA, 0x06, 0xCF, 0x89, 0xCD, 0x82, + 0xCD, 0x85, 0xDA, 0x06, 0xE0, 0xA4, 0xA8, 0xE0, + 0xA4, 0xBC, 0x09, 0x06, 0xE0, 0xA4, 0xB0, 0xE0, + 0xA4, 0xBC, 0x09, 0x06, 0xE0, 0xA4, 0xB3, 0xE0, + 0xA4, 0xBC, 0x09, 0x06, 0xE0, 0xB1, 0x86, 0xE0, + 0xB1, 0x96, 0x85, 0x06, 0xE0, 0xB7, 0x99, 0xE0, + // Bytes 3f00 - 3f3f + 0xB7, 0x8A, 0x11, 0x06, 0xE3, 0x81, 0x86, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x8B, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x8D, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x8F, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x91, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x93, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x95, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x97, 0xE3, + // Bytes 3f40 - 3f7f + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x99, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x9B, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x9D, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0x9F, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xA1, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xA4, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xA6, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xA8, 0xE3, + // Bytes 3f80 - 3fbf + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xAF, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xAF, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x81, 0xB2, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xB2, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x81, 0xB5, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xB5, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x81, 0xB8, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xB8, 0xE3, + // Bytes 3fc0 - 3fff + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x81, 0xBB, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x81, 0xBB, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x82, 0x9D, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xA6, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xAB, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xAD, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xAF, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xB1, 0xE3, + // Bytes 4000 - 403f + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xB3, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xB5, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xB7, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xB9, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xBB, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xBD, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x82, 0xBF, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x81, 0xE3, + // Bytes 4040 - 407f + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x84, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x86, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x88, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x8F, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x8F, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x83, 0x92, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x92, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x83, 0x95, 0xE3, + // Bytes 4080 - 40bf + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x95, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x83, 0x98, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x98, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x83, 0x9B, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0x9B, 0xE3, + 0x82, 0x9A, 0x0D, 0x06, 0xE3, 0x83, 0xAF, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0xB0, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0xB1, 0xE3, + // Bytes 40c0 - 40ff + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0xB2, 0xE3, + 0x82, 0x99, 0x0D, 0x06, 0xE3, 0x83, 0xBD, 0xE3, + 0x82, 0x99, 0x0D, 0x08, 0xCE, 0x91, 0xCC, 0x93, + 0xCC, 0x80, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0x91, + 0xCC, 0x93, 0xCC, 0x81, 0xCD, 0x85, 0xDB, 0x08, + 0xCE, 0x91, 0xCC, 0x93, 0xCD, 0x82, 0xCD, 0x85, + 0xDB, 0x08, 0xCE, 0x91, 0xCC, 0x94, 0xCC, 0x80, + 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0x91, 0xCC, 0x94, + // Bytes 4100 - 413f + 0xCC, 0x81, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0x91, + 0xCC, 0x94, 0xCD, 0x82, 0xCD, 0x85, 0xDB, 0x08, + 0xCE, 0x97, 0xCC, 0x93, 0xCC, 0x80, 0xCD, 0x85, + 0xDB, 0x08, 0xCE, 0x97, 0xCC, 0x93, 0xCC, 0x81, + 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0x97, 0xCC, 0x93, + 0xCD, 0x82, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0x97, + 0xCC, 0x94, 0xCC, 0x80, 0xCD, 0x85, 0xDB, 0x08, + 0xCE, 0x97, 0xCC, 0x94, 0xCC, 0x81, 0xCD, 0x85, + // Bytes 4140 - 417f + 0xDB, 0x08, 0xCE, 0x97, 0xCC, 0x94, 0xCD, 0x82, + 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xA9, 0xCC, 0x93, + 0xCC, 0x80, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xA9, + 0xCC, 0x93, 0xCC, 0x81, 0xCD, 0x85, 0xDB, 0x08, + 0xCE, 0xA9, 0xCC, 0x93, 0xCD, 0x82, 0xCD, 0x85, + 0xDB, 0x08, 0xCE, 0xA9, 0xCC, 0x94, 0xCC, 0x80, + 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xA9, 0xCC, 0x94, + 0xCC, 0x81, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xA9, + // Bytes 4180 - 41bf + 0xCC, 0x94, 0xCD, 0x82, 0xCD, 0x85, 0xDB, 0x08, + 0xCE, 0xB1, 0xCC, 0x93, 0xCC, 0x80, 0xCD, 0x85, + 0xDB, 0x08, 0xCE, 0xB1, 0xCC, 0x93, 0xCC, 0x81, + 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xB1, 0xCC, 0x93, + 0xCD, 0x82, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xB1, + 0xCC, 0x94, 0xCC, 0x80, 0xCD, 0x85, 0xDB, 0x08, + 0xCE, 0xB1, 0xCC, 0x94, 0xCC, 0x81, 0xCD, 0x85, + 0xDB, 0x08, 0xCE, 0xB1, 0xCC, 0x94, 0xCD, 0x82, + // Bytes 41c0 - 41ff + 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xB7, 0xCC, 0x93, + 0xCC, 0x80, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xB7, + 0xCC, 0x93, 0xCC, 0x81, 0xCD, 0x85, 0xDB, 0x08, + 0xCE, 0xB7, 0xCC, 0x93, 0xCD, 0x82, 0xCD, 0x85, + 0xDB, 0x08, 0xCE, 0xB7, 0xCC, 0x94, 0xCC, 0x80, + 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xB7, 0xCC, 0x94, + 0xCC, 0x81, 0xCD, 0x85, 0xDB, 0x08, 0xCE, 0xB7, + 0xCC, 0x94, 0xCD, 0x82, 0xCD, 0x85, 0xDB, 0x08, + // Bytes 4200 - 423f + 0xCF, 0x89, 0xCC, 0x93, 0xCC, 0x80, 0xCD, 0x85, + 0xDB, 0x08, 0xCF, 0x89, 0xCC, 0x93, 0xCC, 0x81, + 0xCD, 0x85, 0xDB, 0x08, 0xCF, 0x89, 0xCC, 0x93, + 0xCD, 0x82, 0xCD, 0x85, 0xDB, 0x08, 0xCF, 0x89, + 0xCC, 0x94, 0xCC, 0x80, 0xCD, 0x85, 0xDB, 0x08, + 0xCF, 0x89, 0xCC, 0x94, 0xCC, 0x81, 0xCD, 0x85, + 0xDB, 0x08, 0xCF, 0x89, 0xCC, 0x94, 0xCD, 0x82, + 0xCD, 0x85, 0xDB, 0x08, 0xF0, 0x91, 0x82, 0x99, + // Bytes 4240 - 427f + 0xF0, 0x91, 0x82, 0xBA, 0x09, 0x08, 0xF0, 0x91, + 0x82, 0x9B, 0xF0, 0x91, 0x82, 0xBA, 0x09, 0x08, + 0xF0, 0x91, 0x82, 0xA5, 0xF0, 0x91, 0x82, 0xBA, + 0x09, 0x42, 0xC2, 0xB4, 0x01, 0x43, 0x20, 0xCC, + 0x81, 0xC9, 0x43, 0x20, 0xCC, 0x83, 0xC9, 0x43, + 0x20, 0xCC, 0x84, 0xC9, 0x43, 0x20, 0xCC, 0x85, + 0xC9, 0x43, 0x20, 0xCC, 0x86, 0xC9, 0x43, 0x20, + 0xCC, 0x87, 0xC9, 0x43, 0x20, 0xCC, 0x88, 0xC9, + // Bytes 4280 - 42bf + 0x43, 0x20, 0xCC, 0x8A, 0xC9, 0x43, 0x20, 0xCC, + 0x8B, 0xC9, 0x43, 0x20, 0xCC, 0x93, 0xC9, 0x43, + 0x20, 0xCC, 0x94, 0xC9, 0x43, 0x20, 0xCC, 0xA7, + 0xA5, 0x43, 0x20, 0xCC, 0xA8, 0xA5, 0x43, 0x20, + 0xCC, 0xB3, 0xB5, 0x43, 0x20, 0xCD, 0x82, 0xC9, + 0x43, 0x20, 0xCD, 0x85, 0xD9, 0x43, 0x20, 0xD9, + 0x8B, 0x59, 0x43, 0x20, 0xD9, 0x8C, 0x5D, 0x43, + 0x20, 0xD9, 0x8D, 0x61, 0x43, 0x20, 0xD9, 0x8E, + // Bytes 42c0 - 42ff + 0x65, 0x43, 0x20, 0xD9, 0x8F, 0x69, 0x43, 0x20, + 0xD9, 0x90, 0x6D, 0x43, 0x20, 0xD9, 0x91, 0x71, + 0x43, 0x20, 0xD9, 0x92, 0x75, 0x43, 0x41, 0xCC, + 0x8A, 0xC9, 0x43, 0x73, 0xCC, 0x87, 0xC9, 0x44, + 0x20, 0xE3, 0x82, 0x99, 0x0D, 0x44, 0x20, 0xE3, + 0x82, 0x9A, 0x0D, 0x44, 0xC2, 0xA8, 0xCC, 0x81, + 0xCA, 0x44, 0xCE, 0x91, 0xCC, 0x81, 0xC9, 0x44, + 0xCE, 0x95, 0xCC, 0x81, 0xC9, 0x44, 0xCE, 0x97, + // Bytes 4300 - 433f + 0xCC, 0x81, 0xC9, 0x44, 0xCE, 0x99, 0xCC, 0x81, + 0xC9, 0x44, 0xCE, 0x9F, 0xCC, 0x81, 0xC9, 0x44, + 0xCE, 0xA5, 0xCC, 0x81, 0xC9, 0x44, 0xCE, 0xA5, + 0xCC, 0x88, 0xC9, 0x44, 0xCE, 0xA9, 0xCC, 0x81, + 0xC9, 0x44, 0xCE, 0xB1, 0xCC, 0x81, 0xC9, 0x44, + 0xCE, 0xB5, 0xCC, 0x81, 0xC9, 0x44, 0xCE, 0xB7, + 0xCC, 0x81, 0xC9, 0x44, 0xCE, 0xB9, 0xCC, 0x81, + 0xC9, 0x44, 0xCE, 0xBF, 0xCC, 0x81, 0xC9, 0x44, + // Bytes 4340 - 437f + 0xCF, 0x85, 0xCC, 0x81, 0xC9, 0x44, 0xCF, 0x89, + 0xCC, 0x81, 0xC9, 0x44, 0xD7, 0x90, 0xD6, 0xB7, + 0x31, 0x44, 0xD7, 0x90, 0xD6, 0xB8, 0x35, 0x44, + 0xD7, 0x90, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x91, + 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x91, 0xD6, 0xBF, + 0x49, 0x44, 0xD7, 0x92, 0xD6, 0xBC, 0x41, 0x44, + 0xD7, 0x93, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x94, + 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x95, 0xD6, 0xB9, + // Bytes 4380 - 43bf + 0x39, 0x44, 0xD7, 0x95, 0xD6, 0xBC, 0x41, 0x44, + 0xD7, 0x96, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x98, + 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x99, 0xD6, 0xB4, + 0x25, 0x44, 0xD7, 0x99, 0xD6, 0xBC, 0x41, 0x44, + 0xD7, 0x9A, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x9B, + 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0x9B, 0xD6, 0xBF, + 0x49, 0x44, 0xD7, 0x9C, 0xD6, 0xBC, 0x41, 0x44, + 0xD7, 0x9E, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0xA0, + // Bytes 43c0 - 43ff + 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0xA1, 0xD6, 0xBC, + 0x41, 0x44, 0xD7, 0xA3, 0xD6, 0xBC, 0x41, 0x44, + 0xD7, 0xA4, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0xA4, + 0xD6, 0xBF, 0x49, 0x44, 0xD7, 0xA6, 0xD6, 0xBC, + 0x41, 0x44, 0xD7, 0xA7, 0xD6, 0xBC, 0x41, 0x44, + 0xD7, 0xA8, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0xA9, + 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0xA9, 0xD7, 0x81, + 0x4D, 0x44, 0xD7, 0xA9, 0xD7, 0x82, 0x51, 0x44, + // Bytes 4400 - 443f + 0xD7, 0xAA, 0xD6, 0xBC, 0x41, 0x44, 0xD7, 0xB2, + 0xD6, 0xB7, 0x31, 0x44, 0xD8, 0xA7, 0xD9, 0x8B, + 0x59, 0x44, 0xD8, 0xA7, 0xD9, 0x93, 0xC9, 0x44, + 0xD8, 0xA7, 0xD9, 0x94, 0xC9, 0x44, 0xD8, 0xA7, + 0xD9, 0x95, 0xB5, 0x44, 0xD8, 0xB0, 0xD9, 0xB0, + 0x79, 0x44, 0xD8, 0xB1, 0xD9, 0xB0, 0x79, 0x44, + 0xD9, 0x80, 0xD9, 0x8B, 0x59, 0x44, 0xD9, 0x80, + 0xD9, 0x8E, 0x65, 0x44, 0xD9, 0x80, 0xD9, 0x8F, + // Bytes 4440 - 447f + 0x69, 0x44, 0xD9, 0x80, 0xD9, 0x90, 0x6D, 0x44, + 0xD9, 0x80, 0xD9, 0x91, 0x71, 0x44, 0xD9, 0x80, + 0xD9, 0x92, 0x75, 0x44, 0xD9, 0x87, 0xD9, 0xB0, + 0x79, 0x44, 0xD9, 0x88, 0xD9, 0x94, 0xC9, 0x44, + 0xD9, 0x89, 0xD9, 0xB0, 0x79, 0x44, 0xD9, 0x8A, + 0xD9, 0x94, 0xC9, 0x44, 0xDB, 0x92, 0xD9, 0x94, + 0xC9, 0x44, 0xDB, 0x95, 0xD9, 0x94, 0xC9, 0x45, + 0x20, 0xCC, 0x88, 0xCC, 0x80, 0xCA, 0x45, 0x20, + // Bytes 4480 - 44bf + 0xCC, 0x88, 0xCC, 0x81, 0xCA, 0x45, 0x20, 0xCC, + 0x88, 0xCD, 0x82, 0xCA, 0x45, 0x20, 0xCC, 0x93, + 0xCC, 0x80, 0xCA, 0x45, 0x20, 0xCC, 0x93, 0xCC, + 0x81, 0xCA, 0x45, 0x20, 0xCC, 0x93, 0xCD, 0x82, + 0xCA, 0x45, 0x20, 0xCC, 0x94, 0xCC, 0x80, 0xCA, + 0x45, 0x20, 0xCC, 0x94, 0xCC, 0x81, 0xCA, 0x45, + 0x20, 0xCC, 0x94, 0xCD, 0x82, 0xCA, 0x45, 0x20, + 0xD9, 0x8C, 0xD9, 0x91, 0x72, 0x45, 0x20, 0xD9, + // Bytes 44c0 - 44ff + 0x8D, 0xD9, 0x91, 0x72, 0x45, 0x20, 0xD9, 0x8E, + 0xD9, 0x91, 0x72, 0x45, 0x20, 0xD9, 0x8F, 0xD9, + 0x91, 0x72, 0x45, 0x20, 0xD9, 0x90, 0xD9, 0x91, + 0x72, 0x45, 0x20, 0xD9, 0x91, 0xD9, 0xB0, 0x7A, + 0x45, 0xE2, 0xAB, 0x9D, 0xCC, 0xB8, 0x05, 0x46, + 0xCE, 0xB9, 0xCC, 0x88, 0xCC, 0x81, 0xCA, 0x46, + 0xCF, 0x85, 0xCC, 0x88, 0xCC, 0x81, 0xCA, 0x46, + 0xD7, 0xA9, 0xD6, 0xBC, 0xD7, 0x81, 0x4E, 0x46, + // Bytes 4500 - 453f + 0xD7, 0xA9, 0xD6, 0xBC, 0xD7, 0x82, 0x52, 0x46, + 0xD9, 0x80, 0xD9, 0x8E, 0xD9, 0x91, 0x72, 0x46, + 0xD9, 0x80, 0xD9, 0x8F, 0xD9, 0x91, 0x72, 0x46, + 0xD9, 0x80, 0xD9, 0x90, 0xD9, 0x91, 0x72, 0x46, + 0xE0, 0xA4, 0x95, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + 0xE0, 0xA4, 0x96, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + 0xE0, 0xA4, 0x97, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + 0xE0, 0xA4, 0x9C, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + // Bytes 4540 - 457f + 0xE0, 0xA4, 0xA1, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + 0xE0, 0xA4, 0xA2, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + 0xE0, 0xA4, 0xAB, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + 0xE0, 0xA4, 0xAF, 0xE0, 0xA4, 0xBC, 0x09, 0x46, + 0xE0, 0xA6, 0xA1, 0xE0, 0xA6, 0xBC, 0x09, 0x46, + 0xE0, 0xA6, 0xA2, 0xE0, 0xA6, 0xBC, 0x09, 0x46, + 0xE0, 0xA6, 0xAF, 0xE0, 0xA6, 0xBC, 0x09, 0x46, + 0xE0, 0xA8, 0x96, 0xE0, 0xA8, 0xBC, 0x09, 0x46, + // Bytes 4580 - 45bf + 0xE0, 0xA8, 0x97, 0xE0, 0xA8, 0xBC, 0x09, 0x46, + 0xE0, 0xA8, 0x9C, 0xE0, 0xA8, 0xBC, 0x09, 0x46, + 0xE0, 0xA8, 0xAB, 0xE0, 0xA8, 0xBC, 0x09, 0x46, + 0xE0, 0xA8, 0xB2, 0xE0, 0xA8, 0xBC, 0x09, 0x46, + 0xE0, 0xA8, 0xB8, 0xE0, 0xA8, 0xBC, 0x09, 0x46, + 0xE0, 0xAC, 0xA1, 0xE0, 0xAC, 0xBC, 0x09, 0x46, + 0xE0, 0xAC, 0xA2, 0xE0, 0xAC, 0xBC, 0x09, 0x46, + 0xE0, 0xBE, 0xB2, 0xE0, 0xBE, 0x80, 0x9D, 0x46, + // Bytes 45c0 - 45ff + 0xE0, 0xBE, 0xB3, 0xE0, 0xBE, 0x80, 0x9D, 0x46, + 0xE3, 0x83, 0x86, 0xE3, 0x82, 0x99, 0x0D, 0x48, + 0xF0, 0x9D, 0x85, 0x97, 0xF0, 0x9D, 0x85, 0xA5, + 0xAD, 0x48, 0xF0, 0x9D, 0x85, 0x98, 0xF0, 0x9D, + 0x85, 0xA5, 0xAD, 0x48, 0xF0, 0x9D, 0x86, 0xB9, + 0xF0, 0x9D, 0x85, 0xA5, 0xAD, 0x48, 0xF0, 0x9D, + 0x86, 0xBA, 0xF0, 0x9D, 0x85, 0xA5, 0xAD, 0x49, + 0xE0, 0xBE, 0xB2, 0xE0, 0xBD, 0xB1, 0xE0, 0xBE, + // Bytes 4600 - 463f + 0x80, 0x9E, 0x49, 0xE0, 0xBE, 0xB3, 0xE0, 0xBD, + 0xB1, 0xE0, 0xBE, 0x80, 0x9E, 0x4C, 0xF0, 0x9D, + 0x85, 0x98, 0xF0, 0x9D, 0x85, 0xA5, 0xF0, 0x9D, + 0x85, 0xAE, 0xAE, 0x4C, 0xF0, 0x9D, 0x85, 0x98, + 0xF0, 0x9D, 0x85, 0xA5, 0xF0, 0x9D, 0x85, 0xAF, + 0xAE, 0x4C, 0xF0, 0x9D, 0x85, 0x98, 0xF0, 0x9D, + 0x85, 0xA5, 0xF0, 0x9D, 0x85, 0xB0, 0xAE, 0x4C, + 0xF0, 0x9D, 0x85, 0x98, 0xF0, 0x9D, 0x85, 0xA5, + // Bytes 4640 - 467f + 0xF0, 0x9D, 0x85, 0xB1, 0xAE, 0x4C, 0xF0, 0x9D, + 0x85, 0x98, 0xF0, 0x9D, 0x85, 0xA5, 0xF0, 0x9D, + 0x85, 0xB2, 0xAE, 0x4C, 0xF0, 0x9D, 0x86, 0xB9, + 0xF0, 0x9D, 0x85, 0xA5, 0xF0, 0x9D, 0x85, 0xAE, + 0xAE, 0x4C, 0xF0, 0x9D, 0x86, 0xB9, 0xF0, 0x9D, + 0x85, 0xA5, 0xF0, 0x9D, 0x85, 0xAF, 0xAE, 0x4C, + 0xF0, 0x9D, 0x86, 0xBA, 0xF0, 0x9D, 0x85, 0xA5, + 0xF0, 0x9D, 0x85, 0xAE, 0xAE, 0x4C, 0xF0, 0x9D, + // Bytes 4680 - 46bf + 0x86, 0xBA, 0xF0, 0x9D, 0x85, 0xA5, 0xF0, 0x9D, + 0x85, 0xAF, 0xAE, 0x83, 0x41, 0xCC, 0x82, 0xC9, + 0x83, 0x41, 0xCC, 0x86, 0xC9, 0x83, 0x41, 0xCC, + 0x87, 0xC9, 0x83, 0x41, 0xCC, 0x88, 0xC9, 0x83, + 0x41, 0xCC, 0x8A, 0xC9, 0x83, 0x41, 0xCC, 0xA3, + 0xB5, 0x83, 0x43, 0xCC, 0xA7, 0xA5, 0x83, 0x45, + 0xCC, 0x82, 0xC9, 0x83, 0x45, 0xCC, 0x84, 0xC9, + 0x83, 0x45, 0xCC, 0xA3, 0xB5, 0x83, 0x45, 0xCC, + // Bytes 46c0 - 46ff + 0xA7, 0xA5, 0x83, 0x49, 0xCC, 0x88, 0xC9, 0x83, + 0x4C, 0xCC, 0xA3, 0xB5, 0x83, 0x4F, 0xCC, 0x82, + 0xC9, 0x83, 0x4F, 0xCC, 0x83, 0xC9, 0x83, 0x4F, + 0xCC, 0x84, 0xC9, 0x83, 0x4F, 0xCC, 0x87, 0xC9, + 0x83, 0x4F, 0xCC, 0x88, 0xC9, 0x83, 0x4F, 0xCC, + 0x9B, 0xAD, 0x83, 0x4F, 0xCC, 0xA3, 0xB5, 0x83, + 0x4F, 0xCC, 0xA8, 0xA5, 0x83, 0x52, 0xCC, 0xA3, + 0xB5, 0x83, 0x53, 0xCC, 0x81, 0xC9, 0x83, 0x53, + // Bytes 4700 - 473f + 0xCC, 0x8C, 0xC9, 0x83, 0x53, 0xCC, 0xA3, 0xB5, + 0x83, 0x55, 0xCC, 0x83, 0xC9, 0x83, 0x55, 0xCC, + 0x84, 0xC9, 0x83, 0x55, 0xCC, 0x88, 0xC9, 0x83, + 0x55, 0xCC, 0x9B, 0xAD, 0x83, 0x61, 0xCC, 0x82, + 0xC9, 0x83, 0x61, 0xCC, 0x86, 0xC9, 0x83, 0x61, + 0xCC, 0x87, 0xC9, 0x83, 0x61, 0xCC, 0x88, 0xC9, + 0x83, 0x61, 0xCC, 0x8A, 0xC9, 0x83, 0x61, 0xCC, + 0xA3, 0xB5, 0x83, 0x63, 0xCC, 0xA7, 0xA5, 0x83, + // Bytes 4740 - 477f + 0x65, 0xCC, 0x82, 0xC9, 0x83, 0x65, 0xCC, 0x84, + 0xC9, 0x83, 0x65, 0xCC, 0xA3, 0xB5, 0x83, 0x65, + 0xCC, 0xA7, 0xA5, 0x83, 0x69, 0xCC, 0x88, 0xC9, + 0x83, 0x6C, 0xCC, 0xA3, 0xB5, 0x83, 0x6F, 0xCC, + 0x82, 0xC9, 0x83, 0x6F, 0xCC, 0x83, 0xC9, 0x83, + 0x6F, 0xCC, 0x84, 0xC9, 0x83, 0x6F, 0xCC, 0x87, + 0xC9, 0x83, 0x6F, 0xCC, 0x88, 0xC9, 0x83, 0x6F, + 0xCC, 0x9B, 0xAD, 0x83, 0x6F, 0xCC, 0xA3, 0xB5, + // Bytes 4780 - 47bf + 0x83, 0x6F, 0xCC, 0xA8, 0xA5, 0x83, 0x72, 0xCC, + 0xA3, 0xB5, 0x83, 0x73, 0xCC, 0x81, 0xC9, 0x83, + 0x73, 0xCC, 0x8C, 0xC9, 0x83, 0x73, 0xCC, 0xA3, + 0xB5, 0x83, 0x75, 0xCC, 0x83, 0xC9, 0x83, 0x75, + 0xCC, 0x84, 0xC9, 0x83, 0x75, 0xCC, 0x88, 0xC9, + 0x83, 0x75, 0xCC, 0x9B, 0xAD, 0x84, 0xCE, 0x91, + 0xCC, 0x93, 0xC9, 0x84, 0xCE, 0x91, 0xCC, 0x94, + 0xC9, 0x84, 0xCE, 0x95, 0xCC, 0x93, 0xC9, 0x84, + // Bytes 47c0 - 47ff + 0xCE, 0x95, 0xCC, 0x94, 0xC9, 0x84, 0xCE, 0x97, + 0xCC, 0x93, 0xC9, 0x84, 0xCE, 0x97, 0xCC, 0x94, + 0xC9, 0x84, 0xCE, 0x99, 0xCC, 0x93, 0xC9, 0x84, + 0xCE, 0x99, 0xCC, 0x94, 0xC9, 0x84, 0xCE, 0x9F, + 0xCC, 0x93, 0xC9, 0x84, 0xCE, 0x9F, 0xCC, 0x94, + 0xC9, 0x84, 0xCE, 0xA5, 0xCC, 0x94, 0xC9, 0x84, + 0xCE, 0xA9, 0xCC, 0x93, 0xC9, 0x84, 0xCE, 0xA9, + 0xCC, 0x94, 0xC9, 0x84, 0xCE, 0xB1, 0xCC, 0x80, + // Bytes 4800 - 483f + 0xC9, 0x84, 0xCE, 0xB1, 0xCC, 0x81, 0xC9, 0x84, + 0xCE, 0xB1, 0xCC, 0x93, 0xC9, 0x84, 0xCE, 0xB1, + 0xCC, 0x94, 0xC9, 0x84, 0xCE, 0xB1, 0xCD, 0x82, + 0xC9, 0x84, 0xCE, 0xB5, 0xCC, 0x93, 0xC9, 0x84, + 0xCE, 0xB5, 0xCC, 0x94, 0xC9, 0x84, 0xCE, 0xB7, + 0xCC, 0x80, 0xC9, 0x84, 0xCE, 0xB7, 0xCC, 0x81, + 0xC9, 0x84, 0xCE, 0xB7, 0xCC, 0x93, 0xC9, 0x84, + 0xCE, 0xB7, 0xCC, 0x94, 0xC9, 0x84, 0xCE, 0xB7, + // Bytes 4840 - 487f + 0xCD, 0x82, 0xC9, 0x84, 0xCE, 0xB9, 0xCC, 0x88, + 0xC9, 0x84, 0xCE, 0xB9, 0xCC, 0x93, 0xC9, 0x84, + 0xCE, 0xB9, 0xCC, 0x94, 0xC9, 0x84, 0xCE, 0xBF, + 0xCC, 0x93, 0xC9, 0x84, 0xCE, 0xBF, 0xCC, 0x94, + 0xC9, 0x84, 0xCF, 0x85, 0xCC, 0x88, 0xC9, 0x84, + 0xCF, 0x85, 0xCC, 0x93, 0xC9, 0x84, 0xCF, 0x85, + 0xCC, 0x94, 0xC9, 0x84, 0xCF, 0x89, 0xCC, 0x80, + 0xC9, 0x84, 0xCF, 0x89, 0xCC, 0x81, 0xC9, 0x84, + // Bytes 4880 - 48bf + 0xCF, 0x89, 0xCC, 0x93, 0xC9, 0x84, 0xCF, 0x89, + 0xCC, 0x94, 0xC9, 0x84, 0xCF, 0x89, 0xCD, 0x82, + 0xC9, 0x86, 0xCE, 0x91, 0xCC, 0x93, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0x91, 0xCC, 0x93, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0x91, 0xCC, 0x93, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0x91, 0xCC, 0x94, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0x91, 0xCC, 0x94, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0x91, 0xCC, 0x94, 0xCD, 0x82, + // Bytes 48c0 - 48ff + 0xCA, 0x86, 0xCE, 0x97, 0xCC, 0x93, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0x97, 0xCC, 0x93, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0x97, 0xCC, 0x93, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0x97, 0xCC, 0x94, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0x97, 0xCC, 0x94, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0x97, 0xCC, 0x94, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0xA9, 0xCC, 0x93, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0xA9, 0xCC, 0x93, 0xCC, 0x81, + // Bytes 4900 - 493f + 0xCA, 0x86, 0xCE, 0xA9, 0xCC, 0x93, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0xA9, 0xCC, 0x94, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0xA9, 0xCC, 0x94, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0xA9, 0xCC, 0x94, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0xB1, 0xCC, 0x93, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0xB1, 0xCC, 0x93, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0xB1, 0xCC, 0x93, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0xB1, 0xCC, 0x94, 0xCC, 0x80, + // Bytes 4940 - 497f + 0xCA, 0x86, 0xCE, 0xB1, 0xCC, 0x94, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0xB1, 0xCC, 0x94, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0xB7, 0xCC, 0x93, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0xB7, 0xCC, 0x93, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0xB7, 0xCC, 0x93, 0xCD, 0x82, + 0xCA, 0x86, 0xCE, 0xB7, 0xCC, 0x94, 0xCC, 0x80, + 0xCA, 0x86, 0xCE, 0xB7, 0xCC, 0x94, 0xCC, 0x81, + 0xCA, 0x86, 0xCE, 0xB7, 0xCC, 0x94, 0xCD, 0x82, + // Bytes 4980 - 49bf + 0xCA, 0x86, 0xCF, 0x89, 0xCC, 0x93, 0xCC, 0x80, + 0xCA, 0x86, 0xCF, 0x89, 0xCC, 0x93, 0xCC, 0x81, + 0xCA, 0x86, 0xCF, 0x89, 0xCC, 0x93, 0xCD, 0x82, + 0xCA, 0x86, 0xCF, 0x89, 0xCC, 0x94, 0xCC, 0x80, + 0xCA, 0x86, 0xCF, 0x89, 0xCC, 0x94, 0xCC, 0x81, + 0xCA, 0x86, 0xCF, 0x89, 0xCC, 0x94, 0xCD, 0x82, + 0xCA, 0x42, 0xCC, 0x80, 0xC9, 0x32, 0x42, 0xCC, + 0x81, 0xC9, 0x32, 0x42, 0xCC, 0x93, 0xC9, 0x32, + // Bytes 49c0 - 49ff + 0x43, 0xE1, 0x85, 0xA1, 0x01, 0x00, 0x43, 0xE1, + 0x85, 0xA2, 0x01, 0x00, 0x43, 0xE1, 0x85, 0xA3, + 0x01, 0x00, 0x43, 0xE1, 0x85, 0xA4, 0x01, 0x00, + 0x43, 0xE1, 0x85, 0xA5, 0x01, 0x00, 0x43, 0xE1, + 0x85, 0xA6, 0x01, 0x00, 0x43, 0xE1, 0x85, 0xA7, + 0x01, 0x00, 0x43, 0xE1, 0x85, 0xA8, 0x01, 0x00, + 0x43, 0xE1, 0x85, 0xA9, 0x01, 0x00, 0x43, 0xE1, + 0x85, 0xAA, 0x01, 0x00, 0x43, 0xE1, 0x85, 0xAB, + // Bytes 4a00 - 4a3f + 0x01, 0x00, 0x43, 0xE1, 0x85, 0xAC, 0x01, 0x00, + 0x43, 0xE1, 0x85, 0xAD, 0x01, 0x00, 0x43, 0xE1, + 0x85, 0xAE, 0x01, 0x00, 0x43, 0xE1, 0x85, 0xAF, + 0x01, 0x00, 0x43, 0xE1, 0x85, 0xB0, 0x01, 0x00, + 0x43, 0xE1, 0x85, 0xB1, 0x01, 0x00, 0x43, 0xE1, + 0x85, 0xB2, 0x01, 0x00, 0x43, 0xE1, 0x85, 0xB3, + 0x01, 0x00, 0x43, 0xE1, 0x85, 0xB4, 0x01, 0x00, + 0x43, 0xE1, 0x85, 0xB5, 0x01, 0x00, 0x43, 0xE1, + // Bytes 4a40 - 4a7f + 0x86, 0xAA, 0x01, 0x00, 0x43, 0xE1, 0x86, 0xAC, + 0x01, 0x00, 0x43, 0xE1, 0x86, 0xAD, 0x01, 0x00, + 0x43, 0xE1, 0x86, 0xB0, 0x01, 0x00, 0x43, 0xE1, + 0x86, 0xB1, 0x01, 0x00, 0x43, 0xE1, 0x86, 0xB2, + 0x01, 0x00, 0x43, 0xE1, 0x86, 0xB3, 0x01, 0x00, + 0x43, 0xE1, 0x86, 0xB4, 0x01, 0x00, 0x43, 0xE1, + 0x86, 0xB5, 0x01, 0x00, 0x44, 0xCC, 0x88, 0xCC, + 0x81, 0xCA, 0x32, 0x43, 0xE3, 0x82, 0x99, 0x0D, + // Bytes 4a80 - 4abf + 0x03, 0x43, 0xE3, 0x82, 0x9A, 0x0D, 0x03, 0x46, + 0xE0, 0xBD, 0xB1, 0xE0, 0xBD, 0xB2, 0x9E, 0x26, + 0x46, 0xE0, 0xBD, 0xB1, 0xE0, 0xBD, 0xB4, 0xA2, + 0x26, 0x46, 0xE0, 0xBD, 0xB1, 0xE0, 0xBE, 0x80, + 0x9E, 0x26, 0x00, 0x01, +} + +// lookup returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *nfcTrie) lookup(s []byte) (v uint16, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return nfcValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := nfcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := nfcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := nfcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = nfcIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *nfcTrie) lookupUnsafe(s []byte) uint16 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return nfcValues[c0] + } + i := nfcIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = nfcIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = nfcIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// lookupString returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *nfcTrie) lookupString(s string) (v uint16, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return nfcValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := nfcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := nfcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := nfcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = nfcIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupStringUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *nfcTrie) lookupStringUnsafe(s string) uint16 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return nfcValues[c0] + } + i := nfcIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = nfcIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = nfcIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// nfcTrie. Total size: 10610 bytes (10.36 KiB). Checksum: 95e8869a9f81e5e6. +type nfcTrie struct{} + +func newNfcTrie(i int) *nfcTrie { + return &nfcTrie{} +} + +// lookupValue determines the type of block n and looks up the value for b. +func (t *nfcTrie) lookupValue(n uint32, b byte) uint16 { + switch { + case n < 46: + return uint16(nfcValues[n<<6+uint32(b)]) + default: + n -= 46 + return uint16(nfcSparse.lookup(n, b)) + } +} + +// nfcValues: 48 blocks, 3072 entries, 6144 bytes +// The third block is the zero block. +var nfcValues = [3072]uint16{ + // Block 0x0, offset 0x0 + 0x3c: 0xa000, 0x3d: 0xa000, 0x3e: 0xa000, + // Block 0x1, offset 0x40 + 0x41: 0xa000, 0x42: 0xa000, 0x43: 0xa000, 0x44: 0xa000, 0x45: 0xa000, + 0x46: 0xa000, 0x47: 0xa000, 0x48: 0xa000, 0x49: 0xa000, 0x4a: 0xa000, 0x4b: 0xa000, + 0x4c: 0xa000, 0x4d: 0xa000, 0x4e: 0xa000, 0x4f: 0xa000, 0x50: 0xa000, + 0x52: 0xa000, 0x53: 0xa000, 0x54: 0xa000, 0x55: 0xa000, 0x56: 0xa000, 0x57: 0xa000, + 0x58: 0xa000, 0x59: 0xa000, 0x5a: 0xa000, + 0x61: 0xa000, 0x62: 0xa000, 0x63: 0xa000, + 0x64: 0xa000, 0x65: 0xa000, 0x66: 0xa000, 0x67: 0xa000, 0x68: 0xa000, 0x69: 0xa000, + 0x6a: 0xa000, 0x6b: 0xa000, 0x6c: 0xa000, 0x6d: 0xa000, 0x6e: 0xa000, 0x6f: 0xa000, + 0x70: 0xa000, 0x72: 0xa000, 0x73: 0xa000, 0x74: 0xa000, 0x75: 0xa000, + 0x76: 0xa000, 0x77: 0xa000, 0x78: 0xa000, 0x79: 0xa000, 0x7a: 0xa000, + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xc0: 0x2f72, 0xc1: 0x2f77, 0xc2: 0x468b, 0xc3: 0x2f7c, 0xc4: 0x469a, 0xc5: 0x469f, + 0xc6: 0xa000, 0xc7: 0x46a9, 0xc8: 0x2fe5, 0xc9: 0x2fea, 0xca: 0x46ae, 0xcb: 0x2ffe, + 0xcc: 0x3071, 0xcd: 0x3076, 0xce: 0x307b, 0xcf: 0x46c2, 0xd1: 0x3107, + 0xd2: 0x312a, 0xd3: 0x312f, 0xd4: 0x46cc, 0xd5: 0x46d1, 0xd6: 0x46e0, + 0xd8: 0xa000, 0xd9: 0x31b6, 0xda: 0x31bb, 0xdb: 0x31c0, 0xdc: 0x4712, 0xdd: 0x3238, + 0xe0: 0x327e, 0xe1: 0x3283, 0xe2: 0x471c, 0xe3: 0x3288, + 0xe4: 0x472b, 0xe5: 0x4730, 0xe6: 0xa000, 0xe7: 0x473a, 0xe8: 0x32f1, 0xe9: 0x32f6, + 0xea: 0x473f, 0xeb: 0x330a, 0xec: 0x3382, 0xed: 0x3387, 0xee: 0x338c, 0xef: 0x4753, + 0xf1: 0x3418, 0xf2: 0x343b, 0xf3: 0x3440, 0xf4: 0x475d, 0xf5: 0x4762, + 0xf6: 0x4771, 0xf8: 0xa000, 0xf9: 0x34cc, 0xfa: 0x34d1, 0xfb: 0x34d6, + 0xfc: 0x47a3, 0xfd: 0x3553, 0xff: 0x356c, + // Block 0x4, offset 0x100 + 0x100: 0x2f81, 0x101: 0x328d, 0x102: 0x4690, 0x103: 0x4721, 0x104: 0x2f9f, 0x105: 0x32ab, + 0x106: 0x2fb3, 0x107: 0x32bf, 0x108: 0x2fb8, 0x109: 0x32c4, 0x10a: 0x2fbd, 0x10b: 0x32c9, + 0x10c: 0x2fc2, 0x10d: 0x32ce, 0x10e: 0x2fcc, 0x10f: 0x32d8, + 0x112: 0x46b3, 0x113: 0x4744, 0x114: 0x2ff4, 0x115: 0x3300, 0x116: 0x2ff9, 0x117: 0x3305, + 0x118: 0x3017, 0x119: 0x3323, 0x11a: 0x3008, 0x11b: 0x3314, 0x11c: 0x3030, 0x11d: 0x333c, + 0x11e: 0x303a, 0x11f: 0x3346, 0x120: 0x303f, 0x121: 0x334b, 0x122: 0x3049, 0x123: 0x3355, + 0x124: 0x304e, 0x125: 0x335a, 0x128: 0x3080, 0x129: 0x3391, + 0x12a: 0x3085, 0x12b: 0x3396, 0x12c: 0x308a, 0x12d: 0x339b, 0x12e: 0x30ad, 0x12f: 0x33b9, + 0x130: 0x308f, 0x134: 0x30b7, 0x135: 0x33c3, + 0x136: 0x30cb, 0x137: 0x33dc, 0x139: 0x30d5, 0x13a: 0x33e6, 0x13b: 0x30df, + 0x13c: 0x33f0, 0x13d: 0x30da, 0x13e: 0x33eb, + // Block 0x5, offset 0x140 + 0x143: 0x3102, 0x144: 0x3413, 0x145: 0x311b, + 0x146: 0x342c, 0x147: 0x3111, 0x148: 0x3422, + 0x14c: 0x46d6, 0x14d: 0x4767, 0x14e: 0x3134, 0x14f: 0x3445, 0x150: 0x313e, 0x151: 0x344f, + 0x154: 0x315c, 0x155: 0x346d, 0x156: 0x3175, 0x157: 0x3486, + 0x158: 0x3166, 0x159: 0x3477, 0x15a: 0x46f9, 0x15b: 0x478a, 0x15c: 0x317f, 0x15d: 0x3490, + 0x15e: 0x318e, 0x15f: 0x349f, 0x160: 0x46fe, 0x161: 0x478f, 0x162: 0x31a7, 0x163: 0x34bd, + 0x164: 0x3198, 0x165: 0x34ae, 0x168: 0x4708, 0x169: 0x4799, + 0x16a: 0x470d, 0x16b: 0x479e, 0x16c: 0x31c5, 0x16d: 0x34db, 0x16e: 0x31cf, 0x16f: 0x34e5, + 0x170: 0x31d4, 0x171: 0x34ea, 0x172: 0x31f2, 0x173: 0x3508, 0x174: 0x3215, 0x175: 0x352b, + 0x176: 0x323d, 0x177: 0x3558, 0x178: 0x3251, 0x179: 0x3260, 0x17a: 0x3580, 0x17b: 0x326a, + 0x17c: 0x358a, 0x17d: 0x326f, 0x17e: 0x358f, 0x17f: 0xa000, + // Block 0x6, offset 0x180 + 0x184: 0x8100, 0x185: 0x8100, + 0x186: 0x8100, + 0x18d: 0x2f8b, 0x18e: 0x3297, 0x18f: 0x3099, 0x190: 0x33a5, 0x191: 0x3143, + 0x192: 0x3454, 0x193: 0x31d9, 0x194: 0x34ef, 0x195: 0x39d2, 0x196: 0x3b61, 0x197: 0x39cb, + 0x198: 0x3b5a, 0x199: 0x39d9, 0x19a: 0x3b68, 0x19b: 0x39c4, 0x19c: 0x3b53, + 0x19e: 0x38b3, 0x19f: 0x3a42, 0x1a0: 0x38ac, 0x1a1: 0x3a3b, 0x1a2: 0x35b6, 0x1a3: 0x35c8, + 0x1a6: 0x3044, 0x1a7: 0x3350, 0x1a8: 0x30c1, 0x1a9: 0x33d2, + 0x1aa: 0x46ef, 0x1ab: 0x4780, 0x1ac: 0x3993, 0x1ad: 0x3b22, 0x1ae: 0x35da, 0x1af: 0x35e0, + 0x1b0: 0x33c8, 0x1b4: 0x302b, 0x1b5: 0x3337, + 0x1b8: 0x30fd, 0x1b9: 0x340e, 0x1ba: 0x38ba, 0x1bb: 0x3a49, + 0x1bc: 0x35b0, 0x1bd: 0x35c2, 0x1be: 0x35bc, 0x1bf: 0x35ce, + // Block 0x7, offset 0x1c0 + 0x1c0: 0x2f90, 0x1c1: 0x329c, 0x1c2: 0x2f95, 0x1c3: 0x32a1, 0x1c4: 0x300d, 0x1c5: 0x3319, + 0x1c6: 0x3012, 0x1c7: 0x331e, 0x1c8: 0x309e, 0x1c9: 0x33aa, 0x1ca: 0x30a3, 0x1cb: 0x33af, + 0x1cc: 0x3148, 0x1cd: 0x3459, 0x1ce: 0x314d, 0x1cf: 0x345e, 0x1d0: 0x316b, 0x1d1: 0x347c, + 0x1d2: 0x3170, 0x1d3: 0x3481, 0x1d4: 0x31de, 0x1d5: 0x34f4, 0x1d6: 0x31e3, 0x1d7: 0x34f9, + 0x1d8: 0x3189, 0x1d9: 0x349a, 0x1da: 0x31a2, 0x1db: 0x34b8, + 0x1de: 0x305d, 0x1df: 0x3369, + 0x1e6: 0x4695, 0x1e7: 0x4726, 0x1e8: 0x46bd, 0x1e9: 0x474e, + 0x1ea: 0x3962, 0x1eb: 0x3af1, 0x1ec: 0x393f, 0x1ed: 0x3ace, 0x1ee: 0x46db, 0x1ef: 0x476c, + 0x1f0: 0x395b, 0x1f1: 0x3aea, 0x1f2: 0x3247, 0x1f3: 0x3562, + // Block 0x8, offset 0x200 + 0x200: 0x9932, 0x201: 0x9932, 0x202: 0x9932, 0x203: 0x9932, 0x204: 0x9932, 0x205: 0x8132, + 0x206: 0x9932, 0x207: 0x9932, 0x208: 0x9932, 0x209: 0x9932, 0x20a: 0x9932, 0x20b: 0x9932, + 0x20c: 0x9932, 0x20d: 0x8132, 0x20e: 0x8132, 0x20f: 0x9932, 0x210: 0x8132, 0x211: 0x9932, + 0x212: 0x8132, 0x213: 0x9932, 0x214: 0x9932, 0x215: 0x8133, 0x216: 0x812d, 0x217: 0x812d, + 0x218: 0x812d, 0x219: 0x812d, 0x21a: 0x8133, 0x21b: 0x992b, 0x21c: 0x812d, 0x21d: 0x812d, + 0x21e: 0x812d, 0x21f: 0x812d, 0x220: 0x812d, 0x221: 0x8129, 0x222: 0x8129, 0x223: 0x992d, + 0x224: 0x992d, 0x225: 0x992d, 0x226: 0x992d, 0x227: 0x9929, 0x228: 0x9929, 0x229: 0x812d, + 0x22a: 0x812d, 0x22b: 0x812d, 0x22c: 0x812d, 0x22d: 0x992d, 0x22e: 0x992d, 0x22f: 0x812d, + 0x230: 0x992d, 0x231: 0x992d, 0x232: 0x812d, 0x233: 0x812d, 0x234: 0x8101, 0x235: 0x8101, + 0x236: 0x8101, 0x237: 0x8101, 0x238: 0x9901, 0x239: 0x812d, 0x23a: 0x812d, 0x23b: 0x812d, + 0x23c: 0x812d, 0x23d: 0x8132, 0x23e: 0x8132, 0x23f: 0x8132, + // Block 0x9, offset 0x240 + 0x240: 0x49b1, 0x241: 0x49b6, 0x242: 0x9932, 0x243: 0x49bb, 0x244: 0x4a74, 0x245: 0x9936, + 0x246: 0x8132, 0x247: 0x812d, 0x248: 0x812d, 0x249: 0x812d, 0x24a: 0x8132, 0x24b: 0x8132, + 0x24c: 0x8132, 0x24d: 0x812d, 0x24e: 0x812d, 0x250: 0x8132, 0x251: 0x8132, + 0x252: 0x8132, 0x253: 0x812d, 0x254: 0x812d, 0x255: 0x812d, 0x256: 0x812d, 0x257: 0x8132, + 0x258: 0x8133, 0x259: 0x812d, 0x25a: 0x812d, 0x25b: 0x8132, 0x25c: 0x8134, 0x25d: 0x8135, + 0x25e: 0x8135, 0x25f: 0x8134, 0x260: 0x8135, 0x261: 0x8135, 0x262: 0x8134, 0x263: 0x8132, + 0x264: 0x8132, 0x265: 0x8132, 0x266: 0x8132, 0x267: 0x8132, 0x268: 0x8132, 0x269: 0x8132, + 0x26a: 0x8132, 0x26b: 0x8132, 0x26c: 0x8132, 0x26d: 0x8132, 0x26e: 0x8132, 0x26f: 0x8132, + 0x274: 0x0170, + 0x27a: 0x8100, + 0x27e: 0x0037, + // Block 0xa, offset 0x280 + 0x284: 0x8100, 0x285: 0x35a4, + 0x286: 0x35ec, 0x287: 0x00ce, 0x288: 0x360a, 0x289: 0x3616, 0x28a: 0x3628, + 0x28c: 0x3646, 0x28e: 0x3658, 0x28f: 0x3676, 0x290: 0x3e0b, 0x291: 0xa000, + 0x295: 0xa000, 0x297: 0xa000, + 0x299: 0xa000, + 0x29f: 0xa000, 0x2a1: 0xa000, + 0x2a5: 0xa000, 0x2a9: 0xa000, + 0x2aa: 0x363a, 0x2ab: 0x366a, 0x2ac: 0x4801, 0x2ad: 0x369a, 0x2ae: 0x482b, 0x2af: 0x36ac, + 0x2b0: 0x3e73, 0x2b1: 0xa000, 0x2b5: 0xa000, + 0x2b7: 0xa000, 0x2b9: 0xa000, + 0x2bf: 0xa000, + // Block 0xb, offset 0x2c0 + 0x2c0: 0x3724, 0x2c1: 0x3730, 0x2c3: 0x371e, + 0x2c6: 0xa000, 0x2c7: 0x370c, + 0x2cc: 0x3760, 0x2cd: 0x3748, 0x2ce: 0x3772, 0x2d0: 0xa000, + 0x2d3: 0xa000, 0x2d5: 0xa000, 0x2d6: 0xa000, 0x2d7: 0xa000, + 0x2d8: 0xa000, 0x2d9: 0x3754, 0x2da: 0xa000, + 0x2de: 0xa000, 0x2e3: 0xa000, + 0x2e7: 0xa000, + 0x2eb: 0xa000, 0x2ed: 0xa000, + 0x2f0: 0xa000, 0x2f3: 0xa000, 0x2f5: 0xa000, + 0x2f6: 0xa000, 0x2f7: 0xa000, 0x2f8: 0xa000, 0x2f9: 0x37d8, 0x2fa: 0xa000, + 0x2fe: 0xa000, + // Block 0xc, offset 0x300 + 0x301: 0x3736, 0x302: 0x37ba, + 0x310: 0x3712, 0x311: 0x3796, + 0x312: 0x3718, 0x313: 0x379c, 0x316: 0x372a, 0x317: 0x37ae, + 0x318: 0xa000, 0x319: 0xa000, 0x31a: 0x382c, 0x31b: 0x3832, 0x31c: 0x373c, 0x31d: 0x37c0, + 0x31e: 0x3742, 0x31f: 0x37c6, 0x322: 0x374e, 0x323: 0x37d2, + 0x324: 0x375a, 0x325: 0x37de, 0x326: 0x3766, 0x327: 0x37ea, 0x328: 0xa000, 0x329: 0xa000, + 0x32a: 0x3838, 0x32b: 0x383e, 0x32c: 0x3790, 0x32d: 0x3814, 0x32e: 0x376c, 0x32f: 0x37f0, + 0x330: 0x3778, 0x331: 0x37fc, 0x332: 0x377e, 0x333: 0x3802, 0x334: 0x3784, 0x335: 0x3808, + 0x338: 0x378a, 0x339: 0x380e, + // Block 0xd, offset 0x340 + 0x351: 0x812d, + 0x352: 0x8132, 0x353: 0x8132, 0x354: 0x8132, 0x355: 0x8132, 0x356: 0x812d, 0x357: 0x8132, + 0x358: 0x8132, 0x359: 0x8132, 0x35a: 0x812e, 0x35b: 0x812d, 0x35c: 0x8132, 0x35d: 0x8132, + 0x35e: 0x8132, 0x35f: 0x8132, 0x360: 0x8132, 0x361: 0x8132, 0x362: 0x812d, 0x363: 0x812d, + 0x364: 0x812d, 0x365: 0x812d, 0x366: 0x812d, 0x367: 0x812d, 0x368: 0x8132, 0x369: 0x8132, + 0x36a: 0x812d, 0x36b: 0x8132, 0x36c: 0x8132, 0x36d: 0x812e, 0x36e: 0x8131, 0x36f: 0x8132, + 0x370: 0x8105, 0x371: 0x8106, 0x372: 0x8107, 0x373: 0x8108, 0x374: 0x8109, 0x375: 0x810a, + 0x376: 0x810b, 0x377: 0x810c, 0x378: 0x810d, 0x379: 0x810e, 0x37a: 0x810e, 0x37b: 0x810f, + 0x37c: 0x8110, 0x37d: 0x8111, 0x37f: 0x8112, + // Block 0xe, offset 0x380 + 0x388: 0xa000, 0x38a: 0xa000, 0x38b: 0x8116, + 0x38c: 0x8117, 0x38d: 0x8118, 0x38e: 0x8119, 0x38f: 0x811a, 0x390: 0x811b, 0x391: 0x811c, + 0x392: 0x811d, 0x393: 0x9932, 0x394: 0x9932, 0x395: 0x992d, 0x396: 0x812d, 0x397: 0x8132, + 0x398: 0x8132, 0x399: 0x8132, 0x39a: 0x8132, 0x39b: 0x8132, 0x39c: 0x812d, 0x39d: 0x8132, + 0x39e: 0x8132, 0x39f: 0x812d, + 0x3b0: 0x811e, + // Block 0xf, offset 0x3c0 + 0x3d3: 0x812d, 0x3d4: 0x8132, 0x3d5: 0x8132, 0x3d6: 0x8132, 0x3d7: 0x8132, + 0x3d8: 0x8132, 0x3d9: 0x8132, 0x3da: 0x8132, 0x3db: 0x8132, 0x3dc: 0x8132, 0x3dd: 0x8132, + 0x3de: 0x8132, 0x3df: 0x8132, 0x3e0: 0x8132, 0x3e1: 0x8132, 0x3e3: 0x812d, + 0x3e4: 0x8132, 0x3e5: 0x8132, 0x3e6: 0x812d, 0x3e7: 0x8132, 0x3e8: 0x8132, 0x3e9: 0x812d, + 0x3ea: 0x8132, 0x3eb: 0x8132, 0x3ec: 0x8132, 0x3ed: 0x812d, 0x3ee: 0x812d, 0x3ef: 0x812d, + 0x3f0: 0x8116, 0x3f1: 0x8117, 0x3f2: 0x8118, 0x3f3: 0x8132, 0x3f4: 0x8132, 0x3f5: 0x8132, + 0x3f6: 0x812d, 0x3f7: 0x8132, 0x3f8: 0x8132, 0x3f9: 0x812d, 0x3fa: 0x812d, 0x3fb: 0x8132, + 0x3fc: 0x8132, 0x3fd: 0x8132, 0x3fe: 0x8132, 0x3ff: 0x8132, + // Block 0x10, offset 0x400 + 0x405: 0xa000, + 0x406: 0x2d29, 0x407: 0xa000, 0x408: 0x2d31, 0x409: 0xa000, 0x40a: 0x2d39, 0x40b: 0xa000, + 0x40c: 0x2d41, 0x40d: 0xa000, 0x40e: 0x2d49, 0x411: 0xa000, + 0x412: 0x2d51, + 0x434: 0x8102, 0x435: 0x9900, + 0x43a: 0xa000, 0x43b: 0x2d59, + 0x43c: 0xa000, 0x43d: 0x2d61, 0x43e: 0xa000, 0x43f: 0xa000, + // Block 0x11, offset 0x440 + 0x440: 0x8132, 0x441: 0x8132, 0x442: 0x812d, 0x443: 0x8132, 0x444: 0x8132, 0x445: 0x8132, + 0x446: 0x8132, 0x447: 0x8132, 0x448: 0x8132, 0x449: 0x8132, 0x44a: 0x812d, 0x44b: 0x8132, + 0x44c: 0x8132, 0x44d: 0x8135, 0x44e: 0x812a, 0x44f: 0x812d, 0x450: 0x8129, 0x451: 0x8132, + 0x452: 0x8132, 0x453: 0x8132, 0x454: 0x8132, 0x455: 0x8132, 0x456: 0x8132, 0x457: 0x8132, + 0x458: 0x8132, 0x459: 0x8132, 0x45a: 0x8132, 0x45b: 0x8132, 0x45c: 0x8132, 0x45d: 0x8132, + 0x45e: 0x8132, 0x45f: 0x8132, 0x460: 0x8132, 0x461: 0x8132, 0x462: 0x8132, 0x463: 0x8132, + 0x464: 0x8132, 0x465: 0x8132, 0x466: 0x8132, 0x467: 0x8132, 0x468: 0x8132, 0x469: 0x8132, + 0x46a: 0x8132, 0x46b: 0x8132, 0x46c: 0x8132, 0x46d: 0x8132, 0x46e: 0x8132, 0x46f: 0x8132, + 0x470: 0x8132, 0x471: 0x8132, 0x472: 0x8132, 0x473: 0x8132, 0x474: 0x8132, 0x475: 0x8132, + 0x476: 0x8133, 0x477: 0x8131, 0x478: 0x8131, 0x479: 0x812d, 0x47b: 0x8132, + 0x47c: 0x8134, 0x47d: 0x812d, 0x47e: 0x8132, 0x47f: 0x812d, + // Block 0x12, offset 0x480 + 0x480: 0x2f9a, 0x481: 0x32a6, 0x482: 0x2fa4, 0x483: 0x32b0, 0x484: 0x2fa9, 0x485: 0x32b5, + 0x486: 0x2fae, 0x487: 0x32ba, 0x488: 0x38cf, 0x489: 0x3a5e, 0x48a: 0x2fc7, 0x48b: 0x32d3, + 0x48c: 0x2fd1, 0x48d: 0x32dd, 0x48e: 0x2fe0, 0x48f: 0x32ec, 0x490: 0x2fd6, 0x491: 0x32e2, + 0x492: 0x2fdb, 0x493: 0x32e7, 0x494: 0x38f2, 0x495: 0x3a81, 0x496: 0x38f9, 0x497: 0x3a88, + 0x498: 0x301c, 0x499: 0x3328, 0x49a: 0x3021, 0x49b: 0x332d, 0x49c: 0x3907, 0x49d: 0x3a96, + 0x49e: 0x3026, 0x49f: 0x3332, 0x4a0: 0x3035, 0x4a1: 0x3341, 0x4a2: 0x3053, 0x4a3: 0x335f, + 0x4a4: 0x3062, 0x4a5: 0x336e, 0x4a6: 0x3058, 0x4a7: 0x3364, 0x4a8: 0x3067, 0x4a9: 0x3373, + 0x4aa: 0x306c, 0x4ab: 0x3378, 0x4ac: 0x30b2, 0x4ad: 0x33be, 0x4ae: 0x390e, 0x4af: 0x3a9d, + 0x4b0: 0x30bc, 0x4b1: 0x33cd, 0x4b2: 0x30c6, 0x4b3: 0x33d7, 0x4b4: 0x30d0, 0x4b5: 0x33e1, + 0x4b6: 0x46c7, 0x4b7: 0x4758, 0x4b8: 0x3915, 0x4b9: 0x3aa4, 0x4ba: 0x30e9, 0x4bb: 0x33fa, + 0x4bc: 0x30e4, 0x4bd: 0x33f5, 0x4be: 0x30ee, 0x4bf: 0x33ff, + // Block 0x13, offset 0x4c0 + 0x4c0: 0x30f3, 0x4c1: 0x3404, 0x4c2: 0x30f8, 0x4c3: 0x3409, 0x4c4: 0x310c, 0x4c5: 0x341d, + 0x4c6: 0x3116, 0x4c7: 0x3427, 0x4c8: 0x3125, 0x4c9: 0x3436, 0x4ca: 0x3120, 0x4cb: 0x3431, + 0x4cc: 0x3938, 0x4cd: 0x3ac7, 0x4ce: 0x3946, 0x4cf: 0x3ad5, 0x4d0: 0x394d, 0x4d1: 0x3adc, + 0x4d2: 0x3954, 0x4d3: 0x3ae3, 0x4d4: 0x3152, 0x4d5: 0x3463, 0x4d6: 0x3157, 0x4d7: 0x3468, + 0x4d8: 0x3161, 0x4d9: 0x3472, 0x4da: 0x46f4, 0x4db: 0x4785, 0x4dc: 0x399a, 0x4dd: 0x3b29, + 0x4de: 0x317a, 0x4df: 0x348b, 0x4e0: 0x3184, 0x4e1: 0x3495, 0x4e2: 0x4703, 0x4e3: 0x4794, + 0x4e4: 0x39a1, 0x4e5: 0x3b30, 0x4e6: 0x39a8, 0x4e7: 0x3b37, 0x4e8: 0x39af, 0x4e9: 0x3b3e, + 0x4ea: 0x3193, 0x4eb: 0x34a4, 0x4ec: 0x319d, 0x4ed: 0x34b3, 0x4ee: 0x31b1, 0x4ef: 0x34c7, + 0x4f0: 0x31ac, 0x4f1: 0x34c2, 0x4f2: 0x31ed, 0x4f3: 0x3503, 0x4f4: 0x31fc, 0x4f5: 0x3512, + 0x4f6: 0x31f7, 0x4f7: 0x350d, 0x4f8: 0x39b6, 0x4f9: 0x3b45, 0x4fa: 0x39bd, 0x4fb: 0x3b4c, + 0x4fc: 0x3201, 0x4fd: 0x3517, 0x4fe: 0x3206, 0x4ff: 0x351c, + // Block 0x14, offset 0x500 + 0x500: 0x320b, 0x501: 0x3521, 0x502: 0x3210, 0x503: 0x3526, 0x504: 0x321f, 0x505: 0x3535, + 0x506: 0x321a, 0x507: 0x3530, 0x508: 0x3224, 0x509: 0x353f, 0x50a: 0x3229, 0x50b: 0x3544, + 0x50c: 0x322e, 0x50d: 0x3549, 0x50e: 0x324c, 0x50f: 0x3567, 0x510: 0x3265, 0x511: 0x3585, + 0x512: 0x3274, 0x513: 0x3594, 0x514: 0x3279, 0x515: 0x3599, 0x516: 0x337d, 0x517: 0x34a9, + 0x518: 0x353a, 0x519: 0x3576, 0x51b: 0x35d4, + 0x520: 0x46a4, 0x521: 0x4735, 0x522: 0x2f86, 0x523: 0x3292, + 0x524: 0x387b, 0x525: 0x3a0a, 0x526: 0x3874, 0x527: 0x3a03, 0x528: 0x3889, 0x529: 0x3a18, + 0x52a: 0x3882, 0x52b: 0x3a11, 0x52c: 0x38c1, 0x52d: 0x3a50, 0x52e: 0x3897, 0x52f: 0x3a26, + 0x530: 0x3890, 0x531: 0x3a1f, 0x532: 0x38a5, 0x533: 0x3a34, 0x534: 0x389e, 0x535: 0x3a2d, + 0x536: 0x38c8, 0x537: 0x3a57, 0x538: 0x46b8, 0x539: 0x4749, 0x53a: 0x3003, 0x53b: 0x330f, + 0x53c: 0x2fef, 0x53d: 0x32fb, 0x53e: 0x38dd, 0x53f: 0x3a6c, + // Block 0x15, offset 0x540 + 0x540: 0x38d6, 0x541: 0x3a65, 0x542: 0x38eb, 0x543: 0x3a7a, 0x544: 0x38e4, 0x545: 0x3a73, + 0x546: 0x3900, 0x547: 0x3a8f, 0x548: 0x3094, 0x549: 0x33a0, 0x54a: 0x30a8, 0x54b: 0x33b4, + 0x54c: 0x46ea, 0x54d: 0x477b, 0x54e: 0x3139, 0x54f: 0x344a, 0x550: 0x3923, 0x551: 0x3ab2, + 0x552: 0x391c, 0x553: 0x3aab, 0x554: 0x3931, 0x555: 0x3ac0, 0x556: 0x392a, 0x557: 0x3ab9, + 0x558: 0x398c, 0x559: 0x3b1b, 0x55a: 0x3970, 0x55b: 0x3aff, 0x55c: 0x3969, 0x55d: 0x3af8, + 0x55e: 0x397e, 0x55f: 0x3b0d, 0x560: 0x3977, 0x561: 0x3b06, 0x562: 0x3985, 0x563: 0x3b14, + 0x564: 0x31e8, 0x565: 0x34fe, 0x566: 0x31ca, 0x567: 0x34e0, 0x568: 0x39e7, 0x569: 0x3b76, + 0x56a: 0x39e0, 0x56b: 0x3b6f, 0x56c: 0x39f5, 0x56d: 0x3b84, 0x56e: 0x39ee, 0x56f: 0x3b7d, + 0x570: 0x39fc, 0x571: 0x3b8b, 0x572: 0x3233, 0x573: 0x354e, 0x574: 0x325b, 0x575: 0x357b, + 0x576: 0x3256, 0x577: 0x3571, 0x578: 0x3242, 0x579: 0x355d, + // Block 0x16, offset 0x580 + 0x580: 0x4807, 0x581: 0x480d, 0x582: 0x4921, 0x583: 0x4939, 0x584: 0x4929, 0x585: 0x4941, + 0x586: 0x4931, 0x587: 0x4949, 0x588: 0x47ad, 0x589: 0x47b3, 0x58a: 0x4891, 0x58b: 0x48a9, + 0x58c: 0x4899, 0x58d: 0x48b1, 0x58e: 0x48a1, 0x58f: 0x48b9, 0x590: 0x4819, 0x591: 0x481f, + 0x592: 0x3dbb, 0x593: 0x3dcb, 0x594: 0x3dc3, 0x595: 0x3dd3, + 0x598: 0x47b9, 0x599: 0x47bf, 0x59a: 0x3ceb, 0x59b: 0x3cfb, 0x59c: 0x3cf3, 0x59d: 0x3d03, + 0x5a0: 0x4831, 0x5a1: 0x4837, 0x5a2: 0x4951, 0x5a3: 0x4969, + 0x5a4: 0x4959, 0x5a5: 0x4971, 0x5a6: 0x4961, 0x5a7: 0x4979, 0x5a8: 0x47c5, 0x5a9: 0x47cb, + 0x5aa: 0x48c1, 0x5ab: 0x48d9, 0x5ac: 0x48c9, 0x5ad: 0x48e1, 0x5ae: 0x48d1, 0x5af: 0x48e9, + 0x5b0: 0x4849, 0x5b1: 0x484f, 0x5b2: 0x3e1b, 0x5b3: 0x3e33, 0x5b4: 0x3e23, 0x5b5: 0x3e3b, + 0x5b6: 0x3e2b, 0x5b7: 0x3e43, 0x5b8: 0x47d1, 0x5b9: 0x47d7, 0x5ba: 0x3d1b, 0x5bb: 0x3d33, + 0x5bc: 0x3d23, 0x5bd: 0x3d3b, 0x5be: 0x3d2b, 0x5bf: 0x3d43, + // Block 0x17, offset 0x5c0 + 0x5c0: 0x4855, 0x5c1: 0x485b, 0x5c2: 0x3e4b, 0x5c3: 0x3e5b, 0x5c4: 0x3e53, 0x5c5: 0x3e63, + 0x5c8: 0x47dd, 0x5c9: 0x47e3, 0x5ca: 0x3d4b, 0x5cb: 0x3d5b, + 0x5cc: 0x3d53, 0x5cd: 0x3d63, 0x5d0: 0x4867, 0x5d1: 0x486d, + 0x5d2: 0x3e83, 0x5d3: 0x3e9b, 0x5d4: 0x3e8b, 0x5d5: 0x3ea3, 0x5d6: 0x3e93, 0x5d7: 0x3eab, + 0x5d9: 0x47e9, 0x5db: 0x3d6b, 0x5dd: 0x3d73, + 0x5df: 0x3d7b, 0x5e0: 0x487f, 0x5e1: 0x4885, 0x5e2: 0x4981, 0x5e3: 0x4999, + 0x5e4: 0x4989, 0x5e5: 0x49a1, 0x5e6: 0x4991, 0x5e7: 0x49a9, 0x5e8: 0x47ef, 0x5e9: 0x47f5, + 0x5ea: 0x48f1, 0x5eb: 0x4909, 0x5ec: 0x48f9, 0x5ed: 0x4911, 0x5ee: 0x4901, 0x5ef: 0x4919, + 0x5f0: 0x47fb, 0x5f1: 0x4321, 0x5f2: 0x3694, 0x5f3: 0x4327, 0x5f4: 0x4825, 0x5f5: 0x432d, + 0x5f6: 0x36a6, 0x5f7: 0x4333, 0x5f8: 0x36c4, 0x5f9: 0x4339, 0x5fa: 0x36dc, 0x5fb: 0x433f, + 0x5fc: 0x4873, 0x5fd: 0x4345, + // Block 0x18, offset 0x600 + 0x600: 0x3da3, 0x601: 0x3dab, 0x602: 0x4187, 0x603: 0x41a5, 0x604: 0x4191, 0x605: 0x41af, + 0x606: 0x419b, 0x607: 0x41b9, 0x608: 0x3cdb, 0x609: 0x3ce3, 0x60a: 0x40d3, 0x60b: 0x40f1, + 0x60c: 0x40dd, 0x60d: 0x40fb, 0x60e: 0x40e7, 0x60f: 0x4105, 0x610: 0x3deb, 0x611: 0x3df3, + 0x612: 0x41c3, 0x613: 0x41e1, 0x614: 0x41cd, 0x615: 0x41eb, 0x616: 0x41d7, 0x617: 0x41f5, + 0x618: 0x3d0b, 0x619: 0x3d13, 0x61a: 0x410f, 0x61b: 0x412d, 0x61c: 0x4119, 0x61d: 0x4137, + 0x61e: 0x4123, 0x61f: 0x4141, 0x620: 0x3ec3, 0x621: 0x3ecb, 0x622: 0x41ff, 0x623: 0x421d, + 0x624: 0x4209, 0x625: 0x4227, 0x626: 0x4213, 0x627: 0x4231, 0x628: 0x3d83, 0x629: 0x3d8b, + 0x62a: 0x414b, 0x62b: 0x4169, 0x62c: 0x4155, 0x62d: 0x4173, 0x62e: 0x415f, 0x62f: 0x417d, + 0x630: 0x3688, 0x631: 0x3682, 0x632: 0x3d93, 0x633: 0x368e, 0x634: 0x3d9b, + 0x636: 0x4813, 0x637: 0x3db3, 0x638: 0x35f8, 0x639: 0x35f2, 0x63a: 0x35e6, 0x63b: 0x42f1, + 0x63c: 0x35fe, 0x63d: 0x8100, 0x63e: 0x01d3, 0x63f: 0xa100, + // Block 0x19, offset 0x640 + 0x640: 0x8100, 0x641: 0x35aa, 0x642: 0x3ddb, 0x643: 0x36a0, 0x644: 0x3de3, + 0x646: 0x483d, 0x647: 0x3dfb, 0x648: 0x3604, 0x649: 0x42f7, 0x64a: 0x3610, 0x64b: 0x42fd, + 0x64c: 0x361c, 0x64d: 0x3b92, 0x64e: 0x3b99, 0x64f: 0x3ba0, 0x650: 0x36b8, 0x651: 0x36b2, + 0x652: 0x3e03, 0x653: 0x44e7, 0x656: 0x36be, 0x657: 0x3e13, + 0x658: 0x3634, 0x659: 0x362e, 0x65a: 0x3622, 0x65b: 0x4303, 0x65d: 0x3ba7, + 0x65e: 0x3bae, 0x65f: 0x3bb5, 0x660: 0x36ee, 0x661: 0x36e8, 0x662: 0x3e6b, 0x663: 0x44ef, + 0x664: 0x36d0, 0x665: 0x36d6, 0x666: 0x36f4, 0x667: 0x3e7b, 0x668: 0x3664, 0x669: 0x365e, + 0x66a: 0x3652, 0x66b: 0x430f, 0x66c: 0x364c, 0x66d: 0x359e, 0x66e: 0x42eb, 0x66f: 0x0081, + 0x672: 0x3eb3, 0x673: 0x36fa, 0x674: 0x3ebb, + 0x676: 0x488b, 0x677: 0x3ed3, 0x678: 0x3640, 0x679: 0x4309, 0x67a: 0x3670, 0x67b: 0x431b, + 0x67c: 0x367c, 0x67d: 0x4259, 0x67e: 0xa100, + // Block 0x1a, offset 0x680 + 0x681: 0x3c09, 0x683: 0xa000, 0x684: 0x3c10, 0x685: 0xa000, + 0x687: 0x3c17, 0x688: 0xa000, 0x689: 0x3c1e, + 0x68d: 0xa000, + 0x6a0: 0x2f68, 0x6a1: 0xa000, 0x6a2: 0x3c2c, + 0x6a4: 0xa000, 0x6a5: 0xa000, + 0x6ad: 0x3c25, 0x6ae: 0x2f63, 0x6af: 0x2f6d, + 0x6b0: 0x3c33, 0x6b1: 0x3c3a, 0x6b2: 0xa000, 0x6b3: 0xa000, 0x6b4: 0x3c41, 0x6b5: 0x3c48, + 0x6b6: 0xa000, 0x6b7: 0xa000, 0x6b8: 0x3c4f, 0x6b9: 0x3c56, 0x6ba: 0xa000, 0x6bb: 0xa000, + 0x6bc: 0xa000, 0x6bd: 0xa000, + // Block 0x1b, offset 0x6c0 + 0x6c0: 0x3c5d, 0x6c1: 0x3c64, 0x6c2: 0xa000, 0x6c3: 0xa000, 0x6c4: 0x3c79, 0x6c5: 0x3c80, + 0x6c6: 0xa000, 0x6c7: 0xa000, 0x6c8: 0x3c87, 0x6c9: 0x3c8e, + 0x6d1: 0xa000, + 0x6d2: 0xa000, + 0x6e2: 0xa000, + 0x6e8: 0xa000, 0x6e9: 0xa000, + 0x6eb: 0xa000, 0x6ec: 0x3ca3, 0x6ed: 0x3caa, 0x6ee: 0x3cb1, 0x6ef: 0x3cb8, + 0x6f2: 0xa000, 0x6f3: 0xa000, 0x6f4: 0xa000, 0x6f5: 0xa000, + // Block 0x1c, offset 0x700 + 0x706: 0xa000, 0x70b: 0xa000, + 0x70c: 0x3f0b, 0x70d: 0xa000, 0x70e: 0x3f13, 0x70f: 0xa000, 0x710: 0x3f1b, 0x711: 0xa000, + 0x712: 0x3f23, 0x713: 0xa000, 0x714: 0x3f2b, 0x715: 0xa000, 0x716: 0x3f33, 0x717: 0xa000, + 0x718: 0x3f3b, 0x719: 0xa000, 0x71a: 0x3f43, 0x71b: 0xa000, 0x71c: 0x3f4b, 0x71d: 0xa000, + 0x71e: 0x3f53, 0x71f: 0xa000, 0x720: 0x3f5b, 0x721: 0xa000, 0x722: 0x3f63, + 0x724: 0xa000, 0x725: 0x3f6b, 0x726: 0xa000, 0x727: 0x3f73, 0x728: 0xa000, 0x729: 0x3f7b, + 0x72f: 0xa000, + 0x730: 0x3f83, 0x731: 0x3f8b, 0x732: 0xa000, 0x733: 0x3f93, 0x734: 0x3f9b, 0x735: 0xa000, + 0x736: 0x3fa3, 0x737: 0x3fab, 0x738: 0xa000, 0x739: 0x3fb3, 0x73a: 0x3fbb, 0x73b: 0xa000, + 0x73c: 0x3fc3, 0x73d: 0x3fcb, + // Block 0x1d, offset 0x740 + 0x754: 0x3f03, + 0x759: 0x9903, 0x75a: 0x9903, 0x75b: 0x8100, 0x75c: 0x8100, 0x75d: 0xa000, + 0x75e: 0x3fd3, + 0x766: 0xa000, + 0x76b: 0xa000, 0x76c: 0x3fe3, 0x76d: 0xa000, 0x76e: 0x3feb, 0x76f: 0xa000, + 0x770: 0x3ff3, 0x771: 0xa000, 0x772: 0x3ffb, 0x773: 0xa000, 0x774: 0x4003, 0x775: 0xa000, + 0x776: 0x400b, 0x777: 0xa000, 0x778: 0x4013, 0x779: 0xa000, 0x77a: 0x401b, 0x77b: 0xa000, + 0x77c: 0x4023, 0x77d: 0xa000, 0x77e: 0x402b, 0x77f: 0xa000, + // Block 0x1e, offset 0x780 + 0x780: 0x4033, 0x781: 0xa000, 0x782: 0x403b, 0x784: 0xa000, 0x785: 0x4043, + 0x786: 0xa000, 0x787: 0x404b, 0x788: 0xa000, 0x789: 0x4053, + 0x78f: 0xa000, 0x790: 0x405b, 0x791: 0x4063, + 0x792: 0xa000, 0x793: 0x406b, 0x794: 0x4073, 0x795: 0xa000, 0x796: 0x407b, 0x797: 0x4083, + 0x798: 0xa000, 0x799: 0x408b, 0x79a: 0x4093, 0x79b: 0xa000, 0x79c: 0x409b, 0x79d: 0x40a3, + 0x7af: 0xa000, + 0x7b0: 0xa000, 0x7b1: 0xa000, 0x7b2: 0xa000, 0x7b4: 0x3fdb, + 0x7b7: 0x40ab, 0x7b8: 0x40b3, 0x7b9: 0x40bb, 0x7ba: 0x40c3, + 0x7bd: 0xa000, 0x7be: 0x40cb, + // Block 0x1f, offset 0x7c0 + 0x7c0: 0x1377, 0x7c1: 0x0cfb, 0x7c2: 0x13d3, 0x7c3: 0x139f, 0x7c4: 0x0e57, 0x7c5: 0x06eb, + 0x7c6: 0x08df, 0x7c7: 0x162b, 0x7c8: 0x162b, 0x7c9: 0x0a0b, 0x7ca: 0x145f, 0x7cb: 0x0943, + 0x7cc: 0x0a07, 0x7cd: 0x0bef, 0x7ce: 0x0fcf, 0x7cf: 0x115f, 0x7d0: 0x1297, 0x7d1: 0x12d3, + 0x7d2: 0x1307, 0x7d3: 0x141b, 0x7d4: 0x0d73, 0x7d5: 0x0dff, 0x7d6: 0x0eab, 0x7d7: 0x0f43, + 0x7d8: 0x125f, 0x7d9: 0x1447, 0x7da: 0x1573, 0x7db: 0x070f, 0x7dc: 0x08b3, 0x7dd: 0x0d87, + 0x7de: 0x0ecf, 0x7df: 0x1293, 0x7e0: 0x15c3, 0x7e1: 0x0ab3, 0x7e2: 0x0e77, 0x7e3: 0x1283, + 0x7e4: 0x1317, 0x7e5: 0x0c23, 0x7e6: 0x11bb, 0x7e7: 0x12df, 0x7e8: 0x0b1f, 0x7e9: 0x0d0f, + 0x7ea: 0x0e17, 0x7eb: 0x0f1b, 0x7ec: 0x1427, 0x7ed: 0x074f, 0x7ee: 0x07e7, 0x7ef: 0x0853, + 0x7f0: 0x0c8b, 0x7f1: 0x0d7f, 0x7f2: 0x0ecb, 0x7f3: 0x0fef, 0x7f4: 0x1177, 0x7f5: 0x128b, + 0x7f6: 0x12a3, 0x7f7: 0x13c7, 0x7f8: 0x14ef, 0x7f9: 0x15a3, 0x7fa: 0x15bf, 0x7fb: 0x102b, + 0x7fc: 0x106b, 0x7fd: 0x1123, 0x7fe: 0x1243, 0x7ff: 0x147b, + // Block 0x20, offset 0x800 + 0x800: 0x15cb, 0x801: 0x134b, 0x802: 0x09c7, 0x803: 0x0b3b, 0x804: 0x10db, 0x805: 0x119b, + 0x806: 0x0eff, 0x807: 0x1033, 0x808: 0x1397, 0x809: 0x14e7, 0x80a: 0x09c3, 0x80b: 0x0a8f, + 0x80c: 0x0d77, 0x80d: 0x0e2b, 0x80e: 0x0e5f, 0x80f: 0x1113, 0x810: 0x113b, 0x811: 0x14a7, + 0x812: 0x084f, 0x813: 0x11a7, 0x814: 0x07f3, 0x815: 0x07ef, 0x816: 0x1097, 0x817: 0x1127, + 0x818: 0x125b, 0x819: 0x14af, 0x81a: 0x1367, 0x81b: 0x0c27, 0x81c: 0x0d73, 0x81d: 0x1357, + 0x81e: 0x06f7, 0x81f: 0x0a63, 0x820: 0x0b93, 0x821: 0x0f2f, 0x822: 0x0faf, 0x823: 0x0873, + 0x824: 0x103b, 0x825: 0x075f, 0x826: 0x0b77, 0x827: 0x06d7, 0x828: 0x0deb, 0x829: 0x0ca3, + 0x82a: 0x110f, 0x82b: 0x08c7, 0x82c: 0x09b3, 0x82d: 0x0ffb, 0x82e: 0x1263, 0x82f: 0x133b, + 0x830: 0x0db7, 0x831: 0x13f7, 0x832: 0x0de3, 0x833: 0x0c37, 0x834: 0x121b, 0x835: 0x0c57, + 0x836: 0x0fab, 0x837: 0x072b, 0x838: 0x07a7, 0x839: 0x07eb, 0x83a: 0x0d53, 0x83b: 0x10fb, + 0x83c: 0x11f3, 0x83d: 0x1347, 0x83e: 0x145b, 0x83f: 0x085b, + // Block 0x21, offset 0x840 + 0x840: 0x090f, 0x841: 0x0a17, 0x842: 0x0b2f, 0x843: 0x0cbf, 0x844: 0x0e7b, 0x845: 0x103f, + 0x846: 0x1497, 0x847: 0x157b, 0x848: 0x15cf, 0x849: 0x15e7, 0x84a: 0x0837, 0x84b: 0x0cf3, + 0x84c: 0x0da3, 0x84d: 0x13eb, 0x84e: 0x0afb, 0x84f: 0x0bd7, 0x850: 0x0bf3, 0x851: 0x0c83, + 0x852: 0x0e6b, 0x853: 0x0eb7, 0x854: 0x0f67, 0x855: 0x108b, 0x856: 0x112f, 0x857: 0x1193, + 0x858: 0x13db, 0x859: 0x126b, 0x85a: 0x1403, 0x85b: 0x147f, 0x85c: 0x080f, 0x85d: 0x083b, + 0x85e: 0x0923, 0x85f: 0x0ea7, 0x860: 0x12f3, 0x861: 0x133b, 0x862: 0x0b1b, 0x863: 0x0b8b, + 0x864: 0x0c4f, 0x865: 0x0daf, 0x866: 0x10d7, 0x867: 0x0f23, 0x868: 0x073b, 0x869: 0x097f, + 0x86a: 0x0a63, 0x86b: 0x0ac7, 0x86c: 0x0b97, 0x86d: 0x0f3f, 0x86e: 0x0f5b, 0x86f: 0x116b, + 0x870: 0x118b, 0x871: 0x1463, 0x872: 0x14e3, 0x873: 0x14f3, 0x874: 0x152f, 0x875: 0x0753, + 0x876: 0x107f, 0x877: 0x144f, 0x878: 0x14cb, 0x879: 0x0baf, 0x87a: 0x0717, 0x87b: 0x0777, + 0x87c: 0x0a67, 0x87d: 0x0a87, 0x87e: 0x0caf, 0x87f: 0x0d73, + // Block 0x22, offset 0x880 + 0x880: 0x0ec3, 0x881: 0x0fcb, 0x882: 0x1277, 0x883: 0x1417, 0x884: 0x1623, 0x885: 0x0ce3, + 0x886: 0x14a3, 0x887: 0x0833, 0x888: 0x0d2f, 0x889: 0x0d3b, 0x88a: 0x0e0f, 0x88b: 0x0e47, + 0x88c: 0x0f4b, 0x88d: 0x0fa7, 0x88e: 0x1027, 0x88f: 0x110b, 0x890: 0x153b, 0x891: 0x07af, + 0x892: 0x0c03, 0x893: 0x14b3, 0x894: 0x0767, 0x895: 0x0aab, 0x896: 0x0e2f, 0x897: 0x13df, + 0x898: 0x0b67, 0x899: 0x0bb7, 0x89a: 0x0d43, 0x89b: 0x0f2f, 0x89c: 0x14bb, 0x89d: 0x0817, + 0x89e: 0x08ff, 0x89f: 0x0a97, 0x8a0: 0x0cd3, 0x8a1: 0x0d1f, 0x8a2: 0x0d5f, 0x8a3: 0x0df3, + 0x8a4: 0x0f47, 0x8a5: 0x0fbb, 0x8a6: 0x1157, 0x8a7: 0x12f7, 0x8a8: 0x1303, 0x8a9: 0x1457, + 0x8aa: 0x14d7, 0x8ab: 0x0883, 0x8ac: 0x0e4b, 0x8ad: 0x0903, 0x8ae: 0x0ec7, 0x8af: 0x0f6b, + 0x8b0: 0x1287, 0x8b1: 0x14bf, 0x8b2: 0x15ab, 0x8b3: 0x15d3, 0x8b4: 0x0d37, 0x8b5: 0x0e27, + 0x8b6: 0x11c3, 0x8b7: 0x10b7, 0x8b8: 0x10c3, 0x8b9: 0x10e7, 0x8ba: 0x0f17, 0x8bb: 0x0e9f, + 0x8bc: 0x1363, 0x8bd: 0x0733, 0x8be: 0x122b, 0x8bf: 0x081b, + // Block 0x23, offset 0x8c0 + 0x8c0: 0x080b, 0x8c1: 0x0b0b, 0x8c2: 0x0c2b, 0x8c3: 0x10f3, 0x8c4: 0x0a53, 0x8c5: 0x0e03, + 0x8c6: 0x0cef, 0x8c7: 0x13e7, 0x8c8: 0x12e7, 0x8c9: 0x14ab, 0x8ca: 0x1323, 0x8cb: 0x0b27, + 0x8cc: 0x0787, 0x8cd: 0x095b, 0x8d0: 0x09af, + 0x8d2: 0x0cdf, 0x8d5: 0x07f7, 0x8d6: 0x0f1f, 0x8d7: 0x0fe3, + 0x8d8: 0x1047, 0x8d9: 0x1063, 0x8da: 0x1067, 0x8db: 0x107b, 0x8dc: 0x14fb, 0x8dd: 0x10eb, + 0x8de: 0x116f, 0x8e0: 0x128f, 0x8e2: 0x1353, + 0x8e5: 0x1407, 0x8e6: 0x1433, + 0x8ea: 0x154f, 0x8eb: 0x1553, 0x8ec: 0x1557, 0x8ed: 0x15bb, 0x8ee: 0x142b, 0x8ef: 0x14c7, + 0x8f0: 0x0757, 0x8f1: 0x077b, 0x8f2: 0x078f, 0x8f3: 0x084b, 0x8f4: 0x0857, 0x8f5: 0x0897, + 0x8f6: 0x094b, 0x8f7: 0x0967, 0x8f8: 0x096f, 0x8f9: 0x09ab, 0x8fa: 0x09b7, 0x8fb: 0x0a93, + 0x8fc: 0x0a9b, 0x8fd: 0x0ba3, 0x8fe: 0x0bcb, 0x8ff: 0x0bd3, + // Block 0x24, offset 0x900 + 0x900: 0x0beb, 0x901: 0x0c97, 0x902: 0x0cc7, 0x903: 0x0ce7, 0x904: 0x0d57, 0x905: 0x0e1b, + 0x906: 0x0e37, 0x907: 0x0e67, 0x908: 0x0ebb, 0x909: 0x0edb, 0x90a: 0x0f4f, 0x90b: 0x102f, + 0x90c: 0x104b, 0x90d: 0x1053, 0x90e: 0x104f, 0x90f: 0x1057, 0x910: 0x105b, 0x911: 0x105f, + 0x912: 0x1073, 0x913: 0x1077, 0x914: 0x109b, 0x915: 0x10af, 0x916: 0x10cb, 0x917: 0x112f, + 0x918: 0x1137, 0x919: 0x113f, 0x91a: 0x1153, 0x91b: 0x117b, 0x91c: 0x11cb, 0x91d: 0x11ff, + 0x91e: 0x11ff, 0x91f: 0x1267, 0x920: 0x130f, 0x921: 0x1327, 0x922: 0x135b, 0x923: 0x135f, + 0x924: 0x13a3, 0x925: 0x13a7, 0x926: 0x13ff, 0x927: 0x1407, 0x928: 0x14db, 0x929: 0x151f, + 0x92a: 0x1537, 0x92b: 0x0b9b, 0x92c: 0x171e, 0x92d: 0x11e3, + 0x930: 0x06df, 0x931: 0x07e3, 0x932: 0x07a3, 0x933: 0x074b, 0x934: 0x078b, 0x935: 0x07b7, + 0x936: 0x0847, 0x937: 0x0863, 0x938: 0x094b, 0x939: 0x0937, 0x93a: 0x0947, 0x93b: 0x0963, + 0x93c: 0x09af, 0x93d: 0x09bf, 0x93e: 0x0a03, 0x93f: 0x0a0f, + // Block 0x25, offset 0x940 + 0x940: 0x0a2b, 0x941: 0x0a3b, 0x942: 0x0b23, 0x943: 0x0b2b, 0x944: 0x0b5b, 0x945: 0x0b7b, + 0x946: 0x0bab, 0x947: 0x0bc3, 0x948: 0x0bb3, 0x949: 0x0bd3, 0x94a: 0x0bc7, 0x94b: 0x0beb, + 0x94c: 0x0c07, 0x94d: 0x0c5f, 0x94e: 0x0c6b, 0x94f: 0x0c73, 0x950: 0x0c9b, 0x951: 0x0cdf, + 0x952: 0x0d0f, 0x953: 0x0d13, 0x954: 0x0d27, 0x955: 0x0da7, 0x956: 0x0db7, 0x957: 0x0e0f, + 0x958: 0x0e5b, 0x959: 0x0e53, 0x95a: 0x0e67, 0x95b: 0x0e83, 0x95c: 0x0ebb, 0x95d: 0x1013, + 0x95e: 0x0edf, 0x95f: 0x0f13, 0x960: 0x0f1f, 0x961: 0x0f5f, 0x962: 0x0f7b, 0x963: 0x0f9f, + 0x964: 0x0fc3, 0x965: 0x0fc7, 0x966: 0x0fe3, 0x967: 0x0fe7, 0x968: 0x0ff7, 0x969: 0x100b, + 0x96a: 0x1007, 0x96b: 0x1037, 0x96c: 0x10b3, 0x96d: 0x10cb, 0x96e: 0x10e3, 0x96f: 0x111b, + 0x970: 0x112f, 0x971: 0x114b, 0x972: 0x117b, 0x973: 0x122f, 0x974: 0x1257, 0x975: 0x12cb, + 0x976: 0x1313, 0x977: 0x131f, 0x978: 0x1327, 0x979: 0x133f, 0x97a: 0x1353, 0x97b: 0x1343, + 0x97c: 0x135b, 0x97d: 0x1357, 0x97e: 0x134f, 0x97f: 0x135f, + // Block 0x26, offset 0x980 + 0x980: 0x136b, 0x981: 0x13a7, 0x982: 0x13e3, 0x983: 0x1413, 0x984: 0x144b, 0x985: 0x146b, + 0x986: 0x14b7, 0x987: 0x14db, 0x988: 0x14fb, 0x989: 0x150f, 0x98a: 0x151f, 0x98b: 0x152b, + 0x98c: 0x1537, 0x98d: 0x158b, 0x98e: 0x162b, 0x98f: 0x16b5, 0x990: 0x16b0, 0x991: 0x16e2, + 0x992: 0x0607, 0x993: 0x062f, 0x994: 0x0633, 0x995: 0x1764, 0x996: 0x1791, 0x997: 0x1809, + 0x998: 0x1617, 0x999: 0x1627, + // Block 0x27, offset 0x9c0 + 0x9c0: 0x06fb, 0x9c1: 0x06f3, 0x9c2: 0x0703, 0x9c3: 0x1647, 0x9c4: 0x0747, 0x9c5: 0x0757, + 0x9c6: 0x075b, 0x9c7: 0x0763, 0x9c8: 0x076b, 0x9c9: 0x076f, 0x9ca: 0x077b, 0x9cb: 0x0773, + 0x9cc: 0x05b3, 0x9cd: 0x165b, 0x9ce: 0x078f, 0x9cf: 0x0793, 0x9d0: 0x0797, 0x9d1: 0x07b3, + 0x9d2: 0x164c, 0x9d3: 0x05b7, 0x9d4: 0x079f, 0x9d5: 0x07bf, 0x9d6: 0x1656, 0x9d7: 0x07cf, + 0x9d8: 0x07d7, 0x9d9: 0x0737, 0x9da: 0x07df, 0x9db: 0x07e3, 0x9dc: 0x1831, 0x9dd: 0x07ff, + 0x9de: 0x0807, 0x9df: 0x05bf, 0x9e0: 0x081f, 0x9e1: 0x0823, 0x9e2: 0x082b, 0x9e3: 0x082f, + 0x9e4: 0x05c3, 0x9e5: 0x0847, 0x9e6: 0x084b, 0x9e7: 0x0857, 0x9e8: 0x0863, 0x9e9: 0x0867, + 0x9ea: 0x086b, 0x9eb: 0x0873, 0x9ec: 0x0893, 0x9ed: 0x0897, 0x9ee: 0x089f, 0x9ef: 0x08af, + 0x9f0: 0x08b7, 0x9f1: 0x08bb, 0x9f2: 0x08bb, 0x9f3: 0x08bb, 0x9f4: 0x166a, 0x9f5: 0x0e93, + 0x9f6: 0x08cf, 0x9f7: 0x08d7, 0x9f8: 0x166f, 0x9f9: 0x08e3, 0x9fa: 0x08eb, 0x9fb: 0x08f3, + 0x9fc: 0x091b, 0x9fd: 0x0907, 0x9fe: 0x0913, 0x9ff: 0x0917, + // Block 0x28, offset 0xa00 + 0xa00: 0x091f, 0xa01: 0x0927, 0xa02: 0x092b, 0xa03: 0x0933, 0xa04: 0x093b, 0xa05: 0x093f, + 0xa06: 0x093f, 0xa07: 0x0947, 0xa08: 0x094f, 0xa09: 0x0953, 0xa0a: 0x095f, 0xa0b: 0x0983, + 0xa0c: 0x0967, 0xa0d: 0x0987, 0xa0e: 0x096b, 0xa0f: 0x0973, 0xa10: 0x080b, 0xa11: 0x09cf, + 0xa12: 0x0997, 0xa13: 0x099b, 0xa14: 0x099f, 0xa15: 0x0993, 0xa16: 0x09a7, 0xa17: 0x09a3, + 0xa18: 0x09bb, 0xa19: 0x1674, 0xa1a: 0x09d7, 0xa1b: 0x09db, 0xa1c: 0x09e3, 0xa1d: 0x09ef, + 0xa1e: 0x09f7, 0xa1f: 0x0a13, 0xa20: 0x1679, 0xa21: 0x167e, 0xa22: 0x0a1f, 0xa23: 0x0a23, + 0xa24: 0x0a27, 0xa25: 0x0a1b, 0xa26: 0x0a2f, 0xa27: 0x05c7, 0xa28: 0x05cb, 0xa29: 0x0a37, + 0xa2a: 0x0a3f, 0xa2b: 0x0a3f, 0xa2c: 0x1683, 0xa2d: 0x0a5b, 0xa2e: 0x0a5f, 0xa2f: 0x0a63, + 0xa30: 0x0a6b, 0xa31: 0x1688, 0xa32: 0x0a73, 0xa33: 0x0a77, 0xa34: 0x0b4f, 0xa35: 0x0a7f, + 0xa36: 0x05cf, 0xa37: 0x0a8b, 0xa38: 0x0a9b, 0xa39: 0x0aa7, 0xa3a: 0x0aa3, 0xa3b: 0x1692, + 0xa3c: 0x0aaf, 0xa3d: 0x1697, 0xa3e: 0x0abb, 0xa3f: 0x0ab7, + // Block 0x29, offset 0xa40 + 0xa40: 0x0abf, 0xa41: 0x0acf, 0xa42: 0x0ad3, 0xa43: 0x05d3, 0xa44: 0x0ae3, 0xa45: 0x0aeb, + 0xa46: 0x0aef, 0xa47: 0x0af3, 0xa48: 0x05d7, 0xa49: 0x169c, 0xa4a: 0x05db, 0xa4b: 0x0b0f, + 0xa4c: 0x0b13, 0xa4d: 0x0b17, 0xa4e: 0x0b1f, 0xa4f: 0x1863, 0xa50: 0x0b37, 0xa51: 0x16a6, + 0xa52: 0x16a6, 0xa53: 0x11d7, 0xa54: 0x0b47, 0xa55: 0x0b47, 0xa56: 0x05df, 0xa57: 0x16c9, + 0xa58: 0x179b, 0xa59: 0x0b57, 0xa5a: 0x0b5f, 0xa5b: 0x05e3, 0xa5c: 0x0b73, 0xa5d: 0x0b83, + 0xa5e: 0x0b87, 0xa5f: 0x0b8f, 0xa60: 0x0b9f, 0xa61: 0x05eb, 0xa62: 0x05e7, 0xa63: 0x0ba3, + 0xa64: 0x16ab, 0xa65: 0x0ba7, 0xa66: 0x0bbb, 0xa67: 0x0bbf, 0xa68: 0x0bc3, 0xa69: 0x0bbf, + 0xa6a: 0x0bcf, 0xa6b: 0x0bd3, 0xa6c: 0x0be3, 0xa6d: 0x0bdb, 0xa6e: 0x0bdf, 0xa6f: 0x0be7, + 0xa70: 0x0beb, 0xa71: 0x0bef, 0xa72: 0x0bfb, 0xa73: 0x0bff, 0xa74: 0x0c17, 0xa75: 0x0c1f, + 0xa76: 0x0c2f, 0xa77: 0x0c43, 0xa78: 0x16ba, 0xa79: 0x0c3f, 0xa7a: 0x0c33, 0xa7b: 0x0c4b, + 0xa7c: 0x0c53, 0xa7d: 0x0c67, 0xa7e: 0x16bf, 0xa7f: 0x0c6f, + // Block 0x2a, offset 0xa80 + 0xa80: 0x0c63, 0xa81: 0x0c5b, 0xa82: 0x05ef, 0xa83: 0x0c77, 0xa84: 0x0c7f, 0xa85: 0x0c87, + 0xa86: 0x0c7b, 0xa87: 0x05f3, 0xa88: 0x0c97, 0xa89: 0x0c9f, 0xa8a: 0x16c4, 0xa8b: 0x0ccb, + 0xa8c: 0x0cff, 0xa8d: 0x0cdb, 0xa8e: 0x05ff, 0xa8f: 0x0ce7, 0xa90: 0x05fb, 0xa91: 0x05f7, + 0xa92: 0x07c3, 0xa93: 0x07c7, 0xa94: 0x0d03, 0xa95: 0x0ceb, 0xa96: 0x11ab, 0xa97: 0x0663, + 0xa98: 0x0d0f, 0xa99: 0x0d13, 0xa9a: 0x0d17, 0xa9b: 0x0d2b, 0xa9c: 0x0d23, 0xa9d: 0x16dd, + 0xa9e: 0x0603, 0xa9f: 0x0d3f, 0xaa0: 0x0d33, 0xaa1: 0x0d4f, 0xaa2: 0x0d57, 0xaa3: 0x16e7, + 0xaa4: 0x0d5b, 0xaa5: 0x0d47, 0xaa6: 0x0d63, 0xaa7: 0x0607, 0xaa8: 0x0d67, 0xaa9: 0x0d6b, + 0xaaa: 0x0d6f, 0xaab: 0x0d7b, 0xaac: 0x16ec, 0xaad: 0x0d83, 0xaae: 0x060b, 0xaaf: 0x0d8f, + 0xab0: 0x16f1, 0xab1: 0x0d93, 0xab2: 0x060f, 0xab3: 0x0d9f, 0xab4: 0x0dab, 0xab5: 0x0db7, + 0xab6: 0x0dbb, 0xab7: 0x16f6, 0xab8: 0x168d, 0xab9: 0x16fb, 0xaba: 0x0ddb, 0xabb: 0x1700, + 0xabc: 0x0de7, 0xabd: 0x0def, 0xabe: 0x0ddf, 0xabf: 0x0dfb, + // Block 0x2b, offset 0xac0 + 0xac0: 0x0e0b, 0xac1: 0x0e1b, 0xac2: 0x0e0f, 0xac3: 0x0e13, 0xac4: 0x0e1f, 0xac5: 0x0e23, + 0xac6: 0x1705, 0xac7: 0x0e07, 0xac8: 0x0e3b, 0xac9: 0x0e3f, 0xaca: 0x0613, 0xacb: 0x0e53, + 0xacc: 0x0e4f, 0xacd: 0x170a, 0xace: 0x0e33, 0xacf: 0x0e6f, 0xad0: 0x170f, 0xad1: 0x1714, + 0xad2: 0x0e73, 0xad3: 0x0e87, 0xad4: 0x0e83, 0xad5: 0x0e7f, 0xad6: 0x0617, 0xad7: 0x0e8b, + 0xad8: 0x0e9b, 0xad9: 0x0e97, 0xada: 0x0ea3, 0xadb: 0x1651, 0xadc: 0x0eb3, 0xadd: 0x1719, + 0xade: 0x0ebf, 0xadf: 0x1723, 0xae0: 0x0ed3, 0xae1: 0x0edf, 0xae2: 0x0ef3, 0xae3: 0x1728, + 0xae4: 0x0f07, 0xae5: 0x0f0b, 0xae6: 0x172d, 0xae7: 0x1732, 0xae8: 0x0f27, 0xae9: 0x0f37, + 0xaea: 0x061b, 0xaeb: 0x0f3b, 0xaec: 0x061f, 0xaed: 0x061f, 0xaee: 0x0f53, 0xaef: 0x0f57, + 0xaf0: 0x0f5f, 0xaf1: 0x0f63, 0xaf2: 0x0f6f, 0xaf3: 0x0623, 0xaf4: 0x0f87, 0xaf5: 0x1737, + 0xaf6: 0x0fa3, 0xaf7: 0x173c, 0xaf8: 0x0faf, 0xaf9: 0x16a1, 0xafa: 0x0fbf, 0xafb: 0x1741, + 0xafc: 0x1746, 0xafd: 0x174b, 0xafe: 0x0627, 0xaff: 0x062b, + // Block 0x2c, offset 0xb00 + 0xb00: 0x0ff7, 0xb01: 0x1755, 0xb02: 0x1750, 0xb03: 0x175a, 0xb04: 0x175f, 0xb05: 0x0fff, + 0xb06: 0x1003, 0xb07: 0x1003, 0xb08: 0x100b, 0xb09: 0x0633, 0xb0a: 0x100f, 0xb0b: 0x0637, + 0xb0c: 0x063b, 0xb0d: 0x1769, 0xb0e: 0x1023, 0xb0f: 0x102b, 0xb10: 0x1037, 0xb11: 0x063f, + 0xb12: 0x176e, 0xb13: 0x105b, 0xb14: 0x1773, 0xb15: 0x1778, 0xb16: 0x107b, 0xb17: 0x1093, + 0xb18: 0x0643, 0xb19: 0x109b, 0xb1a: 0x109f, 0xb1b: 0x10a3, 0xb1c: 0x177d, 0xb1d: 0x1782, + 0xb1e: 0x1782, 0xb1f: 0x10bb, 0xb20: 0x0647, 0xb21: 0x1787, 0xb22: 0x10cf, 0xb23: 0x10d3, + 0xb24: 0x064b, 0xb25: 0x178c, 0xb26: 0x10ef, 0xb27: 0x064f, 0xb28: 0x10ff, 0xb29: 0x10f7, + 0xb2a: 0x1107, 0xb2b: 0x1796, 0xb2c: 0x111f, 0xb2d: 0x0653, 0xb2e: 0x112b, 0xb2f: 0x1133, + 0xb30: 0x1143, 0xb31: 0x0657, 0xb32: 0x17a0, 0xb33: 0x17a5, 0xb34: 0x065b, 0xb35: 0x17aa, + 0xb36: 0x115b, 0xb37: 0x17af, 0xb38: 0x1167, 0xb39: 0x1173, 0xb3a: 0x117b, 0xb3b: 0x17b4, + 0xb3c: 0x17b9, 0xb3d: 0x118f, 0xb3e: 0x17be, 0xb3f: 0x1197, + // Block 0x2d, offset 0xb40 + 0xb40: 0x16ce, 0xb41: 0x065f, 0xb42: 0x11af, 0xb43: 0x11b3, 0xb44: 0x0667, 0xb45: 0x11b7, + 0xb46: 0x0a33, 0xb47: 0x17c3, 0xb48: 0x17c8, 0xb49: 0x16d3, 0xb4a: 0x16d8, 0xb4b: 0x11d7, + 0xb4c: 0x11db, 0xb4d: 0x13f3, 0xb4e: 0x066b, 0xb4f: 0x1207, 0xb50: 0x1203, 0xb51: 0x120b, + 0xb52: 0x083f, 0xb53: 0x120f, 0xb54: 0x1213, 0xb55: 0x1217, 0xb56: 0x121f, 0xb57: 0x17cd, + 0xb58: 0x121b, 0xb59: 0x1223, 0xb5a: 0x1237, 0xb5b: 0x123b, 0xb5c: 0x1227, 0xb5d: 0x123f, + 0xb5e: 0x1253, 0xb5f: 0x1267, 0xb60: 0x1233, 0xb61: 0x1247, 0xb62: 0x124b, 0xb63: 0x124f, + 0xb64: 0x17d2, 0xb65: 0x17dc, 0xb66: 0x17d7, 0xb67: 0x066f, 0xb68: 0x126f, 0xb69: 0x1273, + 0xb6a: 0x127b, 0xb6b: 0x17f0, 0xb6c: 0x127f, 0xb6d: 0x17e1, 0xb6e: 0x0673, 0xb6f: 0x0677, + 0xb70: 0x17e6, 0xb71: 0x17eb, 0xb72: 0x067b, 0xb73: 0x129f, 0xb74: 0x12a3, 0xb75: 0x12a7, + 0xb76: 0x12ab, 0xb77: 0x12b7, 0xb78: 0x12b3, 0xb79: 0x12bf, 0xb7a: 0x12bb, 0xb7b: 0x12cb, + 0xb7c: 0x12c3, 0xb7d: 0x12c7, 0xb7e: 0x12cf, 0xb7f: 0x067f, + // Block 0x2e, offset 0xb80 + 0xb80: 0x12d7, 0xb81: 0x12db, 0xb82: 0x0683, 0xb83: 0x12eb, 0xb84: 0x12ef, 0xb85: 0x17f5, + 0xb86: 0x12fb, 0xb87: 0x12ff, 0xb88: 0x0687, 0xb89: 0x130b, 0xb8a: 0x05bb, 0xb8b: 0x17fa, + 0xb8c: 0x17ff, 0xb8d: 0x068b, 0xb8e: 0x068f, 0xb8f: 0x1337, 0xb90: 0x134f, 0xb91: 0x136b, + 0xb92: 0x137b, 0xb93: 0x1804, 0xb94: 0x138f, 0xb95: 0x1393, 0xb96: 0x13ab, 0xb97: 0x13b7, + 0xb98: 0x180e, 0xb99: 0x1660, 0xb9a: 0x13c3, 0xb9b: 0x13bf, 0xb9c: 0x13cb, 0xb9d: 0x1665, + 0xb9e: 0x13d7, 0xb9f: 0x13e3, 0xba0: 0x1813, 0xba1: 0x1818, 0xba2: 0x1423, 0xba3: 0x142f, + 0xba4: 0x1437, 0xba5: 0x181d, 0xba6: 0x143b, 0xba7: 0x1467, 0xba8: 0x1473, 0xba9: 0x1477, + 0xbaa: 0x146f, 0xbab: 0x1483, 0xbac: 0x1487, 0xbad: 0x1822, 0xbae: 0x1493, 0xbaf: 0x0693, + 0xbb0: 0x149b, 0xbb1: 0x1827, 0xbb2: 0x0697, 0xbb3: 0x14d3, 0xbb4: 0x0ac3, 0xbb5: 0x14eb, + 0xbb6: 0x182c, 0xbb7: 0x1836, 0xbb8: 0x069b, 0xbb9: 0x069f, 0xbba: 0x1513, 0xbbb: 0x183b, + 0xbbc: 0x06a3, 0xbbd: 0x1840, 0xbbe: 0x152b, 0xbbf: 0x152b, + // Block 0x2f, offset 0xbc0 + 0xbc0: 0x1533, 0xbc1: 0x1845, 0xbc2: 0x154b, 0xbc3: 0x06a7, 0xbc4: 0x155b, 0xbc5: 0x1567, + 0xbc6: 0x156f, 0xbc7: 0x1577, 0xbc8: 0x06ab, 0xbc9: 0x184a, 0xbca: 0x158b, 0xbcb: 0x15a7, + 0xbcc: 0x15b3, 0xbcd: 0x06af, 0xbce: 0x06b3, 0xbcf: 0x15b7, 0xbd0: 0x184f, 0xbd1: 0x06b7, + 0xbd2: 0x1854, 0xbd3: 0x1859, 0xbd4: 0x185e, 0xbd5: 0x15db, 0xbd6: 0x06bb, 0xbd7: 0x15ef, + 0xbd8: 0x15f7, 0xbd9: 0x15fb, 0xbda: 0x1603, 0xbdb: 0x160b, 0xbdc: 0x1613, 0xbdd: 0x1868, +} + +// nfcIndex: 22 blocks, 1408 entries, 1408 bytes +// Block 0 is the zero block. +var nfcIndex = [1408]uint8{ + // Block 0x0, offset 0x0 + // Block 0x1, offset 0x40 + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xc2: 0x2e, 0xc3: 0x01, 0xc4: 0x02, 0xc5: 0x03, 0xc6: 0x2f, 0xc7: 0x04, + 0xc8: 0x05, 0xca: 0x30, 0xcb: 0x31, 0xcc: 0x06, 0xcd: 0x07, 0xce: 0x08, 0xcf: 0x32, + 0xd0: 0x09, 0xd1: 0x33, 0xd2: 0x34, 0xd3: 0x0a, 0xd6: 0x0b, 0xd7: 0x35, + 0xd8: 0x36, 0xd9: 0x0c, 0xdb: 0x37, 0xdc: 0x38, 0xdd: 0x39, 0xdf: 0x3a, + 0xe0: 0x02, 0xe1: 0x03, 0xe2: 0x04, 0xe3: 0x05, + 0xea: 0x06, 0xeb: 0x07, 0xec: 0x08, 0xed: 0x09, 0xef: 0x0a, + 0xf0: 0x13, + // Block 0x4, offset 0x100 + 0x120: 0x3b, 0x121: 0x3c, 0x123: 0x0d, 0x124: 0x3d, 0x125: 0x3e, 0x126: 0x3f, 0x127: 0x40, + 0x128: 0x41, 0x129: 0x42, 0x12a: 0x43, 0x12b: 0x44, 0x12c: 0x3f, 0x12d: 0x45, 0x12e: 0x46, 0x12f: 0x47, + 0x131: 0x48, 0x132: 0x49, 0x133: 0x4a, 0x134: 0x4b, 0x135: 0x4c, 0x137: 0x4d, + 0x138: 0x4e, 0x139: 0x4f, 0x13a: 0x50, 0x13b: 0x51, 0x13c: 0x52, 0x13d: 0x53, 0x13e: 0x54, 0x13f: 0x55, + // Block 0x5, offset 0x140 + 0x140: 0x56, 0x142: 0x57, 0x144: 0x58, 0x145: 0x59, 0x146: 0x5a, 0x147: 0x5b, + 0x14d: 0x5c, + 0x15c: 0x5d, 0x15f: 0x5e, + 0x162: 0x5f, 0x164: 0x60, + 0x168: 0x61, 0x169: 0x62, 0x16a: 0x63, 0x16c: 0x0e, 0x16d: 0x64, 0x16e: 0x65, 0x16f: 0x66, + 0x170: 0x67, 0x173: 0x68, 0x177: 0x0f, + 0x178: 0x10, 0x179: 0x11, 0x17a: 0x12, 0x17b: 0x13, 0x17c: 0x14, 0x17d: 0x15, 0x17e: 0x16, 0x17f: 0x17, + // Block 0x6, offset 0x180 + 0x180: 0x69, 0x183: 0x6a, 0x184: 0x6b, 0x186: 0x6c, 0x187: 0x6d, + 0x188: 0x6e, 0x189: 0x18, 0x18a: 0x19, 0x18b: 0x6f, 0x18c: 0x70, + 0x1ab: 0x71, + 0x1b3: 0x72, 0x1b5: 0x73, 0x1b7: 0x74, + // Block 0x7, offset 0x1c0 + 0x1c0: 0x75, 0x1c1: 0x1a, 0x1c2: 0x1b, 0x1c3: 0x1c, 0x1c4: 0x76, 0x1c5: 0x77, + 0x1c9: 0x78, 0x1cc: 0x79, 0x1cd: 0x7a, + // Block 0x8, offset 0x200 + 0x219: 0x7b, 0x21a: 0x7c, 0x21b: 0x7d, + 0x220: 0x7e, 0x223: 0x7f, 0x224: 0x80, 0x225: 0x81, 0x226: 0x82, 0x227: 0x83, + 0x22a: 0x84, 0x22b: 0x85, 0x22f: 0x86, + 0x230: 0x87, 0x231: 0x88, 0x232: 0x89, 0x233: 0x8a, 0x234: 0x8b, 0x235: 0x8c, 0x236: 0x8d, 0x237: 0x87, + 0x238: 0x88, 0x239: 0x89, 0x23a: 0x8a, 0x23b: 0x8b, 0x23c: 0x8c, 0x23d: 0x8d, 0x23e: 0x87, 0x23f: 0x88, + // Block 0x9, offset 0x240 + 0x240: 0x89, 0x241: 0x8a, 0x242: 0x8b, 0x243: 0x8c, 0x244: 0x8d, 0x245: 0x87, 0x246: 0x88, 0x247: 0x89, + 0x248: 0x8a, 0x249: 0x8b, 0x24a: 0x8c, 0x24b: 0x8d, 0x24c: 0x87, 0x24d: 0x88, 0x24e: 0x89, 0x24f: 0x8a, + 0x250: 0x8b, 0x251: 0x8c, 0x252: 0x8d, 0x253: 0x87, 0x254: 0x88, 0x255: 0x89, 0x256: 0x8a, 0x257: 0x8b, + 0x258: 0x8c, 0x259: 0x8d, 0x25a: 0x87, 0x25b: 0x88, 0x25c: 0x89, 0x25d: 0x8a, 0x25e: 0x8b, 0x25f: 0x8c, + 0x260: 0x8d, 0x261: 0x87, 0x262: 0x88, 0x263: 0x89, 0x264: 0x8a, 0x265: 0x8b, 0x266: 0x8c, 0x267: 0x8d, + 0x268: 0x87, 0x269: 0x88, 0x26a: 0x89, 0x26b: 0x8a, 0x26c: 0x8b, 0x26d: 0x8c, 0x26e: 0x8d, 0x26f: 0x87, + 0x270: 0x88, 0x271: 0x89, 0x272: 0x8a, 0x273: 0x8b, 0x274: 0x8c, 0x275: 0x8d, 0x276: 0x87, 0x277: 0x88, + 0x278: 0x89, 0x279: 0x8a, 0x27a: 0x8b, 0x27b: 0x8c, 0x27c: 0x8d, 0x27d: 0x87, 0x27e: 0x88, 0x27f: 0x89, + // Block 0xa, offset 0x280 + 0x280: 0x8a, 0x281: 0x8b, 0x282: 0x8c, 0x283: 0x8d, 0x284: 0x87, 0x285: 0x88, 0x286: 0x89, 0x287: 0x8a, + 0x288: 0x8b, 0x289: 0x8c, 0x28a: 0x8d, 0x28b: 0x87, 0x28c: 0x88, 0x28d: 0x89, 0x28e: 0x8a, 0x28f: 0x8b, + 0x290: 0x8c, 0x291: 0x8d, 0x292: 0x87, 0x293: 0x88, 0x294: 0x89, 0x295: 0x8a, 0x296: 0x8b, 0x297: 0x8c, + 0x298: 0x8d, 0x299: 0x87, 0x29a: 0x88, 0x29b: 0x89, 0x29c: 0x8a, 0x29d: 0x8b, 0x29e: 0x8c, 0x29f: 0x8d, + 0x2a0: 0x87, 0x2a1: 0x88, 0x2a2: 0x89, 0x2a3: 0x8a, 0x2a4: 0x8b, 0x2a5: 0x8c, 0x2a6: 0x8d, 0x2a7: 0x87, + 0x2a8: 0x88, 0x2a9: 0x89, 0x2aa: 0x8a, 0x2ab: 0x8b, 0x2ac: 0x8c, 0x2ad: 0x8d, 0x2ae: 0x87, 0x2af: 0x88, + 0x2b0: 0x89, 0x2b1: 0x8a, 0x2b2: 0x8b, 0x2b3: 0x8c, 0x2b4: 0x8d, 0x2b5: 0x87, 0x2b6: 0x88, 0x2b7: 0x89, + 0x2b8: 0x8a, 0x2b9: 0x8b, 0x2ba: 0x8c, 0x2bb: 0x8d, 0x2bc: 0x87, 0x2bd: 0x88, 0x2be: 0x89, 0x2bf: 0x8a, + // Block 0xb, offset 0x2c0 + 0x2c0: 0x8b, 0x2c1: 0x8c, 0x2c2: 0x8d, 0x2c3: 0x87, 0x2c4: 0x88, 0x2c5: 0x89, 0x2c6: 0x8a, 0x2c7: 0x8b, + 0x2c8: 0x8c, 0x2c9: 0x8d, 0x2ca: 0x87, 0x2cb: 0x88, 0x2cc: 0x89, 0x2cd: 0x8a, 0x2ce: 0x8b, 0x2cf: 0x8c, + 0x2d0: 0x8d, 0x2d1: 0x87, 0x2d2: 0x88, 0x2d3: 0x89, 0x2d4: 0x8a, 0x2d5: 0x8b, 0x2d6: 0x8c, 0x2d7: 0x8d, + 0x2d8: 0x87, 0x2d9: 0x88, 0x2da: 0x89, 0x2db: 0x8a, 0x2dc: 0x8b, 0x2dd: 0x8c, 0x2de: 0x8e, + // Block 0xc, offset 0x300 + 0x324: 0x1d, 0x325: 0x1e, 0x326: 0x1f, 0x327: 0x20, + 0x328: 0x21, 0x329: 0x22, 0x32a: 0x23, 0x32b: 0x24, 0x32c: 0x8f, 0x32d: 0x90, 0x32e: 0x91, + 0x331: 0x92, 0x332: 0x93, 0x333: 0x94, 0x334: 0x95, + 0x338: 0x96, 0x339: 0x97, 0x33a: 0x98, 0x33b: 0x99, 0x33e: 0x9a, 0x33f: 0x9b, + // Block 0xd, offset 0x340 + 0x347: 0x9c, + 0x34b: 0x9d, 0x34d: 0x9e, + 0x368: 0x9f, 0x36b: 0xa0, + 0x374: 0xa1, + 0x37d: 0xa2, + // Block 0xe, offset 0x380 + 0x381: 0xa3, 0x382: 0xa4, 0x384: 0xa5, 0x385: 0x82, 0x387: 0xa6, + 0x388: 0xa7, 0x38b: 0xa8, 0x38c: 0xa9, 0x38d: 0xaa, + 0x391: 0xab, 0x392: 0xac, 0x393: 0xad, 0x396: 0xae, 0x397: 0xaf, + 0x398: 0x73, 0x39a: 0xb0, 0x39c: 0xb1, + 0x3a0: 0xb2, 0x3a7: 0xb3, + 0x3a8: 0xb4, 0x3a9: 0xb5, 0x3aa: 0xb6, + 0x3b0: 0x73, 0x3b5: 0xb7, 0x3b6: 0xb8, + // Block 0xf, offset 0x3c0 + 0x3eb: 0xb9, 0x3ec: 0xba, + // Block 0x10, offset 0x400 + 0x432: 0xbb, + // Block 0x11, offset 0x440 + 0x445: 0xbc, 0x446: 0xbd, 0x447: 0xbe, + 0x449: 0xbf, + // Block 0x12, offset 0x480 + 0x480: 0xc0, 0x484: 0xba, + 0x48b: 0xc1, + 0x4a3: 0xc2, 0x4a5: 0xc3, + // Block 0x13, offset 0x4c0 + 0x4c8: 0xc4, + // Block 0x14, offset 0x500 + 0x520: 0x25, 0x521: 0x26, 0x522: 0x27, 0x523: 0x28, 0x524: 0x29, 0x525: 0x2a, 0x526: 0x2b, 0x527: 0x2c, + 0x528: 0x2d, + // Block 0x15, offset 0x540 + 0x550: 0x0b, 0x551: 0x0c, 0x556: 0x0d, + 0x55b: 0x0e, 0x55d: 0x0f, 0x55e: 0x10, 0x55f: 0x11, + 0x56f: 0x12, +} + +// nfcSparseOffset: 151 entries, 302 bytes +var nfcSparseOffset = []uint16{0x0, 0x5, 0x9, 0xb, 0xd, 0x18, 0x28, 0x2a, 0x2f, 0x3a, 0x49, 0x56, 0x5e, 0x63, 0x68, 0x6a, 0x72, 0x79, 0x7c, 0x84, 0x88, 0x8c, 0x8e, 0x90, 0x99, 0x9d, 0xa4, 0xa9, 0xac, 0xb6, 0xb9, 0xc0, 0xc8, 0xcb, 0xcd, 0xd0, 0xd2, 0xd7, 0xe8, 0xf4, 0xf6, 0xfc, 0xfe, 0x100, 0x102, 0x104, 0x106, 0x108, 0x10b, 0x10e, 0x110, 0x113, 0x116, 0x11a, 0x11f, 0x128, 0x12a, 0x12d, 0x12f, 0x13a, 0x13e, 0x14c, 0x14f, 0x155, 0x15b, 0x166, 0x16a, 0x16c, 0x16e, 0x170, 0x172, 0x174, 0x17a, 0x17e, 0x180, 0x182, 0x18a, 0x18e, 0x191, 0x193, 0x195, 0x197, 0x19a, 0x19c, 0x19e, 0x1a0, 0x1a2, 0x1a8, 0x1ab, 0x1ad, 0x1b4, 0x1ba, 0x1c0, 0x1c8, 0x1ce, 0x1d4, 0x1da, 0x1de, 0x1ec, 0x1f5, 0x1f8, 0x1fb, 0x1fd, 0x200, 0x202, 0x206, 0x20b, 0x20d, 0x20f, 0x214, 0x21a, 0x21c, 0x21e, 0x220, 0x226, 0x229, 0x22b, 0x231, 0x234, 0x23c, 0x243, 0x246, 0x249, 0x24b, 0x24e, 0x256, 0x25a, 0x261, 0x264, 0x26a, 0x26c, 0x26f, 0x271, 0x274, 0x276, 0x278, 0x27a, 0x27c, 0x27f, 0x281, 0x283, 0x285, 0x287, 0x294, 0x29e, 0x2a0, 0x2a2, 0x2a8, 0x2aa, 0x2ac, 0x2af} + +// nfcSparseValues: 689 entries, 2756 bytes +var nfcSparseValues = [689]valueRange{ + // Block 0x0, offset 0x0 + {value: 0x0000, lo: 0x04}, + {value: 0xa100, lo: 0xa8, hi: 0xa8}, + {value: 0x8100, lo: 0xaf, hi: 0xaf}, + {value: 0x8100, lo: 0xb4, hi: 0xb4}, + {value: 0x8100, lo: 0xb8, hi: 0xb8}, + // Block 0x1, offset 0x5 + {value: 0x0091, lo: 0x03}, + {value: 0x46e5, lo: 0xa0, hi: 0xa1}, + {value: 0x4717, lo: 0xaf, hi: 0xb0}, + {value: 0xa000, lo: 0xb7, hi: 0xb7}, + // Block 0x2, offset 0x9 + {value: 0x0000, lo: 0x01}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + // Block 0x3, offset 0xb + {value: 0x0000, lo: 0x01}, + {value: 0x8100, lo: 0x98, hi: 0x9d}, + // Block 0x4, offset 0xd + {value: 0x0006, lo: 0x0a}, + {value: 0xa000, lo: 0x81, hi: 0x81}, + {value: 0xa000, lo: 0x85, hi: 0x85}, + {value: 0xa000, lo: 0x89, hi: 0x89}, + {value: 0x4843, lo: 0x8a, hi: 0x8a}, + {value: 0x4861, lo: 0x8b, hi: 0x8b}, + {value: 0x36ca, lo: 0x8c, hi: 0x8c}, + {value: 0x36e2, lo: 0x8d, hi: 0x8d}, + {value: 0x4879, lo: 0x8e, hi: 0x8e}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0x3700, lo: 0x93, hi: 0x94}, + // Block 0x5, offset 0x18 + {value: 0x0000, lo: 0x0f}, + {value: 0xa000, lo: 0x83, hi: 0x83}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0xa000, lo: 0x8b, hi: 0x8b}, + {value: 0xa000, lo: 0x8d, hi: 0x8d}, + {value: 0x37a8, lo: 0x90, hi: 0x90}, + {value: 0x37b4, lo: 0x91, hi: 0x91}, + {value: 0x37a2, lo: 0x93, hi: 0x93}, + {value: 0xa000, lo: 0x96, hi: 0x96}, + {value: 0x381a, lo: 0x97, hi: 0x97}, + {value: 0x37e4, lo: 0x9c, hi: 0x9c}, + {value: 0x37cc, lo: 0x9d, hi: 0x9d}, + {value: 0x37f6, lo: 0x9e, hi: 0x9e}, + {value: 0xa000, lo: 0xb4, hi: 0xb5}, + {value: 0x3820, lo: 0xb6, hi: 0xb6}, + {value: 0x3826, lo: 0xb7, hi: 0xb7}, + // Block 0x6, offset 0x28 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0x83, hi: 0x87}, + // Block 0x7, offset 0x2a + {value: 0x0001, lo: 0x04}, + {value: 0x8113, lo: 0x81, hi: 0x82}, + {value: 0x8132, lo: 0x84, hi: 0x84}, + {value: 0x812d, lo: 0x85, hi: 0x85}, + {value: 0x810d, lo: 0x87, hi: 0x87}, + // Block 0x8, offset 0x2f + {value: 0x0000, lo: 0x0a}, + {value: 0x8132, lo: 0x90, hi: 0x97}, + {value: 0x8119, lo: 0x98, hi: 0x98}, + {value: 0x811a, lo: 0x99, hi: 0x99}, + {value: 0x811b, lo: 0x9a, hi: 0x9a}, + {value: 0x3844, lo: 0xa2, hi: 0xa2}, + {value: 0x384a, lo: 0xa3, hi: 0xa3}, + {value: 0x3856, lo: 0xa4, hi: 0xa4}, + {value: 0x3850, lo: 0xa5, hi: 0xa5}, + {value: 0x385c, lo: 0xa6, hi: 0xa6}, + {value: 0xa000, lo: 0xa7, hi: 0xa7}, + // Block 0x9, offset 0x3a + {value: 0x0000, lo: 0x0e}, + {value: 0x386e, lo: 0x80, hi: 0x80}, + {value: 0xa000, lo: 0x81, hi: 0x81}, + {value: 0x3862, lo: 0x82, hi: 0x82}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0x3868, lo: 0x93, hi: 0x93}, + {value: 0xa000, lo: 0x95, hi: 0x95}, + {value: 0x8132, lo: 0x96, hi: 0x9c}, + {value: 0x8132, lo: 0x9f, hi: 0xa2}, + {value: 0x812d, lo: 0xa3, hi: 0xa3}, + {value: 0x8132, lo: 0xa4, hi: 0xa4}, + {value: 0x8132, lo: 0xa7, hi: 0xa8}, + {value: 0x812d, lo: 0xaa, hi: 0xaa}, + {value: 0x8132, lo: 0xab, hi: 0xac}, + {value: 0x812d, lo: 0xad, hi: 0xad}, + // Block 0xa, offset 0x49 + {value: 0x0000, lo: 0x0c}, + {value: 0x811f, lo: 0x91, hi: 0x91}, + {value: 0x8132, lo: 0xb0, hi: 0xb0}, + {value: 0x812d, lo: 0xb1, hi: 0xb1}, + {value: 0x8132, lo: 0xb2, hi: 0xb3}, + {value: 0x812d, lo: 0xb4, hi: 0xb4}, + {value: 0x8132, lo: 0xb5, hi: 0xb6}, + {value: 0x812d, lo: 0xb7, hi: 0xb9}, + {value: 0x8132, lo: 0xba, hi: 0xba}, + {value: 0x812d, lo: 0xbb, hi: 0xbc}, + {value: 0x8132, lo: 0xbd, hi: 0xbd}, + {value: 0x812d, lo: 0xbe, hi: 0xbe}, + {value: 0x8132, lo: 0xbf, hi: 0xbf}, + // Block 0xb, offset 0x56 + {value: 0x0005, lo: 0x07}, + {value: 0x8132, lo: 0x80, hi: 0x80}, + {value: 0x8132, lo: 0x81, hi: 0x81}, + {value: 0x812d, lo: 0x82, hi: 0x83}, + {value: 0x812d, lo: 0x84, hi: 0x85}, + {value: 0x812d, lo: 0x86, hi: 0x87}, + {value: 0x812d, lo: 0x88, hi: 0x89}, + {value: 0x8132, lo: 0x8a, hi: 0x8a}, + // Block 0xc, offset 0x5e + {value: 0x0000, lo: 0x04}, + {value: 0x8132, lo: 0xab, hi: 0xb1}, + {value: 0x812d, lo: 0xb2, hi: 0xb2}, + {value: 0x8132, lo: 0xb3, hi: 0xb3}, + {value: 0x812d, lo: 0xbd, hi: 0xbd}, + // Block 0xd, offset 0x63 + {value: 0x0000, lo: 0x04}, + {value: 0x8132, lo: 0x96, hi: 0x99}, + {value: 0x8132, lo: 0x9b, hi: 0xa3}, + {value: 0x8132, lo: 0xa5, hi: 0xa7}, + {value: 0x8132, lo: 0xa9, hi: 0xad}, + // Block 0xe, offset 0x68 + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x99, hi: 0x9b}, + // Block 0xf, offset 0x6a + {value: 0x0000, lo: 0x07}, + {value: 0xa000, lo: 0xa8, hi: 0xa8}, + {value: 0x3edb, lo: 0xa9, hi: 0xa9}, + {value: 0xa000, lo: 0xb0, hi: 0xb0}, + {value: 0x3ee3, lo: 0xb1, hi: 0xb1}, + {value: 0xa000, lo: 0xb3, hi: 0xb3}, + {value: 0x3eeb, lo: 0xb4, hi: 0xb4}, + {value: 0x9902, lo: 0xbc, hi: 0xbc}, + // Block 0x10, offset 0x72 + {value: 0x0008, lo: 0x06}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x8132, lo: 0x91, hi: 0x91}, + {value: 0x812d, lo: 0x92, hi: 0x92}, + {value: 0x8132, lo: 0x93, hi: 0x93}, + {value: 0x8132, lo: 0x94, hi: 0x94}, + {value: 0x451f, lo: 0x98, hi: 0x9f}, + // Block 0x11, offset 0x79 + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x12, offset 0x7c + {value: 0x0008, lo: 0x07}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0x2ca1, lo: 0x8b, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + {value: 0x455f, lo: 0x9c, hi: 0x9d}, + {value: 0x456f, lo: 0x9f, hi: 0x9f}, + {value: 0x8132, lo: 0xbe, hi: 0xbe}, + // Block 0x13, offset 0x84 + {value: 0x0000, lo: 0x03}, + {value: 0x4597, lo: 0xb3, hi: 0xb3}, + {value: 0x459f, lo: 0xb6, hi: 0xb6}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + // Block 0x14, offset 0x88 + {value: 0x0008, lo: 0x03}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x4577, lo: 0x99, hi: 0x9b}, + {value: 0x458f, lo: 0x9e, hi: 0x9e}, + // Block 0x15, offset 0x8c + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + // Block 0x16, offset 0x8e + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + // Block 0x17, offset 0x90 + {value: 0x0000, lo: 0x08}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0x2cb9, lo: 0x88, hi: 0x88}, + {value: 0x2cb1, lo: 0x8b, hi: 0x8b}, + {value: 0x2cc1, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x96, hi: 0x97}, + {value: 0x45a7, lo: 0x9c, hi: 0x9c}, + {value: 0x45af, lo: 0x9d, hi: 0x9d}, + // Block 0x18, offset 0x99 + {value: 0x0000, lo: 0x03}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0x2cc9, lo: 0x94, hi: 0x94}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x19, offset 0x9d + {value: 0x0000, lo: 0x06}, + {value: 0xa000, lo: 0x86, hi: 0x87}, + {value: 0x2cd1, lo: 0x8a, hi: 0x8a}, + {value: 0x2ce1, lo: 0x8b, hi: 0x8b}, + {value: 0x2cd9, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + // Block 0x1a, offset 0xa4 + {value: 0x1801, lo: 0x04}, + {value: 0xa000, lo: 0x86, hi: 0x86}, + {value: 0x3ef3, lo: 0x88, hi: 0x88}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x8120, lo: 0x95, hi: 0x96}, + // Block 0x1b, offset 0xa9 + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + {value: 0xa000, lo: 0xbf, hi: 0xbf}, + // Block 0x1c, offset 0xac + {value: 0x0000, lo: 0x09}, + {value: 0x2ce9, lo: 0x80, hi: 0x80}, + {value: 0x9900, lo: 0x82, hi: 0x82}, + {value: 0xa000, lo: 0x86, hi: 0x86}, + {value: 0x2cf1, lo: 0x87, hi: 0x87}, + {value: 0x2cf9, lo: 0x88, hi: 0x88}, + {value: 0x2f53, lo: 0x8a, hi: 0x8a}, + {value: 0x2ddb, lo: 0x8b, hi: 0x8b}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x95, hi: 0x96}, + // Block 0x1d, offset 0xb6 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xbb, hi: 0xbc}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x1e, offset 0xb9 + {value: 0x0000, lo: 0x06}, + {value: 0xa000, lo: 0x86, hi: 0x87}, + {value: 0x2d01, lo: 0x8a, hi: 0x8a}, + {value: 0x2d11, lo: 0x8b, hi: 0x8b}, + {value: 0x2d09, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + // Block 0x1f, offset 0xc0 + {value: 0x6be7, lo: 0x07}, + {value: 0x9904, lo: 0x8a, hi: 0x8a}, + {value: 0x9900, lo: 0x8f, hi: 0x8f}, + {value: 0xa000, lo: 0x99, hi: 0x99}, + {value: 0x3efb, lo: 0x9a, hi: 0x9a}, + {value: 0x2f5b, lo: 0x9c, hi: 0x9c}, + {value: 0x2de6, lo: 0x9d, hi: 0x9d}, + {value: 0x2d19, lo: 0x9e, hi: 0x9f}, + // Block 0x20, offset 0xc8 + {value: 0x0000, lo: 0x02}, + {value: 0x8122, lo: 0xb8, hi: 0xb9}, + {value: 0x8104, lo: 0xba, hi: 0xba}, + // Block 0x21, offset 0xcb + {value: 0x0000, lo: 0x01}, + {value: 0x8123, lo: 0x88, hi: 0x8b}, + // Block 0x22, offset 0xcd + {value: 0x0000, lo: 0x02}, + {value: 0x8124, lo: 0xb8, hi: 0xb9}, + {value: 0x8104, lo: 0xba, hi: 0xba}, + // Block 0x23, offset 0xd0 + {value: 0x0000, lo: 0x01}, + {value: 0x8125, lo: 0x88, hi: 0x8b}, + // Block 0x24, offset 0xd2 + {value: 0x0000, lo: 0x04}, + {value: 0x812d, lo: 0x98, hi: 0x99}, + {value: 0x812d, lo: 0xb5, hi: 0xb5}, + {value: 0x812d, lo: 0xb7, hi: 0xb7}, + {value: 0x812b, lo: 0xb9, hi: 0xb9}, + // Block 0x25, offset 0xd7 + {value: 0x0000, lo: 0x10}, + {value: 0x2647, lo: 0x83, hi: 0x83}, + {value: 0x264e, lo: 0x8d, hi: 0x8d}, + {value: 0x2655, lo: 0x92, hi: 0x92}, + {value: 0x265c, lo: 0x97, hi: 0x97}, + {value: 0x2663, lo: 0x9c, hi: 0x9c}, + {value: 0x2640, lo: 0xa9, hi: 0xa9}, + {value: 0x8126, lo: 0xb1, hi: 0xb1}, + {value: 0x8127, lo: 0xb2, hi: 0xb2}, + {value: 0x4a87, lo: 0xb3, hi: 0xb3}, + {value: 0x8128, lo: 0xb4, hi: 0xb4}, + {value: 0x4a90, lo: 0xb5, hi: 0xb5}, + {value: 0x45b7, lo: 0xb6, hi: 0xb6}, + {value: 0x8200, lo: 0xb7, hi: 0xb7}, + {value: 0x45bf, lo: 0xb8, hi: 0xb8}, + {value: 0x8200, lo: 0xb9, hi: 0xb9}, + {value: 0x8127, lo: 0xba, hi: 0xbd}, + // Block 0x26, offset 0xe8 + {value: 0x0000, lo: 0x0b}, + {value: 0x8127, lo: 0x80, hi: 0x80}, + {value: 0x4a99, lo: 0x81, hi: 0x81}, + {value: 0x8132, lo: 0x82, hi: 0x83}, + {value: 0x8104, lo: 0x84, hi: 0x84}, + {value: 0x8132, lo: 0x86, hi: 0x87}, + {value: 0x2671, lo: 0x93, hi: 0x93}, + {value: 0x2678, lo: 0x9d, hi: 0x9d}, + {value: 0x267f, lo: 0xa2, hi: 0xa2}, + {value: 0x2686, lo: 0xa7, hi: 0xa7}, + {value: 0x268d, lo: 0xac, hi: 0xac}, + {value: 0x266a, lo: 0xb9, hi: 0xb9}, + // Block 0x27, offset 0xf4 + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x86, hi: 0x86}, + // Block 0x28, offset 0xf6 + {value: 0x0000, lo: 0x05}, + {value: 0xa000, lo: 0xa5, hi: 0xa5}, + {value: 0x2d21, lo: 0xa6, hi: 0xa6}, + {value: 0x9900, lo: 0xae, hi: 0xae}, + {value: 0x8102, lo: 0xb7, hi: 0xb7}, + {value: 0x8104, lo: 0xb9, hi: 0xba}, + // Block 0x29, offset 0xfc + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x8d, hi: 0x8d}, + // Block 0x2a, offset 0xfe + {value: 0x0000, lo: 0x01}, + {value: 0xa000, lo: 0x80, hi: 0x92}, + // Block 0x2b, offset 0x100 + {value: 0x0000, lo: 0x01}, + {value: 0xb900, lo: 0xa1, hi: 0xb5}, + // Block 0x2c, offset 0x102 + {value: 0x0000, lo: 0x01}, + {value: 0x9900, lo: 0xa8, hi: 0xbf}, + // Block 0x2d, offset 0x104 + {value: 0x0000, lo: 0x01}, + {value: 0x9900, lo: 0x80, hi: 0x82}, + // Block 0x2e, offset 0x106 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0x9d, hi: 0x9f}, + // Block 0x2f, offset 0x108 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x94, hi: 0x94}, + {value: 0x8104, lo: 0xb4, hi: 0xb4}, + // Block 0x30, offset 0x10b + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x92, hi: 0x92}, + {value: 0x8132, lo: 0x9d, hi: 0x9d}, + // Block 0x31, offset 0x10e + {value: 0x0000, lo: 0x01}, + {value: 0x8131, lo: 0xa9, hi: 0xa9}, + // Block 0x32, offset 0x110 + {value: 0x0004, lo: 0x02}, + {value: 0x812e, lo: 0xb9, hi: 0xba}, + {value: 0x812d, lo: 0xbb, hi: 0xbb}, + // Block 0x33, offset 0x113 + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0x97, hi: 0x97}, + {value: 0x812d, lo: 0x98, hi: 0x98}, + // Block 0x34, offset 0x116 + {value: 0x0000, lo: 0x03}, + {value: 0x8104, lo: 0xa0, hi: 0xa0}, + {value: 0x8132, lo: 0xb5, hi: 0xbc}, + {value: 0x812d, lo: 0xbf, hi: 0xbf}, + // Block 0x35, offset 0x11a + {value: 0x0000, lo: 0x04}, + {value: 0x8132, lo: 0xb0, hi: 0xb4}, + {value: 0x812d, lo: 0xb5, hi: 0xba}, + {value: 0x8132, lo: 0xbb, hi: 0xbc}, + {value: 0x812d, lo: 0xbd, hi: 0xbd}, + // Block 0x36, offset 0x11f + {value: 0x0000, lo: 0x08}, + {value: 0x2d69, lo: 0x80, hi: 0x80}, + {value: 0x2d71, lo: 0x81, hi: 0x81}, + {value: 0xa000, lo: 0x82, hi: 0x82}, + {value: 0x2d79, lo: 0x83, hi: 0x83}, + {value: 0x8104, lo: 0x84, hi: 0x84}, + {value: 0x8132, lo: 0xab, hi: 0xab}, + {value: 0x812d, lo: 0xac, hi: 0xac}, + {value: 0x8132, lo: 0xad, hi: 0xb3}, + // Block 0x37, offset 0x128 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xaa, hi: 0xab}, + // Block 0x38, offset 0x12a + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xa6, hi: 0xa6}, + {value: 0x8104, lo: 0xb2, hi: 0xb3}, + // Block 0x39, offset 0x12d + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0xb7, hi: 0xb7}, + // Block 0x3a, offset 0x12f + {value: 0x0000, lo: 0x0a}, + {value: 0x8132, lo: 0x90, hi: 0x92}, + {value: 0x8101, lo: 0x94, hi: 0x94}, + {value: 0x812d, lo: 0x95, hi: 0x99}, + {value: 0x8132, lo: 0x9a, hi: 0x9b}, + {value: 0x812d, lo: 0x9c, hi: 0x9f}, + {value: 0x8132, lo: 0xa0, hi: 0xa0}, + {value: 0x8101, lo: 0xa2, hi: 0xa8}, + {value: 0x812d, lo: 0xad, hi: 0xad}, + {value: 0x8132, lo: 0xb4, hi: 0xb4}, + {value: 0x8132, lo: 0xb8, hi: 0xb9}, + // Block 0x3b, offset 0x13a + {value: 0x0004, lo: 0x03}, + {value: 0x0433, lo: 0x80, hi: 0x81}, + {value: 0x8100, lo: 0x97, hi: 0x97}, + {value: 0x8100, lo: 0xbe, hi: 0xbe}, + // Block 0x3c, offset 0x13e + {value: 0x0000, lo: 0x0d}, + {value: 0x8132, lo: 0x90, hi: 0x91}, + {value: 0x8101, lo: 0x92, hi: 0x93}, + {value: 0x8132, lo: 0x94, hi: 0x97}, + {value: 0x8101, lo: 0x98, hi: 0x9a}, + {value: 0x8132, lo: 0x9b, hi: 0x9c}, + {value: 0x8132, lo: 0xa1, hi: 0xa1}, + {value: 0x8101, lo: 0xa5, hi: 0xa6}, + {value: 0x8132, lo: 0xa7, hi: 0xa7}, + {value: 0x812d, lo: 0xa8, hi: 0xa8}, + {value: 0x8132, lo: 0xa9, hi: 0xa9}, + {value: 0x8101, lo: 0xaa, hi: 0xab}, + {value: 0x812d, lo: 0xac, hi: 0xaf}, + {value: 0x8132, lo: 0xb0, hi: 0xb0}, + // Block 0x3d, offset 0x14c + {value: 0x427e, lo: 0x02}, + {value: 0x01b8, lo: 0xa6, hi: 0xa6}, + {value: 0x0057, lo: 0xaa, hi: 0xab}, + // Block 0x3e, offset 0x14f + {value: 0x0007, lo: 0x05}, + {value: 0xa000, lo: 0x90, hi: 0x90}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0xa000, lo: 0x94, hi: 0x94}, + {value: 0x3bbc, lo: 0x9a, hi: 0x9b}, + {value: 0x3bca, lo: 0xae, hi: 0xae}, + // Block 0x3f, offset 0x155 + {value: 0x000e, lo: 0x05}, + {value: 0x3bd1, lo: 0x8d, hi: 0x8e}, + {value: 0x3bd8, lo: 0x8f, hi: 0x8f}, + {value: 0xa000, lo: 0x90, hi: 0x90}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0xa000, lo: 0x94, hi: 0x94}, + // Block 0x40, offset 0x15b + {value: 0x6405, lo: 0x0a}, + {value: 0xa000, lo: 0x83, hi: 0x83}, + {value: 0x3be6, lo: 0x84, hi: 0x84}, + {value: 0xa000, lo: 0x88, hi: 0x88}, + {value: 0x3bed, lo: 0x89, hi: 0x89}, + {value: 0xa000, lo: 0x8b, hi: 0x8b}, + {value: 0x3bf4, lo: 0x8c, hi: 0x8c}, + {value: 0xa000, lo: 0xa3, hi: 0xa3}, + {value: 0x3bfb, lo: 0xa4, hi: 0xa5}, + {value: 0x3c02, lo: 0xa6, hi: 0xa6}, + {value: 0xa000, lo: 0xbc, hi: 0xbc}, + // Block 0x41, offset 0x166 + {value: 0x0007, lo: 0x03}, + {value: 0x3c6b, lo: 0xa0, hi: 0xa1}, + {value: 0x3c95, lo: 0xa2, hi: 0xa3}, + {value: 0x3cbf, lo: 0xaa, hi: 0xad}, + // Block 0x42, offset 0x16a + {value: 0x0004, lo: 0x01}, + {value: 0x048b, lo: 0xa9, hi: 0xaa}, + // Block 0x43, offset 0x16c + {value: 0x0000, lo: 0x01}, + {value: 0x44e0, lo: 0x9c, hi: 0x9c}, + // Block 0x44, offset 0x16e + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xaf, hi: 0xb1}, + // Block 0x45, offset 0x170 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x46, offset 0x172 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xa0, hi: 0xbf}, + // Block 0x47, offset 0x174 + {value: 0x0000, lo: 0x05}, + {value: 0x812c, lo: 0xaa, hi: 0xaa}, + {value: 0x8131, lo: 0xab, hi: 0xab}, + {value: 0x8133, lo: 0xac, hi: 0xac}, + {value: 0x812e, lo: 0xad, hi: 0xad}, + {value: 0x812f, lo: 0xae, hi: 0xaf}, + // Block 0x48, offset 0x17a + {value: 0x0000, lo: 0x03}, + {value: 0x4aa2, lo: 0xb3, hi: 0xb3}, + {value: 0x4aa2, lo: 0xb5, hi: 0xb6}, + {value: 0x4aa2, lo: 0xba, hi: 0xbf}, + // Block 0x49, offset 0x17e + {value: 0x0000, lo: 0x01}, + {value: 0x4aa2, lo: 0x8f, hi: 0xa3}, + // Block 0x4a, offset 0x180 + {value: 0x0000, lo: 0x01}, + {value: 0x8100, lo: 0xae, hi: 0xbe}, + // Block 0x4b, offset 0x182 + {value: 0x0000, lo: 0x07}, + {value: 0x8100, lo: 0x84, hi: 0x84}, + {value: 0x8100, lo: 0x87, hi: 0x87}, + {value: 0x8100, lo: 0x90, hi: 0x90}, + {value: 0x8100, lo: 0x9e, hi: 0x9e}, + {value: 0x8100, lo: 0xa1, hi: 0xa1}, + {value: 0x8100, lo: 0xb2, hi: 0xb2}, + {value: 0x8100, lo: 0xbb, hi: 0xbb}, + // Block 0x4c, offset 0x18a + {value: 0x0000, lo: 0x03}, + {value: 0x8100, lo: 0x80, hi: 0x80}, + {value: 0x8100, lo: 0x8b, hi: 0x8b}, + {value: 0x8100, lo: 0x8e, hi: 0x8e}, + // Block 0x4d, offset 0x18e + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0xaf, hi: 0xaf}, + {value: 0x8132, lo: 0xb4, hi: 0xbd}, + // Block 0x4e, offset 0x191 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0x9e, hi: 0x9f}, + // Block 0x4f, offset 0x193 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xb0, hi: 0xb1}, + // Block 0x50, offset 0x195 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x86, hi: 0x86}, + // Block 0x51, offset 0x197 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x84, hi: 0x84}, + {value: 0x8132, lo: 0xa0, hi: 0xb1}, + // Block 0x52, offset 0x19a + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0xab, hi: 0xad}, + // Block 0x53, offset 0x19c + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x93, hi: 0x93}, + // Block 0x54, offset 0x19e + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0xb3, hi: 0xb3}, + // Block 0x55, offset 0x1a0 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x80, hi: 0x80}, + // Block 0x56, offset 0x1a2 + {value: 0x0000, lo: 0x05}, + {value: 0x8132, lo: 0xb0, hi: 0xb0}, + {value: 0x8132, lo: 0xb2, hi: 0xb3}, + {value: 0x812d, lo: 0xb4, hi: 0xb4}, + {value: 0x8132, lo: 0xb7, hi: 0xb8}, + {value: 0x8132, lo: 0xbe, hi: 0xbf}, + // Block 0x57, offset 0x1a8 + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0x81, hi: 0x81}, + {value: 0x8104, lo: 0xb6, hi: 0xb6}, + // Block 0x58, offset 0x1ab + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xad, hi: 0xad}, + // Block 0x59, offset 0x1ad + {value: 0x0000, lo: 0x06}, + {value: 0xe500, lo: 0x80, hi: 0x80}, + {value: 0xc600, lo: 0x81, hi: 0x9b}, + {value: 0xe500, lo: 0x9c, hi: 0x9c}, + {value: 0xc600, lo: 0x9d, hi: 0xb7}, + {value: 0xe500, lo: 0xb8, hi: 0xb8}, + {value: 0xc600, lo: 0xb9, hi: 0xbf}, + // Block 0x5a, offset 0x1b4 + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x93}, + {value: 0xe500, lo: 0x94, hi: 0x94}, + {value: 0xc600, lo: 0x95, hi: 0xaf}, + {value: 0xe500, lo: 0xb0, hi: 0xb0}, + {value: 0xc600, lo: 0xb1, hi: 0xbf}, + // Block 0x5b, offset 0x1ba + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x8b}, + {value: 0xe500, lo: 0x8c, hi: 0x8c}, + {value: 0xc600, lo: 0x8d, hi: 0xa7}, + {value: 0xe500, lo: 0xa8, hi: 0xa8}, + {value: 0xc600, lo: 0xa9, hi: 0xbf}, + // Block 0x5c, offset 0x1c0 + {value: 0x0000, lo: 0x07}, + {value: 0xc600, lo: 0x80, hi: 0x83}, + {value: 0xe500, lo: 0x84, hi: 0x84}, + {value: 0xc600, lo: 0x85, hi: 0x9f}, + {value: 0xe500, lo: 0xa0, hi: 0xa0}, + {value: 0xc600, lo: 0xa1, hi: 0xbb}, + {value: 0xe500, lo: 0xbc, hi: 0xbc}, + {value: 0xc600, lo: 0xbd, hi: 0xbf}, + // Block 0x5d, offset 0x1c8 + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x97}, + {value: 0xe500, lo: 0x98, hi: 0x98}, + {value: 0xc600, lo: 0x99, hi: 0xb3}, + {value: 0xe500, lo: 0xb4, hi: 0xb4}, + {value: 0xc600, lo: 0xb5, hi: 0xbf}, + // Block 0x5e, offset 0x1ce + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x8f}, + {value: 0xe500, lo: 0x90, hi: 0x90}, + {value: 0xc600, lo: 0x91, hi: 0xab}, + {value: 0xe500, lo: 0xac, hi: 0xac}, + {value: 0xc600, lo: 0xad, hi: 0xbf}, + // Block 0x5f, offset 0x1d4 + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x87}, + {value: 0xe500, lo: 0x88, hi: 0x88}, + {value: 0xc600, lo: 0x89, hi: 0xa3}, + {value: 0xe500, lo: 0xa4, hi: 0xa4}, + {value: 0xc600, lo: 0xa5, hi: 0xbf}, + // Block 0x60, offset 0x1da + {value: 0x0000, lo: 0x03}, + {value: 0xc600, lo: 0x80, hi: 0x87}, + {value: 0xe500, lo: 0x88, hi: 0x88}, + {value: 0xc600, lo: 0x89, hi: 0xa3}, + // Block 0x61, offset 0x1de + {value: 0x0006, lo: 0x0d}, + {value: 0x4393, lo: 0x9d, hi: 0x9d}, + {value: 0x8115, lo: 0x9e, hi: 0x9e}, + {value: 0x4405, lo: 0x9f, hi: 0x9f}, + {value: 0x43f3, lo: 0xaa, hi: 0xab}, + {value: 0x44f7, lo: 0xac, hi: 0xac}, + {value: 0x44ff, lo: 0xad, hi: 0xad}, + {value: 0x434b, lo: 0xae, hi: 0xb1}, + {value: 0x4369, lo: 0xb2, hi: 0xb4}, + {value: 0x4381, lo: 0xb5, hi: 0xb6}, + {value: 0x438d, lo: 0xb8, hi: 0xb8}, + {value: 0x4399, lo: 0xb9, hi: 0xbb}, + {value: 0x43b1, lo: 0xbc, hi: 0xbc}, + {value: 0x43b7, lo: 0xbe, hi: 0xbe}, + // Block 0x62, offset 0x1ec + {value: 0x0006, lo: 0x08}, + {value: 0x43bd, lo: 0x80, hi: 0x81}, + {value: 0x43c9, lo: 0x83, hi: 0x84}, + {value: 0x43db, lo: 0x86, hi: 0x89}, + {value: 0x43ff, lo: 0x8a, hi: 0x8a}, + {value: 0x437b, lo: 0x8b, hi: 0x8b}, + {value: 0x4363, lo: 0x8c, hi: 0x8c}, + {value: 0x43ab, lo: 0x8d, hi: 0x8d}, + {value: 0x43d5, lo: 0x8e, hi: 0x8e}, + // Block 0x63, offset 0x1f5 + {value: 0x0000, lo: 0x02}, + {value: 0x8100, lo: 0xa4, hi: 0xa5}, + {value: 0x8100, lo: 0xb0, hi: 0xb1}, + // Block 0x64, offset 0x1f8 + {value: 0x0000, lo: 0x02}, + {value: 0x8100, lo: 0x9b, hi: 0x9d}, + {value: 0x8200, lo: 0x9e, hi: 0xa3}, + // Block 0x65, offset 0x1fb + {value: 0x0000, lo: 0x01}, + {value: 0x8100, lo: 0x90, hi: 0x90}, + // Block 0x66, offset 0x1fd + {value: 0x0000, lo: 0x02}, + {value: 0x8100, lo: 0x99, hi: 0x99}, + {value: 0x8200, lo: 0xb2, hi: 0xb4}, + // Block 0x67, offset 0x200 + {value: 0x0000, lo: 0x01}, + {value: 0x8100, lo: 0xbc, hi: 0xbd}, + // Block 0x68, offset 0x202 + {value: 0x0000, lo: 0x03}, + {value: 0x8132, lo: 0xa0, hi: 0xa6}, + {value: 0x812d, lo: 0xa7, hi: 0xad}, + {value: 0x8132, lo: 0xae, hi: 0xaf}, + // Block 0x69, offset 0x206 + {value: 0x0000, lo: 0x04}, + {value: 0x8100, lo: 0x89, hi: 0x8c}, + {value: 0x8100, lo: 0xb0, hi: 0xb2}, + {value: 0x8100, lo: 0xb4, hi: 0xb4}, + {value: 0x8100, lo: 0xb6, hi: 0xbf}, + // Block 0x6a, offset 0x20b + {value: 0x0000, lo: 0x01}, + {value: 0x8100, lo: 0x81, hi: 0x8c}, + // Block 0x6b, offset 0x20d + {value: 0x0000, lo: 0x01}, + {value: 0x8100, lo: 0xb5, hi: 0xba}, + // Block 0x6c, offset 0x20f + {value: 0x0000, lo: 0x04}, + {value: 0x4aa2, lo: 0x9e, hi: 0x9f}, + {value: 0x4aa2, lo: 0xa3, hi: 0xa3}, + {value: 0x4aa2, lo: 0xa5, hi: 0xa6}, + {value: 0x4aa2, lo: 0xaa, hi: 0xaf}, + // Block 0x6d, offset 0x214 + {value: 0x0000, lo: 0x05}, + {value: 0x4aa2, lo: 0x82, hi: 0x87}, + {value: 0x4aa2, lo: 0x8a, hi: 0x8f}, + {value: 0x4aa2, lo: 0x92, hi: 0x97}, + {value: 0x4aa2, lo: 0x9a, hi: 0x9c}, + {value: 0x8100, lo: 0xa3, hi: 0xa3}, + // Block 0x6e, offset 0x21a + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0xbd, hi: 0xbd}, + // Block 0x6f, offset 0x21c + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0xa0, hi: 0xa0}, + // Block 0x70, offset 0x21e + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xb6, hi: 0xba}, + // Block 0x71, offset 0x220 + {value: 0x002c, lo: 0x05}, + {value: 0x812d, lo: 0x8d, hi: 0x8d}, + {value: 0x8132, lo: 0x8f, hi: 0x8f}, + {value: 0x8132, lo: 0xb8, hi: 0xb8}, + {value: 0x8101, lo: 0xb9, hi: 0xba}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x72, offset 0x226 + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0xa5, hi: 0xa5}, + {value: 0x812d, lo: 0xa6, hi: 0xa6}, + // Block 0x73, offset 0x229 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xa4, hi: 0xa7}, + // Block 0x74, offset 0x22b + {value: 0x0000, lo: 0x05}, + {value: 0x812d, lo: 0x86, hi: 0x87}, + {value: 0x8132, lo: 0x88, hi: 0x8a}, + {value: 0x812d, lo: 0x8b, hi: 0x8b}, + {value: 0x8132, lo: 0x8c, hi: 0x8c}, + {value: 0x812d, lo: 0x8d, hi: 0x90}, + // Block 0x75, offset 0x231 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x86, hi: 0x86}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x76, offset 0x234 + {value: 0x17fe, lo: 0x07}, + {value: 0xa000, lo: 0x99, hi: 0x99}, + {value: 0x423b, lo: 0x9a, hi: 0x9a}, + {value: 0xa000, lo: 0x9b, hi: 0x9b}, + {value: 0x4245, lo: 0x9c, hi: 0x9c}, + {value: 0xa000, lo: 0xa5, hi: 0xa5}, + {value: 0x424f, lo: 0xab, hi: 0xab}, + {value: 0x8104, lo: 0xb9, hi: 0xba}, + // Block 0x77, offset 0x23c + {value: 0x0000, lo: 0x06}, + {value: 0x8132, lo: 0x80, hi: 0x82}, + {value: 0x9900, lo: 0xa7, hi: 0xa7}, + {value: 0x2d81, lo: 0xae, hi: 0xae}, + {value: 0x2d8b, lo: 0xaf, hi: 0xaf}, + {value: 0xa000, lo: 0xb1, hi: 0xb2}, + {value: 0x8104, lo: 0xb3, hi: 0xb4}, + // Block 0x78, offset 0x243 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x80, hi: 0x80}, + {value: 0x8102, lo: 0x8a, hi: 0x8a}, + // Block 0x79, offset 0x246 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xb5, hi: 0xb5}, + {value: 0x8102, lo: 0xb6, hi: 0xb6}, + // Block 0x7a, offset 0x249 + {value: 0x0002, lo: 0x01}, + {value: 0x8102, lo: 0xa9, hi: 0xaa}, + // Block 0x7b, offset 0x24b + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xbb, hi: 0xbc}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x7c, offset 0x24e + {value: 0x0000, lo: 0x07}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0x2d95, lo: 0x8b, hi: 0x8b}, + {value: 0x2d9f, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + {value: 0x8132, lo: 0xa6, hi: 0xac}, + {value: 0x8132, lo: 0xb0, hi: 0xb4}, + // Block 0x7d, offset 0x256 + {value: 0x0000, lo: 0x03}, + {value: 0x8104, lo: 0x82, hi: 0x82}, + {value: 0x8102, lo: 0x86, hi: 0x86}, + {value: 0x8132, lo: 0x9e, hi: 0x9e}, + // Block 0x7e, offset 0x25a + {value: 0x6b57, lo: 0x06}, + {value: 0x9900, lo: 0xb0, hi: 0xb0}, + {value: 0xa000, lo: 0xb9, hi: 0xb9}, + {value: 0x9900, lo: 0xba, hi: 0xba}, + {value: 0x2db3, lo: 0xbb, hi: 0xbb}, + {value: 0x2da9, lo: 0xbc, hi: 0xbd}, + {value: 0x2dbd, lo: 0xbe, hi: 0xbe}, + // Block 0x7f, offset 0x261 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x82, hi: 0x82}, + {value: 0x8102, lo: 0x83, hi: 0x83}, + // Block 0x80, offset 0x264 + {value: 0x0000, lo: 0x05}, + {value: 0x9900, lo: 0xaf, hi: 0xaf}, + {value: 0xa000, lo: 0xb8, hi: 0xb9}, + {value: 0x2dc7, lo: 0xba, hi: 0xba}, + {value: 0x2dd1, lo: 0xbb, hi: 0xbb}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x81, offset 0x26a + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0x80, hi: 0x80}, + // Block 0x82, offset 0x26c + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xb6, hi: 0xb6}, + {value: 0x8102, lo: 0xb7, hi: 0xb7}, + // Block 0x83, offset 0x26f + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xab, hi: 0xab}, + // Block 0x84, offset 0x271 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xb9, hi: 0xb9}, + {value: 0x8102, lo: 0xba, hi: 0xba}, + // Block 0x85, offset 0x274 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xa0, hi: 0xa0}, + // Block 0x86, offset 0x276 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xb4, hi: 0xb4}, + // Block 0x87, offset 0x278 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x87, hi: 0x87}, + // Block 0x88, offset 0x27a + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x99, hi: 0x99}, + // Block 0x89, offset 0x27c + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0x82, hi: 0x82}, + {value: 0x8104, lo: 0x84, hi: 0x85}, + // Block 0x8a, offset 0x27f + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x97, hi: 0x97}, + // Block 0x8b, offset 0x281 + {value: 0x0000, lo: 0x01}, + {value: 0x8101, lo: 0xb0, hi: 0xb4}, + // Block 0x8c, offset 0x283 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xb0, hi: 0xb6}, + // Block 0x8d, offset 0x285 + {value: 0x0000, lo: 0x01}, + {value: 0x8101, lo: 0x9e, hi: 0x9e}, + // Block 0x8e, offset 0x287 + {value: 0x0000, lo: 0x0c}, + {value: 0x45cf, lo: 0x9e, hi: 0x9e}, + {value: 0x45d9, lo: 0x9f, hi: 0x9f}, + {value: 0x460d, lo: 0xa0, hi: 0xa0}, + {value: 0x461b, lo: 0xa1, hi: 0xa1}, + {value: 0x4629, lo: 0xa2, hi: 0xa2}, + {value: 0x4637, lo: 0xa3, hi: 0xa3}, + {value: 0x4645, lo: 0xa4, hi: 0xa4}, + {value: 0x812b, lo: 0xa5, hi: 0xa6}, + {value: 0x8101, lo: 0xa7, hi: 0xa9}, + {value: 0x8130, lo: 0xad, hi: 0xad}, + {value: 0x812b, lo: 0xae, hi: 0xb2}, + {value: 0x812d, lo: 0xbb, hi: 0xbf}, + // Block 0x8f, offset 0x294 + {value: 0x0000, lo: 0x09}, + {value: 0x812d, lo: 0x80, hi: 0x82}, + {value: 0x8132, lo: 0x85, hi: 0x89}, + {value: 0x812d, lo: 0x8a, hi: 0x8b}, + {value: 0x8132, lo: 0xaa, hi: 0xad}, + {value: 0x45e3, lo: 0xbb, hi: 0xbb}, + {value: 0x45ed, lo: 0xbc, hi: 0xbc}, + {value: 0x4653, lo: 0xbd, hi: 0xbd}, + {value: 0x466f, lo: 0xbe, hi: 0xbe}, + {value: 0x4661, lo: 0xbf, hi: 0xbf}, + // Block 0x90, offset 0x29e + {value: 0x0000, lo: 0x01}, + {value: 0x467d, lo: 0x80, hi: 0x80}, + // Block 0x91, offset 0x2a0 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0x82, hi: 0x84}, + // Block 0x92, offset 0x2a2 + {value: 0x0000, lo: 0x05}, + {value: 0x8132, lo: 0x80, hi: 0x86}, + {value: 0x8132, lo: 0x88, hi: 0x98}, + {value: 0x8132, lo: 0x9b, hi: 0xa1}, + {value: 0x8132, lo: 0xa3, hi: 0xa4}, + {value: 0x8132, lo: 0xa6, hi: 0xaa}, + // Block 0x93, offset 0x2a8 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xac, hi: 0xaf}, + // Block 0x94, offset 0x2aa + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x90, hi: 0x96}, + // Block 0x95, offset 0x2ac + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0x84, hi: 0x89}, + {value: 0x8102, lo: 0x8a, hi: 0x8a}, + // Block 0x96, offset 0x2af + {value: 0x0000, lo: 0x01}, + {value: 0x8100, lo: 0x93, hi: 0x93}, +} + +// lookup returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *nfkcTrie) lookup(s []byte) (v uint16, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return nfkcValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := nfkcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := nfkcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfkcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := nfkcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfkcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = nfkcIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *nfkcTrie) lookupUnsafe(s []byte) uint16 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return nfkcValues[c0] + } + i := nfkcIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = nfkcIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = nfkcIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// lookupString returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *nfkcTrie) lookupString(s string) (v uint16, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return nfkcValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := nfkcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := nfkcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfkcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := nfkcIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = nfkcIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = nfkcIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupStringUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *nfkcTrie) lookupStringUnsafe(s string) uint16 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return nfkcValues[c0] + } + i := nfkcIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = nfkcIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = nfkcIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// nfkcTrie. Total size: 18684 bytes (18.25 KiB). Checksum: 113e23c477adfabd. +type nfkcTrie struct{} + +func newNfkcTrie(i int) *nfkcTrie { + return &nfkcTrie{} +} + +// lookupValue determines the type of block n and looks up the value for b. +func (t *nfkcTrie) lookupValue(n uint32, b byte) uint16 { + switch { + case n < 92: + return uint16(nfkcValues[n<<6+uint32(b)]) + default: + n -= 92 + return uint16(nfkcSparse.lookup(n, b)) + } +} + +// nfkcValues: 94 blocks, 6016 entries, 12032 bytes +// The third block is the zero block. +var nfkcValues = [6016]uint16{ + // Block 0x0, offset 0x0 + 0x3c: 0xa000, 0x3d: 0xa000, 0x3e: 0xa000, + // Block 0x1, offset 0x40 + 0x41: 0xa000, 0x42: 0xa000, 0x43: 0xa000, 0x44: 0xa000, 0x45: 0xa000, + 0x46: 0xa000, 0x47: 0xa000, 0x48: 0xa000, 0x49: 0xa000, 0x4a: 0xa000, 0x4b: 0xa000, + 0x4c: 0xa000, 0x4d: 0xa000, 0x4e: 0xa000, 0x4f: 0xa000, 0x50: 0xa000, + 0x52: 0xa000, 0x53: 0xa000, 0x54: 0xa000, 0x55: 0xa000, 0x56: 0xa000, 0x57: 0xa000, + 0x58: 0xa000, 0x59: 0xa000, 0x5a: 0xa000, + 0x61: 0xa000, 0x62: 0xa000, 0x63: 0xa000, + 0x64: 0xa000, 0x65: 0xa000, 0x66: 0xa000, 0x67: 0xa000, 0x68: 0xa000, 0x69: 0xa000, + 0x6a: 0xa000, 0x6b: 0xa000, 0x6c: 0xa000, 0x6d: 0xa000, 0x6e: 0xa000, 0x6f: 0xa000, + 0x70: 0xa000, 0x72: 0xa000, 0x73: 0xa000, 0x74: 0xa000, 0x75: 0xa000, + 0x76: 0xa000, 0x77: 0xa000, 0x78: 0xa000, 0x79: 0xa000, 0x7a: 0xa000, + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xc0: 0x2f72, 0xc1: 0x2f77, 0xc2: 0x468b, 0xc3: 0x2f7c, 0xc4: 0x469a, 0xc5: 0x469f, + 0xc6: 0xa000, 0xc7: 0x46a9, 0xc8: 0x2fe5, 0xc9: 0x2fea, 0xca: 0x46ae, 0xcb: 0x2ffe, + 0xcc: 0x3071, 0xcd: 0x3076, 0xce: 0x307b, 0xcf: 0x46c2, 0xd1: 0x3107, + 0xd2: 0x312a, 0xd3: 0x312f, 0xd4: 0x46cc, 0xd5: 0x46d1, 0xd6: 0x46e0, + 0xd8: 0xa000, 0xd9: 0x31b6, 0xda: 0x31bb, 0xdb: 0x31c0, 0xdc: 0x4712, 0xdd: 0x3238, + 0xe0: 0x327e, 0xe1: 0x3283, 0xe2: 0x471c, 0xe3: 0x3288, + 0xe4: 0x472b, 0xe5: 0x4730, 0xe6: 0xa000, 0xe7: 0x473a, 0xe8: 0x32f1, 0xe9: 0x32f6, + 0xea: 0x473f, 0xeb: 0x330a, 0xec: 0x3382, 0xed: 0x3387, 0xee: 0x338c, 0xef: 0x4753, + 0xf1: 0x3418, 0xf2: 0x343b, 0xf3: 0x3440, 0xf4: 0x475d, 0xf5: 0x4762, + 0xf6: 0x4771, 0xf8: 0xa000, 0xf9: 0x34cc, 0xfa: 0x34d1, 0xfb: 0x34d6, + 0xfc: 0x47a3, 0xfd: 0x3553, 0xff: 0x356c, + // Block 0x4, offset 0x100 + 0x100: 0x2f81, 0x101: 0x328d, 0x102: 0x4690, 0x103: 0x4721, 0x104: 0x2f9f, 0x105: 0x32ab, + 0x106: 0x2fb3, 0x107: 0x32bf, 0x108: 0x2fb8, 0x109: 0x32c4, 0x10a: 0x2fbd, 0x10b: 0x32c9, + 0x10c: 0x2fc2, 0x10d: 0x32ce, 0x10e: 0x2fcc, 0x10f: 0x32d8, + 0x112: 0x46b3, 0x113: 0x4744, 0x114: 0x2ff4, 0x115: 0x3300, 0x116: 0x2ff9, 0x117: 0x3305, + 0x118: 0x3017, 0x119: 0x3323, 0x11a: 0x3008, 0x11b: 0x3314, 0x11c: 0x3030, 0x11d: 0x333c, + 0x11e: 0x303a, 0x11f: 0x3346, 0x120: 0x303f, 0x121: 0x334b, 0x122: 0x3049, 0x123: 0x3355, + 0x124: 0x304e, 0x125: 0x335a, 0x128: 0x3080, 0x129: 0x3391, + 0x12a: 0x3085, 0x12b: 0x3396, 0x12c: 0x308a, 0x12d: 0x339b, 0x12e: 0x30ad, 0x12f: 0x33b9, + 0x130: 0x308f, 0x132: 0x195d, 0x133: 0x19ea, 0x134: 0x30b7, 0x135: 0x33c3, + 0x136: 0x30cb, 0x137: 0x33dc, 0x139: 0x30d5, 0x13a: 0x33e6, 0x13b: 0x30df, + 0x13c: 0x33f0, 0x13d: 0x30da, 0x13e: 0x33eb, 0x13f: 0x1baf, + // Block 0x5, offset 0x140 + 0x140: 0x1c37, 0x143: 0x3102, 0x144: 0x3413, 0x145: 0x311b, + 0x146: 0x342c, 0x147: 0x3111, 0x148: 0x3422, 0x149: 0x1c5f, + 0x14c: 0x46d6, 0x14d: 0x4767, 0x14e: 0x3134, 0x14f: 0x3445, 0x150: 0x313e, 0x151: 0x344f, + 0x154: 0x315c, 0x155: 0x346d, 0x156: 0x3175, 0x157: 0x3486, + 0x158: 0x3166, 0x159: 0x3477, 0x15a: 0x46f9, 0x15b: 0x478a, 0x15c: 0x317f, 0x15d: 0x3490, + 0x15e: 0x318e, 0x15f: 0x349f, 0x160: 0x46fe, 0x161: 0x478f, 0x162: 0x31a7, 0x163: 0x34bd, + 0x164: 0x3198, 0x165: 0x34ae, 0x168: 0x4708, 0x169: 0x4799, + 0x16a: 0x470d, 0x16b: 0x479e, 0x16c: 0x31c5, 0x16d: 0x34db, 0x16e: 0x31cf, 0x16f: 0x34e5, + 0x170: 0x31d4, 0x171: 0x34ea, 0x172: 0x31f2, 0x173: 0x3508, 0x174: 0x3215, 0x175: 0x352b, + 0x176: 0x323d, 0x177: 0x3558, 0x178: 0x3251, 0x179: 0x3260, 0x17a: 0x3580, 0x17b: 0x326a, + 0x17c: 0x358a, 0x17d: 0x326f, 0x17e: 0x358f, 0x17f: 0x00a7, + // Block 0x6, offset 0x180 + 0x184: 0x2df1, 0x185: 0x2df7, + 0x186: 0x2dfd, 0x187: 0x1972, 0x188: 0x1975, 0x189: 0x1a0b, 0x18a: 0x198a, 0x18b: 0x198d, + 0x18c: 0x1a41, 0x18d: 0x2f8b, 0x18e: 0x3297, 0x18f: 0x3099, 0x190: 0x33a5, 0x191: 0x3143, + 0x192: 0x3454, 0x193: 0x31d9, 0x194: 0x34ef, 0x195: 0x39d2, 0x196: 0x3b61, 0x197: 0x39cb, + 0x198: 0x3b5a, 0x199: 0x39d9, 0x19a: 0x3b68, 0x19b: 0x39c4, 0x19c: 0x3b53, + 0x19e: 0x38b3, 0x19f: 0x3a42, 0x1a0: 0x38ac, 0x1a1: 0x3a3b, 0x1a2: 0x35b6, 0x1a3: 0x35c8, + 0x1a6: 0x3044, 0x1a7: 0x3350, 0x1a8: 0x30c1, 0x1a9: 0x33d2, + 0x1aa: 0x46ef, 0x1ab: 0x4780, 0x1ac: 0x3993, 0x1ad: 0x3b22, 0x1ae: 0x35da, 0x1af: 0x35e0, + 0x1b0: 0x33c8, 0x1b1: 0x1942, 0x1b2: 0x1945, 0x1b3: 0x19d2, 0x1b4: 0x302b, 0x1b5: 0x3337, + 0x1b8: 0x30fd, 0x1b9: 0x340e, 0x1ba: 0x38ba, 0x1bb: 0x3a49, + 0x1bc: 0x35b0, 0x1bd: 0x35c2, 0x1be: 0x35bc, 0x1bf: 0x35ce, + // Block 0x7, offset 0x1c0 + 0x1c0: 0x2f90, 0x1c1: 0x329c, 0x1c2: 0x2f95, 0x1c3: 0x32a1, 0x1c4: 0x300d, 0x1c5: 0x3319, + 0x1c6: 0x3012, 0x1c7: 0x331e, 0x1c8: 0x309e, 0x1c9: 0x33aa, 0x1ca: 0x30a3, 0x1cb: 0x33af, + 0x1cc: 0x3148, 0x1cd: 0x3459, 0x1ce: 0x314d, 0x1cf: 0x345e, 0x1d0: 0x316b, 0x1d1: 0x347c, + 0x1d2: 0x3170, 0x1d3: 0x3481, 0x1d4: 0x31de, 0x1d5: 0x34f4, 0x1d6: 0x31e3, 0x1d7: 0x34f9, + 0x1d8: 0x3189, 0x1d9: 0x349a, 0x1da: 0x31a2, 0x1db: 0x34b8, + 0x1de: 0x305d, 0x1df: 0x3369, + 0x1e6: 0x4695, 0x1e7: 0x4726, 0x1e8: 0x46bd, 0x1e9: 0x474e, + 0x1ea: 0x3962, 0x1eb: 0x3af1, 0x1ec: 0x393f, 0x1ed: 0x3ace, 0x1ee: 0x46db, 0x1ef: 0x476c, + 0x1f0: 0x395b, 0x1f1: 0x3aea, 0x1f2: 0x3247, 0x1f3: 0x3562, + // Block 0x8, offset 0x200 + 0x200: 0x9932, 0x201: 0x9932, 0x202: 0x9932, 0x203: 0x9932, 0x204: 0x9932, 0x205: 0x8132, + 0x206: 0x9932, 0x207: 0x9932, 0x208: 0x9932, 0x209: 0x9932, 0x20a: 0x9932, 0x20b: 0x9932, + 0x20c: 0x9932, 0x20d: 0x8132, 0x20e: 0x8132, 0x20f: 0x9932, 0x210: 0x8132, 0x211: 0x9932, + 0x212: 0x8132, 0x213: 0x9932, 0x214: 0x9932, 0x215: 0x8133, 0x216: 0x812d, 0x217: 0x812d, + 0x218: 0x812d, 0x219: 0x812d, 0x21a: 0x8133, 0x21b: 0x992b, 0x21c: 0x812d, 0x21d: 0x812d, + 0x21e: 0x812d, 0x21f: 0x812d, 0x220: 0x812d, 0x221: 0x8129, 0x222: 0x8129, 0x223: 0x992d, + 0x224: 0x992d, 0x225: 0x992d, 0x226: 0x992d, 0x227: 0x9929, 0x228: 0x9929, 0x229: 0x812d, + 0x22a: 0x812d, 0x22b: 0x812d, 0x22c: 0x812d, 0x22d: 0x992d, 0x22e: 0x992d, 0x22f: 0x812d, + 0x230: 0x992d, 0x231: 0x992d, 0x232: 0x812d, 0x233: 0x812d, 0x234: 0x8101, 0x235: 0x8101, + 0x236: 0x8101, 0x237: 0x8101, 0x238: 0x9901, 0x239: 0x812d, 0x23a: 0x812d, 0x23b: 0x812d, + 0x23c: 0x812d, 0x23d: 0x8132, 0x23e: 0x8132, 0x23f: 0x8132, + // Block 0x9, offset 0x240 + 0x240: 0x49b1, 0x241: 0x49b6, 0x242: 0x9932, 0x243: 0x49bb, 0x244: 0x4a74, 0x245: 0x9936, + 0x246: 0x8132, 0x247: 0x812d, 0x248: 0x812d, 0x249: 0x812d, 0x24a: 0x8132, 0x24b: 0x8132, + 0x24c: 0x8132, 0x24d: 0x812d, 0x24e: 0x812d, 0x250: 0x8132, 0x251: 0x8132, + 0x252: 0x8132, 0x253: 0x812d, 0x254: 0x812d, 0x255: 0x812d, 0x256: 0x812d, 0x257: 0x8132, + 0x258: 0x8133, 0x259: 0x812d, 0x25a: 0x812d, 0x25b: 0x8132, 0x25c: 0x8134, 0x25d: 0x8135, + 0x25e: 0x8135, 0x25f: 0x8134, 0x260: 0x8135, 0x261: 0x8135, 0x262: 0x8134, 0x263: 0x8132, + 0x264: 0x8132, 0x265: 0x8132, 0x266: 0x8132, 0x267: 0x8132, 0x268: 0x8132, 0x269: 0x8132, + 0x26a: 0x8132, 0x26b: 0x8132, 0x26c: 0x8132, 0x26d: 0x8132, 0x26e: 0x8132, 0x26f: 0x8132, + 0x274: 0x0170, + 0x27a: 0x42a8, + 0x27e: 0x0037, + // Block 0xa, offset 0x280 + 0x284: 0x425d, 0x285: 0x447e, + 0x286: 0x35ec, 0x287: 0x00ce, 0x288: 0x360a, 0x289: 0x3616, 0x28a: 0x3628, + 0x28c: 0x3646, 0x28e: 0x3658, 0x28f: 0x3676, 0x290: 0x3e0b, 0x291: 0xa000, + 0x295: 0xa000, 0x297: 0xa000, + 0x299: 0xa000, + 0x29f: 0xa000, 0x2a1: 0xa000, + 0x2a5: 0xa000, 0x2a9: 0xa000, + 0x2aa: 0x363a, 0x2ab: 0x366a, 0x2ac: 0x4801, 0x2ad: 0x369a, 0x2ae: 0x482b, 0x2af: 0x36ac, + 0x2b0: 0x3e73, 0x2b1: 0xa000, 0x2b5: 0xa000, + 0x2b7: 0xa000, 0x2b9: 0xa000, + 0x2bf: 0xa000, + // Block 0xb, offset 0x2c0 + 0x2c1: 0xa000, 0x2c5: 0xa000, + 0x2c9: 0xa000, 0x2ca: 0x4843, 0x2cb: 0x4861, + 0x2cc: 0x36ca, 0x2cd: 0x36e2, 0x2ce: 0x4879, 0x2d0: 0x01be, 0x2d1: 0x01d0, + 0x2d2: 0x01ac, 0x2d3: 0x430f, 0x2d4: 0x4315, 0x2d5: 0x01fa, 0x2d6: 0x01e8, + 0x2f0: 0x01d6, 0x2f1: 0x01eb, 0x2f2: 0x01ee, 0x2f4: 0x0188, 0x2f5: 0x01c7, + 0x2f9: 0x01a6, + // Block 0xc, offset 0x300 + 0x300: 0x3724, 0x301: 0x3730, 0x303: 0x371e, + 0x306: 0xa000, 0x307: 0x370c, + 0x30c: 0x3760, 0x30d: 0x3748, 0x30e: 0x3772, 0x310: 0xa000, + 0x313: 0xa000, 0x315: 0xa000, 0x316: 0xa000, 0x317: 0xa000, + 0x318: 0xa000, 0x319: 0x3754, 0x31a: 0xa000, + 0x31e: 0xa000, 0x323: 0xa000, + 0x327: 0xa000, + 0x32b: 0xa000, 0x32d: 0xa000, + 0x330: 0xa000, 0x333: 0xa000, 0x335: 0xa000, + 0x336: 0xa000, 0x337: 0xa000, 0x338: 0xa000, 0x339: 0x37d8, 0x33a: 0xa000, + 0x33e: 0xa000, + // Block 0xd, offset 0x340 + 0x341: 0x3736, 0x342: 0x37ba, + 0x350: 0x3712, 0x351: 0x3796, + 0x352: 0x3718, 0x353: 0x379c, 0x356: 0x372a, 0x357: 0x37ae, + 0x358: 0xa000, 0x359: 0xa000, 0x35a: 0x382c, 0x35b: 0x3832, 0x35c: 0x373c, 0x35d: 0x37c0, + 0x35e: 0x3742, 0x35f: 0x37c6, 0x362: 0x374e, 0x363: 0x37d2, + 0x364: 0x375a, 0x365: 0x37de, 0x366: 0x3766, 0x367: 0x37ea, 0x368: 0xa000, 0x369: 0xa000, + 0x36a: 0x3838, 0x36b: 0x383e, 0x36c: 0x3790, 0x36d: 0x3814, 0x36e: 0x376c, 0x36f: 0x37f0, + 0x370: 0x3778, 0x371: 0x37fc, 0x372: 0x377e, 0x373: 0x3802, 0x374: 0x3784, 0x375: 0x3808, + 0x378: 0x378a, 0x379: 0x380e, + // Block 0xe, offset 0x380 + 0x387: 0x1d64, + 0x391: 0x812d, + 0x392: 0x8132, 0x393: 0x8132, 0x394: 0x8132, 0x395: 0x8132, 0x396: 0x812d, 0x397: 0x8132, + 0x398: 0x8132, 0x399: 0x8132, 0x39a: 0x812e, 0x39b: 0x812d, 0x39c: 0x8132, 0x39d: 0x8132, + 0x39e: 0x8132, 0x39f: 0x8132, 0x3a0: 0x8132, 0x3a1: 0x8132, 0x3a2: 0x812d, 0x3a3: 0x812d, + 0x3a4: 0x812d, 0x3a5: 0x812d, 0x3a6: 0x812d, 0x3a7: 0x812d, 0x3a8: 0x8132, 0x3a9: 0x8132, + 0x3aa: 0x812d, 0x3ab: 0x8132, 0x3ac: 0x8132, 0x3ad: 0x812e, 0x3ae: 0x8131, 0x3af: 0x8132, + 0x3b0: 0x8105, 0x3b1: 0x8106, 0x3b2: 0x8107, 0x3b3: 0x8108, 0x3b4: 0x8109, 0x3b5: 0x810a, + 0x3b6: 0x810b, 0x3b7: 0x810c, 0x3b8: 0x810d, 0x3b9: 0x810e, 0x3ba: 0x810e, 0x3bb: 0x810f, + 0x3bc: 0x8110, 0x3bd: 0x8111, 0x3bf: 0x8112, + // Block 0xf, offset 0x3c0 + 0x3c8: 0xa000, 0x3ca: 0xa000, 0x3cb: 0x8116, + 0x3cc: 0x8117, 0x3cd: 0x8118, 0x3ce: 0x8119, 0x3cf: 0x811a, 0x3d0: 0x811b, 0x3d1: 0x811c, + 0x3d2: 0x811d, 0x3d3: 0x9932, 0x3d4: 0x9932, 0x3d5: 0x992d, 0x3d6: 0x812d, 0x3d7: 0x8132, + 0x3d8: 0x8132, 0x3d9: 0x8132, 0x3da: 0x8132, 0x3db: 0x8132, 0x3dc: 0x812d, 0x3dd: 0x8132, + 0x3de: 0x8132, 0x3df: 0x812d, + 0x3f0: 0x811e, 0x3f5: 0x1d87, + 0x3f6: 0x2016, 0x3f7: 0x2052, 0x3f8: 0x204d, + // Block 0x10, offset 0x400 + 0x413: 0x812d, 0x414: 0x8132, 0x415: 0x8132, 0x416: 0x8132, 0x417: 0x8132, + 0x418: 0x8132, 0x419: 0x8132, 0x41a: 0x8132, 0x41b: 0x8132, 0x41c: 0x8132, 0x41d: 0x8132, + 0x41e: 0x8132, 0x41f: 0x8132, 0x420: 0x8132, 0x421: 0x8132, 0x423: 0x812d, + 0x424: 0x8132, 0x425: 0x8132, 0x426: 0x812d, 0x427: 0x8132, 0x428: 0x8132, 0x429: 0x812d, + 0x42a: 0x8132, 0x42b: 0x8132, 0x42c: 0x8132, 0x42d: 0x812d, 0x42e: 0x812d, 0x42f: 0x812d, + 0x430: 0x8116, 0x431: 0x8117, 0x432: 0x8118, 0x433: 0x8132, 0x434: 0x8132, 0x435: 0x8132, + 0x436: 0x812d, 0x437: 0x8132, 0x438: 0x8132, 0x439: 0x812d, 0x43a: 0x812d, 0x43b: 0x8132, + 0x43c: 0x8132, 0x43d: 0x8132, 0x43e: 0x8132, 0x43f: 0x8132, + // Block 0x11, offset 0x440 + 0x445: 0xa000, + 0x446: 0x2d29, 0x447: 0xa000, 0x448: 0x2d31, 0x449: 0xa000, 0x44a: 0x2d39, 0x44b: 0xa000, + 0x44c: 0x2d41, 0x44d: 0xa000, 0x44e: 0x2d49, 0x451: 0xa000, + 0x452: 0x2d51, + 0x474: 0x8102, 0x475: 0x9900, + 0x47a: 0xa000, 0x47b: 0x2d59, + 0x47c: 0xa000, 0x47d: 0x2d61, 0x47e: 0xa000, 0x47f: 0xa000, + // Block 0x12, offset 0x480 + 0x480: 0x0069, 0x481: 0x006b, 0x482: 0x006f, 0x483: 0x0083, 0x484: 0x00f5, 0x485: 0x00f8, + 0x486: 0x0413, 0x487: 0x0085, 0x488: 0x0089, 0x489: 0x008b, 0x48a: 0x0104, 0x48b: 0x0107, + 0x48c: 0x010a, 0x48d: 0x008f, 0x48f: 0x0097, 0x490: 0x009b, 0x491: 0x00e0, + 0x492: 0x009f, 0x493: 0x00fe, 0x494: 0x0417, 0x495: 0x041b, 0x496: 0x00a1, 0x497: 0x00a9, + 0x498: 0x00ab, 0x499: 0x0423, 0x49a: 0x012b, 0x49b: 0x00ad, 0x49c: 0x0427, 0x49d: 0x01be, + 0x49e: 0x01c1, 0x49f: 0x01c4, 0x4a0: 0x01fa, 0x4a1: 0x01fd, 0x4a2: 0x0093, 0x4a3: 0x00a5, + 0x4a4: 0x00ab, 0x4a5: 0x00ad, 0x4a6: 0x01be, 0x4a7: 0x01c1, 0x4a8: 0x01eb, 0x4a9: 0x01fa, + 0x4aa: 0x01fd, + 0x4b8: 0x020c, + // Block 0x13, offset 0x4c0 + 0x4db: 0x00fb, 0x4dc: 0x0087, 0x4dd: 0x0101, + 0x4de: 0x00d4, 0x4df: 0x010a, 0x4e0: 0x008d, 0x4e1: 0x010d, 0x4e2: 0x0110, 0x4e3: 0x0116, + 0x4e4: 0x011c, 0x4e5: 0x011f, 0x4e6: 0x0122, 0x4e7: 0x042b, 0x4e8: 0x016a, 0x4e9: 0x0128, + 0x4ea: 0x042f, 0x4eb: 0x016d, 0x4ec: 0x0131, 0x4ed: 0x012e, 0x4ee: 0x0134, 0x4ef: 0x0137, + 0x4f0: 0x013a, 0x4f1: 0x013d, 0x4f2: 0x0140, 0x4f3: 0x014c, 0x4f4: 0x014f, 0x4f5: 0x00ec, + 0x4f6: 0x0152, 0x4f7: 0x0155, 0x4f8: 0x041f, 0x4f9: 0x0158, 0x4fa: 0x015b, 0x4fb: 0x00b5, + 0x4fc: 0x015e, 0x4fd: 0x0161, 0x4fe: 0x0164, 0x4ff: 0x01d0, + // Block 0x14, offset 0x500 + 0x500: 0x8132, 0x501: 0x8132, 0x502: 0x812d, 0x503: 0x8132, 0x504: 0x8132, 0x505: 0x8132, + 0x506: 0x8132, 0x507: 0x8132, 0x508: 0x8132, 0x509: 0x8132, 0x50a: 0x812d, 0x50b: 0x8132, + 0x50c: 0x8132, 0x50d: 0x8135, 0x50e: 0x812a, 0x50f: 0x812d, 0x510: 0x8129, 0x511: 0x8132, + 0x512: 0x8132, 0x513: 0x8132, 0x514: 0x8132, 0x515: 0x8132, 0x516: 0x8132, 0x517: 0x8132, + 0x518: 0x8132, 0x519: 0x8132, 0x51a: 0x8132, 0x51b: 0x8132, 0x51c: 0x8132, 0x51d: 0x8132, + 0x51e: 0x8132, 0x51f: 0x8132, 0x520: 0x8132, 0x521: 0x8132, 0x522: 0x8132, 0x523: 0x8132, + 0x524: 0x8132, 0x525: 0x8132, 0x526: 0x8132, 0x527: 0x8132, 0x528: 0x8132, 0x529: 0x8132, + 0x52a: 0x8132, 0x52b: 0x8132, 0x52c: 0x8132, 0x52d: 0x8132, 0x52e: 0x8132, 0x52f: 0x8132, + 0x530: 0x8132, 0x531: 0x8132, 0x532: 0x8132, 0x533: 0x8132, 0x534: 0x8132, 0x535: 0x8132, + 0x536: 0x8133, 0x537: 0x8131, 0x538: 0x8131, 0x539: 0x812d, 0x53b: 0x8132, + 0x53c: 0x8134, 0x53d: 0x812d, 0x53e: 0x8132, 0x53f: 0x812d, + // Block 0x15, offset 0x540 + 0x540: 0x2f9a, 0x541: 0x32a6, 0x542: 0x2fa4, 0x543: 0x32b0, 0x544: 0x2fa9, 0x545: 0x32b5, + 0x546: 0x2fae, 0x547: 0x32ba, 0x548: 0x38cf, 0x549: 0x3a5e, 0x54a: 0x2fc7, 0x54b: 0x32d3, + 0x54c: 0x2fd1, 0x54d: 0x32dd, 0x54e: 0x2fe0, 0x54f: 0x32ec, 0x550: 0x2fd6, 0x551: 0x32e2, + 0x552: 0x2fdb, 0x553: 0x32e7, 0x554: 0x38f2, 0x555: 0x3a81, 0x556: 0x38f9, 0x557: 0x3a88, + 0x558: 0x301c, 0x559: 0x3328, 0x55a: 0x3021, 0x55b: 0x332d, 0x55c: 0x3907, 0x55d: 0x3a96, + 0x55e: 0x3026, 0x55f: 0x3332, 0x560: 0x3035, 0x561: 0x3341, 0x562: 0x3053, 0x563: 0x335f, + 0x564: 0x3062, 0x565: 0x336e, 0x566: 0x3058, 0x567: 0x3364, 0x568: 0x3067, 0x569: 0x3373, + 0x56a: 0x306c, 0x56b: 0x3378, 0x56c: 0x30b2, 0x56d: 0x33be, 0x56e: 0x390e, 0x56f: 0x3a9d, + 0x570: 0x30bc, 0x571: 0x33cd, 0x572: 0x30c6, 0x573: 0x33d7, 0x574: 0x30d0, 0x575: 0x33e1, + 0x576: 0x46c7, 0x577: 0x4758, 0x578: 0x3915, 0x579: 0x3aa4, 0x57a: 0x30e9, 0x57b: 0x33fa, + 0x57c: 0x30e4, 0x57d: 0x33f5, 0x57e: 0x30ee, 0x57f: 0x33ff, + // Block 0x16, offset 0x580 + 0x580: 0x30f3, 0x581: 0x3404, 0x582: 0x30f8, 0x583: 0x3409, 0x584: 0x310c, 0x585: 0x341d, + 0x586: 0x3116, 0x587: 0x3427, 0x588: 0x3125, 0x589: 0x3436, 0x58a: 0x3120, 0x58b: 0x3431, + 0x58c: 0x3938, 0x58d: 0x3ac7, 0x58e: 0x3946, 0x58f: 0x3ad5, 0x590: 0x394d, 0x591: 0x3adc, + 0x592: 0x3954, 0x593: 0x3ae3, 0x594: 0x3152, 0x595: 0x3463, 0x596: 0x3157, 0x597: 0x3468, + 0x598: 0x3161, 0x599: 0x3472, 0x59a: 0x46f4, 0x59b: 0x4785, 0x59c: 0x399a, 0x59d: 0x3b29, + 0x59e: 0x317a, 0x59f: 0x348b, 0x5a0: 0x3184, 0x5a1: 0x3495, 0x5a2: 0x4703, 0x5a3: 0x4794, + 0x5a4: 0x39a1, 0x5a5: 0x3b30, 0x5a6: 0x39a8, 0x5a7: 0x3b37, 0x5a8: 0x39af, 0x5a9: 0x3b3e, + 0x5aa: 0x3193, 0x5ab: 0x34a4, 0x5ac: 0x319d, 0x5ad: 0x34b3, 0x5ae: 0x31b1, 0x5af: 0x34c7, + 0x5b0: 0x31ac, 0x5b1: 0x34c2, 0x5b2: 0x31ed, 0x5b3: 0x3503, 0x5b4: 0x31fc, 0x5b5: 0x3512, + 0x5b6: 0x31f7, 0x5b7: 0x350d, 0x5b8: 0x39b6, 0x5b9: 0x3b45, 0x5ba: 0x39bd, 0x5bb: 0x3b4c, + 0x5bc: 0x3201, 0x5bd: 0x3517, 0x5be: 0x3206, 0x5bf: 0x351c, + // Block 0x17, offset 0x5c0 + 0x5c0: 0x320b, 0x5c1: 0x3521, 0x5c2: 0x3210, 0x5c3: 0x3526, 0x5c4: 0x321f, 0x5c5: 0x3535, + 0x5c6: 0x321a, 0x5c7: 0x3530, 0x5c8: 0x3224, 0x5c9: 0x353f, 0x5ca: 0x3229, 0x5cb: 0x3544, + 0x5cc: 0x322e, 0x5cd: 0x3549, 0x5ce: 0x324c, 0x5cf: 0x3567, 0x5d0: 0x3265, 0x5d1: 0x3585, + 0x5d2: 0x3274, 0x5d3: 0x3594, 0x5d4: 0x3279, 0x5d5: 0x3599, 0x5d6: 0x337d, 0x5d7: 0x34a9, + 0x5d8: 0x353a, 0x5d9: 0x3576, 0x5da: 0x1be3, 0x5db: 0x42da, + 0x5e0: 0x46a4, 0x5e1: 0x4735, 0x5e2: 0x2f86, 0x5e3: 0x3292, + 0x5e4: 0x387b, 0x5e5: 0x3a0a, 0x5e6: 0x3874, 0x5e7: 0x3a03, 0x5e8: 0x3889, 0x5e9: 0x3a18, + 0x5ea: 0x3882, 0x5eb: 0x3a11, 0x5ec: 0x38c1, 0x5ed: 0x3a50, 0x5ee: 0x3897, 0x5ef: 0x3a26, + 0x5f0: 0x3890, 0x5f1: 0x3a1f, 0x5f2: 0x38a5, 0x5f3: 0x3a34, 0x5f4: 0x389e, 0x5f5: 0x3a2d, + 0x5f6: 0x38c8, 0x5f7: 0x3a57, 0x5f8: 0x46b8, 0x5f9: 0x4749, 0x5fa: 0x3003, 0x5fb: 0x330f, + 0x5fc: 0x2fef, 0x5fd: 0x32fb, 0x5fe: 0x38dd, 0x5ff: 0x3a6c, + // Block 0x18, offset 0x600 + 0x600: 0x38d6, 0x601: 0x3a65, 0x602: 0x38eb, 0x603: 0x3a7a, 0x604: 0x38e4, 0x605: 0x3a73, + 0x606: 0x3900, 0x607: 0x3a8f, 0x608: 0x3094, 0x609: 0x33a0, 0x60a: 0x30a8, 0x60b: 0x33b4, + 0x60c: 0x46ea, 0x60d: 0x477b, 0x60e: 0x3139, 0x60f: 0x344a, 0x610: 0x3923, 0x611: 0x3ab2, + 0x612: 0x391c, 0x613: 0x3aab, 0x614: 0x3931, 0x615: 0x3ac0, 0x616: 0x392a, 0x617: 0x3ab9, + 0x618: 0x398c, 0x619: 0x3b1b, 0x61a: 0x3970, 0x61b: 0x3aff, 0x61c: 0x3969, 0x61d: 0x3af8, + 0x61e: 0x397e, 0x61f: 0x3b0d, 0x620: 0x3977, 0x621: 0x3b06, 0x622: 0x3985, 0x623: 0x3b14, + 0x624: 0x31e8, 0x625: 0x34fe, 0x626: 0x31ca, 0x627: 0x34e0, 0x628: 0x39e7, 0x629: 0x3b76, + 0x62a: 0x39e0, 0x62b: 0x3b6f, 0x62c: 0x39f5, 0x62d: 0x3b84, 0x62e: 0x39ee, 0x62f: 0x3b7d, + 0x630: 0x39fc, 0x631: 0x3b8b, 0x632: 0x3233, 0x633: 0x354e, 0x634: 0x325b, 0x635: 0x357b, + 0x636: 0x3256, 0x637: 0x3571, 0x638: 0x3242, 0x639: 0x355d, + // Block 0x19, offset 0x640 + 0x640: 0x4807, 0x641: 0x480d, 0x642: 0x4921, 0x643: 0x4939, 0x644: 0x4929, 0x645: 0x4941, + 0x646: 0x4931, 0x647: 0x4949, 0x648: 0x47ad, 0x649: 0x47b3, 0x64a: 0x4891, 0x64b: 0x48a9, + 0x64c: 0x4899, 0x64d: 0x48b1, 0x64e: 0x48a1, 0x64f: 0x48b9, 0x650: 0x4819, 0x651: 0x481f, + 0x652: 0x3dbb, 0x653: 0x3dcb, 0x654: 0x3dc3, 0x655: 0x3dd3, + 0x658: 0x47b9, 0x659: 0x47bf, 0x65a: 0x3ceb, 0x65b: 0x3cfb, 0x65c: 0x3cf3, 0x65d: 0x3d03, + 0x660: 0x4831, 0x661: 0x4837, 0x662: 0x4951, 0x663: 0x4969, + 0x664: 0x4959, 0x665: 0x4971, 0x666: 0x4961, 0x667: 0x4979, 0x668: 0x47c5, 0x669: 0x47cb, + 0x66a: 0x48c1, 0x66b: 0x48d9, 0x66c: 0x48c9, 0x66d: 0x48e1, 0x66e: 0x48d1, 0x66f: 0x48e9, + 0x670: 0x4849, 0x671: 0x484f, 0x672: 0x3e1b, 0x673: 0x3e33, 0x674: 0x3e23, 0x675: 0x3e3b, + 0x676: 0x3e2b, 0x677: 0x3e43, 0x678: 0x47d1, 0x679: 0x47d7, 0x67a: 0x3d1b, 0x67b: 0x3d33, + 0x67c: 0x3d23, 0x67d: 0x3d3b, 0x67e: 0x3d2b, 0x67f: 0x3d43, + // Block 0x1a, offset 0x680 + 0x680: 0x4855, 0x681: 0x485b, 0x682: 0x3e4b, 0x683: 0x3e5b, 0x684: 0x3e53, 0x685: 0x3e63, + 0x688: 0x47dd, 0x689: 0x47e3, 0x68a: 0x3d4b, 0x68b: 0x3d5b, + 0x68c: 0x3d53, 0x68d: 0x3d63, 0x690: 0x4867, 0x691: 0x486d, + 0x692: 0x3e83, 0x693: 0x3e9b, 0x694: 0x3e8b, 0x695: 0x3ea3, 0x696: 0x3e93, 0x697: 0x3eab, + 0x699: 0x47e9, 0x69b: 0x3d6b, 0x69d: 0x3d73, + 0x69f: 0x3d7b, 0x6a0: 0x487f, 0x6a1: 0x4885, 0x6a2: 0x4981, 0x6a3: 0x4999, + 0x6a4: 0x4989, 0x6a5: 0x49a1, 0x6a6: 0x4991, 0x6a7: 0x49a9, 0x6a8: 0x47ef, 0x6a9: 0x47f5, + 0x6aa: 0x48f1, 0x6ab: 0x4909, 0x6ac: 0x48f9, 0x6ad: 0x4911, 0x6ae: 0x4901, 0x6af: 0x4919, + 0x6b0: 0x47fb, 0x6b1: 0x4321, 0x6b2: 0x3694, 0x6b3: 0x4327, 0x6b4: 0x4825, 0x6b5: 0x432d, + 0x6b6: 0x36a6, 0x6b7: 0x4333, 0x6b8: 0x36c4, 0x6b9: 0x4339, 0x6ba: 0x36dc, 0x6bb: 0x433f, + 0x6bc: 0x4873, 0x6bd: 0x4345, + // Block 0x1b, offset 0x6c0 + 0x6c0: 0x3da3, 0x6c1: 0x3dab, 0x6c2: 0x4187, 0x6c3: 0x41a5, 0x6c4: 0x4191, 0x6c5: 0x41af, + 0x6c6: 0x419b, 0x6c7: 0x41b9, 0x6c8: 0x3cdb, 0x6c9: 0x3ce3, 0x6ca: 0x40d3, 0x6cb: 0x40f1, + 0x6cc: 0x40dd, 0x6cd: 0x40fb, 0x6ce: 0x40e7, 0x6cf: 0x4105, 0x6d0: 0x3deb, 0x6d1: 0x3df3, + 0x6d2: 0x41c3, 0x6d3: 0x41e1, 0x6d4: 0x41cd, 0x6d5: 0x41eb, 0x6d6: 0x41d7, 0x6d7: 0x41f5, + 0x6d8: 0x3d0b, 0x6d9: 0x3d13, 0x6da: 0x410f, 0x6db: 0x412d, 0x6dc: 0x4119, 0x6dd: 0x4137, + 0x6de: 0x4123, 0x6df: 0x4141, 0x6e0: 0x3ec3, 0x6e1: 0x3ecb, 0x6e2: 0x41ff, 0x6e3: 0x421d, + 0x6e4: 0x4209, 0x6e5: 0x4227, 0x6e6: 0x4213, 0x6e7: 0x4231, 0x6e8: 0x3d83, 0x6e9: 0x3d8b, + 0x6ea: 0x414b, 0x6eb: 0x4169, 0x6ec: 0x4155, 0x6ed: 0x4173, 0x6ee: 0x415f, 0x6ef: 0x417d, + 0x6f0: 0x3688, 0x6f1: 0x3682, 0x6f2: 0x3d93, 0x6f3: 0x368e, 0x6f4: 0x3d9b, + 0x6f6: 0x4813, 0x6f7: 0x3db3, 0x6f8: 0x35f8, 0x6f9: 0x35f2, 0x6fa: 0x35e6, 0x6fb: 0x42f1, + 0x6fc: 0x35fe, 0x6fd: 0x428a, 0x6fe: 0x01d3, 0x6ff: 0x428a, + // Block 0x1c, offset 0x700 + 0x700: 0x42a3, 0x701: 0x4485, 0x702: 0x3ddb, 0x703: 0x36a0, 0x704: 0x3de3, + 0x706: 0x483d, 0x707: 0x3dfb, 0x708: 0x3604, 0x709: 0x42f7, 0x70a: 0x3610, 0x70b: 0x42fd, + 0x70c: 0x361c, 0x70d: 0x448c, 0x70e: 0x4493, 0x70f: 0x449a, 0x710: 0x36b8, 0x711: 0x36b2, + 0x712: 0x3e03, 0x713: 0x44e7, 0x716: 0x36be, 0x717: 0x3e13, + 0x718: 0x3634, 0x719: 0x362e, 0x71a: 0x3622, 0x71b: 0x4303, 0x71d: 0x44a1, + 0x71e: 0x44a8, 0x71f: 0x44af, 0x720: 0x36ee, 0x721: 0x36e8, 0x722: 0x3e6b, 0x723: 0x44ef, + 0x724: 0x36d0, 0x725: 0x36d6, 0x726: 0x36f4, 0x727: 0x3e7b, 0x728: 0x3664, 0x729: 0x365e, + 0x72a: 0x3652, 0x72b: 0x430f, 0x72c: 0x364c, 0x72d: 0x4477, 0x72e: 0x447e, 0x72f: 0x0081, + 0x732: 0x3eb3, 0x733: 0x36fa, 0x734: 0x3ebb, + 0x736: 0x488b, 0x737: 0x3ed3, 0x738: 0x3640, 0x739: 0x4309, 0x73a: 0x3670, 0x73b: 0x431b, + 0x73c: 0x367c, 0x73d: 0x425d, 0x73e: 0x428f, + // Block 0x1d, offset 0x740 + 0x740: 0x1bdb, 0x741: 0x1bdf, 0x742: 0x0047, 0x743: 0x1c57, 0x745: 0x1beb, + 0x746: 0x1bef, 0x747: 0x00e9, 0x749: 0x1c5b, 0x74a: 0x008f, 0x74b: 0x0051, + 0x74c: 0x0051, 0x74d: 0x0051, 0x74e: 0x0091, 0x74f: 0x00da, 0x750: 0x0053, 0x751: 0x0053, + 0x752: 0x0059, 0x753: 0x0099, 0x755: 0x005d, 0x756: 0x1990, + 0x759: 0x0061, 0x75a: 0x0063, 0x75b: 0x0065, 0x75c: 0x0065, 0x75d: 0x0065, + 0x760: 0x19a2, 0x761: 0x1bcb, 0x762: 0x19ab, + 0x764: 0x0075, 0x766: 0x01b8, 0x768: 0x0075, + 0x76a: 0x0057, 0x76b: 0x42d5, 0x76c: 0x0045, 0x76d: 0x0047, 0x76f: 0x008b, + 0x770: 0x004b, 0x771: 0x004d, 0x773: 0x005b, 0x774: 0x009f, 0x775: 0x0215, + 0x776: 0x0218, 0x777: 0x021b, 0x778: 0x021e, 0x779: 0x0093, 0x77b: 0x1b9b, + 0x77c: 0x01e8, 0x77d: 0x01c1, 0x77e: 0x0179, 0x77f: 0x01a0, + // Block 0x1e, offset 0x780 + 0x780: 0x0463, 0x785: 0x0049, + 0x786: 0x0089, 0x787: 0x008b, 0x788: 0x0093, 0x789: 0x0095, + 0x790: 0x2231, 0x791: 0x223d, + 0x792: 0x22f1, 0x793: 0x2219, 0x794: 0x229d, 0x795: 0x2225, 0x796: 0x22a3, 0x797: 0x22bb, + 0x798: 0x22c7, 0x799: 0x222b, 0x79a: 0x22cd, 0x79b: 0x2237, 0x79c: 0x22c1, 0x79d: 0x22d3, + 0x79e: 0x22d9, 0x79f: 0x1cbf, 0x7a0: 0x0053, 0x7a1: 0x195a, 0x7a2: 0x1ba7, 0x7a3: 0x1963, + 0x7a4: 0x006d, 0x7a5: 0x19ae, 0x7a6: 0x1bd3, 0x7a7: 0x1d4b, 0x7a8: 0x1966, 0x7a9: 0x0071, + 0x7aa: 0x19ba, 0x7ab: 0x1bd7, 0x7ac: 0x0059, 0x7ad: 0x0047, 0x7ae: 0x0049, 0x7af: 0x005b, + 0x7b0: 0x0093, 0x7b1: 0x19e7, 0x7b2: 0x1c1b, 0x7b3: 0x19f0, 0x7b4: 0x00ad, 0x7b5: 0x1a65, + 0x7b6: 0x1c4f, 0x7b7: 0x1d5f, 0x7b8: 0x19f3, 0x7b9: 0x00b1, 0x7ba: 0x1a68, 0x7bb: 0x1c53, + 0x7bc: 0x0099, 0x7bd: 0x0087, 0x7be: 0x0089, 0x7bf: 0x009b, + // Block 0x1f, offset 0x7c0 + 0x7c1: 0x3c09, 0x7c3: 0xa000, 0x7c4: 0x3c10, 0x7c5: 0xa000, + 0x7c7: 0x3c17, 0x7c8: 0xa000, 0x7c9: 0x3c1e, + 0x7cd: 0xa000, + 0x7e0: 0x2f68, 0x7e1: 0xa000, 0x7e2: 0x3c2c, + 0x7e4: 0xa000, 0x7e5: 0xa000, + 0x7ed: 0x3c25, 0x7ee: 0x2f63, 0x7ef: 0x2f6d, + 0x7f0: 0x3c33, 0x7f1: 0x3c3a, 0x7f2: 0xa000, 0x7f3: 0xa000, 0x7f4: 0x3c41, 0x7f5: 0x3c48, + 0x7f6: 0xa000, 0x7f7: 0xa000, 0x7f8: 0x3c4f, 0x7f9: 0x3c56, 0x7fa: 0xa000, 0x7fb: 0xa000, + 0x7fc: 0xa000, 0x7fd: 0xa000, + // Block 0x20, offset 0x800 + 0x800: 0x3c5d, 0x801: 0x3c64, 0x802: 0xa000, 0x803: 0xa000, 0x804: 0x3c79, 0x805: 0x3c80, + 0x806: 0xa000, 0x807: 0xa000, 0x808: 0x3c87, 0x809: 0x3c8e, + 0x811: 0xa000, + 0x812: 0xa000, + 0x822: 0xa000, + 0x828: 0xa000, 0x829: 0xa000, + 0x82b: 0xa000, 0x82c: 0x3ca3, 0x82d: 0x3caa, 0x82e: 0x3cb1, 0x82f: 0x3cb8, + 0x832: 0xa000, 0x833: 0xa000, 0x834: 0xa000, 0x835: 0xa000, + // Block 0x21, offset 0x840 + 0x860: 0x0023, 0x861: 0x0025, 0x862: 0x0027, 0x863: 0x0029, + 0x864: 0x002b, 0x865: 0x002d, 0x866: 0x002f, 0x867: 0x0031, 0x868: 0x0033, 0x869: 0x1882, + 0x86a: 0x1885, 0x86b: 0x1888, 0x86c: 0x188b, 0x86d: 0x188e, 0x86e: 0x1891, 0x86f: 0x1894, + 0x870: 0x1897, 0x871: 0x189a, 0x872: 0x189d, 0x873: 0x18a6, 0x874: 0x1a6b, 0x875: 0x1a6f, + 0x876: 0x1a73, 0x877: 0x1a77, 0x878: 0x1a7b, 0x879: 0x1a7f, 0x87a: 0x1a83, 0x87b: 0x1a87, + 0x87c: 0x1a8b, 0x87d: 0x1c83, 0x87e: 0x1c88, 0x87f: 0x1c8d, + // Block 0x22, offset 0x880 + 0x880: 0x1c92, 0x881: 0x1c97, 0x882: 0x1c9c, 0x883: 0x1ca1, 0x884: 0x1ca6, 0x885: 0x1cab, + 0x886: 0x1cb0, 0x887: 0x1cb5, 0x888: 0x187f, 0x889: 0x18a3, 0x88a: 0x18c7, 0x88b: 0x18eb, + 0x88c: 0x190f, 0x88d: 0x1918, 0x88e: 0x191e, 0x88f: 0x1924, 0x890: 0x192a, 0x891: 0x1b63, + 0x892: 0x1b67, 0x893: 0x1b6b, 0x894: 0x1b6f, 0x895: 0x1b73, 0x896: 0x1b77, 0x897: 0x1b7b, + 0x898: 0x1b7f, 0x899: 0x1b83, 0x89a: 0x1b87, 0x89b: 0x1b8b, 0x89c: 0x1af7, 0x89d: 0x1afb, + 0x89e: 0x1aff, 0x89f: 0x1b03, 0x8a0: 0x1b07, 0x8a1: 0x1b0b, 0x8a2: 0x1b0f, 0x8a3: 0x1b13, + 0x8a4: 0x1b17, 0x8a5: 0x1b1b, 0x8a6: 0x1b1f, 0x8a7: 0x1b23, 0x8a8: 0x1b27, 0x8a9: 0x1b2b, + 0x8aa: 0x1b2f, 0x8ab: 0x1b33, 0x8ac: 0x1b37, 0x8ad: 0x1b3b, 0x8ae: 0x1b3f, 0x8af: 0x1b43, + 0x8b0: 0x1b47, 0x8b1: 0x1b4b, 0x8b2: 0x1b4f, 0x8b3: 0x1b53, 0x8b4: 0x1b57, 0x8b5: 0x1b5b, + 0x8b6: 0x0043, 0x8b7: 0x0045, 0x8b8: 0x0047, 0x8b9: 0x0049, 0x8ba: 0x004b, 0x8bb: 0x004d, + 0x8bc: 0x004f, 0x8bd: 0x0051, 0x8be: 0x0053, 0x8bf: 0x0055, + // Block 0x23, offset 0x8c0 + 0x8c0: 0x06bf, 0x8c1: 0x06e3, 0x8c2: 0x06ef, 0x8c3: 0x06ff, 0x8c4: 0x0707, 0x8c5: 0x0713, + 0x8c6: 0x071b, 0x8c7: 0x0723, 0x8c8: 0x072f, 0x8c9: 0x0783, 0x8ca: 0x079b, 0x8cb: 0x07ab, + 0x8cc: 0x07bb, 0x8cd: 0x07cb, 0x8ce: 0x07db, 0x8cf: 0x07fb, 0x8d0: 0x07ff, 0x8d1: 0x0803, + 0x8d2: 0x0837, 0x8d3: 0x085f, 0x8d4: 0x086f, 0x8d5: 0x0877, 0x8d6: 0x087b, 0x8d7: 0x0887, + 0x8d8: 0x08a3, 0x8d9: 0x08a7, 0x8da: 0x08bf, 0x8db: 0x08c3, 0x8dc: 0x08cb, 0x8dd: 0x08db, + 0x8de: 0x0977, 0x8df: 0x098b, 0x8e0: 0x09cb, 0x8e1: 0x09df, 0x8e2: 0x09e7, 0x8e3: 0x09eb, + 0x8e4: 0x09fb, 0x8e5: 0x0a17, 0x8e6: 0x0a43, 0x8e7: 0x0a4f, 0x8e8: 0x0a6f, 0x8e9: 0x0a7b, + 0x8ea: 0x0a7f, 0x8eb: 0x0a83, 0x8ec: 0x0a9b, 0x8ed: 0x0a9f, 0x8ee: 0x0acb, 0x8ef: 0x0ad7, + 0x8f0: 0x0adf, 0x8f1: 0x0ae7, 0x8f2: 0x0af7, 0x8f3: 0x0aff, 0x8f4: 0x0b07, 0x8f5: 0x0b33, + 0x8f6: 0x0b37, 0x8f7: 0x0b3f, 0x8f8: 0x0b43, 0x8f9: 0x0b4b, 0x8fa: 0x0b53, 0x8fb: 0x0b63, + 0x8fc: 0x0b7f, 0x8fd: 0x0bf7, 0x8fe: 0x0c0b, 0x8ff: 0x0c0f, + // Block 0x24, offset 0x900 + 0x900: 0x0c8f, 0x901: 0x0c93, 0x902: 0x0ca7, 0x903: 0x0cab, 0x904: 0x0cb3, 0x905: 0x0cbb, + 0x906: 0x0cc3, 0x907: 0x0ccf, 0x908: 0x0cf7, 0x909: 0x0d07, 0x90a: 0x0d1b, 0x90b: 0x0d8b, + 0x90c: 0x0d97, 0x90d: 0x0da7, 0x90e: 0x0db3, 0x90f: 0x0dbf, 0x910: 0x0dc7, 0x911: 0x0dcb, + 0x912: 0x0dcf, 0x913: 0x0dd3, 0x914: 0x0dd7, 0x915: 0x0e8f, 0x916: 0x0ed7, 0x917: 0x0ee3, + 0x918: 0x0ee7, 0x919: 0x0eeb, 0x91a: 0x0eef, 0x91b: 0x0ef7, 0x91c: 0x0efb, 0x91d: 0x0f0f, + 0x91e: 0x0f2b, 0x91f: 0x0f33, 0x920: 0x0f73, 0x921: 0x0f77, 0x922: 0x0f7f, 0x923: 0x0f83, + 0x924: 0x0f8b, 0x925: 0x0f8f, 0x926: 0x0fb3, 0x927: 0x0fb7, 0x928: 0x0fd3, 0x929: 0x0fd7, + 0x92a: 0x0fdb, 0x92b: 0x0fdf, 0x92c: 0x0ff3, 0x92d: 0x1017, 0x92e: 0x101b, 0x92f: 0x101f, + 0x930: 0x1043, 0x931: 0x1083, 0x932: 0x1087, 0x933: 0x10a7, 0x934: 0x10b7, 0x935: 0x10bf, + 0x936: 0x10df, 0x937: 0x1103, 0x938: 0x1147, 0x939: 0x114f, 0x93a: 0x1163, 0x93b: 0x116f, + 0x93c: 0x1177, 0x93d: 0x117f, 0x93e: 0x1183, 0x93f: 0x1187, + // Block 0x25, offset 0x940 + 0x940: 0x119f, 0x941: 0x11a3, 0x942: 0x11bf, 0x943: 0x11c7, 0x944: 0x11cf, 0x945: 0x11d3, + 0x946: 0x11df, 0x947: 0x11e7, 0x948: 0x11eb, 0x949: 0x11ef, 0x94a: 0x11f7, 0x94b: 0x11fb, + 0x94c: 0x129b, 0x94d: 0x12af, 0x94e: 0x12e3, 0x94f: 0x12e7, 0x950: 0x12ef, 0x951: 0x131b, + 0x952: 0x1323, 0x953: 0x132b, 0x954: 0x1333, 0x955: 0x136f, 0x956: 0x1373, 0x957: 0x137b, + 0x958: 0x137f, 0x959: 0x1383, 0x95a: 0x13af, 0x95b: 0x13b3, 0x95c: 0x13bb, 0x95d: 0x13cf, + 0x95e: 0x13d3, 0x95f: 0x13ef, 0x960: 0x13f7, 0x961: 0x13fb, 0x962: 0x141f, 0x963: 0x143f, + 0x964: 0x1453, 0x965: 0x1457, 0x966: 0x145f, 0x967: 0x148b, 0x968: 0x148f, 0x969: 0x149f, + 0x96a: 0x14c3, 0x96b: 0x14cf, 0x96c: 0x14df, 0x96d: 0x14f7, 0x96e: 0x14ff, 0x96f: 0x1503, + 0x970: 0x1507, 0x971: 0x150b, 0x972: 0x1517, 0x973: 0x151b, 0x974: 0x1523, 0x975: 0x153f, + 0x976: 0x1543, 0x977: 0x1547, 0x978: 0x155f, 0x979: 0x1563, 0x97a: 0x156b, 0x97b: 0x157f, + 0x97c: 0x1583, 0x97d: 0x1587, 0x97e: 0x158f, 0x97f: 0x1593, + // Block 0x26, offset 0x980 + 0x986: 0xa000, 0x98b: 0xa000, + 0x98c: 0x3f0b, 0x98d: 0xa000, 0x98e: 0x3f13, 0x98f: 0xa000, 0x990: 0x3f1b, 0x991: 0xa000, + 0x992: 0x3f23, 0x993: 0xa000, 0x994: 0x3f2b, 0x995: 0xa000, 0x996: 0x3f33, 0x997: 0xa000, + 0x998: 0x3f3b, 0x999: 0xa000, 0x99a: 0x3f43, 0x99b: 0xa000, 0x99c: 0x3f4b, 0x99d: 0xa000, + 0x99e: 0x3f53, 0x99f: 0xa000, 0x9a0: 0x3f5b, 0x9a1: 0xa000, 0x9a2: 0x3f63, + 0x9a4: 0xa000, 0x9a5: 0x3f6b, 0x9a6: 0xa000, 0x9a7: 0x3f73, 0x9a8: 0xa000, 0x9a9: 0x3f7b, + 0x9af: 0xa000, + 0x9b0: 0x3f83, 0x9b1: 0x3f8b, 0x9b2: 0xa000, 0x9b3: 0x3f93, 0x9b4: 0x3f9b, 0x9b5: 0xa000, + 0x9b6: 0x3fa3, 0x9b7: 0x3fab, 0x9b8: 0xa000, 0x9b9: 0x3fb3, 0x9ba: 0x3fbb, 0x9bb: 0xa000, + 0x9bc: 0x3fc3, 0x9bd: 0x3fcb, + // Block 0x27, offset 0x9c0 + 0x9d4: 0x3f03, + 0x9d9: 0x9903, 0x9da: 0x9903, 0x9db: 0x42df, 0x9dc: 0x42e5, 0x9dd: 0xa000, + 0x9de: 0x3fd3, 0x9df: 0x26b7, + 0x9e6: 0xa000, + 0x9eb: 0xa000, 0x9ec: 0x3fe3, 0x9ed: 0xa000, 0x9ee: 0x3feb, 0x9ef: 0xa000, + 0x9f0: 0x3ff3, 0x9f1: 0xa000, 0x9f2: 0x3ffb, 0x9f3: 0xa000, 0x9f4: 0x4003, 0x9f5: 0xa000, + 0x9f6: 0x400b, 0x9f7: 0xa000, 0x9f8: 0x4013, 0x9f9: 0xa000, 0x9fa: 0x401b, 0x9fb: 0xa000, + 0x9fc: 0x4023, 0x9fd: 0xa000, 0x9fe: 0x402b, 0x9ff: 0xa000, + // Block 0x28, offset 0xa00 + 0xa00: 0x4033, 0xa01: 0xa000, 0xa02: 0x403b, 0xa04: 0xa000, 0xa05: 0x4043, + 0xa06: 0xa000, 0xa07: 0x404b, 0xa08: 0xa000, 0xa09: 0x4053, + 0xa0f: 0xa000, 0xa10: 0x405b, 0xa11: 0x4063, + 0xa12: 0xa000, 0xa13: 0x406b, 0xa14: 0x4073, 0xa15: 0xa000, 0xa16: 0x407b, 0xa17: 0x4083, + 0xa18: 0xa000, 0xa19: 0x408b, 0xa1a: 0x4093, 0xa1b: 0xa000, 0xa1c: 0x409b, 0xa1d: 0x40a3, + 0xa2f: 0xa000, + 0xa30: 0xa000, 0xa31: 0xa000, 0xa32: 0xa000, 0xa34: 0x3fdb, + 0xa37: 0x40ab, 0xa38: 0x40b3, 0xa39: 0x40bb, 0xa3a: 0x40c3, + 0xa3d: 0xa000, 0xa3e: 0x40cb, 0xa3f: 0x26cc, + // Block 0x29, offset 0xa40 + 0xa40: 0x0367, 0xa41: 0x032b, 0xa42: 0x032f, 0xa43: 0x0333, 0xa44: 0x037b, 0xa45: 0x0337, + 0xa46: 0x033b, 0xa47: 0x033f, 0xa48: 0x0343, 0xa49: 0x0347, 0xa4a: 0x034b, 0xa4b: 0x034f, + 0xa4c: 0x0353, 0xa4d: 0x0357, 0xa4e: 0x035b, 0xa4f: 0x49c0, 0xa50: 0x49c6, 0xa51: 0x49cc, + 0xa52: 0x49d2, 0xa53: 0x49d8, 0xa54: 0x49de, 0xa55: 0x49e4, 0xa56: 0x49ea, 0xa57: 0x49f0, + 0xa58: 0x49f6, 0xa59: 0x49fc, 0xa5a: 0x4a02, 0xa5b: 0x4a08, 0xa5c: 0x4a0e, 0xa5d: 0x4a14, + 0xa5e: 0x4a1a, 0xa5f: 0x4a20, 0xa60: 0x4a26, 0xa61: 0x4a2c, 0xa62: 0x4a32, 0xa63: 0x4a38, + 0xa64: 0x03c3, 0xa65: 0x035f, 0xa66: 0x0363, 0xa67: 0x03e7, 0xa68: 0x03eb, 0xa69: 0x03ef, + 0xa6a: 0x03f3, 0xa6b: 0x03f7, 0xa6c: 0x03fb, 0xa6d: 0x03ff, 0xa6e: 0x036b, 0xa6f: 0x0403, + 0xa70: 0x0407, 0xa71: 0x036f, 0xa72: 0x0373, 0xa73: 0x0377, 0xa74: 0x037f, 0xa75: 0x0383, + 0xa76: 0x0387, 0xa77: 0x038b, 0xa78: 0x038f, 0xa79: 0x0393, 0xa7a: 0x0397, 0xa7b: 0x039b, + 0xa7c: 0x039f, 0xa7d: 0x03a3, 0xa7e: 0x03a7, 0xa7f: 0x03ab, + // Block 0x2a, offset 0xa80 + 0xa80: 0x03af, 0xa81: 0x03b3, 0xa82: 0x040b, 0xa83: 0x040f, 0xa84: 0x03b7, 0xa85: 0x03bb, + 0xa86: 0x03bf, 0xa87: 0x03c7, 0xa88: 0x03cb, 0xa89: 0x03cf, 0xa8a: 0x03d3, 0xa8b: 0x03d7, + 0xa8c: 0x03db, 0xa8d: 0x03df, 0xa8e: 0x03e3, + 0xa92: 0x06bf, 0xa93: 0x071b, 0xa94: 0x06cb, 0xa95: 0x097b, 0xa96: 0x06cf, 0xa97: 0x06e7, + 0xa98: 0x06d3, 0xa99: 0x0f93, 0xa9a: 0x0707, 0xa9b: 0x06db, 0xa9c: 0x06c3, 0xa9d: 0x09ff, + 0xa9e: 0x098f, 0xa9f: 0x072f, + // Block 0x2b, offset 0xac0 + 0xac0: 0x2057, 0xac1: 0x205d, 0xac2: 0x2063, 0xac3: 0x2069, 0xac4: 0x206f, 0xac5: 0x2075, + 0xac6: 0x207b, 0xac7: 0x2081, 0xac8: 0x2087, 0xac9: 0x208d, 0xaca: 0x2093, 0xacb: 0x2099, + 0xacc: 0x209f, 0xacd: 0x20a5, 0xace: 0x2729, 0xacf: 0x2732, 0xad0: 0x273b, 0xad1: 0x2744, + 0xad2: 0x274d, 0xad3: 0x2756, 0xad4: 0x275f, 0xad5: 0x2768, 0xad6: 0x2771, 0xad7: 0x2783, + 0xad8: 0x278c, 0xad9: 0x2795, 0xada: 0x279e, 0xadb: 0x27a7, 0xadc: 0x277a, 0xadd: 0x2baf, + 0xade: 0x2af0, 0xae0: 0x20ab, 0xae1: 0x20c3, 0xae2: 0x20b7, 0xae3: 0x210b, + 0xae4: 0x20c9, 0xae5: 0x20e7, 0xae6: 0x20b1, 0xae7: 0x20e1, 0xae8: 0x20bd, 0xae9: 0x20f3, + 0xaea: 0x2123, 0xaeb: 0x2141, 0xaec: 0x213b, 0xaed: 0x212f, 0xaee: 0x217d, 0xaef: 0x2111, + 0xaf0: 0x211d, 0xaf1: 0x2135, 0xaf2: 0x2129, 0xaf3: 0x2153, 0xaf4: 0x20ff, 0xaf5: 0x2147, + 0xaf6: 0x2171, 0xaf7: 0x2159, 0xaf8: 0x20ed, 0xaf9: 0x20cf, 0xafa: 0x2105, 0xafb: 0x2117, + 0xafc: 0x214d, 0xafd: 0x20d5, 0xafe: 0x2177, 0xaff: 0x20f9, + // Block 0x2c, offset 0xb00 + 0xb00: 0x215f, 0xb01: 0x20db, 0xb02: 0x2165, 0xb03: 0x216b, 0xb04: 0x092f, 0xb05: 0x0b03, + 0xb06: 0x0ca7, 0xb07: 0x10c7, + 0xb10: 0x1bc7, 0xb11: 0x18a9, + 0xb12: 0x18ac, 0xb13: 0x18af, 0xb14: 0x18b2, 0xb15: 0x18b5, 0xb16: 0x18b8, 0xb17: 0x18bb, + 0xb18: 0x18be, 0xb19: 0x18c1, 0xb1a: 0x18ca, 0xb1b: 0x18cd, 0xb1c: 0x18d0, 0xb1d: 0x18d3, + 0xb1e: 0x18d6, 0xb1f: 0x18d9, 0xb20: 0x0313, 0xb21: 0x031b, 0xb22: 0x031f, 0xb23: 0x0327, + 0xb24: 0x032b, 0xb25: 0x032f, 0xb26: 0x0337, 0xb27: 0x033f, 0xb28: 0x0343, 0xb29: 0x034b, + 0xb2a: 0x034f, 0xb2b: 0x0353, 0xb2c: 0x0357, 0xb2d: 0x035b, 0xb2e: 0x2e1b, 0xb2f: 0x2e23, + 0xb30: 0x2e2b, 0xb31: 0x2e33, 0xb32: 0x2e3b, 0xb33: 0x2e43, 0xb34: 0x2e4b, 0xb35: 0x2e53, + 0xb36: 0x2e63, 0xb37: 0x2e6b, 0xb38: 0x2e73, 0xb39: 0x2e7b, 0xb3a: 0x2e83, 0xb3b: 0x2e8b, + 0xb3c: 0x2ed6, 0xb3d: 0x2e9e, 0xb3e: 0x2e5b, + // Block 0x2d, offset 0xb40 + 0xb40: 0x06bf, 0xb41: 0x071b, 0xb42: 0x06cb, 0xb43: 0x097b, 0xb44: 0x071f, 0xb45: 0x07af, + 0xb46: 0x06c7, 0xb47: 0x07ab, 0xb48: 0x070b, 0xb49: 0x0887, 0xb4a: 0x0d07, 0xb4b: 0x0e8f, + 0xb4c: 0x0dd7, 0xb4d: 0x0d1b, 0xb4e: 0x145f, 0xb4f: 0x098b, 0xb50: 0x0ccf, 0xb51: 0x0d4b, + 0xb52: 0x0d0b, 0xb53: 0x104b, 0xb54: 0x08fb, 0xb55: 0x0f03, 0xb56: 0x1387, 0xb57: 0x105f, + 0xb58: 0x0843, 0xb59: 0x108f, 0xb5a: 0x0f9b, 0xb5b: 0x0a17, 0xb5c: 0x140f, 0xb5d: 0x077f, + 0xb5e: 0x08ab, 0xb5f: 0x0df7, 0xb60: 0x1527, 0xb61: 0x0743, 0xb62: 0x07d3, 0xb63: 0x0d9b, + 0xb64: 0x06cf, 0xb65: 0x06e7, 0xb66: 0x06d3, 0xb67: 0x0adb, 0xb68: 0x08ef, 0xb69: 0x087f, + 0xb6a: 0x0a57, 0xb6b: 0x0a4b, 0xb6c: 0x0feb, 0xb6d: 0x073f, 0xb6e: 0x139b, 0xb6f: 0x089b, + 0xb70: 0x09f3, 0xb71: 0x18dc, 0xb72: 0x18df, 0xb73: 0x18e2, 0xb74: 0x18e5, 0xb75: 0x18ee, + 0xb76: 0x18f1, 0xb77: 0x18f4, 0xb78: 0x18f7, 0xb79: 0x18fa, 0xb7a: 0x18fd, 0xb7b: 0x1900, + 0xb7c: 0x1903, 0xb7d: 0x1906, 0xb7e: 0x1909, 0xb7f: 0x1912, + // Block 0x2e, offset 0xb80 + 0xb80: 0x1cc9, 0xb81: 0x1cd8, 0xb82: 0x1ce7, 0xb83: 0x1cf6, 0xb84: 0x1d05, 0xb85: 0x1d14, + 0xb86: 0x1d23, 0xb87: 0x1d32, 0xb88: 0x1d41, 0xb89: 0x218f, 0xb8a: 0x21a1, 0xb8b: 0x21b3, + 0xb8c: 0x1954, 0xb8d: 0x1c07, 0xb8e: 0x19d5, 0xb8f: 0x1bab, 0xb90: 0x04cb, 0xb91: 0x04d3, + 0xb92: 0x04db, 0xb93: 0x04e3, 0xb94: 0x04eb, 0xb95: 0x04ef, 0xb96: 0x04f3, 0xb97: 0x04f7, + 0xb98: 0x04fb, 0xb99: 0x04ff, 0xb9a: 0x0503, 0xb9b: 0x0507, 0xb9c: 0x050b, 0xb9d: 0x050f, + 0xb9e: 0x0513, 0xb9f: 0x0517, 0xba0: 0x051b, 0xba1: 0x0523, 0xba2: 0x0527, 0xba3: 0x052b, + 0xba4: 0x052f, 0xba5: 0x0533, 0xba6: 0x0537, 0xba7: 0x053b, 0xba8: 0x053f, 0xba9: 0x0543, + 0xbaa: 0x0547, 0xbab: 0x054b, 0xbac: 0x054f, 0xbad: 0x0553, 0xbae: 0x0557, 0xbaf: 0x055b, + 0xbb0: 0x055f, 0xbb1: 0x0563, 0xbb2: 0x0567, 0xbb3: 0x056f, 0xbb4: 0x0577, 0xbb5: 0x057f, + 0xbb6: 0x0583, 0xbb7: 0x0587, 0xbb8: 0x058b, 0xbb9: 0x058f, 0xbba: 0x0593, 0xbbb: 0x0597, + 0xbbc: 0x059b, 0xbbd: 0x059f, 0xbbe: 0x05a3, + // Block 0x2f, offset 0xbc0 + 0xbc0: 0x2b0f, 0xbc1: 0x29ab, 0xbc2: 0x2b1f, 0xbc3: 0x2883, 0xbc4: 0x2ee7, 0xbc5: 0x288d, + 0xbc6: 0x2897, 0xbc7: 0x2f2b, 0xbc8: 0x29b8, 0xbc9: 0x28a1, 0xbca: 0x28ab, 0xbcb: 0x28b5, + 0xbcc: 0x29df, 0xbcd: 0x29ec, 0xbce: 0x29c5, 0xbcf: 0x29d2, 0xbd0: 0x2eac, 0xbd1: 0x29f9, + 0xbd2: 0x2a06, 0xbd3: 0x2bc1, 0xbd4: 0x26be, 0xbd5: 0x2bd4, 0xbd6: 0x2be7, 0xbd7: 0x2b2f, + 0xbd8: 0x2a13, 0xbd9: 0x2bfa, 0xbda: 0x2c0d, 0xbdb: 0x2a20, 0xbdc: 0x28bf, 0xbdd: 0x28c9, + 0xbde: 0x2eba, 0xbdf: 0x2a2d, 0xbe0: 0x2b3f, 0xbe1: 0x2ef8, 0xbe2: 0x28d3, 0xbe3: 0x28dd, + 0xbe4: 0x2a3a, 0xbe5: 0x28e7, 0xbe6: 0x28f1, 0xbe7: 0x26d3, 0xbe8: 0x26da, 0xbe9: 0x28fb, + 0xbea: 0x2905, 0xbeb: 0x2c20, 0xbec: 0x2a47, 0xbed: 0x2b4f, 0xbee: 0x2c33, 0xbef: 0x2a54, + 0xbf0: 0x2919, 0xbf1: 0x290f, 0xbf2: 0x2f3f, 0xbf3: 0x2a61, 0xbf4: 0x2c46, 0xbf5: 0x2923, + 0xbf6: 0x2b5f, 0xbf7: 0x292d, 0xbf8: 0x2a7b, 0xbf9: 0x2937, 0xbfa: 0x2a88, 0xbfb: 0x2f09, + 0xbfc: 0x2a6e, 0xbfd: 0x2b6f, 0xbfe: 0x2a95, 0xbff: 0x26e1, + // Block 0x30, offset 0xc00 + 0xc00: 0x2f1a, 0xc01: 0x2941, 0xc02: 0x294b, 0xc03: 0x2aa2, 0xc04: 0x2955, 0xc05: 0x295f, + 0xc06: 0x2969, 0xc07: 0x2b7f, 0xc08: 0x2aaf, 0xc09: 0x26e8, 0xc0a: 0x2c59, 0xc0b: 0x2e93, + 0xc0c: 0x2b8f, 0xc0d: 0x2abc, 0xc0e: 0x2ec8, 0xc0f: 0x2973, 0xc10: 0x297d, 0xc11: 0x2ac9, + 0xc12: 0x26ef, 0xc13: 0x2ad6, 0xc14: 0x2b9f, 0xc15: 0x26f6, 0xc16: 0x2c6c, 0xc17: 0x2987, + 0xc18: 0x1cba, 0xc19: 0x1cce, 0xc1a: 0x1cdd, 0xc1b: 0x1cec, 0xc1c: 0x1cfb, 0xc1d: 0x1d0a, + 0xc1e: 0x1d19, 0xc1f: 0x1d28, 0xc20: 0x1d37, 0xc21: 0x1d46, 0xc22: 0x2195, 0xc23: 0x21a7, + 0xc24: 0x21b9, 0xc25: 0x21c5, 0xc26: 0x21d1, 0xc27: 0x21dd, 0xc28: 0x21e9, 0xc29: 0x21f5, + 0xc2a: 0x2201, 0xc2b: 0x220d, 0xc2c: 0x2249, 0xc2d: 0x2255, 0xc2e: 0x2261, 0xc2f: 0x226d, + 0xc30: 0x2279, 0xc31: 0x1c17, 0xc32: 0x19c9, 0xc33: 0x1936, 0xc34: 0x1be7, 0xc35: 0x1a4a, + 0xc36: 0x1a59, 0xc37: 0x19cf, 0xc38: 0x1bff, 0xc39: 0x1c03, 0xc3a: 0x1960, 0xc3b: 0x2704, + 0xc3c: 0x2712, 0xc3d: 0x26fd, 0xc3e: 0x270b, 0xc3f: 0x2ae3, + // Block 0x31, offset 0xc40 + 0xc40: 0x1a4d, 0xc41: 0x1a35, 0xc42: 0x1c63, 0xc43: 0x1a1d, 0xc44: 0x19f6, 0xc45: 0x1969, + 0xc46: 0x1978, 0xc47: 0x1948, 0xc48: 0x1bf3, 0xc49: 0x1d55, 0xc4a: 0x1a50, 0xc4b: 0x1a38, + 0xc4c: 0x1c67, 0xc4d: 0x1c73, 0xc4e: 0x1a29, 0xc4f: 0x19ff, 0xc50: 0x1957, 0xc51: 0x1c1f, + 0xc52: 0x1bb3, 0xc53: 0x1b9f, 0xc54: 0x1bcf, 0xc55: 0x1c77, 0xc56: 0x1a2c, 0xc57: 0x19cc, + 0xc58: 0x1a02, 0xc59: 0x19e1, 0xc5a: 0x1a44, 0xc5b: 0x1c7b, 0xc5c: 0x1a2f, 0xc5d: 0x19c3, + 0xc5e: 0x1a05, 0xc5f: 0x1c3f, 0xc60: 0x1bf7, 0xc61: 0x1a17, 0xc62: 0x1c27, 0xc63: 0x1c43, + 0xc64: 0x1bfb, 0xc65: 0x1a1a, 0xc66: 0x1c2b, 0xc67: 0x22eb, 0xc68: 0x22ff, 0xc69: 0x1999, + 0xc6a: 0x1c23, 0xc6b: 0x1bb7, 0xc6c: 0x1ba3, 0xc6d: 0x1c4b, 0xc6e: 0x2719, 0xc6f: 0x27b0, + 0xc70: 0x1a5c, 0xc71: 0x1a47, 0xc72: 0x1c7f, 0xc73: 0x1a32, 0xc74: 0x1a53, 0xc75: 0x1a3b, + 0xc76: 0x1c6b, 0xc77: 0x1a20, 0xc78: 0x19f9, 0xc79: 0x1984, 0xc7a: 0x1a56, 0xc7b: 0x1a3e, + 0xc7c: 0x1c6f, 0xc7d: 0x1a23, 0xc7e: 0x19fc, 0xc7f: 0x1987, + // Block 0x32, offset 0xc80 + 0xc80: 0x1c2f, 0xc81: 0x1bbb, 0xc82: 0x1d50, 0xc83: 0x1939, 0xc84: 0x19bd, 0xc85: 0x19c0, + 0xc86: 0x22f8, 0xc87: 0x1b97, 0xc88: 0x19c6, 0xc89: 0x194b, 0xc8a: 0x19e4, 0xc8b: 0x194e, + 0xc8c: 0x19ed, 0xc8d: 0x196c, 0xc8e: 0x196f, 0xc8f: 0x1a08, 0xc90: 0x1a0e, 0xc91: 0x1a11, + 0xc92: 0x1c33, 0xc93: 0x1a14, 0xc94: 0x1a26, 0xc95: 0x1c3b, 0xc96: 0x1c47, 0xc97: 0x1993, + 0xc98: 0x1d5a, 0xc99: 0x1bbf, 0xc9a: 0x1996, 0xc9b: 0x1a5f, 0xc9c: 0x19a8, 0xc9d: 0x19b7, + 0xc9e: 0x22e5, 0xc9f: 0x22df, 0xca0: 0x1cc4, 0xca1: 0x1cd3, 0xca2: 0x1ce2, 0xca3: 0x1cf1, + 0xca4: 0x1d00, 0xca5: 0x1d0f, 0xca6: 0x1d1e, 0xca7: 0x1d2d, 0xca8: 0x1d3c, 0xca9: 0x2189, + 0xcaa: 0x219b, 0xcab: 0x21ad, 0xcac: 0x21bf, 0xcad: 0x21cb, 0xcae: 0x21d7, 0xcaf: 0x21e3, + 0xcb0: 0x21ef, 0xcb1: 0x21fb, 0xcb2: 0x2207, 0xcb3: 0x2243, 0xcb4: 0x224f, 0xcb5: 0x225b, + 0xcb6: 0x2267, 0xcb7: 0x2273, 0xcb8: 0x227f, 0xcb9: 0x2285, 0xcba: 0x228b, 0xcbb: 0x2291, + 0xcbc: 0x2297, 0xcbd: 0x22a9, 0xcbe: 0x22af, 0xcbf: 0x1c13, + // Block 0x33, offset 0xcc0 + 0xcc0: 0x1377, 0xcc1: 0x0cfb, 0xcc2: 0x13d3, 0xcc3: 0x139f, 0xcc4: 0x0e57, 0xcc5: 0x06eb, + 0xcc6: 0x08df, 0xcc7: 0x162b, 0xcc8: 0x162b, 0xcc9: 0x0a0b, 0xcca: 0x145f, 0xccb: 0x0943, + 0xccc: 0x0a07, 0xccd: 0x0bef, 0xcce: 0x0fcf, 0xccf: 0x115f, 0xcd0: 0x1297, 0xcd1: 0x12d3, + 0xcd2: 0x1307, 0xcd3: 0x141b, 0xcd4: 0x0d73, 0xcd5: 0x0dff, 0xcd6: 0x0eab, 0xcd7: 0x0f43, + 0xcd8: 0x125f, 0xcd9: 0x1447, 0xcda: 0x1573, 0xcdb: 0x070f, 0xcdc: 0x08b3, 0xcdd: 0x0d87, + 0xcde: 0x0ecf, 0xcdf: 0x1293, 0xce0: 0x15c3, 0xce1: 0x0ab3, 0xce2: 0x0e77, 0xce3: 0x1283, + 0xce4: 0x1317, 0xce5: 0x0c23, 0xce6: 0x11bb, 0xce7: 0x12df, 0xce8: 0x0b1f, 0xce9: 0x0d0f, + 0xcea: 0x0e17, 0xceb: 0x0f1b, 0xcec: 0x1427, 0xced: 0x074f, 0xcee: 0x07e7, 0xcef: 0x0853, + 0xcf0: 0x0c8b, 0xcf1: 0x0d7f, 0xcf2: 0x0ecb, 0xcf3: 0x0fef, 0xcf4: 0x1177, 0xcf5: 0x128b, + 0xcf6: 0x12a3, 0xcf7: 0x13c7, 0xcf8: 0x14ef, 0xcf9: 0x15a3, 0xcfa: 0x15bf, 0xcfb: 0x102b, + 0xcfc: 0x106b, 0xcfd: 0x1123, 0xcfe: 0x1243, 0xcff: 0x147b, + // Block 0x34, offset 0xd00 + 0xd00: 0x15cb, 0xd01: 0x134b, 0xd02: 0x09c7, 0xd03: 0x0b3b, 0xd04: 0x10db, 0xd05: 0x119b, + 0xd06: 0x0eff, 0xd07: 0x1033, 0xd08: 0x1397, 0xd09: 0x14e7, 0xd0a: 0x09c3, 0xd0b: 0x0a8f, + 0xd0c: 0x0d77, 0xd0d: 0x0e2b, 0xd0e: 0x0e5f, 0xd0f: 0x1113, 0xd10: 0x113b, 0xd11: 0x14a7, + 0xd12: 0x084f, 0xd13: 0x11a7, 0xd14: 0x07f3, 0xd15: 0x07ef, 0xd16: 0x1097, 0xd17: 0x1127, + 0xd18: 0x125b, 0xd19: 0x14af, 0xd1a: 0x1367, 0xd1b: 0x0c27, 0xd1c: 0x0d73, 0xd1d: 0x1357, + 0xd1e: 0x06f7, 0xd1f: 0x0a63, 0xd20: 0x0b93, 0xd21: 0x0f2f, 0xd22: 0x0faf, 0xd23: 0x0873, + 0xd24: 0x103b, 0xd25: 0x075f, 0xd26: 0x0b77, 0xd27: 0x06d7, 0xd28: 0x0deb, 0xd29: 0x0ca3, + 0xd2a: 0x110f, 0xd2b: 0x08c7, 0xd2c: 0x09b3, 0xd2d: 0x0ffb, 0xd2e: 0x1263, 0xd2f: 0x133b, + 0xd30: 0x0db7, 0xd31: 0x13f7, 0xd32: 0x0de3, 0xd33: 0x0c37, 0xd34: 0x121b, 0xd35: 0x0c57, + 0xd36: 0x0fab, 0xd37: 0x072b, 0xd38: 0x07a7, 0xd39: 0x07eb, 0xd3a: 0x0d53, 0xd3b: 0x10fb, + 0xd3c: 0x11f3, 0xd3d: 0x1347, 0xd3e: 0x145b, 0xd3f: 0x085b, + // Block 0x35, offset 0xd40 + 0xd40: 0x090f, 0xd41: 0x0a17, 0xd42: 0x0b2f, 0xd43: 0x0cbf, 0xd44: 0x0e7b, 0xd45: 0x103f, + 0xd46: 0x1497, 0xd47: 0x157b, 0xd48: 0x15cf, 0xd49: 0x15e7, 0xd4a: 0x0837, 0xd4b: 0x0cf3, + 0xd4c: 0x0da3, 0xd4d: 0x13eb, 0xd4e: 0x0afb, 0xd4f: 0x0bd7, 0xd50: 0x0bf3, 0xd51: 0x0c83, + 0xd52: 0x0e6b, 0xd53: 0x0eb7, 0xd54: 0x0f67, 0xd55: 0x108b, 0xd56: 0x112f, 0xd57: 0x1193, + 0xd58: 0x13db, 0xd59: 0x126b, 0xd5a: 0x1403, 0xd5b: 0x147f, 0xd5c: 0x080f, 0xd5d: 0x083b, + 0xd5e: 0x0923, 0xd5f: 0x0ea7, 0xd60: 0x12f3, 0xd61: 0x133b, 0xd62: 0x0b1b, 0xd63: 0x0b8b, + 0xd64: 0x0c4f, 0xd65: 0x0daf, 0xd66: 0x10d7, 0xd67: 0x0f23, 0xd68: 0x073b, 0xd69: 0x097f, + 0xd6a: 0x0a63, 0xd6b: 0x0ac7, 0xd6c: 0x0b97, 0xd6d: 0x0f3f, 0xd6e: 0x0f5b, 0xd6f: 0x116b, + 0xd70: 0x118b, 0xd71: 0x1463, 0xd72: 0x14e3, 0xd73: 0x14f3, 0xd74: 0x152f, 0xd75: 0x0753, + 0xd76: 0x107f, 0xd77: 0x144f, 0xd78: 0x14cb, 0xd79: 0x0baf, 0xd7a: 0x0717, 0xd7b: 0x0777, + 0xd7c: 0x0a67, 0xd7d: 0x0a87, 0xd7e: 0x0caf, 0xd7f: 0x0d73, + // Block 0x36, offset 0xd80 + 0xd80: 0x0ec3, 0xd81: 0x0fcb, 0xd82: 0x1277, 0xd83: 0x1417, 0xd84: 0x1623, 0xd85: 0x0ce3, + 0xd86: 0x14a3, 0xd87: 0x0833, 0xd88: 0x0d2f, 0xd89: 0x0d3b, 0xd8a: 0x0e0f, 0xd8b: 0x0e47, + 0xd8c: 0x0f4b, 0xd8d: 0x0fa7, 0xd8e: 0x1027, 0xd8f: 0x110b, 0xd90: 0x153b, 0xd91: 0x07af, + 0xd92: 0x0c03, 0xd93: 0x14b3, 0xd94: 0x0767, 0xd95: 0x0aab, 0xd96: 0x0e2f, 0xd97: 0x13df, + 0xd98: 0x0b67, 0xd99: 0x0bb7, 0xd9a: 0x0d43, 0xd9b: 0x0f2f, 0xd9c: 0x14bb, 0xd9d: 0x0817, + 0xd9e: 0x08ff, 0xd9f: 0x0a97, 0xda0: 0x0cd3, 0xda1: 0x0d1f, 0xda2: 0x0d5f, 0xda3: 0x0df3, + 0xda4: 0x0f47, 0xda5: 0x0fbb, 0xda6: 0x1157, 0xda7: 0x12f7, 0xda8: 0x1303, 0xda9: 0x1457, + 0xdaa: 0x14d7, 0xdab: 0x0883, 0xdac: 0x0e4b, 0xdad: 0x0903, 0xdae: 0x0ec7, 0xdaf: 0x0f6b, + 0xdb0: 0x1287, 0xdb1: 0x14bf, 0xdb2: 0x15ab, 0xdb3: 0x15d3, 0xdb4: 0x0d37, 0xdb5: 0x0e27, + 0xdb6: 0x11c3, 0xdb7: 0x10b7, 0xdb8: 0x10c3, 0xdb9: 0x10e7, 0xdba: 0x0f17, 0xdbb: 0x0e9f, + 0xdbc: 0x1363, 0xdbd: 0x0733, 0xdbe: 0x122b, 0xdbf: 0x081b, + // Block 0x37, offset 0xdc0 + 0xdc0: 0x080b, 0xdc1: 0x0b0b, 0xdc2: 0x0c2b, 0xdc3: 0x10f3, 0xdc4: 0x0a53, 0xdc5: 0x0e03, + 0xdc6: 0x0cef, 0xdc7: 0x13e7, 0xdc8: 0x12e7, 0xdc9: 0x14ab, 0xdca: 0x1323, 0xdcb: 0x0b27, + 0xdcc: 0x0787, 0xdcd: 0x095b, 0xdd0: 0x09af, + 0xdd2: 0x0cdf, 0xdd5: 0x07f7, 0xdd6: 0x0f1f, 0xdd7: 0x0fe3, + 0xdd8: 0x1047, 0xdd9: 0x1063, 0xdda: 0x1067, 0xddb: 0x107b, 0xddc: 0x14fb, 0xddd: 0x10eb, + 0xdde: 0x116f, 0xde0: 0x128f, 0xde2: 0x1353, + 0xde5: 0x1407, 0xde6: 0x1433, + 0xdea: 0x154f, 0xdeb: 0x1553, 0xdec: 0x1557, 0xded: 0x15bb, 0xdee: 0x142b, 0xdef: 0x14c7, + 0xdf0: 0x0757, 0xdf1: 0x077b, 0xdf2: 0x078f, 0xdf3: 0x084b, 0xdf4: 0x0857, 0xdf5: 0x0897, + 0xdf6: 0x094b, 0xdf7: 0x0967, 0xdf8: 0x096f, 0xdf9: 0x09ab, 0xdfa: 0x09b7, 0xdfb: 0x0a93, + 0xdfc: 0x0a9b, 0xdfd: 0x0ba3, 0xdfe: 0x0bcb, 0xdff: 0x0bd3, + // Block 0x38, offset 0xe00 + 0xe00: 0x0beb, 0xe01: 0x0c97, 0xe02: 0x0cc7, 0xe03: 0x0ce7, 0xe04: 0x0d57, 0xe05: 0x0e1b, + 0xe06: 0x0e37, 0xe07: 0x0e67, 0xe08: 0x0ebb, 0xe09: 0x0edb, 0xe0a: 0x0f4f, 0xe0b: 0x102f, + 0xe0c: 0x104b, 0xe0d: 0x1053, 0xe0e: 0x104f, 0xe0f: 0x1057, 0xe10: 0x105b, 0xe11: 0x105f, + 0xe12: 0x1073, 0xe13: 0x1077, 0xe14: 0x109b, 0xe15: 0x10af, 0xe16: 0x10cb, 0xe17: 0x112f, + 0xe18: 0x1137, 0xe19: 0x113f, 0xe1a: 0x1153, 0xe1b: 0x117b, 0xe1c: 0x11cb, 0xe1d: 0x11ff, + 0xe1e: 0x11ff, 0xe1f: 0x1267, 0xe20: 0x130f, 0xe21: 0x1327, 0xe22: 0x135b, 0xe23: 0x135f, + 0xe24: 0x13a3, 0xe25: 0x13a7, 0xe26: 0x13ff, 0xe27: 0x1407, 0xe28: 0x14db, 0xe29: 0x151f, + 0xe2a: 0x1537, 0xe2b: 0x0b9b, 0xe2c: 0x171e, 0xe2d: 0x11e3, + 0xe30: 0x06df, 0xe31: 0x07e3, 0xe32: 0x07a3, 0xe33: 0x074b, 0xe34: 0x078b, 0xe35: 0x07b7, + 0xe36: 0x0847, 0xe37: 0x0863, 0xe38: 0x094b, 0xe39: 0x0937, 0xe3a: 0x0947, 0xe3b: 0x0963, + 0xe3c: 0x09af, 0xe3d: 0x09bf, 0xe3e: 0x0a03, 0xe3f: 0x0a0f, + // Block 0x39, offset 0xe40 + 0xe40: 0x0a2b, 0xe41: 0x0a3b, 0xe42: 0x0b23, 0xe43: 0x0b2b, 0xe44: 0x0b5b, 0xe45: 0x0b7b, + 0xe46: 0x0bab, 0xe47: 0x0bc3, 0xe48: 0x0bb3, 0xe49: 0x0bd3, 0xe4a: 0x0bc7, 0xe4b: 0x0beb, + 0xe4c: 0x0c07, 0xe4d: 0x0c5f, 0xe4e: 0x0c6b, 0xe4f: 0x0c73, 0xe50: 0x0c9b, 0xe51: 0x0cdf, + 0xe52: 0x0d0f, 0xe53: 0x0d13, 0xe54: 0x0d27, 0xe55: 0x0da7, 0xe56: 0x0db7, 0xe57: 0x0e0f, + 0xe58: 0x0e5b, 0xe59: 0x0e53, 0xe5a: 0x0e67, 0xe5b: 0x0e83, 0xe5c: 0x0ebb, 0xe5d: 0x1013, + 0xe5e: 0x0edf, 0xe5f: 0x0f13, 0xe60: 0x0f1f, 0xe61: 0x0f5f, 0xe62: 0x0f7b, 0xe63: 0x0f9f, + 0xe64: 0x0fc3, 0xe65: 0x0fc7, 0xe66: 0x0fe3, 0xe67: 0x0fe7, 0xe68: 0x0ff7, 0xe69: 0x100b, + 0xe6a: 0x1007, 0xe6b: 0x1037, 0xe6c: 0x10b3, 0xe6d: 0x10cb, 0xe6e: 0x10e3, 0xe6f: 0x111b, + 0xe70: 0x112f, 0xe71: 0x114b, 0xe72: 0x117b, 0xe73: 0x122f, 0xe74: 0x1257, 0xe75: 0x12cb, + 0xe76: 0x1313, 0xe77: 0x131f, 0xe78: 0x1327, 0xe79: 0x133f, 0xe7a: 0x1353, 0xe7b: 0x1343, + 0xe7c: 0x135b, 0xe7d: 0x1357, 0xe7e: 0x134f, 0xe7f: 0x135f, + // Block 0x3a, offset 0xe80 + 0xe80: 0x136b, 0xe81: 0x13a7, 0xe82: 0x13e3, 0xe83: 0x1413, 0xe84: 0x144b, 0xe85: 0x146b, + 0xe86: 0x14b7, 0xe87: 0x14db, 0xe88: 0x14fb, 0xe89: 0x150f, 0xe8a: 0x151f, 0xe8b: 0x152b, + 0xe8c: 0x1537, 0xe8d: 0x158b, 0xe8e: 0x162b, 0xe8f: 0x16b5, 0xe90: 0x16b0, 0xe91: 0x16e2, + 0xe92: 0x0607, 0xe93: 0x062f, 0xe94: 0x0633, 0xe95: 0x1764, 0xe96: 0x1791, 0xe97: 0x1809, + 0xe98: 0x1617, 0xe99: 0x1627, + // Block 0x3b, offset 0xec0 + 0xec0: 0x19d8, 0xec1: 0x19db, 0xec2: 0x19de, 0xec3: 0x1c0b, 0xec4: 0x1c0f, 0xec5: 0x1a62, + 0xec6: 0x1a62, + 0xed3: 0x1d78, 0xed4: 0x1d69, 0xed5: 0x1d6e, 0xed6: 0x1d7d, 0xed7: 0x1d73, + 0xedd: 0x4393, + 0xede: 0x8115, 0xedf: 0x4405, 0xee0: 0x022d, 0xee1: 0x0215, 0xee2: 0x021e, 0xee3: 0x0221, + 0xee4: 0x0224, 0xee5: 0x0227, 0xee6: 0x022a, 0xee7: 0x0230, 0xee8: 0x0233, 0xee9: 0x0017, + 0xeea: 0x43f3, 0xeeb: 0x43f9, 0xeec: 0x44f7, 0xeed: 0x44ff, 0xeee: 0x434b, 0xeef: 0x4351, + 0xef0: 0x4357, 0xef1: 0x435d, 0xef2: 0x4369, 0xef3: 0x436f, 0xef4: 0x4375, 0xef5: 0x4381, + 0xef6: 0x4387, 0xef8: 0x438d, 0xef9: 0x4399, 0xefa: 0x439f, 0xefb: 0x43a5, + 0xefc: 0x43b1, 0xefe: 0x43b7, + // Block 0x3c, offset 0xf00 + 0xf00: 0x43bd, 0xf01: 0x43c3, 0xf03: 0x43c9, 0xf04: 0x43cf, + 0xf06: 0x43db, 0xf07: 0x43e1, 0xf08: 0x43e7, 0xf09: 0x43ed, 0xf0a: 0x43ff, 0xf0b: 0x437b, + 0xf0c: 0x4363, 0xf0d: 0x43ab, 0xf0e: 0x43d5, 0xf0f: 0x1d82, 0xf10: 0x0299, 0xf11: 0x0299, + 0xf12: 0x02a2, 0xf13: 0x02a2, 0xf14: 0x02a2, 0xf15: 0x02a2, 0xf16: 0x02a5, 0xf17: 0x02a5, + 0xf18: 0x02a5, 0xf19: 0x02a5, 0xf1a: 0x02ab, 0xf1b: 0x02ab, 0xf1c: 0x02ab, 0xf1d: 0x02ab, + 0xf1e: 0x029f, 0xf1f: 0x029f, 0xf20: 0x029f, 0xf21: 0x029f, 0xf22: 0x02a8, 0xf23: 0x02a8, + 0xf24: 0x02a8, 0xf25: 0x02a8, 0xf26: 0x029c, 0xf27: 0x029c, 0xf28: 0x029c, 0xf29: 0x029c, + 0xf2a: 0x02cf, 0xf2b: 0x02cf, 0xf2c: 0x02cf, 0xf2d: 0x02cf, 0xf2e: 0x02d2, 0xf2f: 0x02d2, + 0xf30: 0x02d2, 0xf31: 0x02d2, 0xf32: 0x02b1, 0xf33: 0x02b1, 0xf34: 0x02b1, 0xf35: 0x02b1, + 0xf36: 0x02ae, 0xf37: 0x02ae, 0xf38: 0x02ae, 0xf39: 0x02ae, 0xf3a: 0x02b4, 0xf3b: 0x02b4, + 0xf3c: 0x02b4, 0xf3d: 0x02b4, 0xf3e: 0x02b7, 0xf3f: 0x02b7, + // Block 0x3d, offset 0xf40 + 0xf40: 0x02b7, 0xf41: 0x02b7, 0xf42: 0x02c0, 0xf43: 0x02c0, 0xf44: 0x02bd, 0xf45: 0x02bd, + 0xf46: 0x02c3, 0xf47: 0x02c3, 0xf48: 0x02ba, 0xf49: 0x02ba, 0xf4a: 0x02c9, 0xf4b: 0x02c9, + 0xf4c: 0x02c6, 0xf4d: 0x02c6, 0xf4e: 0x02d5, 0xf4f: 0x02d5, 0xf50: 0x02d5, 0xf51: 0x02d5, + 0xf52: 0x02db, 0xf53: 0x02db, 0xf54: 0x02db, 0xf55: 0x02db, 0xf56: 0x02e1, 0xf57: 0x02e1, + 0xf58: 0x02e1, 0xf59: 0x02e1, 0xf5a: 0x02de, 0xf5b: 0x02de, 0xf5c: 0x02de, 0xf5d: 0x02de, + 0xf5e: 0x02e4, 0xf5f: 0x02e4, 0xf60: 0x02e7, 0xf61: 0x02e7, 0xf62: 0x02e7, 0xf63: 0x02e7, + 0xf64: 0x4471, 0xf65: 0x4471, 0xf66: 0x02ed, 0xf67: 0x02ed, 0xf68: 0x02ed, 0xf69: 0x02ed, + 0xf6a: 0x02ea, 0xf6b: 0x02ea, 0xf6c: 0x02ea, 0xf6d: 0x02ea, 0xf6e: 0x0308, 0xf6f: 0x0308, + 0xf70: 0x446b, 0xf71: 0x446b, + // Block 0x3e, offset 0xf80 + 0xf93: 0x02d8, 0xf94: 0x02d8, 0xf95: 0x02d8, 0xf96: 0x02d8, 0xf97: 0x02f6, + 0xf98: 0x02f6, 0xf99: 0x02f3, 0xf9a: 0x02f3, 0xf9b: 0x02f9, 0xf9c: 0x02f9, 0xf9d: 0x2052, + 0xf9e: 0x02ff, 0xf9f: 0x02ff, 0xfa0: 0x02f0, 0xfa1: 0x02f0, 0xfa2: 0x02fc, 0xfa3: 0x02fc, + 0xfa4: 0x0305, 0xfa5: 0x0305, 0xfa6: 0x0305, 0xfa7: 0x0305, 0xfa8: 0x028d, 0xfa9: 0x028d, + 0xfaa: 0x25ad, 0xfab: 0x25ad, 0xfac: 0x261d, 0xfad: 0x261d, 0xfae: 0x25ec, 0xfaf: 0x25ec, + 0xfb0: 0x2608, 0xfb1: 0x2608, 0xfb2: 0x2601, 0xfb3: 0x2601, 0xfb4: 0x260f, 0xfb5: 0x260f, + 0xfb6: 0x2616, 0xfb7: 0x2616, 0xfb8: 0x2616, 0xfb9: 0x25f3, 0xfba: 0x25f3, 0xfbb: 0x25f3, + 0xfbc: 0x0302, 0xfbd: 0x0302, 0xfbe: 0x0302, 0xfbf: 0x0302, + // Block 0x3f, offset 0xfc0 + 0xfc0: 0x25b4, 0xfc1: 0x25bb, 0xfc2: 0x25d7, 0xfc3: 0x25f3, 0xfc4: 0x25fa, 0xfc5: 0x1d8c, + 0xfc6: 0x1d91, 0xfc7: 0x1d96, 0xfc8: 0x1da5, 0xfc9: 0x1db4, 0xfca: 0x1db9, 0xfcb: 0x1dbe, + 0xfcc: 0x1dc3, 0xfcd: 0x1dc8, 0xfce: 0x1dd7, 0xfcf: 0x1de6, 0xfd0: 0x1deb, 0xfd1: 0x1df0, + 0xfd2: 0x1dff, 0xfd3: 0x1e0e, 0xfd4: 0x1e13, 0xfd5: 0x1e18, 0xfd6: 0x1e1d, 0xfd7: 0x1e2c, + 0xfd8: 0x1e31, 0xfd9: 0x1e40, 0xfda: 0x1e45, 0xfdb: 0x1e4a, 0xfdc: 0x1e59, 0xfdd: 0x1e5e, + 0xfde: 0x1e63, 0xfdf: 0x1e6d, 0xfe0: 0x1ea9, 0xfe1: 0x1eb8, 0xfe2: 0x1ec7, 0xfe3: 0x1ecc, + 0xfe4: 0x1ed1, 0xfe5: 0x1edb, 0xfe6: 0x1eea, 0xfe7: 0x1eef, 0xfe8: 0x1efe, 0xfe9: 0x1f03, + 0xfea: 0x1f08, 0xfeb: 0x1f17, 0xfec: 0x1f1c, 0xfed: 0x1f2b, 0xfee: 0x1f30, 0xfef: 0x1f35, + 0xff0: 0x1f3a, 0xff1: 0x1f3f, 0xff2: 0x1f44, 0xff3: 0x1f49, 0xff4: 0x1f4e, 0xff5: 0x1f53, + 0xff6: 0x1f58, 0xff7: 0x1f5d, 0xff8: 0x1f62, 0xff9: 0x1f67, 0xffa: 0x1f6c, 0xffb: 0x1f71, + 0xffc: 0x1f76, 0xffd: 0x1f7b, 0xffe: 0x1f80, 0xfff: 0x1f8a, + // Block 0x40, offset 0x1000 + 0x1000: 0x1f8f, 0x1001: 0x1f94, 0x1002: 0x1f99, 0x1003: 0x1fa3, 0x1004: 0x1fa8, 0x1005: 0x1fb2, + 0x1006: 0x1fb7, 0x1007: 0x1fbc, 0x1008: 0x1fc1, 0x1009: 0x1fc6, 0x100a: 0x1fcb, 0x100b: 0x1fd0, + 0x100c: 0x1fd5, 0x100d: 0x1fda, 0x100e: 0x1fe9, 0x100f: 0x1ff8, 0x1010: 0x1ffd, 0x1011: 0x2002, + 0x1012: 0x2007, 0x1013: 0x200c, 0x1014: 0x2011, 0x1015: 0x201b, 0x1016: 0x2020, 0x1017: 0x2025, + 0x1018: 0x2034, 0x1019: 0x2043, 0x101a: 0x2048, 0x101b: 0x4423, 0x101c: 0x4429, 0x101d: 0x445f, + 0x101e: 0x44b6, 0x101f: 0x44bd, 0x1020: 0x44c4, 0x1021: 0x44cb, 0x1022: 0x44d2, 0x1023: 0x44d9, + 0x1024: 0x25c9, 0x1025: 0x25d0, 0x1026: 0x25d7, 0x1027: 0x25de, 0x1028: 0x25f3, 0x1029: 0x25fa, + 0x102a: 0x1d9b, 0x102b: 0x1da0, 0x102c: 0x1da5, 0x102d: 0x1daa, 0x102e: 0x1db4, 0x102f: 0x1db9, + 0x1030: 0x1dcd, 0x1031: 0x1dd2, 0x1032: 0x1dd7, 0x1033: 0x1ddc, 0x1034: 0x1de6, 0x1035: 0x1deb, + 0x1036: 0x1df5, 0x1037: 0x1dfa, 0x1038: 0x1dff, 0x1039: 0x1e04, 0x103a: 0x1e0e, 0x103b: 0x1e13, + 0x103c: 0x1f3f, 0x103d: 0x1f44, 0x103e: 0x1f53, 0x103f: 0x1f58, + // Block 0x41, offset 0x1040 + 0x1040: 0x1f5d, 0x1041: 0x1f71, 0x1042: 0x1f76, 0x1043: 0x1f7b, 0x1044: 0x1f80, 0x1045: 0x1f99, + 0x1046: 0x1fa3, 0x1047: 0x1fa8, 0x1048: 0x1fad, 0x1049: 0x1fc1, 0x104a: 0x1fdf, 0x104b: 0x1fe4, + 0x104c: 0x1fe9, 0x104d: 0x1fee, 0x104e: 0x1ff8, 0x104f: 0x1ffd, 0x1050: 0x445f, 0x1051: 0x202a, + 0x1052: 0x202f, 0x1053: 0x2034, 0x1054: 0x2039, 0x1055: 0x2043, 0x1056: 0x2048, 0x1057: 0x25b4, + 0x1058: 0x25bb, 0x1059: 0x25c2, 0x105a: 0x25d7, 0x105b: 0x25e5, 0x105c: 0x1d8c, 0x105d: 0x1d91, + 0x105e: 0x1d96, 0x105f: 0x1da5, 0x1060: 0x1daf, 0x1061: 0x1dbe, 0x1062: 0x1dc3, 0x1063: 0x1dc8, + 0x1064: 0x1dd7, 0x1065: 0x1de1, 0x1066: 0x1dff, 0x1067: 0x1e18, 0x1068: 0x1e1d, 0x1069: 0x1e2c, + 0x106a: 0x1e31, 0x106b: 0x1e40, 0x106c: 0x1e4a, 0x106d: 0x1e59, 0x106e: 0x1e5e, 0x106f: 0x1e63, + 0x1070: 0x1e6d, 0x1071: 0x1ea9, 0x1072: 0x1eae, 0x1073: 0x1eb8, 0x1074: 0x1ec7, 0x1075: 0x1ecc, + 0x1076: 0x1ed1, 0x1077: 0x1edb, 0x1078: 0x1eea, 0x1079: 0x1efe, 0x107a: 0x1f03, 0x107b: 0x1f08, + 0x107c: 0x1f17, 0x107d: 0x1f1c, 0x107e: 0x1f2b, 0x107f: 0x1f30, + // Block 0x42, offset 0x1080 + 0x1080: 0x1f35, 0x1081: 0x1f3a, 0x1082: 0x1f49, 0x1083: 0x1f4e, 0x1084: 0x1f62, 0x1085: 0x1f67, + 0x1086: 0x1f6c, 0x1087: 0x1f71, 0x1088: 0x1f76, 0x1089: 0x1f8a, 0x108a: 0x1f8f, 0x108b: 0x1f94, + 0x108c: 0x1f99, 0x108d: 0x1f9e, 0x108e: 0x1fb2, 0x108f: 0x1fb7, 0x1090: 0x1fbc, 0x1091: 0x1fc1, + 0x1092: 0x1fd0, 0x1093: 0x1fd5, 0x1094: 0x1fda, 0x1095: 0x1fe9, 0x1096: 0x1ff3, 0x1097: 0x2002, + 0x1098: 0x2007, 0x1099: 0x4453, 0x109a: 0x201b, 0x109b: 0x2020, 0x109c: 0x2025, 0x109d: 0x2034, + 0x109e: 0x203e, 0x109f: 0x25d7, 0x10a0: 0x25e5, 0x10a1: 0x1da5, 0x10a2: 0x1daf, 0x10a3: 0x1dd7, + 0x10a4: 0x1de1, 0x10a5: 0x1dff, 0x10a6: 0x1e09, 0x10a7: 0x1e6d, 0x10a8: 0x1e72, 0x10a9: 0x1e95, + 0x10aa: 0x1e9a, 0x10ab: 0x1f71, 0x10ac: 0x1f76, 0x10ad: 0x1f99, 0x10ae: 0x1fe9, 0x10af: 0x1ff3, + 0x10b0: 0x2034, 0x10b1: 0x203e, 0x10b2: 0x4507, 0x10b3: 0x450f, 0x10b4: 0x4517, 0x10b5: 0x1ef4, + 0x10b6: 0x1ef9, 0x10b7: 0x1f0d, 0x10b8: 0x1f12, 0x10b9: 0x1f21, 0x10ba: 0x1f26, 0x10bb: 0x1e77, + 0x10bc: 0x1e7c, 0x10bd: 0x1e9f, 0x10be: 0x1ea4, 0x10bf: 0x1e36, + // Block 0x43, offset 0x10c0 + 0x10c0: 0x1e3b, 0x10c1: 0x1e22, 0x10c2: 0x1e27, 0x10c3: 0x1e4f, 0x10c4: 0x1e54, 0x10c5: 0x1ebd, + 0x10c6: 0x1ec2, 0x10c7: 0x1ee0, 0x10c8: 0x1ee5, 0x10c9: 0x1e81, 0x10ca: 0x1e86, 0x10cb: 0x1e8b, + 0x10cc: 0x1e95, 0x10cd: 0x1e90, 0x10ce: 0x1e68, 0x10cf: 0x1eb3, 0x10d0: 0x1ed6, 0x10d1: 0x1ef4, + 0x10d2: 0x1ef9, 0x10d3: 0x1f0d, 0x10d4: 0x1f12, 0x10d5: 0x1f21, 0x10d6: 0x1f26, 0x10d7: 0x1e77, + 0x10d8: 0x1e7c, 0x10d9: 0x1e9f, 0x10da: 0x1ea4, 0x10db: 0x1e36, 0x10dc: 0x1e3b, 0x10dd: 0x1e22, + 0x10de: 0x1e27, 0x10df: 0x1e4f, 0x10e0: 0x1e54, 0x10e1: 0x1ebd, 0x10e2: 0x1ec2, 0x10e3: 0x1ee0, + 0x10e4: 0x1ee5, 0x10e5: 0x1e81, 0x10e6: 0x1e86, 0x10e7: 0x1e8b, 0x10e8: 0x1e95, 0x10e9: 0x1e90, + 0x10ea: 0x1e68, 0x10eb: 0x1eb3, 0x10ec: 0x1ed6, 0x10ed: 0x1e81, 0x10ee: 0x1e86, 0x10ef: 0x1e8b, + 0x10f0: 0x1e95, 0x10f1: 0x1e72, 0x10f2: 0x1e9a, 0x10f3: 0x1eef, 0x10f4: 0x1e59, 0x10f5: 0x1e5e, + 0x10f6: 0x1e63, 0x10f7: 0x1e81, 0x10f8: 0x1e86, 0x10f9: 0x1e8b, 0x10fa: 0x1eef, 0x10fb: 0x1efe, + 0x10fc: 0x440b, 0x10fd: 0x440b, + // Block 0x44, offset 0x1100 + 0x1110: 0x2314, 0x1111: 0x2329, + 0x1112: 0x2329, 0x1113: 0x2330, 0x1114: 0x2337, 0x1115: 0x234c, 0x1116: 0x2353, 0x1117: 0x235a, + 0x1118: 0x237d, 0x1119: 0x237d, 0x111a: 0x23a0, 0x111b: 0x2399, 0x111c: 0x23b5, 0x111d: 0x23a7, + 0x111e: 0x23ae, 0x111f: 0x23d1, 0x1120: 0x23d1, 0x1121: 0x23ca, 0x1122: 0x23d8, 0x1123: 0x23d8, + 0x1124: 0x2402, 0x1125: 0x2402, 0x1126: 0x241e, 0x1127: 0x23e6, 0x1128: 0x23e6, 0x1129: 0x23df, + 0x112a: 0x23f4, 0x112b: 0x23f4, 0x112c: 0x23fb, 0x112d: 0x23fb, 0x112e: 0x2425, 0x112f: 0x2433, + 0x1130: 0x2433, 0x1131: 0x243a, 0x1132: 0x243a, 0x1133: 0x2441, 0x1134: 0x2448, 0x1135: 0x244f, + 0x1136: 0x2456, 0x1137: 0x2456, 0x1138: 0x245d, 0x1139: 0x246b, 0x113a: 0x2479, 0x113b: 0x2472, + 0x113c: 0x2480, 0x113d: 0x2480, 0x113e: 0x2495, 0x113f: 0x249c, + // Block 0x45, offset 0x1140 + 0x1140: 0x24cd, 0x1141: 0x24db, 0x1142: 0x24d4, 0x1143: 0x24b8, 0x1144: 0x24b8, 0x1145: 0x24e2, + 0x1146: 0x24e2, 0x1147: 0x24e9, 0x1148: 0x24e9, 0x1149: 0x2513, 0x114a: 0x251a, 0x114b: 0x2521, + 0x114c: 0x24f7, 0x114d: 0x2505, 0x114e: 0x2528, 0x114f: 0x252f, + 0x1152: 0x24fe, 0x1153: 0x2583, 0x1154: 0x258a, 0x1155: 0x2560, 0x1156: 0x2567, 0x1157: 0x254b, + 0x1158: 0x254b, 0x1159: 0x2552, 0x115a: 0x257c, 0x115b: 0x2575, 0x115c: 0x259f, 0x115d: 0x259f, + 0x115e: 0x230d, 0x115f: 0x2322, 0x1160: 0x231b, 0x1161: 0x2345, 0x1162: 0x233e, 0x1163: 0x2368, + 0x1164: 0x2361, 0x1165: 0x238b, 0x1166: 0x236f, 0x1167: 0x2384, 0x1168: 0x23bc, 0x1169: 0x2409, + 0x116a: 0x23ed, 0x116b: 0x242c, 0x116c: 0x24c6, 0x116d: 0x24f0, 0x116e: 0x2598, 0x116f: 0x2591, + 0x1170: 0x25a6, 0x1171: 0x253d, 0x1172: 0x24a3, 0x1173: 0x256e, 0x1174: 0x2495, 0x1175: 0x24cd, + 0x1176: 0x2464, 0x1177: 0x24b1, 0x1178: 0x2544, 0x1179: 0x2536, 0x117a: 0x24bf, 0x117b: 0x24aa, + 0x117c: 0x24bf, 0x117d: 0x2544, 0x117e: 0x2376, 0x117f: 0x2392, + // Block 0x46, offset 0x1180 + 0x1180: 0x250c, 0x1181: 0x2487, 0x1182: 0x2306, 0x1183: 0x24aa, 0x1184: 0x244f, 0x1185: 0x241e, + 0x1186: 0x23c3, 0x1187: 0x2559, + 0x11b0: 0x2417, 0x11b1: 0x248e, 0x11b2: 0x27c2, 0x11b3: 0x27b9, 0x11b4: 0x27ef, 0x11b5: 0x27dd, + 0x11b6: 0x27cb, 0x11b7: 0x27e6, 0x11b8: 0x27f8, 0x11b9: 0x2410, 0x11ba: 0x2c7f, 0x11bb: 0x2aff, + 0x11bc: 0x27d4, + // Block 0x47, offset 0x11c0 + 0x11d0: 0x0019, 0x11d1: 0x0483, + 0x11d2: 0x0487, 0x11d3: 0x0035, 0x11d4: 0x0037, 0x11d5: 0x0003, 0x11d6: 0x003f, 0x11d7: 0x04bf, + 0x11d8: 0x04c3, 0x11d9: 0x1b5f, + 0x11e0: 0x8132, 0x11e1: 0x8132, 0x11e2: 0x8132, 0x11e3: 0x8132, + 0x11e4: 0x8132, 0x11e5: 0x8132, 0x11e6: 0x8132, 0x11e7: 0x812d, 0x11e8: 0x812d, 0x11e9: 0x812d, + 0x11ea: 0x812d, 0x11eb: 0x812d, 0x11ec: 0x812d, 0x11ed: 0x812d, 0x11ee: 0x8132, 0x11ef: 0x8132, + 0x11f0: 0x1873, 0x11f1: 0x0443, 0x11f2: 0x043f, 0x11f3: 0x007f, 0x11f4: 0x007f, 0x11f5: 0x0011, + 0x11f6: 0x0013, 0x11f7: 0x00b7, 0x11f8: 0x00bb, 0x11f9: 0x04b7, 0x11fa: 0x04bb, 0x11fb: 0x04ab, + 0x11fc: 0x04af, 0x11fd: 0x0493, 0x11fe: 0x0497, 0x11ff: 0x048b, + // Block 0x48, offset 0x1200 + 0x1200: 0x048f, 0x1201: 0x049b, 0x1202: 0x049f, 0x1203: 0x04a3, 0x1204: 0x04a7, + 0x1207: 0x0077, 0x1208: 0x007b, 0x1209: 0x426c, 0x120a: 0x426c, 0x120b: 0x426c, + 0x120c: 0x426c, 0x120d: 0x007f, 0x120e: 0x007f, 0x120f: 0x007f, 0x1210: 0x0019, 0x1211: 0x0483, + 0x1212: 0x001d, 0x1214: 0x0037, 0x1215: 0x0035, 0x1216: 0x003f, 0x1217: 0x0003, + 0x1218: 0x0443, 0x1219: 0x0011, 0x121a: 0x0013, 0x121b: 0x00b7, 0x121c: 0x00bb, 0x121d: 0x04b7, + 0x121e: 0x04bb, 0x121f: 0x0007, 0x1220: 0x000d, 0x1221: 0x0015, 0x1222: 0x0017, 0x1223: 0x001b, + 0x1224: 0x0039, 0x1225: 0x003d, 0x1226: 0x003b, 0x1228: 0x0079, 0x1229: 0x0009, + 0x122a: 0x000b, 0x122b: 0x0041, + 0x1230: 0x42ad, 0x1231: 0x442f, 0x1232: 0x42b2, 0x1234: 0x42b7, + 0x1236: 0x42bc, 0x1237: 0x4435, 0x1238: 0x42c1, 0x1239: 0x443b, 0x123a: 0x42c6, 0x123b: 0x4441, + 0x123c: 0x42cb, 0x123d: 0x4447, 0x123e: 0x42d0, 0x123f: 0x444d, + // Block 0x49, offset 0x1240 + 0x1240: 0x0236, 0x1241: 0x4411, 0x1242: 0x4411, 0x1243: 0x4417, 0x1244: 0x4417, 0x1245: 0x4459, + 0x1246: 0x4459, 0x1247: 0x441d, 0x1248: 0x441d, 0x1249: 0x4465, 0x124a: 0x4465, 0x124b: 0x4465, + 0x124c: 0x4465, 0x124d: 0x0239, 0x124e: 0x0239, 0x124f: 0x023c, 0x1250: 0x023c, 0x1251: 0x023c, + 0x1252: 0x023c, 0x1253: 0x023f, 0x1254: 0x023f, 0x1255: 0x0242, 0x1256: 0x0242, 0x1257: 0x0242, + 0x1258: 0x0242, 0x1259: 0x0245, 0x125a: 0x0245, 0x125b: 0x0245, 0x125c: 0x0245, 0x125d: 0x0248, + 0x125e: 0x0248, 0x125f: 0x0248, 0x1260: 0x0248, 0x1261: 0x024b, 0x1262: 0x024b, 0x1263: 0x024b, + 0x1264: 0x024b, 0x1265: 0x024e, 0x1266: 0x024e, 0x1267: 0x024e, 0x1268: 0x024e, 0x1269: 0x0251, + 0x126a: 0x0251, 0x126b: 0x0254, 0x126c: 0x0254, 0x126d: 0x0257, 0x126e: 0x0257, 0x126f: 0x025a, + 0x1270: 0x025a, 0x1271: 0x025d, 0x1272: 0x025d, 0x1273: 0x025d, 0x1274: 0x025d, 0x1275: 0x0260, + 0x1276: 0x0260, 0x1277: 0x0260, 0x1278: 0x0260, 0x1279: 0x0263, 0x127a: 0x0263, 0x127b: 0x0263, + 0x127c: 0x0263, 0x127d: 0x0266, 0x127e: 0x0266, 0x127f: 0x0266, + // Block 0x4a, offset 0x1280 + 0x1280: 0x0266, 0x1281: 0x0269, 0x1282: 0x0269, 0x1283: 0x0269, 0x1284: 0x0269, 0x1285: 0x026c, + 0x1286: 0x026c, 0x1287: 0x026c, 0x1288: 0x026c, 0x1289: 0x026f, 0x128a: 0x026f, 0x128b: 0x026f, + 0x128c: 0x026f, 0x128d: 0x0272, 0x128e: 0x0272, 0x128f: 0x0272, 0x1290: 0x0272, 0x1291: 0x0275, + 0x1292: 0x0275, 0x1293: 0x0275, 0x1294: 0x0275, 0x1295: 0x0278, 0x1296: 0x0278, 0x1297: 0x0278, + 0x1298: 0x0278, 0x1299: 0x027b, 0x129a: 0x027b, 0x129b: 0x027b, 0x129c: 0x027b, 0x129d: 0x027e, + 0x129e: 0x027e, 0x129f: 0x027e, 0x12a0: 0x027e, 0x12a1: 0x0281, 0x12a2: 0x0281, 0x12a3: 0x0281, + 0x12a4: 0x0281, 0x12a5: 0x0284, 0x12a6: 0x0284, 0x12a7: 0x0284, 0x12a8: 0x0284, 0x12a9: 0x0287, + 0x12aa: 0x0287, 0x12ab: 0x0287, 0x12ac: 0x0287, 0x12ad: 0x028a, 0x12ae: 0x028a, 0x12af: 0x028d, + 0x12b0: 0x028d, 0x12b1: 0x0290, 0x12b2: 0x0290, 0x12b3: 0x0290, 0x12b4: 0x0290, 0x12b5: 0x2e03, + 0x12b6: 0x2e03, 0x12b7: 0x2e0b, 0x12b8: 0x2e0b, 0x12b9: 0x2e13, 0x12ba: 0x2e13, 0x12bb: 0x1f85, + 0x12bc: 0x1f85, + // Block 0x4b, offset 0x12c0 + 0x12c0: 0x0081, 0x12c1: 0x0083, 0x12c2: 0x0085, 0x12c3: 0x0087, 0x12c4: 0x0089, 0x12c5: 0x008b, + 0x12c6: 0x008d, 0x12c7: 0x008f, 0x12c8: 0x0091, 0x12c9: 0x0093, 0x12ca: 0x0095, 0x12cb: 0x0097, + 0x12cc: 0x0099, 0x12cd: 0x009b, 0x12ce: 0x009d, 0x12cf: 0x009f, 0x12d0: 0x00a1, 0x12d1: 0x00a3, + 0x12d2: 0x00a5, 0x12d3: 0x00a7, 0x12d4: 0x00a9, 0x12d5: 0x00ab, 0x12d6: 0x00ad, 0x12d7: 0x00af, + 0x12d8: 0x00b1, 0x12d9: 0x00b3, 0x12da: 0x00b5, 0x12db: 0x00b7, 0x12dc: 0x00b9, 0x12dd: 0x00bb, + 0x12de: 0x00bd, 0x12df: 0x0477, 0x12e0: 0x047b, 0x12e1: 0x0487, 0x12e2: 0x049b, 0x12e3: 0x049f, + 0x12e4: 0x0483, 0x12e5: 0x05ab, 0x12e6: 0x05a3, 0x12e7: 0x04c7, 0x12e8: 0x04cf, 0x12e9: 0x04d7, + 0x12ea: 0x04df, 0x12eb: 0x04e7, 0x12ec: 0x056b, 0x12ed: 0x0573, 0x12ee: 0x057b, 0x12ef: 0x051f, + 0x12f0: 0x05af, 0x12f1: 0x04cb, 0x12f2: 0x04d3, 0x12f3: 0x04db, 0x12f4: 0x04e3, 0x12f5: 0x04eb, + 0x12f6: 0x04ef, 0x12f7: 0x04f3, 0x12f8: 0x04f7, 0x12f9: 0x04fb, 0x12fa: 0x04ff, 0x12fb: 0x0503, + 0x12fc: 0x0507, 0x12fd: 0x050b, 0x12fe: 0x050f, 0x12ff: 0x0513, + // Block 0x4c, offset 0x1300 + 0x1300: 0x0517, 0x1301: 0x051b, 0x1302: 0x0523, 0x1303: 0x0527, 0x1304: 0x052b, 0x1305: 0x052f, + 0x1306: 0x0533, 0x1307: 0x0537, 0x1308: 0x053b, 0x1309: 0x053f, 0x130a: 0x0543, 0x130b: 0x0547, + 0x130c: 0x054b, 0x130d: 0x054f, 0x130e: 0x0553, 0x130f: 0x0557, 0x1310: 0x055b, 0x1311: 0x055f, + 0x1312: 0x0563, 0x1313: 0x0567, 0x1314: 0x056f, 0x1315: 0x0577, 0x1316: 0x057f, 0x1317: 0x0583, + 0x1318: 0x0587, 0x1319: 0x058b, 0x131a: 0x058f, 0x131b: 0x0593, 0x131c: 0x0597, 0x131d: 0x05a7, + 0x131e: 0x4a7b, 0x131f: 0x4a81, 0x1320: 0x03c3, 0x1321: 0x0313, 0x1322: 0x0317, 0x1323: 0x4a3e, + 0x1324: 0x031b, 0x1325: 0x4a44, 0x1326: 0x4a4a, 0x1327: 0x031f, 0x1328: 0x0323, 0x1329: 0x0327, + 0x132a: 0x4a50, 0x132b: 0x4a56, 0x132c: 0x4a5c, 0x132d: 0x4a62, 0x132e: 0x4a68, 0x132f: 0x4a6e, + 0x1330: 0x0367, 0x1331: 0x032b, 0x1332: 0x032f, 0x1333: 0x0333, 0x1334: 0x037b, 0x1335: 0x0337, + 0x1336: 0x033b, 0x1337: 0x033f, 0x1338: 0x0343, 0x1339: 0x0347, 0x133a: 0x034b, 0x133b: 0x034f, + 0x133c: 0x0353, 0x133d: 0x0357, 0x133e: 0x035b, + // Block 0x4d, offset 0x1340 + 0x1342: 0x49c0, 0x1343: 0x49c6, 0x1344: 0x49cc, 0x1345: 0x49d2, + 0x1346: 0x49d8, 0x1347: 0x49de, 0x134a: 0x49e4, 0x134b: 0x49ea, + 0x134c: 0x49f0, 0x134d: 0x49f6, 0x134e: 0x49fc, 0x134f: 0x4a02, + 0x1352: 0x4a08, 0x1353: 0x4a0e, 0x1354: 0x4a14, 0x1355: 0x4a1a, 0x1356: 0x4a20, 0x1357: 0x4a26, + 0x135a: 0x4a2c, 0x135b: 0x4a32, 0x135c: 0x4a38, + 0x1360: 0x00bf, 0x1361: 0x00c2, 0x1362: 0x00cb, 0x1363: 0x4267, + 0x1364: 0x00c8, 0x1365: 0x00c5, 0x1366: 0x0447, 0x1368: 0x046b, 0x1369: 0x044b, + 0x136a: 0x044f, 0x136b: 0x0453, 0x136c: 0x0457, 0x136d: 0x046f, 0x136e: 0x0473, + // Block 0x4e, offset 0x1380 + 0x1380: 0x0063, 0x1381: 0x0065, 0x1382: 0x0067, 0x1383: 0x0069, 0x1384: 0x006b, 0x1385: 0x006d, + 0x1386: 0x006f, 0x1387: 0x0071, 0x1388: 0x0073, 0x1389: 0x0075, 0x138a: 0x0083, 0x138b: 0x0085, + 0x138c: 0x0087, 0x138d: 0x0089, 0x138e: 0x008b, 0x138f: 0x008d, 0x1390: 0x008f, 0x1391: 0x0091, + 0x1392: 0x0093, 0x1393: 0x0095, 0x1394: 0x0097, 0x1395: 0x0099, 0x1396: 0x009b, 0x1397: 0x009d, + 0x1398: 0x009f, 0x1399: 0x00a1, 0x139a: 0x00a3, 0x139b: 0x00a5, 0x139c: 0x00a7, 0x139d: 0x00a9, + 0x139e: 0x00ab, 0x139f: 0x00ad, 0x13a0: 0x00af, 0x13a1: 0x00b1, 0x13a2: 0x00b3, 0x13a3: 0x00b5, + 0x13a4: 0x00dd, 0x13a5: 0x00f2, 0x13a8: 0x0173, 0x13a9: 0x0176, + 0x13aa: 0x0179, 0x13ab: 0x017c, 0x13ac: 0x017f, 0x13ad: 0x0182, 0x13ae: 0x0185, 0x13af: 0x0188, + 0x13b0: 0x018b, 0x13b1: 0x018e, 0x13b2: 0x0191, 0x13b3: 0x0194, 0x13b4: 0x0197, 0x13b5: 0x019a, + 0x13b6: 0x019d, 0x13b7: 0x01a0, 0x13b8: 0x01a3, 0x13b9: 0x0188, 0x13ba: 0x01a6, 0x13bb: 0x01a9, + 0x13bc: 0x01ac, 0x13bd: 0x01af, 0x13be: 0x01b2, 0x13bf: 0x01b5, + // Block 0x4f, offset 0x13c0 + 0x13c0: 0x01fd, 0x13c1: 0x0200, 0x13c2: 0x0203, 0x13c3: 0x045b, 0x13c4: 0x01c7, 0x13c5: 0x01d0, + 0x13c6: 0x01d6, 0x13c7: 0x01fa, 0x13c8: 0x01eb, 0x13c9: 0x01e8, 0x13ca: 0x0206, 0x13cb: 0x0209, + 0x13ce: 0x0021, 0x13cf: 0x0023, 0x13d0: 0x0025, 0x13d1: 0x0027, + 0x13d2: 0x0029, 0x13d3: 0x002b, 0x13d4: 0x002d, 0x13d5: 0x002f, 0x13d6: 0x0031, 0x13d7: 0x0033, + 0x13d8: 0x0021, 0x13d9: 0x0023, 0x13da: 0x0025, 0x13db: 0x0027, 0x13dc: 0x0029, 0x13dd: 0x002b, + 0x13de: 0x002d, 0x13df: 0x002f, 0x13e0: 0x0031, 0x13e1: 0x0033, 0x13e2: 0x0021, 0x13e3: 0x0023, + 0x13e4: 0x0025, 0x13e5: 0x0027, 0x13e6: 0x0029, 0x13e7: 0x002b, 0x13e8: 0x002d, 0x13e9: 0x002f, + 0x13ea: 0x0031, 0x13eb: 0x0033, 0x13ec: 0x0021, 0x13ed: 0x0023, 0x13ee: 0x0025, 0x13ef: 0x0027, + 0x13f0: 0x0029, 0x13f1: 0x002b, 0x13f2: 0x002d, 0x13f3: 0x002f, 0x13f4: 0x0031, 0x13f5: 0x0033, + 0x13f6: 0x0021, 0x13f7: 0x0023, 0x13f8: 0x0025, 0x13f9: 0x0027, 0x13fa: 0x0029, 0x13fb: 0x002b, + 0x13fc: 0x002d, 0x13fd: 0x002f, 0x13fe: 0x0031, 0x13ff: 0x0033, + // Block 0x50, offset 0x1400 + 0x1400: 0x0239, 0x1401: 0x023c, 0x1402: 0x0248, 0x1403: 0x0251, 0x1405: 0x028a, + 0x1406: 0x025a, 0x1407: 0x024b, 0x1408: 0x0269, 0x1409: 0x0290, 0x140a: 0x027b, 0x140b: 0x027e, + 0x140c: 0x0281, 0x140d: 0x0284, 0x140e: 0x025d, 0x140f: 0x026f, 0x1410: 0x0275, 0x1411: 0x0263, + 0x1412: 0x0278, 0x1413: 0x0257, 0x1414: 0x0260, 0x1415: 0x0242, 0x1416: 0x0245, 0x1417: 0x024e, + 0x1418: 0x0254, 0x1419: 0x0266, 0x141a: 0x026c, 0x141b: 0x0272, 0x141c: 0x0293, 0x141d: 0x02e4, + 0x141e: 0x02cc, 0x141f: 0x0296, 0x1421: 0x023c, 0x1422: 0x0248, + 0x1424: 0x0287, 0x1427: 0x024b, 0x1429: 0x0290, + 0x142a: 0x027b, 0x142b: 0x027e, 0x142c: 0x0281, 0x142d: 0x0284, 0x142e: 0x025d, 0x142f: 0x026f, + 0x1430: 0x0275, 0x1431: 0x0263, 0x1432: 0x0278, 0x1434: 0x0260, 0x1435: 0x0242, + 0x1436: 0x0245, 0x1437: 0x024e, 0x1439: 0x0266, 0x143b: 0x0272, + // Block 0x51, offset 0x1440 + 0x1442: 0x0248, + 0x1447: 0x024b, 0x1449: 0x0290, 0x144b: 0x027e, + 0x144d: 0x0284, 0x144e: 0x025d, 0x144f: 0x026f, 0x1451: 0x0263, + 0x1452: 0x0278, 0x1454: 0x0260, 0x1457: 0x024e, + 0x1459: 0x0266, 0x145b: 0x0272, 0x145d: 0x02e4, + 0x145f: 0x0296, 0x1461: 0x023c, 0x1462: 0x0248, + 0x1464: 0x0287, 0x1467: 0x024b, 0x1468: 0x0269, 0x1469: 0x0290, + 0x146a: 0x027b, 0x146c: 0x0281, 0x146d: 0x0284, 0x146e: 0x025d, 0x146f: 0x026f, + 0x1470: 0x0275, 0x1471: 0x0263, 0x1472: 0x0278, 0x1474: 0x0260, 0x1475: 0x0242, + 0x1476: 0x0245, 0x1477: 0x024e, 0x1479: 0x0266, 0x147a: 0x026c, 0x147b: 0x0272, + 0x147c: 0x0293, 0x147e: 0x02cc, + // Block 0x52, offset 0x1480 + 0x1480: 0x0239, 0x1481: 0x023c, 0x1482: 0x0248, 0x1483: 0x0251, 0x1484: 0x0287, 0x1485: 0x028a, + 0x1486: 0x025a, 0x1487: 0x024b, 0x1488: 0x0269, 0x1489: 0x0290, 0x148b: 0x027e, + 0x148c: 0x0281, 0x148d: 0x0284, 0x148e: 0x025d, 0x148f: 0x026f, 0x1490: 0x0275, 0x1491: 0x0263, + 0x1492: 0x0278, 0x1493: 0x0257, 0x1494: 0x0260, 0x1495: 0x0242, 0x1496: 0x0245, 0x1497: 0x024e, + 0x1498: 0x0254, 0x1499: 0x0266, 0x149a: 0x026c, 0x149b: 0x0272, + 0x14a1: 0x023c, 0x14a2: 0x0248, 0x14a3: 0x0251, + 0x14a5: 0x028a, 0x14a6: 0x025a, 0x14a7: 0x024b, 0x14a8: 0x0269, 0x14a9: 0x0290, + 0x14ab: 0x027e, 0x14ac: 0x0281, 0x14ad: 0x0284, 0x14ae: 0x025d, 0x14af: 0x026f, + 0x14b0: 0x0275, 0x14b1: 0x0263, 0x14b2: 0x0278, 0x14b3: 0x0257, 0x14b4: 0x0260, 0x14b5: 0x0242, + 0x14b6: 0x0245, 0x14b7: 0x024e, 0x14b8: 0x0254, 0x14b9: 0x0266, 0x14ba: 0x026c, 0x14bb: 0x0272, + // Block 0x53, offset 0x14c0 + 0x14c0: 0x1879, 0x14c1: 0x1876, 0x14c2: 0x187c, 0x14c3: 0x18a0, 0x14c4: 0x18c4, 0x14c5: 0x18e8, + 0x14c6: 0x190c, 0x14c7: 0x1915, 0x14c8: 0x191b, 0x14c9: 0x1921, 0x14ca: 0x1927, + 0x14d0: 0x1a8f, 0x14d1: 0x1a93, + 0x14d2: 0x1a97, 0x14d3: 0x1a9b, 0x14d4: 0x1a9f, 0x14d5: 0x1aa3, 0x14d6: 0x1aa7, 0x14d7: 0x1aab, + 0x14d8: 0x1aaf, 0x14d9: 0x1ab3, 0x14da: 0x1ab7, 0x14db: 0x1abb, 0x14dc: 0x1abf, 0x14dd: 0x1ac3, + 0x14de: 0x1ac7, 0x14df: 0x1acb, 0x14e0: 0x1acf, 0x14e1: 0x1ad3, 0x14e2: 0x1ad7, 0x14e3: 0x1adb, + 0x14e4: 0x1adf, 0x14e5: 0x1ae3, 0x14e6: 0x1ae7, 0x14e7: 0x1aeb, 0x14e8: 0x1aef, 0x14e9: 0x1af3, + 0x14ea: 0x2721, 0x14eb: 0x0047, 0x14ec: 0x0065, 0x14ed: 0x193c, 0x14ee: 0x19b4, + 0x14f0: 0x0043, 0x14f1: 0x0045, 0x14f2: 0x0047, 0x14f3: 0x0049, 0x14f4: 0x004b, 0x14f5: 0x004d, + 0x14f6: 0x004f, 0x14f7: 0x0051, 0x14f8: 0x0053, 0x14f9: 0x0055, 0x14fa: 0x0057, 0x14fb: 0x0059, + 0x14fc: 0x005b, 0x14fd: 0x005d, 0x14fe: 0x005f, 0x14ff: 0x0061, + // Block 0x54, offset 0x1500 + 0x1500: 0x26b0, 0x1501: 0x26c5, 0x1502: 0x0503, + 0x1510: 0x0c0f, 0x1511: 0x0a47, + 0x1512: 0x08d3, 0x1513: 0x45c7, 0x1514: 0x071b, 0x1515: 0x09ef, 0x1516: 0x132f, 0x1517: 0x09ff, + 0x1518: 0x0727, 0x1519: 0x0cd7, 0x151a: 0x0eaf, 0x151b: 0x0caf, 0x151c: 0x0827, 0x151d: 0x0b6b, + 0x151e: 0x07bf, 0x151f: 0x0cb7, 0x1520: 0x0813, 0x1521: 0x1117, 0x1522: 0x0f83, 0x1523: 0x138b, + 0x1524: 0x09d3, 0x1525: 0x090b, 0x1526: 0x0e63, 0x1527: 0x0c1b, 0x1528: 0x0c47, 0x1529: 0x06bf, + 0x152a: 0x06cb, 0x152b: 0x140b, 0x152c: 0x0adb, 0x152d: 0x06e7, 0x152e: 0x08ef, 0x152f: 0x0c3b, + 0x1530: 0x13b3, 0x1531: 0x0c13, 0x1532: 0x106f, 0x1533: 0x10ab, 0x1534: 0x08f7, 0x1535: 0x0e43, + 0x1536: 0x0d0b, 0x1537: 0x0d07, 0x1538: 0x0f97, 0x1539: 0x082b, 0x153a: 0x0957, 0x153b: 0x1443, + // Block 0x55, offset 0x1540 + 0x1540: 0x06fb, 0x1541: 0x06f3, 0x1542: 0x0703, 0x1543: 0x1647, 0x1544: 0x0747, 0x1545: 0x0757, + 0x1546: 0x075b, 0x1547: 0x0763, 0x1548: 0x076b, 0x1549: 0x076f, 0x154a: 0x077b, 0x154b: 0x0773, + 0x154c: 0x05b3, 0x154d: 0x165b, 0x154e: 0x078f, 0x154f: 0x0793, 0x1550: 0x0797, 0x1551: 0x07b3, + 0x1552: 0x164c, 0x1553: 0x05b7, 0x1554: 0x079f, 0x1555: 0x07bf, 0x1556: 0x1656, 0x1557: 0x07cf, + 0x1558: 0x07d7, 0x1559: 0x0737, 0x155a: 0x07df, 0x155b: 0x07e3, 0x155c: 0x1831, 0x155d: 0x07ff, + 0x155e: 0x0807, 0x155f: 0x05bf, 0x1560: 0x081f, 0x1561: 0x0823, 0x1562: 0x082b, 0x1563: 0x082f, + 0x1564: 0x05c3, 0x1565: 0x0847, 0x1566: 0x084b, 0x1567: 0x0857, 0x1568: 0x0863, 0x1569: 0x0867, + 0x156a: 0x086b, 0x156b: 0x0873, 0x156c: 0x0893, 0x156d: 0x0897, 0x156e: 0x089f, 0x156f: 0x08af, + 0x1570: 0x08b7, 0x1571: 0x08bb, 0x1572: 0x08bb, 0x1573: 0x08bb, 0x1574: 0x166a, 0x1575: 0x0e93, + 0x1576: 0x08cf, 0x1577: 0x08d7, 0x1578: 0x166f, 0x1579: 0x08e3, 0x157a: 0x08eb, 0x157b: 0x08f3, + 0x157c: 0x091b, 0x157d: 0x0907, 0x157e: 0x0913, 0x157f: 0x0917, + // Block 0x56, offset 0x1580 + 0x1580: 0x091f, 0x1581: 0x0927, 0x1582: 0x092b, 0x1583: 0x0933, 0x1584: 0x093b, 0x1585: 0x093f, + 0x1586: 0x093f, 0x1587: 0x0947, 0x1588: 0x094f, 0x1589: 0x0953, 0x158a: 0x095f, 0x158b: 0x0983, + 0x158c: 0x0967, 0x158d: 0x0987, 0x158e: 0x096b, 0x158f: 0x0973, 0x1590: 0x080b, 0x1591: 0x09cf, + 0x1592: 0x0997, 0x1593: 0x099b, 0x1594: 0x099f, 0x1595: 0x0993, 0x1596: 0x09a7, 0x1597: 0x09a3, + 0x1598: 0x09bb, 0x1599: 0x1674, 0x159a: 0x09d7, 0x159b: 0x09db, 0x159c: 0x09e3, 0x159d: 0x09ef, + 0x159e: 0x09f7, 0x159f: 0x0a13, 0x15a0: 0x1679, 0x15a1: 0x167e, 0x15a2: 0x0a1f, 0x15a3: 0x0a23, + 0x15a4: 0x0a27, 0x15a5: 0x0a1b, 0x15a6: 0x0a2f, 0x15a7: 0x05c7, 0x15a8: 0x05cb, 0x15a9: 0x0a37, + 0x15aa: 0x0a3f, 0x15ab: 0x0a3f, 0x15ac: 0x1683, 0x15ad: 0x0a5b, 0x15ae: 0x0a5f, 0x15af: 0x0a63, + 0x15b0: 0x0a6b, 0x15b1: 0x1688, 0x15b2: 0x0a73, 0x15b3: 0x0a77, 0x15b4: 0x0b4f, 0x15b5: 0x0a7f, + 0x15b6: 0x05cf, 0x15b7: 0x0a8b, 0x15b8: 0x0a9b, 0x15b9: 0x0aa7, 0x15ba: 0x0aa3, 0x15bb: 0x1692, + 0x15bc: 0x0aaf, 0x15bd: 0x1697, 0x15be: 0x0abb, 0x15bf: 0x0ab7, + // Block 0x57, offset 0x15c0 + 0x15c0: 0x0abf, 0x15c1: 0x0acf, 0x15c2: 0x0ad3, 0x15c3: 0x05d3, 0x15c4: 0x0ae3, 0x15c5: 0x0aeb, + 0x15c6: 0x0aef, 0x15c7: 0x0af3, 0x15c8: 0x05d7, 0x15c9: 0x169c, 0x15ca: 0x05db, 0x15cb: 0x0b0f, + 0x15cc: 0x0b13, 0x15cd: 0x0b17, 0x15ce: 0x0b1f, 0x15cf: 0x1863, 0x15d0: 0x0b37, 0x15d1: 0x16a6, + 0x15d2: 0x16a6, 0x15d3: 0x11d7, 0x15d4: 0x0b47, 0x15d5: 0x0b47, 0x15d6: 0x05df, 0x15d7: 0x16c9, + 0x15d8: 0x179b, 0x15d9: 0x0b57, 0x15da: 0x0b5f, 0x15db: 0x05e3, 0x15dc: 0x0b73, 0x15dd: 0x0b83, + 0x15de: 0x0b87, 0x15df: 0x0b8f, 0x15e0: 0x0b9f, 0x15e1: 0x05eb, 0x15e2: 0x05e7, 0x15e3: 0x0ba3, + 0x15e4: 0x16ab, 0x15e5: 0x0ba7, 0x15e6: 0x0bbb, 0x15e7: 0x0bbf, 0x15e8: 0x0bc3, 0x15e9: 0x0bbf, + 0x15ea: 0x0bcf, 0x15eb: 0x0bd3, 0x15ec: 0x0be3, 0x15ed: 0x0bdb, 0x15ee: 0x0bdf, 0x15ef: 0x0be7, + 0x15f0: 0x0beb, 0x15f1: 0x0bef, 0x15f2: 0x0bfb, 0x15f3: 0x0bff, 0x15f4: 0x0c17, 0x15f5: 0x0c1f, + 0x15f6: 0x0c2f, 0x15f7: 0x0c43, 0x15f8: 0x16ba, 0x15f9: 0x0c3f, 0x15fa: 0x0c33, 0x15fb: 0x0c4b, + 0x15fc: 0x0c53, 0x15fd: 0x0c67, 0x15fe: 0x16bf, 0x15ff: 0x0c6f, + // Block 0x58, offset 0x1600 + 0x1600: 0x0c63, 0x1601: 0x0c5b, 0x1602: 0x05ef, 0x1603: 0x0c77, 0x1604: 0x0c7f, 0x1605: 0x0c87, + 0x1606: 0x0c7b, 0x1607: 0x05f3, 0x1608: 0x0c97, 0x1609: 0x0c9f, 0x160a: 0x16c4, 0x160b: 0x0ccb, + 0x160c: 0x0cff, 0x160d: 0x0cdb, 0x160e: 0x05ff, 0x160f: 0x0ce7, 0x1610: 0x05fb, 0x1611: 0x05f7, + 0x1612: 0x07c3, 0x1613: 0x07c7, 0x1614: 0x0d03, 0x1615: 0x0ceb, 0x1616: 0x11ab, 0x1617: 0x0663, + 0x1618: 0x0d0f, 0x1619: 0x0d13, 0x161a: 0x0d17, 0x161b: 0x0d2b, 0x161c: 0x0d23, 0x161d: 0x16dd, + 0x161e: 0x0603, 0x161f: 0x0d3f, 0x1620: 0x0d33, 0x1621: 0x0d4f, 0x1622: 0x0d57, 0x1623: 0x16e7, + 0x1624: 0x0d5b, 0x1625: 0x0d47, 0x1626: 0x0d63, 0x1627: 0x0607, 0x1628: 0x0d67, 0x1629: 0x0d6b, + 0x162a: 0x0d6f, 0x162b: 0x0d7b, 0x162c: 0x16ec, 0x162d: 0x0d83, 0x162e: 0x060b, 0x162f: 0x0d8f, + 0x1630: 0x16f1, 0x1631: 0x0d93, 0x1632: 0x060f, 0x1633: 0x0d9f, 0x1634: 0x0dab, 0x1635: 0x0db7, + 0x1636: 0x0dbb, 0x1637: 0x16f6, 0x1638: 0x168d, 0x1639: 0x16fb, 0x163a: 0x0ddb, 0x163b: 0x1700, + 0x163c: 0x0de7, 0x163d: 0x0def, 0x163e: 0x0ddf, 0x163f: 0x0dfb, + // Block 0x59, offset 0x1640 + 0x1640: 0x0e0b, 0x1641: 0x0e1b, 0x1642: 0x0e0f, 0x1643: 0x0e13, 0x1644: 0x0e1f, 0x1645: 0x0e23, + 0x1646: 0x1705, 0x1647: 0x0e07, 0x1648: 0x0e3b, 0x1649: 0x0e3f, 0x164a: 0x0613, 0x164b: 0x0e53, + 0x164c: 0x0e4f, 0x164d: 0x170a, 0x164e: 0x0e33, 0x164f: 0x0e6f, 0x1650: 0x170f, 0x1651: 0x1714, + 0x1652: 0x0e73, 0x1653: 0x0e87, 0x1654: 0x0e83, 0x1655: 0x0e7f, 0x1656: 0x0617, 0x1657: 0x0e8b, + 0x1658: 0x0e9b, 0x1659: 0x0e97, 0x165a: 0x0ea3, 0x165b: 0x1651, 0x165c: 0x0eb3, 0x165d: 0x1719, + 0x165e: 0x0ebf, 0x165f: 0x1723, 0x1660: 0x0ed3, 0x1661: 0x0edf, 0x1662: 0x0ef3, 0x1663: 0x1728, + 0x1664: 0x0f07, 0x1665: 0x0f0b, 0x1666: 0x172d, 0x1667: 0x1732, 0x1668: 0x0f27, 0x1669: 0x0f37, + 0x166a: 0x061b, 0x166b: 0x0f3b, 0x166c: 0x061f, 0x166d: 0x061f, 0x166e: 0x0f53, 0x166f: 0x0f57, + 0x1670: 0x0f5f, 0x1671: 0x0f63, 0x1672: 0x0f6f, 0x1673: 0x0623, 0x1674: 0x0f87, 0x1675: 0x1737, + 0x1676: 0x0fa3, 0x1677: 0x173c, 0x1678: 0x0faf, 0x1679: 0x16a1, 0x167a: 0x0fbf, 0x167b: 0x1741, + 0x167c: 0x1746, 0x167d: 0x174b, 0x167e: 0x0627, 0x167f: 0x062b, + // Block 0x5a, offset 0x1680 + 0x1680: 0x0ff7, 0x1681: 0x1755, 0x1682: 0x1750, 0x1683: 0x175a, 0x1684: 0x175f, 0x1685: 0x0fff, + 0x1686: 0x1003, 0x1687: 0x1003, 0x1688: 0x100b, 0x1689: 0x0633, 0x168a: 0x100f, 0x168b: 0x0637, + 0x168c: 0x063b, 0x168d: 0x1769, 0x168e: 0x1023, 0x168f: 0x102b, 0x1690: 0x1037, 0x1691: 0x063f, + 0x1692: 0x176e, 0x1693: 0x105b, 0x1694: 0x1773, 0x1695: 0x1778, 0x1696: 0x107b, 0x1697: 0x1093, + 0x1698: 0x0643, 0x1699: 0x109b, 0x169a: 0x109f, 0x169b: 0x10a3, 0x169c: 0x177d, 0x169d: 0x1782, + 0x169e: 0x1782, 0x169f: 0x10bb, 0x16a0: 0x0647, 0x16a1: 0x1787, 0x16a2: 0x10cf, 0x16a3: 0x10d3, + 0x16a4: 0x064b, 0x16a5: 0x178c, 0x16a6: 0x10ef, 0x16a7: 0x064f, 0x16a8: 0x10ff, 0x16a9: 0x10f7, + 0x16aa: 0x1107, 0x16ab: 0x1796, 0x16ac: 0x111f, 0x16ad: 0x0653, 0x16ae: 0x112b, 0x16af: 0x1133, + 0x16b0: 0x1143, 0x16b1: 0x0657, 0x16b2: 0x17a0, 0x16b3: 0x17a5, 0x16b4: 0x065b, 0x16b5: 0x17aa, + 0x16b6: 0x115b, 0x16b7: 0x17af, 0x16b8: 0x1167, 0x16b9: 0x1173, 0x16ba: 0x117b, 0x16bb: 0x17b4, + 0x16bc: 0x17b9, 0x16bd: 0x118f, 0x16be: 0x17be, 0x16bf: 0x1197, + // Block 0x5b, offset 0x16c0 + 0x16c0: 0x16ce, 0x16c1: 0x065f, 0x16c2: 0x11af, 0x16c3: 0x11b3, 0x16c4: 0x0667, 0x16c5: 0x11b7, + 0x16c6: 0x0a33, 0x16c7: 0x17c3, 0x16c8: 0x17c8, 0x16c9: 0x16d3, 0x16ca: 0x16d8, 0x16cb: 0x11d7, + 0x16cc: 0x11db, 0x16cd: 0x13f3, 0x16ce: 0x066b, 0x16cf: 0x1207, 0x16d0: 0x1203, 0x16d1: 0x120b, + 0x16d2: 0x083f, 0x16d3: 0x120f, 0x16d4: 0x1213, 0x16d5: 0x1217, 0x16d6: 0x121f, 0x16d7: 0x17cd, + 0x16d8: 0x121b, 0x16d9: 0x1223, 0x16da: 0x1237, 0x16db: 0x123b, 0x16dc: 0x1227, 0x16dd: 0x123f, + 0x16de: 0x1253, 0x16df: 0x1267, 0x16e0: 0x1233, 0x16e1: 0x1247, 0x16e2: 0x124b, 0x16e3: 0x124f, + 0x16e4: 0x17d2, 0x16e5: 0x17dc, 0x16e6: 0x17d7, 0x16e7: 0x066f, 0x16e8: 0x126f, 0x16e9: 0x1273, + 0x16ea: 0x127b, 0x16eb: 0x17f0, 0x16ec: 0x127f, 0x16ed: 0x17e1, 0x16ee: 0x0673, 0x16ef: 0x0677, + 0x16f0: 0x17e6, 0x16f1: 0x17eb, 0x16f2: 0x067b, 0x16f3: 0x129f, 0x16f4: 0x12a3, 0x16f5: 0x12a7, + 0x16f6: 0x12ab, 0x16f7: 0x12b7, 0x16f8: 0x12b3, 0x16f9: 0x12bf, 0x16fa: 0x12bb, 0x16fb: 0x12cb, + 0x16fc: 0x12c3, 0x16fd: 0x12c7, 0x16fe: 0x12cf, 0x16ff: 0x067f, + // Block 0x5c, offset 0x1700 + 0x1700: 0x12d7, 0x1701: 0x12db, 0x1702: 0x0683, 0x1703: 0x12eb, 0x1704: 0x12ef, 0x1705: 0x17f5, + 0x1706: 0x12fb, 0x1707: 0x12ff, 0x1708: 0x0687, 0x1709: 0x130b, 0x170a: 0x05bb, 0x170b: 0x17fa, + 0x170c: 0x17ff, 0x170d: 0x068b, 0x170e: 0x068f, 0x170f: 0x1337, 0x1710: 0x134f, 0x1711: 0x136b, + 0x1712: 0x137b, 0x1713: 0x1804, 0x1714: 0x138f, 0x1715: 0x1393, 0x1716: 0x13ab, 0x1717: 0x13b7, + 0x1718: 0x180e, 0x1719: 0x1660, 0x171a: 0x13c3, 0x171b: 0x13bf, 0x171c: 0x13cb, 0x171d: 0x1665, + 0x171e: 0x13d7, 0x171f: 0x13e3, 0x1720: 0x1813, 0x1721: 0x1818, 0x1722: 0x1423, 0x1723: 0x142f, + 0x1724: 0x1437, 0x1725: 0x181d, 0x1726: 0x143b, 0x1727: 0x1467, 0x1728: 0x1473, 0x1729: 0x1477, + 0x172a: 0x146f, 0x172b: 0x1483, 0x172c: 0x1487, 0x172d: 0x1822, 0x172e: 0x1493, 0x172f: 0x0693, + 0x1730: 0x149b, 0x1731: 0x1827, 0x1732: 0x0697, 0x1733: 0x14d3, 0x1734: 0x0ac3, 0x1735: 0x14eb, + 0x1736: 0x182c, 0x1737: 0x1836, 0x1738: 0x069b, 0x1739: 0x069f, 0x173a: 0x1513, 0x173b: 0x183b, + 0x173c: 0x06a3, 0x173d: 0x1840, 0x173e: 0x152b, 0x173f: 0x152b, + // Block 0x5d, offset 0x1740 + 0x1740: 0x1533, 0x1741: 0x1845, 0x1742: 0x154b, 0x1743: 0x06a7, 0x1744: 0x155b, 0x1745: 0x1567, + 0x1746: 0x156f, 0x1747: 0x1577, 0x1748: 0x06ab, 0x1749: 0x184a, 0x174a: 0x158b, 0x174b: 0x15a7, + 0x174c: 0x15b3, 0x174d: 0x06af, 0x174e: 0x06b3, 0x174f: 0x15b7, 0x1750: 0x184f, 0x1751: 0x06b7, + 0x1752: 0x1854, 0x1753: 0x1859, 0x1754: 0x185e, 0x1755: 0x15db, 0x1756: 0x06bb, 0x1757: 0x15ef, + 0x1758: 0x15f7, 0x1759: 0x15fb, 0x175a: 0x1603, 0x175b: 0x160b, 0x175c: 0x1613, 0x175d: 0x1868, +} + +// nfkcIndex: 22 blocks, 1408 entries, 2816 bytes +// Block 0 is the zero block. +var nfkcIndex = [1408]uint16{ + // Block 0x0, offset 0x0 + // Block 0x1, offset 0x40 + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xc2: 0x5c, 0xc3: 0x01, 0xc4: 0x02, 0xc5: 0x03, 0xc6: 0x5d, 0xc7: 0x04, + 0xc8: 0x05, 0xca: 0x5e, 0xcb: 0x5f, 0xcc: 0x06, 0xcd: 0x07, 0xce: 0x08, 0xcf: 0x09, + 0xd0: 0x0a, 0xd1: 0x60, 0xd2: 0x61, 0xd3: 0x0b, 0xd6: 0x0c, 0xd7: 0x62, + 0xd8: 0x63, 0xd9: 0x0d, 0xdb: 0x64, 0xdc: 0x65, 0xdd: 0x66, 0xdf: 0x67, + 0xe0: 0x02, 0xe1: 0x03, 0xe2: 0x04, 0xe3: 0x05, + 0xea: 0x06, 0xeb: 0x07, 0xec: 0x08, 0xed: 0x09, 0xef: 0x0a, + 0xf0: 0x13, + // Block 0x4, offset 0x100 + 0x120: 0x68, 0x121: 0x69, 0x123: 0x0e, 0x124: 0x6a, 0x125: 0x6b, 0x126: 0x6c, 0x127: 0x6d, + 0x128: 0x6e, 0x129: 0x6f, 0x12a: 0x70, 0x12b: 0x71, 0x12c: 0x6c, 0x12d: 0x72, 0x12e: 0x73, 0x12f: 0x74, + 0x131: 0x75, 0x132: 0x76, 0x133: 0x77, 0x134: 0x78, 0x135: 0x79, 0x137: 0x7a, + 0x138: 0x7b, 0x139: 0x7c, 0x13a: 0x7d, 0x13b: 0x7e, 0x13c: 0x7f, 0x13d: 0x80, 0x13e: 0x81, 0x13f: 0x82, + // Block 0x5, offset 0x140 + 0x140: 0x83, 0x142: 0x84, 0x143: 0x85, 0x144: 0x86, 0x145: 0x87, 0x146: 0x88, 0x147: 0x89, + 0x14d: 0x8a, + 0x15c: 0x8b, 0x15f: 0x8c, + 0x162: 0x8d, 0x164: 0x8e, + 0x168: 0x8f, 0x169: 0x90, 0x16a: 0x91, 0x16c: 0x0f, 0x16d: 0x92, 0x16e: 0x93, 0x16f: 0x94, + 0x170: 0x95, 0x173: 0x96, 0x174: 0x97, 0x175: 0x10, 0x176: 0x11, 0x177: 0x12, + 0x178: 0x13, 0x179: 0x14, 0x17a: 0x15, 0x17b: 0x16, 0x17c: 0x17, 0x17d: 0x18, 0x17e: 0x19, 0x17f: 0x1a, + // Block 0x6, offset 0x180 + 0x180: 0x98, 0x181: 0x99, 0x182: 0x9a, 0x183: 0x9b, 0x184: 0x1b, 0x185: 0x1c, 0x186: 0x9c, 0x187: 0x9d, + 0x188: 0x9e, 0x189: 0x1d, 0x18a: 0x1e, 0x18b: 0x9f, 0x18c: 0xa0, + 0x191: 0x1f, 0x192: 0x20, 0x193: 0xa1, + 0x1a8: 0xa2, 0x1a9: 0xa3, 0x1ab: 0xa4, + 0x1b1: 0xa5, 0x1b3: 0xa6, 0x1b5: 0xa7, 0x1b7: 0xa8, + 0x1ba: 0xa9, 0x1bb: 0xaa, 0x1bc: 0x21, 0x1bd: 0x22, 0x1be: 0x23, 0x1bf: 0xab, + // Block 0x7, offset 0x1c0 + 0x1c0: 0xac, 0x1c1: 0x24, 0x1c2: 0x25, 0x1c3: 0x26, 0x1c4: 0xad, 0x1c5: 0x27, 0x1c6: 0x28, + 0x1c8: 0x29, 0x1c9: 0x2a, 0x1ca: 0x2b, 0x1cb: 0x2c, 0x1cc: 0x2d, 0x1cd: 0x2e, 0x1ce: 0x2f, 0x1cf: 0x30, + // Block 0x8, offset 0x200 + 0x219: 0xae, 0x21a: 0xaf, 0x21b: 0xb0, 0x21d: 0xb1, 0x21f: 0xb2, + 0x220: 0xb3, 0x223: 0xb4, 0x224: 0xb5, 0x225: 0xb6, 0x226: 0xb7, 0x227: 0xb8, + 0x22a: 0xb9, 0x22b: 0xba, 0x22d: 0xbb, 0x22f: 0xbc, + 0x230: 0xbd, 0x231: 0xbe, 0x232: 0xbf, 0x233: 0xc0, 0x234: 0xc1, 0x235: 0xc2, 0x236: 0xc3, 0x237: 0xbd, + 0x238: 0xbe, 0x239: 0xbf, 0x23a: 0xc0, 0x23b: 0xc1, 0x23c: 0xc2, 0x23d: 0xc3, 0x23e: 0xbd, 0x23f: 0xbe, + // Block 0x9, offset 0x240 + 0x240: 0xbf, 0x241: 0xc0, 0x242: 0xc1, 0x243: 0xc2, 0x244: 0xc3, 0x245: 0xbd, 0x246: 0xbe, 0x247: 0xbf, + 0x248: 0xc0, 0x249: 0xc1, 0x24a: 0xc2, 0x24b: 0xc3, 0x24c: 0xbd, 0x24d: 0xbe, 0x24e: 0xbf, 0x24f: 0xc0, + 0x250: 0xc1, 0x251: 0xc2, 0x252: 0xc3, 0x253: 0xbd, 0x254: 0xbe, 0x255: 0xbf, 0x256: 0xc0, 0x257: 0xc1, + 0x258: 0xc2, 0x259: 0xc3, 0x25a: 0xbd, 0x25b: 0xbe, 0x25c: 0xbf, 0x25d: 0xc0, 0x25e: 0xc1, 0x25f: 0xc2, + 0x260: 0xc3, 0x261: 0xbd, 0x262: 0xbe, 0x263: 0xbf, 0x264: 0xc0, 0x265: 0xc1, 0x266: 0xc2, 0x267: 0xc3, + 0x268: 0xbd, 0x269: 0xbe, 0x26a: 0xbf, 0x26b: 0xc0, 0x26c: 0xc1, 0x26d: 0xc2, 0x26e: 0xc3, 0x26f: 0xbd, + 0x270: 0xbe, 0x271: 0xbf, 0x272: 0xc0, 0x273: 0xc1, 0x274: 0xc2, 0x275: 0xc3, 0x276: 0xbd, 0x277: 0xbe, + 0x278: 0xbf, 0x279: 0xc0, 0x27a: 0xc1, 0x27b: 0xc2, 0x27c: 0xc3, 0x27d: 0xbd, 0x27e: 0xbe, 0x27f: 0xbf, + // Block 0xa, offset 0x280 + 0x280: 0xc0, 0x281: 0xc1, 0x282: 0xc2, 0x283: 0xc3, 0x284: 0xbd, 0x285: 0xbe, 0x286: 0xbf, 0x287: 0xc0, + 0x288: 0xc1, 0x289: 0xc2, 0x28a: 0xc3, 0x28b: 0xbd, 0x28c: 0xbe, 0x28d: 0xbf, 0x28e: 0xc0, 0x28f: 0xc1, + 0x290: 0xc2, 0x291: 0xc3, 0x292: 0xbd, 0x293: 0xbe, 0x294: 0xbf, 0x295: 0xc0, 0x296: 0xc1, 0x297: 0xc2, + 0x298: 0xc3, 0x299: 0xbd, 0x29a: 0xbe, 0x29b: 0xbf, 0x29c: 0xc0, 0x29d: 0xc1, 0x29e: 0xc2, 0x29f: 0xc3, + 0x2a0: 0xbd, 0x2a1: 0xbe, 0x2a2: 0xbf, 0x2a3: 0xc0, 0x2a4: 0xc1, 0x2a5: 0xc2, 0x2a6: 0xc3, 0x2a7: 0xbd, + 0x2a8: 0xbe, 0x2a9: 0xbf, 0x2aa: 0xc0, 0x2ab: 0xc1, 0x2ac: 0xc2, 0x2ad: 0xc3, 0x2ae: 0xbd, 0x2af: 0xbe, + 0x2b0: 0xbf, 0x2b1: 0xc0, 0x2b2: 0xc1, 0x2b3: 0xc2, 0x2b4: 0xc3, 0x2b5: 0xbd, 0x2b6: 0xbe, 0x2b7: 0xbf, + 0x2b8: 0xc0, 0x2b9: 0xc1, 0x2ba: 0xc2, 0x2bb: 0xc3, 0x2bc: 0xbd, 0x2bd: 0xbe, 0x2be: 0xbf, 0x2bf: 0xc0, + // Block 0xb, offset 0x2c0 + 0x2c0: 0xc1, 0x2c1: 0xc2, 0x2c2: 0xc3, 0x2c3: 0xbd, 0x2c4: 0xbe, 0x2c5: 0xbf, 0x2c6: 0xc0, 0x2c7: 0xc1, + 0x2c8: 0xc2, 0x2c9: 0xc3, 0x2ca: 0xbd, 0x2cb: 0xbe, 0x2cc: 0xbf, 0x2cd: 0xc0, 0x2ce: 0xc1, 0x2cf: 0xc2, + 0x2d0: 0xc3, 0x2d1: 0xbd, 0x2d2: 0xbe, 0x2d3: 0xbf, 0x2d4: 0xc0, 0x2d5: 0xc1, 0x2d6: 0xc2, 0x2d7: 0xc3, + 0x2d8: 0xbd, 0x2d9: 0xbe, 0x2da: 0xbf, 0x2db: 0xc0, 0x2dc: 0xc1, 0x2dd: 0xc2, 0x2de: 0xc4, + // Block 0xc, offset 0x300 + 0x324: 0x31, 0x325: 0x32, 0x326: 0x33, 0x327: 0x34, + 0x328: 0x35, 0x329: 0x36, 0x32a: 0x37, 0x32b: 0x38, 0x32c: 0x39, 0x32d: 0x3a, 0x32e: 0x3b, 0x32f: 0x3c, + 0x330: 0x3d, 0x331: 0x3e, 0x332: 0x3f, 0x333: 0x40, 0x334: 0x41, 0x335: 0x42, 0x336: 0x43, 0x337: 0x44, + 0x338: 0x45, 0x339: 0x46, 0x33a: 0x47, 0x33b: 0x48, 0x33c: 0xc5, 0x33d: 0x49, 0x33e: 0x4a, 0x33f: 0x4b, + // Block 0xd, offset 0x340 + 0x347: 0xc6, + 0x34b: 0xc7, 0x34d: 0xc8, + 0x368: 0xc9, 0x36b: 0xca, + 0x374: 0xcb, + 0x37d: 0xcc, + // Block 0xe, offset 0x380 + 0x381: 0xcd, 0x382: 0xce, 0x384: 0xcf, 0x385: 0xb7, 0x387: 0xd0, + 0x388: 0xd1, 0x38b: 0xd2, 0x38c: 0xd3, 0x38d: 0xd4, + 0x391: 0xd5, 0x392: 0xd6, 0x393: 0xd7, 0x396: 0xd8, 0x397: 0xd9, + 0x398: 0xda, 0x39a: 0xdb, 0x39c: 0xdc, + 0x3a0: 0xdd, 0x3a7: 0xde, + 0x3a8: 0xdf, 0x3a9: 0xe0, 0x3aa: 0xe1, + 0x3b0: 0xda, 0x3b5: 0xe2, 0x3b6: 0xe3, + // Block 0xf, offset 0x3c0 + 0x3eb: 0xe4, 0x3ec: 0xe5, + // Block 0x10, offset 0x400 + 0x432: 0xe6, + // Block 0x11, offset 0x440 + 0x445: 0xe7, 0x446: 0xe8, 0x447: 0xe9, + 0x449: 0xea, + 0x450: 0xeb, 0x451: 0xec, 0x452: 0xed, 0x453: 0xee, 0x454: 0xef, 0x455: 0xf0, 0x456: 0xf1, 0x457: 0xf2, + 0x458: 0xf3, 0x459: 0xf4, 0x45a: 0x4c, 0x45b: 0xf5, 0x45c: 0xf6, 0x45d: 0xf7, 0x45e: 0xf8, 0x45f: 0x4d, + // Block 0x12, offset 0x480 + 0x480: 0xf9, 0x484: 0xe5, + 0x48b: 0xfa, + 0x4a3: 0xfb, 0x4a5: 0xfc, + 0x4b8: 0x4e, 0x4b9: 0x4f, 0x4ba: 0x50, + // Block 0x13, offset 0x4c0 + 0x4c4: 0x51, 0x4c5: 0xfd, 0x4c6: 0xfe, + 0x4c8: 0x52, 0x4c9: 0xff, + // Block 0x14, offset 0x500 + 0x520: 0x53, 0x521: 0x54, 0x522: 0x55, 0x523: 0x56, 0x524: 0x57, 0x525: 0x58, 0x526: 0x59, 0x527: 0x5a, + 0x528: 0x5b, + // Block 0x15, offset 0x540 + 0x550: 0x0b, 0x551: 0x0c, 0x556: 0x0d, + 0x55b: 0x0e, 0x55d: 0x0f, 0x55e: 0x10, 0x55f: 0x11, + 0x56f: 0x12, +} + +// nfkcSparseOffset: 164 entries, 328 bytes +var nfkcSparseOffset = []uint16{0x0, 0xe, 0x12, 0x1b, 0x25, 0x35, 0x37, 0x3c, 0x47, 0x56, 0x63, 0x6b, 0x70, 0x75, 0x77, 0x7f, 0x86, 0x89, 0x91, 0x95, 0x99, 0x9b, 0x9d, 0xa6, 0xaa, 0xb1, 0xb6, 0xb9, 0xc3, 0xc6, 0xcd, 0xd5, 0xd9, 0xdb, 0xdf, 0xe3, 0xe9, 0xfa, 0x106, 0x108, 0x10e, 0x110, 0x112, 0x114, 0x116, 0x118, 0x11a, 0x11c, 0x11f, 0x122, 0x124, 0x127, 0x12a, 0x12e, 0x133, 0x13c, 0x13e, 0x141, 0x143, 0x14e, 0x159, 0x167, 0x175, 0x185, 0x193, 0x19a, 0x1a0, 0x1af, 0x1b3, 0x1b5, 0x1b9, 0x1bb, 0x1be, 0x1c0, 0x1c3, 0x1c5, 0x1c8, 0x1ca, 0x1cc, 0x1ce, 0x1da, 0x1e4, 0x1ee, 0x1f1, 0x1f5, 0x1f7, 0x1f9, 0x1fb, 0x1fd, 0x200, 0x202, 0x204, 0x206, 0x208, 0x20e, 0x211, 0x215, 0x217, 0x21e, 0x224, 0x22a, 0x232, 0x238, 0x23e, 0x244, 0x248, 0x24a, 0x24c, 0x24e, 0x250, 0x256, 0x259, 0x25b, 0x261, 0x264, 0x26c, 0x273, 0x276, 0x279, 0x27b, 0x27e, 0x286, 0x28a, 0x291, 0x294, 0x29a, 0x29c, 0x29e, 0x2a1, 0x2a3, 0x2a6, 0x2a8, 0x2aa, 0x2ac, 0x2ae, 0x2b1, 0x2b3, 0x2b5, 0x2b7, 0x2b9, 0x2c6, 0x2d0, 0x2d2, 0x2d4, 0x2d8, 0x2dd, 0x2e9, 0x2ee, 0x2f7, 0x2fd, 0x302, 0x306, 0x30b, 0x30f, 0x31f, 0x32d, 0x33b, 0x349, 0x34f, 0x351, 0x353, 0x356, 0x361, 0x363} + +// nfkcSparseValues: 877 entries, 3508 bytes +var nfkcSparseValues = [877]valueRange{ + // Block 0x0, offset 0x0 + {value: 0x0002, lo: 0x0d}, + {value: 0x0001, lo: 0xa0, hi: 0xa0}, + {value: 0x427b, lo: 0xa8, hi: 0xa8}, + {value: 0x0083, lo: 0xaa, hi: 0xaa}, + {value: 0x4267, lo: 0xaf, hi: 0xaf}, + {value: 0x0025, lo: 0xb2, hi: 0xb3}, + {value: 0x425d, lo: 0xb4, hi: 0xb4}, + {value: 0x01dc, lo: 0xb5, hi: 0xb5}, + {value: 0x4294, lo: 0xb8, hi: 0xb8}, + {value: 0x0023, lo: 0xb9, hi: 0xb9}, + {value: 0x009f, lo: 0xba, hi: 0xba}, + {value: 0x221f, lo: 0xbc, hi: 0xbc}, + {value: 0x2213, lo: 0xbd, hi: 0xbd}, + {value: 0x22b5, lo: 0xbe, hi: 0xbe}, + // Block 0x1, offset 0xe + {value: 0x0091, lo: 0x03}, + {value: 0x46e5, lo: 0xa0, hi: 0xa1}, + {value: 0x4717, lo: 0xaf, hi: 0xb0}, + {value: 0xa000, lo: 0xb7, hi: 0xb7}, + // Block 0x2, offset 0x12 + {value: 0x0003, lo: 0x08}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0x0091, lo: 0xb0, hi: 0xb0}, + {value: 0x0119, lo: 0xb1, hi: 0xb1}, + {value: 0x0095, lo: 0xb2, hi: 0xb2}, + {value: 0x00a5, lo: 0xb3, hi: 0xb3}, + {value: 0x0143, lo: 0xb4, hi: 0xb6}, + {value: 0x00af, lo: 0xb7, hi: 0xb7}, + {value: 0x00b3, lo: 0xb8, hi: 0xb8}, + // Block 0x3, offset 0x1b + {value: 0x000a, lo: 0x09}, + {value: 0x4271, lo: 0x98, hi: 0x98}, + {value: 0x4276, lo: 0x99, hi: 0x9a}, + {value: 0x4299, lo: 0x9b, hi: 0x9b}, + {value: 0x4262, lo: 0x9c, hi: 0x9c}, + {value: 0x4285, lo: 0x9d, hi: 0x9d}, + {value: 0x0113, lo: 0xa0, hi: 0xa0}, + {value: 0x0099, lo: 0xa1, hi: 0xa1}, + {value: 0x00a7, lo: 0xa2, hi: 0xa3}, + {value: 0x0167, lo: 0xa4, hi: 0xa4}, + // Block 0x4, offset 0x25 + {value: 0x0000, lo: 0x0f}, + {value: 0xa000, lo: 0x83, hi: 0x83}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0xa000, lo: 0x8b, hi: 0x8b}, + {value: 0xa000, lo: 0x8d, hi: 0x8d}, + {value: 0x37a8, lo: 0x90, hi: 0x90}, + {value: 0x37b4, lo: 0x91, hi: 0x91}, + {value: 0x37a2, lo: 0x93, hi: 0x93}, + {value: 0xa000, lo: 0x96, hi: 0x96}, + {value: 0x381a, lo: 0x97, hi: 0x97}, + {value: 0x37e4, lo: 0x9c, hi: 0x9c}, + {value: 0x37cc, lo: 0x9d, hi: 0x9d}, + {value: 0x37f6, lo: 0x9e, hi: 0x9e}, + {value: 0xa000, lo: 0xb4, hi: 0xb5}, + {value: 0x3820, lo: 0xb6, hi: 0xb6}, + {value: 0x3826, lo: 0xb7, hi: 0xb7}, + // Block 0x5, offset 0x35 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0x83, hi: 0x87}, + // Block 0x6, offset 0x37 + {value: 0x0001, lo: 0x04}, + {value: 0x8113, lo: 0x81, hi: 0x82}, + {value: 0x8132, lo: 0x84, hi: 0x84}, + {value: 0x812d, lo: 0x85, hi: 0x85}, + {value: 0x810d, lo: 0x87, hi: 0x87}, + // Block 0x7, offset 0x3c + {value: 0x0000, lo: 0x0a}, + {value: 0x8132, lo: 0x90, hi: 0x97}, + {value: 0x8119, lo: 0x98, hi: 0x98}, + {value: 0x811a, lo: 0x99, hi: 0x99}, + {value: 0x811b, lo: 0x9a, hi: 0x9a}, + {value: 0x3844, lo: 0xa2, hi: 0xa2}, + {value: 0x384a, lo: 0xa3, hi: 0xa3}, + {value: 0x3856, lo: 0xa4, hi: 0xa4}, + {value: 0x3850, lo: 0xa5, hi: 0xa5}, + {value: 0x385c, lo: 0xa6, hi: 0xa6}, + {value: 0xa000, lo: 0xa7, hi: 0xa7}, + // Block 0x8, offset 0x47 + {value: 0x0000, lo: 0x0e}, + {value: 0x386e, lo: 0x80, hi: 0x80}, + {value: 0xa000, lo: 0x81, hi: 0x81}, + {value: 0x3862, lo: 0x82, hi: 0x82}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0x3868, lo: 0x93, hi: 0x93}, + {value: 0xa000, lo: 0x95, hi: 0x95}, + {value: 0x8132, lo: 0x96, hi: 0x9c}, + {value: 0x8132, lo: 0x9f, hi: 0xa2}, + {value: 0x812d, lo: 0xa3, hi: 0xa3}, + {value: 0x8132, lo: 0xa4, hi: 0xa4}, + {value: 0x8132, lo: 0xa7, hi: 0xa8}, + {value: 0x812d, lo: 0xaa, hi: 0xaa}, + {value: 0x8132, lo: 0xab, hi: 0xac}, + {value: 0x812d, lo: 0xad, hi: 0xad}, + // Block 0x9, offset 0x56 + {value: 0x0000, lo: 0x0c}, + {value: 0x811f, lo: 0x91, hi: 0x91}, + {value: 0x8132, lo: 0xb0, hi: 0xb0}, + {value: 0x812d, lo: 0xb1, hi: 0xb1}, + {value: 0x8132, lo: 0xb2, hi: 0xb3}, + {value: 0x812d, lo: 0xb4, hi: 0xb4}, + {value: 0x8132, lo: 0xb5, hi: 0xb6}, + {value: 0x812d, lo: 0xb7, hi: 0xb9}, + {value: 0x8132, lo: 0xba, hi: 0xba}, + {value: 0x812d, lo: 0xbb, hi: 0xbc}, + {value: 0x8132, lo: 0xbd, hi: 0xbd}, + {value: 0x812d, lo: 0xbe, hi: 0xbe}, + {value: 0x8132, lo: 0xbf, hi: 0xbf}, + // Block 0xa, offset 0x63 + {value: 0x0005, lo: 0x07}, + {value: 0x8132, lo: 0x80, hi: 0x80}, + {value: 0x8132, lo: 0x81, hi: 0x81}, + {value: 0x812d, lo: 0x82, hi: 0x83}, + {value: 0x812d, lo: 0x84, hi: 0x85}, + {value: 0x812d, lo: 0x86, hi: 0x87}, + {value: 0x812d, lo: 0x88, hi: 0x89}, + {value: 0x8132, lo: 0x8a, hi: 0x8a}, + // Block 0xb, offset 0x6b + {value: 0x0000, lo: 0x04}, + {value: 0x8132, lo: 0xab, hi: 0xb1}, + {value: 0x812d, lo: 0xb2, hi: 0xb2}, + {value: 0x8132, lo: 0xb3, hi: 0xb3}, + {value: 0x812d, lo: 0xbd, hi: 0xbd}, + // Block 0xc, offset 0x70 + {value: 0x0000, lo: 0x04}, + {value: 0x8132, lo: 0x96, hi: 0x99}, + {value: 0x8132, lo: 0x9b, hi: 0xa3}, + {value: 0x8132, lo: 0xa5, hi: 0xa7}, + {value: 0x8132, lo: 0xa9, hi: 0xad}, + // Block 0xd, offset 0x75 + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x99, hi: 0x9b}, + // Block 0xe, offset 0x77 + {value: 0x0000, lo: 0x07}, + {value: 0xa000, lo: 0xa8, hi: 0xa8}, + {value: 0x3edb, lo: 0xa9, hi: 0xa9}, + {value: 0xa000, lo: 0xb0, hi: 0xb0}, + {value: 0x3ee3, lo: 0xb1, hi: 0xb1}, + {value: 0xa000, lo: 0xb3, hi: 0xb3}, + {value: 0x3eeb, lo: 0xb4, hi: 0xb4}, + {value: 0x9902, lo: 0xbc, hi: 0xbc}, + // Block 0xf, offset 0x7f + {value: 0x0008, lo: 0x06}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x8132, lo: 0x91, hi: 0x91}, + {value: 0x812d, lo: 0x92, hi: 0x92}, + {value: 0x8132, lo: 0x93, hi: 0x93}, + {value: 0x8132, lo: 0x94, hi: 0x94}, + {value: 0x451f, lo: 0x98, hi: 0x9f}, + // Block 0x10, offset 0x86 + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x11, offset 0x89 + {value: 0x0008, lo: 0x07}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0x2ca1, lo: 0x8b, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + {value: 0x455f, lo: 0x9c, hi: 0x9d}, + {value: 0x456f, lo: 0x9f, hi: 0x9f}, + {value: 0x8132, lo: 0xbe, hi: 0xbe}, + // Block 0x12, offset 0x91 + {value: 0x0000, lo: 0x03}, + {value: 0x4597, lo: 0xb3, hi: 0xb3}, + {value: 0x459f, lo: 0xb6, hi: 0xb6}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + // Block 0x13, offset 0x95 + {value: 0x0008, lo: 0x03}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x4577, lo: 0x99, hi: 0x9b}, + {value: 0x458f, lo: 0x9e, hi: 0x9e}, + // Block 0x14, offset 0x99 + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + // Block 0x15, offset 0x9b + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + // Block 0x16, offset 0x9d + {value: 0x0000, lo: 0x08}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0x2cb9, lo: 0x88, hi: 0x88}, + {value: 0x2cb1, lo: 0x8b, hi: 0x8b}, + {value: 0x2cc1, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x96, hi: 0x97}, + {value: 0x45a7, lo: 0x9c, hi: 0x9c}, + {value: 0x45af, lo: 0x9d, hi: 0x9d}, + // Block 0x17, offset 0xa6 + {value: 0x0000, lo: 0x03}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0x2cc9, lo: 0x94, hi: 0x94}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x18, offset 0xaa + {value: 0x0000, lo: 0x06}, + {value: 0xa000, lo: 0x86, hi: 0x87}, + {value: 0x2cd1, lo: 0x8a, hi: 0x8a}, + {value: 0x2ce1, lo: 0x8b, hi: 0x8b}, + {value: 0x2cd9, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + // Block 0x19, offset 0xb1 + {value: 0x1801, lo: 0x04}, + {value: 0xa000, lo: 0x86, hi: 0x86}, + {value: 0x3ef3, lo: 0x88, hi: 0x88}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x8120, lo: 0x95, hi: 0x96}, + // Block 0x1a, offset 0xb6 + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xbc, hi: 0xbc}, + {value: 0xa000, lo: 0xbf, hi: 0xbf}, + // Block 0x1b, offset 0xb9 + {value: 0x0000, lo: 0x09}, + {value: 0x2ce9, lo: 0x80, hi: 0x80}, + {value: 0x9900, lo: 0x82, hi: 0x82}, + {value: 0xa000, lo: 0x86, hi: 0x86}, + {value: 0x2cf1, lo: 0x87, hi: 0x87}, + {value: 0x2cf9, lo: 0x88, hi: 0x88}, + {value: 0x2f53, lo: 0x8a, hi: 0x8a}, + {value: 0x2ddb, lo: 0x8b, hi: 0x8b}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x95, hi: 0x96}, + // Block 0x1c, offset 0xc3 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xbb, hi: 0xbc}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x1d, offset 0xc6 + {value: 0x0000, lo: 0x06}, + {value: 0xa000, lo: 0x86, hi: 0x87}, + {value: 0x2d01, lo: 0x8a, hi: 0x8a}, + {value: 0x2d11, lo: 0x8b, hi: 0x8b}, + {value: 0x2d09, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + // Block 0x1e, offset 0xcd + {value: 0x6be7, lo: 0x07}, + {value: 0x9904, lo: 0x8a, hi: 0x8a}, + {value: 0x9900, lo: 0x8f, hi: 0x8f}, + {value: 0xa000, lo: 0x99, hi: 0x99}, + {value: 0x3efb, lo: 0x9a, hi: 0x9a}, + {value: 0x2f5b, lo: 0x9c, hi: 0x9c}, + {value: 0x2de6, lo: 0x9d, hi: 0x9d}, + {value: 0x2d19, lo: 0x9e, hi: 0x9f}, + // Block 0x1f, offset 0xd5 + {value: 0x0000, lo: 0x03}, + {value: 0x2624, lo: 0xb3, hi: 0xb3}, + {value: 0x8122, lo: 0xb8, hi: 0xb9}, + {value: 0x8104, lo: 0xba, hi: 0xba}, + // Block 0x20, offset 0xd9 + {value: 0x0000, lo: 0x01}, + {value: 0x8123, lo: 0x88, hi: 0x8b}, + // Block 0x21, offset 0xdb + {value: 0x0000, lo: 0x03}, + {value: 0x2639, lo: 0xb3, hi: 0xb3}, + {value: 0x8124, lo: 0xb8, hi: 0xb9}, + {value: 0x8104, lo: 0xba, hi: 0xba}, + // Block 0x22, offset 0xdf + {value: 0x0000, lo: 0x03}, + {value: 0x8125, lo: 0x88, hi: 0x8b}, + {value: 0x262b, lo: 0x9c, hi: 0x9c}, + {value: 0x2632, lo: 0x9d, hi: 0x9d}, + // Block 0x23, offset 0xe3 + {value: 0x0000, lo: 0x05}, + {value: 0x030b, lo: 0x8c, hi: 0x8c}, + {value: 0x812d, lo: 0x98, hi: 0x99}, + {value: 0x812d, lo: 0xb5, hi: 0xb5}, + {value: 0x812d, lo: 0xb7, hi: 0xb7}, + {value: 0x812b, lo: 0xb9, hi: 0xb9}, + // Block 0x24, offset 0xe9 + {value: 0x0000, lo: 0x10}, + {value: 0x2647, lo: 0x83, hi: 0x83}, + {value: 0x264e, lo: 0x8d, hi: 0x8d}, + {value: 0x2655, lo: 0x92, hi: 0x92}, + {value: 0x265c, lo: 0x97, hi: 0x97}, + {value: 0x2663, lo: 0x9c, hi: 0x9c}, + {value: 0x2640, lo: 0xa9, hi: 0xa9}, + {value: 0x8126, lo: 0xb1, hi: 0xb1}, + {value: 0x8127, lo: 0xb2, hi: 0xb2}, + {value: 0x4a87, lo: 0xb3, hi: 0xb3}, + {value: 0x8128, lo: 0xb4, hi: 0xb4}, + {value: 0x4a90, lo: 0xb5, hi: 0xb5}, + {value: 0x45b7, lo: 0xb6, hi: 0xb6}, + {value: 0x45f7, lo: 0xb7, hi: 0xb7}, + {value: 0x45bf, lo: 0xb8, hi: 0xb8}, + {value: 0x4602, lo: 0xb9, hi: 0xb9}, + {value: 0x8127, lo: 0xba, hi: 0xbd}, + // Block 0x25, offset 0xfa + {value: 0x0000, lo: 0x0b}, + {value: 0x8127, lo: 0x80, hi: 0x80}, + {value: 0x4a99, lo: 0x81, hi: 0x81}, + {value: 0x8132, lo: 0x82, hi: 0x83}, + {value: 0x8104, lo: 0x84, hi: 0x84}, + {value: 0x8132, lo: 0x86, hi: 0x87}, + {value: 0x2671, lo: 0x93, hi: 0x93}, + {value: 0x2678, lo: 0x9d, hi: 0x9d}, + {value: 0x267f, lo: 0xa2, hi: 0xa2}, + {value: 0x2686, lo: 0xa7, hi: 0xa7}, + {value: 0x268d, lo: 0xac, hi: 0xac}, + {value: 0x266a, lo: 0xb9, hi: 0xb9}, + // Block 0x26, offset 0x106 + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x86, hi: 0x86}, + // Block 0x27, offset 0x108 + {value: 0x0000, lo: 0x05}, + {value: 0xa000, lo: 0xa5, hi: 0xa5}, + {value: 0x2d21, lo: 0xa6, hi: 0xa6}, + {value: 0x9900, lo: 0xae, hi: 0xae}, + {value: 0x8102, lo: 0xb7, hi: 0xb7}, + {value: 0x8104, lo: 0xb9, hi: 0xba}, + // Block 0x28, offset 0x10e + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x8d, hi: 0x8d}, + // Block 0x29, offset 0x110 + {value: 0x0000, lo: 0x01}, + {value: 0x030f, lo: 0xbc, hi: 0xbc}, + // Block 0x2a, offset 0x112 + {value: 0x0000, lo: 0x01}, + {value: 0xa000, lo: 0x80, hi: 0x92}, + // Block 0x2b, offset 0x114 + {value: 0x0000, lo: 0x01}, + {value: 0xb900, lo: 0xa1, hi: 0xb5}, + // Block 0x2c, offset 0x116 + {value: 0x0000, lo: 0x01}, + {value: 0x9900, lo: 0xa8, hi: 0xbf}, + // Block 0x2d, offset 0x118 + {value: 0x0000, lo: 0x01}, + {value: 0x9900, lo: 0x80, hi: 0x82}, + // Block 0x2e, offset 0x11a + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0x9d, hi: 0x9f}, + // Block 0x2f, offset 0x11c + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x94, hi: 0x94}, + {value: 0x8104, lo: 0xb4, hi: 0xb4}, + // Block 0x30, offset 0x11f + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x92, hi: 0x92}, + {value: 0x8132, lo: 0x9d, hi: 0x9d}, + // Block 0x31, offset 0x122 + {value: 0x0000, lo: 0x01}, + {value: 0x8131, lo: 0xa9, hi: 0xa9}, + // Block 0x32, offset 0x124 + {value: 0x0004, lo: 0x02}, + {value: 0x812e, lo: 0xb9, hi: 0xba}, + {value: 0x812d, lo: 0xbb, hi: 0xbb}, + // Block 0x33, offset 0x127 + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0x97, hi: 0x97}, + {value: 0x812d, lo: 0x98, hi: 0x98}, + // Block 0x34, offset 0x12a + {value: 0x0000, lo: 0x03}, + {value: 0x8104, lo: 0xa0, hi: 0xa0}, + {value: 0x8132, lo: 0xb5, hi: 0xbc}, + {value: 0x812d, lo: 0xbf, hi: 0xbf}, + // Block 0x35, offset 0x12e + {value: 0x0000, lo: 0x04}, + {value: 0x8132, lo: 0xb0, hi: 0xb4}, + {value: 0x812d, lo: 0xb5, hi: 0xba}, + {value: 0x8132, lo: 0xbb, hi: 0xbc}, + {value: 0x812d, lo: 0xbd, hi: 0xbd}, + // Block 0x36, offset 0x133 + {value: 0x0000, lo: 0x08}, + {value: 0x2d69, lo: 0x80, hi: 0x80}, + {value: 0x2d71, lo: 0x81, hi: 0x81}, + {value: 0xa000, lo: 0x82, hi: 0x82}, + {value: 0x2d79, lo: 0x83, hi: 0x83}, + {value: 0x8104, lo: 0x84, hi: 0x84}, + {value: 0x8132, lo: 0xab, hi: 0xab}, + {value: 0x812d, lo: 0xac, hi: 0xac}, + {value: 0x8132, lo: 0xad, hi: 0xb3}, + // Block 0x37, offset 0x13c + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xaa, hi: 0xab}, + // Block 0x38, offset 0x13e + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xa6, hi: 0xa6}, + {value: 0x8104, lo: 0xb2, hi: 0xb3}, + // Block 0x39, offset 0x141 + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0xb7, hi: 0xb7}, + // Block 0x3a, offset 0x143 + {value: 0x0000, lo: 0x0a}, + {value: 0x8132, lo: 0x90, hi: 0x92}, + {value: 0x8101, lo: 0x94, hi: 0x94}, + {value: 0x812d, lo: 0x95, hi: 0x99}, + {value: 0x8132, lo: 0x9a, hi: 0x9b}, + {value: 0x812d, lo: 0x9c, hi: 0x9f}, + {value: 0x8132, lo: 0xa0, hi: 0xa0}, + {value: 0x8101, lo: 0xa2, hi: 0xa8}, + {value: 0x812d, lo: 0xad, hi: 0xad}, + {value: 0x8132, lo: 0xb4, hi: 0xb4}, + {value: 0x8132, lo: 0xb8, hi: 0xb9}, + // Block 0x3b, offset 0x14e + {value: 0x0002, lo: 0x0a}, + {value: 0x0043, lo: 0xac, hi: 0xac}, + {value: 0x00d1, lo: 0xad, hi: 0xad}, + {value: 0x0045, lo: 0xae, hi: 0xae}, + {value: 0x0049, lo: 0xb0, hi: 0xb1}, + {value: 0x00e6, lo: 0xb2, hi: 0xb2}, + {value: 0x004f, lo: 0xb3, hi: 0xba}, + {value: 0x005f, lo: 0xbc, hi: 0xbc}, + {value: 0x00ef, lo: 0xbd, hi: 0xbd}, + {value: 0x0061, lo: 0xbe, hi: 0xbe}, + {value: 0x0065, lo: 0xbf, hi: 0xbf}, + // Block 0x3c, offset 0x159 + {value: 0x0000, lo: 0x0d}, + {value: 0x0001, lo: 0x80, hi: 0x8a}, + {value: 0x043b, lo: 0x91, hi: 0x91}, + {value: 0x429e, lo: 0x97, hi: 0x97}, + {value: 0x001d, lo: 0xa4, hi: 0xa4}, + {value: 0x1873, lo: 0xa5, hi: 0xa5}, + {value: 0x1b5f, lo: 0xa6, hi: 0xa6}, + {value: 0x0001, lo: 0xaf, hi: 0xaf}, + {value: 0x2694, lo: 0xb3, hi: 0xb3}, + {value: 0x2801, lo: 0xb4, hi: 0xb4}, + {value: 0x269b, lo: 0xb6, hi: 0xb6}, + {value: 0x280b, lo: 0xb7, hi: 0xb7}, + {value: 0x186d, lo: 0xbc, hi: 0xbc}, + {value: 0x426c, lo: 0xbe, hi: 0xbe}, + // Block 0x3d, offset 0x167 + {value: 0x0002, lo: 0x0d}, + {value: 0x1933, lo: 0x87, hi: 0x87}, + {value: 0x1930, lo: 0x88, hi: 0x88}, + {value: 0x1870, lo: 0x89, hi: 0x89}, + {value: 0x2991, lo: 0x97, hi: 0x97}, + {value: 0x0001, lo: 0x9f, hi: 0x9f}, + {value: 0x0021, lo: 0xb0, hi: 0xb0}, + {value: 0x0093, lo: 0xb1, hi: 0xb1}, + {value: 0x0029, lo: 0xb4, hi: 0xb9}, + {value: 0x0017, lo: 0xba, hi: 0xba}, + {value: 0x0467, lo: 0xbb, hi: 0xbb}, + {value: 0x003b, lo: 0xbc, hi: 0xbc}, + {value: 0x0011, lo: 0xbd, hi: 0xbe}, + {value: 0x009d, lo: 0xbf, hi: 0xbf}, + // Block 0x3e, offset 0x175 + {value: 0x0002, lo: 0x0f}, + {value: 0x0021, lo: 0x80, hi: 0x89}, + {value: 0x0017, lo: 0x8a, hi: 0x8a}, + {value: 0x0467, lo: 0x8b, hi: 0x8b}, + {value: 0x003b, lo: 0x8c, hi: 0x8c}, + {value: 0x0011, lo: 0x8d, hi: 0x8e}, + {value: 0x0083, lo: 0x90, hi: 0x90}, + {value: 0x008b, lo: 0x91, hi: 0x91}, + {value: 0x009f, lo: 0x92, hi: 0x92}, + {value: 0x00b1, lo: 0x93, hi: 0x93}, + {value: 0x0104, lo: 0x94, hi: 0x94}, + {value: 0x0091, lo: 0x95, hi: 0x95}, + {value: 0x0097, lo: 0x96, hi: 0x99}, + {value: 0x00a1, lo: 0x9a, hi: 0x9a}, + {value: 0x00a7, lo: 0x9b, hi: 0x9c}, + {value: 0x199c, lo: 0xa8, hi: 0xa8}, + // Block 0x3f, offset 0x185 + {value: 0x0000, lo: 0x0d}, + {value: 0x8132, lo: 0x90, hi: 0x91}, + {value: 0x8101, lo: 0x92, hi: 0x93}, + {value: 0x8132, lo: 0x94, hi: 0x97}, + {value: 0x8101, lo: 0x98, hi: 0x9a}, + {value: 0x8132, lo: 0x9b, hi: 0x9c}, + {value: 0x8132, lo: 0xa1, hi: 0xa1}, + {value: 0x8101, lo: 0xa5, hi: 0xa6}, + {value: 0x8132, lo: 0xa7, hi: 0xa7}, + {value: 0x812d, lo: 0xa8, hi: 0xa8}, + {value: 0x8132, lo: 0xa9, hi: 0xa9}, + {value: 0x8101, lo: 0xaa, hi: 0xab}, + {value: 0x812d, lo: 0xac, hi: 0xaf}, + {value: 0x8132, lo: 0xb0, hi: 0xb0}, + // Block 0x40, offset 0x193 + {value: 0x0007, lo: 0x06}, + {value: 0x2183, lo: 0x89, hi: 0x89}, + {value: 0xa000, lo: 0x90, hi: 0x90}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0xa000, lo: 0x94, hi: 0x94}, + {value: 0x3bbc, lo: 0x9a, hi: 0x9b}, + {value: 0x3bca, lo: 0xae, hi: 0xae}, + // Block 0x41, offset 0x19a + {value: 0x000e, lo: 0x05}, + {value: 0x3bd1, lo: 0x8d, hi: 0x8e}, + {value: 0x3bd8, lo: 0x8f, hi: 0x8f}, + {value: 0xa000, lo: 0x90, hi: 0x90}, + {value: 0xa000, lo: 0x92, hi: 0x92}, + {value: 0xa000, lo: 0x94, hi: 0x94}, + // Block 0x42, offset 0x1a0 + {value: 0x0173, lo: 0x0e}, + {value: 0xa000, lo: 0x83, hi: 0x83}, + {value: 0x3be6, lo: 0x84, hi: 0x84}, + {value: 0xa000, lo: 0x88, hi: 0x88}, + {value: 0x3bed, lo: 0x89, hi: 0x89}, + {value: 0xa000, lo: 0x8b, hi: 0x8b}, + {value: 0x3bf4, lo: 0x8c, hi: 0x8c}, + {value: 0xa000, lo: 0xa3, hi: 0xa3}, + {value: 0x3bfb, lo: 0xa4, hi: 0xa4}, + {value: 0xa000, lo: 0xa5, hi: 0xa5}, + {value: 0x3c02, lo: 0xa6, hi: 0xa6}, + {value: 0x26a2, lo: 0xac, hi: 0xad}, + {value: 0x26a9, lo: 0xaf, hi: 0xaf}, + {value: 0x281f, lo: 0xb0, hi: 0xb0}, + {value: 0xa000, lo: 0xbc, hi: 0xbc}, + // Block 0x43, offset 0x1af + {value: 0x0007, lo: 0x03}, + {value: 0x3c6b, lo: 0xa0, hi: 0xa1}, + {value: 0x3c95, lo: 0xa2, hi: 0xa3}, + {value: 0x3cbf, lo: 0xaa, hi: 0xad}, + // Block 0x44, offset 0x1b3 + {value: 0x0004, lo: 0x01}, + {value: 0x048b, lo: 0xa9, hi: 0xaa}, + // Block 0x45, offset 0x1b5 + {value: 0x0002, lo: 0x03}, + {value: 0x0057, lo: 0x80, hi: 0x8f}, + {value: 0x0083, lo: 0x90, hi: 0xa9}, + {value: 0x0021, lo: 0xaa, hi: 0xaa}, + // Block 0x46, offset 0x1b9 + {value: 0x0000, lo: 0x01}, + {value: 0x299e, lo: 0x8c, hi: 0x8c}, + // Block 0x47, offset 0x1bb + {value: 0x0266, lo: 0x02}, + {value: 0x1b8f, lo: 0xb4, hi: 0xb4}, + {value: 0x192d, lo: 0xb5, hi: 0xb6}, + // Block 0x48, offset 0x1be + {value: 0x0000, lo: 0x01}, + {value: 0x44e0, lo: 0x9c, hi: 0x9c}, + // Block 0x49, offset 0x1c0 + {value: 0x0000, lo: 0x02}, + {value: 0x0095, lo: 0xbc, hi: 0xbc}, + {value: 0x006d, lo: 0xbd, hi: 0xbd}, + // Block 0x4a, offset 0x1c3 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xaf, hi: 0xb1}, + // Block 0x4b, offset 0x1c5 + {value: 0x0000, lo: 0x02}, + {value: 0x047f, lo: 0xaf, hi: 0xaf}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x4c, offset 0x1c8 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xa0, hi: 0xbf}, + // Block 0x4d, offset 0x1ca + {value: 0x0000, lo: 0x01}, + {value: 0x0dc3, lo: 0x9f, hi: 0x9f}, + // Block 0x4e, offset 0x1cc + {value: 0x0000, lo: 0x01}, + {value: 0x162f, lo: 0xb3, hi: 0xb3}, + // Block 0x4f, offset 0x1ce + {value: 0x0004, lo: 0x0b}, + {value: 0x1597, lo: 0x80, hi: 0x82}, + {value: 0x15af, lo: 0x83, hi: 0x83}, + {value: 0x15c7, lo: 0x84, hi: 0x85}, + {value: 0x15d7, lo: 0x86, hi: 0x89}, + {value: 0x15eb, lo: 0x8a, hi: 0x8c}, + {value: 0x15ff, lo: 0x8d, hi: 0x8d}, + {value: 0x1607, lo: 0x8e, hi: 0x8e}, + {value: 0x160f, lo: 0x8f, hi: 0x90}, + {value: 0x161b, lo: 0x91, hi: 0x93}, + {value: 0x162b, lo: 0x94, hi: 0x94}, + {value: 0x1633, lo: 0x95, hi: 0x95}, + // Block 0x50, offset 0x1da + {value: 0x0004, lo: 0x09}, + {value: 0x0001, lo: 0x80, hi: 0x80}, + {value: 0x812c, lo: 0xaa, hi: 0xaa}, + {value: 0x8131, lo: 0xab, hi: 0xab}, + {value: 0x8133, lo: 0xac, hi: 0xac}, + {value: 0x812e, lo: 0xad, hi: 0xad}, + {value: 0x812f, lo: 0xae, hi: 0xae}, + {value: 0x812f, lo: 0xaf, hi: 0xaf}, + {value: 0x04b3, lo: 0xb6, hi: 0xb6}, + {value: 0x0887, lo: 0xb8, hi: 0xba}, + // Block 0x51, offset 0x1e4 + {value: 0x0006, lo: 0x09}, + {value: 0x0313, lo: 0xb1, hi: 0xb1}, + {value: 0x0317, lo: 0xb2, hi: 0xb2}, + {value: 0x4a3e, lo: 0xb3, hi: 0xb3}, + {value: 0x031b, lo: 0xb4, hi: 0xb4}, + {value: 0x4a44, lo: 0xb5, hi: 0xb6}, + {value: 0x031f, lo: 0xb7, hi: 0xb7}, + {value: 0x0323, lo: 0xb8, hi: 0xb8}, + {value: 0x0327, lo: 0xb9, hi: 0xb9}, + {value: 0x4a50, lo: 0xba, hi: 0xbf}, + // Block 0x52, offset 0x1ee + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0xaf, hi: 0xaf}, + {value: 0x8132, lo: 0xb4, hi: 0xbd}, + // Block 0x53, offset 0x1f1 + {value: 0x0000, lo: 0x03}, + {value: 0x020f, lo: 0x9c, hi: 0x9c}, + {value: 0x0212, lo: 0x9d, hi: 0x9d}, + {value: 0x8132, lo: 0x9e, hi: 0x9f}, + // Block 0x54, offset 0x1f5 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xb0, hi: 0xb1}, + // Block 0x55, offset 0x1f7 + {value: 0x0000, lo: 0x01}, + {value: 0x163b, lo: 0xb0, hi: 0xb0}, + // Block 0x56, offset 0x1f9 + {value: 0x000c, lo: 0x01}, + {value: 0x00d7, lo: 0xb8, hi: 0xb9}, + // Block 0x57, offset 0x1fb + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x86, hi: 0x86}, + // Block 0x58, offset 0x1fd + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x84, hi: 0x84}, + {value: 0x8132, lo: 0xa0, hi: 0xb1}, + // Block 0x59, offset 0x200 + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0xab, hi: 0xad}, + // Block 0x5a, offset 0x202 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x93, hi: 0x93}, + // Block 0x5b, offset 0x204 + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0xb3, hi: 0xb3}, + // Block 0x5c, offset 0x206 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x80, hi: 0x80}, + // Block 0x5d, offset 0x208 + {value: 0x0000, lo: 0x05}, + {value: 0x8132, lo: 0xb0, hi: 0xb0}, + {value: 0x8132, lo: 0xb2, hi: 0xb3}, + {value: 0x812d, lo: 0xb4, hi: 0xb4}, + {value: 0x8132, lo: 0xb7, hi: 0xb8}, + {value: 0x8132, lo: 0xbe, hi: 0xbf}, + // Block 0x5e, offset 0x20e + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0x81, hi: 0x81}, + {value: 0x8104, lo: 0xb6, hi: 0xb6}, + // Block 0x5f, offset 0x211 + {value: 0x0008, lo: 0x03}, + {value: 0x1637, lo: 0x9c, hi: 0x9d}, + {value: 0x0125, lo: 0x9e, hi: 0x9e}, + {value: 0x1643, lo: 0x9f, hi: 0x9f}, + // Block 0x60, offset 0x215 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xad, hi: 0xad}, + // Block 0x61, offset 0x217 + {value: 0x0000, lo: 0x06}, + {value: 0xe500, lo: 0x80, hi: 0x80}, + {value: 0xc600, lo: 0x81, hi: 0x9b}, + {value: 0xe500, lo: 0x9c, hi: 0x9c}, + {value: 0xc600, lo: 0x9d, hi: 0xb7}, + {value: 0xe500, lo: 0xb8, hi: 0xb8}, + {value: 0xc600, lo: 0xb9, hi: 0xbf}, + // Block 0x62, offset 0x21e + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x93}, + {value: 0xe500, lo: 0x94, hi: 0x94}, + {value: 0xc600, lo: 0x95, hi: 0xaf}, + {value: 0xe500, lo: 0xb0, hi: 0xb0}, + {value: 0xc600, lo: 0xb1, hi: 0xbf}, + // Block 0x63, offset 0x224 + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x8b}, + {value: 0xe500, lo: 0x8c, hi: 0x8c}, + {value: 0xc600, lo: 0x8d, hi: 0xa7}, + {value: 0xe500, lo: 0xa8, hi: 0xa8}, + {value: 0xc600, lo: 0xa9, hi: 0xbf}, + // Block 0x64, offset 0x22a + {value: 0x0000, lo: 0x07}, + {value: 0xc600, lo: 0x80, hi: 0x83}, + {value: 0xe500, lo: 0x84, hi: 0x84}, + {value: 0xc600, lo: 0x85, hi: 0x9f}, + {value: 0xe500, lo: 0xa0, hi: 0xa0}, + {value: 0xc600, lo: 0xa1, hi: 0xbb}, + {value: 0xe500, lo: 0xbc, hi: 0xbc}, + {value: 0xc600, lo: 0xbd, hi: 0xbf}, + // Block 0x65, offset 0x232 + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x97}, + {value: 0xe500, lo: 0x98, hi: 0x98}, + {value: 0xc600, lo: 0x99, hi: 0xb3}, + {value: 0xe500, lo: 0xb4, hi: 0xb4}, + {value: 0xc600, lo: 0xb5, hi: 0xbf}, + // Block 0x66, offset 0x238 + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x8f}, + {value: 0xe500, lo: 0x90, hi: 0x90}, + {value: 0xc600, lo: 0x91, hi: 0xab}, + {value: 0xe500, lo: 0xac, hi: 0xac}, + {value: 0xc600, lo: 0xad, hi: 0xbf}, + // Block 0x67, offset 0x23e + {value: 0x0000, lo: 0x05}, + {value: 0xc600, lo: 0x80, hi: 0x87}, + {value: 0xe500, lo: 0x88, hi: 0x88}, + {value: 0xc600, lo: 0x89, hi: 0xa3}, + {value: 0xe500, lo: 0xa4, hi: 0xa4}, + {value: 0xc600, lo: 0xa5, hi: 0xbf}, + // Block 0x68, offset 0x244 + {value: 0x0000, lo: 0x03}, + {value: 0xc600, lo: 0x80, hi: 0x87}, + {value: 0xe500, lo: 0x88, hi: 0x88}, + {value: 0xc600, lo: 0x89, hi: 0xa3}, + // Block 0x69, offset 0x248 + {value: 0x0002, lo: 0x01}, + {value: 0x0003, lo: 0x81, hi: 0xbf}, + // Block 0x6a, offset 0x24a + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0xbd, hi: 0xbd}, + // Block 0x6b, offset 0x24c + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0xa0, hi: 0xa0}, + // Block 0x6c, offset 0x24e + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xb6, hi: 0xba}, + // Block 0x6d, offset 0x250 + {value: 0x002c, lo: 0x05}, + {value: 0x812d, lo: 0x8d, hi: 0x8d}, + {value: 0x8132, lo: 0x8f, hi: 0x8f}, + {value: 0x8132, lo: 0xb8, hi: 0xb8}, + {value: 0x8101, lo: 0xb9, hi: 0xba}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x6e, offset 0x256 + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0xa5, hi: 0xa5}, + {value: 0x812d, lo: 0xa6, hi: 0xa6}, + // Block 0x6f, offset 0x259 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xa4, hi: 0xa7}, + // Block 0x70, offset 0x25b + {value: 0x0000, lo: 0x05}, + {value: 0x812d, lo: 0x86, hi: 0x87}, + {value: 0x8132, lo: 0x88, hi: 0x8a}, + {value: 0x812d, lo: 0x8b, hi: 0x8b}, + {value: 0x8132, lo: 0x8c, hi: 0x8c}, + {value: 0x812d, lo: 0x8d, hi: 0x90}, + // Block 0x71, offset 0x261 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x86, hi: 0x86}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x72, offset 0x264 + {value: 0x17fe, lo: 0x07}, + {value: 0xa000, lo: 0x99, hi: 0x99}, + {value: 0x423b, lo: 0x9a, hi: 0x9a}, + {value: 0xa000, lo: 0x9b, hi: 0x9b}, + {value: 0x4245, lo: 0x9c, hi: 0x9c}, + {value: 0xa000, lo: 0xa5, hi: 0xa5}, + {value: 0x424f, lo: 0xab, hi: 0xab}, + {value: 0x8104, lo: 0xb9, hi: 0xba}, + // Block 0x73, offset 0x26c + {value: 0x0000, lo: 0x06}, + {value: 0x8132, lo: 0x80, hi: 0x82}, + {value: 0x9900, lo: 0xa7, hi: 0xa7}, + {value: 0x2d81, lo: 0xae, hi: 0xae}, + {value: 0x2d8b, lo: 0xaf, hi: 0xaf}, + {value: 0xa000, lo: 0xb1, hi: 0xb2}, + {value: 0x8104, lo: 0xb3, hi: 0xb4}, + // Block 0x74, offset 0x273 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x80, hi: 0x80}, + {value: 0x8102, lo: 0x8a, hi: 0x8a}, + // Block 0x75, offset 0x276 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xb5, hi: 0xb5}, + {value: 0x8102, lo: 0xb6, hi: 0xb6}, + // Block 0x76, offset 0x279 + {value: 0x0002, lo: 0x01}, + {value: 0x8102, lo: 0xa9, hi: 0xaa}, + // Block 0x77, offset 0x27b + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0xbb, hi: 0xbc}, + {value: 0x9900, lo: 0xbe, hi: 0xbe}, + // Block 0x78, offset 0x27e + {value: 0x0000, lo: 0x07}, + {value: 0xa000, lo: 0x87, hi: 0x87}, + {value: 0x2d95, lo: 0x8b, hi: 0x8b}, + {value: 0x2d9f, lo: 0x8c, hi: 0x8c}, + {value: 0x8104, lo: 0x8d, hi: 0x8d}, + {value: 0x9900, lo: 0x97, hi: 0x97}, + {value: 0x8132, lo: 0xa6, hi: 0xac}, + {value: 0x8132, lo: 0xb0, hi: 0xb4}, + // Block 0x79, offset 0x286 + {value: 0x0000, lo: 0x03}, + {value: 0x8104, lo: 0x82, hi: 0x82}, + {value: 0x8102, lo: 0x86, hi: 0x86}, + {value: 0x8132, lo: 0x9e, hi: 0x9e}, + // Block 0x7a, offset 0x28a + {value: 0x6b57, lo: 0x06}, + {value: 0x9900, lo: 0xb0, hi: 0xb0}, + {value: 0xa000, lo: 0xb9, hi: 0xb9}, + {value: 0x9900, lo: 0xba, hi: 0xba}, + {value: 0x2db3, lo: 0xbb, hi: 0xbb}, + {value: 0x2da9, lo: 0xbc, hi: 0xbd}, + {value: 0x2dbd, lo: 0xbe, hi: 0xbe}, + // Block 0x7b, offset 0x291 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0x82, hi: 0x82}, + {value: 0x8102, lo: 0x83, hi: 0x83}, + // Block 0x7c, offset 0x294 + {value: 0x0000, lo: 0x05}, + {value: 0x9900, lo: 0xaf, hi: 0xaf}, + {value: 0xa000, lo: 0xb8, hi: 0xb9}, + {value: 0x2dc7, lo: 0xba, hi: 0xba}, + {value: 0x2dd1, lo: 0xbb, hi: 0xbb}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x7d, offset 0x29a + {value: 0x0000, lo: 0x01}, + {value: 0x8102, lo: 0x80, hi: 0x80}, + // Block 0x7e, offset 0x29c + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xbf, hi: 0xbf}, + // Block 0x7f, offset 0x29e + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xb6, hi: 0xb6}, + {value: 0x8102, lo: 0xb7, hi: 0xb7}, + // Block 0x80, offset 0x2a1 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xab, hi: 0xab}, + // Block 0x81, offset 0x2a3 + {value: 0x0000, lo: 0x02}, + {value: 0x8104, lo: 0xb9, hi: 0xb9}, + {value: 0x8102, lo: 0xba, hi: 0xba}, + // Block 0x82, offset 0x2a6 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xa0, hi: 0xa0}, + // Block 0x83, offset 0x2a8 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0xb4, hi: 0xb4}, + // Block 0x84, offset 0x2aa + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x87, hi: 0x87}, + // Block 0x85, offset 0x2ac + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x99, hi: 0x99}, + // Block 0x86, offset 0x2ae + {value: 0x0000, lo: 0x02}, + {value: 0x8102, lo: 0x82, hi: 0x82}, + {value: 0x8104, lo: 0x84, hi: 0x85}, + // Block 0x87, offset 0x2b1 + {value: 0x0000, lo: 0x01}, + {value: 0x8104, lo: 0x97, hi: 0x97}, + // Block 0x88, offset 0x2b3 + {value: 0x0000, lo: 0x01}, + {value: 0x8101, lo: 0xb0, hi: 0xb4}, + // Block 0x89, offset 0x2b5 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xb0, hi: 0xb6}, + // Block 0x8a, offset 0x2b7 + {value: 0x0000, lo: 0x01}, + {value: 0x8101, lo: 0x9e, hi: 0x9e}, + // Block 0x8b, offset 0x2b9 + {value: 0x0000, lo: 0x0c}, + {value: 0x45cf, lo: 0x9e, hi: 0x9e}, + {value: 0x45d9, lo: 0x9f, hi: 0x9f}, + {value: 0x460d, lo: 0xa0, hi: 0xa0}, + {value: 0x461b, lo: 0xa1, hi: 0xa1}, + {value: 0x4629, lo: 0xa2, hi: 0xa2}, + {value: 0x4637, lo: 0xa3, hi: 0xa3}, + {value: 0x4645, lo: 0xa4, hi: 0xa4}, + {value: 0x812b, lo: 0xa5, hi: 0xa6}, + {value: 0x8101, lo: 0xa7, hi: 0xa9}, + {value: 0x8130, lo: 0xad, hi: 0xad}, + {value: 0x812b, lo: 0xae, hi: 0xb2}, + {value: 0x812d, lo: 0xbb, hi: 0xbf}, + // Block 0x8c, offset 0x2c6 + {value: 0x0000, lo: 0x09}, + {value: 0x812d, lo: 0x80, hi: 0x82}, + {value: 0x8132, lo: 0x85, hi: 0x89}, + {value: 0x812d, lo: 0x8a, hi: 0x8b}, + {value: 0x8132, lo: 0xaa, hi: 0xad}, + {value: 0x45e3, lo: 0xbb, hi: 0xbb}, + {value: 0x45ed, lo: 0xbc, hi: 0xbc}, + {value: 0x4653, lo: 0xbd, hi: 0xbd}, + {value: 0x466f, lo: 0xbe, hi: 0xbe}, + {value: 0x4661, lo: 0xbf, hi: 0xbf}, + // Block 0x8d, offset 0x2d0 + {value: 0x0000, lo: 0x01}, + {value: 0x467d, lo: 0x80, hi: 0x80}, + // Block 0x8e, offset 0x2d2 + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0x82, hi: 0x84}, + // Block 0x8f, offset 0x2d4 + {value: 0x0002, lo: 0x03}, + {value: 0x0043, lo: 0x80, hi: 0x99}, + {value: 0x0083, lo: 0x9a, hi: 0xb3}, + {value: 0x0043, lo: 0xb4, hi: 0xbf}, + // Block 0x90, offset 0x2d8 + {value: 0x0002, lo: 0x04}, + {value: 0x005b, lo: 0x80, hi: 0x8d}, + {value: 0x0083, lo: 0x8e, hi: 0x94}, + {value: 0x0093, lo: 0x96, hi: 0xa7}, + {value: 0x0043, lo: 0xa8, hi: 0xbf}, + // Block 0x91, offset 0x2dd + {value: 0x0002, lo: 0x0b}, + {value: 0x0073, lo: 0x80, hi: 0x81}, + {value: 0x0083, lo: 0x82, hi: 0x9b}, + {value: 0x0043, lo: 0x9c, hi: 0x9c}, + {value: 0x0047, lo: 0x9e, hi: 0x9f}, + {value: 0x004f, lo: 0xa2, hi: 0xa2}, + {value: 0x0055, lo: 0xa5, hi: 0xa6}, + {value: 0x005d, lo: 0xa9, hi: 0xac}, + {value: 0x0067, lo: 0xae, hi: 0xb5}, + {value: 0x0083, lo: 0xb6, hi: 0xb9}, + {value: 0x008d, lo: 0xbb, hi: 0xbb}, + {value: 0x0091, lo: 0xbd, hi: 0xbf}, + // Block 0x92, offset 0x2e9 + {value: 0x0002, lo: 0x04}, + {value: 0x0097, lo: 0x80, hi: 0x83}, + {value: 0x00a1, lo: 0x85, hi: 0x8f}, + {value: 0x0043, lo: 0x90, hi: 0xa9}, + {value: 0x0083, lo: 0xaa, hi: 0xbf}, + // Block 0x93, offset 0x2ee + {value: 0x0002, lo: 0x08}, + {value: 0x00af, lo: 0x80, hi: 0x83}, + {value: 0x0043, lo: 0x84, hi: 0x85}, + {value: 0x0049, lo: 0x87, hi: 0x8a}, + {value: 0x0055, lo: 0x8d, hi: 0x94}, + {value: 0x0067, lo: 0x96, hi: 0x9c}, + {value: 0x0083, lo: 0x9e, hi: 0xb7}, + {value: 0x0043, lo: 0xb8, hi: 0xb9}, + {value: 0x0049, lo: 0xbb, hi: 0xbe}, + // Block 0x94, offset 0x2f7 + {value: 0x0002, lo: 0x05}, + {value: 0x0053, lo: 0x80, hi: 0x84}, + {value: 0x005f, lo: 0x86, hi: 0x86}, + {value: 0x0067, lo: 0x8a, hi: 0x90}, + {value: 0x0083, lo: 0x92, hi: 0xab}, + {value: 0x0043, lo: 0xac, hi: 0xbf}, + // Block 0x95, offset 0x2fd + {value: 0x0002, lo: 0x04}, + {value: 0x006b, lo: 0x80, hi: 0x85}, + {value: 0x0083, lo: 0x86, hi: 0x9f}, + {value: 0x0043, lo: 0xa0, hi: 0xb9}, + {value: 0x0083, lo: 0xba, hi: 0xbf}, + // Block 0x96, offset 0x302 + {value: 0x0002, lo: 0x03}, + {value: 0x008f, lo: 0x80, hi: 0x93}, + {value: 0x0043, lo: 0x94, hi: 0xad}, + {value: 0x0083, lo: 0xae, hi: 0xbf}, + // Block 0x97, offset 0x306 + {value: 0x0002, lo: 0x04}, + {value: 0x00a7, lo: 0x80, hi: 0x87}, + {value: 0x0043, lo: 0x88, hi: 0xa1}, + {value: 0x0083, lo: 0xa2, hi: 0xbb}, + {value: 0x0043, lo: 0xbc, hi: 0xbf}, + // Block 0x98, offset 0x30b + {value: 0x0002, lo: 0x03}, + {value: 0x004b, lo: 0x80, hi: 0x95}, + {value: 0x0083, lo: 0x96, hi: 0xaf}, + {value: 0x0043, lo: 0xb0, hi: 0xbf}, + // Block 0x99, offset 0x30f + {value: 0x0003, lo: 0x0f}, + {value: 0x01b8, lo: 0x80, hi: 0x80}, + {value: 0x045f, lo: 0x81, hi: 0x81}, + {value: 0x01bb, lo: 0x82, hi: 0x9a}, + {value: 0x045b, lo: 0x9b, hi: 0x9b}, + {value: 0x01c7, lo: 0x9c, hi: 0x9c}, + {value: 0x01d0, lo: 0x9d, hi: 0x9d}, + {value: 0x01d6, lo: 0x9e, hi: 0x9e}, + {value: 0x01fa, lo: 0x9f, hi: 0x9f}, + {value: 0x01eb, lo: 0xa0, hi: 0xa0}, + {value: 0x01e8, lo: 0xa1, hi: 0xa1}, + {value: 0x0173, lo: 0xa2, hi: 0xb2}, + {value: 0x0188, lo: 0xb3, hi: 0xb3}, + {value: 0x01a6, lo: 0xb4, hi: 0xba}, + {value: 0x045f, lo: 0xbb, hi: 0xbb}, + {value: 0x01bb, lo: 0xbc, hi: 0xbf}, + // Block 0x9a, offset 0x31f + {value: 0x0003, lo: 0x0d}, + {value: 0x01c7, lo: 0x80, hi: 0x94}, + {value: 0x045b, lo: 0x95, hi: 0x95}, + {value: 0x01c7, lo: 0x96, hi: 0x96}, + {value: 0x01d0, lo: 0x97, hi: 0x97}, + {value: 0x01d6, lo: 0x98, hi: 0x98}, + {value: 0x01fa, lo: 0x99, hi: 0x99}, + {value: 0x01eb, lo: 0x9a, hi: 0x9a}, + {value: 0x01e8, lo: 0x9b, hi: 0x9b}, + {value: 0x0173, lo: 0x9c, hi: 0xac}, + {value: 0x0188, lo: 0xad, hi: 0xad}, + {value: 0x01a6, lo: 0xae, hi: 0xb4}, + {value: 0x045f, lo: 0xb5, hi: 0xb5}, + {value: 0x01bb, lo: 0xb6, hi: 0xbf}, + // Block 0x9b, offset 0x32d + {value: 0x0003, lo: 0x0d}, + {value: 0x01d9, lo: 0x80, hi: 0x8e}, + {value: 0x045b, lo: 0x8f, hi: 0x8f}, + {value: 0x01c7, lo: 0x90, hi: 0x90}, + {value: 0x01d0, lo: 0x91, hi: 0x91}, + {value: 0x01d6, lo: 0x92, hi: 0x92}, + {value: 0x01fa, lo: 0x93, hi: 0x93}, + {value: 0x01eb, lo: 0x94, hi: 0x94}, + {value: 0x01e8, lo: 0x95, hi: 0x95}, + {value: 0x0173, lo: 0x96, hi: 0xa6}, + {value: 0x0188, lo: 0xa7, hi: 0xa7}, + {value: 0x01a6, lo: 0xa8, hi: 0xae}, + {value: 0x045f, lo: 0xaf, hi: 0xaf}, + {value: 0x01bb, lo: 0xb0, hi: 0xbf}, + // Block 0x9c, offset 0x33b + {value: 0x0003, lo: 0x0d}, + {value: 0x01eb, lo: 0x80, hi: 0x88}, + {value: 0x045b, lo: 0x89, hi: 0x89}, + {value: 0x01c7, lo: 0x8a, hi: 0x8a}, + {value: 0x01d0, lo: 0x8b, hi: 0x8b}, + {value: 0x01d6, lo: 0x8c, hi: 0x8c}, + {value: 0x01fa, lo: 0x8d, hi: 0x8d}, + {value: 0x01eb, lo: 0x8e, hi: 0x8e}, + {value: 0x01e8, lo: 0x8f, hi: 0x8f}, + {value: 0x0173, lo: 0x90, hi: 0xa0}, + {value: 0x0188, lo: 0xa1, hi: 0xa1}, + {value: 0x01a6, lo: 0xa2, hi: 0xa8}, + {value: 0x045f, lo: 0xa9, hi: 0xa9}, + {value: 0x01bb, lo: 0xaa, hi: 0xbf}, + // Block 0x9d, offset 0x349 + {value: 0x0000, lo: 0x05}, + {value: 0x8132, lo: 0x80, hi: 0x86}, + {value: 0x8132, lo: 0x88, hi: 0x98}, + {value: 0x8132, lo: 0x9b, hi: 0xa1}, + {value: 0x8132, lo: 0xa3, hi: 0xa4}, + {value: 0x8132, lo: 0xa6, hi: 0xaa}, + // Block 0x9e, offset 0x34f + {value: 0x0000, lo: 0x01}, + {value: 0x8132, lo: 0xac, hi: 0xaf}, + // Block 0x9f, offset 0x351 + {value: 0x0000, lo: 0x01}, + {value: 0x812d, lo: 0x90, hi: 0x96}, + // Block 0xa0, offset 0x353 + {value: 0x0000, lo: 0x02}, + {value: 0x8132, lo: 0x84, hi: 0x89}, + {value: 0x8102, lo: 0x8a, hi: 0x8a}, + // Block 0xa1, offset 0x356 + {value: 0x0002, lo: 0x0a}, + {value: 0x0063, lo: 0x80, hi: 0x89}, + {value: 0x1951, lo: 0x8a, hi: 0x8a}, + {value: 0x1984, lo: 0x8b, hi: 0x8b}, + {value: 0x199f, lo: 0x8c, hi: 0x8c}, + {value: 0x19a5, lo: 0x8d, hi: 0x8d}, + {value: 0x1bc3, lo: 0x8e, hi: 0x8e}, + {value: 0x19b1, lo: 0x8f, hi: 0x8f}, + {value: 0x197b, lo: 0xaa, hi: 0xaa}, + {value: 0x197e, lo: 0xab, hi: 0xab}, + {value: 0x1981, lo: 0xac, hi: 0xac}, + // Block 0xa2, offset 0x361 + {value: 0x0000, lo: 0x01}, + {value: 0x193f, lo: 0x90, hi: 0x90}, + // Block 0xa3, offset 0x363 + {value: 0x0028, lo: 0x09}, + {value: 0x2865, lo: 0x80, hi: 0x80}, + {value: 0x2829, lo: 0x81, hi: 0x81}, + {value: 0x2833, lo: 0x82, hi: 0x82}, + {value: 0x2847, lo: 0x83, hi: 0x84}, + {value: 0x2851, lo: 0x85, hi: 0x86}, + {value: 0x283d, lo: 0x87, hi: 0x87}, + {value: 0x285b, lo: 0x88, hi: 0x88}, + {value: 0x0b6f, lo: 0x90, hi: 0x90}, + {value: 0x08e7, lo: 0x91, hi: 0x91}, +} + +// recompMap: 7520 bytes (entries only) +var recompMap map[uint32]rune +var recompMapOnce sync.Once + +const recompMapPacked = "" + + "\x00A\x03\x00\x00\x00\x00\xc0" + // 0x00410300: 0x000000C0 + "\x00A\x03\x01\x00\x00\x00\xc1" + // 0x00410301: 0x000000C1 + "\x00A\x03\x02\x00\x00\x00\xc2" + // 0x00410302: 0x000000C2 + "\x00A\x03\x03\x00\x00\x00\xc3" + // 0x00410303: 0x000000C3 + "\x00A\x03\b\x00\x00\x00\xc4" + // 0x00410308: 0x000000C4 + "\x00A\x03\n\x00\x00\x00\xc5" + // 0x0041030A: 0x000000C5 + "\x00C\x03'\x00\x00\x00\xc7" + // 0x00430327: 0x000000C7 + "\x00E\x03\x00\x00\x00\x00\xc8" + // 0x00450300: 0x000000C8 + "\x00E\x03\x01\x00\x00\x00\xc9" + // 0x00450301: 0x000000C9 + "\x00E\x03\x02\x00\x00\x00\xca" + // 0x00450302: 0x000000CA + "\x00E\x03\b\x00\x00\x00\xcb" + // 0x00450308: 0x000000CB + "\x00I\x03\x00\x00\x00\x00\xcc" + // 0x00490300: 0x000000CC + "\x00I\x03\x01\x00\x00\x00\xcd" + // 0x00490301: 0x000000CD + "\x00I\x03\x02\x00\x00\x00\xce" + // 0x00490302: 0x000000CE + "\x00I\x03\b\x00\x00\x00\xcf" + // 0x00490308: 0x000000CF + "\x00N\x03\x03\x00\x00\x00\xd1" + // 0x004E0303: 0x000000D1 + "\x00O\x03\x00\x00\x00\x00\xd2" + // 0x004F0300: 0x000000D2 + "\x00O\x03\x01\x00\x00\x00\xd3" + // 0x004F0301: 0x000000D3 + "\x00O\x03\x02\x00\x00\x00\xd4" + // 0x004F0302: 0x000000D4 + "\x00O\x03\x03\x00\x00\x00\xd5" + // 0x004F0303: 0x000000D5 + "\x00O\x03\b\x00\x00\x00\xd6" + // 0x004F0308: 0x000000D6 + "\x00U\x03\x00\x00\x00\x00\xd9" + // 0x00550300: 0x000000D9 + "\x00U\x03\x01\x00\x00\x00\xda" + // 0x00550301: 0x000000DA + "\x00U\x03\x02\x00\x00\x00\xdb" + // 0x00550302: 0x000000DB + "\x00U\x03\b\x00\x00\x00\xdc" + // 0x00550308: 0x000000DC + "\x00Y\x03\x01\x00\x00\x00\xdd" + // 0x00590301: 0x000000DD + "\x00a\x03\x00\x00\x00\x00\xe0" + // 0x00610300: 0x000000E0 + "\x00a\x03\x01\x00\x00\x00\xe1" + // 0x00610301: 0x000000E1 + "\x00a\x03\x02\x00\x00\x00\xe2" + // 0x00610302: 0x000000E2 + "\x00a\x03\x03\x00\x00\x00\xe3" + // 0x00610303: 0x000000E3 + "\x00a\x03\b\x00\x00\x00\xe4" + // 0x00610308: 0x000000E4 + "\x00a\x03\n\x00\x00\x00\xe5" + // 0x0061030A: 0x000000E5 + "\x00c\x03'\x00\x00\x00\xe7" + // 0x00630327: 0x000000E7 + "\x00e\x03\x00\x00\x00\x00\xe8" + // 0x00650300: 0x000000E8 + "\x00e\x03\x01\x00\x00\x00\xe9" + // 0x00650301: 0x000000E9 + "\x00e\x03\x02\x00\x00\x00\xea" + // 0x00650302: 0x000000EA + "\x00e\x03\b\x00\x00\x00\xeb" + // 0x00650308: 0x000000EB + "\x00i\x03\x00\x00\x00\x00\xec" + // 0x00690300: 0x000000EC + "\x00i\x03\x01\x00\x00\x00\xed" + // 0x00690301: 0x000000ED + "\x00i\x03\x02\x00\x00\x00\xee" + // 0x00690302: 0x000000EE + "\x00i\x03\b\x00\x00\x00\xef" + // 0x00690308: 0x000000EF + "\x00n\x03\x03\x00\x00\x00\xf1" + // 0x006E0303: 0x000000F1 + "\x00o\x03\x00\x00\x00\x00\xf2" + // 0x006F0300: 0x000000F2 + "\x00o\x03\x01\x00\x00\x00\xf3" + // 0x006F0301: 0x000000F3 + "\x00o\x03\x02\x00\x00\x00\xf4" + // 0x006F0302: 0x000000F4 + "\x00o\x03\x03\x00\x00\x00\xf5" + // 0x006F0303: 0x000000F5 + "\x00o\x03\b\x00\x00\x00\xf6" + // 0x006F0308: 0x000000F6 + "\x00u\x03\x00\x00\x00\x00\xf9" + // 0x00750300: 0x000000F9 + "\x00u\x03\x01\x00\x00\x00\xfa" + // 0x00750301: 0x000000FA + "\x00u\x03\x02\x00\x00\x00\xfb" + // 0x00750302: 0x000000FB + "\x00u\x03\b\x00\x00\x00\xfc" + // 0x00750308: 0x000000FC + "\x00y\x03\x01\x00\x00\x00\xfd" + // 0x00790301: 0x000000FD + "\x00y\x03\b\x00\x00\x00\xff" + // 0x00790308: 0x000000FF + "\x00A\x03\x04\x00\x00\x01\x00" + // 0x00410304: 0x00000100 + "\x00a\x03\x04\x00\x00\x01\x01" + // 0x00610304: 0x00000101 + "\x00A\x03\x06\x00\x00\x01\x02" + // 0x00410306: 0x00000102 + "\x00a\x03\x06\x00\x00\x01\x03" + // 0x00610306: 0x00000103 + "\x00A\x03(\x00\x00\x01\x04" + // 0x00410328: 0x00000104 + "\x00a\x03(\x00\x00\x01\x05" + // 0x00610328: 0x00000105 + "\x00C\x03\x01\x00\x00\x01\x06" + // 0x00430301: 0x00000106 + "\x00c\x03\x01\x00\x00\x01\a" + // 0x00630301: 0x00000107 + "\x00C\x03\x02\x00\x00\x01\b" + // 0x00430302: 0x00000108 + "\x00c\x03\x02\x00\x00\x01\t" + // 0x00630302: 0x00000109 + "\x00C\x03\a\x00\x00\x01\n" + // 0x00430307: 0x0000010A + "\x00c\x03\a\x00\x00\x01\v" + // 0x00630307: 0x0000010B + "\x00C\x03\f\x00\x00\x01\f" + // 0x0043030C: 0x0000010C + "\x00c\x03\f\x00\x00\x01\r" + // 0x0063030C: 0x0000010D + "\x00D\x03\f\x00\x00\x01\x0e" + // 0x0044030C: 0x0000010E + "\x00d\x03\f\x00\x00\x01\x0f" + // 0x0064030C: 0x0000010F + "\x00E\x03\x04\x00\x00\x01\x12" + // 0x00450304: 0x00000112 + "\x00e\x03\x04\x00\x00\x01\x13" + // 0x00650304: 0x00000113 + "\x00E\x03\x06\x00\x00\x01\x14" + // 0x00450306: 0x00000114 + "\x00e\x03\x06\x00\x00\x01\x15" + // 0x00650306: 0x00000115 + "\x00E\x03\a\x00\x00\x01\x16" + // 0x00450307: 0x00000116 + "\x00e\x03\a\x00\x00\x01\x17" + // 0x00650307: 0x00000117 + "\x00E\x03(\x00\x00\x01\x18" + // 0x00450328: 0x00000118 + "\x00e\x03(\x00\x00\x01\x19" + // 0x00650328: 0x00000119 + "\x00E\x03\f\x00\x00\x01\x1a" + // 0x0045030C: 0x0000011A + "\x00e\x03\f\x00\x00\x01\x1b" + // 0x0065030C: 0x0000011B + "\x00G\x03\x02\x00\x00\x01\x1c" + // 0x00470302: 0x0000011C + "\x00g\x03\x02\x00\x00\x01\x1d" + // 0x00670302: 0x0000011D + "\x00G\x03\x06\x00\x00\x01\x1e" + // 0x00470306: 0x0000011E + "\x00g\x03\x06\x00\x00\x01\x1f" + // 0x00670306: 0x0000011F + "\x00G\x03\a\x00\x00\x01 " + // 0x00470307: 0x00000120 + "\x00g\x03\a\x00\x00\x01!" + // 0x00670307: 0x00000121 + "\x00G\x03'\x00\x00\x01\"" + // 0x00470327: 0x00000122 + "\x00g\x03'\x00\x00\x01#" + // 0x00670327: 0x00000123 + "\x00H\x03\x02\x00\x00\x01$" + // 0x00480302: 0x00000124 + "\x00h\x03\x02\x00\x00\x01%" + // 0x00680302: 0x00000125 + "\x00I\x03\x03\x00\x00\x01(" + // 0x00490303: 0x00000128 + "\x00i\x03\x03\x00\x00\x01)" + // 0x00690303: 0x00000129 + "\x00I\x03\x04\x00\x00\x01*" + // 0x00490304: 0x0000012A + "\x00i\x03\x04\x00\x00\x01+" + // 0x00690304: 0x0000012B + "\x00I\x03\x06\x00\x00\x01," + // 0x00490306: 0x0000012C + "\x00i\x03\x06\x00\x00\x01-" + // 0x00690306: 0x0000012D + "\x00I\x03(\x00\x00\x01." + // 0x00490328: 0x0000012E + "\x00i\x03(\x00\x00\x01/" + // 0x00690328: 0x0000012F + "\x00I\x03\a\x00\x00\x010" + // 0x00490307: 0x00000130 + "\x00J\x03\x02\x00\x00\x014" + // 0x004A0302: 0x00000134 + "\x00j\x03\x02\x00\x00\x015" + // 0x006A0302: 0x00000135 + "\x00K\x03'\x00\x00\x016" + // 0x004B0327: 0x00000136 + "\x00k\x03'\x00\x00\x017" + // 0x006B0327: 0x00000137 + "\x00L\x03\x01\x00\x00\x019" + // 0x004C0301: 0x00000139 + "\x00l\x03\x01\x00\x00\x01:" + // 0x006C0301: 0x0000013A + "\x00L\x03'\x00\x00\x01;" + // 0x004C0327: 0x0000013B + "\x00l\x03'\x00\x00\x01<" + // 0x006C0327: 0x0000013C + "\x00L\x03\f\x00\x00\x01=" + // 0x004C030C: 0x0000013D + "\x00l\x03\f\x00\x00\x01>" + // 0x006C030C: 0x0000013E + "\x00N\x03\x01\x00\x00\x01C" + // 0x004E0301: 0x00000143 + "\x00n\x03\x01\x00\x00\x01D" + // 0x006E0301: 0x00000144 + "\x00N\x03'\x00\x00\x01E" + // 0x004E0327: 0x00000145 + "\x00n\x03'\x00\x00\x01F" + // 0x006E0327: 0x00000146 + "\x00N\x03\f\x00\x00\x01G" + // 0x004E030C: 0x00000147 + "\x00n\x03\f\x00\x00\x01H" + // 0x006E030C: 0x00000148 + "\x00O\x03\x04\x00\x00\x01L" + // 0x004F0304: 0x0000014C + "\x00o\x03\x04\x00\x00\x01M" + // 0x006F0304: 0x0000014D + "\x00O\x03\x06\x00\x00\x01N" + // 0x004F0306: 0x0000014E + "\x00o\x03\x06\x00\x00\x01O" + // 0x006F0306: 0x0000014F + "\x00O\x03\v\x00\x00\x01P" + // 0x004F030B: 0x00000150 + "\x00o\x03\v\x00\x00\x01Q" + // 0x006F030B: 0x00000151 + "\x00R\x03\x01\x00\x00\x01T" + // 0x00520301: 0x00000154 + "\x00r\x03\x01\x00\x00\x01U" + // 0x00720301: 0x00000155 + "\x00R\x03'\x00\x00\x01V" + // 0x00520327: 0x00000156 + "\x00r\x03'\x00\x00\x01W" + // 0x00720327: 0x00000157 + "\x00R\x03\f\x00\x00\x01X" + // 0x0052030C: 0x00000158 + "\x00r\x03\f\x00\x00\x01Y" + // 0x0072030C: 0x00000159 + "\x00S\x03\x01\x00\x00\x01Z" + // 0x00530301: 0x0000015A + "\x00s\x03\x01\x00\x00\x01[" + // 0x00730301: 0x0000015B + "\x00S\x03\x02\x00\x00\x01\\" + // 0x00530302: 0x0000015C + "\x00s\x03\x02\x00\x00\x01]" + // 0x00730302: 0x0000015D + "\x00S\x03'\x00\x00\x01^" + // 0x00530327: 0x0000015E + "\x00s\x03'\x00\x00\x01_" + // 0x00730327: 0x0000015F + "\x00S\x03\f\x00\x00\x01`" + // 0x0053030C: 0x00000160 + "\x00s\x03\f\x00\x00\x01a" + // 0x0073030C: 0x00000161 + "\x00T\x03'\x00\x00\x01b" + // 0x00540327: 0x00000162 + "\x00t\x03'\x00\x00\x01c" + // 0x00740327: 0x00000163 + "\x00T\x03\f\x00\x00\x01d" + // 0x0054030C: 0x00000164 + "\x00t\x03\f\x00\x00\x01e" + // 0x0074030C: 0x00000165 + "\x00U\x03\x03\x00\x00\x01h" + // 0x00550303: 0x00000168 + "\x00u\x03\x03\x00\x00\x01i" + // 0x00750303: 0x00000169 + "\x00U\x03\x04\x00\x00\x01j" + // 0x00550304: 0x0000016A + "\x00u\x03\x04\x00\x00\x01k" + // 0x00750304: 0x0000016B + "\x00U\x03\x06\x00\x00\x01l" + // 0x00550306: 0x0000016C + "\x00u\x03\x06\x00\x00\x01m" + // 0x00750306: 0x0000016D + "\x00U\x03\n\x00\x00\x01n" + // 0x0055030A: 0x0000016E + "\x00u\x03\n\x00\x00\x01o" + // 0x0075030A: 0x0000016F + "\x00U\x03\v\x00\x00\x01p" + // 0x0055030B: 0x00000170 + "\x00u\x03\v\x00\x00\x01q" + // 0x0075030B: 0x00000171 + "\x00U\x03(\x00\x00\x01r" + // 0x00550328: 0x00000172 + "\x00u\x03(\x00\x00\x01s" + // 0x00750328: 0x00000173 + "\x00W\x03\x02\x00\x00\x01t" + // 0x00570302: 0x00000174 + "\x00w\x03\x02\x00\x00\x01u" + // 0x00770302: 0x00000175 + "\x00Y\x03\x02\x00\x00\x01v" + // 0x00590302: 0x00000176 + "\x00y\x03\x02\x00\x00\x01w" + // 0x00790302: 0x00000177 + "\x00Y\x03\b\x00\x00\x01x" + // 0x00590308: 0x00000178 + "\x00Z\x03\x01\x00\x00\x01y" + // 0x005A0301: 0x00000179 + "\x00z\x03\x01\x00\x00\x01z" + // 0x007A0301: 0x0000017A + "\x00Z\x03\a\x00\x00\x01{" + // 0x005A0307: 0x0000017B + "\x00z\x03\a\x00\x00\x01|" + // 0x007A0307: 0x0000017C + "\x00Z\x03\f\x00\x00\x01}" + // 0x005A030C: 0x0000017D + "\x00z\x03\f\x00\x00\x01~" + // 0x007A030C: 0x0000017E + "\x00O\x03\x1b\x00\x00\x01\xa0" + // 0x004F031B: 0x000001A0 + "\x00o\x03\x1b\x00\x00\x01\xa1" + // 0x006F031B: 0x000001A1 + "\x00U\x03\x1b\x00\x00\x01\xaf" + // 0x0055031B: 0x000001AF + "\x00u\x03\x1b\x00\x00\x01\xb0" + // 0x0075031B: 0x000001B0 + "\x00A\x03\f\x00\x00\x01\xcd" + // 0x0041030C: 0x000001CD + "\x00a\x03\f\x00\x00\x01\xce" + // 0x0061030C: 0x000001CE + "\x00I\x03\f\x00\x00\x01\xcf" + // 0x0049030C: 0x000001CF + "\x00i\x03\f\x00\x00\x01\xd0" + // 0x0069030C: 0x000001D0 + "\x00O\x03\f\x00\x00\x01\xd1" + // 0x004F030C: 0x000001D1 + "\x00o\x03\f\x00\x00\x01\xd2" + // 0x006F030C: 0x000001D2 + "\x00U\x03\f\x00\x00\x01\xd3" + // 0x0055030C: 0x000001D3 + "\x00u\x03\f\x00\x00\x01\xd4" + // 0x0075030C: 0x000001D4 + "\x00\xdc\x03\x04\x00\x00\x01\xd5" + // 0x00DC0304: 0x000001D5 + "\x00\xfc\x03\x04\x00\x00\x01\xd6" + // 0x00FC0304: 0x000001D6 + "\x00\xdc\x03\x01\x00\x00\x01\xd7" + // 0x00DC0301: 0x000001D7 + "\x00\xfc\x03\x01\x00\x00\x01\xd8" + // 0x00FC0301: 0x000001D8 + "\x00\xdc\x03\f\x00\x00\x01\xd9" + // 0x00DC030C: 0x000001D9 + "\x00\xfc\x03\f\x00\x00\x01\xda" + // 0x00FC030C: 0x000001DA + "\x00\xdc\x03\x00\x00\x00\x01\xdb" + // 0x00DC0300: 0x000001DB + "\x00\xfc\x03\x00\x00\x00\x01\xdc" + // 0x00FC0300: 0x000001DC + "\x00\xc4\x03\x04\x00\x00\x01\xde" + // 0x00C40304: 0x000001DE + "\x00\xe4\x03\x04\x00\x00\x01\xdf" + // 0x00E40304: 0x000001DF + "\x02&\x03\x04\x00\x00\x01\xe0" + // 0x02260304: 0x000001E0 + "\x02'\x03\x04\x00\x00\x01\xe1" + // 0x02270304: 0x000001E1 + "\x00\xc6\x03\x04\x00\x00\x01\xe2" + // 0x00C60304: 0x000001E2 + "\x00\xe6\x03\x04\x00\x00\x01\xe3" + // 0x00E60304: 0x000001E3 + "\x00G\x03\f\x00\x00\x01\xe6" + // 0x0047030C: 0x000001E6 + "\x00g\x03\f\x00\x00\x01\xe7" + // 0x0067030C: 0x000001E7 + "\x00K\x03\f\x00\x00\x01\xe8" + // 0x004B030C: 0x000001E8 + "\x00k\x03\f\x00\x00\x01\xe9" + // 0x006B030C: 0x000001E9 + "\x00O\x03(\x00\x00\x01\xea" + // 0x004F0328: 0x000001EA + "\x00o\x03(\x00\x00\x01\xeb" + // 0x006F0328: 0x000001EB + "\x01\xea\x03\x04\x00\x00\x01\xec" + // 0x01EA0304: 0x000001EC + "\x01\xeb\x03\x04\x00\x00\x01\xed" + // 0x01EB0304: 0x000001ED + "\x01\xb7\x03\f\x00\x00\x01\xee" + // 0x01B7030C: 0x000001EE + "\x02\x92\x03\f\x00\x00\x01\xef" + // 0x0292030C: 0x000001EF + "\x00j\x03\f\x00\x00\x01\xf0" + // 0x006A030C: 0x000001F0 + "\x00G\x03\x01\x00\x00\x01\xf4" + // 0x00470301: 0x000001F4 + "\x00g\x03\x01\x00\x00\x01\xf5" + // 0x00670301: 0x000001F5 + "\x00N\x03\x00\x00\x00\x01\xf8" + // 0x004E0300: 0x000001F8 + "\x00n\x03\x00\x00\x00\x01\xf9" + // 0x006E0300: 0x000001F9 + "\x00\xc5\x03\x01\x00\x00\x01\xfa" + // 0x00C50301: 0x000001FA + "\x00\xe5\x03\x01\x00\x00\x01\xfb" + // 0x00E50301: 0x000001FB + "\x00\xc6\x03\x01\x00\x00\x01\xfc" + // 0x00C60301: 0x000001FC + "\x00\xe6\x03\x01\x00\x00\x01\xfd" + // 0x00E60301: 0x000001FD + "\x00\xd8\x03\x01\x00\x00\x01\xfe" + // 0x00D80301: 0x000001FE + "\x00\xf8\x03\x01\x00\x00\x01\xff" + // 0x00F80301: 0x000001FF + "\x00A\x03\x0f\x00\x00\x02\x00" + // 0x0041030F: 0x00000200 + "\x00a\x03\x0f\x00\x00\x02\x01" + // 0x0061030F: 0x00000201 + "\x00A\x03\x11\x00\x00\x02\x02" + // 0x00410311: 0x00000202 + "\x00a\x03\x11\x00\x00\x02\x03" + // 0x00610311: 0x00000203 + "\x00E\x03\x0f\x00\x00\x02\x04" + // 0x0045030F: 0x00000204 + "\x00e\x03\x0f\x00\x00\x02\x05" + // 0x0065030F: 0x00000205 + "\x00E\x03\x11\x00\x00\x02\x06" + // 0x00450311: 0x00000206 + "\x00e\x03\x11\x00\x00\x02\a" + // 0x00650311: 0x00000207 + "\x00I\x03\x0f\x00\x00\x02\b" + // 0x0049030F: 0x00000208 + "\x00i\x03\x0f\x00\x00\x02\t" + // 0x0069030F: 0x00000209 + "\x00I\x03\x11\x00\x00\x02\n" + // 0x00490311: 0x0000020A + "\x00i\x03\x11\x00\x00\x02\v" + // 0x00690311: 0x0000020B + "\x00O\x03\x0f\x00\x00\x02\f" + // 0x004F030F: 0x0000020C + "\x00o\x03\x0f\x00\x00\x02\r" + // 0x006F030F: 0x0000020D + "\x00O\x03\x11\x00\x00\x02\x0e" + // 0x004F0311: 0x0000020E + "\x00o\x03\x11\x00\x00\x02\x0f" + // 0x006F0311: 0x0000020F + "\x00R\x03\x0f\x00\x00\x02\x10" + // 0x0052030F: 0x00000210 + "\x00r\x03\x0f\x00\x00\x02\x11" + // 0x0072030F: 0x00000211 + "\x00R\x03\x11\x00\x00\x02\x12" + // 0x00520311: 0x00000212 + "\x00r\x03\x11\x00\x00\x02\x13" + // 0x00720311: 0x00000213 + "\x00U\x03\x0f\x00\x00\x02\x14" + // 0x0055030F: 0x00000214 + "\x00u\x03\x0f\x00\x00\x02\x15" + // 0x0075030F: 0x00000215 + "\x00U\x03\x11\x00\x00\x02\x16" + // 0x00550311: 0x00000216 + "\x00u\x03\x11\x00\x00\x02\x17" + // 0x00750311: 0x00000217 + "\x00S\x03&\x00\x00\x02\x18" + // 0x00530326: 0x00000218 + "\x00s\x03&\x00\x00\x02\x19" + // 0x00730326: 0x00000219 + "\x00T\x03&\x00\x00\x02\x1a" + // 0x00540326: 0x0000021A + "\x00t\x03&\x00\x00\x02\x1b" + // 0x00740326: 0x0000021B + "\x00H\x03\f\x00\x00\x02\x1e" + // 0x0048030C: 0x0000021E + "\x00h\x03\f\x00\x00\x02\x1f" + // 0x0068030C: 0x0000021F + "\x00A\x03\a\x00\x00\x02&" + // 0x00410307: 0x00000226 + "\x00a\x03\a\x00\x00\x02'" + // 0x00610307: 0x00000227 + "\x00E\x03'\x00\x00\x02(" + // 0x00450327: 0x00000228 + "\x00e\x03'\x00\x00\x02)" + // 0x00650327: 0x00000229 + "\x00\xd6\x03\x04\x00\x00\x02*" + // 0x00D60304: 0x0000022A + "\x00\xf6\x03\x04\x00\x00\x02+" + // 0x00F60304: 0x0000022B + "\x00\xd5\x03\x04\x00\x00\x02," + // 0x00D50304: 0x0000022C + "\x00\xf5\x03\x04\x00\x00\x02-" + // 0x00F50304: 0x0000022D + "\x00O\x03\a\x00\x00\x02." + // 0x004F0307: 0x0000022E + "\x00o\x03\a\x00\x00\x02/" + // 0x006F0307: 0x0000022F + "\x02.\x03\x04\x00\x00\x020" + // 0x022E0304: 0x00000230 + "\x02/\x03\x04\x00\x00\x021" + // 0x022F0304: 0x00000231 + "\x00Y\x03\x04\x00\x00\x022" + // 0x00590304: 0x00000232 + "\x00y\x03\x04\x00\x00\x023" + // 0x00790304: 0x00000233 + "\x00\xa8\x03\x01\x00\x00\x03\x85" + // 0x00A80301: 0x00000385 + "\x03\x91\x03\x01\x00\x00\x03\x86" + // 0x03910301: 0x00000386 + "\x03\x95\x03\x01\x00\x00\x03\x88" + // 0x03950301: 0x00000388 + "\x03\x97\x03\x01\x00\x00\x03\x89" + // 0x03970301: 0x00000389 + "\x03\x99\x03\x01\x00\x00\x03\x8a" + // 0x03990301: 0x0000038A + "\x03\x9f\x03\x01\x00\x00\x03\x8c" + // 0x039F0301: 0x0000038C + "\x03\xa5\x03\x01\x00\x00\x03\x8e" + // 0x03A50301: 0x0000038E + "\x03\xa9\x03\x01\x00\x00\x03\x8f" + // 0x03A90301: 0x0000038F + "\x03\xca\x03\x01\x00\x00\x03\x90" + // 0x03CA0301: 0x00000390 + "\x03\x99\x03\b\x00\x00\x03\xaa" + // 0x03990308: 0x000003AA + "\x03\xa5\x03\b\x00\x00\x03\xab" + // 0x03A50308: 0x000003AB + "\x03\xb1\x03\x01\x00\x00\x03\xac" + // 0x03B10301: 0x000003AC + "\x03\xb5\x03\x01\x00\x00\x03\xad" + // 0x03B50301: 0x000003AD + "\x03\xb7\x03\x01\x00\x00\x03\xae" + // 0x03B70301: 0x000003AE + "\x03\xb9\x03\x01\x00\x00\x03\xaf" + // 0x03B90301: 0x000003AF + "\x03\xcb\x03\x01\x00\x00\x03\xb0" + // 0x03CB0301: 0x000003B0 + "\x03\xb9\x03\b\x00\x00\x03\xca" + // 0x03B90308: 0x000003CA + "\x03\xc5\x03\b\x00\x00\x03\xcb" + // 0x03C50308: 0x000003CB + "\x03\xbf\x03\x01\x00\x00\x03\xcc" + // 0x03BF0301: 0x000003CC + "\x03\xc5\x03\x01\x00\x00\x03\xcd" + // 0x03C50301: 0x000003CD + "\x03\xc9\x03\x01\x00\x00\x03\xce" + // 0x03C90301: 0x000003CE + "\x03\xd2\x03\x01\x00\x00\x03\xd3" + // 0x03D20301: 0x000003D3 + "\x03\xd2\x03\b\x00\x00\x03\xd4" + // 0x03D20308: 0x000003D4 + "\x04\x15\x03\x00\x00\x00\x04\x00" + // 0x04150300: 0x00000400 + "\x04\x15\x03\b\x00\x00\x04\x01" + // 0x04150308: 0x00000401 + "\x04\x13\x03\x01\x00\x00\x04\x03" + // 0x04130301: 0x00000403 + "\x04\x06\x03\b\x00\x00\x04\a" + // 0x04060308: 0x00000407 + "\x04\x1a\x03\x01\x00\x00\x04\f" + // 0x041A0301: 0x0000040C + "\x04\x18\x03\x00\x00\x00\x04\r" + // 0x04180300: 0x0000040D + "\x04#\x03\x06\x00\x00\x04\x0e" + // 0x04230306: 0x0000040E + "\x04\x18\x03\x06\x00\x00\x04\x19" + // 0x04180306: 0x00000419 + "\x048\x03\x06\x00\x00\x049" + // 0x04380306: 0x00000439 + "\x045\x03\x00\x00\x00\x04P" + // 0x04350300: 0x00000450 + "\x045\x03\b\x00\x00\x04Q" + // 0x04350308: 0x00000451 + "\x043\x03\x01\x00\x00\x04S" + // 0x04330301: 0x00000453 + "\x04V\x03\b\x00\x00\x04W" + // 0x04560308: 0x00000457 + "\x04:\x03\x01\x00\x00\x04\\" + // 0x043A0301: 0x0000045C + "\x048\x03\x00\x00\x00\x04]" + // 0x04380300: 0x0000045D + "\x04C\x03\x06\x00\x00\x04^" + // 0x04430306: 0x0000045E + "\x04t\x03\x0f\x00\x00\x04v" + // 0x0474030F: 0x00000476 + "\x04u\x03\x0f\x00\x00\x04w" + // 0x0475030F: 0x00000477 + "\x04\x16\x03\x06\x00\x00\x04\xc1" + // 0x04160306: 0x000004C1 + "\x046\x03\x06\x00\x00\x04\xc2" + // 0x04360306: 0x000004C2 + "\x04\x10\x03\x06\x00\x00\x04\xd0" + // 0x04100306: 0x000004D0 + "\x040\x03\x06\x00\x00\x04\xd1" + // 0x04300306: 0x000004D1 + "\x04\x10\x03\b\x00\x00\x04\xd2" + // 0x04100308: 0x000004D2 + "\x040\x03\b\x00\x00\x04\xd3" + // 0x04300308: 0x000004D3 + "\x04\x15\x03\x06\x00\x00\x04\xd6" + // 0x04150306: 0x000004D6 + "\x045\x03\x06\x00\x00\x04\xd7" + // 0x04350306: 0x000004D7 + "\x04\xd8\x03\b\x00\x00\x04\xda" + // 0x04D80308: 0x000004DA + "\x04\xd9\x03\b\x00\x00\x04\xdb" + // 0x04D90308: 0x000004DB + "\x04\x16\x03\b\x00\x00\x04\xdc" + // 0x04160308: 0x000004DC + "\x046\x03\b\x00\x00\x04\xdd" + // 0x04360308: 0x000004DD + "\x04\x17\x03\b\x00\x00\x04\xde" + // 0x04170308: 0x000004DE + "\x047\x03\b\x00\x00\x04\xdf" + // 0x04370308: 0x000004DF + "\x04\x18\x03\x04\x00\x00\x04\xe2" + // 0x04180304: 0x000004E2 + "\x048\x03\x04\x00\x00\x04\xe3" + // 0x04380304: 0x000004E3 + "\x04\x18\x03\b\x00\x00\x04\xe4" + // 0x04180308: 0x000004E4 + "\x048\x03\b\x00\x00\x04\xe5" + // 0x04380308: 0x000004E5 + "\x04\x1e\x03\b\x00\x00\x04\xe6" + // 0x041E0308: 0x000004E6 + "\x04>\x03\b\x00\x00\x04\xe7" + // 0x043E0308: 0x000004E7 + "\x04\xe8\x03\b\x00\x00\x04\xea" + // 0x04E80308: 0x000004EA + "\x04\xe9\x03\b\x00\x00\x04\xeb" + // 0x04E90308: 0x000004EB + "\x04-\x03\b\x00\x00\x04\xec" + // 0x042D0308: 0x000004EC + "\x04M\x03\b\x00\x00\x04\xed" + // 0x044D0308: 0x000004ED + "\x04#\x03\x04\x00\x00\x04\xee" + // 0x04230304: 0x000004EE + "\x04C\x03\x04\x00\x00\x04\xef" + // 0x04430304: 0x000004EF + "\x04#\x03\b\x00\x00\x04\xf0" + // 0x04230308: 0x000004F0 + "\x04C\x03\b\x00\x00\x04\xf1" + // 0x04430308: 0x000004F1 + "\x04#\x03\v\x00\x00\x04\xf2" + // 0x0423030B: 0x000004F2 + "\x04C\x03\v\x00\x00\x04\xf3" + // 0x0443030B: 0x000004F3 + "\x04'\x03\b\x00\x00\x04\xf4" + // 0x04270308: 0x000004F4 + "\x04G\x03\b\x00\x00\x04\xf5" + // 0x04470308: 0x000004F5 + "\x04+\x03\b\x00\x00\x04\xf8" + // 0x042B0308: 0x000004F8 + "\x04K\x03\b\x00\x00\x04\xf9" + // 0x044B0308: 0x000004F9 + "\x06'\x06S\x00\x00\x06\"" + // 0x06270653: 0x00000622 + "\x06'\x06T\x00\x00\x06#" + // 0x06270654: 0x00000623 + "\x06H\x06T\x00\x00\x06$" + // 0x06480654: 0x00000624 + "\x06'\x06U\x00\x00\x06%" + // 0x06270655: 0x00000625 + "\x06J\x06T\x00\x00\x06&" + // 0x064A0654: 0x00000626 + "\x06\xd5\x06T\x00\x00\x06\xc0" + // 0x06D50654: 0x000006C0 + "\x06\xc1\x06T\x00\x00\x06\xc2" + // 0x06C10654: 0x000006C2 + "\x06\xd2\x06T\x00\x00\x06\xd3" + // 0x06D20654: 0x000006D3 + "\t(\t<\x00\x00\t)" + // 0x0928093C: 0x00000929 + "\t0\t<\x00\x00\t1" + // 0x0930093C: 0x00000931 + "\t3\t<\x00\x00\t4" + // 0x0933093C: 0x00000934 + "\t\xc7\t\xbe\x00\x00\t\xcb" + // 0x09C709BE: 0x000009CB + "\t\xc7\t\xd7\x00\x00\t\xcc" + // 0x09C709D7: 0x000009CC + "\vG\vV\x00\x00\vH" + // 0x0B470B56: 0x00000B48 + "\vG\v>\x00\x00\vK" + // 0x0B470B3E: 0x00000B4B + "\vG\vW\x00\x00\vL" + // 0x0B470B57: 0x00000B4C + "\v\x92\v\xd7\x00\x00\v\x94" + // 0x0B920BD7: 0x00000B94 + "\v\xc6\v\xbe\x00\x00\v\xca" + // 0x0BC60BBE: 0x00000BCA + "\v\xc7\v\xbe\x00\x00\v\xcb" + // 0x0BC70BBE: 0x00000BCB + "\v\xc6\v\xd7\x00\x00\v\xcc" + // 0x0BC60BD7: 0x00000BCC + "\fF\fV\x00\x00\fH" + // 0x0C460C56: 0x00000C48 + "\f\xbf\f\xd5\x00\x00\f\xc0" + // 0x0CBF0CD5: 0x00000CC0 + "\f\xc6\f\xd5\x00\x00\f\xc7" + // 0x0CC60CD5: 0x00000CC7 + "\f\xc6\f\xd6\x00\x00\f\xc8" + // 0x0CC60CD6: 0x00000CC8 + "\f\xc6\f\xc2\x00\x00\f\xca" + // 0x0CC60CC2: 0x00000CCA + "\f\xca\f\xd5\x00\x00\f\xcb" + // 0x0CCA0CD5: 0x00000CCB + "\rF\r>\x00\x00\rJ" + // 0x0D460D3E: 0x00000D4A + "\rG\r>\x00\x00\rK" + // 0x0D470D3E: 0x00000D4B + "\rF\rW\x00\x00\rL" + // 0x0D460D57: 0x00000D4C + "\r\xd9\r\xca\x00\x00\r\xda" + // 0x0DD90DCA: 0x00000DDA + "\r\xd9\r\xcf\x00\x00\r\xdc" + // 0x0DD90DCF: 0x00000DDC + "\r\xdc\r\xca\x00\x00\r\xdd" + // 0x0DDC0DCA: 0x00000DDD + "\r\xd9\r\xdf\x00\x00\r\xde" + // 0x0DD90DDF: 0x00000DDE + "\x10%\x10.\x00\x00\x10&" + // 0x1025102E: 0x00001026 + "\x1b\x05\x1b5\x00\x00\x1b\x06" + // 0x1B051B35: 0x00001B06 + "\x1b\a\x1b5\x00\x00\x1b\b" + // 0x1B071B35: 0x00001B08 + "\x1b\t\x1b5\x00\x00\x1b\n" + // 0x1B091B35: 0x00001B0A + "\x1b\v\x1b5\x00\x00\x1b\f" + // 0x1B0B1B35: 0x00001B0C + "\x1b\r\x1b5\x00\x00\x1b\x0e" + // 0x1B0D1B35: 0x00001B0E + "\x1b\x11\x1b5\x00\x00\x1b\x12" + // 0x1B111B35: 0x00001B12 + "\x1b:\x1b5\x00\x00\x1b;" + // 0x1B3A1B35: 0x00001B3B + "\x1b<\x1b5\x00\x00\x1b=" + // 0x1B3C1B35: 0x00001B3D + "\x1b>\x1b5\x00\x00\x1b@" + // 0x1B3E1B35: 0x00001B40 + "\x1b?\x1b5\x00\x00\x1bA" + // 0x1B3F1B35: 0x00001B41 + "\x1bB\x1b5\x00\x00\x1bC" + // 0x1B421B35: 0x00001B43 + "\x00A\x03%\x00\x00\x1e\x00" + // 0x00410325: 0x00001E00 + "\x00a\x03%\x00\x00\x1e\x01" + // 0x00610325: 0x00001E01 + "\x00B\x03\a\x00\x00\x1e\x02" + // 0x00420307: 0x00001E02 + "\x00b\x03\a\x00\x00\x1e\x03" + // 0x00620307: 0x00001E03 + "\x00B\x03#\x00\x00\x1e\x04" + // 0x00420323: 0x00001E04 + "\x00b\x03#\x00\x00\x1e\x05" + // 0x00620323: 0x00001E05 + "\x00B\x031\x00\x00\x1e\x06" + // 0x00420331: 0x00001E06 + "\x00b\x031\x00\x00\x1e\a" + // 0x00620331: 0x00001E07 + "\x00\xc7\x03\x01\x00\x00\x1e\b" + // 0x00C70301: 0x00001E08 + "\x00\xe7\x03\x01\x00\x00\x1e\t" + // 0x00E70301: 0x00001E09 + "\x00D\x03\a\x00\x00\x1e\n" + // 0x00440307: 0x00001E0A + "\x00d\x03\a\x00\x00\x1e\v" + // 0x00640307: 0x00001E0B + "\x00D\x03#\x00\x00\x1e\f" + // 0x00440323: 0x00001E0C + "\x00d\x03#\x00\x00\x1e\r" + // 0x00640323: 0x00001E0D + "\x00D\x031\x00\x00\x1e\x0e" + // 0x00440331: 0x00001E0E + "\x00d\x031\x00\x00\x1e\x0f" + // 0x00640331: 0x00001E0F + "\x00D\x03'\x00\x00\x1e\x10" + // 0x00440327: 0x00001E10 + "\x00d\x03'\x00\x00\x1e\x11" + // 0x00640327: 0x00001E11 + "\x00D\x03-\x00\x00\x1e\x12" + // 0x0044032D: 0x00001E12 + "\x00d\x03-\x00\x00\x1e\x13" + // 0x0064032D: 0x00001E13 + "\x01\x12\x03\x00\x00\x00\x1e\x14" + // 0x01120300: 0x00001E14 + "\x01\x13\x03\x00\x00\x00\x1e\x15" + // 0x01130300: 0x00001E15 + "\x01\x12\x03\x01\x00\x00\x1e\x16" + // 0x01120301: 0x00001E16 + "\x01\x13\x03\x01\x00\x00\x1e\x17" + // 0x01130301: 0x00001E17 + "\x00E\x03-\x00\x00\x1e\x18" + // 0x0045032D: 0x00001E18 + "\x00e\x03-\x00\x00\x1e\x19" + // 0x0065032D: 0x00001E19 + "\x00E\x030\x00\x00\x1e\x1a" + // 0x00450330: 0x00001E1A + "\x00e\x030\x00\x00\x1e\x1b" + // 0x00650330: 0x00001E1B + "\x02(\x03\x06\x00\x00\x1e\x1c" + // 0x02280306: 0x00001E1C + "\x02)\x03\x06\x00\x00\x1e\x1d" + // 0x02290306: 0x00001E1D + "\x00F\x03\a\x00\x00\x1e\x1e" + // 0x00460307: 0x00001E1E + "\x00f\x03\a\x00\x00\x1e\x1f" + // 0x00660307: 0x00001E1F + "\x00G\x03\x04\x00\x00\x1e " + // 0x00470304: 0x00001E20 + "\x00g\x03\x04\x00\x00\x1e!" + // 0x00670304: 0x00001E21 + "\x00H\x03\a\x00\x00\x1e\"" + // 0x00480307: 0x00001E22 + "\x00h\x03\a\x00\x00\x1e#" + // 0x00680307: 0x00001E23 + "\x00H\x03#\x00\x00\x1e$" + // 0x00480323: 0x00001E24 + "\x00h\x03#\x00\x00\x1e%" + // 0x00680323: 0x00001E25 + "\x00H\x03\b\x00\x00\x1e&" + // 0x00480308: 0x00001E26 + "\x00h\x03\b\x00\x00\x1e'" + // 0x00680308: 0x00001E27 + "\x00H\x03'\x00\x00\x1e(" + // 0x00480327: 0x00001E28 + "\x00h\x03'\x00\x00\x1e)" + // 0x00680327: 0x00001E29 + "\x00H\x03.\x00\x00\x1e*" + // 0x0048032E: 0x00001E2A + "\x00h\x03.\x00\x00\x1e+" + // 0x0068032E: 0x00001E2B + "\x00I\x030\x00\x00\x1e," + // 0x00490330: 0x00001E2C + "\x00i\x030\x00\x00\x1e-" + // 0x00690330: 0x00001E2D + "\x00\xcf\x03\x01\x00\x00\x1e." + // 0x00CF0301: 0x00001E2E + "\x00\xef\x03\x01\x00\x00\x1e/" + // 0x00EF0301: 0x00001E2F + "\x00K\x03\x01\x00\x00\x1e0" + // 0x004B0301: 0x00001E30 + "\x00k\x03\x01\x00\x00\x1e1" + // 0x006B0301: 0x00001E31 + "\x00K\x03#\x00\x00\x1e2" + // 0x004B0323: 0x00001E32 + "\x00k\x03#\x00\x00\x1e3" + // 0x006B0323: 0x00001E33 + "\x00K\x031\x00\x00\x1e4" + // 0x004B0331: 0x00001E34 + "\x00k\x031\x00\x00\x1e5" + // 0x006B0331: 0x00001E35 + "\x00L\x03#\x00\x00\x1e6" + // 0x004C0323: 0x00001E36 + "\x00l\x03#\x00\x00\x1e7" + // 0x006C0323: 0x00001E37 + "\x1e6\x03\x04\x00\x00\x1e8" + // 0x1E360304: 0x00001E38 + "\x1e7\x03\x04\x00\x00\x1e9" + // 0x1E370304: 0x00001E39 + "\x00L\x031\x00\x00\x1e:" + // 0x004C0331: 0x00001E3A + "\x00l\x031\x00\x00\x1e;" + // 0x006C0331: 0x00001E3B + "\x00L\x03-\x00\x00\x1e<" + // 0x004C032D: 0x00001E3C + "\x00l\x03-\x00\x00\x1e=" + // 0x006C032D: 0x00001E3D + "\x00M\x03\x01\x00\x00\x1e>" + // 0x004D0301: 0x00001E3E + "\x00m\x03\x01\x00\x00\x1e?" + // 0x006D0301: 0x00001E3F + "\x00M\x03\a\x00\x00\x1e@" + // 0x004D0307: 0x00001E40 + "\x00m\x03\a\x00\x00\x1eA" + // 0x006D0307: 0x00001E41 + "\x00M\x03#\x00\x00\x1eB" + // 0x004D0323: 0x00001E42 + "\x00m\x03#\x00\x00\x1eC" + // 0x006D0323: 0x00001E43 + "\x00N\x03\a\x00\x00\x1eD" + // 0x004E0307: 0x00001E44 + "\x00n\x03\a\x00\x00\x1eE" + // 0x006E0307: 0x00001E45 + "\x00N\x03#\x00\x00\x1eF" + // 0x004E0323: 0x00001E46 + "\x00n\x03#\x00\x00\x1eG" + // 0x006E0323: 0x00001E47 + "\x00N\x031\x00\x00\x1eH" + // 0x004E0331: 0x00001E48 + "\x00n\x031\x00\x00\x1eI" + // 0x006E0331: 0x00001E49 + "\x00N\x03-\x00\x00\x1eJ" + // 0x004E032D: 0x00001E4A + "\x00n\x03-\x00\x00\x1eK" + // 0x006E032D: 0x00001E4B + "\x00\xd5\x03\x01\x00\x00\x1eL" + // 0x00D50301: 0x00001E4C + "\x00\xf5\x03\x01\x00\x00\x1eM" + // 0x00F50301: 0x00001E4D + "\x00\xd5\x03\b\x00\x00\x1eN" + // 0x00D50308: 0x00001E4E + "\x00\xf5\x03\b\x00\x00\x1eO" + // 0x00F50308: 0x00001E4F + "\x01L\x03\x00\x00\x00\x1eP" + // 0x014C0300: 0x00001E50 + "\x01M\x03\x00\x00\x00\x1eQ" + // 0x014D0300: 0x00001E51 + "\x01L\x03\x01\x00\x00\x1eR" + // 0x014C0301: 0x00001E52 + "\x01M\x03\x01\x00\x00\x1eS" + // 0x014D0301: 0x00001E53 + "\x00P\x03\x01\x00\x00\x1eT" + // 0x00500301: 0x00001E54 + "\x00p\x03\x01\x00\x00\x1eU" + // 0x00700301: 0x00001E55 + "\x00P\x03\a\x00\x00\x1eV" + // 0x00500307: 0x00001E56 + "\x00p\x03\a\x00\x00\x1eW" + // 0x00700307: 0x00001E57 + "\x00R\x03\a\x00\x00\x1eX" + // 0x00520307: 0x00001E58 + "\x00r\x03\a\x00\x00\x1eY" + // 0x00720307: 0x00001E59 + "\x00R\x03#\x00\x00\x1eZ" + // 0x00520323: 0x00001E5A + "\x00r\x03#\x00\x00\x1e[" + // 0x00720323: 0x00001E5B + "\x1eZ\x03\x04\x00\x00\x1e\\" + // 0x1E5A0304: 0x00001E5C + "\x1e[\x03\x04\x00\x00\x1e]" + // 0x1E5B0304: 0x00001E5D + "\x00R\x031\x00\x00\x1e^" + // 0x00520331: 0x00001E5E + "\x00r\x031\x00\x00\x1e_" + // 0x00720331: 0x00001E5F + "\x00S\x03\a\x00\x00\x1e`" + // 0x00530307: 0x00001E60 + "\x00s\x03\a\x00\x00\x1ea" + // 0x00730307: 0x00001E61 + "\x00S\x03#\x00\x00\x1eb" + // 0x00530323: 0x00001E62 + "\x00s\x03#\x00\x00\x1ec" + // 0x00730323: 0x00001E63 + "\x01Z\x03\a\x00\x00\x1ed" + // 0x015A0307: 0x00001E64 + "\x01[\x03\a\x00\x00\x1ee" + // 0x015B0307: 0x00001E65 + "\x01`\x03\a\x00\x00\x1ef" + // 0x01600307: 0x00001E66 + "\x01a\x03\a\x00\x00\x1eg" + // 0x01610307: 0x00001E67 + "\x1eb\x03\a\x00\x00\x1eh" + // 0x1E620307: 0x00001E68 + "\x1ec\x03\a\x00\x00\x1ei" + // 0x1E630307: 0x00001E69 + "\x00T\x03\a\x00\x00\x1ej" + // 0x00540307: 0x00001E6A + "\x00t\x03\a\x00\x00\x1ek" + // 0x00740307: 0x00001E6B + "\x00T\x03#\x00\x00\x1el" + // 0x00540323: 0x00001E6C + "\x00t\x03#\x00\x00\x1em" + // 0x00740323: 0x00001E6D + "\x00T\x031\x00\x00\x1en" + // 0x00540331: 0x00001E6E + "\x00t\x031\x00\x00\x1eo" + // 0x00740331: 0x00001E6F + "\x00T\x03-\x00\x00\x1ep" + // 0x0054032D: 0x00001E70 + "\x00t\x03-\x00\x00\x1eq" + // 0x0074032D: 0x00001E71 + "\x00U\x03$\x00\x00\x1er" + // 0x00550324: 0x00001E72 + "\x00u\x03$\x00\x00\x1es" + // 0x00750324: 0x00001E73 + "\x00U\x030\x00\x00\x1et" + // 0x00550330: 0x00001E74 + "\x00u\x030\x00\x00\x1eu" + // 0x00750330: 0x00001E75 + "\x00U\x03-\x00\x00\x1ev" + // 0x0055032D: 0x00001E76 + "\x00u\x03-\x00\x00\x1ew" + // 0x0075032D: 0x00001E77 + "\x01h\x03\x01\x00\x00\x1ex" + // 0x01680301: 0x00001E78 + "\x01i\x03\x01\x00\x00\x1ey" + // 0x01690301: 0x00001E79 + "\x01j\x03\b\x00\x00\x1ez" + // 0x016A0308: 0x00001E7A + "\x01k\x03\b\x00\x00\x1e{" + // 0x016B0308: 0x00001E7B + "\x00V\x03\x03\x00\x00\x1e|" + // 0x00560303: 0x00001E7C + "\x00v\x03\x03\x00\x00\x1e}" + // 0x00760303: 0x00001E7D + "\x00V\x03#\x00\x00\x1e~" + // 0x00560323: 0x00001E7E + "\x00v\x03#\x00\x00\x1e\u007f" + // 0x00760323: 0x00001E7F + "\x00W\x03\x00\x00\x00\x1e\x80" + // 0x00570300: 0x00001E80 + "\x00w\x03\x00\x00\x00\x1e\x81" + // 0x00770300: 0x00001E81 + "\x00W\x03\x01\x00\x00\x1e\x82" + // 0x00570301: 0x00001E82 + "\x00w\x03\x01\x00\x00\x1e\x83" + // 0x00770301: 0x00001E83 + "\x00W\x03\b\x00\x00\x1e\x84" + // 0x00570308: 0x00001E84 + "\x00w\x03\b\x00\x00\x1e\x85" + // 0x00770308: 0x00001E85 + "\x00W\x03\a\x00\x00\x1e\x86" + // 0x00570307: 0x00001E86 + "\x00w\x03\a\x00\x00\x1e\x87" + // 0x00770307: 0x00001E87 + "\x00W\x03#\x00\x00\x1e\x88" + // 0x00570323: 0x00001E88 + "\x00w\x03#\x00\x00\x1e\x89" + // 0x00770323: 0x00001E89 + "\x00X\x03\a\x00\x00\x1e\x8a" + // 0x00580307: 0x00001E8A + "\x00x\x03\a\x00\x00\x1e\x8b" + // 0x00780307: 0x00001E8B + "\x00X\x03\b\x00\x00\x1e\x8c" + // 0x00580308: 0x00001E8C + "\x00x\x03\b\x00\x00\x1e\x8d" + // 0x00780308: 0x00001E8D + "\x00Y\x03\a\x00\x00\x1e\x8e" + // 0x00590307: 0x00001E8E + "\x00y\x03\a\x00\x00\x1e\x8f" + // 0x00790307: 0x00001E8F + "\x00Z\x03\x02\x00\x00\x1e\x90" + // 0x005A0302: 0x00001E90 + "\x00z\x03\x02\x00\x00\x1e\x91" + // 0x007A0302: 0x00001E91 + "\x00Z\x03#\x00\x00\x1e\x92" + // 0x005A0323: 0x00001E92 + "\x00z\x03#\x00\x00\x1e\x93" + // 0x007A0323: 0x00001E93 + "\x00Z\x031\x00\x00\x1e\x94" + // 0x005A0331: 0x00001E94 + "\x00z\x031\x00\x00\x1e\x95" + // 0x007A0331: 0x00001E95 + "\x00h\x031\x00\x00\x1e\x96" + // 0x00680331: 0x00001E96 + "\x00t\x03\b\x00\x00\x1e\x97" + // 0x00740308: 0x00001E97 + "\x00w\x03\n\x00\x00\x1e\x98" + // 0x0077030A: 0x00001E98 + "\x00y\x03\n\x00\x00\x1e\x99" + // 0x0079030A: 0x00001E99 + "\x01\u007f\x03\a\x00\x00\x1e\x9b" + // 0x017F0307: 0x00001E9B + "\x00A\x03#\x00\x00\x1e\xa0" + // 0x00410323: 0x00001EA0 + "\x00a\x03#\x00\x00\x1e\xa1" + // 0x00610323: 0x00001EA1 + "\x00A\x03\t\x00\x00\x1e\xa2" + // 0x00410309: 0x00001EA2 + "\x00a\x03\t\x00\x00\x1e\xa3" + // 0x00610309: 0x00001EA3 + "\x00\xc2\x03\x01\x00\x00\x1e\xa4" + // 0x00C20301: 0x00001EA4 + "\x00\xe2\x03\x01\x00\x00\x1e\xa5" + // 0x00E20301: 0x00001EA5 + "\x00\xc2\x03\x00\x00\x00\x1e\xa6" + // 0x00C20300: 0x00001EA6 + "\x00\xe2\x03\x00\x00\x00\x1e\xa7" + // 0x00E20300: 0x00001EA7 + "\x00\xc2\x03\t\x00\x00\x1e\xa8" + // 0x00C20309: 0x00001EA8 + "\x00\xe2\x03\t\x00\x00\x1e\xa9" + // 0x00E20309: 0x00001EA9 + "\x00\xc2\x03\x03\x00\x00\x1e\xaa" + // 0x00C20303: 0x00001EAA + "\x00\xe2\x03\x03\x00\x00\x1e\xab" + // 0x00E20303: 0x00001EAB + "\x1e\xa0\x03\x02\x00\x00\x1e\xac" + // 0x1EA00302: 0x00001EAC + "\x1e\xa1\x03\x02\x00\x00\x1e\xad" + // 0x1EA10302: 0x00001EAD + "\x01\x02\x03\x01\x00\x00\x1e\xae" + // 0x01020301: 0x00001EAE + "\x01\x03\x03\x01\x00\x00\x1e\xaf" + // 0x01030301: 0x00001EAF + "\x01\x02\x03\x00\x00\x00\x1e\xb0" + // 0x01020300: 0x00001EB0 + "\x01\x03\x03\x00\x00\x00\x1e\xb1" + // 0x01030300: 0x00001EB1 + "\x01\x02\x03\t\x00\x00\x1e\xb2" + // 0x01020309: 0x00001EB2 + "\x01\x03\x03\t\x00\x00\x1e\xb3" + // 0x01030309: 0x00001EB3 + "\x01\x02\x03\x03\x00\x00\x1e\xb4" + // 0x01020303: 0x00001EB4 + "\x01\x03\x03\x03\x00\x00\x1e\xb5" + // 0x01030303: 0x00001EB5 + "\x1e\xa0\x03\x06\x00\x00\x1e\xb6" + // 0x1EA00306: 0x00001EB6 + "\x1e\xa1\x03\x06\x00\x00\x1e\xb7" + // 0x1EA10306: 0x00001EB7 + "\x00E\x03#\x00\x00\x1e\xb8" + // 0x00450323: 0x00001EB8 + "\x00e\x03#\x00\x00\x1e\xb9" + // 0x00650323: 0x00001EB9 + "\x00E\x03\t\x00\x00\x1e\xba" + // 0x00450309: 0x00001EBA + "\x00e\x03\t\x00\x00\x1e\xbb" + // 0x00650309: 0x00001EBB + "\x00E\x03\x03\x00\x00\x1e\xbc" + // 0x00450303: 0x00001EBC + "\x00e\x03\x03\x00\x00\x1e\xbd" + // 0x00650303: 0x00001EBD + "\x00\xca\x03\x01\x00\x00\x1e\xbe" + // 0x00CA0301: 0x00001EBE + "\x00\xea\x03\x01\x00\x00\x1e\xbf" + // 0x00EA0301: 0x00001EBF + "\x00\xca\x03\x00\x00\x00\x1e\xc0" + // 0x00CA0300: 0x00001EC0 + "\x00\xea\x03\x00\x00\x00\x1e\xc1" + // 0x00EA0300: 0x00001EC1 + "\x00\xca\x03\t\x00\x00\x1e\xc2" + // 0x00CA0309: 0x00001EC2 + "\x00\xea\x03\t\x00\x00\x1e\xc3" + // 0x00EA0309: 0x00001EC3 + "\x00\xca\x03\x03\x00\x00\x1e\xc4" + // 0x00CA0303: 0x00001EC4 + "\x00\xea\x03\x03\x00\x00\x1e\xc5" + // 0x00EA0303: 0x00001EC5 + "\x1e\xb8\x03\x02\x00\x00\x1e\xc6" + // 0x1EB80302: 0x00001EC6 + "\x1e\xb9\x03\x02\x00\x00\x1e\xc7" + // 0x1EB90302: 0x00001EC7 + "\x00I\x03\t\x00\x00\x1e\xc8" + // 0x00490309: 0x00001EC8 + "\x00i\x03\t\x00\x00\x1e\xc9" + // 0x00690309: 0x00001EC9 + "\x00I\x03#\x00\x00\x1e\xca" + // 0x00490323: 0x00001ECA + "\x00i\x03#\x00\x00\x1e\xcb" + // 0x00690323: 0x00001ECB + "\x00O\x03#\x00\x00\x1e\xcc" + // 0x004F0323: 0x00001ECC + "\x00o\x03#\x00\x00\x1e\xcd" + // 0x006F0323: 0x00001ECD + "\x00O\x03\t\x00\x00\x1e\xce" + // 0x004F0309: 0x00001ECE + "\x00o\x03\t\x00\x00\x1e\xcf" + // 0x006F0309: 0x00001ECF + "\x00\xd4\x03\x01\x00\x00\x1e\xd0" + // 0x00D40301: 0x00001ED0 + "\x00\xf4\x03\x01\x00\x00\x1e\xd1" + // 0x00F40301: 0x00001ED1 + "\x00\xd4\x03\x00\x00\x00\x1e\xd2" + // 0x00D40300: 0x00001ED2 + "\x00\xf4\x03\x00\x00\x00\x1e\xd3" + // 0x00F40300: 0x00001ED3 + "\x00\xd4\x03\t\x00\x00\x1e\xd4" + // 0x00D40309: 0x00001ED4 + "\x00\xf4\x03\t\x00\x00\x1e\xd5" + // 0x00F40309: 0x00001ED5 + "\x00\xd4\x03\x03\x00\x00\x1e\xd6" + // 0x00D40303: 0x00001ED6 + "\x00\xf4\x03\x03\x00\x00\x1e\xd7" + // 0x00F40303: 0x00001ED7 + "\x1e\xcc\x03\x02\x00\x00\x1e\xd8" + // 0x1ECC0302: 0x00001ED8 + "\x1e\xcd\x03\x02\x00\x00\x1e\xd9" + // 0x1ECD0302: 0x00001ED9 + "\x01\xa0\x03\x01\x00\x00\x1e\xda" + // 0x01A00301: 0x00001EDA + "\x01\xa1\x03\x01\x00\x00\x1e\xdb" + // 0x01A10301: 0x00001EDB + "\x01\xa0\x03\x00\x00\x00\x1e\xdc" + // 0x01A00300: 0x00001EDC + "\x01\xa1\x03\x00\x00\x00\x1e\xdd" + // 0x01A10300: 0x00001EDD + "\x01\xa0\x03\t\x00\x00\x1e\xde" + // 0x01A00309: 0x00001EDE + "\x01\xa1\x03\t\x00\x00\x1e\xdf" + // 0x01A10309: 0x00001EDF + "\x01\xa0\x03\x03\x00\x00\x1e\xe0" + // 0x01A00303: 0x00001EE0 + "\x01\xa1\x03\x03\x00\x00\x1e\xe1" + // 0x01A10303: 0x00001EE1 + "\x01\xa0\x03#\x00\x00\x1e\xe2" + // 0x01A00323: 0x00001EE2 + "\x01\xa1\x03#\x00\x00\x1e\xe3" + // 0x01A10323: 0x00001EE3 + "\x00U\x03#\x00\x00\x1e\xe4" + // 0x00550323: 0x00001EE4 + "\x00u\x03#\x00\x00\x1e\xe5" + // 0x00750323: 0x00001EE5 + "\x00U\x03\t\x00\x00\x1e\xe6" + // 0x00550309: 0x00001EE6 + "\x00u\x03\t\x00\x00\x1e\xe7" + // 0x00750309: 0x00001EE7 + "\x01\xaf\x03\x01\x00\x00\x1e\xe8" + // 0x01AF0301: 0x00001EE8 + "\x01\xb0\x03\x01\x00\x00\x1e\xe9" + // 0x01B00301: 0x00001EE9 + "\x01\xaf\x03\x00\x00\x00\x1e\xea" + // 0x01AF0300: 0x00001EEA + "\x01\xb0\x03\x00\x00\x00\x1e\xeb" + // 0x01B00300: 0x00001EEB + "\x01\xaf\x03\t\x00\x00\x1e\xec" + // 0x01AF0309: 0x00001EEC + "\x01\xb0\x03\t\x00\x00\x1e\xed" + // 0x01B00309: 0x00001EED + "\x01\xaf\x03\x03\x00\x00\x1e\xee" + // 0x01AF0303: 0x00001EEE + "\x01\xb0\x03\x03\x00\x00\x1e\xef" + // 0x01B00303: 0x00001EEF + "\x01\xaf\x03#\x00\x00\x1e\xf0" + // 0x01AF0323: 0x00001EF0 + "\x01\xb0\x03#\x00\x00\x1e\xf1" + // 0x01B00323: 0x00001EF1 + "\x00Y\x03\x00\x00\x00\x1e\xf2" + // 0x00590300: 0x00001EF2 + "\x00y\x03\x00\x00\x00\x1e\xf3" + // 0x00790300: 0x00001EF3 + "\x00Y\x03#\x00\x00\x1e\xf4" + // 0x00590323: 0x00001EF4 + "\x00y\x03#\x00\x00\x1e\xf5" + // 0x00790323: 0x00001EF5 + "\x00Y\x03\t\x00\x00\x1e\xf6" + // 0x00590309: 0x00001EF6 + "\x00y\x03\t\x00\x00\x1e\xf7" + // 0x00790309: 0x00001EF7 + "\x00Y\x03\x03\x00\x00\x1e\xf8" + // 0x00590303: 0x00001EF8 + "\x00y\x03\x03\x00\x00\x1e\xf9" + // 0x00790303: 0x00001EF9 + "\x03\xb1\x03\x13\x00\x00\x1f\x00" + // 0x03B10313: 0x00001F00 + "\x03\xb1\x03\x14\x00\x00\x1f\x01" + // 0x03B10314: 0x00001F01 + "\x1f\x00\x03\x00\x00\x00\x1f\x02" + // 0x1F000300: 0x00001F02 + "\x1f\x01\x03\x00\x00\x00\x1f\x03" + // 0x1F010300: 0x00001F03 + "\x1f\x00\x03\x01\x00\x00\x1f\x04" + // 0x1F000301: 0x00001F04 + "\x1f\x01\x03\x01\x00\x00\x1f\x05" + // 0x1F010301: 0x00001F05 + "\x1f\x00\x03B\x00\x00\x1f\x06" + // 0x1F000342: 0x00001F06 + "\x1f\x01\x03B\x00\x00\x1f\a" + // 0x1F010342: 0x00001F07 + "\x03\x91\x03\x13\x00\x00\x1f\b" + // 0x03910313: 0x00001F08 + "\x03\x91\x03\x14\x00\x00\x1f\t" + // 0x03910314: 0x00001F09 + "\x1f\b\x03\x00\x00\x00\x1f\n" + // 0x1F080300: 0x00001F0A + "\x1f\t\x03\x00\x00\x00\x1f\v" + // 0x1F090300: 0x00001F0B + "\x1f\b\x03\x01\x00\x00\x1f\f" + // 0x1F080301: 0x00001F0C + "\x1f\t\x03\x01\x00\x00\x1f\r" + // 0x1F090301: 0x00001F0D + "\x1f\b\x03B\x00\x00\x1f\x0e" + // 0x1F080342: 0x00001F0E + "\x1f\t\x03B\x00\x00\x1f\x0f" + // 0x1F090342: 0x00001F0F + "\x03\xb5\x03\x13\x00\x00\x1f\x10" + // 0x03B50313: 0x00001F10 + "\x03\xb5\x03\x14\x00\x00\x1f\x11" + // 0x03B50314: 0x00001F11 + "\x1f\x10\x03\x00\x00\x00\x1f\x12" + // 0x1F100300: 0x00001F12 + "\x1f\x11\x03\x00\x00\x00\x1f\x13" + // 0x1F110300: 0x00001F13 + "\x1f\x10\x03\x01\x00\x00\x1f\x14" + // 0x1F100301: 0x00001F14 + "\x1f\x11\x03\x01\x00\x00\x1f\x15" + // 0x1F110301: 0x00001F15 + "\x03\x95\x03\x13\x00\x00\x1f\x18" + // 0x03950313: 0x00001F18 + "\x03\x95\x03\x14\x00\x00\x1f\x19" + // 0x03950314: 0x00001F19 + "\x1f\x18\x03\x00\x00\x00\x1f\x1a" + // 0x1F180300: 0x00001F1A + "\x1f\x19\x03\x00\x00\x00\x1f\x1b" + // 0x1F190300: 0x00001F1B + "\x1f\x18\x03\x01\x00\x00\x1f\x1c" + // 0x1F180301: 0x00001F1C + "\x1f\x19\x03\x01\x00\x00\x1f\x1d" + // 0x1F190301: 0x00001F1D + "\x03\xb7\x03\x13\x00\x00\x1f " + // 0x03B70313: 0x00001F20 + "\x03\xb7\x03\x14\x00\x00\x1f!" + // 0x03B70314: 0x00001F21 + "\x1f \x03\x00\x00\x00\x1f\"" + // 0x1F200300: 0x00001F22 + "\x1f!\x03\x00\x00\x00\x1f#" + // 0x1F210300: 0x00001F23 + "\x1f \x03\x01\x00\x00\x1f$" + // 0x1F200301: 0x00001F24 + "\x1f!\x03\x01\x00\x00\x1f%" + // 0x1F210301: 0x00001F25 + "\x1f \x03B\x00\x00\x1f&" + // 0x1F200342: 0x00001F26 + "\x1f!\x03B\x00\x00\x1f'" + // 0x1F210342: 0x00001F27 + "\x03\x97\x03\x13\x00\x00\x1f(" + // 0x03970313: 0x00001F28 + "\x03\x97\x03\x14\x00\x00\x1f)" + // 0x03970314: 0x00001F29 + "\x1f(\x03\x00\x00\x00\x1f*" + // 0x1F280300: 0x00001F2A + "\x1f)\x03\x00\x00\x00\x1f+" + // 0x1F290300: 0x00001F2B + "\x1f(\x03\x01\x00\x00\x1f," + // 0x1F280301: 0x00001F2C + "\x1f)\x03\x01\x00\x00\x1f-" + // 0x1F290301: 0x00001F2D + "\x1f(\x03B\x00\x00\x1f." + // 0x1F280342: 0x00001F2E + "\x1f)\x03B\x00\x00\x1f/" + // 0x1F290342: 0x00001F2F + "\x03\xb9\x03\x13\x00\x00\x1f0" + // 0x03B90313: 0x00001F30 + "\x03\xb9\x03\x14\x00\x00\x1f1" + // 0x03B90314: 0x00001F31 + "\x1f0\x03\x00\x00\x00\x1f2" + // 0x1F300300: 0x00001F32 + "\x1f1\x03\x00\x00\x00\x1f3" + // 0x1F310300: 0x00001F33 + "\x1f0\x03\x01\x00\x00\x1f4" + // 0x1F300301: 0x00001F34 + "\x1f1\x03\x01\x00\x00\x1f5" + // 0x1F310301: 0x00001F35 + "\x1f0\x03B\x00\x00\x1f6" + // 0x1F300342: 0x00001F36 + "\x1f1\x03B\x00\x00\x1f7" + // 0x1F310342: 0x00001F37 + "\x03\x99\x03\x13\x00\x00\x1f8" + // 0x03990313: 0x00001F38 + "\x03\x99\x03\x14\x00\x00\x1f9" + // 0x03990314: 0x00001F39 + "\x1f8\x03\x00\x00\x00\x1f:" + // 0x1F380300: 0x00001F3A + "\x1f9\x03\x00\x00\x00\x1f;" + // 0x1F390300: 0x00001F3B + "\x1f8\x03\x01\x00\x00\x1f<" + // 0x1F380301: 0x00001F3C + "\x1f9\x03\x01\x00\x00\x1f=" + // 0x1F390301: 0x00001F3D + "\x1f8\x03B\x00\x00\x1f>" + // 0x1F380342: 0x00001F3E + "\x1f9\x03B\x00\x00\x1f?" + // 0x1F390342: 0x00001F3F + "\x03\xbf\x03\x13\x00\x00\x1f@" + // 0x03BF0313: 0x00001F40 + "\x03\xbf\x03\x14\x00\x00\x1fA" + // 0x03BF0314: 0x00001F41 + "\x1f@\x03\x00\x00\x00\x1fB" + // 0x1F400300: 0x00001F42 + "\x1fA\x03\x00\x00\x00\x1fC" + // 0x1F410300: 0x00001F43 + "\x1f@\x03\x01\x00\x00\x1fD" + // 0x1F400301: 0x00001F44 + "\x1fA\x03\x01\x00\x00\x1fE" + // 0x1F410301: 0x00001F45 + "\x03\x9f\x03\x13\x00\x00\x1fH" + // 0x039F0313: 0x00001F48 + "\x03\x9f\x03\x14\x00\x00\x1fI" + // 0x039F0314: 0x00001F49 + "\x1fH\x03\x00\x00\x00\x1fJ" + // 0x1F480300: 0x00001F4A + "\x1fI\x03\x00\x00\x00\x1fK" + // 0x1F490300: 0x00001F4B + "\x1fH\x03\x01\x00\x00\x1fL" + // 0x1F480301: 0x00001F4C + "\x1fI\x03\x01\x00\x00\x1fM" + // 0x1F490301: 0x00001F4D + "\x03\xc5\x03\x13\x00\x00\x1fP" + // 0x03C50313: 0x00001F50 + "\x03\xc5\x03\x14\x00\x00\x1fQ" + // 0x03C50314: 0x00001F51 + "\x1fP\x03\x00\x00\x00\x1fR" + // 0x1F500300: 0x00001F52 + "\x1fQ\x03\x00\x00\x00\x1fS" + // 0x1F510300: 0x00001F53 + "\x1fP\x03\x01\x00\x00\x1fT" + // 0x1F500301: 0x00001F54 + "\x1fQ\x03\x01\x00\x00\x1fU" + // 0x1F510301: 0x00001F55 + "\x1fP\x03B\x00\x00\x1fV" + // 0x1F500342: 0x00001F56 + "\x1fQ\x03B\x00\x00\x1fW" + // 0x1F510342: 0x00001F57 + "\x03\xa5\x03\x14\x00\x00\x1fY" + // 0x03A50314: 0x00001F59 + "\x1fY\x03\x00\x00\x00\x1f[" + // 0x1F590300: 0x00001F5B + "\x1fY\x03\x01\x00\x00\x1f]" + // 0x1F590301: 0x00001F5D + "\x1fY\x03B\x00\x00\x1f_" + // 0x1F590342: 0x00001F5F + "\x03\xc9\x03\x13\x00\x00\x1f`" + // 0x03C90313: 0x00001F60 + "\x03\xc9\x03\x14\x00\x00\x1fa" + // 0x03C90314: 0x00001F61 + "\x1f`\x03\x00\x00\x00\x1fb" + // 0x1F600300: 0x00001F62 + "\x1fa\x03\x00\x00\x00\x1fc" + // 0x1F610300: 0x00001F63 + "\x1f`\x03\x01\x00\x00\x1fd" + // 0x1F600301: 0x00001F64 + "\x1fa\x03\x01\x00\x00\x1fe" + // 0x1F610301: 0x00001F65 + "\x1f`\x03B\x00\x00\x1ff" + // 0x1F600342: 0x00001F66 + "\x1fa\x03B\x00\x00\x1fg" + // 0x1F610342: 0x00001F67 + "\x03\xa9\x03\x13\x00\x00\x1fh" + // 0x03A90313: 0x00001F68 + "\x03\xa9\x03\x14\x00\x00\x1fi" + // 0x03A90314: 0x00001F69 + "\x1fh\x03\x00\x00\x00\x1fj" + // 0x1F680300: 0x00001F6A + "\x1fi\x03\x00\x00\x00\x1fk" + // 0x1F690300: 0x00001F6B + "\x1fh\x03\x01\x00\x00\x1fl" + // 0x1F680301: 0x00001F6C + "\x1fi\x03\x01\x00\x00\x1fm" + // 0x1F690301: 0x00001F6D + "\x1fh\x03B\x00\x00\x1fn" + // 0x1F680342: 0x00001F6E + "\x1fi\x03B\x00\x00\x1fo" + // 0x1F690342: 0x00001F6F + "\x03\xb1\x03\x00\x00\x00\x1fp" + // 0x03B10300: 0x00001F70 + "\x03\xb5\x03\x00\x00\x00\x1fr" + // 0x03B50300: 0x00001F72 + "\x03\xb7\x03\x00\x00\x00\x1ft" + // 0x03B70300: 0x00001F74 + "\x03\xb9\x03\x00\x00\x00\x1fv" + // 0x03B90300: 0x00001F76 + "\x03\xbf\x03\x00\x00\x00\x1fx" + // 0x03BF0300: 0x00001F78 + "\x03\xc5\x03\x00\x00\x00\x1fz" + // 0x03C50300: 0x00001F7A + "\x03\xc9\x03\x00\x00\x00\x1f|" + // 0x03C90300: 0x00001F7C + "\x1f\x00\x03E\x00\x00\x1f\x80" + // 0x1F000345: 0x00001F80 + "\x1f\x01\x03E\x00\x00\x1f\x81" + // 0x1F010345: 0x00001F81 + "\x1f\x02\x03E\x00\x00\x1f\x82" + // 0x1F020345: 0x00001F82 + "\x1f\x03\x03E\x00\x00\x1f\x83" + // 0x1F030345: 0x00001F83 + "\x1f\x04\x03E\x00\x00\x1f\x84" + // 0x1F040345: 0x00001F84 + "\x1f\x05\x03E\x00\x00\x1f\x85" + // 0x1F050345: 0x00001F85 + "\x1f\x06\x03E\x00\x00\x1f\x86" + // 0x1F060345: 0x00001F86 + "\x1f\a\x03E\x00\x00\x1f\x87" + // 0x1F070345: 0x00001F87 + "\x1f\b\x03E\x00\x00\x1f\x88" + // 0x1F080345: 0x00001F88 + "\x1f\t\x03E\x00\x00\x1f\x89" + // 0x1F090345: 0x00001F89 + "\x1f\n\x03E\x00\x00\x1f\x8a" + // 0x1F0A0345: 0x00001F8A + "\x1f\v\x03E\x00\x00\x1f\x8b" + // 0x1F0B0345: 0x00001F8B + "\x1f\f\x03E\x00\x00\x1f\x8c" + // 0x1F0C0345: 0x00001F8C + "\x1f\r\x03E\x00\x00\x1f\x8d" + // 0x1F0D0345: 0x00001F8D + "\x1f\x0e\x03E\x00\x00\x1f\x8e" + // 0x1F0E0345: 0x00001F8E + "\x1f\x0f\x03E\x00\x00\x1f\x8f" + // 0x1F0F0345: 0x00001F8F + "\x1f \x03E\x00\x00\x1f\x90" + // 0x1F200345: 0x00001F90 + "\x1f!\x03E\x00\x00\x1f\x91" + // 0x1F210345: 0x00001F91 + "\x1f\"\x03E\x00\x00\x1f\x92" + // 0x1F220345: 0x00001F92 + "\x1f#\x03E\x00\x00\x1f\x93" + // 0x1F230345: 0x00001F93 + "\x1f$\x03E\x00\x00\x1f\x94" + // 0x1F240345: 0x00001F94 + "\x1f%\x03E\x00\x00\x1f\x95" + // 0x1F250345: 0x00001F95 + "\x1f&\x03E\x00\x00\x1f\x96" + // 0x1F260345: 0x00001F96 + "\x1f'\x03E\x00\x00\x1f\x97" + // 0x1F270345: 0x00001F97 + "\x1f(\x03E\x00\x00\x1f\x98" + // 0x1F280345: 0x00001F98 + "\x1f)\x03E\x00\x00\x1f\x99" + // 0x1F290345: 0x00001F99 + "\x1f*\x03E\x00\x00\x1f\x9a" + // 0x1F2A0345: 0x00001F9A + "\x1f+\x03E\x00\x00\x1f\x9b" + // 0x1F2B0345: 0x00001F9B + "\x1f,\x03E\x00\x00\x1f\x9c" + // 0x1F2C0345: 0x00001F9C + "\x1f-\x03E\x00\x00\x1f\x9d" + // 0x1F2D0345: 0x00001F9D + "\x1f.\x03E\x00\x00\x1f\x9e" + // 0x1F2E0345: 0x00001F9E + "\x1f/\x03E\x00\x00\x1f\x9f" + // 0x1F2F0345: 0x00001F9F + "\x1f`\x03E\x00\x00\x1f\xa0" + // 0x1F600345: 0x00001FA0 + "\x1fa\x03E\x00\x00\x1f\xa1" + // 0x1F610345: 0x00001FA1 + "\x1fb\x03E\x00\x00\x1f\xa2" + // 0x1F620345: 0x00001FA2 + "\x1fc\x03E\x00\x00\x1f\xa3" + // 0x1F630345: 0x00001FA3 + "\x1fd\x03E\x00\x00\x1f\xa4" + // 0x1F640345: 0x00001FA4 + "\x1fe\x03E\x00\x00\x1f\xa5" + // 0x1F650345: 0x00001FA5 + "\x1ff\x03E\x00\x00\x1f\xa6" + // 0x1F660345: 0x00001FA6 + "\x1fg\x03E\x00\x00\x1f\xa7" + // 0x1F670345: 0x00001FA7 + "\x1fh\x03E\x00\x00\x1f\xa8" + // 0x1F680345: 0x00001FA8 + "\x1fi\x03E\x00\x00\x1f\xa9" + // 0x1F690345: 0x00001FA9 + "\x1fj\x03E\x00\x00\x1f\xaa" + // 0x1F6A0345: 0x00001FAA + "\x1fk\x03E\x00\x00\x1f\xab" + // 0x1F6B0345: 0x00001FAB + "\x1fl\x03E\x00\x00\x1f\xac" + // 0x1F6C0345: 0x00001FAC + "\x1fm\x03E\x00\x00\x1f\xad" + // 0x1F6D0345: 0x00001FAD + "\x1fn\x03E\x00\x00\x1f\xae" + // 0x1F6E0345: 0x00001FAE + "\x1fo\x03E\x00\x00\x1f\xaf" + // 0x1F6F0345: 0x00001FAF + "\x03\xb1\x03\x06\x00\x00\x1f\xb0" + // 0x03B10306: 0x00001FB0 + "\x03\xb1\x03\x04\x00\x00\x1f\xb1" + // 0x03B10304: 0x00001FB1 + "\x1fp\x03E\x00\x00\x1f\xb2" + // 0x1F700345: 0x00001FB2 + "\x03\xb1\x03E\x00\x00\x1f\xb3" + // 0x03B10345: 0x00001FB3 + "\x03\xac\x03E\x00\x00\x1f\xb4" + // 0x03AC0345: 0x00001FB4 + "\x03\xb1\x03B\x00\x00\x1f\xb6" + // 0x03B10342: 0x00001FB6 + "\x1f\xb6\x03E\x00\x00\x1f\xb7" + // 0x1FB60345: 0x00001FB7 + "\x03\x91\x03\x06\x00\x00\x1f\xb8" + // 0x03910306: 0x00001FB8 + "\x03\x91\x03\x04\x00\x00\x1f\xb9" + // 0x03910304: 0x00001FB9 + "\x03\x91\x03\x00\x00\x00\x1f\xba" + // 0x03910300: 0x00001FBA + "\x03\x91\x03E\x00\x00\x1f\xbc" + // 0x03910345: 0x00001FBC + "\x00\xa8\x03B\x00\x00\x1f\xc1" + // 0x00A80342: 0x00001FC1 + "\x1ft\x03E\x00\x00\x1f\xc2" + // 0x1F740345: 0x00001FC2 + "\x03\xb7\x03E\x00\x00\x1f\xc3" + // 0x03B70345: 0x00001FC3 + "\x03\xae\x03E\x00\x00\x1f\xc4" + // 0x03AE0345: 0x00001FC4 + "\x03\xb7\x03B\x00\x00\x1f\xc6" + // 0x03B70342: 0x00001FC6 + "\x1f\xc6\x03E\x00\x00\x1f\xc7" + // 0x1FC60345: 0x00001FC7 + "\x03\x95\x03\x00\x00\x00\x1f\xc8" + // 0x03950300: 0x00001FC8 + "\x03\x97\x03\x00\x00\x00\x1f\xca" + // 0x03970300: 0x00001FCA + "\x03\x97\x03E\x00\x00\x1f\xcc" + // 0x03970345: 0x00001FCC + "\x1f\xbf\x03\x00\x00\x00\x1f\xcd" + // 0x1FBF0300: 0x00001FCD + "\x1f\xbf\x03\x01\x00\x00\x1f\xce" + // 0x1FBF0301: 0x00001FCE + "\x1f\xbf\x03B\x00\x00\x1f\xcf" + // 0x1FBF0342: 0x00001FCF + "\x03\xb9\x03\x06\x00\x00\x1f\xd0" + // 0x03B90306: 0x00001FD0 + "\x03\xb9\x03\x04\x00\x00\x1f\xd1" + // 0x03B90304: 0x00001FD1 + "\x03\xca\x03\x00\x00\x00\x1f\xd2" + // 0x03CA0300: 0x00001FD2 + "\x03\xb9\x03B\x00\x00\x1f\xd6" + // 0x03B90342: 0x00001FD6 + "\x03\xca\x03B\x00\x00\x1f\xd7" + // 0x03CA0342: 0x00001FD7 + "\x03\x99\x03\x06\x00\x00\x1f\xd8" + // 0x03990306: 0x00001FD8 + "\x03\x99\x03\x04\x00\x00\x1f\xd9" + // 0x03990304: 0x00001FD9 + "\x03\x99\x03\x00\x00\x00\x1f\xda" + // 0x03990300: 0x00001FDA + "\x1f\xfe\x03\x00\x00\x00\x1f\xdd" + // 0x1FFE0300: 0x00001FDD + "\x1f\xfe\x03\x01\x00\x00\x1f\xde" + // 0x1FFE0301: 0x00001FDE + "\x1f\xfe\x03B\x00\x00\x1f\xdf" + // 0x1FFE0342: 0x00001FDF + "\x03\xc5\x03\x06\x00\x00\x1f\xe0" + // 0x03C50306: 0x00001FE0 + "\x03\xc5\x03\x04\x00\x00\x1f\xe1" + // 0x03C50304: 0x00001FE1 + "\x03\xcb\x03\x00\x00\x00\x1f\xe2" + // 0x03CB0300: 0x00001FE2 + "\x03\xc1\x03\x13\x00\x00\x1f\xe4" + // 0x03C10313: 0x00001FE4 + "\x03\xc1\x03\x14\x00\x00\x1f\xe5" + // 0x03C10314: 0x00001FE5 + "\x03\xc5\x03B\x00\x00\x1f\xe6" + // 0x03C50342: 0x00001FE6 + "\x03\xcb\x03B\x00\x00\x1f\xe7" + // 0x03CB0342: 0x00001FE7 + "\x03\xa5\x03\x06\x00\x00\x1f\xe8" + // 0x03A50306: 0x00001FE8 + "\x03\xa5\x03\x04\x00\x00\x1f\xe9" + // 0x03A50304: 0x00001FE9 + "\x03\xa5\x03\x00\x00\x00\x1f\xea" + // 0x03A50300: 0x00001FEA + "\x03\xa1\x03\x14\x00\x00\x1f\xec" + // 0x03A10314: 0x00001FEC + "\x00\xa8\x03\x00\x00\x00\x1f\xed" + // 0x00A80300: 0x00001FED + "\x1f|\x03E\x00\x00\x1f\xf2" + // 0x1F7C0345: 0x00001FF2 + "\x03\xc9\x03E\x00\x00\x1f\xf3" + // 0x03C90345: 0x00001FF3 + "\x03\xce\x03E\x00\x00\x1f\xf4" + // 0x03CE0345: 0x00001FF4 + "\x03\xc9\x03B\x00\x00\x1f\xf6" + // 0x03C90342: 0x00001FF6 + "\x1f\xf6\x03E\x00\x00\x1f\xf7" + // 0x1FF60345: 0x00001FF7 + "\x03\x9f\x03\x00\x00\x00\x1f\xf8" + // 0x039F0300: 0x00001FF8 + "\x03\xa9\x03\x00\x00\x00\x1f\xfa" + // 0x03A90300: 0x00001FFA + "\x03\xa9\x03E\x00\x00\x1f\xfc" + // 0x03A90345: 0x00001FFC + "!\x90\x038\x00\x00!\x9a" + // 0x21900338: 0x0000219A + "!\x92\x038\x00\x00!\x9b" + // 0x21920338: 0x0000219B + "!\x94\x038\x00\x00!\xae" + // 0x21940338: 0x000021AE + "!\xd0\x038\x00\x00!\xcd" + // 0x21D00338: 0x000021CD + "!\xd4\x038\x00\x00!\xce" + // 0x21D40338: 0x000021CE + "!\xd2\x038\x00\x00!\xcf" + // 0x21D20338: 0x000021CF + "\"\x03\x038\x00\x00\"\x04" + // 0x22030338: 0x00002204 + "\"\b\x038\x00\x00\"\t" + // 0x22080338: 0x00002209 + "\"\v\x038\x00\x00\"\f" + // 0x220B0338: 0x0000220C + "\"#\x038\x00\x00\"$" + // 0x22230338: 0x00002224 + "\"%\x038\x00\x00\"&" + // 0x22250338: 0x00002226 + "\"<\x038\x00\x00\"A" + // 0x223C0338: 0x00002241 + "\"C\x038\x00\x00\"D" + // 0x22430338: 0x00002244 + "\"E\x038\x00\x00\"G" + // 0x22450338: 0x00002247 + "\"H\x038\x00\x00\"I" + // 0x22480338: 0x00002249 + "\x00=\x038\x00\x00\"`" + // 0x003D0338: 0x00002260 + "\"a\x038\x00\x00\"b" + // 0x22610338: 0x00002262 + "\"M\x038\x00\x00\"m" + // 0x224D0338: 0x0000226D + "\x00<\x038\x00\x00\"n" + // 0x003C0338: 0x0000226E + "\x00>\x038\x00\x00\"o" + // 0x003E0338: 0x0000226F + "\"d\x038\x00\x00\"p" + // 0x22640338: 0x00002270 + "\"e\x038\x00\x00\"q" + // 0x22650338: 0x00002271 + "\"r\x038\x00\x00\"t" + // 0x22720338: 0x00002274 + "\"s\x038\x00\x00\"u" + // 0x22730338: 0x00002275 + "\"v\x038\x00\x00\"x" + // 0x22760338: 0x00002278 + "\"w\x038\x00\x00\"y" + // 0x22770338: 0x00002279 + "\"z\x038\x00\x00\"\x80" + // 0x227A0338: 0x00002280 + "\"{\x038\x00\x00\"\x81" + // 0x227B0338: 0x00002281 + "\"\x82\x038\x00\x00\"\x84" + // 0x22820338: 0x00002284 + "\"\x83\x038\x00\x00\"\x85" + // 0x22830338: 0x00002285 + "\"\x86\x038\x00\x00\"\x88" + // 0x22860338: 0x00002288 + "\"\x87\x038\x00\x00\"\x89" + // 0x22870338: 0x00002289 + "\"\xa2\x038\x00\x00\"\xac" + // 0x22A20338: 0x000022AC + "\"\xa8\x038\x00\x00\"\xad" + // 0x22A80338: 0x000022AD + "\"\xa9\x038\x00\x00\"\xae" + // 0x22A90338: 0x000022AE + "\"\xab\x038\x00\x00\"\xaf" + // 0x22AB0338: 0x000022AF + "\"|\x038\x00\x00\"\xe0" + // 0x227C0338: 0x000022E0 + "\"}\x038\x00\x00\"\xe1" + // 0x227D0338: 0x000022E1 + "\"\x91\x038\x00\x00\"\xe2" + // 0x22910338: 0x000022E2 + "\"\x92\x038\x00\x00\"\xe3" + // 0x22920338: 0x000022E3 + "\"\xb2\x038\x00\x00\"\xea" + // 0x22B20338: 0x000022EA + "\"\xb3\x038\x00\x00\"\xeb" + // 0x22B30338: 0x000022EB + "\"\xb4\x038\x00\x00\"\xec" + // 0x22B40338: 0x000022EC + "\"\xb5\x038\x00\x00\"\xed" + // 0x22B50338: 0x000022ED + "0K0\x99\x00\x000L" + // 0x304B3099: 0x0000304C + "0M0\x99\x00\x000N" + // 0x304D3099: 0x0000304E + "0O0\x99\x00\x000P" + // 0x304F3099: 0x00003050 + "0Q0\x99\x00\x000R" + // 0x30513099: 0x00003052 + "0S0\x99\x00\x000T" + // 0x30533099: 0x00003054 + "0U0\x99\x00\x000V" + // 0x30553099: 0x00003056 + "0W0\x99\x00\x000X" + // 0x30573099: 0x00003058 + "0Y0\x99\x00\x000Z" + // 0x30593099: 0x0000305A + "0[0\x99\x00\x000\\" + // 0x305B3099: 0x0000305C + "0]0\x99\x00\x000^" + // 0x305D3099: 0x0000305E + "0_0\x99\x00\x000`" + // 0x305F3099: 0x00003060 + "0a0\x99\x00\x000b" + // 0x30613099: 0x00003062 + "0d0\x99\x00\x000e" + // 0x30643099: 0x00003065 + "0f0\x99\x00\x000g" + // 0x30663099: 0x00003067 + "0h0\x99\x00\x000i" + // 0x30683099: 0x00003069 + "0o0\x99\x00\x000p" + // 0x306F3099: 0x00003070 + "0o0\x9a\x00\x000q" + // 0x306F309A: 0x00003071 + "0r0\x99\x00\x000s" + // 0x30723099: 0x00003073 + "0r0\x9a\x00\x000t" + // 0x3072309A: 0x00003074 + "0u0\x99\x00\x000v" + // 0x30753099: 0x00003076 + "0u0\x9a\x00\x000w" + // 0x3075309A: 0x00003077 + "0x0\x99\x00\x000y" + // 0x30783099: 0x00003079 + "0x0\x9a\x00\x000z" + // 0x3078309A: 0x0000307A + "0{0\x99\x00\x000|" + // 0x307B3099: 0x0000307C + "0{0\x9a\x00\x000}" + // 0x307B309A: 0x0000307D + "0F0\x99\x00\x000\x94" + // 0x30463099: 0x00003094 + "0\x9d0\x99\x00\x000\x9e" + // 0x309D3099: 0x0000309E + "0\xab0\x99\x00\x000\xac" + // 0x30AB3099: 0x000030AC + "0\xad0\x99\x00\x000\xae" + // 0x30AD3099: 0x000030AE + "0\xaf0\x99\x00\x000\xb0" + // 0x30AF3099: 0x000030B0 + "0\xb10\x99\x00\x000\xb2" + // 0x30B13099: 0x000030B2 + "0\xb30\x99\x00\x000\xb4" + // 0x30B33099: 0x000030B4 + "0\xb50\x99\x00\x000\xb6" + // 0x30B53099: 0x000030B6 + "0\xb70\x99\x00\x000\xb8" + // 0x30B73099: 0x000030B8 + "0\xb90\x99\x00\x000\xba" + // 0x30B93099: 0x000030BA + "0\xbb0\x99\x00\x000\xbc" + // 0x30BB3099: 0x000030BC + "0\xbd0\x99\x00\x000\xbe" + // 0x30BD3099: 0x000030BE + "0\xbf0\x99\x00\x000\xc0" + // 0x30BF3099: 0x000030C0 + "0\xc10\x99\x00\x000\xc2" + // 0x30C13099: 0x000030C2 + "0\xc40\x99\x00\x000\xc5" + // 0x30C43099: 0x000030C5 + "0\xc60\x99\x00\x000\xc7" + // 0x30C63099: 0x000030C7 + "0\xc80\x99\x00\x000\xc9" + // 0x30C83099: 0x000030C9 + "0\xcf0\x99\x00\x000\xd0" + // 0x30CF3099: 0x000030D0 + "0\xcf0\x9a\x00\x000\xd1" + // 0x30CF309A: 0x000030D1 + "0\xd20\x99\x00\x000\xd3" + // 0x30D23099: 0x000030D3 + "0\xd20\x9a\x00\x000\xd4" + // 0x30D2309A: 0x000030D4 + "0\xd50\x99\x00\x000\xd6" + // 0x30D53099: 0x000030D6 + "0\xd50\x9a\x00\x000\xd7" + // 0x30D5309A: 0x000030D7 + "0\xd80\x99\x00\x000\xd9" + // 0x30D83099: 0x000030D9 + "0\xd80\x9a\x00\x000\xda" + // 0x30D8309A: 0x000030DA + "0\xdb0\x99\x00\x000\xdc" + // 0x30DB3099: 0x000030DC + "0\xdb0\x9a\x00\x000\xdd" + // 0x30DB309A: 0x000030DD + "0\xa60\x99\x00\x000\xf4" + // 0x30A63099: 0x000030F4 + "0\xef0\x99\x00\x000\xf7" + // 0x30EF3099: 0x000030F7 + "0\xf00\x99\x00\x000\xf8" + // 0x30F03099: 0x000030F8 + "0\xf10\x99\x00\x000\xf9" + // 0x30F13099: 0x000030F9 + "0\xf20\x99\x00\x000\xfa" + // 0x30F23099: 0x000030FA + "0\xfd0\x99\x00\x000\xfe" + // 0x30FD3099: 0x000030FE + "\x10\x99\x10\xba\x00\x01\x10\x9a" + // 0x109910BA: 0x0001109A + "\x10\x9b\x10\xba\x00\x01\x10\x9c" + // 0x109B10BA: 0x0001109C + "\x10\xa5\x10\xba\x00\x01\x10\xab" + // 0x10A510BA: 0x000110AB + "\x111\x11'\x00\x01\x11." + // 0x11311127: 0x0001112E + "\x112\x11'\x00\x01\x11/" + // 0x11321127: 0x0001112F + "\x13G\x13>\x00\x01\x13K" + // 0x1347133E: 0x0001134B + "\x13G\x13W\x00\x01\x13L" + // 0x13471357: 0x0001134C + "\x14\xb9\x14\xba\x00\x01\x14\xbb" + // 0x14B914BA: 0x000114BB + "\x14\xb9\x14\xb0\x00\x01\x14\xbc" + // 0x14B914B0: 0x000114BC + "\x14\xb9\x14\xbd\x00\x01\x14\xbe" + // 0x14B914BD: 0x000114BE + "\x15\xb8\x15\xaf\x00\x01\x15\xba" + // 0x15B815AF: 0x000115BA + "\x15\xb9\x15\xaf\x00\x01\x15\xbb" + // 0x15B915AF: 0x000115BB + "" + // Total size of tables: 55KB (55977 bytes) diff --git a/vendor/golang.org/x/text/width/tables11.0.0.go b/vendor/golang.org/x/text/width/tables11.0.0.go index d6def0e7..3c75e428 100644 --- a/vendor/golang.org/x/text/width/tables11.0.0.go +++ b/vendor/golang.org/x/text/width/tables11.0.0.go @@ -1,6 +1,6 @@ // Code generated by running "go generate" in golang.org/x/text. DO NOT EDIT. -// +build go1.13 +// +build go1.13,!go1.14 package width diff --git a/vendor/golang.org/x/text/width/tables12.0.0.go b/vendor/golang.org/x/text/width/tables12.0.0.go new file mode 100644 index 00000000..5c859677 --- /dev/null +++ b/vendor/golang.org/x/text/width/tables12.0.0.go @@ -0,0 +1,1350 @@ +// Code generated by running "go generate" in golang.org/x/text. DO NOT EDIT. + +// +build go1.14 + +package width + +// UnicodeVersion is the Unicode version from which the tables in this package are derived. +const UnicodeVersion = "12.0.0" + +// lookup returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *widthTrie) lookup(s []byte) (v uint16, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return widthValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := widthIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := widthIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = widthIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := widthIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = widthIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = widthIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *widthTrie) lookupUnsafe(s []byte) uint16 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return widthValues[c0] + } + i := widthIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = widthIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = widthIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// lookupString returns the trie value for the first UTF-8 encoding in s and +// the width in bytes of this encoding. The size will be 0 if s does not +// hold enough bytes to complete the encoding. len(s) must be greater than 0. +func (t *widthTrie) lookupString(s string) (v uint16, sz int) { + c0 := s[0] + switch { + case c0 < 0x80: // is ASCII + return widthValues[c0], 1 + case c0 < 0xC2: + return 0, 1 // Illegal UTF-8: not a starter, not ASCII. + case c0 < 0xE0: // 2-byte UTF-8 + if len(s) < 2 { + return 0, 0 + } + i := widthIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c1), 2 + case c0 < 0xF0: // 3-byte UTF-8 + if len(s) < 3 { + return 0, 0 + } + i := widthIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = widthIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c2), 3 + case c0 < 0xF8: // 4-byte UTF-8 + if len(s) < 4 { + return 0, 0 + } + i := widthIndex[c0] + c1 := s[1] + if c1 < 0x80 || 0xC0 <= c1 { + return 0, 1 // Illegal UTF-8: not a continuation byte. + } + o := uint32(i)<<6 + uint32(c1) + i = widthIndex[o] + c2 := s[2] + if c2 < 0x80 || 0xC0 <= c2 { + return 0, 2 // Illegal UTF-8: not a continuation byte. + } + o = uint32(i)<<6 + uint32(c2) + i = widthIndex[o] + c3 := s[3] + if c3 < 0x80 || 0xC0 <= c3 { + return 0, 3 // Illegal UTF-8: not a continuation byte. + } + return t.lookupValue(uint32(i), c3), 4 + } + // Illegal rune + return 0, 1 +} + +// lookupStringUnsafe returns the trie value for the first UTF-8 encoding in s. +// s must start with a full and valid UTF-8 encoded rune. +func (t *widthTrie) lookupStringUnsafe(s string) uint16 { + c0 := s[0] + if c0 < 0x80 { // is ASCII + return widthValues[c0] + } + i := widthIndex[c0] + if c0 < 0xE0 { // 2-byte UTF-8 + return t.lookupValue(uint32(i), s[1]) + } + i = widthIndex[uint32(i)<<6+uint32(s[1])] + if c0 < 0xF0 { // 3-byte UTF-8 + return t.lookupValue(uint32(i), s[2]) + } + i = widthIndex[uint32(i)<<6+uint32(s[2])] + if c0 < 0xF8 { // 4-byte UTF-8 + return t.lookupValue(uint32(i), s[3]) + } + return 0 +} + +// widthTrie. Total size: 14720 bytes (14.38 KiB). Checksum: 3f4f2516ded5489b. +type widthTrie struct{} + +func newWidthTrie(i int) *widthTrie { + return &widthTrie{} +} + +// lookupValue determines the type of block n and looks up the value for b. +func (t *widthTrie) lookupValue(n uint32, b byte) uint16 { + switch { + default: + return uint16(widthValues[n<<6+uint32(b)]) + } +} + +// widthValues: 104 blocks, 6656 entries, 13312 bytes +// The third block is the zero block. +var widthValues = [6656]uint16{ + // Block 0x0, offset 0x0 + 0x20: 0x6001, 0x21: 0x6002, 0x22: 0x6002, 0x23: 0x6002, + 0x24: 0x6002, 0x25: 0x6002, 0x26: 0x6002, 0x27: 0x6002, 0x28: 0x6002, 0x29: 0x6002, + 0x2a: 0x6002, 0x2b: 0x6002, 0x2c: 0x6002, 0x2d: 0x6002, 0x2e: 0x6002, 0x2f: 0x6002, + 0x30: 0x6002, 0x31: 0x6002, 0x32: 0x6002, 0x33: 0x6002, 0x34: 0x6002, 0x35: 0x6002, + 0x36: 0x6002, 0x37: 0x6002, 0x38: 0x6002, 0x39: 0x6002, 0x3a: 0x6002, 0x3b: 0x6002, + 0x3c: 0x6002, 0x3d: 0x6002, 0x3e: 0x6002, 0x3f: 0x6002, + // Block 0x1, offset 0x40 + 0x40: 0x6003, 0x41: 0x6003, 0x42: 0x6003, 0x43: 0x6003, 0x44: 0x6003, 0x45: 0x6003, + 0x46: 0x6003, 0x47: 0x6003, 0x48: 0x6003, 0x49: 0x6003, 0x4a: 0x6003, 0x4b: 0x6003, + 0x4c: 0x6003, 0x4d: 0x6003, 0x4e: 0x6003, 0x4f: 0x6003, 0x50: 0x6003, 0x51: 0x6003, + 0x52: 0x6003, 0x53: 0x6003, 0x54: 0x6003, 0x55: 0x6003, 0x56: 0x6003, 0x57: 0x6003, + 0x58: 0x6003, 0x59: 0x6003, 0x5a: 0x6003, 0x5b: 0x6003, 0x5c: 0x6003, 0x5d: 0x6003, + 0x5e: 0x6003, 0x5f: 0x6003, 0x60: 0x6004, 0x61: 0x6004, 0x62: 0x6004, 0x63: 0x6004, + 0x64: 0x6004, 0x65: 0x6004, 0x66: 0x6004, 0x67: 0x6004, 0x68: 0x6004, 0x69: 0x6004, + 0x6a: 0x6004, 0x6b: 0x6004, 0x6c: 0x6004, 0x6d: 0x6004, 0x6e: 0x6004, 0x6f: 0x6004, + 0x70: 0x6004, 0x71: 0x6004, 0x72: 0x6004, 0x73: 0x6004, 0x74: 0x6004, 0x75: 0x6004, + 0x76: 0x6004, 0x77: 0x6004, 0x78: 0x6004, 0x79: 0x6004, 0x7a: 0x6004, 0x7b: 0x6004, + 0x7c: 0x6004, 0x7d: 0x6004, 0x7e: 0x6004, + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xe1: 0x2000, 0xe2: 0x6005, 0xe3: 0x6005, + 0xe4: 0x2000, 0xe5: 0x6006, 0xe6: 0x6005, 0xe7: 0x2000, 0xe8: 0x2000, + 0xea: 0x2000, 0xec: 0x6007, 0xed: 0x2000, 0xee: 0x2000, 0xef: 0x6008, + 0xf0: 0x2000, 0xf1: 0x2000, 0xf2: 0x2000, 0xf3: 0x2000, 0xf4: 0x2000, + 0xf6: 0x2000, 0xf7: 0x2000, 0xf8: 0x2000, 0xf9: 0x2000, 0xfa: 0x2000, + 0xfc: 0x2000, 0xfd: 0x2000, 0xfe: 0x2000, 0xff: 0x2000, + // Block 0x4, offset 0x100 + 0x106: 0x2000, + 0x110: 0x2000, + 0x117: 0x2000, + 0x118: 0x2000, + 0x11e: 0x2000, 0x11f: 0x2000, 0x120: 0x2000, 0x121: 0x2000, + 0x126: 0x2000, 0x128: 0x2000, 0x129: 0x2000, + 0x12a: 0x2000, 0x12c: 0x2000, 0x12d: 0x2000, + 0x130: 0x2000, 0x132: 0x2000, 0x133: 0x2000, + 0x137: 0x2000, 0x138: 0x2000, 0x139: 0x2000, 0x13a: 0x2000, + 0x13c: 0x2000, 0x13e: 0x2000, + // Block 0x5, offset 0x140 + 0x141: 0x2000, + 0x151: 0x2000, + 0x153: 0x2000, + 0x15b: 0x2000, + 0x166: 0x2000, 0x167: 0x2000, + 0x16b: 0x2000, + 0x171: 0x2000, 0x172: 0x2000, 0x173: 0x2000, + 0x178: 0x2000, + 0x17f: 0x2000, + // Block 0x6, offset 0x180 + 0x180: 0x2000, 0x181: 0x2000, 0x182: 0x2000, 0x184: 0x2000, + 0x188: 0x2000, 0x189: 0x2000, 0x18a: 0x2000, 0x18b: 0x2000, + 0x18d: 0x2000, + 0x192: 0x2000, 0x193: 0x2000, + 0x1a6: 0x2000, 0x1a7: 0x2000, + 0x1ab: 0x2000, + // Block 0x7, offset 0x1c0 + 0x1ce: 0x2000, 0x1d0: 0x2000, + 0x1d2: 0x2000, 0x1d4: 0x2000, 0x1d6: 0x2000, + 0x1d8: 0x2000, 0x1da: 0x2000, 0x1dc: 0x2000, + // Block 0x8, offset 0x200 + 0x211: 0x2000, + 0x221: 0x2000, + // Block 0x9, offset 0x240 + 0x244: 0x2000, + 0x247: 0x2000, 0x249: 0x2000, 0x24a: 0x2000, 0x24b: 0x2000, + 0x24d: 0x2000, 0x250: 0x2000, + 0x258: 0x2000, 0x259: 0x2000, 0x25a: 0x2000, 0x25b: 0x2000, 0x25d: 0x2000, + 0x25f: 0x2000, + // Block 0xa, offset 0x280 + 0x280: 0x2000, 0x281: 0x2000, 0x282: 0x2000, 0x283: 0x2000, 0x284: 0x2000, 0x285: 0x2000, + 0x286: 0x2000, 0x287: 0x2000, 0x288: 0x2000, 0x289: 0x2000, 0x28a: 0x2000, 0x28b: 0x2000, + 0x28c: 0x2000, 0x28d: 0x2000, 0x28e: 0x2000, 0x28f: 0x2000, 0x290: 0x2000, 0x291: 0x2000, + 0x292: 0x2000, 0x293: 0x2000, 0x294: 0x2000, 0x295: 0x2000, 0x296: 0x2000, 0x297: 0x2000, + 0x298: 0x2000, 0x299: 0x2000, 0x29a: 0x2000, 0x29b: 0x2000, 0x29c: 0x2000, 0x29d: 0x2000, + 0x29e: 0x2000, 0x29f: 0x2000, 0x2a0: 0x2000, 0x2a1: 0x2000, 0x2a2: 0x2000, 0x2a3: 0x2000, + 0x2a4: 0x2000, 0x2a5: 0x2000, 0x2a6: 0x2000, 0x2a7: 0x2000, 0x2a8: 0x2000, 0x2a9: 0x2000, + 0x2aa: 0x2000, 0x2ab: 0x2000, 0x2ac: 0x2000, 0x2ad: 0x2000, 0x2ae: 0x2000, 0x2af: 0x2000, + 0x2b0: 0x2000, 0x2b1: 0x2000, 0x2b2: 0x2000, 0x2b3: 0x2000, 0x2b4: 0x2000, 0x2b5: 0x2000, + 0x2b6: 0x2000, 0x2b7: 0x2000, 0x2b8: 0x2000, 0x2b9: 0x2000, 0x2ba: 0x2000, 0x2bb: 0x2000, + 0x2bc: 0x2000, 0x2bd: 0x2000, 0x2be: 0x2000, 0x2bf: 0x2000, + // Block 0xb, offset 0x2c0 + 0x2c0: 0x2000, 0x2c1: 0x2000, 0x2c2: 0x2000, 0x2c3: 0x2000, 0x2c4: 0x2000, 0x2c5: 0x2000, + 0x2c6: 0x2000, 0x2c7: 0x2000, 0x2c8: 0x2000, 0x2c9: 0x2000, 0x2ca: 0x2000, 0x2cb: 0x2000, + 0x2cc: 0x2000, 0x2cd: 0x2000, 0x2ce: 0x2000, 0x2cf: 0x2000, 0x2d0: 0x2000, 0x2d1: 0x2000, + 0x2d2: 0x2000, 0x2d3: 0x2000, 0x2d4: 0x2000, 0x2d5: 0x2000, 0x2d6: 0x2000, 0x2d7: 0x2000, + 0x2d8: 0x2000, 0x2d9: 0x2000, 0x2da: 0x2000, 0x2db: 0x2000, 0x2dc: 0x2000, 0x2dd: 0x2000, + 0x2de: 0x2000, 0x2df: 0x2000, 0x2e0: 0x2000, 0x2e1: 0x2000, 0x2e2: 0x2000, 0x2e3: 0x2000, + 0x2e4: 0x2000, 0x2e5: 0x2000, 0x2e6: 0x2000, 0x2e7: 0x2000, 0x2e8: 0x2000, 0x2e9: 0x2000, + 0x2ea: 0x2000, 0x2eb: 0x2000, 0x2ec: 0x2000, 0x2ed: 0x2000, 0x2ee: 0x2000, 0x2ef: 0x2000, + // Block 0xc, offset 0x300 + 0x311: 0x2000, + 0x312: 0x2000, 0x313: 0x2000, 0x314: 0x2000, 0x315: 0x2000, 0x316: 0x2000, 0x317: 0x2000, + 0x318: 0x2000, 0x319: 0x2000, 0x31a: 0x2000, 0x31b: 0x2000, 0x31c: 0x2000, 0x31d: 0x2000, + 0x31e: 0x2000, 0x31f: 0x2000, 0x320: 0x2000, 0x321: 0x2000, 0x323: 0x2000, + 0x324: 0x2000, 0x325: 0x2000, 0x326: 0x2000, 0x327: 0x2000, 0x328: 0x2000, 0x329: 0x2000, + 0x331: 0x2000, 0x332: 0x2000, 0x333: 0x2000, 0x334: 0x2000, 0x335: 0x2000, + 0x336: 0x2000, 0x337: 0x2000, 0x338: 0x2000, 0x339: 0x2000, 0x33a: 0x2000, 0x33b: 0x2000, + 0x33c: 0x2000, 0x33d: 0x2000, 0x33e: 0x2000, 0x33f: 0x2000, + // Block 0xd, offset 0x340 + 0x340: 0x2000, 0x341: 0x2000, 0x343: 0x2000, 0x344: 0x2000, 0x345: 0x2000, + 0x346: 0x2000, 0x347: 0x2000, 0x348: 0x2000, 0x349: 0x2000, + // Block 0xe, offset 0x380 + 0x381: 0x2000, + 0x390: 0x2000, 0x391: 0x2000, + 0x392: 0x2000, 0x393: 0x2000, 0x394: 0x2000, 0x395: 0x2000, 0x396: 0x2000, 0x397: 0x2000, + 0x398: 0x2000, 0x399: 0x2000, 0x39a: 0x2000, 0x39b: 0x2000, 0x39c: 0x2000, 0x39d: 0x2000, + 0x39e: 0x2000, 0x39f: 0x2000, 0x3a0: 0x2000, 0x3a1: 0x2000, 0x3a2: 0x2000, 0x3a3: 0x2000, + 0x3a4: 0x2000, 0x3a5: 0x2000, 0x3a6: 0x2000, 0x3a7: 0x2000, 0x3a8: 0x2000, 0x3a9: 0x2000, + 0x3aa: 0x2000, 0x3ab: 0x2000, 0x3ac: 0x2000, 0x3ad: 0x2000, 0x3ae: 0x2000, 0x3af: 0x2000, + 0x3b0: 0x2000, 0x3b1: 0x2000, 0x3b2: 0x2000, 0x3b3: 0x2000, 0x3b4: 0x2000, 0x3b5: 0x2000, + 0x3b6: 0x2000, 0x3b7: 0x2000, 0x3b8: 0x2000, 0x3b9: 0x2000, 0x3ba: 0x2000, 0x3bb: 0x2000, + 0x3bc: 0x2000, 0x3bd: 0x2000, 0x3be: 0x2000, 0x3bf: 0x2000, + // Block 0xf, offset 0x3c0 + 0x3c0: 0x2000, 0x3c1: 0x2000, 0x3c2: 0x2000, 0x3c3: 0x2000, 0x3c4: 0x2000, 0x3c5: 0x2000, + 0x3c6: 0x2000, 0x3c7: 0x2000, 0x3c8: 0x2000, 0x3c9: 0x2000, 0x3ca: 0x2000, 0x3cb: 0x2000, + 0x3cc: 0x2000, 0x3cd: 0x2000, 0x3ce: 0x2000, 0x3cf: 0x2000, 0x3d1: 0x2000, + // Block 0x10, offset 0x400 + 0x400: 0x4000, 0x401: 0x4000, 0x402: 0x4000, 0x403: 0x4000, 0x404: 0x4000, 0x405: 0x4000, + 0x406: 0x4000, 0x407: 0x4000, 0x408: 0x4000, 0x409: 0x4000, 0x40a: 0x4000, 0x40b: 0x4000, + 0x40c: 0x4000, 0x40d: 0x4000, 0x40e: 0x4000, 0x40f: 0x4000, 0x410: 0x4000, 0x411: 0x4000, + 0x412: 0x4000, 0x413: 0x4000, 0x414: 0x4000, 0x415: 0x4000, 0x416: 0x4000, 0x417: 0x4000, + 0x418: 0x4000, 0x419: 0x4000, 0x41a: 0x4000, 0x41b: 0x4000, 0x41c: 0x4000, 0x41d: 0x4000, + 0x41e: 0x4000, 0x41f: 0x4000, 0x420: 0x4000, 0x421: 0x4000, 0x422: 0x4000, 0x423: 0x4000, + 0x424: 0x4000, 0x425: 0x4000, 0x426: 0x4000, 0x427: 0x4000, 0x428: 0x4000, 0x429: 0x4000, + 0x42a: 0x4000, 0x42b: 0x4000, 0x42c: 0x4000, 0x42d: 0x4000, 0x42e: 0x4000, 0x42f: 0x4000, + 0x430: 0x4000, 0x431: 0x4000, 0x432: 0x4000, 0x433: 0x4000, 0x434: 0x4000, 0x435: 0x4000, + 0x436: 0x4000, 0x437: 0x4000, 0x438: 0x4000, 0x439: 0x4000, 0x43a: 0x4000, 0x43b: 0x4000, + 0x43c: 0x4000, 0x43d: 0x4000, 0x43e: 0x4000, 0x43f: 0x4000, + // Block 0x11, offset 0x440 + 0x440: 0x4000, 0x441: 0x4000, 0x442: 0x4000, 0x443: 0x4000, 0x444: 0x4000, 0x445: 0x4000, + 0x446: 0x4000, 0x447: 0x4000, 0x448: 0x4000, 0x449: 0x4000, 0x44a: 0x4000, 0x44b: 0x4000, + 0x44c: 0x4000, 0x44d: 0x4000, 0x44e: 0x4000, 0x44f: 0x4000, 0x450: 0x4000, 0x451: 0x4000, + 0x452: 0x4000, 0x453: 0x4000, 0x454: 0x4000, 0x455: 0x4000, 0x456: 0x4000, 0x457: 0x4000, + 0x458: 0x4000, 0x459: 0x4000, 0x45a: 0x4000, 0x45b: 0x4000, 0x45c: 0x4000, 0x45d: 0x4000, + 0x45e: 0x4000, 0x45f: 0x4000, + // Block 0x12, offset 0x480 + 0x490: 0x2000, + 0x493: 0x2000, 0x494: 0x2000, 0x495: 0x2000, 0x496: 0x2000, + 0x498: 0x2000, 0x499: 0x2000, 0x49c: 0x2000, 0x49d: 0x2000, + 0x4a0: 0x2000, 0x4a1: 0x2000, 0x4a2: 0x2000, + 0x4a4: 0x2000, 0x4a5: 0x2000, 0x4a6: 0x2000, 0x4a7: 0x2000, + 0x4b0: 0x2000, 0x4b2: 0x2000, 0x4b3: 0x2000, 0x4b5: 0x2000, + 0x4bb: 0x2000, + 0x4be: 0x2000, + // Block 0x13, offset 0x4c0 + 0x4f4: 0x2000, + 0x4ff: 0x2000, + // Block 0x14, offset 0x500 + 0x501: 0x2000, 0x502: 0x2000, 0x503: 0x2000, 0x504: 0x2000, + 0x529: 0xa009, + 0x52c: 0x2000, + // Block 0x15, offset 0x540 + 0x543: 0x2000, 0x545: 0x2000, + 0x549: 0x2000, + 0x553: 0x2000, 0x556: 0x2000, + 0x561: 0x2000, 0x562: 0x2000, + 0x566: 0x2000, + 0x56b: 0x2000, + // Block 0x16, offset 0x580 + 0x593: 0x2000, 0x594: 0x2000, + 0x59b: 0x2000, 0x59c: 0x2000, 0x59d: 0x2000, + 0x59e: 0x2000, 0x5a0: 0x2000, 0x5a1: 0x2000, 0x5a2: 0x2000, 0x5a3: 0x2000, + 0x5a4: 0x2000, 0x5a5: 0x2000, 0x5a6: 0x2000, 0x5a7: 0x2000, 0x5a8: 0x2000, 0x5a9: 0x2000, + 0x5aa: 0x2000, 0x5ab: 0x2000, + 0x5b0: 0x2000, 0x5b1: 0x2000, 0x5b2: 0x2000, 0x5b3: 0x2000, 0x5b4: 0x2000, 0x5b5: 0x2000, + 0x5b6: 0x2000, 0x5b7: 0x2000, 0x5b8: 0x2000, 0x5b9: 0x2000, + // Block 0x17, offset 0x5c0 + 0x5c9: 0x2000, + 0x5d0: 0x200a, 0x5d1: 0x200b, + 0x5d2: 0x200a, 0x5d3: 0x200c, 0x5d4: 0x2000, 0x5d5: 0x2000, 0x5d6: 0x2000, 0x5d7: 0x2000, + 0x5d8: 0x2000, 0x5d9: 0x2000, + 0x5f8: 0x2000, 0x5f9: 0x2000, + // Block 0x18, offset 0x600 + 0x612: 0x2000, 0x614: 0x2000, + 0x627: 0x2000, + // Block 0x19, offset 0x640 + 0x640: 0x2000, 0x642: 0x2000, 0x643: 0x2000, + 0x647: 0x2000, 0x648: 0x2000, 0x64b: 0x2000, + 0x64f: 0x2000, 0x651: 0x2000, + 0x655: 0x2000, + 0x65a: 0x2000, 0x65d: 0x2000, + 0x65e: 0x2000, 0x65f: 0x2000, 0x660: 0x2000, 0x663: 0x2000, + 0x665: 0x2000, 0x667: 0x2000, 0x668: 0x2000, 0x669: 0x2000, + 0x66a: 0x2000, 0x66b: 0x2000, 0x66c: 0x2000, 0x66e: 0x2000, + 0x674: 0x2000, 0x675: 0x2000, + 0x676: 0x2000, 0x677: 0x2000, + 0x67c: 0x2000, 0x67d: 0x2000, + // Block 0x1a, offset 0x680 + 0x688: 0x2000, + 0x68c: 0x2000, + 0x692: 0x2000, + 0x6a0: 0x2000, 0x6a1: 0x2000, + 0x6a4: 0x2000, 0x6a5: 0x2000, 0x6a6: 0x2000, 0x6a7: 0x2000, + 0x6aa: 0x2000, 0x6ab: 0x2000, 0x6ae: 0x2000, 0x6af: 0x2000, + // Block 0x1b, offset 0x6c0 + 0x6c2: 0x2000, 0x6c3: 0x2000, + 0x6c6: 0x2000, 0x6c7: 0x2000, + 0x6d5: 0x2000, + 0x6d9: 0x2000, + 0x6e5: 0x2000, + 0x6ff: 0x2000, + // Block 0x1c, offset 0x700 + 0x712: 0x2000, + 0x71a: 0x4000, 0x71b: 0x4000, + 0x729: 0x4000, + 0x72a: 0x4000, + // Block 0x1d, offset 0x740 + 0x769: 0x4000, + 0x76a: 0x4000, 0x76b: 0x4000, 0x76c: 0x4000, + 0x770: 0x4000, 0x773: 0x4000, + // Block 0x1e, offset 0x780 + 0x7a0: 0x2000, 0x7a1: 0x2000, 0x7a2: 0x2000, 0x7a3: 0x2000, + 0x7a4: 0x2000, 0x7a5: 0x2000, 0x7a6: 0x2000, 0x7a7: 0x2000, 0x7a8: 0x2000, 0x7a9: 0x2000, + 0x7aa: 0x2000, 0x7ab: 0x2000, 0x7ac: 0x2000, 0x7ad: 0x2000, 0x7ae: 0x2000, 0x7af: 0x2000, + 0x7b0: 0x2000, 0x7b1: 0x2000, 0x7b2: 0x2000, 0x7b3: 0x2000, 0x7b4: 0x2000, 0x7b5: 0x2000, + 0x7b6: 0x2000, 0x7b7: 0x2000, 0x7b8: 0x2000, 0x7b9: 0x2000, 0x7ba: 0x2000, 0x7bb: 0x2000, + 0x7bc: 0x2000, 0x7bd: 0x2000, 0x7be: 0x2000, 0x7bf: 0x2000, + // Block 0x1f, offset 0x7c0 + 0x7c0: 0x2000, 0x7c1: 0x2000, 0x7c2: 0x2000, 0x7c3: 0x2000, 0x7c4: 0x2000, 0x7c5: 0x2000, + 0x7c6: 0x2000, 0x7c7: 0x2000, 0x7c8: 0x2000, 0x7c9: 0x2000, 0x7ca: 0x2000, 0x7cb: 0x2000, + 0x7cc: 0x2000, 0x7cd: 0x2000, 0x7ce: 0x2000, 0x7cf: 0x2000, 0x7d0: 0x2000, 0x7d1: 0x2000, + 0x7d2: 0x2000, 0x7d3: 0x2000, 0x7d4: 0x2000, 0x7d5: 0x2000, 0x7d6: 0x2000, 0x7d7: 0x2000, + 0x7d8: 0x2000, 0x7d9: 0x2000, 0x7da: 0x2000, 0x7db: 0x2000, 0x7dc: 0x2000, 0x7dd: 0x2000, + 0x7de: 0x2000, 0x7df: 0x2000, 0x7e0: 0x2000, 0x7e1: 0x2000, 0x7e2: 0x2000, 0x7e3: 0x2000, + 0x7e4: 0x2000, 0x7e5: 0x2000, 0x7e6: 0x2000, 0x7e7: 0x2000, 0x7e8: 0x2000, 0x7e9: 0x2000, + 0x7eb: 0x2000, 0x7ec: 0x2000, 0x7ed: 0x2000, 0x7ee: 0x2000, 0x7ef: 0x2000, + 0x7f0: 0x2000, 0x7f1: 0x2000, 0x7f2: 0x2000, 0x7f3: 0x2000, 0x7f4: 0x2000, 0x7f5: 0x2000, + 0x7f6: 0x2000, 0x7f7: 0x2000, 0x7f8: 0x2000, 0x7f9: 0x2000, 0x7fa: 0x2000, 0x7fb: 0x2000, + 0x7fc: 0x2000, 0x7fd: 0x2000, 0x7fe: 0x2000, 0x7ff: 0x2000, + // Block 0x20, offset 0x800 + 0x800: 0x2000, 0x801: 0x2000, 0x802: 0x200d, 0x803: 0x2000, 0x804: 0x2000, 0x805: 0x2000, + 0x806: 0x2000, 0x807: 0x2000, 0x808: 0x2000, 0x809: 0x2000, 0x80a: 0x2000, 0x80b: 0x2000, + 0x80c: 0x2000, 0x80d: 0x2000, 0x80e: 0x2000, 0x80f: 0x2000, 0x810: 0x2000, 0x811: 0x2000, + 0x812: 0x2000, 0x813: 0x2000, 0x814: 0x2000, 0x815: 0x2000, 0x816: 0x2000, 0x817: 0x2000, + 0x818: 0x2000, 0x819: 0x2000, 0x81a: 0x2000, 0x81b: 0x2000, 0x81c: 0x2000, 0x81d: 0x2000, + 0x81e: 0x2000, 0x81f: 0x2000, 0x820: 0x2000, 0x821: 0x2000, 0x822: 0x2000, 0x823: 0x2000, + 0x824: 0x2000, 0x825: 0x2000, 0x826: 0x2000, 0x827: 0x2000, 0x828: 0x2000, 0x829: 0x2000, + 0x82a: 0x2000, 0x82b: 0x2000, 0x82c: 0x2000, 0x82d: 0x2000, 0x82e: 0x2000, 0x82f: 0x2000, + 0x830: 0x2000, 0x831: 0x2000, 0x832: 0x2000, 0x833: 0x2000, 0x834: 0x2000, 0x835: 0x2000, + 0x836: 0x2000, 0x837: 0x2000, 0x838: 0x2000, 0x839: 0x2000, 0x83a: 0x2000, 0x83b: 0x2000, + 0x83c: 0x2000, 0x83d: 0x2000, 0x83e: 0x2000, 0x83f: 0x2000, + // Block 0x21, offset 0x840 + 0x840: 0x2000, 0x841: 0x2000, 0x842: 0x2000, 0x843: 0x2000, 0x844: 0x2000, 0x845: 0x2000, + 0x846: 0x2000, 0x847: 0x2000, 0x848: 0x2000, 0x849: 0x2000, 0x84a: 0x2000, 0x84b: 0x2000, + 0x850: 0x2000, 0x851: 0x2000, + 0x852: 0x2000, 0x853: 0x2000, 0x854: 0x2000, 0x855: 0x2000, 0x856: 0x2000, 0x857: 0x2000, + 0x858: 0x2000, 0x859: 0x2000, 0x85a: 0x2000, 0x85b: 0x2000, 0x85c: 0x2000, 0x85d: 0x2000, + 0x85e: 0x2000, 0x85f: 0x2000, 0x860: 0x2000, 0x861: 0x2000, 0x862: 0x2000, 0x863: 0x2000, + 0x864: 0x2000, 0x865: 0x2000, 0x866: 0x2000, 0x867: 0x2000, 0x868: 0x2000, 0x869: 0x2000, + 0x86a: 0x2000, 0x86b: 0x2000, 0x86c: 0x2000, 0x86d: 0x2000, 0x86e: 0x2000, 0x86f: 0x2000, + 0x870: 0x2000, 0x871: 0x2000, 0x872: 0x2000, 0x873: 0x2000, + // Block 0x22, offset 0x880 + 0x880: 0x2000, 0x881: 0x2000, 0x882: 0x2000, 0x883: 0x2000, 0x884: 0x2000, 0x885: 0x2000, + 0x886: 0x2000, 0x887: 0x2000, 0x888: 0x2000, 0x889: 0x2000, 0x88a: 0x2000, 0x88b: 0x2000, + 0x88c: 0x2000, 0x88d: 0x2000, 0x88e: 0x2000, 0x88f: 0x2000, + 0x892: 0x2000, 0x893: 0x2000, 0x894: 0x2000, 0x895: 0x2000, + 0x8a0: 0x200e, 0x8a1: 0x2000, 0x8a3: 0x2000, + 0x8a4: 0x2000, 0x8a5: 0x2000, 0x8a6: 0x2000, 0x8a7: 0x2000, 0x8a8: 0x2000, 0x8a9: 0x2000, + 0x8b2: 0x2000, 0x8b3: 0x2000, + 0x8b6: 0x2000, 0x8b7: 0x2000, + 0x8bc: 0x2000, 0x8bd: 0x2000, + // Block 0x23, offset 0x8c0 + 0x8c0: 0x2000, 0x8c1: 0x2000, + 0x8c6: 0x2000, 0x8c7: 0x2000, 0x8c8: 0x2000, 0x8cb: 0x200f, + 0x8ce: 0x2000, 0x8cf: 0x2000, 0x8d0: 0x2000, 0x8d1: 0x2000, + 0x8e2: 0x2000, 0x8e3: 0x2000, + 0x8e4: 0x2000, 0x8e5: 0x2000, + 0x8ef: 0x2000, + 0x8fd: 0x4000, 0x8fe: 0x4000, + // Block 0x24, offset 0x900 + 0x905: 0x2000, + 0x906: 0x2000, 0x909: 0x2000, + 0x90e: 0x2000, 0x90f: 0x2000, + 0x914: 0x4000, 0x915: 0x4000, + 0x91c: 0x2000, + 0x91e: 0x2000, + // Block 0x25, offset 0x940 + 0x940: 0x2000, 0x942: 0x2000, + 0x948: 0x4000, 0x949: 0x4000, 0x94a: 0x4000, 0x94b: 0x4000, + 0x94c: 0x4000, 0x94d: 0x4000, 0x94e: 0x4000, 0x94f: 0x4000, 0x950: 0x4000, 0x951: 0x4000, + 0x952: 0x4000, 0x953: 0x4000, + 0x960: 0x2000, 0x961: 0x2000, 0x963: 0x2000, + 0x964: 0x2000, 0x965: 0x2000, 0x967: 0x2000, 0x968: 0x2000, 0x969: 0x2000, + 0x96a: 0x2000, 0x96c: 0x2000, 0x96d: 0x2000, 0x96f: 0x2000, + 0x97f: 0x4000, + // Block 0x26, offset 0x980 + 0x993: 0x4000, + 0x99e: 0x2000, 0x99f: 0x2000, 0x9a1: 0x4000, + 0x9aa: 0x4000, 0x9ab: 0x4000, + 0x9bd: 0x4000, 0x9be: 0x4000, 0x9bf: 0x2000, + // Block 0x27, offset 0x9c0 + 0x9c4: 0x4000, 0x9c5: 0x4000, + 0x9c6: 0x2000, 0x9c7: 0x2000, 0x9c8: 0x2000, 0x9c9: 0x2000, 0x9ca: 0x2000, 0x9cb: 0x2000, + 0x9cc: 0x2000, 0x9cd: 0x2000, 0x9ce: 0x4000, 0x9cf: 0x2000, 0x9d0: 0x2000, 0x9d1: 0x2000, + 0x9d2: 0x2000, 0x9d3: 0x2000, 0x9d4: 0x4000, 0x9d5: 0x2000, 0x9d6: 0x2000, 0x9d7: 0x2000, + 0x9d8: 0x2000, 0x9d9: 0x2000, 0x9da: 0x2000, 0x9db: 0x2000, 0x9dc: 0x2000, 0x9dd: 0x2000, + 0x9de: 0x2000, 0x9df: 0x2000, 0x9e0: 0x2000, 0x9e1: 0x2000, 0x9e3: 0x2000, + 0x9e8: 0x2000, 0x9e9: 0x2000, + 0x9ea: 0x4000, 0x9eb: 0x2000, 0x9ec: 0x2000, 0x9ed: 0x2000, 0x9ee: 0x2000, 0x9ef: 0x2000, + 0x9f0: 0x2000, 0x9f1: 0x2000, 0x9f2: 0x4000, 0x9f3: 0x4000, 0x9f4: 0x2000, 0x9f5: 0x4000, + 0x9f6: 0x2000, 0x9f7: 0x2000, 0x9f8: 0x2000, 0x9f9: 0x2000, 0x9fa: 0x4000, 0x9fb: 0x2000, + 0x9fc: 0x2000, 0x9fd: 0x4000, 0x9fe: 0x2000, 0x9ff: 0x2000, + // Block 0x28, offset 0xa00 + 0xa05: 0x4000, + 0xa0a: 0x4000, 0xa0b: 0x4000, + 0xa28: 0x4000, + 0xa3d: 0x2000, + // Block 0x29, offset 0xa40 + 0xa4c: 0x4000, 0xa4e: 0x4000, + 0xa53: 0x4000, 0xa54: 0x4000, 0xa55: 0x4000, 0xa57: 0x4000, + 0xa76: 0x2000, 0xa77: 0x2000, 0xa78: 0x2000, 0xa79: 0x2000, 0xa7a: 0x2000, 0xa7b: 0x2000, + 0xa7c: 0x2000, 0xa7d: 0x2000, 0xa7e: 0x2000, 0xa7f: 0x2000, + // Block 0x2a, offset 0xa80 + 0xa95: 0x4000, 0xa96: 0x4000, 0xa97: 0x4000, + 0xab0: 0x4000, + 0xabf: 0x4000, + // Block 0x2b, offset 0xac0 + 0xae6: 0x6000, 0xae7: 0x6000, 0xae8: 0x6000, 0xae9: 0x6000, + 0xaea: 0x6000, 0xaeb: 0x6000, 0xaec: 0x6000, 0xaed: 0x6000, + // Block 0x2c, offset 0xb00 + 0xb05: 0x6010, + 0xb06: 0x6011, + // Block 0x2d, offset 0xb40 + 0xb5b: 0x4000, 0xb5c: 0x4000, + // Block 0x2e, offset 0xb80 + 0xb90: 0x4000, + 0xb95: 0x4000, 0xb96: 0x2000, 0xb97: 0x2000, + 0xb98: 0x2000, 0xb99: 0x2000, + // Block 0x2f, offset 0xbc0 + 0xbc0: 0x4000, 0xbc1: 0x4000, 0xbc2: 0x4000, 0xbc3: 0x4000, 0xbc4: 0x4000, 0xbc5: 0x4000, + 0xbc6: 0x4000, 0xbc7: 0x4000, 0xbc8: 0x4000, 0xbc9: 0x4000, 0xbca: 0x4000, 0xbcb: 0x4000, + 0xbcc: 0x4000, 0xbcd: 0x4000, 0xbce: 0x4000, 0xbcf: 0x4000, 0xbd0: 0x4000, 0xbd1: 0x4000, + 0xbd2: 0x4000, 0xbd3: 0x4000, 0xbd4: 0x4000, 0xbd5: 0x4000, 0xbd6: 0x4000, 0xbd7: 0x4000, + 0xbd8: 0x4000, 0xbd9: 0x4000, 0xbdb: 0x4000, 0xbdc: 0x4000, 0xbdd: 0x4000, + 0xbde: 0x4000, 0xbdf: 0x4000, 0xbe0: 0x4000, 0xbe1: 0x4000, 0xbe2: 0x4000, 0xbe3: 0x4000, + 0xbe4: 0x4000, 0xbe5: 0x4000, 0xbe6: 0x4000, 0xbe7: 0x4000, 0xbe8: 0x4000, 0xbe9: 0x4000, + 0xbea: 0x4000, 0xbeb: 0x4000, 0xbec: 0x4000, 0xbed: 0x4000, 0xbee: 0x4000, 0xbef: 0x4000, + 0xbf0: 0x4000, 0xbf1: 0x4000, 0xbf2: 0x4000, 0xbf3: 0x4000, 0xbf4: 0x4000, 0xbf5: 0x4000, + 0xbf6: 0x4000, 0xbf7: 0x4000, 0xbf8: 0x4000, 0xbf9: 0x4000, 0xbfa: 0x4000, 0xbfb: 0x4000, + 0xbfc: 0x4000, 0xbfd: 0x4000, 0xbfe: 0x4000, 0xbff: 0x4000, + // Block 0x30, offset 0xc00 + 0xc00: 0x4000, 0xc01: 0x4000, 0xc02: 0x4000, 0xc03: 0x4000, 0xc04: 0x4000, 0xc05: 0x4000, + 0xc06: 0x4000, 0xc07: 0x4000, 0xc08: 0x4000, 0xc09: 0x4000, 0xc0a: 0x4000, 0xc0b: 0x4000, + 0xc0c: 0x4000, 0xc0d: 0x4000, 0xc0e: 0x4000, 0xc0f: 0x4000, 0xc10: 0x4000, 0xc11: 0x4000, + 0xc12: 0x4000, 0xc13: 0x4000, 0xc14: 0x4000, 0xc15: 0x4000, 0xc16: 0x4000, 0xc17: 0x4000, + 0xc18: 0x4000, 0xc19: 0x4000, 0xc1a: 0x4000, 0xc1b: 0x4000, 0xc1c: 0x4000, 0xc1d: 0x4000, + 0xc1e: 0x4000, 0xc1f: 0x4000, 0xc20: 0x4000, 0xc21: 0x4000, 0xc22: 0x4000, 0xc23: 0x4000, + 0xc24: 0x4000, 0xc25: 0x4000, 0xc26: 0x4000, 0xc27: 0x4000, 0xc28: 0x4000, 0xc29: 0x4000, + 0xc2a: 0x4000, 0xc2b: 0x4000, 0xc2c: 0x4000, 0xc2d: 0x4000, 0xc2e: 0x4000, 0xc2f: 0x4000, + 0xc30: 0x4000, 0xc31: 0x4000, 0xc32: 0x4000, 0xc33: 0x4000, + // Block 0x31, offset 0xc40 + 0xc40: 0x4000, 0xc41: 0x4000, 0xc42: 0x4000, 0xc43: 0x4000, 0xc44: 0x4000, 0xc45: 0x4000, + 0xc46: 0x4000, 0xc47: 0x4000, 0xc48: 0x4000, 0xc49: 0x4000, 0xc4a: 0x4000, 0xc4b: 0x4000, + 0xc4c: 0x4000, 0xc4d: 0x4000, 0xc4e: 0x4000, 0xc4f: 0x4000, 0xc50: 0x4000, 0xc51: 0x4000, + 0xc52: 0x4000, 0xc53: 0x4000, 0xc54: 0x4000, 0xc55: 0x4000, + 0xc70: 0x4000, 0xc71: 0x4000, 0xc72: 0x4000, 0xc73: 0x4000, 0xc74: 0x4000, 0xc75: 0x4000, + 0xc76: 0x4000, 0xc77: 0x4000, 0xc78: 0x4000, 0xc79: 0x4000, 0xc7a: 0x4000, 0xc7b: 0x4000, + // Block 0x32, offset 0xc80 + 0xc80: 0x9012, 0xc81: 0x4013, 0xc82: 0x4014, 0xc83: 0x4000, 0xc84: 0x4000, 0xc85: 0x4000, + 0xc86: 0x4000, 0xc87: 0x4000, 0xc88: 0x4000, 0xc89: 0x4000, 0xc8a: 0x4000, 0xc8b: 0x4000, + 0xc8c: 0x4015, 0xc8d: 0x4015, 0xc8e: 0x4000, 0xc8f: 0x4000, 0xc90: 0x4000, 0xc91: 0x4000, + 0xc92: 0x4000, 0xc93: 0x4000, 0xc94: 0x4000, 0xc95: 0x4000, 0xc96: 0x4000, 0xc97: 0x4000, + 0xc98: 0x4000, 0xc99: 0x4000, 0xc9a: 0x4000, 0xc9b: 0x4000, 0xc9c: 0x4000, 0xc9d: 0x4000, + 0xc9e: 0x4000, 0xc9f: 0x4000, 0xca0: 0x4000, 0xca1: 0x4000, 0xca2: 0x4000, 0xca3: 0x4000, + 0xca4: 0x4000, 0xca5: 0x4000, 0xca6: 0x4000, 0xca7: 0x4000, 0xca8: 0x4000, 0xca9: 0x4000, + 0xcaa: 0x4000, 0xcab: 0x4000, 0xcac: 0x4000, 0xcad: 0x4000, 0xcae: 0x4000, 0xcaf: 0x4000, + 0xcb0: 0x4000, 0xcb1: 0x4000, 0xcb2: 0x4000, 0xcb3: 0x4000, 0xcb4: 0x4000, 0xcb5: 0x4000, + 0xcb6: 0x4000, 0xcb7: 0x4000, 0xcb8: 0x4000, 0xcb9: 0x4000, 0xcba: 0x4000, 0xcbb: 0x4000, + 0xcbc: 0x4000, 0xcbd: 0x4000, 0xcbe: 0x4000, + // Block 0x33, offset 0xcc0 + 0xcc1: 0x4000, 0xcc2: 0x4000, 0xcc3: 0x4000, 0xcc4: 0x4000, 0xcc5: 0x4000, + 0xcc6: 0x4000, 0xcc7: 0x4000, 0xcc8: 0x4000, 0xcc9: 0x4000, 0xcca: 0x4000, 0xccb: 0x4000, + 0xccc: 0x4000, 0xccd: 0x4000, 0xcce: 0x4000, 0xccf: 0x4000, 0xcd0: 0x4000, 0xcd1: 0x4000, + 0xcd2: 0x4000, 0xcd3: 0x4000, 0xcd4: 0x4000, 0xcd5: 0x4000, 0xcd6: 0x4000, 0xcd7: 0x4000, + 0xcd8: 0x4000, 0xcd9: 0x4000, 0xcda: 0x4000, 0xcdb: 0x4000, 0xcdc: 0x4000, 0xcdd: 0x4000, + 0xcde: 0x4000, 0xcdf: 0x4000, 0xce0: 0x4000, 0xce1: 0x4000, 0xce2: 0x4000, 0xce3: 0x4000, + 0xce4: 0x4000, 0xce5: 0x4000, 0xce6: 0x4000, 0xce7: 0x4000, 0xce8: 0x4000, 0xce9: 0x4000, + 0xcea: 0x4000, 0xceb: 0x4000, 0xcec: 0x4000, 0xced: 0x4000, 0xcee: 0x4000, 0xcef: 0x4000, + 0xcf0: 0x4000, 0xcf1: 0x4000, 0xcf2: 0x4000, 0xcf3: 0x4000, 0xcf4: 0x4000, 0xcf5: 0x4000, + 0xcf6: 0x4000, 0xcf7: 0x4000, 0xcf8: 0x4000, 0xcf9: 0x4000, 0xcfa: 0x4000, 0xcfb: 0x4000, + 0xcfc: 0x4000, 0xcfd: 0x4000, 0xcfe: 0x4000, 0xcff: 0x4000, + // Block 0x34, offset 0xd00 + 0xd00: 0x4000, 0xd01: 0x4000, 0xd02: 0x4000, 0xd03: 0x4000, 0xd04: 0x4000, 0xd05: 0x4000, + 0xd06: 0x4000, 0xd07: 0x4000, 0xd08: 0x4000, 0xd09: 0x4000, 0xd0a: 0x4000, 0xd0b: 0x4000, + 0xd0c: 0x4000, 0xd0d: 0x4000, 0xd0e: 0x4000, 0xd0f: 0x4000, 0xd10: 0x4000, 0xd11: 0x4000, + 0xd12: 0x4000, 0xd13: 0x4000, 0xd14: 0x4000, 0xd15: 0x4000, 0xd16: 0x4000, + 0xd19: 0x4016, 0xd1a: 0x4017, 0xd1b: 0x4000, 0xd1c: 0x4000, 0xd1d: 0x4000, + 0xd1e: 0x4000, 0xd1f: 0x4000, 0xd20: 0x4000, 0xd21: 0x4018, 0xd22: 0x4019, 0xd23: 0x401a, + 0xd24: 0x401b, 0xd25: 0x401c, 0xd26: 0x401d, 0xd27: 0x401e, 0xd28: 0x401f, 0xd29: 0x4020, + 0xd2a: 0x4021, 0xd2b: 0x4022, 0xd2c: 0x4000, 0xd2d: 0x4010, 0xd2e: 0x4000, 0xd2f: 0x4023, + 0xd30: 0x4000, 0xd31: 0x4024, 0xd32: 0x4000, 0xd33: 0x4025, 0xd34: 0x4000, 0xd35: 0x4026, + 0xd36: 0x4000, 0xd37: 0x401a, 0xd38: 0x4000, 0xd39: 0x4027, 0xd3a: 0x4000, 0xd3b: 0x4028, + 0xd3c: 0x4000, 0xd3d: 0x4020, 0xd3e: 0x4000, 0xd3f: 0x4029, + // Block 0x35, offset 0xd40 + 0xd40: 0x4000, 0xd41: 0x402a, 0xd42: 0x4000, 0xd43: 0x402b, 0xd44: 0x402c, 0xd45: 0x4000, + 0xd46: 0x4017, 0xd47: 0x4000, 0xd48: 0x402d, 0xd49: 0x4000, 0xd4a: 0x402e, 0xd4b: 0x402f, + 0xd4c: 0x4030, 0xd4d: 0x4017, 0xd4e: 0x4016, 0xd4f: 0x4017, 0xd50: 0x4000, 0xd51: 0x4000, + 0xd52: 0x4031, 0xd53: 0x4000, 0xd54: 0x4000, 0xd55: 0x4031, 0xd56: 0x4000, 0xd57: 0x4000, + 0xd58: 0x4032, 0xd59: 0x4000, 0xd5a: 0x4000, 0xd5b: 0x4032, 0xd5c: 0x4000, 0xd5d: 0x4000, + 0xd5e: 0x4033, 0xd5f: 0x402e, 0xd60: 0x4034, 0xd61: 0x4035, 0xd62: 0x4034, 0xd63: 0x4036, + 0xd64: 0x4037, 0xd65: 0x4024, 0xd66: 0x4035, 0xd67: 0x4025, 0xd68: 0x4038, 0xd69: 0x4038, + 0xd6a: 0x4039, 0xd6b: 0x4039, 0xd6c: 0x403a, 0xd6d: 0x403a, 0xd6e: 0x4000, 0xd6f: 0x4035, + 0xd70: 0x4000, 0xd71: 0x4000, 0xd72: 0x403b, 0xd73: 0x403c, 0xd74: 0x4000, 0xd75: 0x4000, + 0xd76: 0x4000, 0xd77: 0x4000, 0xd78: 0x4000, 0xd79: 0x4000, 0xd7a: 0x4000, 0xd7b: 0x403d, + 0xd7c: 0x401c, 0xd7d: 0x4000, 0xd7e: 0x4000, 0xd7f: 0x4000, + // Block 0x36, offset 0xd80 + 0xd85: 0x4000, + 0xd86: 0x4000, 0xd87: 0x4000, 0xd88: 0x4000, 0xd89: 0x4000, 0xd8a: 0x4000, 0xd8b: 0x4000, + 0xd8c: 0x4000, 0xd8d: 0x4000, 0xd8e: 0x4000, 0xd8f: 0x4000, 0xd90: 0x4000, 0xd91: 0x4000, + 0xd92: 0x4000, 0xd93: 0x4000, 0xd94: 0x4000, 0xd95: 0x4000, 0xd96: 0x4000, 0xd97: 0x4000, + 0xd98: 0x4000, 0xd99: 0x4000, 0xd9a: 0x4000, 0xd9b: 0x4000, 0xd9c: 0x4000, 0xd9d: 0x4000, + 0xd9e: 0x4000, 0xd9f: 0x4000, 0xda0: 0x4000, 0xda1: 0x4000, 0xda2: 0x4000, 0xda3: 0x4000, + 0xda4: 0x4000, 0xda5: 0x4000, 0xda6: 0x4000, 0xda7: 0x4000, 0xda8: 0x4000, 0xda9: 0x4000, + 0xdaa: 0x4000, 0xdab: 0x4000, 0xdac: 0x4000, 0xdad: 0x4000, 0xdae: 0x4000, 0xdaf: 0x4000, + 0xdb1: 0x403e, 0xdb2: 0x403e, 0xdb3: 0x403e, 0xdb4: 0x403e, 0xdb5: 0x403e, + 0xdb6: 0x403e, 0xdb7: 0x403e, 0xdb8: 0x403e, 0xdb9: 0x403e, 0xdba: 0x403e, 0xdbb: 0x403e, + 0xdbc: 0x403e, 0xdbd: 0x403e, 0xdbe: 0x403e, 0xdbf: 0x403e, + // Block 0x37, offset 0xdc0 + 0xdc0: 0x4037, 0xdc1: 0x4037, 0xdc2: 0x4037, 0xdc3: 0x4037, 0xdc4: 0x4037, 0xdc5: 0x4037, + 0xdc6: 0x4037, 0xdc7: 0x4037, 0xdc8: 0x4037, 0xdc9: 0x4037, 0xdca: 0x4037, 0xdcb: 0x4037, + 0xdcc: 0x4037, 0xdcd: 0x4037, 0xdce: 0x4037, 0xdcf: 0x400e, 0xdd0: 0x403f, 0xdd1: 0x4040, + 0xdd2: 0x4041, 0xdd3: 0x4040, 0xdd4: 0x403f, 0xdd5: 0x4042, 0xdd6: 0x4043, 0xdd7: 0x4044, + 0xdd8: 0x4040, 0xdd9: 0x4041, 0xdda: 0x4040, 0xddb: 0x4045, 0xddc: 0x4009, 0xddd: 0x4045, + 0xdde: 0x4046, 0xddf: 0x4045, 0xde0: 0x4047, 0xde1: 0x400b, 0xde2: 0x400a, 0xde3: 0x400c, + 0xde4: 0x4048, 0xde5: 0x4000, 0xde6: 0x4000, 0xde7: 0x4000, 0xde8: 0x4000, 0xde9: 0x4000, + 0xdea: 0x4000, 0xdeb: 0x4000, 0xdec: 0x4000, 0xded: 0x4000, 0xdee: 0x4000, 0xdef: 0x4000, + 0xdf0: 0x4000, 0xdf1: 0x4000, 0xdf2: 0x4000, 0xdf3: 0x4000, 0xdf4: 0x4000, 0xdf5: 0x4000, + 0xdf6: 0x4000, 0xdf7: 0x4000, 0xdf8: 0x4000, 0xdf9: 0x4000, 0xdfa: 0x4000, 0xdfb: 0x4000, + 0xdfc: 0x4000, 0xdfd: 0x4000, 0xdfe: 0x4000, 0xdff: 0x4000, + // Block 0x38, offset 0xe00 + 0xe00: 0x4000, 0xe01: 0x4000, 0xe02: 0x4000, 0xe03: 0x4000, 0xe04: 0x4000, 0xe05: 0x4000, + 0xe06: 0x4000, 0xe07: 0x4000, 0xe08: 0x4000, 0xe09: 0x4000, 0xe0a: 0x4000, 0xe0b: 0x4000, + 0xe0c: 0x4000, 0xe0d: 0x4000, 0xe0e: 0x4000, 0xe10: 0x4000, 0xe11: 0x4000, + 0xe12: 0x4000, 0xe13: 0x4000, 0xe14: 0x4000, 0xe15: 0x4000, 0xe16: 0x4000, 0xe17: 0x4000, + 0xe18: 0x4000, 0xe19: 0x4000, 0xe1a: 0x4000, 0xe1b: 0x4000, 0xe1c: 0x4000, 0xe1d: 0x4000, + 0xe1e: 0x4000, 0xe1f: 0x4000, 0xe20: 0x4000, 0xe21: 0x4000, 0xe22: 0x4000, 0xe23: 0x4000, + 0xe24: 0x4000, 0xe25: 0x4000, 0xe26: 0x4000, 0xe27: 0x4000, 0xe28: 0x4000, 0xe29: 0x4000, + 0xe2a: 0x4000, 0xe2b: 0x4000, 0xe2c: 0x4000, 0xe2d: 0x4000, 0xe2e: 0x4000, 0xe2f: 0x4000, + 0xe30: 0x4000, 0xe31: 0x4000, 0xe32: 0x4000, 0xe33: 0x4000, 0xe34: 0x4000, 0xe35: 0x4000, + 0xe36: 0x4000, 0xe37: 0x4000, 0xe38: 0x4000, 0xe39: 0x4000, 0xe3a: 0x4000, + // Block 0x39, offset 0xe40 + 0xe40: 0x4000, 0xe41: 0x4000, 0xe42: 0x4000, 0xe43: 0x4000, 0xe44: 0x4000, 0xe45: 0x4000, + 0xe46: 0x4000, 0xe47: 0x4000, 0xe48: 0x4000, 0xe49: 0x4000, 0xe4a: 0x4000, 0xe4b: 0x4000, + 0xe4c: 0x4000, 0xe4d: 0x4000, 0xe4e: 0x4000, 0xe4f: 0x4000, 0xe50: 0x4000, 0xe51: 0x4000, + 0xe52: 0x4000, 0xe53: 0x4000, 0xe54: 0x4000, 0xe55: 0x4000, 0xe56: 0x4000, 0xe57: 0x4000, + 0xe58: 0x4000, 0xe59: 0x4000, 0xe5a: 0x4000, 0xe5b: 0x4000, 0xe5c: 0x4000, 0xe5d: 0x4000, + 0xe5e: 0x4000, 0xe5f: 0x4000, 0xe60: 0x4000, 0xe61: 0x4000, 0xe62: 0x4000, 0xe63: 0x4000, + 0xe70: 0x4000, 0xe71: 0x4000, 0xe72: 0x4000, 0xe73: 0x4000, 0xe74: 0x4000, 0xe75: 0x4000, + 0xe76: 0x4000, 0xe77: 0x4000, 0xe78: 0x4000, 0xe79: 0x4000, 0xe7a: 0x4000, 0xe7b: 0x4000, + 0xe7c: 0x4000, 0xe7d: 0x4000, 0xe7e: 0x4000, 0xe7f: 0x4000, + // Block 0x3a, offset 0xe80 + 0xe80: 0x4000, 0xe81: 0x4000, 0xe82: 0x4000, 0xe83: 0x4000, 0xe84: 0x4000, 0xe85: 0x4000, + 0xe86: 0x4000, 0xe87: 0x4000, 0xe88: 0x4000, 0xe89: 0x4000, 0xe8a: 0x4000, 0xe8b: 0x4000, + 0xe8c: 0x4000, 0xe8d: 0x4000, 0xe8e: 0x4000, 0xe8f: 0x4000, 0xe90: 0x4000, 0xe91: 0x4000, + 0xe92: 0x4000, 0xe93: 0x4000, 0xe94: 0x4000, 0xe95: 0x4000, 0xe96: 0x4000, 0xe97: 0x4000, + 0xe98: 0x4000, 0xe99: 0x4000, 0xe9a: 0x4000, 0xe9b: 0x4000, 0xe9c: 0x4000, 0xe9d: 0x4000, + 0xe9e: 0x4000, 0xea0: 0x4000, 0xea1: 0x4000, 0xea2: 0x4000, 0xea3: 0x4000, + 0xea4: 0x4000, 0xea5: 0x4000, 0xea6: 0x4000, 0xea7: 0x4000, 0xea8: 0x4000, 0xea9: 0x4000, + 0xeaa: 0x4000, 0xeab: 0x4000, 0xeac: 0x4000, 0xead: 0x4000, 0xeae: 0x4000, 0xeaf: 0x4000, + 0xeb0: 0x4000, 0xeb1: 0x4000, 0xeb2: 0x4000, 0xeb3: 0x4000, 0xeb4: 0x4000, 0xeb5: 0x4000, + 0xeb6: 0x4000, 0xeb7: 0x4000, 0xeb8: 0x4000, 0xeb9: 0x4000, 0xeba: 0x4000, 0xebb: 0x4000, + 0xebc: 0x4000, 0xebd: 0x4000, 0xebe: 0x4000, 0xebf: 0x4000, + // Block 0x3b, offset 0xec0 + 0xec0: 0x4000, 0xec1: 0x4000, 0xec2: 0x4000, 0xec3: 0x4000, 0xec4: 0x4000, 0xec5: 0x4000, + 0xec6: 0x4000, 0xec7: 0x4000, 0xec8: 0x2000, 0xec9: 0x2000, 0xeca: 0x2000, 0xecb: 0x2000, + 0xecc: 0x2000, 0xecd: 0x2000, 0xece: 0x2000, 0xecf: 0x2000, 0xed0: 0x4000, 0xed1: 0x4000, + 0xed2: 0x4000, 0xed3: 0x4000, 0xed4: 0x4000, 0xed5: 0x4000, 0xed6: 0x4000, 0xed7: 0x4000, + 0xed8: 0x4000, 0xed9: 0x4000, 0xeda: 0x4000, 0xedb: 0x4000, 0xedc: 0x4000, 0xedd: 0x4000, + 0xede: 0x4000, 0xedf: 0x4000, 0xee0: 0x4000, 0xee1: 0x4000, 0xee2: 0x4000, 0xee3: 0x4000, + 0xee4: 0x4000, 0xee5: 0x4000, 0xee6: 0x4000, 0xee7: 0x4000, 0xee8: 0x4000, 0xee9: 0x4000, + 0xeea: 0x4000, 0xeeb: 0x4000, 0xeec: 0x4000, 0xeed: 0x4000, 0xeee: 0x4000, 0xeef: 0x4000, + 0xef0: 0x4000, 0xef1: 0x4000, 0xef2: 0x4000, 0xef3: 0x4000, 0xef4: 0x4000, 0xef5: 0x4000, + 0xef6: 0x4000, 0xef7: 0x4000, 0xef8: 0x4000, 0xef9: 0x4000, 0xefa: 0x4000, 0xefb: 0x4000, + 0xefc: 0x4000, 0xefd: 0x4000, 0xefe: 0x4000, 0xeff: 0x4000, + // Block 0x3c, offset 0xf00 + 0xf00: 0x4000, 0xf01: 0x4000, 0xf02: 0x4000, 0xf03: 0x4000, 0xf04: 0x4000, 0xf05: 0x4000, + 0xf06: 0x4000, 0xf07: 0x4000, 0xf08: 0x4000, 0xf09: 0x4000, 0xf0a: 0x4000, 0xf0b: 0x4000, + 0xf0c: 0x4000, 0xf0d: 0x4000, 0xf0e: 0x4000, 0xf0f: 0x4000, 0xf10: 0x4000, 0xf11: 0x4000, + 0xf12: 0x4000, 0xf13: 0x4000, 0xf14: 0x4000, 0xf15: 0x4000, 0xf16: 0x4000, 0xf17: 0x4000, + 0xf18: 0x4000, 0xf19: 0x4000, 0xf1a: 0x4000, 0xf1b: 0x4000, 0xf1c: 0x4000, 0xf1d: 0x4000, + 0xf1e: 0x4000, 0xf1f: 0x4000, 0xf20: 0x4000, 0xf21: 0x4000, 0xf22: 0x4000, 0xf23: 0x4000, + 0xf24: 0x4000, 0xf25: 0x4000, 0xf26: 0x4000, 0xf27: 0x4000, 0xf28: 0x4000, 0xf29: 0x4000, + 0xf2a: 0x4000, 0xf2b: 0x4000, 0xf2c: 0x4000, 0xf2d: 0x4000, 0xf2e: 0x4000, 0xf2f: 0x4000, + 0xf30: 0x4000, 0xf31: 0x4000, 0xf32: 0x4000, 0xf33: 0x4000, 0xf34: 0x4000, 0xf35: 0x4000, + 0xf36: 0x4000, 0xf37: 0x4000, 0xf38: 0x4000, 0xf39: 0x4000, 0xf3a: 0x4000, 0xf3b: 0x4000, + 0xf3c: 0x4000, 0xf3d: 0x4000, 0xf3e: 0x4000, + // Block 0x3d, offset 0xf40 + 0xf40: 0x4000, 0xf41: 0x4000, 0xf42: 0x4000, 0xf43: 0x4000, 0xf44: 0x4000, 0xf45: 0x4000, + 0xf46: 0x4000, 0xf47: 0x4000, 0xf48: 0x4000, 0xf49: 0x4000, 0xf4a: 0x4000, 0xf4b: 0x4000, + 0xf4c: 0x4000, 0xf50: 0x4000, 0xf51: 0x4000, + 0xf52: 0x4000, 0xf53: 0x4000, 0xf54: 0x4000, 0xf55: 0x4000, 0xf56: 0x4000, 0xf57: 0x4000, + 0xf58: 0x4000, 0xf59: 0x4000, 0xf5a: 0x4000, 0xf5b: 0x4000, 0xf5c: 0x4000, 0xf5d: 0x4000, + 0xf5e: 0x4000, 0xf5f: 0x4000, 0xf60: 0x4000, 0xf61: 0x4000, 0xf62: 0x4000, 0xf63: 0x4000, + 0xf64: 0x4000, 0xf65: 0x4000, 0xf66: 0x4000, 0xf67: 0x4000, 0xf68: 0x4000, 0xf69: 0x4000, + 0xf6a: 0x4000, 0xf6b: 0x4000, 0xf6c: 0x4000, 0xf6d: 0x4000, 0xf6e: 0x4000, 0xf6f: 0x4000, + 0xf70: 0x4000, 0xf71: 0x4000, 0xf72: 0x4000, 0xf73: 0x4000, 0xf74: 0x4000, 0xf75: 0x4000, + 0xf76: 0x4000, 0xf77: 0x4000, 0xf78: 0x4000, 0xf79: 0x4000, 0xf7a: 0x4000, 0xf7b: 0x4000, + 0xf7c: 0x4000, 0xf7d: 0x4000, 0xf7e: 0x4000, 0xf7f: 0x4000, + // Block 0x3e, offset 0xf80 + 0xf80: 0x4000, 0xf81: 0x4000, 0xf82: 0x4000, 0xf83: 0x4000, 0xf84: 0x4000, 0xf85: 0x4000, + 0xf86: 0x4000, + // Block 0x3f, offset 0xfc0 + 0xfe0: 0x4000, 0xfe1: 0x4000, 0xfe2: 0x4000, 0xfe3: 0x4000, + 0xfe4: 0x4000, 0xfe5: 0x4000, 0xfe6: 0x4000, 0xfe7: 0x4000, 0xfe8: 0x4000, 0xfe9: 0x4000, + 0xfea: 0x4000, 0xfeb: 0x4000, 0xfec: 0x4000, 0xfed: 0x4000, 0xfee: 0x4000, 0xfef: 0x4000, + 0xff0: 0x4000, 0xff1: 0x4000, 0xff2: 0x4000, 0xff3: 0x4000, 0xff4: 0x4000, 0xff5: 0x4000, + 0xff6: 0x4000, 0xff7: 0x4000, 0xff8: 0x4000, 0xff9: 0x4000, 0xffa: 0x4000, 0xffb: 0x4000, + 0xffc: 0x4000, + // Block 0x40, offset 0x1000 + 0x1000: 0x4000, 0x1001: 0x4000, 0x1002: 0x4000, 0x1003: 0x4000, 0x1004: 0x4000, 0x1005: 0x4000, + 0x1006: 0x4000, 0x1007: 0x4000, 0x1008: 0x4000, 0x1009: 0x4000, 0x100a: 0x4000, 0x100b: 0x4000, + 0x100c: 0x4000, 0x100d: 0x4000, 0x100e: 0x4000, 0x100f: 0x4000, 0x1010: 0x4000, 0x1011: 0x4000, + 0x1012: 0x4000, 0x1013: 0x4000, 0x1014: 0x4000, 0x1015: 0x4000, 0x1016: 0x4000, 0x1017: 0x4000, + 0x1018: 0x4000, 0x1019: 0x4000, 0x101a: 0x4000, 0x101b: 0x4000, 0x101c: 0x4000, 0x101d: 0x4000, + 0x101e: 0x4000, 0x101f: 0x4000, 0x1020: 0x4000, 0x1021: 0x4000, 0x1022: 0x4000, 0x1023: 0x4000, + // Block 0x41, offset 0x1040 + 0x1040: 0x2000, 0x1041: 0x2000, 0x1042: 0x2000, 0x1043: 0x2000, 0x1044: 0x2000, 0x1045: 0x2000, + 0x1046: 0x2000, 0x1047: 0x2000, 0x1048: 0x2000, 0x1049: 0x2000, 0x104a: 0x2000, 0x104b: 0x2000, + 0x104c: 0x2000, 0x104d: 0x2000, 0x104e: 0x2000, 0x104f: 0x2000, 0x1050: 0x4000, 0x1051: 0x4000, + 0x1052: 0x4000, 0x1053: 0x4000, 0x1054: 0x4000, 0x1055: 0x4000, 0x1056: 0x4000, 0x1057: 0x4000, + 0x1058: 0x4000, 0x1059: 0x4000, + 0x1070: 0x4000, 0x1071: 0x4000, 0x1072: 0x4000, 0x1073: 0x4000, 0x1074: 0x4000, 0x1075: 0x4000, + 0x1076: 0x4000, 0x1077: 0x4000, 0x1078: 0x4000, 0x1079: 0x4000, 0x107a: 0x4000, 0x107b: 0x4000, + 0x107c: 0x4000, 0x107d: 0x4000, 0x107e: 0x4000, 0x107f: 0x4000, + // Block 0x42, offset 0x1080 + 0x1080: 0x4000, 0x1081: 0x4000, 0x1082: 0x4000, 0x1083: 0x4000, 0x1084: 0x4000, 0x1085: 0x4000, + 0x1086: 0x4000, 0x1087: 0x4000, 0x1088: 0x4000, 0x1089: 0x4000, 0x108a: 0x4000, 0x108b: 0x4000, + 0x108c: 0x4000, 0x108d: 0x4000, 0x108e: 0x4000, 0x108f: 0x4000, 0x1090: 0x4000, 0x1091: 0x4000, + 0x1092: 0x4000, 0x1094: 0x4000, 0x1095: 0x4000, 0x1096: 0x4000, 0x1097: 0x4000, + 0x1098: 0x4000, 0x1099: 0x4000, 0x109a: 0x4000, 0x109b: 0x4000, 0x109c: 0x4000, 0x109d: 0x4000, + 0x109e: 0x4000, 0x109f: 0x4000, 0x10a0: 0x4000, 0x10a1: 0x4000, 0x10a2: 0x4000, 0x10a3: 0x4000, + 0x10a4: 0x4000, 0x10a5: 0x4000, 0x10a6: 0x4000, 0x10a8: 0x4000, 0x10a9: 0x4000, + 0x10aa: 0x4000, 0x10ab: 0x4000, + // Block 0x43, offset 0x10c0 + 0x10c1: 0x9012, 0x10c2: 0x9012, 0x10c3: 0x9012, 0x10c4: 0x9012, 0x10c5: 0x9012, + 0x10c6: 0x9012, 0x10c7: 0x9012, 0x10c8: 0x9012, 0x10c9: 0x9012, 0x10ca: 0x9012, 0x10cb: 0x9012, + 0x10cc: 0x9012, 0x10cd: 0x9012, 0x10ce: 0x9012, 0x10cf: 0x9012, 0x10d0: 0x9012, 0x10d1: 0x9012, + 0x10d2: 0x9012, 0x10d3: 0x9012, 0x10d4: 0x9012, 0x10d5: 0x9012, 0x10d6: 0x9012, 0x10d7: 0x9012, + 0x10d8: 0x9012, 0x10d9: 0x9012, 0x10da: 0x9012, 0x10db: 0x9012, 0x10dc: 0x9012, 0x10dd: 0x9012, + 0x10de: 0x9012, 0x10df: 0x9012, 0x10e0: 0x9049, 0x10e1: 0x9049, 0x10e2: 0x9049, 0x10e3: 0x9049, + 0x10e4: 0x9049, 0x10e5: 0x9049, 0x10e6: 0x9049, 0x10e7: 0x9049, 0x10e8: 0x9049, 0x10e9: 0x9049, + 0x10ea: 0x9049, 0x10eb: 0x9049, 0x10ec: 0x9049, 0x10ed: 0x9049, 0x10ee: 0x9049, 0x10ef: 0x9049, + 0x10f0: 0x9049, 0x10f1: 0x9049, 0x10f2: 0x9049, 0x10f3: 0x9049, 0x10f4: 0x9049, 0x10f5: 0x9049, + 0x10f6: 0x9049, 0x10f7: 0x9049, 0x10f8: 0x9049, 0x10f9: 0x9049, 0x10fa: 0x9049, 0x10fb: 0x9049, + 0x10fc: 0x9049, 0x10fd: 0x9049, 0x10fe: 0x9049, 0x10ff: 0x9049, + // Block 0x44, offset 0x1100 + 0x1100: 0x9049, 0x1101: 0x9049, 0x1102: 0x9049, 0x1103: 0x9049, 0x1104: 0x9049, 0x1105: 0x9049, + 0x1106: 0x9049, 0x1107: 0x9049, 0x1108: 0x9049, 0x1109: 0x9049, 0x110a: 0x9049, 0x110b: 0x9049, + 0x110c: 0x9049, 0x110d: 0x9049, 0x110e: 0x9049, 0x110f: 0x9049, 0x1110: 0x9049, 0x1111: 0x9049, + 0x1112: 0x9049, 0x1113: 0x9049, 0x1114: 0x9049, 0x1115: 0x9049, 0x1116: 0x9049, 0x1117: 0x9049, + 0x1118: 0x9049, 0x1119: 0x9049, 0x111a: 0x9049, 0x111b: 0x9049, 0x111c: 0x9049, 0x111d: 0x9049, + 0x111e: 0x9049, 0x111f: 0x904a, 0x1120: 0x904b, 0x1121: 0xb04c, 0x1122: 0xb04d, 0x1123: 0xb04d, + 0x1124: 0xb04e, 0x1125: 0xb04f, 0x1126: 0xb050, 0x1127: 0xb051, 0x1128: 0xb052, 0x1129: 0xb053, + 0x112a: 0xb054, 0x112b: 0xb055, 0x112c: 0xb056, 0x112d: 0xb057, 0x112e: 0xb058, 0x112f: 0xb059, + 0x1130: 0xb05a, 0x1131: 0xb05b, 0x1132: 0xb05c, 0x1133: 0xb05d, 0x1134: 0xb05e, 0x1135: 0xb05f, + 0x1136: 0xb060, 0x1137: 0xb061, 0x1138: 0xb062, 0x1139: 0xb063, 0x113a: 0xb064, 0x113b: 0xb065, + 0x113c: 0xb052, 0x113d: 0xb066, 0x113e: 0xb067, 0x113f: 0xb055, + // Block 0x45, offset 0x1140 + 0x1140: 0xb068, 0x1141: 0xb069, 0x1142: 0xb06a, 0x1143: 0xb06b, 0x1144: 0xb05a, 0x1145: 0xb056, + 0x1146: 0xb06c, 0x1147: 0xb06d, 0x1148: 0xb06b, 0x1149: 0xb06e, 0x114a: 0xb06b, 0x114b: 0xb06f, + 0x114c: 0xb06f, 0x114d: 0xb070, 0x114e: 0xb070, 0x114f: 0xb071, 0x1150: 0xb056, 0x1151: 0xb072, + 0x1152: 0xb073, 0x1153: 0xb072, 0x1154: 0xb074, 0x1155: 0xb073, 0x1156: 0xb075, 0x1157: 0xb075, + 0x1158: 0xb076, 0x1159: 0xb076, 0x115a: 0xb077, 0x115b: 0xb077, 0x115c: 0xb073, 0x115d: 0xb078, + 0x115e: 0xb079, 0x115f: 0xb067, 0x1160: 0xb07a, 0x1161: 0xb07b, 0x1162: 0xb07b, 0x1163: 0xb07b, + 0x1164: 0xb07b, 0x1165: 0xb07b, 0x1166: 0xb07b, 0x1167: 0xb07b, 0x1168: 0xb07b, 0x1169: 0xb07b, + 0x116a: 0xb07b, 0x116b: 0xb07b, 0x116c: 0xb07b, 0x116d: 0xb07b, 0x116e: 0xb07b, 0x116f: 0xb07b, + 0x1170: 0xb07c, 0x1171: 0xb07c, 0x1172: 0xb07c, 0x1173: 0xb07c, 0x1174: 0xb07c, 0x1175: 0xb07c, + 0x1176: 0xb07c, 0x1177: 0xb07c, 0x1178: 0xb07c, 0x1179: 0xb07c, 0x117a: 0xb07c, 0x117b: 0xb07c, + 0x117c: 0xb07c, 0x117d: 0xb07c, 0x117e: 0xb07c, + // Block 0x46, offset 0x1180 + 0x1182: 0xb07d, 0x1183: 0xb07e, 0x1184: 0xb07f, 0x1185: 0xb080, + 0x1186: 0xb07f, 0x1187: 0xb07e, 0x118a: 0xb081, 0x118b: 0xb082, + 0x118c: 0xb083, 0x118d: 0xb07f, 0x118e: 0xb080, 0x118f: 0xb07f, + 0x1192: 0xb084, 0x1193: 0xb085, 0x1194: 0xb084, 0x1195: 0xb086, 0x1196: 0xb084, 0x1197: 0xb087, + 0x119a: 0xb088, 0x119b: 0xb089, 0x119c: 0xb08a, + 0x11a0: 0x908b, 0x11a1: 0x908b, 0x11a2: 0x908c, 0x11a3: 0x908d, + 0x11a4: 0x908b, 0x11a5: 0x908e, 0x11a6: 0x908f, 0x11a8: 0xb090, 0x11a9: 0xb091, + 0x11aa: 0xb092, 0x11ab: 0xb091, 0x11ac: 0xb093, 0x11ad: 0xb094, 0x11ae: 0xb095, + 0x11bd: 0x2000, + // Block 0x47, offset 0x11c0 + 0x11e0: 0x4000, 0x11e1: 0x4000, 0x11e2: 0x4000, 0x11e3: 0x4000, + // Block 0x48, offset 0x1200 + 0x1200: 0x4000, 0x1201: 0x4000, 0x1202: 0x4000, 0x1203: 0x4000, 0x1204: 0x4000, 0x1205: 0x4000, + 0x1206: 0x4000, 0x1207: 0x4000, 0x1208: 0x4000, 0x1209: 0x4000, 0x120a: 0x4000, 0x120b: 0x4000, + 0x120c: 0x4000, 0x120d: 0x4000, 0x120e: 0x4000, 0x120f: 0x4000, 0x1210: 0x4000, 0x1211: 0x4000, + 0x1212: 0x4000, 0x1213: 0x4000, 0x1214: 0x4000, 0x1215: 0x4000, 0x1216: 0x4000, 0x1217: 0x4000, + 0x1218: 0x4000, 0x1219: 0x4000, 0x121a: 0x4000, 0x121b: 0x4000, 0x121c: 0x4000, 0x121d: 0x4000, + 0x121e: 0x4000, 0x121f: 0x4000, 0x1220: 0x4000, 0x1221: 0x4000, 0x1222: 0x4000, 0x1223: 0x4000, + 0x1224: 0x4000, 0x1225: 0x4000, 0x1226: 0x4000, 0x1227: 0x4000, 0x1228: 0x4000, 0x1229: 0x4000, + 0x122a: 0x4000, 0x122b: 0x4000, 0x122c: 0x4000, 0x122d: 0x4000, 0x122e: 0x4000, 0x122f: 0x4000, + 0x1230: 0x4000, 0x1231: 0x4000, 0x1232: 0x4000, 0x1233: 0x4000, 0x1234: 0x4000, 0x1235: 0x4000, + 0x1236: 0x4000, 0x1237: 0x4000, + // Block 0x49, offset 0x1240 + 0x1240: 0x4000, 0x1241: 0x4000, 0x1242: 0x4000, 0x1243: 0x4000, 0x1244: 0x4000, 0x1245: 0x4000, + 0x1246: 0x4000, 0x1247: 0x4000, 0x1248: 0x4000, 0x1249: 0x4000, 0x124a: 0x4000, 0x124b: 0x4000, + 0x124c: 0x4000, 0x124d: 0x4000, 0x124e: 0x4000, 0x124f: 0x4000, 0x1250: 0x4000, 0x1251: 0x4000, + 0x1252: 0x4000, 0x1253: 0x4000, 0x1254: 0x4000, 0x1255: 0x4000, 0x1256: 0x4000, 0x1257: 0x4000, + 0x1258: 0x4000, 0x1259: 0x4000, 0x125a: 0x4000, 0x125b: 0x4000, 0x125c: 0x4000, 0x125d: 0x4000, + 0x125e: 0x4000, 0x125f: 0x4000, 0x1260: 0x4000, 0x1261: 0x4000, 0x1262: 0x4000, 0x1263: 0x4000, + 0x1264: 0x4000, 0x1265: 0x4000, 0x1266: 0x4000, 0x1267: 0x4000, 0x1268: 0x4000, 0x1269: 0x4000, + 0x126a: 0x4000, 0x126b: 0x4000, 0x126c: 0x4000, 0x126d: 0x4000, 0x126e: 0x4000, 0x126f: 0x4000, + 0x1270: 0x4000, 0x1271: 0x4000, 0x1272: 0x4000, + // Block 0x4a, offset 0x1280 + 0x1280: 0x4000, 0x1281: 0x4000, 0x1282: 0x4000, 0x1283: 0x4000, 0x1284: 0x4000, 0x1285: 0x4000, + 0x1286: 0x4000, 0x1287: 0x4000, 0x1288: 0x4000, 0x1289: 0x4000, 0x128a: 0x4000, 0x128b: 0x4000, + 0x128c: 0x4000, 0x128d: 0x4000, 0x128e: 0x4000, 0x128f: 0x4000, 0x1290: 0x4000, 0x1291: 0x4000, + 0x1292: 0x4000, 0x1293: 0x4000, 0x1294: 0x4000, 0x1295: 0x4000, 0x1296: 0x4000, 0x1297: 0x4000, + 0x1298: 0x4000, 0x1299: 0x4000, 0x129a: 0x4000, 0x129b: 0x4000, 0x129c: 0x4000, 0x129d: 0x4000, + 0x129e: 0x4000, + // Block 0x4b, offset 0x12c0 + 0x12d0: 0x4000, 0x12d1: 0x4000, + 0x12d2: 0x4000, + 0x12e4: 0x4000, 0x12e5: 0x4000, 0x12e6: 0x4000, 0x12e7: 0x4000, + 0x12f0: 0x4000, 0x12f1: 0x4000, 0x12f2: 0x4000, 0x12f3: 0x4000, 0x12f4: 0x4000, 0x12f5: 0x4000, + 0x12f6: 0x4000, 0x12f7: 0x4000, 0x12f8: 0x4000, 0x12f9: 0x4000, 0x12fa: 0x4000, 0x12fb: 0x4000, + 0x12fc: 0x4000, 0x12fd: 0x4000, 0x12fe: 0x4000, 0x12ff: 0x4000, + // Block 0x4c, offset 0x1300 + 0x1300: 0x4000, 0x1301: 0x4000, 0x1302: 0x4000, 0x1303: 0x4000, 0x1304: 0x4000, 0x1305: 0x4000, + 0x1306: 0x4000, 0x1307: 0x4000, 0x1308: 0x4000, 0x1309: 0x4000, 0x130a: 0x4000, 0x130b: 0x4000, + 0x130c: 0x4000, 0x130d: 0x4000, 0x130e: 0x4000, 0x130f: 0x4000, 0x1310: 0x4000, 0x1311: 0x4000, + 0x1312: 0x4000, 0x1313: 0x4000, 0x1314: 0x4000, 0x1315: 0x4000, 0x1316: 0x4000, 0x1317: 0x4000, + 0x1318: 0x4000, 0x1319: 0x4000, 0x131a: 0x4000, 0x131b: 0x4000, 0x131c: 0x4000, 0x131d: 0x4000, + 0x131e: 0x4000, 0x131f: 0x4000, 0x1320: 0x4000, 0x1321: 0x4000, 0x1322: 0x4000, 0x1323: 0x4000, + 0x1324: 0x4000, 0x1325: 0x4000, 0x1326: 0x4000, 0x1327: 0x4000, 0x1328: 0x4000, 0x1329: 0x4000, + 0x132a: 0x4000, 0x132b: 0x4000, 0x132c: 0x4000, 0x132d: 0x4000, 0x132e: 0x4000, 0x132f: 0x4000, + 0x1330: 0x4000, 0x1331: 0x4000, 0x1332: 0x4000, 0x1333: 0x4000, 0x1334: 0x4000, 0x1335: 0x4000, + 0x1336: 0x4000, 0x1337: 0x4000, 0x1338: 0x4000, 0x1339: 0x4000, 0x133a: 0x4000, 0x133b: 0x4000, + // Block 0x4d, offset 0x1340 + 0x1344: 0x4000, + // Block 0x4e, offset 0x1380 + 0x138f: 0x4000, + // Block 0x4f, offset 0x13c0 + 0x13c0: 0x2000, 0x13c1: 0x2000, 0x13c2: 0x2000, 0x13c3: 0x2000, 0x13c4: 0x2000, 0x13c5: 0x2000, + 0x13c6: 0x2000, 0x13c7: 0x2000, 0x13c8: 0x2000, 0x13c9: 0x2000, 0x13ca: 0x2000, + 0x13d0: 0x2000, 0x13d1: 0x2000, + 0x13d2: 0x2000, 0x13d3: 0x2000, 0x13d4: 0x2000, 0x13d5: 0x2000, 0x13d6: 0x2000, 0x13d7: 0x2000, + 0x13d8: 0x2000, 0x13d9: 0x2000, 0x13da: 0x2000, 0x13db: 0x2000, 0x13dc: 0x2000, 0x13dd: 0x2000, + 0x13de: 0x2000, 0x13df: 0x2000, 0x13e0: 0x2000, 0x13e1: 0x2000, 0x13e2: 0x2000, 0x13e3: 0x2000, + 0x13e4: 0x2000, 0x13e5: 0x2000, 0x13e6: 0x2000, 0x13e7: 0x2000, 0x13e8: 0x2000, 0x13e9: 0x2000, + 0x13ea: 0x2000, 0x13eb: 0x2000, 0x13ec: 0x2000, 0x13ed: 0x2000, + 0x13f0: 0x2000, 0x13f1: 0x2000, 0x13f2: 0x2000, 0x13f3: 0x2000, 0x13f4: 0x2000, 0x13f5: 0x2000, + 0x13f6: 0x2000, 0x13f7: 0x2000, 0x13f8: 0x2000, 0x13f9: 0x2000, 0x13fa: 0x2000, 0x13fb: 0x2000, + 0x13fc: 0x2000, 0x13fd: 0x2000, 0x13fe: 0x2000, 0x13ff: 0x2000, + // Block 0x50, offset 0x1400 + 0x1400: 0x2000, 0x1401: 0x2000, 0x1402: 0x2000, 0x1403: 0x2000, 0x1404: 0x2000, 0x1405: 0x2000, + 0x1406: 0x2000, 0x1407: 0x2000, 0x1408: 0x2000, 0x1409: 0x2000, 0x140a: 0x2000, 0x140b: 0x2000, + 0x140c: 0x2000, 0x140d: 0x2000, 0x140e: 0x2000, 0x140f: 0x2000, 0x1410: 0x2000, 0x1411: 0x2000, + 0x1412: 0x2000, 0x1413: 0x2000, 0x1414: 0x2000, 0x1415: 0x2000, 0x1416: 0x2000, 0x1417: 0x2000, + 0x1418: 0x2000, 0x1419: 0x2000, 0x141a: 0x2000, 0x141b: 0x2000, 0x141c: 0x2000, 0x141d: 0x2000, + 0x141e: 0x2000, 0x141f: 0x2000, 0x1420: 0x2000, 0x1421: 0x2000, 0x1422: 0x2000, 0x1423: 0x2000, + 0x1424: 0x2000, 0x1425: 0x2000, 0x1426: 0x2000, 0x1427: 0x2000, 0x1428: 0x2000, 0x1429: 0x2000, + 0x1430: 0x2000, 0x1431: 0x2000, 0x1432: 0x2000, 0x1433: 0x2000, 0x1434: 0x2000, 0x1435: 0x2000, + 0x1436: 0x2000, 0x1437: 0x2000, 0x1438: 0x2000, 0x1439: 0x2000, 0x143a: 0x2000, 0x143b: 0x2000, + 0x143c: 0x2000, 0x143d: 0x2000, 0x143e: 0x2000, 0x143f: 0x2000, + // Block 0x51, offset 0x1440 + 0x1440: 0x2000, 0x1441: 0x2000, 0x1442: 0x2000, 0x1443: 0x2000, 0x1444: 0x2000, 0x1445: 0x2000, + 0x1446: 0x2000, 0x1447: 0x2000, 0x1448: 0x2000, 0x1449: 0x2000, 0x144a: 0x2000, 0x144b: 0x2000, + 0x144c: 0x2000, 0x144d: 0x2000, 0x144e: 0x4000, 0x144f: 0x2000, 0x1450: 0x2000, 0x1451: 0x4000, + 0x1452: 0x4000, 0x1453: 0x4000, 0x1454: 0x4000, 0x1455: 0x4000, 0x1456: 0x4000, 0x1457: 0x4000, + 0x1458: 0x4000, 0x1459: 0x4000, 0x145a: 0x4000, 0x145b: 0x2000, 0x145c: 0x2000, 0x145d: 0x2000, + 0x145e: 0x2000, 0x145f: 0x2000, 0x1460: 0x2000, 0x1461: 0x2000, 0x1462: 0x2000, 0x1463: 0x2000, + 0x1464: 0x2000, 0x1465: 0x2000, 0x1466: 0x2000, 0x1467: 0x2000, 0x1468: 0x2000, 0x1469: 0x2000, + 0x146a: 0x2000, 0x146b: 0x2000, 0x146c: 0x2000, + // Block 0x52, offset 0x1480 + 0x1480: 0x4000, 0x1481: 0x4000, 0x1482: 0x4000, + 0x1490: 0x4000, 0x1491: 0x4000, + 0x1492: 0x4000, 0x1493: 0x4000, 0x1494: 0x4000, 0x1495: 0x4000, 0x1496: 0x4000, 0x1497: 0x4000, + 0x1498: 0x4000, 0x1499: 0x4000, 0x149a: 0x4000, 0x149b: 0x4000, 0x149c: 0x4000, 0x149d: 0x4000, + 0x149e: 0x4000, 0x149f: 0x4000, 0x14a0: 0x4000, 0x14a1: 0x4000, 0x14a2: 0x4000, 0x14a3: 0x4000, + 0x14a4: 0x4000, 0x14a5: 0x4000, 0x14a6: 0x4000, 0x14a7: 0x4000, 0x14a8: 0x4000, 0x14a9: 0x4000, + 0x14aa: 0x4000, 0x14ab: 0x4000, 0x14ac: 0x4000, 0x14ad: 0x4000, 0x14ae: 0x4000, 0x14af: 0x4000, + 0x14b0: 0x4000, 0x14b1: 0x4000, 0x14b2: 0x4000, 0x14b3: 0x4000, 0x14b4: 0x4000, 0x14b5: 0x4000, + 0x14b6: 0x4000, 0x14b7: 0x4000, 0x14b8: 0x4000, 0x14b9: 0x4000, 0x14ba: 0x4000, 0x14bb: 0x4000, + // Block 0x53, offset 0x14c0 + 0x14c0: 0x4000, 0x14c1: 0x4000, 0x14c2: 0x4000, 0x14c3: 0x4000, 0x14c4: 0x4000, 0x14c5: 0x4000, + 0x14c6: 0x4000, 0x14c7: 0x4000, 0x14c8: 0x4000, + 0x14d0: 0x4000, 0x14d1: 0x4000, + 0x14e0: 0x4000, 0x14e1: 0x4000, 0x14e2: 0x4000, 0x14e3: 0x4000, + 0x14e4: 0x4000, 0x14e5: 0x4000, + // Block 0x54, offset 0x1500 + 0x1500: 0x4000, 0x1501: 0x4000, 0x1502: 0x4000, 0x1503: 0x4000, 0x1504: 0x4000, 0x1505: 0x4000, + 0x1506: 0x4000, 0x1507: 0x4000, 0x1508: 0x4000, 0x1509: 0x4000, 0x150a: 0x4000, 0x150b: 0x4000, + 0x150c: 0x4000, 0x150d: 0x4000, 0x150e: 0x4000, 0x150f: 0x4000, 0x1510: 0x4000, 0x1511: 0x4000, + 0x1512: 0x4000, 0x1513: 0x4000, 0x1514: 0x4000, 0x1515: 0x4000, 0x1516: 0x4000, 0x1517: 0x4000, + 0x1518: 0x4000, 0x1519: 0x4000, 0x151a: 0x4000, 0x151b: 0x4000, 0x151c: 0x4000, 0x151d: 0x4000, + 0x151e: 0x4000, 0x151f: 0x4000, 0x1520: 0x4000, + 0x152d: 0x4000, 0x152e: 0x4000, 0x152f: 0x4000, + 0x1530: 0x4000, 0x1531: 0x4000, 0x1532: 0x4000, 0x1533: 0x4000, 0x1534: 0x4000, 0x1535: 0x4000, + 0x1537: 0x4000, 0x1538: 0x4000, 0x1539: 0x4000, 0x153a: 0x4000, 0x153b: 0x4000, + 0x153c: 0x4000, 0x153d: 0x4000, 0x153e: 0x4000, 0x153f: 0x4000, + // Block 0x55, offset 0x1540 + 0x1540: 0x4000, 0x1541: 0x4000, 0x1542: 0x4000, 0x1543: 0x4000, 0x1544: 0x4000, 0x1545: 0x4000, + 0x1546: 0x4000, 0x1547: 0x4000, 0x1548: 0x4000, 0x1549: 0x4000, 0x154a: 0x4000, 0x154b: 0x4000, + 0x154c: 0x4000, 0x154d: 0x4000, 0x154e: 0x4000, 0x154f: 0x4000, 0x1550: 0x4000, 0x1551: 0x4000, + 0x1552: 0x4000, 0x1553: 0x4000, 0x1554: 0x4000, 0x1555: 0x4000, 0x1556: 0x4000, 0x1557: 0x4000, + 0x1558: 0x4000, 0x1559: 0x4000, 0x155a: 0x4000, 0x155b: 0x4000, 0x155c: 0x4000, 0x155d: 0x4000, + 0x155e: 0x4000, 0x155f: 0x4000, 0x1560: 0x4000, 0x1561: 0x4000, 0x1562: 0x4000, 0x1563: 0x4000, + 0x1564: 0x4000, 0x1565: 0x4000, 0x1566: 0x4000, 0x1567: 0x4000, 0x1568: 0x4000, 0x1569: 0x4000, + 0x156a: 0x4000, 0x156b: 0x4000, 0x156c: 0x4000, 0x156d: 0x4000, 0x156e: 0x4000, 0x156f: 0x4000, + 0x1570: 0x4000, 0x1571: 0x4000, 0x1572: 0x4000, 0x1573: 0x4000, 0x1574: 0x4000, 0x1575: 0x4000, + 0x1576: 0x4000, 0x1577: 0x4000, 0x1578: 0x4000, 0x1579: 0x4000, 0x157a: 0x4000, 0x157b: 0x4000, + 0x157c: 0x4000, 0x157e: 0x4000, 0x157f: 0x4000, + // Block 0x56, offset 0x1580 + 0x1580: 0x4000, 0x1581: 0x4000, 0x1582: 0x4000, 0x1583: 0x4000, 0x1584: 0x4000, 0x1585: 0x4000, + 0x1586: 0x4000, 0x1587: 0x4000, 0x1588: 0x4000, 0x1589: 0x4000, 0x158a: 0x4000, 0x158b: 0x4000, + 0x158c: 0x4000, 0x158d: 0x4000, 0x158e: 0x4000, 0x158f: 0x4000, 0x1590: 0x4000, 0x1591: 0x4000, + 0x1592: 0x4000, 0x1593: 0x4000, + 0x15a0: 0x4000, 0x15a1: 0x4000, 0x15a2: 0x4000, 0x15a3: 0x4000, + 0x15a4: 0x4000, 0x15a5: 0x4000, 0x15a6: 0x4000, 0x15a7: 0x4000, 0x15a8: 0x4000, 0x15a9: 0x4000, + 0x15aa: 0x4000, 0x15ab: 0x4000, 0x15ac: 0x4000, 0x15ad: 0x4000, 0x15ae: 0x4000, 0x15af: 0x4000, + 0x15b0: 0x4000, 0x15b1: 0x4000, 0x15b2: 0x4000, 0x15b3: 0x4000, 0x15b4: 0x4000, 0x15b5: 0x4000, + 0x15b6: 0x4000, 0x15b7: 0x4000, 0x15b8: 0x4000, 0x15b9: 0x4000, 0x15ba: 0x4000, 0x15bb: 0x4000, + 0x15bc: 0x4000, 0x15bd: 0x4000, 0x15be: 0x4000, 0x15bf: 0x4000, + // Block 0x57, offset 0x15c0 + 0x15c0: 0x4000, 0x15c1: 0x4000, 0x15c2: 0x4000, 0x15c3: 0x4000, 0x15c4: 0x4000, 0x15c5: 0x4000, + 0x15c6: 0x4000, 0x15c7: 0x4000, 0x15c8: 0x4000, 0x15c9: 0x4000, 0x15ca: 0x4000, + 0x15cf: 0x4000, 0x15d0: 0x4000, 0x15d1: 0x4000, + 0x15d2: 0x4000, 0x15d3: 0x4000, + 0x15e0: 0x4000, 0x15e1: 0x4000, 0x15e2: 0x4000, 0x15e3: 0x4000, + 0x15e4: 0x4000, 0x15e5: 0x4000, 0x15e6: 0x4000, 0x15e7: 0x4000, 0x15e8: 0x4000, 0x15e9: 0x4000, + 0x15ea: 0x4000, 0x15eb: 0x4000, 0x15ec: 0x4000, 0x15ed: 0x4000, 0x15ee: 0x4000, 0x15ef: 0x4000, + 0x15f0: 0x4000, 0x15f4: 0x4000, + 0x15f8: 0x4000, 0x15f9: 0x4000, 0x15fa: 0x4000, 0x15fb: 0x4000, + 0x15fc: 0x4000, 0x15fd: 0x4000, 0x15fe: 0x4000, 0x15ff: 0x4000, + // Block 0x58, offset 0x1600 + 0x1600: 0x4000, 0x1602: 0x4000, 0x1603: 0x4000, 0x1604: 0x4000, 0x1605: 0x4000, + 0x1606: 0x4000, 0x1607: 0x4000, 0x1608: 0x4000, 0x1609: 0x4000, 0x160a: 0x4000, 0x160b: 0x4000, + 0x160c: 0x4000, 0x160d: 0x4000, 0x160e: 0x4000, 0x160f: 0x4000, 0x1610: 0x4000, 0x1611: 0x4000, + 0x1612: 0x4000, 0x1613: 0x4000, 0x1614: 0x4000, 0x1615: 0x4000, 0x1616: 0x4000, 0x1617: 0x4000, + 0x1618: 0x4000, 0x1619: 0x4000, 0x161a: 0x4000, 0x161b: 0x4000, 0x161c: 0x4000, 0x161d: 0x4000, + 0x161e: 0x4000, 0x161f: 0x4000, 0x1620: 0x4000, 0x1621: 0x4000, 0x1622: 0x4000, 0x1623: 0x4000, + 0x1624: 0x4000, 0x1625: 0x4000, 0x1626: 0x4000, 0x1627: 0x4000, 0x1628: 0x4000, 0x1629: 0x4000, + 0x162a: 0x4000, 0x162b: 0x4000, 0x162c: 0x4000, 0x162d: 0x4000, 0x162e: 0x4000, 0x162f: 0x4000, + 0x1630: 0x4000, 0x1631: 0x4000, 0x1632: 0x4000, 0x1633: 0x4000, 0x1634: 0x4000, 0x1635: 0x4000, + 0x1636: 0x4000, 0x1637: 0x4000, 0x1638: 0x4000, 0x1639: 0x4000, 0x163a: 0x4000, 0x163b: 0x4000, + 0x163c: 0x4000, 0x163d: 0x4000, 0x163e: 0x4000, 0x163f: 0x4000, + // Block 0x59, offset 0x1640 + 0x1640: 0x4000, 0x1641: 0x4000, 0x1642: 0x4000, 0x1643: 0x4000, 0x1644: 0x4000, 0x1645: 0x4000, + 0x1646: 0x4000, 0x1647: 0x4000, 0x1648: 0x4000, 0x1649: 0x4000, 0x164a: 0x4000, 0x164b: 0x4000, + 0x164c: 0x4000, 0x164d: 0x4000, 0x164e: 0x4000, 0x164f: 0x4000, 0x1650: 0x4000, 0x1651: 0x4000, + 0x1652: 0x4000, 0x1653: 0x4000, 0x1654: 0x4000, 0x1655: 0x4000, 0x1656: 0x4000, 0x1657: 0x4000, + 0x1658: 0x4000, 0x1659: 0x4000, 0x165a: 0x4000, 0x165b: 0x4000, 0x165c: 0x4000, 0x165d: 0x4000, + 0x165e: 0x4000, 0x165f: 0x4000, 0x1660: 0x4000, 0x1661: 0x4000, 0x1662: 0x4000, 0x1663: 0x4000, + 0x1664: 0x4000, 0x1665: 0x4000, 0x1666: 0x4000, 0x1667: 0x4000, 0x1668: 0x4000, 0x1669: 0x4000, + 0x166a: 0x4000, 0x166b: 0x4000, 0x166c: 0x4000, 0x166d: 0x4000, 0x166e: 0x4000, 0x166f: 0x4000, + 0x1670: 0x4000, 0x1671: 0x4000, 0x1672: 0x4000, 0x1673: 0x4000, 0x1674: 0x4000, 0x1675: 0x4000, + 0x1676: 0x4000, 0x1677: 0x4000, 0x1678: 0x4000, 0x1679: 0x4000, 0x167a: 0x4000, 0x167b: 0x4000, + 0x167c: 0x4000, 0x167f: 0x4000, + // Block 0x5a, offset 0x1680 + 0x1680: 0x4000, 0x1681: 0x4000, 0x1682: 0x4000, 0x1683: 0x4000, 0x1684: 0x4000, 0x1685: 0x4000, + 0x1686: 0x4000, 0x1687: 0x4000, 0x1688: 0x4000, 0x1689: 0x4000, 0x168a: 0x4000, 0x168b: 0x4000, + 0x168c: 0x4000, 0x168d: 0x4000, 0x168e: 0x4000, 0x168f: 0x4000, 0x1690: 0x4000, 0x1691: 0x4000, + 0x1692: 0x4000, 0x1693: 0x4000, 0x1694: 0x4000, 0x1695: 0x4000, 0x1696: 0x4000, 0x1697: 0x4000, + 0x1698: 0x4000, 0x1699: 0x4000, 0x169a: 0x4000, 0x169b: 0x4000, 0x169c: 0x4000, 0x169d: 0x4000, + 0x169e: 0x4000, 0x169f: 0x4000, 0x16a0: 0x4000, 0x16a1: 0x4000, 0x16a2: 0x4000, 0x16a3: 0x4000, + 0x16a4: 0x4000, 0x16a5: 0x4000, 0x16a6: 0x4000, 0x16a7: 0x4000, 0x16a8: 0x4000, 0x16a9: 0x4000, + 0x16aa: 0x4000, 0x16ab: 0x4000, 0x16ac: 0x4000, 0x16ad: 0x4000, 0x16ae: 0x4000, 0x16af: 0x4000, + 0x16b0: 0x4000, 0x16b1: 0x4000, 0x16b2: 0x4000, 0x16b3: 0x4000, 0x16b4: 0x4000, 0x16b5: 0x4000, + 0x16b6: 0x4000, 0x16b7: 0x4000, 0x16b8: 0x4000, 0x16b9: 0x4000, 0x16ba: 0x4000, 0x16bb: 0x4000, + 0x16bc: 0x4000, 0x16bd: 0x4000, + // Block 0x5b, offset 0x16c0 + 0x16cb: 0x4000, + 0x16cc: 0x4000, 0x16cd: 0x4000, 0x16ce: 0x4000, 0x16d0: 0x4000, 0x16d1: 0x4000, + 0x16d2: 0x4000, 0x16d3: 0x4000, 0x16d4: 0x4000, 0x16d5: 0x4000, 0x16d6: 0x4000, 0x16d7: 0x4000, + 0x16d8: 0x4000, 0x16d9: 0x4000, 0x16da: 0x4000, 0x16db: 0x4000, 0x16dc: 0x4000, 0x16dd: 0x4000, + 0x16de: 0x4000, 0x16df: 0x4000, 0x16e0: 0x4000, 0x16e1: 0x4000, 0x16e2: 0x4000, 0x16e3: 0x4000, + 0x16e4: 0x4000, 0x16e5: 0x4000, 0x16e6: 0x4000, 0x16e7: 0x4000, + 0x16fa: 0x4000, + // Block 0x5c, offset 0x1700 + 0x1715: 0x4000, 0x1716: 0x4000, + 0x1724: 0x4000, + // Block 0x5d, offset 0x1740 + 0x177b: 0x4000, + 0x177c: 0x4000, 0x177d: 0x4000, 0x177e: 0x4000, 0x177f: 0x4000, + // Block 0x5e, offset 0x1780 + 0x1780: 0x4000, 0x1781: 0x4000, 0x1782: 0x4000, 0x1783: 0x4000, 0x1784: 0x4000, 0x1785: 0x4000, + 0x1786: 0x4000, 0x1787: 0x4000, 0x1788: 0x4000, 0x1789: 0x4000, 0x178a: 0x4000, 0x178b: 0x4000, + 0x178c: 0x4000, 0x178d: 0x4000, 0x178e: 0x4000, 0x178f: 0x4000, + // Block 0x5f, offset 0x17c0 + 0x17c0: 0x4000, 0x17c1: 0x4000, 0x17c2: 0x4000, 0x17c3: 0x4000, 0x17c4: 0x4000, 0x17c5: 0x4000, + 0x17cc: 0x4000, 0x17d0: 0x4000, 0x17d1: 0x4000, + 0x17d2: 0x4000, 0x17d5: 0x4000, + 0x17eb: 0x4000, 0x17ec: 0x4000, + 0x17f4: 0x4000, 0x17f5: 0x4000, + 0x17f6: 0x4000, 0x17f7: 0x4000, 0x17f8: 0x4000, 0x17f9: 0x4000, 0x17fa: 0x4000, + // Block 0x60, offset 0x1800 + 0x1820: 0x4000, 0x1821: 0x4000, 0x1822: 0x4000, 0x1823: 0x4000, + 0x1824: 0x4000, 0x1825: 0x4000, 0x1826: 0x4000, 0x1827: 0x4000, 0x1828: 0x4000, 0x1829: 0x4000, + 0x182a: 0x4000, 0x182b: 0x4000, + // Block 0x61, offset 0x1840 + 0x184d: 0x4000, 0x184e: 0x4000, 0x184f: 0x4000, 0x1850: 0x4000, 0x1851: 0x4000, + 0x1852: 0x4000, 0x1853: 0x4000, 0x1854: 0x4000, 0x1855: 0x4000, 0x1856: 0x4000, 0x1857: 0x4000, + 0x1858: 0x4000, 0x1859: 0x4000, 0x185a: 0x4000, 0x185b: 0x4000, 0x185c: 0x4000, 0x185d: 0x4000, + 0x185e: 0x4000, 0x185f: 0x4000, 0x1860: 0x4000, 0x1861: 0x4000, 0x1862: 0x4000, 0x1863: 0x4000, + 0x1864: 0x4000, 0x1865: 0x4000, 0x1866: 0x4000, 0x1867: 0x4000, 0x1868: 0x4000, 0x1869: 0x4000, + 0x186a: 0x4000, 0x186b: 0x4000, 0x186c: 0x4000, 0x186d: 0x4000, 0x186e: 0x4000, 0x186f: 0x4000, + 0x1870: 0x4000, 0x1871: 0x4000, 0x1872: 0x4000, 0x1873: 0x4000, 0x1874: 0x4000, 0x1875: 0x4000, + 0x1876: 0x4000, 0x1877: 0x4000, 0x1878: 0x4000, 0x1879: 0x4000, 0x187a: 0x4000, 0x187b: 0x4000, + 0x187c: 0x4000, 0x187d: 0x4000, 0x187e: 0x4000, 0x187f: 0x4000, + // Block 0x62, offset 0x1880 + 0x1880: 0x4000, 0x1881: 0x4000, 0x1882: 0x4000, 0x1883: 0x4000, 0x1884: 0x4000, 0x1885: 0x4000, + 0x1886: 0x4000, 0x1887: 0x4000, 0x1888: 0x4000, 0x1889: 0x4000, 0x188a: 0x4000, 0x188b: 0x4000, + 0x188c: 0x4000, 0x188d: 0x4000, 0x188e: 0x4000, 0x188f: 0x4000, 0x1890: 0x4000, 0x1891: 0x4000, + 0x1892: 0x4000, 0x1893: 0x4000, 0x1894: 0x4000, 0x1895: 0x4000, 0x1896: 0x4000, 0x1897: 0x4000, + 0x1898: 0x4000, 0x1899: 0x4000, 0x189a: 0x4000, 0x189b: 0x4000, 0x189c: 0x4000, 0x189d: 0x4000, + 0x189e: 0x4000, 0x189f: 0x4000, 0x18a0: 0x4000, 0x18a1: 0x4000, 0x18a2: 0x4000, 0x18a3: 0x4000, + 0x18a4: 0x4000, 0x18a5: 0x4000, 0x18a6: 0x4000, 0x18a7: 0x4000, 0x18a8: 0x4000, 0x18a9: 0x4000, + 0x18aa: 0x4000, 0x18ab: 0x4000, 0x18ac: 0x4000, 0x18ad: 0x4000, 0x18ae: 0x4000, 0x18af: 0x4000, + 0x18b0: 0x4000, 0x18b1: 0x4000, 0x18b3: 0x4000, 0x18b4: 0x4000, 0x18b5: 0x4000, + 0x18b6: 0x4000, 0x18ba: 0x4000, 0x18bb: 0x4000, + 0x18bc: 0x4000, 0x18bd: 0x4000, 0x18be: 0x4000, 0x18bf: 0x4000, + // Block 0x63, offset 0x18c0 + 0x18c0: 0x4000, 0x18c1: 0x4000, 0x18c2: 0x4000, 0x18c3: 0x4000, 0x18c4: 0x4000, 0x18c5: 0x4000, + 0x18c6: 0x4000, 0x18c7: 0x4000, 0x18c8: 0x4000, 0x18c9: 0x4000, 0x18ca: 0x4000, 0x18cb: 0x4000, + 0x18cc: 0x4000, 0x18cd: 0x4000, 0x18ce: 0x4000, 0x18cf: 0x4000, 0x18d0: 0x4000, 0x18d1: 0x4000, + 0x18d2: 0x4000, 0x18d3: 0x4000, 0x18d4: 0x4000, 0x18d5: 0x4000, 0x18d6: 0x4000, 0x18d7: 0x4000, + 0x18d8: 0x4000, 0x18d9: 0x4000, 0x18da: 0x4000, 0x18db: 0x4000, 0x18dc: 0x4000, 0x18dd: 0x4000, + 0x18de: 0x4000, 0x18df: 0x4000, 0x18e0: 0x4000, 0x18e1: 0x4000, 0x18e2: 0x4000, + 0x18e5: 0x4000, 0x18e6: 0x4000, 0x18e7: 0x4000, 0x18e8: 0x4000, 0x18e9: 0x4000, + 0x18ea: 0x4000, 0x18ee: 0x4000, 0x18ef: 0x4000, + 0x18f0: 0x4000, 0x18f1: 0x4000, 0x18f2: 0x4000, 0x18f3: 0x4000, 0x18f4: 0x4000, 0x18f5: 0x4000, + 0x18f6: 0x4000, 0x18f7: 0x4000, 0x18f8: 0x4000, 0x18f9: 0x4000, 0x18fa: 0x4000, 0x18fb: 0x4000, + 0x18fc: 0x4000, 0x18fd: 0x4000, 0x18fe: 0x4000, 0x18ff: 0x4000, + // Block 0x64, offset 0x1900 + 0x1900: 0x4000, 0x1901: 0x4000, 0x1902: 0x4000, 0x1903: 0x4000, 0x1904: 0x4000, 0x1905: 0x4000, + 0x1906: 0x4000, 0x1907: 0x4000, 0x1908: 0x4000, 0x1909: 0x4000, 0x190a: 0x4000, + 0x190d: 0x4000, 0x190e: 0x4000, 0x190f: 0x4000, 0x1910: 0x4000, 0x1911: 0x4000, + 0x1912: 0x4000, 0x1913: 0x4000, 0x1914: 0x4000, 0x1915: 0x4000, 0x1916: 0x4000, 0x1917: 0x4000, + 0x1918: 0x4000, 0x1919: 0x4000, 0x191a: 0x4000, 0x191b: 0x4000, 0x191c: 0x4000, 0x191d: 0x4000, + 0x191e: 0x4000, 0x191f: 0x4000, 0x1920: 0x4000, 0x1921: 0x4000, 0x1922: 0x4000, 0x1923: 0x4000, + 0x1924: 0x4000, 0x1925: 0x4000, 0x1926: 0x4000, 0x1927: 0x4000, 0x1928: 0x4000, 0x1929: 0x4000, + 0x192a: 0x4000, 0x192b: 0x4000, 0x192c: 0x4000, 0x192d: 0x4000, 0x192e: 0x4000, 0x192f: 0x4000, + 0x1930: 0x4000, 0x1931: 0x4000, 0x1932: 0x4000, 0x1933: 0x4000, 0x1934: 0x4000, 0x1935: 0x4000, + 0x1936: 0x4000, 0x1937: 0x4000, 0x1938: 0x4000, 0x1939: 0x4000, 0x193a: 0x4000, 0x193b: 0x4000, + 0x193c: 0x4000, 0x193d: 0x4000, 0x193e: 0x4000, 0x193f: 0x4000, + // Block 0x65, offset 0x1940 + 0x1970: 0x4000, 0x1971: 0x4000, 0x1972: 0x4000, 0x1973: 0x4000, + 0x1978: 0x4000, 0x1979: 0x4000, 0x197a: 0x4000, + // Block 0x66, offset 0x1980 + 0x1980: 0x4000, 0x1981: 0x4000, 0x1982: 0x4000, + 0x1990: 0x4000, 0x1991: 0x4000, + 0x1992: 0x4000, 0x1993: 0x4000, 0x1994: 0x4000, 0x1995: 0x4000, + // Block 0x67, offset 0x19c0 + 0x19c0: 0x2000, 0x19c1: 0x2000, 0x19c2: 0x2000, 0x19c3: 0x2000, 0x19c4: 0x2000, 0x19c5: 0x2000, + 0x19c6: 0x2000, 0x19c7: 0x2000, 0x19c8: 0x2000, 0x19c9: 0x2000, 0x19ca: 0x2000, 0x19cb: 0x2000, + 0x19cc: 0x2000, 0x19cd: 0x2000, 0x19ce: 0x2000, 0x19cf: 0x2000, 0x19d0: 0x2000, 0x19d1: 0x2000, + 0x19d2: 0x2000, 0x19d3: 0x2000, 0x19d4: 0x2000, 0x19d5: 0x2000, 0x19d6: 0x2000, 0x19d7: 0x2000, + 0x19d8: 0x2000, 0x19d9: 0x2000, 0x19da: 0x2000, 0x19db: 0x2000, 0x19dc: 0x2000, 0x19dd: 0x2000, + 0x19de: 0x2000, 0x19df: 0x2000, 0x19e0: 0x2000, 0x19e1: 0x2000, 0x19e2: 0x2000, 0x19e3: 0x2000, + 0x19e4: 0x2000, 0x19e5: 0x2000, 0x19e6: 0x2000, 0x19e7: 0x2000, 0x19e8: 0x2000, 0x19e9: 0x2000, + 0x19ea: 0x2000, 0x19eb: 0x2000, 0x19ec: 0x2000, 0x19ed: 0x2000, 0x19ee: 0x2000, 0x19ef: 0x2000, + 0x19f0: 0x2000, 0x19f1: 0x2000, 0x19f2: 0x2000, 0x19f3: 0x2000, 0x19f4: 0x2000, 0x19f5: 0x2000, + 0x19f6: 0x2000, 0x19f7: 0x2000, 0x19f8: 0x2000, 0x19f9: 0x2000, 0x19fa: 0x2000, 0x19fb: 0x2000, + 0x19fc: 0x2000, 0x19fd: 0x2000, +} + +// widthIndex: 22 blocks, 1408 entries, 1408 bytes +// Block 0 is the zero block. +var widthIndex = [1408]uint8{ + // Block 0x0, offset 0x0 + // Block 0x1, offset 0x40 + // Block 0x2, offset 0x80 + // Block 0x3, offset 0xc0 + 0xc2: 0x01, 0xc3: 0x02, 0xc4: 0x03, 0xc5: 0x04, 0xc7: 0x05, + 0xc9: 0x06, 0xcb: 0x07, 0xcc: 0x08, 0xcd: 0x09, 0xce: 0x0a, 0xcf: 0x0b, + 0xd0: 0x0c, 0xd1: 0x0d, + 0xe1: 0x02, 0xe2: 0x03, 0xe3: 0x04, 0xe4: 0x05, 0xe5: 0x06, 0xe6: 0x06, 0xe7: 0x06, + 0xe8: 0x06, 0xe9: 0x06, 0xea: 0x07, 0xeb: 0x06, 0xec: 0x06, 0xed: 0x08, 0xee: 0x09, 0xef: 0x0a, + 0xf0: 0x0f, 0xf3: 0x12, 0xf4: 0x13, + // Block 0x4, offset 0x100 + 0x104: 0x0e, 0x105: 0x0f, + // Block 0x5, offset 0x140 + 0x140: 0x10, 0x141: 0x11, 0x142: 0x12, 0x144: 0x13, 0x145: 0x14, 0x146: 0x15, 0x147: 0x16, + 0x148: 0x17, 0x149: 0x18, 0x14a: 0x19, 0x14c: 0x1a, 0x14f: 0x1b, + 0x151: 0x1c, 0x152: 0x08, 0x153: 0x1d, 0x154: 0x1e, 0x155: 0x1f, 0x156: 0x20, 0x157: 0x21, + 0x158: 0x22, 0x159: 0x23, 0x15a: 0x24, 0x15b: 0x25, 0x15c: 0x26, 0x15d: 0x27, 0x15e: 0x28, 0x15f: 0x29, + 0x166: 0x2a, + 0x16c: 0x2b, 0x16d: 0x2c, + 0x17a: 0x2d, 0x17b: 0x2e, 0x17c: 0x0e, 0x17d: 0x0e, 0x17e: 0x0e, 0x17f: 0x2f, + // Block 0x6, offset 0x180 + 0x180: 0x30, 0x181: 0x31, 0x182: 0x32, 0x183: 0x33, 0x184: 0x34, 0x185: 0x35, 0x186: 0x36, 0x187: 0x37, + 0x188: 0x38, 0x189: 0x39, 0x18a: 0x0e, 0x18b: 0x3a, 0x18c: 0x0e, 0x18d: 0x0e, 0x18e: 0x0e, 0x18f: 0x0e, + 0x190: 0x0e, 0x191: 0x0e, 0x192: 0x0e, 0x193: 0x0e, 0x194: 0x0e, 0x195: 0x0e, 0x196: 0x0e, 0x197: 0x0e, + 0x198: 0x0e, 0x199: 0x0e, 0x19a: 0x0e, 0x19b: 0x0e, 0x19c: 0x0e, 0x19d: 0x0e, 0x19e: 0x0e, 0x19f: 0x0e, + 0x1a0: 0x0e, 0x1a1: 0x0e, 0x1a2: 0x0e, 0x1a3: 0x0e, 0x1a4: 0x0e, 0x1a5: 0x0e, 0x1a6: 0x0e, 0x1a7: 0x0e, + 0x1a8: 0x0e, 0x1a9: 0x0e, 0x1aa: 0x0e, 0x1ab: 0x0e, 0x1ac: 0x0e, 0x1ad: 0x0e, 0x1ae: 0x0e, 0x1af: 0x0e, + 0x1b0: 0x0e, 0x1b1: 0x0e, 0x1b2: 0x0e, 0x1b3: 0x0e, 0x1b4: 0x0e, 0x1b5: 0x0e, 0x1b6: 0x0e, 0x1b7: 0x0e, + 0x1b8: 0x0e, 0x1b9: 0x0e, 0x1ba: 0x0e, 0x1bb: 0x0e, 0x1bc: 0x0e, 0x1bd: 0x0e, 0x1be: 0x0e, 0x1bf: 0x0e, + // Block 0x7, offset 0x1c0 + 0x1c0: 0x0e, 0x1c1: 0x0e, 0x1c2: 0x0e, 0x1c3: 0x0e, 0x1c4: 0x0e, 0x1c5: 0x0e, 0x1c6: 0x0e, 0x1c7: 0x0e, + 0x1c8: 0x0e, 0x1c9: 0x0e, 0x1ca: 0x0e, 0x1cb: 0x0e, 0x1cc: 0x0e, 0x1cd: 0x0e, 0x1ce: 0x0e, 0x1cf: 0x0e, + 0x1d0: 0x0e, 0x1d1: 0x0e, 0x1d2: 0x0e, 0x1d3: 0x0e, 0x1d4: 0x0e, 0x1d5: 0x0e, 0x1d6: 0x0e, 0x1d7: 0x0e, + 0x1d8: 0x0e, 0x1d9: 0x0e, 0x1da: 0x0e, 0x1db: 0x0e, 0x1dc: 0x0e, 0x1dd: 0x0e, 0x1de: 0x0e, 0x1df: 0x0e, + 0x1e0: 0x0e, 0x1e1: 0x0e, 0x1e2: 0x0e, 0x1e3: 0x0e, 0x1e4: 0x0e, 0x1e5: 0x0e, 0x1e6: 0x0e, 0x1e7: 0x0e, + 0x1e8: 0x0e, 0x1e9: 0x0e, 0x1ea: 0x0e, 0x1eb: 0x0e, 0x1ec: 0x0e, 0x1ed: 0x0e, 0x1ee: 0x0e, 0x1ef: 0x0e, + 0x1f0: 0x0e, 0x1f1: 0x0e, 0x1f2: 0x0e, 0x1f3: 0x0e, 0x1f4: 0x0e, 0x1f5: 0x0e, 0x1f6: 0x0e, + 0x1f8: 0x0e, 0x1f9: 0x0e, 0x1fa: 0x0e, 0x1fb: 0x0e, 0x1fc: 0x0e, 0x1fd: 0x0e, 0x1fe: 0x0e, 0x1ff: 0x0e, + // Block 0x8, offset 0x200 + 0x200: 0x0e, 0x201: 0x0e, 0x202: 0x0e, 0x203: 0x0e, 0x204: 0x0e, 0x205: 0x0e, 0x206: 0x0e, 0x207: 0x0e, + 0x208: 0x0e, 0x209: 0x0e, 0x20a: 0x0e, 0x20b: 0x0e, 0x20c: 0x0e, 0x20d: 0x0e, 0x20e: 0x0e, 0x20f: 0x0e, + 0x210: 0x0e, 0x211: 0x0e, 0x212: 0x0e, 0x213: 0x0e, 0x214: 0x0e, 0x215: 0x0e, 0x216: 0x0e, 0x217: 0x0e, + 0x218: 0x0e, 0x219: 0x0e, 0x21a: 0x0e, 0x21b: 0x0e, 0x21c: 0x0e, 0x21d: 0x0e, 0x21e: 0x0e, 0x21f: 0x0e, + 0x220: 0x0e, 0x221: 0x0e, 0x222: 0x0e, 0x223: 0x0e, 0x224: 0x0e, 0x225: 0x0e, 0x226: 0x0e, 0x227: 0x0e, + 0x228: 0x0e, 0x229: 0x0e, 0x22a: 0x0e, 0x22b: 0x0e, 0x22c: 0x0e, 0x22d: 0x0e, 0x22e: 0x0e, 0x22f: 0x0e, + 0x230: 0x0e, 0x231: 0x0e, 0x232: 0x0e, 0x233: 0x0e, 0x234: 0x0e, 0x235: 0x0e, 0x236: 0x0e, 0x237: 0x0e, + 0x238: 0x0e, 0x239: 0x0e, 0x23a: 0x0e, 0x23b: 0x0e, 0x23c: 0x0e, 0x23d: 0x0e, 0x23e: 0x0e, 0x23f: 0x0e, + // Block 0x9, offset 0x240 + 0x240: 0x0e, 0x241: 0x0e, 0x242: 0x0e, 0x243: 0x0e, 0x244: 0x0e, 0x245: 0x0e, 0x246: 0x0e, 0x247: 0x0e, + 0x248: 0x0e, 0x249: 0x0e, 0x24a: 0x0e, 0x24b: 0x0e, 0x24c: 0x0e, 0x24d: 0x0e, 0x24e: 0x0e, 0x24f: 0x0e, + 0x250: 0x0e, 0x251: 0x0e, 0x252: 0x3b, 0x253: 0x3c, + 0x265: 0x3d, + 0x270: 0x0e, 0x271: 0x0e, 0x272: 0x0e, 0x273: 0x0e, 0x274: 0x0e, 0x275: 0x0e, 0x276: 0x0e, 0x277: 0x0e, + 0x278: 0x0e, 0x279: 0x0e, 0x27a: 0x0e, 0x27b: 0x0e, 0x27c: 0x0e, 0x27d: 0x0e, 0x27e: 0x0e, 0x27f: 0x0e, + // Block 0xa, offset 0x280 + 0x280: 0x0e, 0x281: 0x0e, 0x282: 0x0e, 0x283: 0x0e, 0x284: 0x0e, 0x285: 0x0e, 0x286: 0x0e, 0x287: 0x0e, + 0x288: 0x0e, 0x289: 0x0e, 0x28a: 0x0e, 0x28b: 0x0e, 0x28c: 0x0e, 0x28d: 0x0e, 0x28e: 0x0e, 0x28f: 0x0e, + 0x290: 0x0e, 0x291: 0x0e, 0x292: 0x0e, 0x293: 0x0e, 0x294: 0x0e, 0x295: 0x0e, 0x296: 0x0e, 0x297: 0x0e, + 0x298: 0x0e, 0x299: 0x0e, 0x29a: 0x0e, 0x29b: 0x0e, 0x29c: 0x0e, 0x29d: 0x0e, 0x29e: 0x3e, + // Block 0xb, offset 0x2c0 + 0x2c0: 0x08, 0x2c1: 0x08, 0x2c2: 0x08, 0x2c3: 0x08, 0x2c4: 0x08, 0x2c5: 0x08, 0x2c6: 0x08, 0x2c7: 0x08, + 0x2c8: 0x08, 0x2c9: 0x08, 0x2ca: 0x08, 0x2cb: 0x08, 0x2cc: 0x08, 0x2cd: 0x08, 0x2ce: 0x08, 0x2cf: 0x08, + 0x2d0: 0x08, 0x2d1: 0x08, 0x2d2: 0x08, 0x2d3: 0x08, 0x2d4: 0x08, 0x2d5: 0x08, 0x2d6: 0x08, 0x2d7: 0x08, + 0x2d8: 0x08, 0x2d9: 0x08, 0x2da: 0x08, 0x2db: 0x08, 0x2dc: 0x08, 0x2dd: 0x08, 0x2de: 0x08, 0x2df: 0x08, + 0x2e0: 0x08, 0x2e1: 0x08, 0x2e2: 0x08, 0x2e3: 0x08, 0x2e4: 0x08, 0x2e5: 0x08, 0x2e6: 0x08, 0x2e7: 0x08, + 0x2e8: 0x08, 0x2e9: 0x08, 0x2ea: 0x08, 0x2eb: 0x08, 0x2ec: 0x08, 0x2ed: 0x08, 0x2ee: 0x08, 0x2ef: 0x08, + 0x2f0: 0x08, 0x2f1: 0x08, 0x2f2: 0x08, 0x2f3: 0x08, 0x2f4: 0x08, 0x2f5: 0x08, 0x2f6: 0x08, 0x2f7: 0x08, + 0x2f8: 0x08, 0x2f9: 0x08, 0x2fa: 0x08, 0x2fb: 0x08, 0x2fc: 0x08, 0x2fd: 0x08, 0x2fe: 0x08, 0x2ff: 0x08, + // Block 0xc, offset 0x300 + 0x300: 0x08, 0x301: 0x08, 0x302: 0x08, 0x303: 0x08, 0x304: 0x08, 0x305: 0x08, 0x306: 0x08, 0x307: 0x08, + 0x308: 0x08, 0x309: 0x08, 0x30a: 0x08, 0x30b: 0x08, 0x30c: 0x08, 0x30d: 0x08, 0x30e: 0x08, 0x30f: 0x08, + 0x310: 0x08, 0x311: 0x08, 0x312: 0x08, 0x313: 0x08, 0x314: 0x08, 0x315: 0x08, 0x316: 0x08, 0x317: 0x08, + 0x318: 0x08, 0x319: 0x08, 0x31a: 0x08, 0x31b: 0x08, 0x31c: 0x08, 0x31d: 0x08, 0x31e: 0x08, 0x31f: 0x08, + 0x320: 0x08, 0x321: 0x08, 0x322: 0x08, 0x323: 0x08, 0x324: 0x0e, 0x325: 0x0e, 0x326: 0x0e, 0x327: 0x0e, + 0x328: 0x0e, 0x329: 0x0e, 0x32a: 0x0e, 0x32b: 0x0e, + 0x338: 0x3f, 0x339: 0x40, 0x33c: 0x41, 0x33d: 0x42, 0x33e: 0x43, 0x33f: 0x44, + // Block 0xd, offset 0x340 + 0x37f: 0x45, + // Block 0xe, offset 0x380 + 0x380: 0x0e, 0x381: 0x0e, 0x382: 0x0e, 0x383: 0x0e, 0x384: 0x0e, 0x385: 0x0e, 0x386: 0x0e, 0x387: 0x0e, + 0x388: 0x0e, 0x389: 0x0e, 0x38a: 0x0e, 0x38b: 0x0e, 0x38c: 0x0e, 0x38d: 0x0e, 0x38e: 0x0e, 0x38f: 0x0e, + 0x390: 0x0e, 0x391: 0x0e, 0x392: 0x0e, 0x393: 0x0e, 0x394: 0x0e, 0x395: 0x0e, 0x396: 0x0e, 0x397: 0x0e, + 0x398: 0x0e, 0x399: 0x0e, 0x39a: 0x0e, 0x39b: 0x0e, 0x39c: 0x0e, 0x39d: 0x0e, 0x39e: 0x0e, 0x39f: 0x46, + 0x3a0: 0x0e, 0x3a1: 0x0e, 0x3a2: 0x0e, 0x3a3: 0x0e, 0x3a4: 0x0e, 0x3a5: 0x0e, 0x3a6: 0x0e, 0x3a7: 0x0e, + 0x3a8: 0x0e, 0x3a9: 0x0e, 0x3aa: 0x0e, 0x3ab: 0x47, + // Block 0xf, offset 0x3c0 + 0x3c0: 0x0e, 0x3c1: 0x0e, 0x3c2: 0x0e, 0x3c3: 0x0e, 0x3c4: 0x48, 0x3c5: 0x49, 0x3c6: 0x0e, 0x3c7: 0x0e, + 0x3c8: 0x0e, 0x3c9: 0x0e, 0x3ca: 0x0e, 0x3cb: 0x4a, + // Block 0x10, offset 0x400 + 0x400: 0x4b, 0x403: 0x4c, 0x404: 0x4d, 0x405: 0x4e, 0x406: 0x4f, + 0x408: 0x50, 0x409: 0x51, 0x40c: 0x52, 0x40d: 0x53, 0x40e: 0x54, 0x40f: 0x55, + 0x410: 0x3a, 0x411: 0x56, 0x412: 0x0e, 0x413: 0x57, 0x414: 0x58, 0x415: 0x59, 0x416: 0x5a, 0x417: 0x5b, + 0x418: 0x0e, 0x419: 0x5c, 0x41a: 0x0e, 0x41b: 0x5d, 0x41f: 0x5e, + 0x424: 0x5f, 0x425: 0x60, 0x426: 0x61, 0x427: 0x62, + 0x429: 0x63, 0x42a: 0x64, + // Block 0x11, offset 0x440 + 0x456: 0x0b, 0x457: 0x06, + 0x458: 0x0c, 0x45b: 0x0d, 0x45f: 0x0e, + 0x460: 0x06, 0x461: 0x06, 0x462: 0x06, 0x463: 0x06, 0x464: 0x06, 0x465: 0x06, 0x466: 0x06, 0x467: 0x06, + 0x468: 0x06, 0x469: 0x06, 0x46a: 0x06, 0x46b: 0x06, 0x46c: 0x06, 0x46d: 0x06, 0x46e: 0x06, 0x46f: 0x06, + 0x470: 0x06, 0x471: 0x06, 0x472: 0x06, 0x473: 0x06, 0x474: 0x06, 0x475: 0x06, 0x476: 0x06, 0x477: 0x06, + 0x478: 0x06, 0x479: 0x06, 0x47a: 0x06, 0x47b: 0x06, 0x47c: 0x06, 0x47d: 0x06, 0x47e: 0x06, 0x47f: 0x06, + // Block 0x12, offset 0x480 + 0x484: 0x08, 0x485: 0x08, 0x486: 0x08, 0x487: 0x09, + // Block 0x13, offset 0x4c0 + 0x4c0: 0x08, 0x4c1: 0x08, 0x4c2: 0x08, 0x4c3: 0x08, 0x4c4: 0x08, 0x4c5: 0x08, 0x4c6: 0x08, 0x4c7: 0x08, + 0x4c8: 0x08, 0x4c9: 0x08, 0x4ca: 0x08, 0x4cb: 0x08, 0x4cc: 0x08, 0x4cd: 0x08, 0x4ce: 0x08, 0x4cf: 0x08, + 0x4d0: 0x08, 0x4d1: 0x08, 0x4d2: 0x08, 0x4d3: 0x08, 0x4d4: 0x08, 0x4d5: 0x08, 0x4d6: 0x08, 0x4d7: 0x08, + 0x4d8: 0x08, 0x4d9: 0x08, 0x4da: 0x08, 0x4db: 0x08, 0x4dc: 0x08, 0x4dd: 0x08, 0x4de: 0x08, 0x4df: 0x08, + 0x4e0: 0x08, 0x4e1: 0x08, 0x4e2: 0x08, 0x4e3: 0x08, 0x4e4: 0x08, 0x4e5: 0x08, 0x4e6: 0x08, 0x4e7: 0x08, + 0x4e8: 0x08, 0x4e9: 0x08, 0x4ea: 0x08, 0x4eb: 0x08, 0x4ec: 0x08, 0x4ed: 0x08, 0x4ee: 0x08, 0x4ef: 0x08, + 0x4f0: 0x08, 0x4f1: 0x08, 0x4f2: 0x08, 0x4f3: 0x08, 0x4f4: 0x08, 0x4f5: 0x08, 0x4f6: 0x08, 0x4f7: 0x08, + 0x4f8: 0x08, 0x4f9: 0x08, 0x4fa: 0x08, 0x4fb: 0x08, 0x4fc: 0x08, 0x4fd: 0x08, 0x4fe: 0x08, 0x4ff: 0x65, + // Block 0x14, offset 0x500 + 0x520: 0x10, + 0x530: 0x09, 0x531: 0x09, 0x532: 0x09, 0x533: 0x09, 0x534: 0x09, 0x535: 0x09, 0x536: 0x09, 0x537: 0x09, + 0x538: 0x09, 0x539: 0x09, 0x53a: 0x09, 0x53b: 0x09, 0x53c: 0x09, 0x53d: 0x09, 0x53e: 0x09, 0x53f: 0x11, + // Block 0x15, offset 0x540 + 0x540: 0x09, 0x541: 0x09, 0x542: 0x09, 0x543: 0x09, 0x544: 0x09, 0x545: 0x09, 0x546: 0x09, 0x547: 0x09, + 0x548: 0x09, 0x549: 0x09, 0x54a: 0x09, 0x54b: 0x09, 0x54c: 0x09, 0x54d: 0x09, 0x54e: 0x09, 0x54f: 0x11, +} + +// inverseData contains 4-byte entries of the following format: +// <0 padding> +// The last byte of the UTF-8-encoded rune is xor-ed with the last byte of the +// UTF-8 encoding of the original rune. Mappings often have the following +// pattern: +// A -> A (U+FF21 -> U+0041) +// B -> B (U+FF22 -> U+0042) +// ... +// By xor-ing the last byte the same entry can be shared by many mappings. This +// reduces the total number of distinct entries by about two thirds. +// The resulting entry for the aforementioned mappings is +// { 0x01, 0xE0, 0x00, 0x00 } +// Using this entry to map U+FF21 (UTF-8 [EF BC A1]), we get +// E0 ^ A1 = 41. +// Similarly, for U+FF22 (UTF-8 [EF BC A2]), we get +// E0 ^ A2 = 42. +// Note that because of the xor-ing, the byte sequence stored in the entry is +// not valid UTF-8. +var inverseData = [150][4]byte{ + {0x00, 0x00, 0x00, 0x00}, + {0x03, 0xe3, 0x80, 0xa0}, + {0x03, 0xef, 0xbc, 0xa0}, + {0x03, 0xef, 0xbc, 0xe0}, + {0x03, 0xef, 0xbd, 0xe0}, + {0x03, 0xef, 0xbf, 0x02}, + {0x03, 0xef, 0xbf, 0x00}, + {0x03, 0xef, 0xbf, 0x0e}, + {0x03, 0xef, 0xbf, 0x0c}, + {0x03, 0xef, 0xbf, 0x0f}, + {0x03, 0xef, 0xbf, 0x39}, + {0x03, 0xef, 0xbf, 0x3b}, + {0x03, 0xef, 0xbf, 0x3f}, + {0x03, 0xef, 0xbf, 0x2a}, + {0x03, 0xef, 0xbf, 0x0d}, + {0x03, 0xef, 0xbf, 0x25}, + {0x03, 0xef, 0xbd, 0x1a}, + {0x03, 0xef, 0xbd, 0x26}, + {0x01, 0xa0, 0x00, 0x00}, + {0x03, 0xef, 0xbd, 0x25}, + {0x03, 0xef, 0xbd, 0x23}, + {0x03, 0xef, 0xbd, 0x2e}, + {0x03, 0xef, 0xbe, 0x07}, + {0x03, 0xef, 0xbe, 0x05}, + {0x03, 0xef, 0xbd, 0x06}, + {0x03, 0xef, 0xbd, 0x13}, + {0x03, 0xef, 0xbd, 0x0b}, + {0x03, 0xef, 0xbd, 0x16}, + {0x03, 0xef, 0xbd, 0x0c}, + {0x03, 0xef, 0xbd, 0x15}, + {0x03, 0xef, 0xbd, 0x0d}, + {0x03, 0xef, 0xbd, 0x1c}, + {0x03, 0xef, 0xbd, 0x02}, + {0x03, 0xef, 0xbd, 0x1f}, + {0x03, 0xef, 0xbd, 0x1d}, + {0x03, 0xef, 0xbd, 0x17}, + {0x03, 0xef, 0xbd, 0x08}, + {0x03, 0xef, 0xbd, 0x09}, + {0x03, 0xef, 0xbd, 0x0e}, + {0x03, 0xef, 0xbd, 0x04}, + {0x03, 0xef, 0xbd, 0x05}, + {0x03, 0xef, 0xbe, 0x3f}, + {0x03, 0xef, 0xbe, 0x00}, + {0x03, 0xef, 0xbd, 0x2c}, + {0x03, 0xef, 0xbe, 0x06}, + {0x03, 0xef, 0xbe, 0x0c}, + {0x03, 0xef, 0xbe, 0x0f}, + {0x03, 0xef, 0xbe, 0x0d}, + {0x03, 0xef, 0xbe, 0x0b}, + {0x03, 0xef, 0xbe, 0x19}, + {0x03, 0xef, 0xbe, 0x15}, + {0x03, 0xef, 0xbe, 0x11}, + {0x03, 0xef, 0xbe, 0x31}, + {0x03, 0xef, 0xbe, 0x33}, + {0x03, 0xef, 0xbd, 0x0f}, + {0x03, 0xef, 0xbe, 0x30}, + {0x03, 0xef, 0xbe, 0x3e}, + {0x03, 0xef, 0xbe, 0x32}, + {0x03, 0xef, 0xbe, 0x36}, + {0x03, 0xef, 0xbd, 0x14}, + {0x03, 0xef, 0xbe, 0x2e}, + {0x03, 0xef, 0xbd, 0x1e}, + {0x03, 0xef, 0xbe, 0x10}, + {0x03, 0xef, 0xbf, 0x13}, + {0x03, 0xef, 0xbf, 0x15}, + {0x03, 0xef, 0xbf, 0x17}, + {0x03, 0xef, 0xbf, 0x1f}, + {0x03, 0xef, 0xbf, 0x1d}, + {0x03, 0xef, 0xbf, 0x1b}, + {0x03, 0xef, 0xbf, 0x09}, + {0x03, 0xef, 0xbf, 0x0b}, + {0x03, 0xef, 0xbf, 0x37}, + {0x03, 0xef, 0xbe, 0x04}, + {0x01, 0xe0, 0x00, 0x00}, + {0x03, 0xe2, 0xa6, 0x1a}, + {0x03, 0xe2, 0xa6, 0x26}, + {0x03, 0xe3, 0x80, 0x23}, + {0x03, 0xe3, 0x80, 0x2e}, + {0x03, 0xe3, 0x80, 0x25}, + {0x03, 0xe3, 0x83, 0x1e}, + {0x03, 0xe3, 0x83, 0x14}, + {0x03, 0xe3, 0x82, 0x06}, + {0x03, 0xe3, 0x82, 0x0b}, + {0x03, 0xe3, 0x82, 0x0c}, + {0x03, 0xe3, 0x82, 0x0d}, + {0x03, 0xe3, 0x82, 0x02}, + {0x03, 0xe3, 0x83, 0x0f}, + {0x03, 0xe3, 0x83, 0x08}, + {0x03, 0xe3, 0x83, 0x09}, + {0x03, 0xe3, 0x83, 0x2c}, + {0x03, 0xe3, 0x83, 0x0c}, + {0x03, 0xe3, 0x82, 0x13}, + {0x03, 0xe3, 0x82, 0x16}, + {0x03, 0xe3, 0x82, 0x15}, + {0x03, 0xe3, 0x82, 0x1c}, + {0x03, 0xe3, 0x82, 0x1f}, + {0x03, 0xe3, 0x82, 0x1d}, + {0x03, 0xe3, 0x82, 0x1a}, + {0x03, 0xe3, 0x82, 0x17}, + {0x03, 0xe3, 0x82, 0x08}, + {0x03, 0xe3, 0x82, 0x09}, + {0x03, 0xe3, 0x82, 0x0e}, + {0x03, 0xe3, 0x82, 0x04}, + {0x03, 0xe3, 0x82, 0x05}, + {0x03, 0xe3, 0x82, 0x3f}, + {0x03, 0xe3, 0x83, 0x00}, + {0x03, 0xe3, 0x83, 0x06}, + {0x03, 0xe3, 0x83, 0x05}, + {0x03, 0xe3, 0x83, 0x0d}, + {0x03, 0xe3, 0x83, 0x0b}, + {0x03, 0xe3, 0x83, 0x07}, + {0x03, 0xe3, 0x83, 0x19}, + {0x03, 0xe3, 0x83, 0x15}, + {0x03, 0xe3, 0x83, 0x11}, + {0x03, 0xe3, 0x83, 0x31}, + {0x03, 0xe3, 0x83, 0x33}, + {0x03, 0xe3, 0x83, 0x30}, + {0x03, 0xe3, 0x83, 0x3e}, + {0x03, 0xe3, 0x83, 0x32}, + {0x03, 0xe3, 0x83, 0x36}, + {0x03, 0xe3, 0x83, 0x2e}, + {0x03, 0xe3, 0x82, 0x07}, + {0x03, 0xe3, 0x85, 0x04}, + {0x03, 0xe3, 0x84, 0x10}, + {0x03, 0xe3, 0x85, 0x30}, + {0x03, 0xe3, 0x85, 0x0d}, + {0x03, 0xe3, 0x85, 0x13}, + {0x03, 0xe3, 0x85, 0x15}, + {0x03, 0xe3, 0x85, 0x17}, + {0x03, 0xe3, 0x85, 0x1f}, + {0x03, 0xe3, 0x85, 0x1d}, + {0x03, 0xe3, 0x85, 0x1b}, + {0x03, 0xe3, 0x85, 0x09}, + {0x03, 0xe3, 0x85, 0x0f}, + {0x03, 0xe3, 0x85, 0x0b}, + {0x03, 0xe3, 0x85, 0x37}, + {0x03, 0xe3, 0x85, 0x3b}, + {0x03, 0xe3, 0x85, 0x39}, + {0x03, 0xe3, 0x85, 0x3f}, + {0x02, 0xc2, 0x02, 0x00}, + {0x02, 0xc2, 0x0e, 0x00}, + {0x02, 0xc2, 0x0c, 0x00}, + {0x02, 0xc2, 0x00, 0x00}, + {0x03, 0xe2, 0x82, 0x0f}, + {0x03, 0xe2, 0x94, 0x2a}, + {0x03, 0xe2, 0x86, 0x39}, + {0x03, 0xe2, 0x86, 0x3b}, + {0x03, 0xe2, 0x86, 0x3f}, + {0x03, 0xe2, 0x96, 0x0d}, + {0x03, 0xe2, 0x97, 0x25}, +} + +// Total table size 15320 bytes (14KiB) diff --git a/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go b/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go index c988461b..e79a5388 100644 --- a/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go @@ -1,26 +1,45 @@ +// Copyright 2020 Google LLC +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + // Code generated by protoc-gen-go. DO NOT EDIT. +// versions: +// protoc-gen-go v1.25.0 +// protoc v3.13.0 // source: google/rpc/status.proto package status import ( - fmt "fmt" - math "math" + reflect "reflect" + sync "sync" proto "github.com/golang/protobuf/proto" - any "github.com/golang/protobuf/ptypes/any" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + anypb "google.golang.org/protobuf/types/known/anypb" ) -// Reference imports to suppress errors if they are not otherwise used. -var _ = proto.Marshal -var _ = fmt.Errorf -var _ = math.Inf +const ( + // Verify that this generated code is sufficiently up-to-date. + _ = protoimpl.EnforceVersion(20 - protoimpl.MinVersion) + // Verify that runtime/protoimpl is sufficiently up-to-date. + _ = protoimpl.EnforceVersion(protoimpl.MaxVersion - 20) +) -// This is a compile-time assertion to ensure that this generated file -// is compatible with the proto package it is being compiled against. -// A compilation error at this line likely means your copy of the -// proto package needs to be updated. -const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package +// This is a compile-time assertion that a sufficiently up-to-date version +// of the legacy proto package is being used. +const _ = proto.ProtoPackageIsVersion4 // The `Status` type defines a logical error model that is suitable for // different programming environments, including REST APIs and RPC APIs. It is @@ -30,6 +49,10 @@ const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package // You can find out more about this error model and how to work with it in the // [API Design Guide](https://cloud.google.com/apis/design/errors). type Status struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + // The status code, which should be an enum value of [google.rpc.Code][google.rpc.Code]. Code int32 `protobuf:"varint,1,opt,name=code,proto3" json:"code,omitempty"` // A developer-facing error message, which should be in English. Any @@ -38,80 +61,146 @@ type Status struct { Message string `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` // A list of messages that carry the error details. There is a common set of // message types for APIs to use. - Details []*any.Any `protobuf:"bytes,3,rep,name=details,proto3" json:"details,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + Details []*anypb.Any `protobuf:"bytes,3,rep,name=details,proto3" json:"details,omitempty"` } -func (m *Status) Reset() { *m = Status{} } -func (m *Status) String() string { return proto.CompactTextString(m) } -func (*Status) ProtoMessage() {} -func (*Status) Descriptor() ([]byte, []int) { - return fileDescriptor_24d244abaf643bfe, []int{0} +func (x *Status) Reset() { + *x = Status{} + if protoimpl.UnsafeEnabled { + mi := &file_google_rpc_status_proto_msgTypes[0] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } } -func (m *Status) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Status.Unmarshal(m, b) -} -func (m *Status) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Status.Marshal(b, m, deterministic) +func (x *Status) String() string { + return protoimpl.X.MessageStringOf(x) } -func (m *Status) XXX_Merge(src proto.Message) { - xxx_messageInfo_Status.Merge(m, src) -} -func (m *Status) XXX_Size() int { - return xxx_messageInfo_Status.Size(m) -} -func (m *Status) XXX_DiscardUnknown() { - xxx_messageInfo_Status.DiscardUnknown(m) + +func (*Status) ProtoMessage() {} + +func (x *Status) ProtoReflect() protoreflect.Message { + mi := &file_google_rpc_status_proto_msgTypes[0] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) } -var xxx_messageInfo_Status proto.InternalMessageInfo +// Deprecated: Use Status.ProtoReflect.Descriptor instead. +func (*Status) Descriptor() ([]byte, []int) { + return file_google_rpc_status_proto_rawDescGZIP(), []int{0} +} -func (m *Status) GetCode() int32 { - if m != nil { - return m.Code +func (x *Status) GetCode() int32 { + if x != nil { + return x.Code } return 0 } -func (m *Status) GetMessage() string { - if m != nil { - return m.Message +func (x *Status) GetMessage() string { + if x != nil { + return x.Message } return "" } -func (m *Status) GetDetails() []*any.Any { - if m != nil { - return m.Details +func (x *Status) GetDetails() []*anypb.Any { + if x != nil { + return x.Details } return nil } -func init() { - proto.RegisterType((*Status)(nil), "google.rpc.Status") +var File_google_rpc_status_proto protoreflect.FileDescriptor + +var file_google_rpc_status_proto_rawDesc = []byte{ + 0x0a, 0x17, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x72, 0x70, 0x63, 0x2f, 0x73, 0x74, 0x61, + 0x74, 0x75, 0x73, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x0a, 0x67, 0x6f, 0x6f, 0x67, 0x6c, + 0x65, 0x2e, 0x72, 0x70, 0x63, 0x1a, 0x19, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, + 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x61, 0x6e, 0x79, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, + 0x22, 0x66, 0x0a, 0x06, 0x53, 0x74, 0x61, 0x74, 0x75, 0x73, 0x12, 0x12, 0x0a, 0x04, 0x63, 0x6f, + 0x64, 0x65, 0x18, 0x01, 0x20, 0x01, 0x28, 0x05, 0x52, 0x04, 0x63, 0x6f, 0x64, 0x65, 0x12, 0x18, + 0x0a, 0x07, 0x6d, 0x65, 0x73, 0x73, 0x61, 0x67, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x09, 0x52, + 0x07, 0x6d, 0x65, 0x73, 0x73, 0x61, 0x67, 0x65, 0x12, 0x2e, 0x0a, 0x07, 0x64, 0x65, 0x74, 0x61, + 0x69, 0x6c, 0x73, 0x18, 0x03, 0x20, 0x03, 0x28, 0x0b, 0x32, 0x14, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x41, 0x6e, 0x79, 0x52, + 0x07, 0x64, 0x65, 0x74, 0x61, 0x69, 0x6c, 0x73, 0x42, 0x61, 0x0a, 0x0e, 0x63, 0x6f, 0x6d, 0x2e, + 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x72, 0x70, 0x63, 0x42, 0x0b, 0x53, 0x74, 0x61, 0x74, + 0x75, 0x73, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, 0x37, 0x67, 0x6f, 0x6f, 0x67, 0x6c, + 0x65, 0x2e, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2e, 0x6f, 0x72, 0x67, 0x2f, 0x67, 0x65, 0x6e, + 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x2f, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x61, 0x70, 0x69, 0x73, + 0x2f, 0x72, 0x70, 0x63, 0x2f, 0x73, 0x74, 0x61, 0x74, 0x75, 0x73, 0x3b, 0x73, 0x74, 0x61, 0x74, + 0x75, 0x73, 0xf8, 0x01, 0x01, 0xa2, 0x02, 0x03, 0x52, 0x50, 0x43, 0x62, 0x06, 0x70, 0x72, 0x6f, + 0x74, 0x6f, 0x33, +} + +var ( + file_google_rpc_status_proto_rawDescOnce sync.Once + file_google_rpc_status_proto_rawDescData = file_google_rpc_status_proto_rawDesc +) + +func file_google_rpc_status_proto_rawDescGZIP() []byte { + file_google_rpc_status_proto_rawDescOnce.Do(func() { + file_google_rpc_status_proto_rawDescData = protoimpl.X.CompressGZIP(file_google_rpc_status_proto_rawDescData) + }) + return file_google_rpc_status_proto_rawDescData } -func init() { - proto.RegisterFile("google/rpc/status.proto", fileDescriptor_24d244abaf643bfe) +var file_google_rpc_status_proto_msgTypes = make([]protoimpl.MessageInfo, 1) +var file_google_rpc_status_proto_goTypes = []interface{}{ + (*Status)(nil), // 0: google.rpc.Status + (*anypb.Any)(nil), // 1: google.protobuf.Any +} +var file_google_rpc_status_proto_depIdxs = []int32{ + 1, // 0: google.rpc.Status.details:type_name -> google.protobuf.Any + 1, // [1:1] is the sub-list for method output_type + 1, // [1:1] is the sub-list for method input_type + 1, // [1:1] is the sub-list for extension type_name + 1, // [1:1] is the sub-list for extension extendee + 0, // [0:1] is the sub-list for field type_name } -var fileDescriptor_24d244abaf643bfe = []byte{ - // 212 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x12, 0x4f, 0xcf, 0xcf, 0x4f, - 0xcf, 0x49, 0xd5, 0x2f, 0x2a, 0x48, 0xd6, 0x2f, 0x2e, 0x49, 0x2c, 0x29, 0x2d, 0xd6, 0x2b, 0x28, - 0xca, 0x2f, 0xc9, 0x17, 0xe2, 0x82, 0x48, 0xe8, 0x15, 0x15, 0x24, 0x4b, 0x49, 0x42, 0x15, 0x81, - 0x65, 0x92, 0x4a, 0xd3, 0xf4, 0x13, 0xf3, 0x2a, 0x21, 0xca, 0x94, 0xd2, 0xb8, 0xd8, 0x82, 0xc1, - 0xda, 0x84, 0x84, 0xb8, 0x58, 0x92, 0xf3, 0x53, 0x52, 0x25, 0x18, 0x15, 0x18, 0x35, 0x58, 0x83, - 0xc0, 0x6c, 0x21, 0x09, 0x2e, 0xf6, 0xdc, 0xd4, 0xe2, 0xe2, 0xc4, 0xf4, 0x54, 0x09, 0x26, 0x05, - 0x46, 0x0d, 0xce, 0x20, 0x18, 0x57, 0x48, 0x8f, 0x8b, 0x3d, 0x25, 0xb5, 0x24, 0x31, 0x33, 0xa7, - 0x58, 0x82, 0x59, 0x81, 0x59, 0x83, 0xdb, 0x48, 0x44, 0x0f, 0x6a, 0x21, 0xcc, 0x12, 0x3d, 0xc7, - 0xbc, 0xca, 0x20, 0x98, 0x22, 0xa7, 0x44, 0x2e, 0xbe, 0xe4, 0xfc, 0x5c, 0x3d, 0x84, 0xa3, 0x9c, - 0xb8, 0x21, 0xf6, 0x06, 0x80, 0x94, 0x07, 0x30, 0x46, 0x99, 0x43, 0xa5, 0xd2, 0xf3, 0x73, 0x12, - 0xf3, 0xd2, 0xf5, 0xf2, 0x8b, 0xd2, 0xf5, 0xd3, 0x53, 0xf3, 0xc0, 0x86, 0xe9, 0x43, 0xa4, 0x12, - 0x0b, 0x32, 0x8b, 0x91, 0xfc, 0x69, 0x0d, 0xa1, 0x7e, 0x30, 0x32, 0x2e, 0x62, 0x62, 0x0e, 0x0a, - 0x70, 0x4e, 0x62, 0x03, 0x2b, 0x36, 0x06, 0x04, 0x00, 0x00, 0xff, 0xff, 0xb9, 0x28, 0x45, 0xb1, - 0x13, 0x01, 0x00, 0x00, +func init() { file_google_rpc_status_proto_init() } +func file_google_rpc_status_proto_init() { + if File_google_rpc_status_proto != nil { + return + } + if !protoimpl.UnsafeEnabled { + file_google_rpc_status_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Status); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + } + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_google_rpc_status_proto_rawDesc, + NumEnums: 0, + NumMessages: 1, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_google_rpc_status_proto_goTypes, + DependencyIndexes: file_google_rpc_status_proto_depIdxs, + MessageInfos: file_google_rpc_status_proto_msgTypes, + }.Build() + File_google_rpc_status_proto = out.File + file_google_rpc_status_proto_rawDesc = nil + file_google_rpc_status_proto_goTypes = nil + file_google_rpc_status_proto_depIdxs = nil } diff --git a/vendor/google.golang.org/protobuf/AUTHORS b/vendor/google.golang.org/protobuf/AUTHORS new file mode 100644 index 00000000..2b00ddba --- /dev/null +++ b/vendor/google.golang.org/protobuf/AUTHORS @@ -0,0 +1,3 @@ +# This source code refers to The Go Authors for copyright purposes. +# The master list of authors is in the main Go distribution, +# visible at https://tip.golang.org/AUTHORS. diff --git a/vendor/google.golang.org/protobuf/CONTRIBUTORS b/vendor/google.golang.org/protobuf/CONTRIBUTORS new file mode 100644 index 00000000..1fbd3e97 --- /dev/null +++ b/vendor/google.golang.org/protobuf/CONTRIBUTORS @@ -0,0 +1,3 @@ +# This source code was written by the Go contributors. +# The master list of contributors is in the main Go distribution, +# visible at https://tip.golang.org/CONTRIBUTORS. diff --git a/vendor/google.golang.org/protobuf/LICENSE b/vendor/google.golang.org/protobuf/LICENSE new file mode 100644 index 00000000..49ea0f92 --- /dev/null +++ b/vendor/google.golang.org/protobuf/LICENSE @@ -0,0 +1,27 @@ +Copyright (c) 2018 The Go Authors. All rights reserved. + +Redistribution and use in source and binary forms, with or without +modification, are permitted provided that the following conditions are +met: + + * Redistributions of source code must retain the above copyright +notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above +copyright notice, this list of conditions and the following disclaimer +in the documentation and/or other materials provided with the +distribution. + * Neither the name of Google Inc. nor the names of its +contributors may be used to endorse or promote products derived from +this software without specific prior written permission. + +THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. diff --git a/vendor/google.golang.org/protobuf/PATENTS b/vendor/google.golang.org/protobuf/PATENTS new file mode 100644 index 00000000..73309904 --- /dev/null +++ b/vendor/google.golang.org/protobuf/PATENTS @@ -0,0 +1,22 @@ +Additional IP Rights Grant (Patents) + +"This implementation" means the copyrightable works distributed by +Google as part of the Go project. + +Google hereby grants to You a perpetual, worldwide, non-exclusive, +no-charge, royalty-free, irrevocable (except as stated in this section) +patent license to make, have made, use, offer to sell, sell, import, +transfer and otherwise run, modify and propagate the contents of this +implementation of Go, where such license applies only to those patent +claims, both currently owned or controlled by Google and acquired in +the future, licensable by Google that are necessarily infringed by this +implementation of Go. This grant does not include claims that would be +infringed only as a consequence of further modification of this +implementation. If you or your agent or exclusive licensee institute or +order or agree to the institution of patent litigation against any +entity (including a cross-claim or counterclaim in a lawsuit) alleging +that this implementation of Go or any code incorporated within this +implementation of Go constitutes direct or contributory patent +infringement, or inducement of patent infringement, then any patent +rights granted to you under this License for this implementation of Go +shall terminate as of the date such litigation is filed. diff --git a/vendor/google.golang.org/protobuf/encoding/prototext/decode.go b/vendor/google.golang.org/protobuf/encoding/prototext/decode.go new file mode 100644 index 00000000..c2f8f28f --- /dev/null +++ b/vendor/google.golang.org/protobuf/encoding/prototext/decode.go @@ -0,0 +1,796 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package prototext + +import ( + "fmt" + "strings" + "unicode/utf8" + + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/encoding/text" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/fieldnum" + "google.golang.org/protobuf/internal/flags" + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/internal/set" + "google.golang.org/protobuf/internal/strs" + "google.golang.org/protobuf/proto" + pref "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +// Unmarshal reads the given []byte into the given proto.Message. +func Unmarshal(b []byte, m proto.Message) error { + return UnmarshalOptions{}.Unmarshal(b, m) +} + +// UnmarshalOptions is a configurable textproto format unmarshaler. +type UnmarshalOptions struct { + pragma.NoUnkeyedLiterals + + // AllowPartial accepts input for messages that will result in missing + // required fields. If AllowPartial is false (the default), Unmarshal will + // return error if there are any missing required fields. + AllowPartial bool + + // DiscardUnknown specifies whether to ignore unknown fields when parsing. + // An unknown field is any field whose field name or field number does not + // resolve to any known or extension field in the message. + // By default, unmarshal rejects unknown fields as an error. + DiscardUnknown bool + + // Resolver is used for looking up types when unmarshaling + // google.protobuf.Any messages or extension fields. + // If nil, this defaults to using protoregistry.GlobalTypes. + Resolver interface { + protoregistry.MessageTypeResolver + protoregistry.ExtensionTypeResolver + } +} + +// Unmarshal reads the given []byte and populates the given proto.Message using options in +// UnmarshalOptions object. +func (o UnmarshalOptions) Unmarshal(b []byte, m proto.Message) error { + return o.unmarshal(b, m) +} + +// unmarshal is a centralized function that all unmarshal operations go through. +// For profiling purposes, avoid changing the name of this function or +// introducing other code paths for unmarshal that do not go through this. +func (o UnmarshalOptions) unmarshal(b []byte, m proto.Message) error { + proto.Reset(m) + + if o.Resolver == nil { + o.Resolver = protoregistry.GlobalTypes + } + + dec := decoder{text.NewDecoder(b), o} + if err := dec.unmarshalMessage(m.ProtoReflect(), false); err != nil { + return err + } + if o.AllowPartial { + return nil + } + return proto.CheckInitialized(m) +} + +type decoder struct { + *text.Decoder + opts UnmarshalOptions +} + +// newError returns an error object with position info. +func (d decoder) newError(pos int, f string, x ...interface{}) error { + line, column := d.Position(pos) + head := fmt.Sprintf("(line %d:%d): ", line, column) + return errors.New(head+f, x...) +} + +// unexpectedTokenError returns a syntax error for the given unexpected token. +func (d decoder) unexpectedTokenError(tok text.Token) error { + return d.syntaxError(tok.Pos(), "unexpected token: %s", tok.RawString()) +} + +// syntaxError returns a syntax error for given position. +func (d decoder) syntaxError(pos int, f string, x ...interface{}) error { + line, column := d.Position(pos) + head := fmt.Sprintf("syntax error (line %d:%d): ", line, column) + return errors.New(head+f, x...) +} + +// unmarshalMessage unmarshals into the given protoreflect.Message. +func (d decoder) unmarshalMessage(m pref.Message, checkDelims bool) error { + messageDesc := m.Descriptor() + if !flags.ProtoLegacy && messageset.IsMessageSet(messageDesc) { + return errors.New("no support for proto1 MessageSets") + } + + if messageDesc.FullName() == "google.protobuf.Any" { + return d.unmarshalAny(m, checkDelims) + } + + if checkDelims { + tok, err := d.Read() + if err != nil { + return err + } + + if tok.Kind() != text.MessageOpen { + return d.unexpectedTokenError(tok) + } + } + + var seenNums set.Ints + var seenOneofs set.Ints + fieldDescs := messageDesc.Fields() + + for { + // Read field name. + tok, err := d.Read() + if err != nil { + return err + } + switch typ := tok.Kind(); typ { + case text.Name: + // Continue below. + case text.EOF: + if checkDelims { + return text.ErrUnexpectedEOF + } + return nil + default: + if checkDelims && typ == text.MessageClose { + return nil + } + return d.unexpectedTokenError(tok) + } + + // Resolve the field descriptor. + var name pref.Name + var fd pref.FieldDescriptor + var xt pref.ExtensionType + var xtErr error + var isFieldNumberName bool + + switch tok.NameKind() { + case text.IdentName: + name = pref.Name(tok.IdentName()) + fd = fieldDescs.ByName(name) + if fd == nil { + // The proto name of a group field is in all lowercase, + // while the textproto field name is the group message name. + gd := fieldDescs.ByName(pref.Name(strings.ToLower(string(name)))) + if gd != nil && gd.Kind() == pref.GroupKind && gd.Message().Name() == name { + fd = gd + } + } else if fd.Kind() == pref.GroupKind && fd.Message().Name() != name { + fd = nil // reset since field name is actually the message name + } + + case text.TypeName: + // Handle extensions only. This code path is not for Any. + xt, xtErr = d.findExtension(pref.FullName(tok.TypeName())) + + case text.FieldNumber: + isFieldNumberName = true + num := pref.FieldNumber(tok.FieldNumber()) + if !num.IsValid() { + return d.newError(tok.Pos(), "invalid field number: %d", num) + } + fd = fieldDescs.ByNumber(num) + if fd == nil { + xt, xtErr = d.opts.Resolver.FindExtensionByNumber(messageDesc.FullName(), num) + } + } + + if xt != nil { + fd = xt.TypeDescriptor() + if !messageDesc.ExtensionRanges().Has(fd.Number()) || fd.ContainingMessage().FullName() != messageDesc.FullName() { + return d.newError(tok.Pos(), "message %v cannot be extended by %v", messageDesc.FullName(), fd.FullName()) + } + } else if xtErr != nil && xtErr != protoregistry.NotFound { + return d.newError(tok.Pos(), "unable to resolve [%s]: %v", tok.RawString(), xtErr) + } + if flags.ProtoLegacy { + if fd != nil && fd.IsWeak() && fd.Message().IsPlaceholder() { + fd = nil // reset since the weak reference is not linked in + } + } + + // Handle unknown fields. + if fd == nil { + if d.opts.DiscardUnknown || messageDesc.ReservedNames().Has(name) { + d.skipValue() + continue + } + return d.newError(tok.Pos(), "unknown field: %v", tok.RawString()) + } + + // Handle fields identified by field number. + if isFieldNumberName { + // TODO: Add an option to permit parsing field numbers. + // + // This requires careful thought as the MarshalOptions.EmitUnknown + // option allows formatting unknown fields as the field number and the + // best-effort textual representation of the field value. In that case, + // it may not be possible to unmarshal the value from a parser that does + // have information about the unknown field. + return d.newError(tok.Pos(), "cannot specify field by number: %v", tok.RawString()) + } + + switch { + case fd.IsList(): + kind := fd.Kind() + if kind != pref.MessageKind && kind != pref.GroupKind && !tok.HasSeparator() { + return d.syntaxError(tok.Pos(), "missing field separator :") + } + + list := m.Mutable(fd).List() + if err := d.unmarshalList(fd, list); err != nil { + return err + } + + case fd.IsMap(): + mmap := m.Mutable(fd).Map() + if err := d.unmarshalMap(fd, mmap); err != nil { + return err + } + + default: + kind := fd.Kind() + if kind != pref.MessageKind && kind != pref.GroupKind && !tok.HasSeparator() { + return d.syntaxError(tok.Pos(), "missing field separator :") + } + + // If field is a oneof, check if it has already been set. + if od := fd.ContainingOneof(); od != nil { + idx := uint64(od.Index()) + if seenOneofs.Has(idx) { + return d.newError(tok.Pos(), "error parsing %q, oneof %v is already set", tok.RawString(), od.FullName()) + } + seenOneofs.Set(idx) + } + + num := uint64(fd.Number()) + if seenNums.Has(num) { + return d.newError(tok.Pos(), "non-repeated field %q is repeated", tok.RawString()) + } + + if err := d.unmarshalSingular(fd, m); err != nil { + return err + } + seenNums.Set(num) + } + } + + return nil +} + +// findExtension returns protoreflect.ExtensionType from the Resolver if found. +func (d decoder) findExtension(xtName pref.FullName) (pref.ExtensionType, error) { + xt, err := d.opts.Resolver.FindExtensionByName(xtName) + if err == nil { + return xt, nil + } + return messageset.FindMessageSetExtension(d.opts.Resolver, xtName) +} + +// unmarshalSingular unmarshals a non-repeated field value specified by the +// given FieldDescriptor. +func (d decoder) unmarshalSingular(fd pref.FieldDescriptor, m pref.Message) error { + var val pref.Value + var err error + switch fd.Kind() { + case pref.MessageKind, pref.GroupKind: + val = m.NewField(fd) + err = d.unmarshalMessage(val.Message(), true) + default: + val, err = d.unmarshalScalar(fd) + } + if err == nil { + m.Set(fd, val) + } + return err +} + +// unmarshalScalar unmarshals a scalar/enum protoreflect.Value specified by the +// given FieldDescriptor. +func (d decoder) unmarshalScalar(fd pref.FieldDescriptor) (pref.Value, error) { + tok, err := d.Read() + if err != nil { + return pref.Value{}, err + } + + if tok.Kind() != text.Scalar { + return pref.Value{}, d.unexpectedTokenError(tok) + } + + kind := fd.Kind() + switch kind { + case pref.BoolKind: + if b, ok := tok.Bool(); ok { + return pref.ValueOfBool(b), nil + } + + case pref.Int32Kind, pref.Sint32Kind, pref.Sfixed32Kind: + if n, ok := tok.Int32(); ok { + return pref.ValueOfInt32(n), nil + } + + case pref.Int64Kind, pref.Sint64Kind, pref.Sfixed64Kind: + if n, ok := tok.Int64(); ok { + return pref.ValueOfInt64(n), nil + } + + case pref.Uint32Kind, pref.Fixed32Kind: + if n, ok := tok.Uint32(); ok { + return pref.ValueOfUint32(n), nil + } + + case pref.Uint64Kind, pref.Fixed64Kind: + if n, ok := tok.Uint64(); ok { + return pref.ValueOfUint64(n), nil + } + + case pref.FloatKind: + if n, ok := tok.Float32(); ok { + return pref.ValueOfFloat32(n), nil + } + + case pref.DoubleKind: + if n, ok := tok.Float64(); ok { + return pref.ValueOfFloat64(n), nil + } + + case pref.StringKind: + if s, ok := tok.String(); ok { + if strs.EnforceUTF8(fd) && !utf8.ValidString(s) { + return pref.Value{}, d.newError(tok.Pos(), "contains invalid UTF-8") + } + return pref.ValueOfString(s), nil + } + + case pref.BytesKind: + if b, ok := tok.String(); ok { + return pref.ValueOfBytes([]byte(b)), nil + } + + case pref.EnumKind: + if lit, ok := tok.Enum(); ok { + // Lookup EnumNumber based on name. + if enumVal := fd.Enum().Values().ByName(pref.Name(lit)); enumVal != nil { + return pref.ValueOfEnum(enumVal.Number()), nil + } + } + if num, ok := tok.Int32(); ok { + return pref.ValueOfEnum(pref.EnumNumber(num)), nil + } + + default: + panic(fmt.Sprintf("invalid scalar kind %v", kind)) + } + + return pref.Value{}, d.newError(tok.Pos(), "invalid value for %v type: %v", kind, tok.RawString()) +} + +// unmarshalList unmarshals into given protoreflect.List. A list value can +// either be in [] syntax or simply just a single scalar/message value. +func (d decoder) unmarshalList(fd pref.FieldDescriptor, list pref.List) error { + tok, err := d.Peek() + if err != nil { + return err + } + + switch fd.Kind() { + case pref.MessageKind, pref.GroupKind: + switch tok.Kind() { + case text.ListOpen: + d.Read() + for { + tok, err := d.Peek() + if err != nil { + return err + } + + switch tok.Kind() { + case text.ListClose: + d.Read() + return nil + case text.MessageOpen: + pval := list.NewElement() + if err := d.unmarshalMessage(pval.Message(), true); err != nil { + return err + } + list.Append(pval) + default: + return d.unexpectedTokenError(tok) + } + } + + case text.MessageOpen: + pval := list.NewElement() + if err := d.unmarshalMessage(pval.Message(), true); err != nil { + return err + } + list.Append(pval) + return nil + } + + default: + switch tok.Kind() { + case text.ListOpen: + d.Read() + for { + tok, err := d.Peek() + if err != nil { + return err + } + + switch tok.Kind() { + case text.ListClose: + d.Read() + return nil + case text.Scalar: + pval, err := d.unmarshalScalar(fd) + if err != nil { + return err + } + list.Append(pval) + default: + return d.unexpectedTokenError(tok) + } + } + + case text.Scalar: + pval, err := d.unmarshalScalar(fd) + if err != nil { + return err + } + list.Append(pval) + return nil + } + } + + return d.unexpectedTokenError(tok) +} + +// unmarshalMap unmarshals into given protoreflect.Map. A map value is a +// textproto message containing {key: , value: }. +func (d decoder) unmarshalMap(fd pref.FieldDescriptor, mmap pref.Map) error { + // Determine ahead whether map entry is a scalar type or a message type in + // order to call the appropriate unmarshalMapValue func inside + // unmarshalMapEntry. + var unmarshalMapValue func() (pref.Value, error) + switch fd.MapValue().Kind() { + case pref.MessageKind, pref.GroupKind: + unmarshalMapValue = func() (pref.Value, error) { + pval := mmap.NewValue() + if err := d.unmarshalMessage(pval.Message(), true); err != nil { + return pref.Value{}, err + } + return pval, nil + } + default: + unmarshalMapValue = func() (pref.Value, error) { + return d.unmarshalScalar(fd.MapValue()) + } + } + + tok, err := d.Read() + if err != nil { + return err + } + switch tok.Kind() { + case text.MessageOpen: + return d.unmarshalMapEntry(fd, mmap, unmarshalMapValue) + + case text.ListOpen: + for { + tok, err := d.Read() + if err != nil { + return err + } + switch tok.Kind() { + case text.ListClose: + return nil + case text.MessageOpen: + if err := d.unmarshalMapEntry(fd, mmap, unmarshalMapValue); err != nil { + return err + } + default: + return d.unexpectedTokenError(tok) + } + } + + default: + return d.unexpectedTokenError(tok) + } +} + +// unmarshalMap unmarshals into given protoreflect.Map. A map value is a +// textproto message containing {key: , value: }. +func (d decoder) unmarshalMapEntry(fd pref.FieldDescriptor, mmap pref.Map, unmarshalMapValue func() (pref.Value, error)) error { + var key pref.MapKey + var pval pref.Value +Loop: + for { + // Read field name. + tok, err := d.Read() + if err != nil { + return err + } + switch tok.Kind() { + case text.Name: + if tok.NameKind() != text.IdentName { + if !d.opts.DiscardUnknown { + return d.newError(tok.Pos(), "unknown map entry field %q", tok.RawString()) + } + d.skipValue() + continue Loop + } + // Continue below. + case text.MessageClose: + break Loop + default: + return d.unexpectedTokenError(tok) + } + + name := tok.IdentName() + switch name { + case "key": + if !tok.HasSeparator() { + return d.syntaxError(tok.Pos(), "missing field separator :") + } + if key.IsValid() { + return d.newError(tok.Pos(), `map entry "key" cannot be repeated`) + } + val, err := d.unmarshalScalar(fd.MapKey()) + if err != nil { + return err + } + key = val.MapKey() + + case "value": + if kind := fd.MapValue().Kind(); (kind != pref.MessageKind) && (kind != pref.GroupKind) { + if !tok.HasSeparator() { + return d.syntaxError(tok.Pos(), "missing field separator :") + } + } + if pval.IsValid() { + return d.newError(tok.Pos(), `map entry "value" cannot be repeated`) + } + pval, err = unmarshalMapValue() + if err != nil { + return err + } + + default: + if !d.opts.DiscardUnknown { + return d.newError(tok.Pos(), "unknown map entry field %q", name) + } + d.skipValue() + } + } + + if !key.IsValid() { + key = fd.MapKey().Default().MapKey() + } + if !pval.IsValid() { + switch fd.MapValue().Kind() { + case pref.MessageKind, pref.GroupKind: + // If value field is not set for message/group types, construct an + // empty one as default. + pval = mmap.NewValue() + default: + pval = fd.MapValue().Default() + } + } + mmap.Set(key, pval) + return nil +} + +// unmarshalAny unmarshals an Any textproto. It can either be in expanded form +// or non-expanded form. +func (d decoder) unmarshalAny(m pref.Message, checkDelims bool) error { + var typeURL string + var bValue []byte + + // hasFields tracks which valid fields have been seen in the loop below in + // order to flag an error if there are duplicates or conflicts. It may + // contain the strings "type_url", "value" and "expanded". The literal + // "expanded" is used to indicate that the expanded form has been + // encountered already. + hasFields := map[string]bool{} + + if checkDelims { + tok, err := d.Read() + if err != nil { + return err + } + + if tok.Kind() != text.MessageOpen { + return d.unexpectedTokenError(tok) + } + } + +Loop: + for { + // Read field name. Can only have 3 possible field names, i.e. type_url, + // value and type URL name inside []. + tok, err := d.Read() + if err != nil { + return err + } + if typ := tok.Kind(); typ != text.Name { + if checkDelims { + if typ == text.MessageClose { + break Loop + } + } else if typ == text.EOF { + break Loop + } + return d.unexpectedTokenError(tok) + } + + switch tok.NameKind() { + case text.IdentName: + // Both type_url and value fields require field separator :. + if !tok.HasSeparator() { + return d.syntaxError(tok.Pos(), "missing field separator :") + } + + switch tok.IdentName() { + case "type_url": + if hasFields["type_url"] { + return d.newError(tok.Pos(), "duplicate Any type_url field") + } + if hasFields["expanded"] { + return d.newError(tok.Pos(), "conflict with [%s] field", typeURL) + } + tok, err := d.Read() + if err != nil { + return err + } + var ok bool + typeURL, ok = tok.String() + if !ok { + return d.newError(tok.Pos(), "invalid Any type_url: %v", tok.RawString()) + } + hasFields["type_url"] = true + + case "value": + if hasFields["value"] { + return d.newError(tok.Pos(), "duplicate Any value field") + } + if hasFields["expanded"] { + return d.newError(tok.Pos(), "conflict with [%s] field", typeURL) + } + tok, err := d.Read() + if err != nil { + return err + } + s, ok := tok.String() + if !ok { + return d.newError(tok.Pos(), "invalid Any value: %v", tok.RawString()) + } + bValue = []byte(s) + hasFields["value"] = true + + default: + if !d.opts.DiscardUnknown { + return d.newError(tok.Pos(), "invalid field name %q in google.protobuf.Any message", tok.RawString()) + } + } + + case text.TypeName: + if hasFields["expanded"] { + return d.newError(tok.Pos(), "cannot have more than one type") + } + if hasFields["type_url"] { + return d.newError(tok.Pos(), "conflict with type_url field") + } + typeURL = tok.TypeName() + var err error + bValue, err = d.unmarshalExpandedAny(typeURL, tok.Pos()) + if err != nil { + return err + } + hasFields["expanded"] = true + + default: + if !d.opts.DiscardUnknown { + return d.newError(tok.Pos(), "invalid field name %q in google.protobuf.Any message", tok.RawString()) + } + } + } + + fds := m.Descriptor().Fields() + if len(typeURL) > 0 { + m.Set(fds.ByNumber(fieldnum.Any_TypeUrl), pref.ValueOfString(typeURL)) + } + if len(bValue) > 0 { + m.Set(fds.ByNumber(fieldnum.Any_Value), pref.ValueOfBytes(bValue)) + } + return nil +} + +func (d decoder) unmarshalExpandedAny(typeURL string, pos int) ([]byte, error) { + mt, err := d.opts.Resolver.FindMessageByURL(typeURL) + if err != nil { + return nil, d.newError(pos, "unable to resolve message [%v]: %v", typeURL, err) + } + // Create new message for the embedded message type and unmarshal the value + // field into it. + m := mt.New() + if err := d.unmarshalMessage(m, true); err != nil { + return nil, err + } + // Serialize the embedded message and return the resulting bytes. + b, err := proto.MarshalOptions{ + AllowPartial: true, // Never check required fields inside an Any. + Deterministic: true, + }.Marshal(m.Interface()) + if err != nil { + return nil, d.newError(pos, "error in marshaling message into Any.value: %v", err) + } + return b, nil +} + +// skipValue makes the decoder parse a field value in order to advance the read +// to the next field. It relies on Read returning an error if the types are not +// in valid sequence. +func (d decoder) skipValue() error { + tok, err := d.Read() + if err != nil { + return err + } + // Only need to continue reading for messages and lists. + switch tok.Kind() { + case text.MessageOpen: + return d.skipMessageValue() + + case text.ListOpen: + for { + tok, err := d.Read() + if err != nil { + return err + } + switch tok.Kind() { + case text.ListClose: + return nil + case text.MessageOpen: + return d.skipMessageValue() + default: + // Skip items. This will not validate whether skipped values are + // of the same type or not, same behavior as C++ + // TextFormat::Parser::AllowUnknownField(true) version 3.8.0. + if err := d.skipValue(); err != nil { + return err + } + } + } + } + return nil +} + +// skipMessageValue makes the decoder parse and skip over all fields in a +// message. It assumes that the previous read type is MessageOpen. +func (d decoder) skipMessageValue() error { + for { + tok, err := d.Read() + if err != nil { + return err + } + switch tok.Kind() { + case text.MessageClose: + return nil + case text.Name: + if err := d.skipValue(); err != nil { + return err + } + } + } +} diff --git a/vendor/google.golang.org/protobuf/encoding/prototext/doc.go b/vendor/google.golang.org/protobuf/encoding/prototext/doc.go new file mode 100644 index 00000000..162b4f98 --- /dev/null +++ b/vendor/google.golang.org/protobuf/encoding/prototext/doc.go @@ -0,0 +1,7 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package prototext marshals and unmarshals protocol buffer messages as the +// textproto format. +package prototext diff --git a/vendor/google.golang.org/protobuf/encoding/prototext/encode.go b/vendor/google.golang.org/protobuf/encoding/prototext/encode.go new file mode 100644 index 00000000..41e5c773 --- /dev/null +++ b/vendor/google.golang.org/protobuf/encoding/prototext/encode.go @@ -0,0 +1,433 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package prototext + +import ( + "fmt" + "sort" + "strconv" + "unicode/utf8" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/encoding/text" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/fieldnum" + "google.golang.org/protobuf/internal/flags" + "google.golang.org/protobuf/internal/mapsort" + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/internal/strs" + "google.golang.org/protobuf/proto" + pref "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +const defaultIndent = " " + +// Format formats the message as a multiline string. +// This function is only intended for human consumption and ignores errors. +// Do not depend on the output being stable. It may change over time across +// different versions of the program. +func Format(m proto.Message) string { + return MarshalOptions{Multiline: true}.Format(m) +} + +// Marshal writes the given proto.Message in textproto format using default +// options. Do not depend on the output being stable. It may change over time +// across different versions of the program. +func Marshal(m proto.Message) ([]byte, error) { + return MarshalOptions{}.Marshal(m) +} + +// MarshalOptions is a configurable text format marshaler. +type MarshalOptions struct { + pragma.NoUnkeyedLiterals + + // Multiline specifies whether the marshaler should format the output in + // indented-form with every textual element on a new line. + // If Indent is an empty string, then an arbitrary indent is chosen. + Multiline bool + + // Indent specifies the set of indentation characters to use in a multiline + // formatted output such that every entry is preceded by Indent and + // terminated by a newline. If non-empty, then Multiline is treated as true. + // Indent can only be composed of space or tab characters. + Indent string + + // EmitASCII specifies whether to format strings and bytes as ASCII only + // as opposed to using UTF-8 encoding when possible. + EmitASCII bool + + // allowInvalidUTF8 specifies whether to permit the encoding of strings + // with invalid UTF-8. This is unexported as it is intended to only + // be specified by the Format method. + allowInvalidUTF8 bool + + // AllowPartial allows messages that have missing required fields to marshal + // without returning an error. If AllowPartial is false (the default), + // Marshal will return error if there are any missing required fields. + AllowPartial bool + + // EmitUnknown specifies whether to emit unknown fields in the output. + // If specified, the unmarshaler may be unable to parse the output. + // The default is to exclude unknown fields. + EmitUnknown bool + + // Resolver is used for looking up types when expanding google.protobuf.Any + // messages. If nil, this defaults to using protoregistry.GlobalTypes. + Resolver interface { + protoregistry.ExtensionTypeResolver + protoregistry.MessageTypeResolver + } +} + +// Format formats the message as a string. +// This method is only intended for human consumption and ignores errors. +// Do not depend on the output being stable. It may change over time across +// different versions of the program. +func (o MarshalOptions) Format(m proto.Message) string { + if m == nil || !m.ProtoReflect().IsValid() { + return "" // invalid syntax, but okay since this is for debugging + } + o.allowInvalidUTF8 = true + o.AllowPartial = true + o.EmitUnknown = true + b, _ := o.Marshal(m) + return string(b) +} + +// Marshal writes the given proto.Message in textproto format using options in +// MarshalOptions object. Do not depend on the output being stable. It may +// change over time across different versions of the program. +func (o MarshalOptions) Marshal(m proto.Message) ([]byte, error) { + return o.marshal(m) +} + +// marshal is a centralized function that all marshal operations go through. +// For profiling purposes, avoid changing the name of this function or +// introducing other code paths for marshal that do not go through this. +func (o MarshalOptions) marshal(m proto.Message) ([]byte, error) { + var delims = [2]byte{'{', '}'} + + if o.Multiline && o.Indent == "" { + o.Indent = defaultIndent + } + if o.Resolver == nil { + o.Resolver = protoregistry.GlobalTypes + } + + internalEnc, err := text.NewEncoder(o.Indent, delims, o.EmitASCII) + if err != nil { + return nil, err + } + + // Treat nil message interface as an empty message, + // in which case there is nothing to output. + if m == nil { + return []byte{}, nil + } + + enc := encoder{internalEnc, o} + err = enc.marshalMessage(m.ProtoReflect(), false) + if err != nil { + return nil, err + } + out := enc.Bytes() + if len(o.Indent) > 0 && len(out) > 0 { + out = append(out, '\n') + } + if o.AllowPartial { + return out, nil + } + return out, proto.CheckInitialized(m) +} + +type encoder struct { + *text.Encoder + opts MarshalOptions +} + +// marshalMessage marshals the given protoreflect.Message. +func (e encoder) marshalMessage(m pref.Message, inclDelims bool) error { + messageDesc := m.Descriptor() + if !flags.ProtoLegacy && messageset.IsMessageSet(messageDesc) { + return errors.New("no support for proto1 MessageSets") + } + + if inclDelims { + e.StartMessage() + defer e.EndMessage() + } + + // Handle Any expansion. + if messageDesc.FullName() == "google.protobuf.Any" { + if e.marshalAny(m) { + return nil + } + // If unable to expand, continue on to marshal Any as a regular message. + } + + // Marshal known fields. + fieldDescs := messageDesc.Fields() + size := fieldDescs.Len() + for i := 0; i < size; { + fd := fieldDescs.Get(i) + if od := fd.ContainingOneof(); od != nil { + fd = m.WhichOneof(od) + i += od.Fields().Len() + } else { + i++ + } + + if fd == nil || !m.Has(fd) { + continue + } + + name := fd.Name() + // Use type name for group field name. + if fd.Kind() == pref.GroupKind { + name = fd.Message().Name() + } + val := m.Get(fd) + if err := e.marshalField(string(name), val, fd); err != nil { + return err + } + } + + // Marshal extensions. + if err := e.marshalExtensions(m); err != nil { + return err + } + + // Marshal unknown fields. + if e.opts.EmitUnknown { + e.marshalUnknown(m.GetUnknown()) + } + + return nil +} + +// marshalField marshals the given field with protoreflect.Value. +func (e encoder) marshalField(name string, val pref.Value, fd pref.FieldDescriptor) error { + switch { + case fd.IsList(): + return e.marshalList(name, val.List(), fd) + case fd.IsMap(): + return e.marshalMap(name, val.Map(), fd) + default: + e.WriteName(name) + return e.marshalSingular(val, fd) + } +} + +// marshalSingular marshals the given non-repeated field value. This includes +// all scalar types, enums, messages, and groups. +func (e encoder) marshalSingular(val pref.Value, fd pref.FieldDescriptor) error { + kind := fd.Kind() + switch kind { + case pref.BoolKind: + e.WriteBool(val.Bool()) + + case pref.StringKind: + s := val.String() + if !e.opts.allowInvalidUTF8 && strs.EnforceUTF8(fd) && !utf8.ValidString(s) { + return errors.InvalidUTF8(string(fd.FullName())) + } + e.WriteString(s) + + case pref.Int32Kind, pref.Int64Kind, + pref.Sint32Kind, pref.Sint64Kind, + pref.Sfixed32Kind, pref.Sfixed64Kind: + e.WriteInt(val.Int()) + + case pref.Uint32Kind, pref.Uint64Kind, + pref.Fixed32Kind, pref.Fixed64Kind: + e.WriteUint(val.Uint()) + + case pref.FloatKind: + // Encoder.WriteFloat handles the special numbers NaN and infinites. + e.WriteFloat(val.Float(), 32) + + case pref.DoubleKind: + // Encoder.WriteFloat handles the special numbers NaN and infinites. + e.WriteFloat(val.Float(), 64) + + case pref.BytesKind: + e.WriteString(string(val.Bytes())) + + case pref.EnumKind: + num := val.Enum() + if desc := fd.Enum().Values().ByNumber(num); desc != nil { + e.WriteLiteral(string(desc.Name())) + } else { + // Use numeric value if there is no enum description. + e.WriteInt(int64(num)) + } + + case pref.MessageKind, pref.GroupKind: + return e.marshalMessage(val.Message(), true) + + default: + panic(fmt.Sprintf("%v has unknown kind: %v", fd.FullName(), kind)) + } + return nil +} + +// marshalList marshals the given protoreflect.List as multiple name-value fields. +func (e encoder) marshalList(name string, list pref.List, fd pref.FieldDescriptor) error { + size := list.Len() + for i := 0; i < size; i++ { + e.WriteName(name) + if err := e.marshalSingular(list.Get(i), fd); err != nil { + return err + } + } + return nil +} + +// marshalMap marshals the given protoreflect.Map as multiple name-value fields. +func (e encoder) marshalMap(name string, mmap pref.Map, fd pref.FieldDescriptor) error { + var err error + mapsort.Range(mmap, fd.MapKey().Kind(), func(key pref.MapKey, val pref.Value) bool { + e.WriteName(name) + e.StartMessage() + defer e.EndMessage() + + e.WriteName("key") + err = e.marshalSingular(key.Value(), fd.MapKey()) + if err != nil { + return false + } + + e.WriteName("value") + err = e.marshalSingular(val, fd.MapValue()) + if err != nil { + return false + } + return true + }) + return err +} + +// marshalExtensions marshals extension fields. +func (e encoder) marshalExtensions(m pref.Message) error { + type entry struct { + key string + value pref.Value + desc pref.FieldDescriptor + } + + // Get a sorted list based on field key first. + var entries []entry + m.Range(func(fd pref.FieldDescriptor, v pref.Value) bool { + if !fd.IsExtension() { + return true + } + // For MessageSet extensions, the name used is the parent message. + name := fd.FullName() + if messageset.IsMessageSetExtension(fd) { + name = name.Parent() + } + entries = append(entries, entry{ + key: string(name), + value: v, + desc: fd, + }) + return true + }) + // Sort extensions lexicographically. + sort.Slice(entries, func(i, j int) bool { + return entries[i].key < entries[j].key + }) + + // Write out sorted list. + for _, entry := range entries { + // Extension field name is the proto field name enclosed in []. + name := "[" + entry.key + "]" + if err := e.marshalField(name, entry.value, entry.desc); err != nil { + return err + } + } + return nil +} + +// marshalUnknown parses the given []byte and marshals fields out. +// This function assumes proper encoding in the given []byte. +func (e encoder) marshalUnknown(b []byte) { + const dec = 10 + const hex = 16 + for len(b) > 0 { + num, wtype, n := protowire.ConsumeTag(b) + b = b[n:] + e.WriteName(strconv.FormatInt(int64(num), dec)) + + switch wtype { + case protowire.VarintType: + var v uint64 + v, n = protowire.ConsumeVarint(b) + e.WriteUint(v) + case protowire.Fixed32Type: + var v uint32 + v, n = protowire.ConsumeFixed32(b) + e.WriteLiteral("0x" + strconv.FormatUint(uint64(v), hex)) + case protowire.Fixed64Type: + var v uint64 + v, n = protowire.ConsumeFixed64(b) + e.WriteLiteral("0x" + strconv.FormatUint(v, hex)) + case protowire.BytesType: + var v []byte + v, n = protowire.ConsumeBytes(b) + e.WriteString(string(v)) + case protowire.StartGroupType: + e.StartMessage() + var v []byte + v, n = protowire.ConsumeGroup(num, b) + e.marshalUnknown(v) + e.EndMessage() + default: + panic(fmt.Sprintf("prototext: error parsing unknown field wire type: %v", wtype)) + } + + b = b[n:] + } +} + +// marshalAny marshals the given google.protobuf.Any message in expanded form. +// It returns true if it was able to marshal, else false. +func (e encoder) marshalAny(any pref.Message) bool { + // Construct the embedded message. + fds := any.Descriptor().Fields() + fdType := fds.ByNumber(fieldnum.Any_TypeUrl) + typeURL := any.Get(fdType).String() + mt, err := e.opts.Resolver.FindMessageByURL(typeURL) + if err != nil { + return false + } + m := mt.New().Interface() + + // Unmarshal bytes into embedded message. + fdValue := fds.ByNumber(fieldnum.Any_Value) + value := any.Get(fdValue) + err = proto.UnmarshalOptions{ + AllowPartial: true, + Resolver: e.opts.Resolver, + }.Unmarshal(value.Bytes(), m) + if err != nil { + return false + } + + // Get current encoder position. If marshaling fails, reset encoder output + // back to this position. + pos := e.Snapshot() + + // Field name is the proto field name enclosed in []. + e.WriteName("[" + typeURL + "]") + err = e.marshalMessage(m.ProtoReflect(), true) + if err != nil { + e.Reset(pos) + return false + } + return true +} diff --git a/vendor/google.golang.org/protobuf/encoding/protowire/wire.go b/vendor/google.golang.org/protobuf/encoding/protowire/wire.go new file mode 100644 index 00000000..a427f8b7 --- /dev/null +++ b/vendor/google.golang.org/protobuf/encoding/protowire/wire.go @@ -0,0 +1,538 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package protowire parses and formats the raw wire encoding. +// See https://developers.google.com/protocol-buffers/docs/encoding. +// +// For marshaling and unmarshaling entire protobuf messages, +// use the "google.golang.org/protobuf/proto" package instead. +package protowire + +import ( + "io" + "math" + "math/bits" + + "google.golang.org/protobuf/internal/errors" +) + +// Number represents the field number. +type Number int32 + +const ( + MinValidNumber Number = 1 + FirstReservedNumber Number = 19000 + LastReservedNumber Number = 19999 + MaxValidNumber Number = 1<<29 - 1 +) + +// IsValid reports whether the field number is semantically valid. +// +// Note that while numbers within the reserved range are semantically invalid, +// they are syntactically valid in the wire format. +// Implementations may treat records with reserved field numbers as unknown. +func (n Number) IsValid() bool { + return MinValidNumber <= n && n < FirstReservedNumber || LastReservedNumber < n && n <= MaxValidNumber +} + +// Type represents the wire type. +type Type int8 + +const ( + VarintType Type = 0 + Fixed32Type Type = 5 + Fixed64Type Type = 1 + BytesType Type = 2 + StartGroupType Type = 3 + EndGroupType Type = 4 +) + +const ( + _ = -iota + errCodeTruncated + errCodeFieldNumber + errCodeOverflow + errCodeReserved + errCodeEndGroup +) + +var ( + errFieldNumber = errors.New("invalid field number") + errOverflow = errors.New("variable length integer overflow") + errReserved = errors.New("cannot parse reserved wire type") + errEndGroup = errors.New("mismatching end group marker") + errParse = errors.New("parse error") +) + +// ParseError converts an error code into an error value. +// This returns nil if n is a non-negative number. +func ParseError(n int) error { + if n >= 0 { + return nil + } + switch n { + case errCodeTruncated: + return io.ErrUnexpectedEOF + case errCodeFieldNumber: + return errFieldNumber + case errCodeOverflow: + return errOverflow + case errCodeReserved: + return errReserved + case errCodeEndGroup: + return errEndGroup + default: + return errParse + } +} + +// ConsumeField parses an entire field record (both tag and value) and returns +// the field number, the wire type, and the total length. +// This returns a negative length upon an error (see ParseError). +// +// The total length includes the tag header and the end group marker (if the +// field is a group). +func ConsumeField(b []byte) (Number, Type, int) { + num, typ, n := ConsumeTag(b) + if n < 0 { + return 0, 0, n // forward error code + } + m := ConsumeFieldValue(num, typ, b[n:]) + if m < 0 { + return 0, 0, m // forward error code + } + return num, typ, n + m +} + +// ConsumeFieldValue parses a field value and returns its length. +// This assumes that the field Number and wire Type have already been parsed. +// This returns a negative length upon an error (see ParseError). +// +// When parsing a group, the length includes the end group marker and +// the end group is verified to match the starting field number. +func ConsumeFieldValue(num Number, typ Type, b []byte) (n int) { + switch typ { + case VarintType: + _, n = ConsumeVarint(b) + return n + case Fixed32Type: + _, n = ConsumeFixed32(b) + return n + case Fixed64Type: + _, n = ConsumeFixed64(b) + return n + case BytesType: + _, n = ConsumeBytes(b) + return n + case StartGroupType: + n0 := len(b) + for { + num2, typ2, n := ConsumeTag(b) + if n < 0 { + return n // forward error code + } + b = b[n:] + if typ2 == EndGroupType { + if num != num2 { + return errCodeEndGroup + } + return n0 - len(b) + } + + n = ConsumeFieldValue(num2, typ2, b) + if n < 0 { + return n // forward error code + } + b = b[n:] + } + case EndGroupType: + return errCodeEndGroup + default: + return errCodeReserved + } +} + +// AppendTag encodes num and typ as a varint-encoded tag and appends it to b. +func AppendTag(b []byte, num Number, typ Type) []byte { + return AppendVarint(b, EncodeTag(num, typ)) +} + +// ConsumeTag parses b as a varint-encoded tag, reporting its length. +// This returns a negative length upon an error (see ParseError). +func ConsumeTag(b []byte) (Number, Type, int) { + v, n := ConsumeVarint(b) + if n < 0 { + return 0, 0, n // forward error code + } + num, typ := DecodeTag(v) + if num < MinValidNumber { + return 0, 0, errCodeFieldNumber + } + return num, typ, n +} + +func SizeTag(num Number) int { + return SizeVarint(EncodeTag(num, 0)) // wire type has no effect on size +} + +// AppendVarint appends v to b as a varint-encoded uint64. +func AppendVarint(b []byte, v uint64) []byte { + switch { + case v < 1<<7: + b = append(b, byte(v)) + case v < 1<<14: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte(v>>7)) + case v < 1<<21: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte(v>>14)) + case v < 1<<28: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte((v>>14)&0x7f|0x80), + byte(v>>21)) + case v < 1<<35: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte((v>>14)&0x7f|0x80), + byte((v>>21)&0x7f|0x80), + byte(v>>28)) + case v < 1<<42: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte((v>>14)&0x7f|0x80), + byte((v>>21)&0x7f|0x80), + byte((v>>28)&0x7f|0x80), + byte(v>>35)) + case v < 1<<49: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte((v>>14)&0x7f|0x80), + byte((v>>21)&0x7f|0x80), + byte((v>>28)&0x7f|0x80), + byte((v>>35)&0x7f|0x80), + byte(v>>42)) + case v < 1<<56: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte((v>>14)&0x7f|0x80), + byte((v>>21)&0x7f|0x80), + byte((v>>28)&0x7f|0x80), + byte((v>>35)&0x7f|0x80), + byte((v>>42)&0x7f|0x80), + byte(v>>49)) + case v < 1<<63: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte((v>>14)&0x7f|0x80), + byte((v>>21)&0x7f|0x80), + byte((v>>28)&0x7f|0x80), + byte((v>>35)&0x7f|0x80), + byte((v>>42)&0x7f|0x80), + byte((v>>49)&0x7f|0x80), + byte(v>>56)) + default: + b = append(b, + byte((v>>0)&0x7f|0x80), + byte((v>>7)&0x7f|0x80), + byte((v>>14)&0x7f|0x80), + byte((v>>21)&0x7f|0x80), + byte((v>>28)&0x7f|0x80), + byte((v>>35)&0x7f|0x80), + byte((v>>42)&0x7f|0x80), + byte((v>>49)&0x7f|0x80), + byte((v>>56)&0x7f|0x80), + 1) + } + return b +} + +// ConsumeVarint parses b as a varint-encoded uint64, reporting its length. +// This returns a negative length upon an error (see ParseError). +func ConsumeVarint(b []byte) (v uint64, n int) { + var y uint64 + if len(b) <= 0 { + return 0, errCodeTruncated + } + v = uint64(b[0]) + if v < 0x80 { + return v, 1 + } + v -= 0x80 + + if len(b) <= 1 { + return 0, errCodeTruncated + } + y = uint64(b[1]) + v += y << 7 + if y < 0x80 { + return v, 2 + } + v -= 0x80 << 7 + + if len(b) <= 2 { + return 0, errCodeTruncated + } + y = uint64(b[2]) + v += y << 14 + if y < 0x80 { + return v, 3 + } + v -= 0x80 << 14 + + if len(b) <= 3 { + return 0, errCodeTruncated + } + y = uint64(b[3]) + v += y << 21 + if y < 0x80 { + return v, 4 + } + v -= 0x80 << 21 + + if len(b) <= 4 { + return 0, errCodeTruncated + } + y = uint64(b[4]) + v += y << 28 + if y < 0x80 { + return v, 5 + } + v -= 0x80 << 28 + + if len(b) <= 5 { + return 0, errCodeTruncated + } + y = uint64(b[5]) + v += y << 35 + if y < 0x80 { + return v, 6 + } + v -= 0x80 << 35 + + if len(b) <= 6 { + return 0, errCodeTruncated + } + y = uint64(b[6]) + v += y << 42 + if y < 0x80 { + return v, 7 + } + v -= 0x80 << 42 + + if len(b) <= 7 { + return 0, errCodeTruncated + } + y = uint64(b[7]) + v += y << 49 + if y < 0x80 { + return v, 8 + } + v -= 0x80 << 49 + + if len(b) <= 8 { + return 0, errCodeTruncated + } + y = uint64(b[8]) + v += y << 56 + if y < 0x80 { + return v, 9 + } + v -= 0x80 << 56 + + if len(b) <= 9 { + return 0, errCodeTruncated + } + y = uint64(b[9]) + v += y << 63 + if y < 2 { + return v, 10 + } + return 0, errCodeOverflow +} + +// SizeVarint returns the encoded size of a varint. +// The size is guaranteed to be within 1 and 10, inclusive. +func SizeVarint(v uint64) int { + // This computes 1 + (bits.Len64(v)-1)/7. + // 9/64 is a good enough approximation of 1/7 + return int(9*uint32(bits.Len64(v))+64) / 64 +} + +// AppendFixed32 appends v to b as a little-endian uint32. +func AppendFixed32(b []byte, v uint32) []byte { + return append(b, + byte(v>>0), + byte(v>>8), + byte(v>>16), + byte(v>>24)) +} + +// ConsumeFixed32 parses b as a little-endian uint32, reporting its length. +// This returns a negative length upon an error (see ParseError). +func ConsumeFixed32(b []byte) (v uint32, n int) { + if len(b) < 4 { + return 0, errCodeTruncated + } + v = uint32(b[0])<<0 | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24 + return v, 4 +} + +// SizeFixed32 returns the encoded size of a fixed32; which is always 4. +func SizeFixed32() int { + return 4 +} + +// AppendFixed64 appends v to b as a little-endian uint64. +func AppendFixed64(b []byte, v uint64) []byte { + return append(b, + byte(v>>0), + byte(v>>8), + byte(v>>16), + byte(v>>24), + byte(v>>32), + byte(v>>40), + byte(v>>48), + byte(v>>56)) +} + +// ConsumeFixed64 parses b as a little-endian uint64, reporting its length. +// This returns a negative length upon an error (see ParseError). +func ConsumeFixed64(b []byte) (v uint64, n int) { + if len(b) < 8 { + return 0, errCodeTruncated + } + v = uint64(b[0])<<0 | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 | uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56 + return v, 8 +} + +// SizeFixed64 returns the encoded size of a fixed64; which is always 8. +func SizeFixed64() int { + return 8 +} + +// AppendBytes appends v to b as a length-prefixed bytes value. +func AppendBytes(b []byte, v []byte) []byte { + return append(AppendVarint(b, uint64(len(v))), v...) +} + +// ConsumeBytes parses b as a length-prefixed bytes value, reporting its length. +// This returns a negative length upon an error (see ParseError). +func ConsumeBytes(b []byte) (v []byte, n int) { + m, n := ConsumeVarint(b) + if n < 0 { + return nil, n // forward error code + } + if m > uint64(len(b[n:])) { + return nil, errCodeTruncated + } + return b[n:][:m], n + int(m) +} + +// SizeBytes returns the encoded size of a length-prefixed bytes value, +// given only the length. +func SizeBytes(n int) int { + return SizeVarint(uint64(n)) + n +} + +// AppendString appends v to b as a length-prefixed bytes value. +func AppendString(b []byte, v string) []byte { + return append(AppendVarint(b, uint64(len(v))), v...) +} + +// ConsumeString parses b as a length-prefixed bytes value, reporting its length. +// This returns a negative length upon an error (see ParseError). +func ConsumeString(b []byte) (v string, n int) { + bb, n := ConsumeBytes(b) + return string(bb), n +} + +// AppendGroup appends v to b as group value, with a trailing end group marker. +// The value v must not contain the end marker. +func AppendGroup(b []byte, num Number, v []byte) []byte { + return AppendVarint(append(b, v...), EncodeTag(num, EndGroupType)) +} + +// ConsumeGroup parses b as a group value until the trailing end group marker, +// and verifies that the end marker matches the provided num. The value v +// does not contain the end marker, while the length does contain the end marker. +// This returns a negative length upon an error (see ParseError). +func ConsumeGroup(num Number, b []byte) (v []byte, n int) { + n = ConsumeFieldValue(num, StartGroupType, b) + if n < 0 { + return nil, n // forward error code + } + b = b[:n] + + // Truncate off end group marker, but need to handle denormalized varints. + // Assuming end marker is never 0 (which is always the case since + // EndGroupType is non-zero), we can truncate all trailing bytes where the + // lower 7 bits are all zero (implying that the varint is denormalized). + for len(b) > 0 && b[len(b)-1]&0x7f == 0 { + b = b[:len(b)-1] + } + b = b[:len(b)-SizeTag(num)] + return b, n +} + +// SizeGroup returns the encoded size of a group, given only the length. +func SizeGroup(num Number, n int) int { + return n + SizeTag(num) +} + +// DecodeTag decodes the field Number and wire Type from its unified form. +// The Number is -1 if the decoded field number overflows int32. +// Other than overflow, this does not check for field number validity. +func DecodeTag(x uint64) (Number, Type) { + // NOTE: MessageSet allows for larger field numbers than normal. + if x>>3 > uint64(math.MaxInt32) { + return -1, 0 + } + return Number(x >> 3), Type(x & 7) +} + +// EncodeTag encodes the field Number and wire Type into its unified form. +func EncodeTag(num Number, typ Type) uint64 { + return uint64(num)<<3 | uint64(typ&7) +} + +// DecodeZigZag decodes a zig-zag-encoded uint64 as an int64. +// Input: {…, 5, 3, 1, 0, 2, 4, 6, …} +// Output: {…, -3, -2, -1, 0, +1, +2, +3, …} +func DecodeZigZag(x uint64) int64 { + return int64(x>>1) ^ int64(x)<<63>>63 +} + +// EncodeZigZag encodes an int64 as a zig-zag-encoded uint64. +// Input: {…, -3, -2, -1, 0, +1, +2, +3, …} +// Output: {…, 5, 3, 1, 0, 2, 4, 6, …} +func EncodeZigZag(x int64) uint64 { + return uint64(x<<1) ^ uint64(x>>63) +} + +// DecodeBool decodes a uint64 as a bool. +// Input: { 0, 1, 2, …} +// Output: {false, true, true, …} +func DecodeBool(x uint64) bool { + return x != 0 +} + +// EncodeBool encodes a bool as a uint64. +// Input: {false, true} +// Output: { 0, 1} +func EncodeBool(x bool) uint64 { + if x { + return 1 + } + return 0 +} diff --git a/vendor/google.golang.org/protobuf/internal/descfmt/stringer.go b/vendor/google.golang.org/protobuf/internal/descfmt/stringer.go new file mode 100644 index 00000000..e7af0fe0 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/descfmt/stringer.go @@ -0,0 +1,316 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package descfmt provides functionality to format descriptors. +package descfmt + +import ( + "fmt" + "io" + "reflect" + "strconv" + "strings" + + "google.golang.org/protobuf/internal/detrand" + "google.golang.org/protobuf/internal/pragma" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type list interface { + Len() int + pragma.DoNotImplement +} + +func FormatList(s fmt.State, r rune, vs list) { + io.WriteString(s, formatListOpt(vs, true, r == 'v' && (s.Flag('+') || s.Flag('#')))) +} +func formatListOpt(vs list, isRoot, allowMulti bool) string { + start, end := "[", "]" + if isRoot { + var name string + switch vs.(type) { + case pref.Names: + name = "Names" + case pref.FieldNumbers: + name = "FieldNumbers" + case pref.FieldRanges: + name = "FieldRanges" + case pref.EnumRanges: + name = "EnumRanges" + case pref.FileImports: + name = "FileImports" + case pref.Descriptor: + name = reflect.ValueOf(vs).MethodByName("Get").Type().Out(0).Name() + "s" + } + start, end = name+"{", "}" + } + + var ss []string + switch vs := vs.(type) { + case pref.Names: + for i := 0; i < vs.Len(); i++ { + ss = append(ss, fmt.Sprint(vs.Get(i))) + } + return start + joinStrings(ss, false) + end + case pref.FieldNumbers: + for i := 0; i < vs.Len(); i++ { + ss = append(ss, fmt.Sprint(vs.Get(i))) + } + return start + joinStrings(ss, false) + end + case pref.FieldRanges: + for i := 0; i < vs.Len(); i++ { + r := vs.Get(i) + if r[0]+1 == r[1] { + ss = append(ss, fmt.Sprintf("%d", r[0])) + } else { + ss = append(ss, fmt.Sprintf("%d:%d", r[0], r[1])) // enum ranges are end exclusive + } + } + return start + joinStrings(ss, false) + end + case pref.EnumRanges: + for i := 0; i < vs.Len(); i++ { + r := vs.Get(i) + if r[0] == r[1] { + ss = append(ss, fmt.Sprintf("%d", r[0])) + } else { + ss = append(ss, fmt.Sprintf("%d:%d", r[0], int64(r[1])+1)) // enum ranges are end inclusive + } + } + return start + joinStrings(ss, false) + end + case pref.FileImports: + for i := 0; i < vs.Len(); i++ { + var rs records + rs.Append(reflect.ValueOf(vs.Get(i)), "Path", "Package", "IsPublic", "IsWeak") + ss = append(ss, "{"+rs.Join()+"}") + } + return start + joinStrings(ss, allowMulti) + end + default: + _, isEnumValue := vs.(pref.EnumValueDescriptors) + for i := 0; i < vs.Len(); i++ { + m := reflect.ValueOf(vs).MethodByName("Get") + v := m.Call([]reflect.Value{reflect.ValueOf(i)})[0].Interface() + ss = append(ss, formatDescOpt(v.(pref.Descriptor), false, allowMulti && !isEnumValue)) + } + return start + joinStrings(ss, allowMulti && isEnumValue) + end + } +} + +// descriptorAccessors is a list of accessors to print for each descriptor. +// +// Do not print all accessors since some contain redundant information, +// while others are pointers that we do not want to follow since the descriptor +// is actually a cyclic graph. +// +// Using a list allows us to print the accessors in a sensible order. +var descriptorAccessors = map[reflect.Type][]string{ + reflect.TypeOf((*pref.FileDescriptor)(nil)).Elem(): {"Path", "Package", "Imports", "Messages", "Enums", "Extensions", "Services"}, + reflect.TypeOf((*pref.MessageDescriptor)(nil)).Elem(): {"IsMapEntry", "Fields", "Oneofs", "ReservedNames", "ReservedRanges", "RequiredNumbers", "ExtensionRanges", "Messages", "Enums", "Extensions"}, + reflect.TypeOf((*pref.FieldDescriptor)(nil)).Elem(): {"Number", "Cardinality", "Kind", "HasJSONName", "JSONName", "HasPresence", "IsExtension", "IsPacked", "IsWeak", "IsList", "IsMap", "MapKey", "MapValue", "HasDefault", "Default", "ContainingOneof", "ContainingMessage", "Message", "Enum"}, + reflect.TypeOf((*pref.OneofDescriptor)(nil)).Elem(): {"Fields"}, // not directly used; must keep in sync with formatDescOpt + reflect.TypeOf((*pref.EnumDescriptor)(nil)).Elem(): {"Values", "ReservedNames", "ReservedRanges"}, + reflect.TypeOf((*pref.EnumValueDescriptor)(nil)).Elem(): {"Number"}, + reflect.TypeOf((*pref.ServiceDescriptor)(nil)).Elem(): {"Methods"}, + reflect.TypeOf((*pref.MethodDescriptor)(nil)).Elem(): {"Input", "Output", "IsStreamingClient", "IsStreamingServer"}, +} + +func FormatDesc(s fmt.State, r rune, t pref.Descriptor) { + io.WriteString(s, formatDescOpt(t, true, r == 'v' && (s.Flag('+') || s.Flag('#')))) +} +func formatDescOpt(t pref.Descriptor, isRoot, allowMulti bool) string { + rv := reflect.ValueOf(t) + rt := rv.MethodByName("ProtoType").Type().In(0) + + start, end := "{", "}" + if isRoot { + start = rt.Name() + "{" + } + + _, isFile := t.(pref.FileDescriptor) + rs := records{allowMulti: allowMulti} + if t.IsPlaceholder() { + if isFile { + rs.Append(rv, "Path", "Package", "IsPlaceholder") + } else { + rs.Append(rv, "FullName", "IsPlaceholder") + } + } else { + switch { + case isFile: + rs.Append(rv, "Syntax") + case isRoot: + rs.Append(rv, "Syntax", "FullName") + default: + rs.Append(rv, "Name") + } + switch t := t.(type) { + case pref.FieldDescriptor: + for _, s := range descriptorAccessors[rt] { + switch s { + case "MapKey": + if k := t.MapKey(); k != nil { + rs.recs = append(rs.recs, [2]string{"MapKey", k.Kind().String()}) + } + case "MapValue": + if v := t.MapValue(); v != nil { + switch v.Kind() { + case pref.EnumKind: + rs.recs = append(rs.recs, [2]string{"MapValue", string(v.Enum().FullName())}) + case pref.MessageKind, pref.GroupKind: + rs.recs = append(rs.recs, [2]string{"MapValue", string(v.Message().FullName())}) + default: + rs.recs = append(rs.recs, [2]string{"MapValue", v.Kind().String()}) + } + } + case "ContainingOneof": + if od := t.ContainingOneof(); od != nil { + rs.recs = append(rs.recs, [2]string{"Oneof", string(od.Name())}) + } + case "ContainingMessage": + if t.IsExtension() { + rs.recs = append(rs.recs, [2]string{"Extendee", string(t.ContainingMessage().FullName())}) + } + case "Message": + if !t.IsMap() { + rs.Append(rv, s) + } + default: + rs.Append(rv, s) + } + } + case pref.OneofDescriptor: + var ss []string + fs := t.Fields() + for i := 0; i < fs.Len(); i++ { + ss = append(ss, string(fs.Get(i).Name())) + } + if len(ss) > 0 { + rs.recs = append(rs.recs, [2]string{"Fields", "[" + joinStrings(ss, false) + "]"}) + } + default: + rs.Append(rv, descriptorAccessors[rt]...) + } + if rv.MethodByName("GoType").IsValid() { + rs.Append(rv, "GoType") + } + } + return start + rs.Join() + end +} + +type records struct { + recs [][2]string + allowMulti bool +} + +func (rs *records) Append(v reflect.Value, accessors ...string) { + for _, a := range accessors { + var rv reflect.Value + if m := v.MethodByName(a); m.IsValid() { + rv = m.Call(nil)[0] + } + if v.Kind() == reflect.Struct && !rv.IsValid() { + rv = v.FieldByName(a) + } + if !rv.IsValid() { + panic(fmt.Sprintf("unknown accessor: %v.%s", v.Type(), a)) + } + if _, ok := rv.Interface().(pref.Value); ok { + rv = rv.MethodByName("Interface").Call(nil)[0] + if !rv.IsNil() { + rv = rv.Elem() + } + } + + // Ignore zero values. + var isZero bool + switch rv.Kind() { + case reflect.Interface, reflect.Slice: + isZero = rv.IsNil() + case reflect.Bool: + isZero = rv.Bool() == false + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + isZero = rv.Int() == 0 + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64: + isZero = rv.Uint() == 0 + case reflect.String: + isZero = rv.String() == "" + } + if n, ok := rv.Interface().(list); ok { + isZero = n.Len() == 0 + } + if isZero { + continue + } + + // Format the value. + var s string + v := rv.Interface() + switch v := v.(type) { + case list: + s = formatListOpt(v, false, rs.allowMulti) + case pref.FieldDescriptor, pref.OneofDescriptor, pref.EnumValueDescriptor, pref.MethodDescriptor: + s = string(v.(pref.Descriptor).Name()) + case pref.Descriptor: + s = string(v.FullName()) + case string: + s = strconv.Quote(v) + case []byte: + s = fmt.Sprintf("%q", v) + default: + s = fmt.Sprint(v) + } + rs.recs = append(rs.recs, [2]string{a, s}) + } +} + +func (rs *records) Join() string { + var ss []string + + // In single line mode, simply join all records with commas. + if !rs.allowMulti { + for _, r := range rs.recs { + ss = append(ss, r[0]+formatColon(0)+r[1]) + } + return joinStrings(ss, false) + } + + // In allowMulti line mode, align single line records for more readable output. + var maxLen int + flush := func(i int) { + for _, r := range rs.recs[len(ss):i] { + ss = append(ss, r[0]+formatColon(maxLen-len(r[0]))+r[1]) + } + maxLen = 0 + } + for i, r := range rs.recs { + if isMulti := strings.Contains(r[1], "\n"); isMulti { + flush(i) + ss = append(ss, r[0]+formatColon(0)+strings.Join(strings.Split(r[1], "\n"), "\n\t")) + } else if maxLen < len(r[0]) { + maxLen = len(r[0]) + } + } + flush(len(rs.recs)) + return joinStrings(ss, true) +} + +func formatColon(padding int) string { + // Deliberately introduce instability into the debug output to + // discourage users from performing string comparisons. + // This provides us flexibility to change the output in the future. + if detrand.Bool() { + return ":" + strings.Repeat(" ", 1+padding) // use non-breaking spaces (U+00a0) + } else { + return ":" + strings.Repeat(" ", 1+padding) // use regular spaces (U+0020) + } +} + +func joinStrings(ss []string, isMulti bool) string { + if len(ss) == 0 { + return "" + } + if isMulti { + return "\n\t" + strings.Join(ss, "\n\t") + "\n" + } + return strings.Join(ss, ", ") +} diff --git a/vendor/google.golang.org/protobuf/internal/descopts/options.go b/vendor/google.golang.org/protobuf/internal/descopts/options.go new file mode 100644 index 00000000..8401be8c --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/descopts/options.go @@ -0,0 +1,29 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package descopts contains the nil pointers to concrete descriptor options. +// +// This package exists as a form of reverse dependency injection so that certain +// packages (e.g., internal/filedesc and internal/filetype can avoid a direct +// dependency on the descriptor proto package). +package descopts + +import pref "google.golang.org/protobuf/reflect/protoreflect" + +// These variables are set by the init function in descriptor.pb.go via logic +// in internal/filetype. In other words, so long as the descriptor proto package +// is linked in, these variables will be populated. +// +// Each variable is populated with a nil pointer to the options struct. +var ( + File pref.ProtoMessage + Enum pref.ProtoMessage + EnumValue pref.ProtoMessage + Message pref.ProtoMessage + Field pref.ProtoMessage + Oneof pref.ProtoMessage + ExtensionRange pref.ProtoMessage + Service pref.ProtoMessage + Method pref.ProtoMessage +) diff --git a/vendor/google.golang.org/protobuf/internal/detrand/rand.go b/vendor/google.golang.org/protobuf/internal/detrand/rand.go new file mode 100644 index 00000000..a904dd1f --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/detrand/rand.go @@ -0,0 +1,61 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package detrand provides deterministically random functionality. +// +// The pseudo-randomness of these functions is seeded by the program binary +// itself and guarantees that the output does not change within a program, +// while ensuring that the output is unstable across different builds. +package detrand + +import ( + "encoding/binary" + "hash/fnv" + "os" +) + +// Disable disables detrand such that all functions returns the zero value. +// This function is not concurrent-safe and must be called during program init. +func Disable() { + randSeed = 0 +} + +// Bool returns a deterministically random boolean. +func Bool() bool { + return randSeed%2 == 1 +} + +// randSeed is a best-effort at an approximate hash of the Go binary. +var randSeed = binaryHash() + +func binaryHash() uint64 { + // Open the Go binary. + s, err := os.Executable() + if err != nil { + return 0 + } + f, err := os.Open(s) + if err != nil { + return 0 + } + defer f.Close() + + // Hash the size and several samples of the Go binary. + const numSamples = 8 + var buf [64]byte + h := fnv.New64() + fi, err := f.Stat() + if err != nil { + return 0 + } + binary.LittleEndian.PutUint64(buf[:8], uint64(fi.Size())) + h.Write(buf[:8]) + for i := int64(0); i < numSamples; i++ { + if _, err := f.ReadAt(buf[:], i*fi.Size()/numSamples); err != nil { + return 0 + } + h.Write(buf[:]) + } + return h.Sum64() +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/defval/default.go b/vendor/google.golang.org/protobuf/internal/encoding/defval/default.go new file mode 100644 index 00000000..fdd9b13f --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/defval/default.go @@ -0,0 +1,213 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package defval marshals and unmarshals textual forms of default values. +// +// This package handles both the form historically used in Go struct field tags +// and also the form used by google.protobuf.FieldDescriptorProto.default_value +// since they differ in superficial ways. +package defval + +import ( + "fmt" + "math" + "strconv" + + ptext "google.golang.org/protobuf/internal/encoding/text" + errors "google.golang.org/protobuf/internal/errors" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +// Format is the serialization format used to represent the default value. +type Format int + +const ( + _ Format = iota + + // Descriptor uses the serialization format that protoc uses with the + // google.protobuf.FieldDescriptorProto.default_value field. + Descriptor + + // GoTag uses the historical serialization format in Go struct field tags. + GoTag +) + +// Unmarshal deserializes the default string s according to the given kind k. +// When k is an enum, a list of enum value descriptors must be provided. +func Unmarshal(s string, k pref.Kind, evs pref.EnumValueDescriptors, f Format) (pref.Value, pref.EnumValueDescriptor, error) { + switch k { + case pref.BoolKind: + if f == GoTag { + switch s { + case "1": + return pref.ValueOfBool(true), nil, nil + case "0": + return pref.ValueOfBool(false), nil, nil + } + } else { + switch s { + case "true": + return pref.ValueOfBool(true), nil, nil + case "false": + return pref.ValueOfBool(false), nil, nil + } + } + case pref.EnumKind: + if f == GoTag { + // Go tags use the numeric form of the enum value. + if n, err := strconv.ParseInt(s, 10, 32); err == nil { + if ev := evs.ByNumber(pref.EnumNumber(n)); ev != nil { + return pref.ValueOfEnum(ev.Number()), ev, nil + } + } + } else { + // Descriptor default_value use the enum identifier. + ev := evs.ByName(pref.Name(s)) + if ev != nil { + return pref.ValueOfEnum(ev.Number()), ev, nil + } + } + case pref.Int32Kind, pref.Sint32Kind, pref.Sfixed32Kind: + if v, err := strconv.ParseInt(s, 10, 32); err == nil { + return pref.ValueOfInt32(int32(v)), nil, nil + } + case pref.Int64Kind, pref.Sint64Kind, pref.Sfixed64Kind: + if v, err := strconv.ParseInt(s, 10, 64); err == nil { + return pref.ValueOfInt64(int64(v)), nil, nil + } + case pref.Uint32Kind, pref.Fixed32Kind: + if v, err := strconv.ParseUint(s, 10, 32); err == nil { + return pref.ValueOfUint32(uint32(v)), nil, nil + } + case pref.Uint64Kind, pref.Fixed64Kind: + if v, err := strconv.ParseUint(s, 10, 64); err == nil { + return pref.ValueOfUint64(uint64(v)), nil, nil + } + case pref.FloatKind, pref.DoubleKind: + var v float64 + var err error + switch s { + case "-inf": + v = math.Inf(-1) + case "inf": + v = math.Inf(+1) + case "nan": + v = math.NaN() + default: + v, err = strconv.ParseFloat(s, 64) + } + if err == nil { + if k == pref.FloatKind { + return pref.ValueOfFloat32(float32(v)), nil, nil + } else { + return pref.ValueOfFloat64(float64(v)), nil, nil + } + } + case pref.StringKind: + // String values are already unescaped and can be used as is. + return pref.ValueOfString(s), nil, nil + case pref.BytesKind: + if b, ok := unmarshalBytes(s); ok { + return pref.ValueOfBytes(b), nil, nil + } + } + return pref.Value{}, nil, errors.New("could not parse value for %v: %q", k, s) +} + +// Marshal serializes v as the default string according to the given kind k. +// When specifying the Descriptor format for an enum kind, the associated +// enum value descriptor must be provided. +func Marshal(v pref.Value, ev pref.EnumValueDescriptor, k pref.Kind, f Format) (string, error) { + switch k { + case pref.BoolKind: + if f == GoTag { + if v.Bool() { + return "1", nil + } else { + return "0", nil + } + } else { + if v.Bool() { + return "true", nil + } else { + return "false", nil + } + } + case pref.EnumKind: + if f == GoTag { + return strconv.FormatInt(int64(v.Enum()), 10), nil + } else { + return string(ev.Name()), nil + } + case pref.Int32Kind, pref.Sint32Kind, pref.Sfixed32Kind, pref.Int64Kind, pref.Sint64Kind, pref.Sfixed64Kind: + return strconv.FormatInt(v.Int(), 10), nil + case pref.Uint32Kind, pref.Fixed32Kind, pref.Uint64Kind, pref.Fixed64Kind: + return strconv.FormatUint(v.Uint(), 10), nil + case pref.FloatKind, pref.DoubleKind: + f := v.Float() + switch { + case math.IsInf(f, -1): + return "-inf", nil + case math.IsInf(f, +1): + return "inf", nil + case math.IsNaN(f): + return "nan", nil + default: + if k == pref.FloatKind { + return strconv.FormatFloat(f, 'g', -1, 32), nil + } else { + return strconv.FormatFloat(f, 'g', -1, 64), nil + } + } + case pref.StringKind: + // String values are serialized as is without any escaping. + return v.String(), nil + case pref.BytesKind: + if s, ok := marshalBytes(v.Bytes()); ok { + return s, nil + } + } + return "", errors.New("could not format value for %v: %v", k, v) +} + +// unmarshalBytes deserializes bytes by applying C unescaping. +func unmarshalBytes(s string) ([]byte, bool) { + // Bytes values use the same escaping as the text format, + // however they lack the surrounding double quotes. + v, err := ptext.UnmarshalString(`"` + s + `"`) + if err != nil { + return nil, false + } + return []byte(v), true +} + +// marshalBytes serializes bytes by using C escaping. +// To match the exact output of protoc, this is identical to the +// CEscape function in strutil.cc of the protoc source code. +func marshalBytes(b []byte) (string, bool) { + var s []byte + for _, c := range b { + switch c { + case '\n': + s = append(s, `\n`...) + case '\r': + s = append(s, `\r`...) + case '\t': + s = append(s, `\t`...) + case '"': + s = append(s, `\"`...) + case '\'': + s = append(s, `\'`...) + case '\\': + s = append(s, `\\`...) + default: + if printableASCII := c >= 0x20 && c <= 0x7e; printableASCII { + s = append(s, c) + } else { + s = append(s, fmt.Sprintf(`\%03o`, c)...) + } + } + } + return string(s), true +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/messageset/messageset.go b/vendor/google.golang.org/protobuf/internal/encoding/messageset/messageset.go new file mode 100644 index 00000000..b1eeea50 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/messageset/messageset.go @@ -0,0 +1,258 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package messageset encodes and decodes the obsolete MessageSet wire format. +package messageset + +import ( + "math" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/errors" + pref "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" +) + +// The MessageSet wire format is equivalent to a message defiend as follows, +// where each Item defines an extension field with a field number of 'type_id' +// and content of 'message'. MessageSet extensions must be non-repeated message +// fields. +// +// message MessageSet { +// repeated group Item = 1 { +// required int32 type_id = 2; +// required string message = 3; +// } +// } +const ( + FieldItem = protowire.Number(1) + FieldTypeID = protowire.Number(2) + FieldMessage = protowire.Number(3) +) + +// ExtensionName is the field name for extensions of MessageSet. +// +// A valid MessageSet extension must be of the form: +// message MyMessage { +// extend proto2.bridge.MessageSet { +// optional MyMessage message_set_extension = 1234; +// } +// ... +// } +const ExtensionName = "message_set_extension" + +// IsMessageSet returns whether the message uses the MessageSet wire format. +func IsMessageSet(md pref.MessageDescriptor) bool { + xmd, ok := md.(interface{ IsMessageSet() bool }) + return ok && xmd.IsMessageSet() +} + +// IsMessageSetExtension reports this field extends a MessageSet. +func IsMessageSetExtension(fd pref.FieldDescriptor) bool { + if fd.Name() != ExtensionName { + return false + } + if fd.FullName().Parent() != fd.Message().FullName() { + return false + } + return IsMessageSet(fd.ContainingMessage()) +} + +// FindMessageSetExtension locates a MessageSet extension field by name. +// In text and JSON formats, the extension name used is the message itself. +// The extension field name is derived by appending ExtensionName. +func FindMessageSetExtension(r preg.ExtensionTypeResolver, s pref.FullName) (pref.ExtensionType, error) { + name := s.Append(ExtensionName) + xt, err := r.FindExtensionByName(name) + if err != nil { + if err == preg.NotFound { + return nil, err + } + return nil, errors.Wrap(err, "%q", name) + } + if !IsMessageSetExtension(xt.TypeDescriptor()) { + return nil, preg.NotFound + } + return xt, nil +} + +// SizeField returns the size of a MessageSet item field containing an extension +// with the given field number, not counting the contents of the message subfield. +func SizeField(num protowire.Number) int { + return 2*protowire.SizeTag(FieldItem) + protowire.SizeTag(FieldTypeID) + protowire.SizeVarint(uint64(num)) +} + +// Unmarshal parses a MessageSet. +// +// It calls fn with the type ID and value of each item in the MessageSet. +// Unknown fields are discarded. +// +// If wantLen is true, the item values include the varint length prefix. +// This is ugly, but simplifies the fast-path decoder in internal/impl. +func Unmarshal(b []byte, wantLen bool, fn func(typeID protowire.Number, value []byte) error) error { + for len(b) > 0 { + num, wtyp, n := protowire.ConsumeTag(b) + if n < 0 { + return protowire.ParseError(n) + } + b = b[n:] + if num != FieldItem || wtyp != protowire.StartGroupType { + n := protowire.ConsumeFieldValue(num, wtyp, b) + if n < 0 { + return protowire.ParseError(n) + } + b = b[n:] + continue + } + typeID, value, n, err := ConsumeFieldValue(b, wantLen) + if err != nil { + return err + } + b = b[n:] + if typeID == 0 { + continue + } + if err := fn(typeID, value); err != nil { + return err + } + } + return nil +} + +// ConsumeFieldValue parses b as a MessageSet item field value until and including +// the trailing end group marker. It assumes the start group tag has already been parsed. +// It returns the contents of the type_id and message subfields and the total +// item length. +// +// If wantLen is true, the returned message value includes the length prefix. +func ConsumeFieldValue(b []byte, wantLen bool) (typeid protowire.Number, message []byte, n int, err error) { + ilen := len(b) + for { + num, wtyp, n := protowire.ConsumeTag(b) + if n < 0 { + return 0, nil, 0, protowire.ParseError(n) + } + b = b[n:] + switch { + case num == FieldItem && wtyp == protowire.EndGroupType: + if wantLen && len(message) == 0 { + // The message field was missing, which should never happen. + // Be prepared for this case anyway. + message = protowire.AppendVarint(message, 0) + } + return typeid, message, ilen - len(b), nil + case num == FieldTypeID && wtyp == protowire.VarintType: + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, nil, 0, protowire.ParseError(n) + } + b = b[n:] + if v < 1 || v > math.MaxInt32 { + return 0, nil, 0, errors.New("invalid type_id in message set") + } + typeid = protowire.Number(v) + case num == FieldMessage && wtyp == protowire.BytesType: + m, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, nil, 0, protowire.ParseError(n) + } + if message == nil { + if wantLen { + message = b[:n:n] + } else { + message = m[:len(m):len(m)] + } + } else { + // This case should never happen in practice, but handle it for + // correctness: The MessageSet item contains multiple message + // fields, which need to be merged. + // + // In the case where we're returning the length, this becomes + // quite inefficient since we need to strip the length off + // the existing data and reconstruct it with the combined length. + if wantLen { + _, nn := protowire.ConsumeVarint(message) + m0 := message[nn:] + message = nil + message = protowire.AppendVarint(message, uint64(len(m0)+len(m))) + message = append(message, m0...) + message = append(message, m...) + } else { + message = append(message, m...) + } + } + b = b[n:] + default: + // We have no place to put it, so we just ignore unknown fields. + n := protowire.ConsumeFieldValue(num, wtyp, b) + if n < 0 { + return 0, nil, 0, protowire.ParseError(n) + } + b = b[n:] + } + } +} + +// AppendFieldStart appends the start of a MessageSet item field containing +// an extension with the given number. The caller must add the message +// subfield (including the tag). +func AppendFieldStart(b []byte, num protowire.Number) []byte { + b = protowire.AppendTag(b, FieldItem, protowire.StartGroupType) + b = protowire.AppendTag(b, FieldTypeID, protowire.VarintType) + b = protowire.AppendVarint(b, uint64(num)) + return b +} + +// AppendFieldEnd appends the trailing end group marker for a MessageSet item field. +func AppendFieldEnd(b []byte) []byte { + return protowire.AppendTag(b, FieldItem, protowire.EndGroupType) +} + +// SizeUnknown returns the size of an unknown fields section in MessageSet format. +// +// See AppendUnknown. +func SizeUnknown(unknown []byte) (size int) { + for len(unknown) > 0 { + num, typ, n := protowire.ConsumeTag(unknown) + if n < 0 || typ != protowire.BytesType { + return 0 + } + unknown = unknown[n:] + _, n = protowire.ConsumeBytes(unknown) + if n < 0 { + return 0 + } + unknown = unknown[n:] + size += SizeField(num) + protowire.SizeTag(FieldMessage) + n + } + return size +} + +// AppendUnknown appends unknown fields to b in MessageSet format. +// +// For historic reasons, unresolved items in a MessageSet are stored in a +// message's unknown fields section in non-MessageSet format. That is, an +// unknown item with typeID T and value V appears in the unknown fields as +// a field with number T and value V. +// +// This function converts the unknown fields back into MessageSet form. +func AppendUnknown(b, unknown []byte) ([]byte, error) { + for len(unknown) > 0 { + num, typ, n := protowire.ConsumeTag(unknown) + if n < 0 || typ != protowire.BytesType { + return nil, errors.New("invalid data in message set unknown fields") + } + unknown = unknown[n:] + _, n = protowire.ConsumeBytes(unknown) + if n < 0 { + return nil, errors.New("invalid data in message set unknown fields") + } + b = AppendFieldStart(b, num) + b = protowire.AppendTag(b, FieldMessage, protowire.BytesType) + b = append(b, unknown[:n]...) + b = AppendFieldEnd(b) + unknown = unknown[n:] + } + return b, nil +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/tag/tag.go b/vendor/google.golang.org/protobuf/internal/encoding/tag/tag.go new file mode 100644 index 00000000..16c02d7b --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/tag/tag.go @@ -0,0 +1,207 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package tag marshals and unmarshals the legacy struct tags as generated +// by historical versions of protoc-gen-go. +package tag + +import ( + "reflect" + "strconv" + "strings" + + defval "google.golang.org/protobuf/internal/encoding/defval" + fdesc "google.golang.org/protobuf/internal/filedesc" + "google.golang.org/protobuf/internal/strs" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +var byteType = reflect.TypeOf(byte(0)) + +// Unmarshal decodes the tag into a prototype.Field. +// +// The goType is needed to determine the original protoreflect.Kind since the +// tag does not record sufficient information to determine that. +// The type is the underlying field type (e.g., a repeated field may be +// represented by []T, but the Go type passed in is just T). +// A list of enum value descriptors must be provided for enum fields. +// This does not populate the Enum or Message (except for weak message). +// +// This function is a best effort attempt; parsing errors are ignored. +func Unmarshal(tag string, goType reflect.Type, evs pref.EnumValueDescriptors) pref.FieldDescriptor { + f := new(fdesc.Field) + f.L0.ParentFile = fdesc.SurrogateProto2 + for len(tag) > 0 { + i := strings.IndexByte(tag, ',') + if i < 0 { + i = len(tag) + } + switch s := tag[:i]; { + case strings.HasPrefix(s, "name="): + f.L0.FullName = pref.FullName(s[len("name="):]) + case strings.Trim(s, "0123456789") == "": + n, _ := strconv.ParseUint(s, 10, 32) + f.L1.Number = pref.FieldNumber(n) + case s == "opt": + f.L1.Cardinality = pref.Optional + case s == "req": + f.L1.Cardinality = pref.Required + case s == "rep": + f.L1.Cardinality = pref.Repeated + case s == "varint": + switch goType.Kind() { + case reflect.Bool: + f.L1.Kind = pref.BoolKind + case reflect.Int32: + f.L1.Kind = pref.Int32Kind + case reflect.Int64: + f.L1.Kind = pref.Int64Kind + case reflect.Uint32: + f.L1.Kind = pref.Uint32Kind + case reflect.Uint64: + f.L1.Kind = pref.Uint64Kind + } + case s == "zigzag32": + if goType.Kind() == reflect.Int32 { + f.L1.Kind = pref.Sint32Kind + } + case s == "zigzag64": + if goType.Kind() == reflect.Int64 { + f.L1.Kind = pref.Sint64Kind + } + case s == "fixed32": + switch goType.Kind() { + case reflect.Int32: + f.L1.Kind = pref.Sfixed32Kind + case reflect.Uint32: + f.L1.Kind = pref.Fixed32Kind + case reflect.Float32: + f.L1.Kind = pref.FloatKind + } + case s == "fixed64": + switch goType.Kind() { + case reflect.Int64: + f.L1.Kind = pref.Sfixed64Kind + case reflect.Uint64: + f.L1.Kind = pref.Fixed64Kind + case reflect.Float64: + f.L1.Kind = pref.DoubleKind + } + case s == "bytes": + switch { + case goType.Kind() == reflect.String: + f.L1.Kind = pref.StringKind + case goType.Kind() == reflect.Slice && goType.Elem() == byteType: + f.L1.Kind = pref.BytesKind + default: + f.L1.Kind = pref.MessageKind + } + case s == "group": + f.L1.Kind = pref.GroupKind + case strings.HasPrefix(s, "enum="): + f.L1.Kind = pref.EnumKind + case strings.HasPrefix(s, "json="): + jsonName := s[len("json="):] + if jsonName != strs.JSONCamelCase(string(f.L0.FullName.Name())) { + f.L1.JSONName.Init(jsonName) + } + case s == "packed": + f.L1.HasPacked = true + f.L1.IsPacked = true + case strings.HasPrefix(s, "weak="): + f.L1.IsWeak = true + f.L1.Message = fdesc.PlaceholderMessage(pref.FullName(s[len("weak="):])) + case strings.HasPrefix(s, "def="): + // The default tag is special in that everything afterwards is the + // default regardless of the presence of commas. + s, i = tag[len("def="):], len(tag) + v, ev, _ := defval.Unmarshal(s, f.L1.Kind, evs, defval.GoTag) + f.L1.Default = fdesc.DefaultValue(v, ev) + case s == "proto3": + f.L0.ParentFile = fdesc.SurrogateProto3 + } + tag = strings.TrimPrefix(tag[i:], ",") + } + + // The generator uses the group message name instead of the field name. + // We obtain the real field name by lowercasing the group name. + if f.L1.Kind == pref.GroupKind { + f.L0.FullName = pref.FullName(strings.ToLower(string(f.L0.FullName))) + } + return f +} + +// Marshal encodes the protoreflect.FieldDescriptor as a tag. +// +// The enumName must be provided if the kind is an enum. +// Historically, the formulation of the enum "name" was the proto package +// dot-concatenated with the generated Go identifier for the enum type. +// Depending on the context on how Marshal is called, there are different ways +// through which that information is determined. As such it is the caller's +// responsibility to provide a function to obtain that information. +func Marshal(fd pref.FieldDescriptor, enumName string) string { + var tag []string + switch fd.Kind() { + case pref.BoolKind, pref.EnumKind, pref.Int32Kind, pref.Uint32Kind, pref.Int64Kind, pref.Uint64Kind: + tag = append(tag, "varint") + case pref.Sint32Kind: + tag = append(tag, "zigzag32") + case pref.Sint64Kind: + tag = append(tag, "zigzag64") + case pref.Sfixed32Kind, pref.Fixed32Kind, pref.FloatKind: + tag = append(tag, "fixed32") + case pref.Sfixed64Kind, pref.Fixed64Kind, pref.DoubleKind: + tag = append(tag, "fixed64") + case pref.StringKind, pref.BytesKind, pref.MessageKind: + tag = append(tag, "bytes") + case pref.GroupKind: + tag = append(tag, "group") + } + tag = append(tag, strconv.Itoa(int(fd.Number()))) + switch fd.Cardinality() { + case pref.Optional: + tag = append(tag, "opt") + case pref.Required: + tag = append(tag, "req") + case pref.Repeated: + tag = append(tag, "rep") + } + if fd.IsPacked() { + tag = append(tag, "packed") + } + name := string(fd.Name()) + if fd.Kind() == pref.GroupKind { + // The name of the FieldDescriptor for a group field is + // lowercased. To find the original capitalization, we + // look in the field's MessageType. + name = string(fd.Message().Name()) + } + tag = append(tag, "name="+name) + if jsonName := fd.JSONName(); jsonName != "" && jsonName != name && !fd.IsExtension() { + // NOTE: The jsonName != name condition is suspect, but it preserve + // the exact same semantics from the previous generator. + tag = append(tag, "json="+jsonName) + } + if fd.IsWeak() { + tag = append(tag, "weak="+string(fd.Message().FullName())) + } + // The previous implementation does not tag extension fields as proto3, + // even when the field is defined in a proto3 file. Match that behavior + // for consistency. + if fd.Syntax() == pref.Proto3 && !fd.IsExtension() { + tag = append(tag, "proto3") + } + if fd.Kind() == pref.EnumKind && enumName != "" { + tag = append(tag, "enum="+enumName) + } + if fd.ContainingOneof() != nil { + tag = append(tag, "oneof") + } + // This must appear last in the tag, since commas in strings aren't escaped. + if fd.HasDefault() { + def, _ := defval.Marshal(fd.Default(), fd.DefaultEnumValue(), fd.Kind(), defval.GoTag) + tag = append(tag, "def="+def) + } + return strings.Join(tag, ",") +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/text/decode.go b/vendor/google.golang.org/protobuf/internal/encoding/text/decode.go new file mode 100644 index 00000000..eb10ea10 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/text/decode.go @@ -0,0 +1,665 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package text + +import ( + "bytes" + "fmt" + "io" + "regexp" + "strconv" + "unicode/utf8" + + "google.golang.org/protobuf/internal/errors" +) + +// Decoder is a token-based textproto decoder. +type Decoder struct { + // lastCall is last method called, either readCall or peekCall. + // Initial value is readCall. + lastCall call + + // lastToken contains the last read token. + lastToken Token + + // lastErr contains the last read error. + lastErr error + + // openStack is a stack containing the byte characters for MessageOpen and + // ListOpen kinds. The top of stack represents the message or the list that + // the current token is nested in. An empty stack means the current token is + // at the top level message. The characters '{' and '<' both represent the + // MessageOpen kind. + openStack []byte + + // orig is used in reporting line and column. + orig []byte + // in contains the unconsumed input. + in []byte +} + +// NewDecoder returns a Decoder to read the given []byte. +func NewDecoder(b []byte) *Decoder { + return &Decoder{orig: b, in: b} +} + +// ErrUnexpectedEOF means that EOF was encountered in the middle of the input. +var ErrUnexpectedEOF = errors.New("%v", io.ErrUnexpectedEOF) + +// call specifies which Decoder method was invoked. +type call uint8 + +const ( + readCall call = iota + peekCall +) + +// Peek looks ahead and returns the next token and error without advancing a read. +func (d *Decoder) Peek() (Token, error) { + defer func() { d.lastCall = peekCall }() + if d.lastCall == readCall { + d.lastToken, d.lastErr = d.Read() + } + return d.lastToken, d.lastErr +} + +// Read returns the next token. +// It will return an error if there is no valid token. +func (d *Decoder) Read() (Token, error) { + defer func() { d.lastCall = readCall }() + if d.lastCall == peekCall { + return d.lastToken, d.lastErr + } + + tok, err := d.parseNext(d.lastToken.kind) + if err != nil { + return Token{}, err + } + + switch tok.kind { + case comma, semicolon: + tok, err = d.parseNext(tok.kind) + if err != nil { + return Token{}, err + } + } + d.lastToken = tok + return tok, nil +} + +const ( + mismatchedFmt = "mismatched close character %q" + unexpectedFmt = "unexpected character %q" +) + +// parseNext parses the next Token based on given last kind. +func (d *Decoder) parseNext(lastKind Kind) (Token, error) { + // Trim leading spaces. + d.consume(0) + isEOF := false + if len(d.in) == 0 { + isEOF = true + } + + switch lastKind { + case EOF: + return d.consumeToken(EOF, 0, 0), nil + + case bof: + // Start of top level message. Next token can be EOF or Name. + if isEOF { + return d.consumeToken(EOF, 0, 0), nil + } + return d.parseFieldName() + + case Name: + // Next token can be MessageOpen, ListOpen or Scalar. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case '{', '<': + d.pushOpenStack(ch) + return d.consumeToken(MessageOpen, 1, 0), nil + case '[': + d.pushOpenStack(ch) + return d.consumeToken(ListOpen, 1, 0), nil + default: + return d.parseScalar() + } + + case Scalar: + openKind, closeCh := d.currentOpenKind() + switch openKind { + case bof: + // Top level message. + // Next token can be EOF, comma, semicolon or Name. + if isEOF { + return d.consumeToken(EOF, 0, 0), nil + } + switch d.in[0] { + case ',': + return d.consumeToken(comma, 1, 0), nil + case ';': + return d.consumeToken(semicolon, 1, 0), nil + default: + return d.parseFieldName() + } + + case MessageOpen: + // Next token can be MessageClose, comma, semicolon or Name. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case closeCh: + d.popOpenStack() + return d.consumeToken(MessageClose, 1, 0), nil + case otherCloseChar[closeCh]: + return Token{}, d.newSyntaxError(mismatchedFmt, ch) + case ',': + return d.consumeToken(comma, 1, 0), nil + case ';': + return d.consumeToken(semicolon, 1, 0), nil + default: + return d.parseFieldName() + } + + case ListOpen: + // Next token can be ListClose or comma. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case ']': + d.popOpenStack() + return d.consumeToken(ListClose, 1, 0), nil + case ',': + return d.consumeToken(comma, 1, 0), nil + default: + return Token{}, d.newSyntaxError(unexpectedFmt, ch) + } + } + + case MessageOpen: + // Next token can be MessageClose or Name. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + _, closeCh := d.currentOpenKind() + switch ch := d.in[0]; ch { + case closeCh: + d.popOpenStack() + return d.consumeToken(MessageClose, 1, 0), nil + case otherCloseChar[closeCh]: + return Token{}, d.newSyntaxError(mismatchedFmt, ch) + default: + return d.parseFieldName() + } + + case MessageClose: + openKind, closeCh := d.currentOpenKind() + switch openKind { + case bof: + // Top level message. + // Next token can be EOF, comma, semicolon or Name. + if isEOF { + return d.consumeToken(EOF, 0, 0), nil + } + switch ch := d.in[0]; ch { + case ',': + return d.consumeToken(comma, 1, 0), nil + case ';': + return d.consumeToken(semicolon, 1, 0), nil + default: + return d.parseFieldName() + } + + case MessageOpen: + // Next token can be MessageClose, comma, semicolon or Name. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case closeCh: + d.popOpenStack() + return d.consumeToken(MessageClose, 1, 0), nil + case otherCloseChar[closeCh]: + return Token{}, d.newSyntaxError(mismatchedFmt, ch) + case ',': + return d.consumeToken(comma, 1, 0), nil + case ';': + return d.consumeToken(semicolon, 1, 0), nil + default: + return d.parseFieldName() + } + + case ListOpen: + // Next token can be ListClose or comma + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case closeCh: + d.popOpenStack() + return d.consumeToken(ListClose, 1, 0), nil + case ',': + return d.consumeToken(comma, 1, 0), nil + default: + return Token{}, d.newSyntaxError(unexpectedFmt, ch) + } + } + + case ListOpen: + // Next token can be ListClose, MessageStart or Scalar. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case ']': + d.popOpenStack() + return d.consumeToken(ListClose, 1, 0), nil + case '{', '<': + d.pushOpenStack(ch) + return d.consumeToken(MessageOpen, 1, 0), nil + default: + return d.parseScalar() + } + + case ListClose: + openKind, closeCh := d.currentOpenKind() + switch openKind { + case bof: + // Top level message. + // Next token can be EOF, comma, semicolon or Name. + if isEOF { + return d.consumeToken(EOF, 0, 0), nil + } + switch ch := d.in[0]; ch { + case ',': + return d.consumeToken(comma, 1, 0), nil + case ';': + return d.consumeToken(semicolon, 1, 0), nil + default: + return d.parseFieldName() + } + + case MessageOpen: + // Next token can be MessageClose, comma, semicolon or Name. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case closeCh: + d.popOpenStack() + return d.consumeToken(MessageClose, 1, 0), nil + case otherCloseChar[closeCh]: + return Token{}, d.newSyntaxError(mismatchedFmt, ch) + case ',': + return d.consumeToken(comma, 1, 0), nil + case ';': + return d.consumeToken(semicolon, 1, 0), nil + default: + return d.parseFieldName() + } + + default: + // It is not possible to have this case. Let it panic below. + } + + case comma, semicolon: + openKind, closeCh := d.currentOpenKind() + switch openKind { + case bof: + // Top level message. Next token can be EOF or Name. + if isEOF { + return d.consumeToken(EOF, 0, 0), nil + } + return d.parseFieldName() + + case MessageOpen: + // Next token can be MessageClose or Name. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case closeCh: + d.popOpenStack() + return d.consumeToken(MessageClose, 1, 0), nil + case otherCloseChar[closeCh]: + return Token{}, d.newSyntaxError(mismatchedFmt, ch) + default: + return d.parseFieldName() + } + + case ListOpen: + if lastKind == semicolon { + // It is not be possible to have this case as logic here + // should not have produced a semicolon Token when inside a + // list. Let it panic below. + break + } + // Next token can be MessageOpen or Scalar. + if isEOF { + return Token{}, ErrUnexpectedEOF + } + switch ch := d.in[0]; ch { + case '{', '<': + d.pushOpenStack(ch) + return d.consumeToken(MessageOpen, 1, 0), nil + default: + return d.parseScalar() + } + } + } + + line, column := d.Position(len(d.orig) - len(d.in)) + panic(fmt.Sprintf("Decoder.parseNext: bug at handling line %d:%d with lastKind=%v", line, column, lastKind)) +} + +var otherCloseChar = map[byte]byte{ + '}': '>', + '>': '}', +} + +// currentOpenKind indicates whether current position is inside a message, list +// or top-level message by returning MessageOpen, ListOpen or bof respectively. +// If the returned kind is either a MessageOpen or ListOpen, it also returns the +// corresponding closing character. +func (d *Decoder) currentOpenKind() (Kind, byte) { + if len(d.openStack) == 0 { + return bof, 0 + } + openCh := d.openStack[len(d.openStack)-1] + switch openCh { + case '{': + return MessageOpen, '}' + case '<': + return MessageOpen, '>' + case '[': + return ListOpen, ']' + } + panic(fmt.Sprintf("Decoder: openStack contains invalid byte %s", string(openCh))) +} + +func (d *Decoder) pushOpenStack(ch byte) { + d.openStack = append(d.openStack, ch) +} + +func (d *Decoder) popOpenStack() { + d.openStack = d.openStack[:len(d.openStack)-1] +} + +// parseFieldName parses field name and separator. +func (d *Decoder) parseFieldName() (tok Token, err error) { + defer func() { + if err == nil && d.tryConsumeChar(':') { + tok.attrs |= hasSeparator + } + }() + + // Extension or Any type URL. + if d.in[0] == '[' { + return d.parseTypeName() + } + + // Identifier. + if size := parseIdent(d.in, false); size > 0 { + return d.consumeToken(Name, size, uint8(IdentName)), nil + } + + // Field number. Identify if input is a valid number that is not negative + // and is decimal integer within 32-bit range. + if num := parseNumber(d.in); num.size > 0 { + if !num.neg && num.kind == numDec { + if _, err := strconv.ParseInt(string(d.in[:num.size]), 10, 32); err == nil { + return d.consumeToken(Name, num.size, uint8(FieldNumber)), nil + } + } + return Token{}, d.newSyntaxError("invalid field number: %s", d.in[:num.size]) + } + + return Token{}, d.newSyntaxError("invalid field name: %s", errRegexp.Find(d.in)) +} + +// parseTypeName parses Any type URL or extension field name. The name is +// enclosed in [ and ] characters. The C++ parser does not handle many legal URL +// strings. This implementation is more liberal and allows for the pattern +// ^[-_a-zA-Z0-9]+([./][-_a-zA-Z0-9]+)*`). Whitespaces and comments are allowed +// in between [ ], '.', '/' and the sub names. +func (d *Decoder) parseTypeName() (Token, error) { + startPos := len(d.orig) - len(d.in) + // Use alias s to advance first in order to use d.in for error handling. + // Caller already checks for [ as first character. + s := consume(d.in[1:], 0) + if len(s) == 0 { + return Token{}, ErrUnexpectedEOF + } + + var name []byte + for len(s) > 0 && isTypeNameChar(s[0]) { + name = append(name, s[0]) + s = s[1:] + } + s = consume(s, 0) + + var closed bool + for len(s) > 0 && !closed { + switch { + case s[0] == ']': + s = s[1:] + closed = true + + case s[0] == '/', s[0] == '.': + if len(name) > 0 && (name[len(name)-1] == '/' || name[len(name)-1] == '.') { + return Token{}, d.newSyntaxError("invalid type URL/extension field name: %s", + d.orig[startPos:len(d.orig)-len(s)+1]) + } + name = append(name, s[0]) + s = s[1:] + s = consume(s, 0) + for len(s) > 0 && isTypeNameChar(s[0]) { + name = append(name, s[0]) + s = s[1:] + } + s = consume(s, 0) + + default: + return Token{}, d.newSyntaxError( + "invalid type URL/extension field name: %s", d.orig[startPos:len(d.orig)-len(s)+1]) + } + } + + if !closed { + return Token{}, ErrUnexpectedEOF + } + + // First character cannot be '.'. Last character cannot be '.' or '/'. + size := len(name) + if size == 0 || name[0] == '.' || name[size-1] == '.' || name[size-1] == '/' { + return Token{}, d.newSyntaxError("invalid type URL/extension field name: %s", + d.orig[startPos:len(d.orig)-len(s)]) + } + + d.in = s + endPos := len(d.orig) - len(d.in) + d.consume(0) + + return Token{ + kind: Name, + attrs: uint8(TypeName), + pos: startPos, + raw: d.orig[startPos:endPos], + str: string(name), + }, nil +} + +func isTypeNameChar(b byte) bool { + return (b == '-' || b == '_' || + ('0' <= b && b <= '9') || + ('a' <= b && b <= 'z') || + ('A' <= b && b <= 'Z')) +} + +func isWhiteSpace(b byte) bool { + switch b { + case ' ', '\n', '\r', '\t': + return true + default: + return false + } +} + +// parseIdent parses an unquoted proto identifier and returns size. +// If allowNeg is true, it allows '-' to be the first character in the +// identifier. This is used when parsing literal values like -infinity, etc. +// Regular expression matches an identifier: `^[_a-zA-Z][_a-zA-Z0-9]*` +func parseIdent(input []byte, allowNeg bool) int { + var size int + + s := input + if len(s) == 0 { + return 0 + } + + if allowNeg && s[0] == '-' { + s = s[1:] + size++ + if len(s) == 0 { + return 0 + } + } + + switch { + case s[0] == '_', + 'a' <= s[0] && s[0] <= 'z', + 'A' <= s[0] && s[0] <= 'Z': + s = s[1:] + size++ + default: + return 0 + } + + for len(s) > 0 && (s[0] == '_' || + 'a' <= s[0] && s[0] <= 'z' || + 'A' <= s[0] && s[0] <= 'Z' || + '0' <= s[0] && s[0] <= '9') { + s = s[1:] + size++ + } + + if len(s) > 0 && !isDelim(s[0]) { + return 0 + } + + return size +} + +// parseScalar parses for a string, literal or number value. +func (d *Decoder) parseScalar() (Token, error) { + if d.in[0] == '"' || d.in[0] == '\'' { + return d.parseStringValue() + } + + if tok, ok := d.parseLiteralValue(); ok { + return tok, nil + } + + if tok, ok := d.parseNumberValue(); ok { + return tok, nil + } + + return Token{}, d.newSyntaxError("invalid scalar value: %s", errRegexp.Find(d.in)) +} + +// parseLiteralValue parses a literal value. A literal value is used for +// bools, special floats and enums. This function simply identifies that the +// field value is a literal. +func (d *Decoder) parseLiteralValue() (Token, bool) { + size := parseIdent(d.in, true) + if size == 0 { + return Token{}, false + } + return d.consumeToken(Scalar, size, literalValue), true +} + +// consumeToken constructs a Token for given Kind from d.in and consumes given +// size-length from it. +func (d *Decoder) consumeToken(kind Kind, size int, attrs uint8) Token { + // Important to compute raw and pos before consuming. + tok := Token{ + kind: kind, + attrs: attrs, + pos: len(d.orig) - len(d.in), + raw: d.in[:size], + } + d.consume(size) + return tok +} + +// newSyntaxError returns a syntax error with line and column information for +// current position. +func (d *Decoder) newSyntaxError(f string, x ...interface{}) error { + e := errors.New(f, x...) + line, column := d.Position(len(d.orig) - len(d.in)) + return errors.New("syntax error (line %d:%d): %v", line, column, e) +} + +// Position returns line and column number of given index of the original input. +// It will panic if index is out of range. +func (d *Decoder) Position(idx int) (line int, column int) { + b := d.orig[:idx] + line = bytes.Count(b, []byte("\n")) + 1 + if i := bytes.LastIndexByte(b, '\n'); i >= 0 { + b = b[i+1:] + } + column = utf8.RuneCount(b) + 1 // ignore multi-rune characters + return line, column +} + +func (d *Decoder) tryConsumeChar(c byte) bool { + if len(d.in) > 0 && d.in[0] == c { + d.consume(1) + return true + } + return false +} + +// consume consumes n bytes of input and any subsequent whitespace or comments. +func (d *Decoder) consume(n int) { + d.in = consume(d.in, n) + return +} + +// consume consumes n bytes of input and any subsequent whitespace or comments. +func consume(b []byte, n int) []byte { + b = b[n:] + for len(b) > 0 { + switch b[0] { + case ' ', '\n', '\r', '\t': + b = b[1:] + case '#': + if i := bytes.IndexByte(b, '\n'); i >= 0 { + b = b[i+len("\n"):] + } else { + b = nil + } + default: + return b + } + } + return b +} + +// Any sequence that looks like a non-delimiter (for error reporting). +var errRegexp = regexp.MustCompile(`^([-+._a-zA-Z0-9\/]+|.)`) + +// isDelim returns true if given byte is a delimiter character. +func isDelim(c byte) bool { + return !(c == '-' || c == '+' || c == '.' || c == '_' || + ('a' <= c && c <= 'z') || + ('A' <= c && c <= 'Z') || + ('0' <= c && c <= '9')) +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/text/decode_number.go b/vendor/google.golang.org/protobuf/internal/encoding/text/decode_number.go new file mode 100644 index 00000000..f2d90b78 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/text/decode_number.go @@ -0,0 +1,190 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package text + +// parseNumberValue parses a number from the input and returns a Token object. +func (d *Decoder) parseNumberValue() (Token, bool) { + in := d.in + num := parseNumber(in) + if num.size == 0 { + return Token{}, false + } + numAttrs := num.kind + if num.neg { + numAttrs |= isNegative + } + strSize := num.size + last := num.size - 1 + if num.kind == numFloat && (d.in[last] == 'f' || d.in[last] == 'F') { + strSize = last + } + tok := Token{ + kind: Scalar, + attrs: numberValue, + pos: len(d.orig) - len(d.in), + raw: d.in[:num.size], + str: string(d.in[:strSize]), + numAttrs: numAttrs, + } + d.consume(num.size) + return tok, true +} + +const ( + numDec uint8 = (1 << iota) / 2 + numHex + numOct + numFloat +) + +// number is the result of parsing out a valid number from parseNumber. It +// contains data for doing float or integer conversion via the strconv package +// in conjunction with the input bytes. +type number struct { + kind uint8 + neg bool + size int +} + +// parseNumber constructs a number object from given input. It allows for the +// following patterns: +// integer: ^-?([1-9][0-9]*|0[xX][0-9a-fA-F]+|0[0-7]*) +// float: ^-?((0|[1-9][0-9]*)?([.][0-9]*)?([eE][+-]?[0-9]+)?[fF]?) +// It also returns the number of parsed bytes for the given number, 0 if it is +// not a number. +func parseNumber(input []byte) number { + kind := numDec + var size int + var neg bool + + s := input + if len(s) == 0 { + return number{} + } + + // Optional - + if s[0] == '-' { + neg = true + s = s[1:] + size++ + if len(s) == 0 { + return number{} + } + } + + // C++ allows for whitespace and comments in between the negative sign and + // the rest of the number. This logic currently does not but is consistent + // with v1. + + switch { + case s[0] == '0': + if len(s) > 1 { + switch { + case s[1] == 'x' || s[1] == 'X': + // Parse as hex number. + kind = numHex + n := 2 + s = s[2:] + for len(s) > 0 && (('0' <= s[0] && s[0] <= '9') || + ('a' <= s[0] && s[0] <= 'f') || + ('A' <= s[0] && s[0] <= 'F')) { + s = s[1:] + n++ + } + if n == 2 { + return number{} + } + size += n + + case '0' <= s[1] && s[1] <= '7': + // Parse as octal number. + kind = numOct + n := 2 + s = s[2:] + for len(s) > 0 && '0' <= s[0] && s[0] <= '7' { + s = s[1:] + n++ + } + size += n + } + + if kind&(numHex|numOct) > 0 { + if len(s) > 0 && !isDelim(s[0]) { + return number{} + } + return number{kind: kind, neg: neg, size: size} + } + } + s = s[1:] + size++ + + case '1' <= s[0] && s[0] <= '9': + n := 1 + s = s[1:] + for len(s) > 0 && '0' <= s[0] && s[0] <= '9' { + s = s[1:] + n++ + } + size += n + + case s[0] == '.': + // Set kind to numFloat to signify the intent to parse as float. And + // that it needs to have other digits after '.'. + kind = numFloat + + default: + return number{} + } + + // . followed by 0 or more digits. + if len(s) > 0 && s[0] == '.' { + n := 1 + s = s[1:] + // If decimal point was before any digits, it should be followed by + // other digits. + if len(s) == 0 && kind == numFloat { + return number{} + } + for len(s) > 0 && '0' <= s[0] && s[0] <= '9' { + s = s[1:] + n++ + } + size += n + kind = numFloat + } + + // e or E followed by an optional - or + and 1 or more digits. + if len(s) >= 2 && (s[0] == 'e' || s[0] == 'E') { + kind = numFloat + s = s[1:] + n := 1 + if s[0] == '+' || s[0] == '-' { + s = s[1:] + n++ + if len(s) == 0 { + return number{} + } + } + for len(s) > 0 && '0' <= s[0] && s[0] <= '9' { + s = s[1:] + n++ + } + size += n + } + + // Optional suffix f or F for floats. + if len(s) > 0 && (s[0] == 'f' || s[0] == 'F') { + kind = numFloat + s = s[1:] + size++ + } + + // Check that next byte is a delimiter or it is at the end. + if len(s) > 0 && !isDelim(s[0]) { + return number{} + } + + return number{kind: kind, neg: neg, size: size} +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/text/decode_string.go b/vendor/google.golang.org/protobuf/internal/encoding/text/decode_string.go new file mode 100644 index 00000000..d4d34902 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/text/decode_string.go @@ -0,0 +1,161 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package text + +import ( + "bytes" + "strconv" + "strings" + "unicode" + "unicode/utf16" + "unicode/utf8" + + "google.golang.org/protobuf/internal/strs" +) + +// parseStringValue parses string field token. +// This differs from parseString since the text format allows +// multiple back-to-back string literals where they are semantically treated +// as a single large string with all values concatenated. +// +// E.g., `"foo" "bar" "baz"` => "foobarbaz" +func (d *Decoder) parseStringValue() (Token, error) { + // Note that the ending quote is sufficient to unambiguously mark the end + // of a string. Thus, the text grammar does not require intervening + // whitespace or control characters in-between strings. + // Thus, the following is valid: + // `"foo"'bar'"baz"` => "foobarbaz" + in0 := d.in + var ss []string + for len(d.in) > 0 && (d.in[0] == '"' || d.in[0] == '\'') { + s, err := d.parseString() + if err != nil { + return Token{}, err + } + ss = append(ss, s) + } + // d.in already points to the end of the value at this point. + return Token{ + kind: Scalar, + attrs: stringValue, + pos: len(d.orig) - len(in0), + raw: in0[:len(in0)-len(d.in)], + str: strings.Join(ss, ""), + }, nil +} + +// parseString parses a string value enclosed in " or '. +func (d *Decoder) parseString() (string, error) { + in := d.in + if len(in) == 0 { + return "", ErrUnexpectedEOF + } + quote := in[0] + in = in[1:] + i := indexNeedEscapeInBytes(in) + in, out := in[i:], in[:i:i] // set cap to prevent mutations + for len(in) > 0 { + switch r, n := utf8.DecodeRune(in); { + case r == utf8.RuneError && n == 1: + return "", d.newSyntaxError("invalid UTF-8 detected") + case r == 0 || r == '\n': + return "", d.newSyntaxError("invalid character %q in string", r) + case r == rune(quote): + in = in[1:] + d.consume(len(d.in) - len(in)) + return string(out), nil + case r == '\\': + if len(in) < 2 { + return "", ErrUnexpectedEOF + } + switch r := in[1]; r { + case '"', '\'', '\\', '?': + in, out = in[2:], append(out, r) + case 'a': + in, out = in[2:], append(out, '\a') + case 'b': + in, out = in[2:], append(out, '\b') + case 'n': + in, out = in[2:], append(out, '\n') + case 'r': + in, out = in[2:], append(out, '\r') + case 't': + in, out = in[2:], append(out, '\t') + case 'v': + in, out = in[2:], append(out, '\v') + case 'f': + in, out = in[2:], append(out, '\f') + case '0', '1', '2', '3', '4', '5', '6', '7': + // One, two, or three octal characters. + n := len(in[1:]) - len(bytes.TrimLeft(in[1:], "01234567")) + if n > 3 { + n = 3 + } + v, err := strconv.ParseUint(string(in[1:1+n]), 8, 8) + if err != nil { + return "", d.newSyntaxError("invalid octal escape code %q in string", in[:1+n]) + } + in, out = in[1+n:], append(out, byte(v)) + case 'x': + // One or two hexadecimal characters. + n := len(in[2:]) - len(bytes.TrimLeft(in[2:], "0123456789abcdefABCDEF")) + if n > 2 { + n = 2 + } + v, err := strconv.ParseUint(string(in[2:2+n]), 16, 8) + if err != nil { + return "", d.newSyntaxError("invalid hex escape code %q in string", in[:2+n]) + } + in, out = in[2+n:], append(out, byte(v)) + case 'u', 'U': + // Four or eight hexadecimal characters + n := 6 + if r == 'U' { + n = 10 + } + if len(in) < n { + return "", ErrUnexpectedEOF + } + v, err := strconv.ParseUint(string(in[2:n]), 16, 32) + if utf8.MaxRune < v || err != nil { + return "", d.newSyntaxError("invalid Unicode escape code %q in string", in[:n]) + } + in = in[n:] + + r := rune(v) + if utf16.IsSurrogate(r) { + if len(in) < 6 { + return "", ErrUnexpectedEOF + } + v, err := strconv.ParseUint(string(in[2:6]), 16, 16) + r = utf16.DecodeRune(r, rune(v)) + if in[0] != '\\' || in[1] != 'u' || r == unicode.ReplacementChar || err != nil { + return "", d.newSyntaxError("invalid Unicode escape code %q in string", in[:6]) + } + in = in[6:] + } + out = append(out, string(r)...) + default: + return "", d.newSyntaxError("invalid escape code %q in string", in[:2]) + } + default: + i := indexNeedEscapeInBytes(in[n:]) + in, out = in[n+i:], append(out, in[:n+i]...) + } + } + return "", ErrUnexpectedEOF +} + +// indexNeedEscapeInString returns the index of the character that needs +// escaping. If no characters need escaping, this returns the input length. +func indexNeedEscapeInBytes(b []byte) int { return indexNeedEscapeInString(strs.UnsafeString(b)) } + +// UnmarshalString returns an unescaped string given a textproto string value. +// String value needs to contain single or double quotes. This is only used by +// internal/encoding/defval package for unmarshaling bytes. +func UnmarshalString(s string) (string, error) { + d := NewDecoder([]byte(s)) + return d.parseString() +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/text/decode_token.go b/vendor/google.golang.org/protobuf/internal/encoding/text/decode_token.go new file mode 100644 index 00000000..83d2b0d5 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/text/decode_token.go @@ -0,0 +1,373 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package text + +import ( + "bytes" + "fmt" + "math" + "strconv" + "strings" + + "google.golang.org/protobuf/internal/flags" +) + +// Kind represents a token kind expressible in the textproto format. +type Kind uint8 + +// Kind values. +const ( + Invalid Kind = iota + EOF + Name // Name indicates the field name. + Scalar // Scalar are scalar values, e.g. "string", 47, ENUM_LITERAL, true. + MessageOpen + MessageClose + ListOpen + ListClose + + // comma and semi-colon are only for parsing in between values and should not be exposed. + comma + semicolon + + // bof indicates beginning of file, which is the default token + // kind at the beginning of parsing. + bof = Invalid +) + +func (t Kind) String() string { + switch t { + case Invalid: + return "" + case EOF: + return "eof" + case Scalar: + return "scalar" + case Name: + return "name" + case MessageOpen: + return "{" + case MessageClose: + return "}" + case ListOpen: + return "[" + case ListClose: + return "]" + case comma: + return "," + case semicolon: + return ";" + default: + return fmt.Sprintf("", uint8(t)) + } +} + +// NameKind represents different types of field names. +type NameKind uint8 + +// NameKind values. +const ( + IdentName NameKind = iota + 1 + TypeName + FieldNumber +) + +func (t NameKind) String() string { + switch t { + case IdentName: + return "IdentName" + case TypeName: + return "TypeName" + case FieldNumber: + return "FieldNumber" + default: + return fmt.Sprintf("", uint8(t)) + } +} + +// Bit mask in Token.attrs to indicate if a Name token is followed by the +// separator char ':'. The field name separator char is optional for message +// field or repeated message field, but required for all other types. Decoder +// simply indicates whether a Name token is followed by separator or not. It is +// up to the prototext package to validate. +const hasSeparator = 1 << 7 + +// Scalar value types. +const ( + numberValue = iota + 1 + stringValue + literalValue +) + +// Bit mask in Token.numAttrs to indicate that the number is a negative. +const isNegative = 1 << 7 + +// Token provides a parsed token kind and value. Values are provided by the +// different accessor methods. +type Token struct { + // Kind of the Token object. + kind Kind + // attrs contains metadata for the following Kinds: + // Name: hasSeparator bit and one of NameKind. + // Scalar: one of numberValue, stringValue, literalValue. + attrs uint8 + // numAttrs contains metadata for numberValue: + // - highest bit is whether negative or positive. + // - lower bits indicate one of numDec, numHex, numOct, numFloat. + numAttrs uint8 + // pos provides the position of the token in the original input. + pos int + // raw bytes of the serialized token. + // This is a subslice into the original input. + raw []byte + // str contains parsed string for the following: + // - stringValue of Scalar kind + // - numberValue of Scalar kind + // - TypeName of Name kind + str string +} + +// Kind returns the token kind. +func (t Token) Kind() Kind { + return t.kind +} + +// RawString returns the read value in string. +func (t Token) RawString() string { + return string(t.raw) +} + +// Pos returns the token position from the input. +func (t Token) Pos() int { + return t.pos +} + +// NameKind returns IdentName, TypeName or FieldNumber. +// It panics if type is not Name. +func (t Token) NameKind() NameKind { + if t.kind == Name { + return NameKind(t.attrs &^ hasSeparator) + } + panic(fmt.Sprintf("Token is not a Name type: %s", t.kind)) +} + +// HasSeparator returns true if the field name is followed by the separator char +// ':', else false. It panics if type is not Name. +func (t Token) HasSeparator() bool { + if t.kind == Name { + return t.attrs&hasSeparator != 0 + } + panic(fmt.Sprintf("Token is not a Name type: %s", t.kind)) +} + +// IdentName returns the value for IdentName type. +func (t Token) IdentName() string { + if t.kind == Name && t.attrs&uint8(IdentName) != 0 { + return string(t.raw) + } + panic(fmt.Sprintf("Token is not an IdentName: %s:%s", t.kind, NameKind(t.attrs&^hasSeparator))) +} + +// TypeName returns the value for TypeName type. +func (t Token) TypeName() string { + if t.kind == Name && t.attrs&uint8(TypeName) != 0 { + return t.str + } + panic(fmt.Sprintf("Token is not a TypeName: %s:%s", t.kind, NameKind(t.attrs&^hasSeparator))) +} + +// FieldNumber returns the value for FieldNumber type. It returns a +// non-negative int32 value. Caller will still need to validate for the correct +// field number range. +func (t Token) FieldNumber() int32 { + if t.kind != Name || t.attrs&uint8(FieldNumber) == 0 { + panic(fmt.Sprintf("Token is not a FieldNumber: %s:%s", t.kind, NameKind(t.attrs&^hasSeparator))) + } + // Following should not return an error as it had already been called right + // before this Token was constructed. + num, _ := strconv.ParseInt(string(t.raw), 10, 32) + return int32(num) +} + +// String returns the string value for a Scalar type. +func (t Token) String() (string, bool) { + if t.kind != Scalar || t.attrs != stringValue { + return "", false + } + return t.str, true +} + +// Enum returns the literal value for a Scalar type for use as enum literals. +func (t Token) Enum() (string, bool) { + if t.kind != Scalar || t.attrs != literalValue || (len(t.raw) > 0 && t.raw[0] == '-') { + return "", false + } + return string(t.raw), true +} + +// Bool returns the bool value for a Scalar type. +func (t Token) Bool() (bool, bool) { + if t.kind != Scalar { + return false, false + } + switch t.attrs { + case literalValue: + if b, ok := boolLits[string(t.raw)]; ok { + return b, true + } + case numberValue: + // Unsigned integer representation of 0 or 1 is permitted: 00, 0x0, 01, + // 0x1, etc. + n, err := strconv.ParseUint(t.str, 0, 64) + if err == nil { + switch n { + case 0: + return false, true + case 1: + return true, true + } + } + } + return false, false +} + +// These exact boolean literals are the ones supported in C++. +var boolLits = map[string]bool{ + "t": true, + "true": true, + "True": true, + "f": false, + "false": false, + "False": false, +} + +// Uint64 returns the uint64 value for a Scalar type. +func (t Token) Uint64() (uint64, bool) { + if t.kind != Scalar || t.attrs != numberValue || + t.numAttrs&isNegative > 0 || t.numAttrs&numFloat > 0 { + return 0, false + } + n, err := strconv.ParseUint(t.str, 0, 64) + if err != nil { + return 0, false + } + return n, true +} + +// Uint32 returns the uint32 value for a Scalar type. +func (t Token) Uint32() (uint32, bool) { + if t.kind != Scalar || t.attrs != numberValue || + t.numAttrs&isNegative > 0 || t.numAttrs&numFloat > 0 { + return 0, false + } + n, err := strconv.ParseUint(t.str, 0, 32) + if err != nil { + return 0, false + } + return uint32(n), true +} + +// Int64 returns the int64 value for a Scalar type. +func (t Token) Int64() (int64, bool) { + if t.kind != Scalar || t.attrs != numberValue || t.numAttrs&numFloat > 0 { + return 0, false + } + if n, err := strconv.ParseInt(t.str, 0, 64); err == nil { + return n, true + } + // C++ accepts large positive hex numbers as negative values. + // This feature is here for proto1 backwards compatibility purposes. + if flags.ProtoLegacy && (t.numAttrs == numHex) { + if n, err := strconv.ParseUint(t.str, 0, 64); err == nil { + return int64(n), true + } + } + return 0, false +} + +// Int32 returns the int32 value for a Scalar type. +func (t Token) Int32() (int32, bool) { + if t.kind != Scalar || t.attrs != numberValue || t.numAttrs&numFloat > 0 { + return 0, false + } + if n, err := strconv.ParseInt(t.str, 0, 32); err == nil { + return int32(n), true + } + // C++ accepts large positive hex numbers as negative values. + // This feature is here for proto1 backwards compatibility purposes. + if flags.ProtoLegacy && (t.numAttrs == numHex) { + if n, err := strconv.ParseUint(t.str, 0, 32); err == nil { + return int32(n), true + } + } + return 0, false +} + +// Float64 returns the float64 value for a Scalar type. +func (t Token) Float64() (float64, bool) { + if t.kind != Scalar { + return 0, false + } + switch t.attrs { + case literalValue: + if f, ok := floatLits[strings.ToLower(string(t.raw))]; ok { + return f, true + } + case numberValue: + n, err := strconv.ParseFloat(t.str, 64) + if err == nil { + return n, true + } + nerr := err.(*strconv.NumError) + if nerr.Err == strconv.ErrRange { + return n, true + } + } + return 0, false +} + +// Float32 returns the float32 value for a Scalar type. +func (t Token) Float32() (float32, bool) { + if t.kind != Scalar { + return 0, false + } + switch t.attrs { + case literalValue: + if f, ok := floatLits[strings.ToLower(string(t.raw))]; ok { + return float32(f), true + } + case numberValue: + n, err := strconv.ParseFloat(t.str, 64) + if err == nil { + // Overflows are treated as (-)infinity. + return float32(n), true + } + nerr := err.(*strconv.NumError) + if nerr.Err == strconv.ErrRange { + return float32(n), true + } + } + return 0, false +} + +// These are the supported float literals which C++ permits case-insensitive +// variants of these. +var floatLits = map[string]float64{ + "nan": math.NaN(), + "inf": math.Inf(1), + "infinity": math.Inf(1), + "-inf": math.Inf(-1), + "-infinity": math.Inf(-1), +} + +// TokenEquals returns true if given Tokens are equal, else false. +func TokenEquals(x, y Token) bool { + return x.kind == y.kind && + x.attrs == y.attrs && + x.numAttrs == y.numAttrs && + x.pos == y.pos && + bytes.Equal(x.raw, y.raw) && + x.str == y.str +} diff --git a/vendor/google.golang.org/protobuf/internal/encoding/text/doc.go b/vendor/google.golang.org/protobuf/internal/encoding/text/doc.go new file mode 100644 index 00000000..0ce8d6fb --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/text/doc.go @@ -0,0 +1,29 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package text implements the text format for protocol buffers. +// This package has no semantic understanding for protocol buffers and is only +// a parser and composer for the format. +// +// There is no formal specification for the protobuf text format, as such the +// C++ implementation (see google::protobuf::TextFormat) is the reference +// implementation of the text format. +// +// This package is neither a superset nor a subset of the C++ implementation. +// This implementation permits a more liberal grammar in some cases to be +// backwards compatible with the historical Go implementation. +// Future parsings unique to Go should not be added. +// Some grammars allowed by the C++ implementation are deliberately +// not implemented here because they are considered a bug by the protobuf team +// and should not be replicated. +// +// The Go implementation should implement a sufficient amount of the C++ +// grammar such that the default text serialization by C++ can be parsed by Go. +// However, just because the C++ parser accepts some input does not mean that +// the Go implementation should as well. +// +// The text format is almost a superset of JSON except: +// * message keys are not quoted strings, but identifiers +// * the top-level value must be a message without the delimiters +package text diff --git a/vendor/google.golang.org/protobuf/internal/encoding/text/encode.go b/vendor/google.golang.org/protobuf/internal/encoding/text/encode.go new file mode 100644 index 00000000..c4ba1c59 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/encoding/text/encode.go @@ -0,0 +1,267 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package text + +import ( + "math" + "math/bits" + "strconv" + "strings" + "unicode/utf8" + + "google.golang.org/protobuf/internal/detrand" + "google.golang.org/protobuf/internal/errors" +) + +// encType represents an encoding type. +type encType uint8 + +const ( + _ encType = (1 << iota) / 2 + name + scalar + messageOpen + messageClose +) + +// Encoder provides methods to write out textproto constructs and values. The user is +// responsible for producing valid sequences of constructs and values. +type Encoder struct { + encoderState + + indent string + newline string // set to "\n" if len(indent) > 0 + delims [2]byte + outputASCII bool +} + +type encoderState struct { + lastType encType + indents []byte + out []byte +} + +// NewEncoder returns an Encoder. +// +// If indent is a non-empty string, it causes every entry in a List or Message +// to be preceded by the indent and trailed by a newline. +// +// If delims is not the zero value, it controls the delimiter characters used +// for messages (e.g., "{}" vs "<>"). +// +// If outputASCII is true, strings will be serialized in such a way that +// multi-byte UTF-8 sequences are escaped. This property ensures that the +// overall output is ASCII (as opposed to UTF-8). +func NewEncoder(indent string, delims [2]byte, outputASCII bool) (*Encoder, error) { + e := &Encoder{} + if len(indent) > 0 { + if strings.Trim(indent, " \t") != "" { + return nil, errors.New("indent may only be composed of space and tab characters") + } + e.indent = indent + e.newline = "\n" + } + switch delims { + case [2]byte{0, 0}: + e.delims = [2]byte{'{', '}'} + case [2]byte{'{', '}'}, [2]byte{'<', '>'}: + e.delims = delims + default: + return nil, errors.New("delimiters may only be \"{}\" or \"<>\"") + } + e.outputASCII = outputASCII + + return e, nil +} + +// Bytes returns the content of the written bytes. +func (e *Encoder) Bytes() []byte { + return e.out +} + +// StartMessage writes out the '{' or '<' symbol. +func (e *Encoder) StartMessage() { + e.prepareNext(messageOpen) + e.out = append(e.out, e.delims[0]) +} + +// EndMessage writes out the '}' or '>' symbol. +func (e *Encoder) EndMessage() { + e.prepareNext(messageClose) + e.out = append(e.out, e.delims[1]) +} + +// WriteName writes out the field name and the separator ':'. +func (e *Encoder) WriteName(s string) { + e.prepareNext(name) + e.out = append(e.out, s...) + e.out = append(e.out, ':') +} + +// WriteBool writes out the given boolean value. +func (e *Encoder) WriteBool(b bool) { + if b { + e.WriteLiteral("true") + } else { + e.WriteLiteral("false") + } +} + +// WriteString writes out the given string value. +func (e *Encoder) WriteString(s string) { + e.prepareNext(scalar) + e.out = appendString(e.out, s, e.outputASCII) +} + +func appendString(out []byte, in string, outputASCII bool) []byte { + out = append(out, '"') + i := indexNeedEscapeInString(in) + in, out = in[i:], append(out, in[:i]...) + for len(in) > 0 { + switch r, n := utf8.DecodeRuneInString(in); { + case r == utf8.RuneError && n == 1: + // We do not report invalid UTF-8 because strings in the text format + // are used to represent both the proto string and bytes type. + r = rune(in[0]) + fallthrough + case r < ' ' || r == '"' || r == '\\': + out = append(out, '\\') + switch r { + case '"', '\\': + out = append(out, byte(r)) + case '\n': + out = append(out, 'n') + case '\r': + out = append(out, 'r') + case '\t': + out = append(out, 't') + default: + out = append(out, 'x') + out = append(out, "00"[1+(bits.Len32(uint32(r))-1)/4:]...) + out = strconv.AppendUint(out, uint64(r), 16) + } + in = in[n:] + case outputASCII && r >= utf8.RuneSelf: + out = append(out, '\\') + if r <= math.MaxUint16 { + out = append(out, 'u') + out = append(out, "0000"[1+(bits.Len32(uint32(r))-1)/4:]...) + out = strconv.AppendUint(out, uint64(r), 16) + } else { + out = append(out, 'U') + out = append(out, "00000000"[1+(bits.Len32(uint32(r))-1)/4:]...) + out = strconv.AppendUint(out, uint64(r), 16) + } + in = in[n:] + default: + i := indexNeedEscapeInString(in[n:]) + in, out = in[n+i:], append(out, in[:n+i]...) + } + } + out = append(out, '"') + return out +} + +// indexNeedEscapeInString returns the index of the character that needs +// escaping. If no characters need escaping, this returns the input length. +func indexNeedEscapeInString(s string) int { + for i := 0; i < len(s); i++ { + if c := s[i]; c < ' ' || c == '"' || c == '\'' || c == '\\' || c >= utf8.RuneSelf { + return i + } + } + return len(s) +} + +// WriteFloat writes out the given float value for given bitSize. +func (e *Encoder) WriteFloat(n float64, bitSize int) { + e.prepareNext(scalar) + e.out = appendFloat(e.out, n, bitSize) +} + +func appendFloat(out []byte, n float64, bitSize int) []byte { + switch { + case math.IsNaN(n): + return append(out, "nan"...) + case math.IsInf(n, +1): + return append(out, "inf"...) + case math.IsInf(n, -1): + return append(out, "-inf"...) + default: + return strconv.AppendFloat(out, n, 'g', -1, bitSize) + } +} + +// WriteInt writes out the given signed integer value. +func (e *Encoder) WriteInt(n int64) { + e.prepareNext(scalar) + e.out = append(e.out, strconv.FormatInt(n, 10)...) +} + +// WriteUint writes out the given unsigned integer value. +func (e *Encoder) WriteUint(n uint64) { + e.prepareNext(scalar) + e.out = append(e.out, strconv.FormatUint(n, 10)...) +} + +// WriteLiteral writes out the given string as a literal value without quotes. +// This is used for writing enum literal strings. +func (e *Encoder) WriteLiteral(s string) { + e.prepareNext(scalar) + e.out = append(e.out, s...) +} + +// prepareNext adds possible space and indentation for the next value based +// on last encType and indent option. It also updates e.lastType to next. +func (e *Encoder) prepareNext(next encType) { + defer func() { + e.lastType = next + }() + + // Single line. + if len(e.indent) == 0 { + // Add space after each field before the next one. + if e.lastType&(scalar|messageClose) != 0 && next == name { + e.out = append(e.out, ' ') + // Add a random extra space to make output unstable. + if detrand.Bool() { + e.out = append(e.out, ' ') + } + } + return + } + + // Multi-line. + switch { + case e.lastType == name: + e.out = append(e.out, ' ') + // Add a random extra space after name: to make output unstable. + if detrand.Bool() { + e.out = append(e.out, ' ') + } + + case e.lastType == messageOpen && next != messageClose: + e.indents = append(e.indents, e.indent...) + e.out = append(e.out, '\n') + e.out = append(e.out, e.indents...) + + case e.lastType&(scalar|messageClose) != 0: + if next == messageClose { + e.indents = e.indents[:len(e.indents)-len(e.indent)] + } + e.out = append(e.out, '\n') + e.out = append(e.out, e.indents...) + } +} + +// Snapshot returns the current snapshot for use in Reset. +func (e *Encoder) Snapshot() encoderState { + return e.encoderState +} + +// Reset resets the Encoder to the given encoderState from a Snapshot. +func (e *Encoder) Reset(es encoderState) { + e.encoderState = es +} diff --git a/vendor/google.golang.org/protobuf/internal/errors/errors.go b/vendor/google.golang.org/protobuf/internal/errors/errors.go new file mode 100644 index 00000000..20c17b35 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/errors/errors.go @@ -0,0 +1,89 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package errors implements functions to manipulate errors. +package errors + +import ( + "errors" + "fmt" + + "google.golang.org/protobuf/internal/detrand" +) + +// Error is a sentinel matching all errors produced by this package. +var Error = errors.New("protobuf error") + +// New formats a string according to the format specifier and arguments and +// returns an error that has a "proto" prefix. +func New(f string, x ...interface{}) error { + return &prefixError{s: format(f, x...)} +} + +type prefixError struct{ s string } + +var prefix = func() string { + // Deliberately introduce instability into the error message string to + // discourage users from performing error string comparisons. + if detrand.Bool() { + return "proto: " // use non-breaking spaces (U+00a0) + } else { + return "proto: " // use regular spaces (U+0020) + } +}() + +func (e *prefixError) Error() string { + return prefix + e.s +} + +func (e *prefixError) Unwrap() error { + return Error +} + +// Wrap returns an error that has a "proto" prefix, the formatted string described +// by the format specifier and arguments, and a suffix of err. The error wraps err. +func Wrap(err error, f string, x ...interface{}) error { + return &wrapError{ + s: format(f, x...), + err: err, + } +} + +type wrapError struct { + s string + err error +} + +func (e *wrapError) Error() string { + return format("%v%v: %v", prefix, e.s, e.err) +} + +func (e *wrapError) Unwrap() error { + return e.err +} + +func (e *wrapError) Is(target error) bool { + return target == Error +} + +func format(f string, x ...interface{}) string { + // avoid "proto: " prefix when chaining + for i := 0; i < len(x); i++ { + switch e := x[i].(type) { + case *prefixError: + x[i] = e.s + case *wrapError: + x[i] = format("%v: %v", e.s, e.err) + } + } + return fmt.Sprintf(f, x...) +} + +func InvalidUTF8(name string) error { + return New("field %v contains invalid UTF-8", name) +} + +func RequiredNotSet(name string) error { + return New("required field %v not set", name) +} diff --git a/vendor/google.golang.org/protobuf/internal/errors/is_go112.go b/vendor/google.golang.org/protobuf/internal/errors/is_go112.go new file mode 100644 index 00000000..f90e909b --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/errors/is_go112.go @@ -0,0 +1,39 @@ +// Copyright 2020 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !go1.13 + +package errors + +import "reflect" + +// Is is a copy of Go 1.13's errors.Is for use with older Go versions. +func Is(err, target error) bool { + if target == nil { + return err == target + } + + isComparable := reflect.TypeOf(target).Comparable() + for { + if isComparable && err == target { + return true + } + if x, ok := err.(interface{ Is(error) bool }); ok && x.Is(target) { + return true + } + if err = unwrap(err); err == nil { + return false + } + } +} + +func unwrap(err error) error { + u, ok := err.(interface { + Unwrap() error + }) + if !ok { + return nil + } + return u.Unwrap() +} diff --git a/vendor/google.golang.org/protobuf/internal/errors/is_go113.go b/vendor/google.golang.org/protobuf/internal/errors/is_go113.go new file mode 100644 index 00000000..dc05f419 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/errors/is_go113.go @@ -0,0 +1,12 @@ +// Copyright 2020 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build go1.13 + +package errors + +import "errors" + +// Is is errors.Is. +func Is(err, target error) bool { return errors.Is(err, target) } diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/any_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/any_gen.go new file mode 100644 index 00000000..74c5fef2 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/any_gen.go @@ -0,0 +1,13 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.Any. +const ( + Any_TypeUrl = 1 // optional string + Any_Value = 2 // optional bytes +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/api_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/api_gen.go new file mode 100644 index 00000000..9a6b5f29 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/api_gen.go @@ -0,0 +1,35 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.Api. +const ( + Api_Name = 1 // optional string + Api_Methods = 2 // repeated google.protobuf.Method + Api_Options = 3 // repeated google.protobuf.Option + Api_Version = 4 // optional string + Api_SourceContext = 5 // optional google.protobuf.SourceContext + Api_Mixins = 6 // repeated google.protobuf.Mixin + Api_Syntax = 7 // optional google.protobuf.Syntax +) + +// Field numbers for google.protobuf.Method. +const ( + Method_Name = 1 // optional string + Method_RequestTypeUrl = 2 // optional string + Method_RequestStreaming = 3 // optional bool + Method_ResponseTypeUrl = 4 // optional string + Method_ResponseStreaming = 5 // optional bool + Method_Options = 6 // repeated google.protobuf.Option + Method_Syntax = 7 // optional google.protobuf.Syntax +) + +// Field numbers for google.protobuf.Mixin. +const ( + Mixin_Name = 1 // optional string + Mixin_Root = 2 // optional string +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/descriptor_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/descriptor_gen.go new file mode 100644 index 00000000..6e37b59e --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/descriptor_gen.go @@ -0,0 +1,240 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.FileDescriptorSet. +const ( + FileDescriptorSet_File = 1 // repeated google.protobuf.FileDescriptorProto +) + +// Field numbers for google.protobuf.FileDescriptorProto. +const ( + FileDescriptorProto_Name = 1 // optional string + FileDescriptorProto_Package = 2 // optional string + FileDescriptorProto_Dependency = 3 // repeated string + FileDescriptorProto_PublicDependency = 10 // repeated int32 + FileDescriptorProto_WeakDependency = 11 // repeated int32 + FileDescriptorProto_MessageType = 4 // repeated google.protobuf.DescriptorProto + FileDescriptorProto_EnumType = 5 // repeated google.protobuf.EnumDescriptorProto + FileDescriptorProto_Service = 6 // repeated google.protobuf.ServiceDescriptorProto + FileDescriptorProto_Extension = 7 // repeated google.protobuf.FieldDescriptorProto + FileDescriptorProto_Options = 8 // optional google.protobuf.FileOptions + FileDescriptorProto_SourceCodeInfo = 9 // optional google.protobuf.SourceCodeInfo + FileDescriptorProto_Syntax = 12 // optional string +) + +// Field numbers for google.protobuf.DescriptorProto. +const ( + DescriptorProto_Name = 1 // optional string + DescriptorProto_Field = 2 // repeated google.protobuf.FieldDescriptorProto + DescriptorProto_Extension = 6 // repeated google.protobuf.FieldDescriptorProto + DescriptorProto_NestedType = 3 // repeated google.protobuf.DescriptorProto + DescriptorProto_EnumType = 4 // repeated google.protobuf.EnumDescriptorProto + DescriptorProto_ExtensionRange = 5 // repeated google.protobuf.DescriptorProto.ExtensionRange + DescriptorProto_OneofDecl = 8 // repeated google.protobuf.OneofDescriptorProto + DescriptorProto_Options = 7 // optional google.protobuf.MessageOptions + DescriptorProto_ReservedRange = 9 // repeated google.protobuf.DescriptorProto.ReservedRange + DescriptorProto_ReservedName = 10 // repeated string +) + +// Field numbers for google.protobuf.DescriptorProto.ExtensionRange. +const ( + DescriptorProto_ExtensionRange_Start = 1 // optional int32 + DescriptorProto_ExtensionRange_End = 2 // optional int32 + DescriptorProto_ExtensionRange_Options = 3 // optional google.protobuf.ExtensionRangeOptions +) + +// Field numbers for google.protobuf.DescriptorProto.ReservedRange. +const ( + DescriptorProto_ReservedRange_Start = 1 // optional int32 + DescriptorProto_ReservedRange_End = 2 // optional int32 +) + +// Field numbers for google.protobuf.ExtensionRangeOptions. +const ( + ExtensionRangeOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.FieldDescriptorProto. +const ( + FieldDescriptorProto_Name = 1 // optional string + FieldDescriptorProto_Number = 3 // optional int32 + FieldDescriptorProto_Label = 4 // optional google.protobuf.FieldDescriptorProto.Label + FieldDescriptorProto_Type = 5 // optional google.protobuf.FieldDescriptorProto.Type + FieldDescriptorProto_TypeName = 6 // optional string + FieldDescriptorProto_Extendee = 2 // optional string + FieldDescriptorProto_DefaultValue = 7 // optional string + FieldDescriptorProto_OneofIndex = 9 // optional int32 + FieldDescriptorProto_JsonName = 10 // optional string + FieldDescriptorProto_Options = 8 // optional google.protobuf.FieldOptions + FieldDescriptorProto_Proto3Optional = 17 // optional bool +) + +// Field numbers for google.protobuf.OneofDescriptorProto. +const ( + OneofDescriptorProto_Name = 1 // optional string + OneofDescriptorProto_Options = 2 // optional google.protobuf.OneofOptions +) + +// Field numbers for google.protobuf.EnumDescriptorProto. +const ( + EnumDescriptorProto_Name = 1 // optional string + EnumDescriptorProto_Value = 2 // repeated google.protobuf.EnumValueDescriptorProto + EnumDescriptorProto_Options = 3 // optional google.protobuf.EnumOptions + EnumDescriptorProto_ReservedRange = 4 // repeated google.protobuf.EnumDescriptorProto.EnumReservedRange + EnumDescriptorProto_ReservedName = 5 // repeated string +) + +// Field numbers for google.protobuf.EnumDescriptorProto.EnumReservedRange. +const ( + EnumDescriptorProto_EnumReservedRange_Start = 1 // optional int32 + EnumDescriptorProto_EnumReservedRange_End = 2 // optional int32 +) + +// Field numbers for google.protobuf.EnumValueDescriptorProto. +const ( + EnumValueDescriptorProto_Name = 1 // optional string + EnumValueDescriptorProto_Number = 2 // optional int32 + EnumValueDescriptorProto_Options = 3 // optional google.protobuf.EnumValueOptions +) + +// Field numbers for google.protobuf.ServiceDescriptorProto. +const ( + ServiceDescriptorProto_Name = 1 // optional string + ServiceDescriptorProto_Method = 2 // repeated google.protobuf.MethodDescriptorProto + ServiceDescriptorProto_Options = 3 // optional google.protobuf.ServiceOptions +) + +// Field numbers for google.protobuf.MethodDescriptorProto. +const ( + MethodDescriptorProto_Name = 1 // optional string + MethodDescriptorProto_InputType = 2 // optional string + MethodDescriptorProto_OutputType = 3 // optional string + MethodDescriptorProto_Options = 4 // optional google.protobuf.MethodOptions + MethodDescriptorProto_ClientStreaming = 5 // optional bool + MethodDescriptorProto_ServerStreaming = 6 // optional bool +) + +// Field numbers for google.protobuf.FileOptions. +const ( + FileOptions_JavaPackage = 1 // optional string + FileOptions_JavaOuterClassname = 8 // optional string + FileOptions_JavaMultipleFiles = 10 // optional bool + FileOptions_JavaGenerateEqualsAndHash = 20 // optional bool + FileOptions_JavaStringCheckUtf8 = 27 // optional bool + FileOptions_OptimizeFor = 9 // optional google.protobuf.FileOptions.OptimizeMode + FileOptions_GoPackage = 11 // optional string + FileOptions_CcGenericServices = 16 // optional bool + FileOptions_JavaGenericServices = 17 // optional bool + FileOptions_PyGenericServices = 18 // optional bool + FileOptions_PhpGenericServices = 42 // optional bool + FileOptions_Deprecated = 23 // optional bool + FileOptions_CcEnableArenas = 31 // optional bool + FileOptions_ObjcClassPrefix = 36 // optional string + FileOptions_CsharpNamespace = 37 // optional string + FileOptions_SwiftPrefix = 39 // optional string + FileOptions_PhpClassPrefix = 40 // optional string + FileOptions_PhpNamespace = 41 // optional string + FileOptions_PhpMetadataNamespace = 44 // optional string + FileOptions_RubyPackage = 45 // optional string + FileOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.MessageOptions. +const ( + MessageOptions_MessageSetWireFormat = 1 // optional bool + MessageOptions_NoStandardDescriptorAccessor = 2 // optional bool + MessageOptions_Deprecated = 3 // optional bool + MessageOptions_MapEntry = 7 // optional bool + MessageOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.FieldOptions. +const ( + FieldOptions_Ctype = 1 // optional google.protobuf.FieldOptions.CType + FieldOptions_Packed = 2 // optional bool + FieldOptions_Jstype = 6 // optional google.protobuf.FieldOptions.JSType + FieldOptions_Lazy = 5 // optional bool + FieldOptions_Deprecated = 3 // optional bool + FieldOptions_Weak = 10 // optional bool + FieldOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.OneofOptions. +const ( + OneofOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.EnumOptions. +const ( + EnumOptions_AllowAlias = 2 // optional bool + EnumOptions_Deprecated = 3 // optional bool + EnumOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.EnumValueOptions. +const ( + EnumValueOptions_Deprecated = 1 // optional bool + EnumValueOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.ServiceOptions. +const ( + ServiceOptions_Deprecated = 33 // optional bool + ServiceOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.MethodOptions. +const ( + MethodOptions_Deprecated = 33 // optional bool + MethodOptions_IdempotencyLevel = 34 // optional google.protobuf.MethodOptions.IdempotencyLevel + MethodOptions_UninterpretedOption = 999 // repeated google.protobuf.UninterpretedOption +) + +// Field numbers for google.protobuf.UninterpretedOption. +const ( + UninterpretedOption_Name = 2 // repeated google.protobuf.UninterpretedOption.NamePart + UninterpretedOption_IdentifierValue = 3 // optional string + UninterpretedOption_PositiveIntValue = 4 // optional uint64 + UninterpretedOption_NegativeIntValue = 5 // optional int64 + UninterpretedOption_DoubleValue = 6 // optional double + UninterpretedOption_StringValue = 7 // optional bytes + UninterpretedOption_AggregateValue = 8 // optional string +) + +// Field numbers for google.protobuf.UninterpretedOption.NamePart. +const ( + UninterpretedOption_NamePart_NamePart = 1 // required string + UninterpretedOption_NamePart_IsExtension = 2 // required bool +) + +// Field numbers for google.protobuf.SourceCodeInfo. +const ( + SourceCodeInfo_Location = 1 // repeated google.protobuf.SourceCodeInfo.Location +) + +// Field numbers for google.protobuf.SourceCodeInfo.Location. +const ( + SourceCodeInfo_Location_Path = 1 // repeated int32 + SourceCodeInfo_Location_Span = 2 // repeated int32 + SourceCodeInfo_Location_LeadingComments = 3 // optional string + SourceCodeInfo_Location_TrailingComments = 4 // optional string + SourceCodeInfo_Location_LeadingDetachedComments = 6 // repeated string +) + +// Field numbers for google.protobuf.GeneratedCodeInfo. +const ( + GeneratedCodeInfo_Annotation = 1 // repeated google.protobuf.GeneratedCodeInfo.Annotation +) + +// Field numbers for google.protobuf.GeneratedCodeInfo.Annotation. +const ( + GeneratedCodeInfo_Annotation_Path = 1 // repeated int32 + GeneratedCodeInfo_Annotation_SourceFile = 2 // optional string + GeneratedCodeInfo_Annotation_Begin = 3 // optional int32 + GeneratedCodeInfo_Annotation_End = 4 // optional int32 +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/doc.go b/vendor/google.golang.org/protobuf/internal/fieldnum/doc.go new file mode 100644 index 00000000..e5978859 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/doc.go @@ -0,0 +1,7 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package fieldnum contains constants for field numbers of fields in messages +// declared in descriptor.proto and any of the well-known types. +package fieldnum diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/duration_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/duration_gen.go new file mode 100644 index 00000000..8816c735 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/duration_gen.go @@ -0,0 +1,13 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.Duration. +const ( + Duration_Seconds = 1 // optional int64 + Duration_Nanos = 2 // optional int32 +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/empty_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/empty_gen.go new file mode 100644 index 00000000..b5130a6d --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/empty_gen.go @@ -0,0 +1,10 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.Empty. +const () diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/field_mask_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/field_mask_gen.go new file mode 100644 index 00000000..7e3bfa27 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/field_mask_gen.go @@ -0,0 +1,12 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.FieldMask. +const ( + FieldMask_Paths = 1 // repeated string +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/source_context_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/source_context_gen.go new file mode 100644 index 00000000..241972b1 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/source_context_gen.go @@ -0,0 +1,12 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.SourceContext. +const ( + SourceContext_FileName = 1 // optional string +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/struct_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/struct_gen.go new file mode 100644 index 00000000..c460aab4 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/struct_gen.go @@ -0,0 +1,33 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.Struct. +const ( + Struct_Fields = 1 // repeated google.protobuf.Struct.FieldsEntry +) + +// Field numbers for google.protobuf.Struct.FieldsEntry. +const ( + Struct_FieldsEntry_Key = 1 // optional string + Struct_FieldsEntry_Value = 2 // optional google.protobuf.Value +) + +// Field numbers for google.protobuf.Value. +const ( + Value_NullValue = 1 // optional google.protobuf.NullValue + Value_NumberValue = 2 // optional double + Value_StringValue = 3 // optional string + Value_BoolValue = 4 // optional bool + Value_StructValue = 5 // optional google.protobuf.Struct + Value_ListValue = 6 // optional google.protobuf.ListValue +) + +// Field numbers for google.protobuf.ListValue. +const ( + ListValue_Values = 1 // repeated google.protobuf.Value +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/timestamp_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/timestamp_gen.go new file mode 100644 index 00000000..b4346fba --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/timestamp_gen.go @@ -0,0 +1,13 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.Timestamp. +const ( + Timestamp_Seconds = 1 // optional int64 + Timestamp_Nanos = 2 // optional int32 +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/type_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/type_gen.go new file mode 100644 index 00000000..b392e959 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/type_gen.go @@ -0,0 +1,53 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.Type. +const ( + Type_Name = 1 // optional string + Type_Fields = 2 // repeated google.protobuf.Field + Type_Oneofs = 3 // repeated string + Type_Options = 4 // repeated google.protobuf.Option + Type_SourceContext = 5 // optional google.protobuf.SourceContext + Type_Syntax = 6 // optional google.protobuf.Syntax +) + +// Field numbers for google.protobuf.Field. +const ( + Field_Kind = 1 // optional google.protobuf.Field.Kind + Field_Cardinality = 2 // optional google.protobuf.Field.Cardinality + Field_Number = 3 // optional int32 + Field_Name = 4 // optional string + Field_TypeUrl = 6 // optional string + Field_OneofIndex = 7 // optional int32 + Field_Packed = 8 // optional bool + Field_Options = 9 // repeated google.protobuf.Option + Field_JsonName = 10 // optional string + Field_DefaultValue = 11 // optional string +) + +// Field numbers for google.protobuf.Enum. +const ( + Enum_Name = 1 // optional string + Enum_Enumvalue = 2 // repeated google.protobuf.EnumValue + Enum_Options = 3 // repeated google.protobuf.Option + Enum_SourceContext = 4 // optional google.protobuf.SourceContext + Enum_Syntax = 5 // optional google.protobuf.Syntax +) + +// Field numbers for google.protobuf.EnumValue. +const ( + EnumValue_Name = 1 // optional string + EnumValue_Number = 2 // optional int32 + EnumValue_Options = 3 // repeated google.protobuf.Option +) + +// Field numbers for google.protobuf.Option. +const ( + Option_Name = 1 // optional string + Option_Value = 2 // optional google.protobuf.Any +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldnum/wrappers_gen.go b/vendor/google.golang.org/protobuf/internal/fieldnum/wrappers_gen.go new file mode 100644 index 00000000..42f846a9 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldnum/wrappers_gen.go @@ -0,0 +1,52 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-protos. DO NOT EDIT. + +package fieldnum + +// Field numbers for google.protobuf.DoubleValue. +const ( + DoubleValue_Value = 1 // optional double +) + +// Field numbers for google.protobuf.FloatValue. +const ( + FloatValue_Value = 1 // optional float +) + +// Field numbers for google.protobuf.Int64Value. +const ( + Int64Value_Value = 1 // optional int64 +) + +// Field numbers for google.protobuf.UInt64Value. +const ( + UInt64Value_Value = 1 // optional uint64 +) + +// Field numbers for google.protobuf.Int32Value. +const ( + Int32Value_Value = 1 // optional int32 +) + +// Field numbers for google.protobuf.UInt32Value. +const ( + UInt32Value_Value = 1 // optional uint32 +) + +// Field numbers for google.protobuf.BoolValue. +const ( + BoolValue_Value = 1 // optional bool +) + +// Field numbers for google.protobuf.StringValue. +const ( + StringValue_Value = 1 // optional string +) + +// Field numbers for google.protobuf.BytesValue. +const ( + BytesValue_Value = 1 // optional bytes +) diff --git a/vendor/google.golang.org/protobuf/internal/fieldsort/fieldsort.go b/vendor/google.golang.org/protobuf/internal/fieldsort/fieldsort.go new file mode 100644 index 00000000..517c4e2a --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/fieldsort/fieldsort.go @@ -0,0 +1,40 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package fieldsort defines an ordering of fields. +// +// The ordering defined by this package matches the historic behavior of the proto +// package, placing extensions first and oneofs last. +// +// There is no guarantee about stability of the wire encoding, and users should not +// depend on the order defined in this package as it is subject to change without +// notice. +package fieldsort + +import ( + "google.golang.org/protobuf/reflect/protoreflect" +) + +// Less returns true if field a comes before field j in ordered wire marshal output. +func Less(a, b protoreflect.FieldDescriptor) bool { + ea := a.IsExtension() + eb := b.IsExtension() + oa := a.ContainingOneof() + ob := b.ContainingOneof() + switch { + case ea != eb: + return ea + case oa != nil && ob != nil: + if oa == ob { + return a.Number() < b.Number() + } + return oa.Index() < ob.Index() + case oa != nil && !oa.IsSynthetic(): + return false + case ob != nil && !ob.IsSynthetic(): + return true + default: + return a.Number() < b.Number() + } +} diff --git a/vendor/google.golang.org/protobuf/internal/filedesc/build.go b/vendor/google.golang.org/protobuf/internal/filedesc/build.go new file mode 100644 index 00000000..462d384e --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filedesc/build.go @@ -0,0 +1,155 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package filedesc provides functionality for constructing descriptors. +package filedesc + +import ( + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/fieldnum" + "google.golang.org/protobuf/reflect/protoreflect" + pref "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" +) + +// Builder construct a protoreflect.FileDescriptor from the raw descriptor. +type Builder struct { + // GoPackagePath is the Go package path that is invoking this builder. + GoPackagePath string + + // RawDescriptor is the wire-encoded bytes of FileDescriptorProto + // and must be populated. + RawDescriptor []byte + + // NumEnums is the total number of enums declared in the file. + NumEnums int32 + // NumMessages is the total number of messages declared in the file. + // It includes the implicit message declarations for map entries. + NumMessages int32 + // NumExtensions is the total number of extensions declared in the file. + NumExtensions int32 + // NumServices is the total number of services declared in the file. + NumServices int32 + + // TypeResolver resolves extension field types for descriptor options. + // If nil, it uses protoregistry.GlobalTypes. + TypeResolver interface { + preg.ExtensionTypeResolver + } + + // FileRegistry is use to lookup file, enum, and message dependencies. + // Once constructed, the file descriptor is registered here. + // If nil, it uses protoregistry.GlobalFiles. + FileRegistry interface { + FindFileByPath(string) (protoreflect.FileDescriptor, error) + FindDescriptorByName(pref.FullName) (pref.Descriptor, error) + RegisterFile(pref.FileDescriptor) error + } +} + +// resolverByIndex is an interface Builder.FileRegistry may implement. +// If so, it permits looking up an enum or message dependency based on the +// sub-list and element index into filetype.Builder.DependencyIndexes. +type resolverByIndex interface { + FindEnumByIndex(int32, int32, []Enum, []Message) pref.EnumDescriptor + FindMessageByIndex(int32, int32, []Enum, []Message) pref.MessageDescriptor +} + +// Indexes of each sub-list in filetype.Builder.DependencyIndexes. +const ( + listFieldDeps int32 = iota + listExtTargets + listExtDeps + listMethInDeps + listMethOutDeps +) + +// Out is the output of the Builder. +type Out struct { + File pref.FileDescriptor + + // Enums is all enum descriptors in "flattened ordering". + Enums []Enum + // Messages is all message descriptors in "flattened ordering". + // It includes the implicit message declarations for map entries. + Messages []Message + // Extensions is all extension descriptors in "flattened ordering". + Extensions []Extension + // Service is all service descriptors in "flattened ordering". + Services []Service +} + +// Build constructs a FileDescriptor given the parameters set in Builder. +// It assumes that the inputs are well-formed and panics if any inconsistencies +// are encountered. +// +// If NumEnums+NumMessages+NumExtensions+NumServices is zero, +// then Build automatically derives them from the raw descriptor. +func (db Builder) Build() (out Out) { + // Populate the counts if uninitialized. + if db.NumEnums+db.NumMessages+db.NumExtensions+db.NumServices == 0 { + db.unmarshalCounts(db.RawDescriptor, true) + } + + // Initialize resolvers and registries if unpopulated. + if db.TypeResolver == nil { + db.TypeResolver = preg.GlobalTypes + } + if db.FileRegistry == nil { + db.FileRegistry = preg.GlobalFiles + } + + fd := newRawFile(db) + out.File = fd + out.Enums = fd.allEnums + out.Messages = fd.allMessages + out.Extensions = fd.allExtensions + out.Services = fd.allServices + + if err := db.FileRegistry.RegisterFile(fd); err != nil { + panic(err) + } + return out +} + +// unmarshalCounts counts the number of enum, message, extension, and service +// declarations in the raw message, which is either a FileDescriptorProto +// or a MessageDescriptorProto depending on whether isFile is set. +func (db *Builder) unmarshalCounts(b []byte, isFile bool) { + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + if isFile { + switch num { + case fieldnum.FileDescriptorProto_EnumType: + db.NumEnums++ + case fieldnum.FileDescriptorProto_MessageType: + db.unmarshalCounts(v, false) + db.NumMessages++ + case fieldnum.FileDescriptorProto_Extension: + db.NumExtensions++ + case fieldnum.FileDescriptorProto_Service: + db.NumServices++ + } + } else { + switch num { + case fieldnum.DescriptorProto_EnumType: + db.NumEnums++ + case fieldnum.DescriptorProto_NestedType: + db.unmarshalCounts(v, false) + db.NumMessages++ + case fieldnum.DescriptorProto_Extension: + db.NumExtensions++ + } + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} diff --git a/vendor/google.golang.org/protobuf/internal/filedesc/desc.go b/vendor/google.golang.org/protobuf/internal/filedesc/desc.go new file mode 100644 index 00000000..2540befd --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filedesc/desc.go @@ -0,0 +1,613 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package filedesc + +import ( + "bytes" + "fmt" + "sync" + "sync/atomic" + + "google.golang.org/protobuf/internal/descfmt" + "google.golang.org/protobuf/internal/descopts" + "google.golang.org/protobuf/internal/encoding/defval" + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/internal/strs" + pref "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +// The types in this file may have a suffix: +// • L0: Contains fields common to all descriptors (except File) and +// must be initialized up front. +// • L1: Contains fields specific to a descriptor and +// must be initialized up front. +// • L2: Contains fields that are lazily initialized when constructing +// from the raw file descriptor. When constructing as a literal, the L2 +// fields must be initialized up front. +// +// The types are exported so that packages like reflect/protodesc can +// directly construct descriptors. + +type ( + File struct { + fileRaw + L1 FileL1 + + once uint32 // atomically set if L2 is valid + mu sync.Mutex // protects L2 + L2 *FileL2 + } + FileL1 struct { + Syntax pref.Syntax + Path string + Package pref.FullName + + Enums Enums + Messages Messages + Extensions Extensions + Services Services + } + FileL2 struct { + Options func() pref.ProtoMessage + Imports FileImports + Locations SourceLocations + } +) + +func (fd *File) ParentFile() pref.FileDescriptor { return fd } +func (fd *File) Parent() pref.Descriptor { return nil } +func (fd *File) Index() int { return 0 } +func (fd *File) Syntax() pref.Syntax { return fd.L1.Syntax } +func (fd *File) Name() pref.Name { return fd.L1.Package.Name() } +func (fd *File) FullName() pref.FullName { return fd.L1.Package } +func (fd *File) IsPlaceholder() bool { return false } +func (fd *File) Options() pref.ProtoMessage { + if f := fd.lazyInit().Options; f != nil { + return f() + } + return descopts.File +} +func (fd *File) Path() string { return fd.L1.Path } +func (fd *File) Package() pref.FullName { return fd.L1.Package } +func (fd *File) Imports() pref.FileImports { return &fd.lazyInit().Imports } +func (fd *File) Enums() pref.EnumDescriptors { return &fd.L1.Enums } +func (fd *File) Messages() pref.MessageDescriptors { return &fd.L1.Messages } +func (fd *File) Extensions() pref.ExtensionDescriptors { return &fd.L1.Extensions } +func (fd *File) Services() pref.ServiceDescriptors { return &fd.L1.Services } +func (fd *File) SourceLocations() pref.SourceLocations { return &fd.lazyInit().Locations } +func (fd *File) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, fd) } +func (fd *File) ProtoType(pref.FileDescriptor) {} +func (fd *File) ProtoInternal(pragma.DoNotImplement) {} + +func (fd *File) lazyInit() *FileL2 { + if atomic.LoadUint32(&fd.once) == 0 { + fd.lazyInitOnce() + } + return fd.L2 +} + +func (fd *File) lazyInitOnce() { + fd.mu.Lock() + if fd.L2 == nil { + fd.lazyRawInit() // recursively initializes all L2 structures + } + atomic.StoreUint32(&fd.once, 1) + fd.mu.Unlock() +} + +// ProtoLegacyRawDesc is a pseudo-internal API for allowing the v1 code +// to be able to retrieve the raw descriptor. +// +// WARNING: This method is exempt from the compatibility promise and may be +// removed in the future without warning. +func (fd *File) ProtoLegacyRawDesc() []byte { + return fd.builder.RawDescriptor +} + +// GoPackagePath is a pseudo-internal API for determining the Go package path +// that this file descriptor is declared in. +// +// WARNING: This method is exempt from the compatibility promise and may be +// removed in the future without warning. +func (fd *File) GoPackagePath() string { + return fd.builder.GoPackagePath +} + +type ( + Enum struct { + Base + L1 EnumL1 + L2 *EnumL2 // protected by fileDesc.once + } + EnumL1 struct { + eagerValues bool // controls whether EnumL2.Values is already populated + } + EnumL2 struct { + Options func() pref.ProtoMessage + Values EnumValues + ReservedNames Names + ReservedRanges EnumRanges + } + + EnumValue struct { + Base + L1 EnumValueL1 + } + EnumValueL1 struct { + Options func() pref.ProtoMessage + Number pref.EnumNumber + } +) + +func (ed *Enum) Options() pref.ProtoMessage { + if f := ed.lazyInit().Options; f != nil { + return f() + } + return descopts.Enum +} +func (ed *Enum) Values() pref.EnumValueDescriptors { + if ed.L1.eagerValues { + return &ed.L2.Values + } + return &ed.lazyInit().Values +} +func (ed *Enum) ReservedNames() pref.Names { return &ed.lazyInit().ReservedNames } +func (ed *Enum) ReservedRanges() pref.EnumRanges { return &ed.lazyInit().ReservedRanges } +func (ed *Enum) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, ed) } +func (ed *Enum) ProtoType(pref.EnumDescriptor) {} +func (ed *Enum) lazyInit() *EnumL2 { + ed.L0.ParentFile.lazyInit() // implicitly initializes L2 + return ed.L2 +} + +func (ed *EnumValue) Options() pref.ProtoMessage { + if f := ed.L1.Options; f != nil { + return f() + } + return descopts.EnumValue +} +func (ed *EnumValue) Number() pref.EnumNumber { return ed.L1.Number } +func (ed *EnumValue) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, ed) } +func (ed *EnumValue) ProtoType(pref.EnumValueDescriptor) {} + +type ( + Message struct { + Base + L1 MessageL1 + L2 *MessageL2 // protected by fileDesc.once + } + MessageL1 struct { + Enums Enums + Messages Messages + Extensions Extensions + IsMapEntry bool // promoted from google.protobuf.MessageOptions + IsMessageSet bool // promoted from google.protobuf.MessageOptions + } + MessageL2 struct { + Options func() pref.ProtoMessage + Fields Fields + Oneofs Oneofs + ReservedNames Names + ReservedRanges FieldRanges + RequiredNumbers FieldNumbers // must be consistent with Fields.Cardinality + ExtensionRanges FieldRanges + ExtensionRangeOptions []func() pref.ProtoMessage // must be same length as ExtensionRanges + } + + Field struct { + Base + L1 FieldL1 + } + FieldL1 struct { + Options func() pref.ProtoMessage + Number pref.FieldNumber + Cardinality pref.Cardinality // must be consistent with Message.RequiredNumbers + Kind pref.Kind + JSONName jsonName + IsProto3Optional bool // promoted from google.protobuf.FieldDescriptorProto + IsWeak bool // promoted from google.protobuf.FieldOptions + HasPacked bool // promoted from google.protobuf.FieldOptions + IsPacked bool // promoted from google.protobuf.FieldOptions + HasEnforceUTF8 bool // promoted from google.protobuf.FieldOptions + EnforceUTF8 bool // promoted from google.protobuf.FieldOptions + Default defaultValue + ContainingOneof pref.OneofDescriptor // must be consistent with Message.Oneofs.Fields + Enum pref.EnumDescriptor + Message pref.MessageDescriptor + } + + Oneof struct { + Base + L1 OneofL1 + } + OneofL1 struct { + Options func() pref.ProtoMessage + Fields OneofFields // must be consistent with Message.Fields.ContainingOneof + } +) + +func (md *Message) Options() pref.ProtoMessage { + if f := md.lazyInit().Options; f != nil { + return f() + } + return descopts.Message +} +func (md *Message) IsMapEntry() bool { return md.L1.IsMapEntry } +func (md *Message) Fields() pref.FieldDescriptors { return &md.lazyInit().Fields } +func (md *Message) Oneofs() pref.OneofDescriptors { return &md.lazyInit().Oneofs } +func (md *Message) ReservedNames() pref.Names { return &md.lazyInit().ReservedNames } +func (md *Message) ReservedRanges() pref.FieldRanges { return &md.lazyInit().ReservedRanges } +func (md *Message) RequiredNumbers() pref.FieldNumbers { return &md.lazyInit().RequiredNumbers } +func (md *Message) ExtensionRanges() pref.FieldRanges { return &md.lazyInit().ExtensionRanges } +func (md *Message) ExtensionRangeOptions(i int) pref.ProtoMessage { + if f := md.lazyInit().ExtensionRangeOptions[i]; f != nil { + return f() + } + return descopts.ExtensionRange +} +func (md *Message) Enums() pref.EnumDescriptors { return &md.L1.Enums } +func (md *Message) Messages() pref.MessageDescriptors { return &md.L1.Messages } +func (md *Message) Extensions() pref.ExtensionDescriptors { return &md.L1.Extensions } +func (md *Message) ProtoType(pref.MessageDescriptor) {} +func (md *Message) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, md) } +func (md *Message) lazyInit() *MessageL2 { + md.L0.ParentFile.lazyInit() // implicitly initializes L2 + return md.L2 +} + +// IsMessageSet is a pseudo-internal API for checking whether a message +// should serialize in the proto1 message format. +// +// WARNING: This method is exempt from the compatibility promise and may be +// removed in the future without warning. +func (md *Message) IsMessageSet() bool { + return md.L1.IsMessageSet +} + +func (fd *Field) Options() pref.ProtoMessage { + if f := fd.L1.Options; f != nil { + return f() + } + return descopts.Field +} +func (fd *Field) Number() pref.FieldNumber { return fd.L1.Number } +func (fd *Field) Cardinality() pref.Cardinality { return fd.L1.Cardinality } +func (fd *Field) Kind() pref.Kind { return fd.L1.Kind } +func (fd *Field) HasJSONName() bool { return fd.L1.JSONName.has } +func (fd *Field) JSONName() string { return fd.L1.JSONName.get(fd) } +func (fd *Field) HasPresence() bool { + return fd.L1.Cardinality != pref.Repeated && (fd.L0.ParentFile.L1.Syntax == pref.Proto2 || fd.L1.Message != nil || fd.L1.ContainingOneof != nil) +} +func (fd *Field) HasOptionalKeyword() bool { + return (fd.L0.ParentFile.L1.Syntax == pref.Proto2 && fd.L1.Cardinality == pref.Optional && fd.L1.ContainingOneof == nil) || fd.L1.IsProto3Optional +} +func (fd *Field) IsPacked() bool { + if !fd.L1.HasPacked && fd.L0.ParentFile.L1.Syntax != pref.Proto2 && fd.L1.Cardinality == pref.Repeated { + switch fd.L1.Kind { + case pref.StringKind, pref.BytesKind, pref.MessageKind, pref.GroupKind: + default: + return true + } + } + return fd.L1.IsPacked +} +func (fd *Field) IsExtension() bool { return false } +func (fd *Field) IsWeak() bool { return fd.L1.IsWeak } +func (fd *Field) IsList() bool { return fd.Cardinality() == pref.Repeated && !fd.IsMap() } +func (fd *Field) IsMap() bool { return fd.Message() != nil && fd.Message().IsMapEntry() } +func (fd *Field) MapKey() pref.FieldDescriptor { + if !fd.IsMap() { + return nil + } + return fd.Message().Fields().ByNumber(1) +} +func (fd *Field) MapValue() pref.FieldDescriptor { + if !fd.IsMap() { + return nil + } + return fd.Message().Fields().ByNumber(2) +} +func (fd *Field) HasDefault() bool { return fd.L1.Default.has } +func (fd *Field) Default() pref.Value { return fd.L1.Default.get(fd) } +func (fd *Field) DefaultEnumValue() pref.EnumValueDescriptor { return fd.L1.Default.enum } +func (fd *Field) ContainingOneof() pref.OneofDescriptor { return fd.L1.ContainingOneof } +func (fd *Field) ContainingMessage() pref.MessageDescriptor { + return fd.L0.Parent.(pref.MessageDescriptor) +} +func (fd *Field) Enum() pref.EnumDescriptor { + return fd.L1.Enum +} +func (fd *Field) Message() pref.MessageDescriptor { + if fd.L1.IsWeak { + if d, _ := protoregistry.GlobalFiles.FindDescriptorByName(fd.L1.Message.FullName()); d != nil { + return d.(pref.MessageDescriptor) + } + } + return fd.L1.Message +} +func (fd *Field) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, fd) } +func (fd *Field) ProtoType(pref.FieldDescriptor) {} + +// EnforceUTF8 is a pseudo-internal API to determine whether to enforce UTF-8 +// validation for the string field. This exists for Google-internal use only +// since proto3 did not enforce UTF-8 validity prior to the open-source release. +// If this method does not exist, the default is to enforce valid UTF-8. +// +// WARNING: This method is exempt from the compatibility promise and may be +// removed in the future without warning. +func (fd *Field) EnforceUTF8() bool { + if fd.L1.HasEnforceUTF8 { + return fd.L1.EnforceUTF8 + } + return fd.L0.ParentFile.L1.Syntax == pref.Proto3 +} + +func (od *Oneof) IsSynthetic() bool { + return od.L0.ParentFile.L1.Syntax == pref.Proto3 && len(od.L1.Fields.List) == 1 && od.L1.Fields.List[0].HasOptionalKeyword() +} +func (od *Oneof) Options() pref.ProtoMessage { + if f := od.L1.Options; f != nil { + return f() + } + return descopts.Oneof +} +func (od *Oneof) Fields() pref.FieldDescriptors { return &od.L1.Fields } +func (od *Oneof) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, od) } +func (od *Oneof) ProtoType(pref.OneofDescriptor) {} + +type ( + Extension struct { + Base + L1 ExtensionL1 + L2 *ExtensionL2 // protected by fileDesc.once + } + ExtensionL1 struct { + Number pref.FieldNumber + Extendee pref.MessageDescriptor + Cardinality pref.Cardinality + Kind pref.Kind + } + ExtensionL2 struct { + Options func() pref.ProtoMessage + JSONName jsonName + IsProto3Optional bool // promoted from google.protobuf.FieldDescriptorProto + IsPacked bool // promoted from google.protobuf.FieldOptions + Default defaultValue + Enum pref.EnumDescriptor + Message pref.MessageDescriptor + } +) + +func (xd *Extension) Options() pref.ProtoMessage { + if f := xd.lazyInit().Options; f != nil { + return f() + } + return descopts.Field +} +func (xd *Extension) Number() pref.FieldNumber { return xd.L1.Number } +func (xd *Extension) Cardinality() pref.Cardinality { return xd.L1.Cardinality } +func (xd *Extension) Kind() pref.Kind { return xd.L1.Kind } +func (xd *Extension) HasJSONName() bool { return xd.lazyInit().JSONName.has } +func (xd *Extension) JSONName() string { return xd.lazyInit().JSONName.get(xd) } +func (xd *Extension) HasPresence() bool { return xd.L1.Cardinality != pref.Repeated } +func (xd *Extension) HasOptionalKeyword() bool { + return (xd.L0.ParentFile.L1.Syntax == pref.Proto2 && xd.L1.Cardinality == pref.Optional) || xd.lazyInit().IsProto3Optional +} +func (xd *Extension) IsPacked() bool { return xd.lazyInit().IsPacked } +func (xd *Extension) IsExtension() bool { return true } +func (xd *Extension) IsWeak() bool { return false } +func (xd *Extension) IsList() bool { return xd.Cardinality() == pref.Repeated } +func (xd *Extension) IsMap() bool { return false } +func (xd *Extension) MapKey() pref.FieldDescriptor { return nil } +func (xd *Extension) MapValue() pref.FieldDescriptor { return nil } +func (xd *Extension) HasDefault() bool { return xd.lazyInit().Default.has } +func (xd *Extension) Default() pref.Value { return xd.lazyInit().Default.get(xd) } +func (xd *Extension) DefaultEnumValue() pref.EnumValueDescriptor { return xd.lazyInit().Default.enum } +func (xd *Extension) ContainingOneof() pref.OneofDescriptor { return nil } +func (xd *Extension) ContainingMessage() pref.MessageDescriptor { return xd.L1.Extendee } +func (xd *Extension) Enum() pref.EnumDescriptor { return xd.lazyInit().Enum } +func (xd *Extension) Message() pref.MessageDescriptor { return xd.lazyInit().Message } +func (xd *Extension) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, xd) } +func (xd *Extension) ProtoType(pref.FieldDescriptor) {} +func (xd *Extension) ProtoInternal(pragma.DoNotImplement) {} +func (xd *Extension) lazyInit() *ExtensionL2 { + xd.L0.ParentFile.lazyInit() // implicitly initializes L2 + return xd.L2 +} + +type ( + Service struct { + Base + L1 ServiceL1 + L2 *ServiceL2 // protected by fileDesc.once + } + ServiceL1 struct{} + ServiceL2 struct { + Options func() pref.ProtoMessage + Methods Methods + } + + Method struct { + Base + L1 MethodL1 + } + MethodL1 struct { + Options func() pref.ProtoMessage + Input pref.MessageDescriptor + Output pref.MessageDescriptor + IsStreamingClient bool + IsStreamingServer bool + } +) + +func (sd *Service) Options() pref.ProtoMessage { + if f := sd.lazyInit().Options; f != nil { + return f() + } + return descopts.Service +} +func (sd *Service) Methods() pref.MethodDescriptors { return &sd.lazyInit().Methods } +func (sd *Service) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, sd) } +func (sd *Service) ProtoType(pref.ServiceDescriptor) {} +func (sd *Service) ProtoInternal(pragma.DoNotImplement) {} +func (sd *Service) lazyInit() *ServiceL2 { + sd.L0.ParentFile.lazyInit() // implicitly initializes L2 + return sd.L2 +} + +func (md *Method) Options() pref.ProtoMessage { + if f := md.L1.Options; f != nil { + return f() + } + return descopts.Method +} +func (md *Method) Input() pref.MessageDescriptor { return md.L1.Input } +func (md *Method) Output() pref.MessageDescriptor { return md.L1.Output } +func (md *Method) IsStreamingClient() bool { return md.L1.IsStreamingClient } +func (md *Method) IsStreamingServer() bool { return md.L1.IsStreamingServer } +func (md *Method) Format(s fmt.State, r rune) { descfmt.FormatDesc(s, r, md) } +func (md *Method) ProtoType(pref.MethodDescriptor) {} +func (md *Method) ProtoInternal(pragma.DoNotImplement) {} + +// Surrogate files are can be used to create standalone descriptors +// where the syntax is only information derived from the parent file. +var ( + SurrogateProto2 = &File{L1: FileL1{Syntax: pref.Proto2}, L2: &FileL2{}} + SurrogateProto3 = &File{L1: FileL1{Syntax: pref.Proto3}, L2: &FileL2{}} +) + +type ( + Base struct { + L0 BaseL0 + } + BaseL0 struct { + FullName pref.FullName // must be populated + ParentFile *File // must be populated + Parent pref.Descriptor + Index int + } +) + +func (d *Base) Name() pref.Name { return d.L0.FullName.Name() } +func (d *Base) FullName() pref.FullName { return d.L0.FullName } +func (d *Base) ParentFile() pref.FileDescriptor { + if d.L0.ParentFile == SurrogateProto2 || d.L0.ParentFile == SurrogateProto3 { + return nil // surrogate files are not real parents + } + return d.L0.ParentFile +} +func (d *Base) Parent() pref.Descriptor { return d.L0.Parent } +func (d *Base) Index() int { return d.L0.Index } +func (d *Base) Syntax() pref.Syntax { return d.L0.ParentFile.Syntax() } +func (d *Base) IsPlaceholder() bool { return false } +func (d *Base) ProtoInternal(pragma.DoNotImplement) {} + +type jsonName struct { + has bool + once sync.Once + name string +} + +// Init initializes the name. It is exported for use by other internal packages. +func (js *jsonName) Init(s string) { + js.has = true + js.name = s +} + +func (js *jsonName) get(fd pref.FieldDescriptor) string { + if !js.has { + js.once.Do(func() { + js.name = strs.JSONCamelCase(string(fd.Name())) + }) + } + return js.name +} + +func DefaultValue(v pref.Value, ev pref.EnumValueDescriptor) defaultValue { + dv := defaultValue{has: v.IsValid(), val: v, enum: ev} + if b, ok := v.Interface().([]byte); ok { + // Store a copy of the default bytes, so that we can detect + // accidental mutations of the original value. + dv.bytes = append([]byte(nil), b...) + } + return dv +} + +func unmarshalDefault(b []byte, k pref.Kind, pf *File, ed pref.EnumDescriptor) defaultValue { + var evs pref.EnumValueDescriptors + if k == pref.EnumKind { + // If the enum is declared within the same file, be careful not to + // blindly call the Values method, lest we bind ourselves in a deadlock. + if e, ok := ed.(*Enum); ok && e.L0.ParentFile == pf { + evs = &e.L2.Values + } else { + evs = ed.Values() + } + + // If we are unable to resolve the enum dependency, use a placeholder + // enum value since we will not be able to parse the default value. + if ed.IsPlaceholder() && pref.Name(b).IsValid() { + v := pref.ValueOfEnum(0) + ev := PlaceholderEnumValue(ed.FullName().Parent().Append(pref.Name(b))) + return DefaultValue(v, ev) + } + } + + v, ev, err := defval.Unmarshal(string(b), k, evs, defval.Descriptor) + if err != nil { + panic(err) + } + return DefaultValue(v, ev) +} + +type defaultValue struct { + has bool + val pref.Value + enum pref.EnumValueDescriptor + bytes []byte +} + +func (dv *defaultValue) get(fd pref.FieldDescriptor) pref.Value { + // Return the zero value as the default if unpopulated. + if !dv.has { + if fd.Cardinality() == pref.Repeated { + return pref.Value{} + } + switch fd.Kind() { + case pref.BoolKind: + return pref.ValueOfBool(false) + case pref.Int32Kind, pref.Sint32Kind, pref.Sfixed32Kind: + return pref.ValueOfInt32(0) + case pref.Int64Kind, pref.Sint64Kind, pref.Sfixed64Kind: + return pref.ValueOfInt64(0) + case pref.Uint32Kind, pref.Fixed32Kind: + return pref.ValueOfUint32(0) + case pref.Uint64Kind, pref.Fixed64Kind: + return pref.ValueOfUint64(0) + case pref.FloatKind: + return pref.ValueOfFloat32(0) + case pref.DoubleKind: + return pref.ValueOfFloat64(0) + case pref.StringKind: + return pref.ValueOfString("") + case pref.BytesKind: + return pref.ValueOfBytes(nil) + case pref.EnumKind: + if evs := fd.Enum().Values(); evs.Len() > 0 { + return pref.ValueOfEnum(evs.Get(0).Number()) + } + return pref.ValueOfEnum(0) + } + } + + if len(dv.bytes) > 0 && !bytes.Equal(dv.bytes, dv.val.Bytes()) { + // TODO: Avoid panic if we're running with the race detector + // and instead spawn a goroutine that periodically resets + // this value back to the original to induce a race. + panic(fmt.Sprintf("detected mutation on the default bytes for %v", fd.FullName())) + } + return dv.val +} diff --git a/vendor/google.golang.org/protobuf/internal/filedesc/desc_init.go b/vendor/google.golang.org/protobuf/internal/filedesc/desc_init.go new file mode 100644 index 00000000..c0cddf86 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filedesc/desc_init.go @@ -0,0 +1,471 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package filedesc + +import ( + "sync" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/fieldnum" + "google.golang.org/protobuf/internal/strs" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +// fileRaw is a data struct used when initializing a file descriptor from +// a raw FileDescriptorProto. +type fileRaw struct { + builder Builder + allEnums []Enum + allMessages []Message + allExtensions []Extension + allServices []Service +} + +func newRawFile(db Builder) *File { + fd := &File{fileRaw: fileRaw{builder: db}} + fd.initDecls(db.NumEnums, db.NumMessages, db.NumExtensions, db.NumServices) + fd.unmarshalSeed(db.RawDescriptor) + + // Extended message targets are eagerly resolved since registration + // needs this information at program init time. + for i := range fd.allExtensions { + xd := &fd.allExtensions[i] + xd.L1.Extendee = fd.resolveMessageDependency(xd.L1.Extendee, listExtTargets, int32(i)) + } + + fd.checkDecls() + return fd +} + +// initDecls pre-allocates slices for the exact number of enums, messages +// (including map entries), extensions, and services declared in the proto file. +// This is done to avoid regrowing the slice, which would change the address +// for any previously seen declaration. +// +// The alloc methods "allocates" slices by pulling from the capacity. +func (fd *File) initDecls(numEnums, numMessages, numExtensions, numServices int32) { + fd.allEnums = make([]Enum, 0, numEnums) + fd.allMessages = make([]Message, 0, numMessages) + fd.allExtensions = make([]Extension, 0, numExtensions) + fd.allServices = make([]Service, 0, numServices) +} + +func (fd *File) allocEnums(n int) []Enum { + total := len(fd.allEnums) + es := fd.allEnums[total : total+n] + fd.allEnums = fd.allEnums[:total+n] + return es +} +func (fd *File) allocMessages(n int) []Message { + total := len(fd.allMessages) + ms := fd.allMessages[total : total+n] + fd.allMessages = fd.allMessages[:total+n] + return ms +} +func (fd *File) allocExtensions(n int) []Extension { + total := len(fd.allExtensions) + xs := fd.allExtensions[total : total+n] + fd.allExtensions = fd.allExtensions[:total+n] + return xs +} +func (fd *File) allocServices(n int) []Service { + total := len(fd.allServices) + xs := fd.allServices[total : total+n] + fd.allServices = fd.allServices[:total+n] + return xs +} + +// checkDecls performs a sanity check that the expected number of expected +// declarations matches the number that were found in the descriptor proto. +func (fd *File) checkDecls() { + switch { + case len(fd.allEnums) != cap(fd.allEnums): + case len(fd.allMessages) != cap(fd.allMessages): + case len(fd.allExtensions) != cap(fd.allExtensions): + case len(fd.allServices) != cap(fd.allServices): + default: + return + } + panic("mismatching cardinality") +} + +func (fd *File) unmarshalSeed(b []byte) { + sb := getBuilder() + defer putBuilder(sb) + + var prevField pref.FieldNumber + var numEnums, numMessages, numExtensions, numServices int + var posEnums, posMessages, posExtensions, posServices int + b0 := b + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.FileDescriptorProto_Syntax: + switch string(v) { + case "proto2": + fd.L1.Syntax = pref.Proto2 + case "proto3": + fd.L1.Syntax = pref.Proto3 + default: + panic("invalid syntax") + } + case fieldnum.FileDescriptorProto_Name: + fd.L1.Path = sb.MakeString(v) + case fieldnum.FileDescriptorProto_Package: + fd.L1.Package = pref.FullName(sb.MakeString(v)) + case fieldnum.FileDescriptorProto_EnumType: + if prevField != fieldnum.FileDescriptorProto_EnumType { + if numEnums > 0 { + panic("non-contiguous repeated field") + } + posEnums = len(b0) - len(b) - n - m + } + numEnums++ + case fieldnum.FileDescriptorProto_MessageType: + if prevField != fieldnum.FileDescriptorProto_MessageType { + if numMessages > 0 { + panic("non-contiguous repeated field") + } + posMessages = len(b0) - len(b) - n - m + } + numMessages++ + case fieldnum.FileDescriptorProto_Extension: + if prevField != fieldnum.FileDescriptorProto_Extension { + if numExtensions > 0 { + panic("non-contiguous repeated field") + } + posExtensions = len(b0) - len(b) - n - m + } + numExtensions++ + case fieldnum.FileDescriptorProto_Service: + if prevField != fieldnum.FileDescriptorProto_Service { + if numServices > 0 { + panic("non-contiguous repeated field") + } + posServices = len(b0) - len(b) - n - m + } + numServices++ + } + prevField = num + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + prevField = -1 // ignore known field numbers of unknown wire type + } + } + + // If syntax is missing, it is assumed to be proto2. + if fd.L1.Syntax == 0 { + fd.L1.Syntax = pref.Proto2 + } + + // Must allocate all declarations before parsing each descriptor type + // to ensure we handled all descriptors in "flattened ordering". + if numEnums > 0 { + fd.L1.Enums.List = fd.allocEnums(numEnums) + } + if numMessages > 0 { + fd.L1.Messages.List = fd.allocMessages(numMessages) + } + if numExtensions > 0 { + fd.L1.Extensions.List = fd.allocExtensions(numExtensions) + } + if numServices > 0 { + fd.L1.Services.List = fd.allocServices(numServices) + } + + if numEnums > 0 { + b := b0[posEnums:] + for i := range fd.L1.Enums.List { + _, n := protowire.ConsumeVarint(b) + v, m := protowire.ConsumeBytes(b[n:]) + fd.L1.Enums.List[i].unmarshalSeed(v, sb, fd, fd, i) + b = b[n+m:] + } + } + if numMessages > 0 { + b := b0[posMessages:] + for i := range fd.L1.Messages.List { + _, n := protowire.ConsumeVarint(b) + v, m := protowire.ConsumeBytes(b[n:]) + fd.L1.Messages.List[i].unmarshalSeed(v, sb, fd, fd, i) + b = b[n+m:] + } + } + if numExtensions > 0 { + b := b0[posExtensions:] + for i := range fd.L1.Extensions.List { + _, n := protowire.ConsumeVarint(b) + v, m := protowire.ConsumeBytes(b[n:]) + fd.L1.Extensions.List[i].unmarshalSeed(v, sb, fd, fd, i) + b = b[n+m:] + } + } + if numServices > 0 { + b := b0[posServices:] + for i := range fd.L1.Services.List { + _, n := protowire.ConsumeVarint(b) + v, m := protowire.ConsumeBytes(b[n:]) + fd.L1.Services.List[i].unmarshalSeed(v, sb, fd, fd, i) + b = b[n+m:] + } + } +} + +func (ed *Enum) unmarshalSeed(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + ed.L0.ParentFile = pf + ed.L0.Parent = pd + ed.L0.Index = i + + var numValues int + for b := b; len(b) > 0; { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.EnumDescriptorProto_Name: + ed.L0.FullName = appendFullName(sb, pd.FullName(), v) + case fieldnum.EnumDescriptorProto_Value: + numValues++ + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + + // Only construct enum value descriptors for top-level enums since + // they are needed for registration. + if pd != pf { + return + } + ed.L1.eagerValues = true + ed.L2 = new(EnumL2) + ed.L2.Values.List = make([]EnumValue, numValues) + for i := 0; len(b) > 0; { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.EnumDescriptorProto_Value: + ed.L2.Values.List[i].unmarshalFull(v, sb, pf, ed, i) + i++ + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} + +func (md *Message) unmarshalSeed(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + md.L0.ParentFile = pf + md.L0.Parent = pd + md.L0.Index = i + + var prevField pref.FieldNumber + var numEnums, numMessages, numExtensions int + var posEnums, posMessages, posExtensions int + b0 := b + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.DescriptorProto_Name: + md.L0.FullName = appendFullName(sb, pd.FullName(), v) + case fieldnum.DescriptorProto_EnumType: + if prevField != fieldnum.DescriptorProto_EnumType { + if numEnums > 0 { + panic("non-contiguous repeated field") + } + posEnums = len(b0) - len(b) - n - m + } + numEnums++ + case fieldnum.DescriptorProto_NestedType: + if prevField != fieldnum.DescriptorProto_NestedType { + if numMessages > 0 { + panic("non-contiguous repeated field") + } + posMessages = len(b0) - len(b) - n - m + } + numMessages++ + case fieldnum.DescriptorProto_Extension: + if prevField != fieldnum.DescriptorProto_Extension { + if numExtensions > 0 { + panic("non-contiguous repeated field") + } + posExtensions = len(b0) - len(b) - n - m + } + numExtensions++ + case fieldnum.DescriptorProto_Options: + md.unmarshalSeedOptions(v) + } + prevField = num + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + prevField = -1 // ignore known field numbers of unknown wire type + } + } + + // Must allocate all declarations before parsing each descriptor type + // to ensure we handled all descriptors in "flattened ordering". + if numEnums > 0 { + md.L1.Enums.List = pf.allocEnums(numEnums) + } + if numMessages > 0 { + md.L1.Messages.List = pf.allocMessages(numMessages) + } + if numExtensions > 0 { + md.L1.Extensions.List = pf.allocExtensions(numExtensions) + } + + if numEnums > 0 { + b := b0[posEnums:] + for i := range md.L1.Enums.List { + _, n := protowire.ConsumeVarint(b) + v, m := protowire.ConsumeBytes(b[n:]) + md.L1.Enums.List[i].unmarshalSeed(v, sb, pf, md, i) + b = b[n+m:] + } + } + if numMessages > 0 { + b := b0[posMessages:] + for i := range md.L1.Messages.List { + _, n := protowire.ConsumeVarint(b) + v, m := protowire.ConsumeBytes(b[n:]) + md.L1.Messages.List[i].unmarshalSeed(v, sb, pf, md, i) + b = b[n+m:] + } + } + if numExtensions > 0 { + b := b0[posExtensions:] + for i := range md.L1.Extensions.List { + _, n := protowire.ConsumeVarint(b) + v, m := protowire.ConsumeBytes(b[n:]) + md.L1.Extensions.List[i].unmarshalSeed(v, sb, pf, md, i) + b = b[n+m:] + } + } +} + +func (md *Message) unmarshalSeedOptions(b []byte) { + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.MessageOptions_MapEntry: + md.L1.IsMapEntry = protowire.DecodeBool(v) + case fieldnum.MessageOptions_MessageSetWireFormat: + md.L1.IsMessageSet = protowire.DecodeBool(v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} + +func (xd *Extension) unmarshalSeed(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + xd.L0.ParentFile = pf + xd.L0.Parent = pd + xd.L0.Index = i + + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.FieldDescriptorProto_Number: + xd.L1.Number = pref.FieldNumber(v) + case fieldnum.FieldDescriptorProto_Label: + xd.L1.Cardinality = pref.Cardinality(v) + case fieldnum.FieldDescriptorProto_Type: + xd.L1.Kind = pref.Kind(v) + } + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.FieldDescriptorProto_Name: + xd.L0.FullName = appendFullName(sb, pd.FullName(), v) + case fieldnum.FieldDescriptorProto_Extendee: + xd.L1.Extendee = PlaceholderMessage(makeFullName(sb, v)) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} + +func (sd *Service) unmarshalSeed(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + sd.L0.ParentFile = pf + sd.L0.Parent = pd + sd.L0.Index = i + + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.ServiceDescriptorProto_Name: + sd.L0.FullName = appendFullName(sb, pd.FullName(), v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} + +var nameBuilderPool = sync.Pool{ + New: func() interface{} { return new(strs.Builder) }, +} + +func getBuilder() *strs.Builder { + return nameBuilderPool.Get().(*strs.Builder) +} +func putBuilder(b *strs.Builder) { + nameBuilderPool.Put(b) +} + +// makeFullName converts b to a protoreflect.FullName, +// where b must start with a leading dot. +func makeFullName(sb *strs.Builder, b []byte) pref.FullName { + if len(b) == 0 || b[0] != '.' { + panic("name reference must be fully qualified") + } + return pref.FullName(sb.MakeString(b[1:])) +} + +func appendFullName(sb *strs.Builder, prefix pref.FullName, suffix []byte) pref.FullName { + return sb.AppendFullName(prefix, pref.Name(strs.UnsafeString(suffix))) +} diff --git a/vendor/google.golang.org/protobuf/internal/filedesc/desc_lazy.go b/vendor/google.golang.org/protobuf/internal/filedesc/desc_lazy.go new file mode 100644 index 00000000..bc215944 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filedesc/desc_lazy.go @@ -0,0 +1,704 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package filedesc + +import ( + "reflect" + "sync" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/descopts" + "google.golang.org/protobuf/internal/fieldnum" + "google.golang.org/protobuf/internal/strs" + "google.golang.org/protobuf/proto" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +func (fd *File) lazyRawInit() { + fd.unmarshalFull(fd.builder.RawDescriptor) + fd.resolveMessages() + fd.resolveExtensions() + fd.resolveServices() +} + +func (file *File) resolveMessages() { + var depIdx int32 + for i := range file.allMessages { + md := &file.allMessages[i] + + // Resolve message field dependencies. + for j := range md.L2.Fields.List { + fd := &md.L2.Fields.List[j] + + // Weak fields are resolved upon actual use. + if fd.L1.IsWeak { + continue + } + + // Resolve message field dependency. + switch fd.L1.Kind { + case pref.EnumKind: + fd.L1.Enum = file.resolveEnumDependency(fd.L1.Enum, listFieldDeps, depIdx) + depIdx++ + case pref.MessageKind, pref.GroupKind: + fd.L1.Message = file.resolveMessageDependency(fd.L1.Message, listFieldDeps, depIdx) + depIdx++ + } + + // Default is resolved here since it depends on Enum being resolved. + if v := fd.L1.Default.val; v.IsValid() { + fd.L1.Default = unmarshalDefault(v.Bytes(), fd.L1.Kind, file, fd.L1.Enum) + } + } + } +} + +func (file *File) resolveExtensions() { + var depIdx int32 + for i := range file.allExtensions { + xd := &file.allExtensions[i] + + // Resolve extension field dependency. + switch xd.L1.Kind { + case pref.EnumKind: + xd.L2.Enum = file.resolveEnumDependency(xd.L2.Enum, listExtDeps, depIdx) + depIdx++ + case pref.MessageKind, pref.GroupKind: + xd.L2.Message = file.resolveMessageDependency(xd.L2.Message, listExtDeps, depIdx) + depIdx++ + } + + // Default is resolved here since it depends on Enum being resolved. + if v := xd.L2.Default.val; v.IsValid() { + xd.L2.Default = unmarshalDefault(v.Bytes(), xd.L1.Kind, file, xd.L2.Enum) + } + } +} + +func (file *File) resolveServices() { + var depIdx int32 + for i := range file.allServices { + sd := &file.allServices[i] + + // Resolve method dependencies. + for j := range sd.L2.Methods.List { + md := &sd.L2.Methods.List[j] + md.L1.Input = file.resolveMessageDependency(md.L1.Input, listMethInDeps, depIdx) + md.L1.Output = file.resolveMessageDependency(md.L1.Output, listMethOutDeps, depIdx) + depIdx++ + } + } +} + +func (file *File) resolveEnumDependency(ed pref.EnumDescriptor, i, j int32) pref.EnumDescriptor { + r := file.builder.FileRegistry + if r, ok := r.(resolverByIndex); ok { + if ed2 := r.FindEnumByIndex(i, j, file.allEnums, file.allMessages); ed2 != nil { + return ed2 + } + } + for i := range file.allEnums { + if ed2 := &file.allEnums[i]; ed2.L0.FullName == ed.FullName() { + return ed2 + } + } + if d, _ := r.FindDescriptorByName(ed.FullName()); d != nil { + return d.(pref.EnumDescriptor) + } + return ed +} + +func (file *File) resolveMessageDependency(md pref.MessageDescriptor, i, j int32) pref.MessageDescriptor { + r := file.builder.FileRegistry + if r, ok := r.(resolverByIndex); ok { + if md2 := r.FindMessageByIndex(i, j, file.allEnums, file.allMessages); md2 != nil { + return md2 + } + } + for i := range file.allMessages { + if md2 := &file.allMessages[i]; md2.L0.FullName == md.FullName() { + return md2 + } + } + if d, _ := r.FindDescriptorByName(md.FullName()); d != nil { + return d.(pref.MessageDescriptor) + } + return md +} + +func (fd *File) unmarshalFull(b []byte) { + sb := getBuilder() + defer putBuilder(sb) + + var enumIdx, messageIdx, extensionIdx, serviceIdx int + var rawOptions []byte + fd.L2 = new(FileL2) + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.FileDescriptorProto_PublicDependency: + fd.L2.Imports[v].IsPublic = true + case fieldnum.FileDescriptorProto_WeakDependency: + fd.L2.Imports[v].IsWeak = true + } + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.FileDescriptorProto_Dependency: + path := sb.MakeString(v) + imp, _ := fd.builder.FileRegistry.FindFileByPath(path) + if imp == nil { + imp = PlaceholderFile(path) + } + fd.L2.Imports = append(fd.L2.Imports, pref.FileImport{FileDescriptor: imp}) + case fieldnum.FileDescriptorProto_EnumType: + fd.L1.Enums.List[enumIdx].unmarshalFull(v, sb) + enumIdx++ + case fieldnum.FileDescriptorProto_MessageType: + fd.L1.Messages.List[messageIdx].unmarshalFull(v, sb) + messageIdx++ + case fieldnum.FileDescriptorProto_Extension: + fd.L1.Extensions.List[extensionIdx].unmarshalFull(v, sb) + extensionIdx++ + case fieldnum.FileDescriptorProto_Service: + fd.L1.Services.List[serviceIdx].unmarshalFull(v, sb) + serviceIdx++ + case fieldnum.FileDescriptorProto_Options: + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + fd.L2.Options = fd.builder.optionsUnmarshaler(&descopts.File, rawOptions) +} + +func (ed *Enum) unmarshalFull(b []byte, sb *strs.Builder) { + var rawValues [][]byte + var rawOptions []byte + if !ed.L1.eagerValues { + ed.L2 = new(EnumL2) + } + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.EnumDescriptorProto_Value: + rawValues = append(rawValues, v) + case fieldnum.EnumDescriptorProto_ReservedName: + ed.L2.ReservedNames.List = append(ed.L2.ReservedNames.List, pref.Name(sb.MakeString(v))) + case fieldnum.EnumDescriptorProto_ReservedRange: + ed.L2.ReservedRanges.List = append(ed.L2.ReservedRanges.List, unmarshalEnumReservedRange(v)) + case fieldnum.EnumDescriptorProto_Options: + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + if !ed.L1.eagerValues && len(rawValues) > 0 { + ed.L2.Values.List = make([]EnumValue, len(rawValues)) + for i, b := range rawValues { + ed.L2.Values.List[i].unmarshalFull(b, sb, ed.L0.ParentFile, ed, i) + } + } + ed.L2.Options = ed.L0.ParentFile.builder.optionsUnmarshaler(&descopts.Enum, rawOptions) +} + +func unmarshalEnumReservedRange(b []byte) (r [2]pref.EnumNumber) { + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.EnumDescriptorProto_EnumReservedRange_Start: + r[0] = pref.EnumNumber(v) + case fieldnum.EnumDescriptorProto_EnumReservedRange_End: + r[1] = pref.EnumNumber(v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + return r +} + +func (vd *EnumValue) unmarshalFull(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + vd.L0.ParentFile = pf + vd.L0.Parent = pd + vd.L0.Index = i + + var rawOptions []byte + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.EnumValueDescriptorProto_Number: + vd.L1.Number = pref.EnumNumber(v) + } + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.EnumValueDescriptorProto_Name: + // NOTE: Enum values are in the same scope as the enum parent. + vd.L0.FullName = appendFullName(sb, pd.Parent().FullName(), v) + case fieldnum.EnumValueDescriptorProto_Options: + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + vd.L1.Options = pf.builder.optionsUnmarshaler(&descopts.EnumValue, rawOptions) +} + +func (md *Message) unmarshalFull(b []byte, sb *strs.Builder) { + var rawFields, rawOneofs [][]byte + var enumIdx, messageIdx, extensionIdx int + var rawOptions []byte + md.L2 = new(MessageL2) + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.DescriptorProto_Field: + rawFields = append(rawFields, v) + case fieldnum.DescriptorProto_OneofDecl: + rawOneofs = append(rawOneofs, v) + case fieldnum.DescriptorProto_ReservedName: + md.L2.ReservedNames.List = append(md.L2.ReservedNames.List, pref.Name(sb.MakeString(v))) + case fieldnum.DescriptorProto_ReservedRange: + md.L2.ReservedRanges.List = append(md.L2.ReservedRanges.List, unmarshalMessageReservedRange(v)) + case fieldnum.DescriptorProto_ExtensionRange: + r, rawOptions := unmarshalMessageExtensionRange(v) + opts := md.L0.ParentFile.builder.optionsUnmarshaler(&descopts.ExtensionRange, rawOptions) + md.L2.ExtensionRanges.List = append(md.L2.ExtensionRanges.List, r) + md.L2.ExtensionRangeOptions = append(md.L2.ExtensionRangeOptions, opts) + case fieldnum.DescriptorProto_EnumType: + md.L1.Enums.List[enumIdx].unmarshalFull(v, sb) + enumIdx++ + case fieldnum.DescriptorProto_NestedType: + md.L1.Messages.List[messageIdx].unmarshalFull(v, sb) + messageIdx++ + case fieldnum.DescriptorProto_Extension: + md.L1.Extensions.List[extensionIdx].unmarshalFull(v, sb) + extensionIdx++ + case fieldnum.DescriptorProto_Options: + md.unmarshalOptions(v) + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + if len(rawFields) > 0 || len(rawOneofs) > 0 { + md.L2.Fields.List = make([]Field, len(rawFields)) + md.L2.Oneofs.List = make([]Oneof, len(rawOneofs)) + for i, b := range rawFields { + fd := &md.L2.Fields.List[i] + fd.unmarshalFull(b, sb, md.L0.ParentFile, md, i) + if fd.L1.Cardinality == pref.Required { + md.L2.RequiredNumbers.List = append(md.L2.RequiredNumbers.List, fd.L1.Number) + } + } + for i, b := range rawOneofs { + od := &md.L2.Oneofs.List[i] + od.unmarshalFull(b, sb, md.L0.ParentFile, md, i) + } + } + md.L2.Options = md.L0.ParentFile.builder.optionsUnmarshaler(&descopts.Message, rawOptions) +} + +func (md *Message) unmarshalOptions(b []byte) { + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.MessageOptions_MapEntry: + md.L1.IsMapEntry = protowire.DecodeBool(v) + case fieldnum.MessageOptions_MessageSetWireFormat: + md.L1.IsMessageSet = protowire.DecodeBool(v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} + +func unmarshalMessageReservedRange(b []byte) (r [2]pref.FieldNumber) { + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.DescriptorProto_ReservedRange_Start: + r[0] = pref.FieldNumber(v) + case fieldnum.DescriptorProto_ReservedRange_End: + r[1] = pref.FieldNumber(v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + return r +} + +func unmarshalMessageExtensionRange(b []byte) (r [2]pref.FieldNumber, rawOptions []byte) { + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.DescriptorProto_ExtensionRange_Start: + r[0] = pref.FieldNumber(v) + case fieldnum.DescriptorProto_ExtensionRange_End: + r[1] = pref.FieldNumber(v) + } + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.DescriptorProto_ExtensionRange_Options: + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + return r, rawOptions +} + +func (fd *Field) unmarshalFull(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + fd.L0.ParentFile = pf + fd.L0.Parent = pd + fd.L0.Index = i + + var rawTypeName []byte + var rawOptions []byte + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.FieldDescriptorProto_Number: + fd.L1.Number = pref.FieldNumber(v) + case fieldnum.FieldDescriptorProto_Label: + fd.L1.Cardinality = pref.Cardinality(v) + case fieldnum.FieldDescriptorProto_Type: + fd.L1.Kind = pref.Kind(v) + case fieldnum.FieldDescriptorProto_OneofIndex: + // In Message.unmarshalFull, we allocate slices for both + // the field and oneof descriptors before unmarshaling either + // of them. This ensures pointers to slice elements are stable. + od := &pd.(*Message).L2.Oneofs.List[v] + od.L1.Fields.List = append(od.L1.Fields.List, fd) + if fd.L1.ContainingOneof != nil { + panic("oneof type already set") + } + fd.L1.ContainingOneof = od + case fieldnum.FieldDescriptorProto_Proto3Optional: + fd.L1.IsProto3Optional = protowire.DecodeBool(v) + } + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.FieldDescriptorProto_Name: + fd.L0.FullName = appendFullName(sb, pd.FullName(), v) + case fieldnum.FieldDescriptorProto_JsonName: + fd.L1.JSONName.Init(sb.MakeString(v)) + case fieldnum.FieldDescriptorProto_DefaultValue: + fd.L1.Default.val = pref.ValueOfBytes(v) // temporarily store as bytes; later resolved in resolveMessages + case fieldnum.FieldDescriptorProto_TypeName: + rawTypeName = v + case fieldnum.FieldDescriptorProto_Options: + fd.unmarshalOptions(v) + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + if rawTypeName != nil { + name := makeFullName(sb, rawTypeName) + switch fd.L1.Kind { + case pref.EnumKind: + fd.L1.Enum = PlaceholderEnum(name) + case pref.MessageKind, pref.GroupKind: + fd.L1.Message = PlaceholderMessage(name) + } + } + fd.L1.Options = pf.builder.optionsUnmarshaler(&descopts.Field, rawOptions) +} + +func (fd *Field) unmarshalOptions(b []byte) { + const FieldOptions_EnforceUTF8 = 13 + + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.FieldOptions_Packed: + fd.L1.HasPacked = true + fd.L1.IsPacked = protowire.DecodeBool(v) + case fieldnum.FieldOptions_Weak: + fd.L1.IsWeak = protowire.DecodeBool(v) + case FieldOptions_EnforceUTF8: + fd.L1.HasEnforceUTF8 = true + fd.L1.EnforceUTF8 = protowire.DecodeBool(v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} + +func (od *Oneof) unmarshalFull(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + od.L0.ParentFile = pf + od.L0.Parent = pd + od.L0.Index = i + + var rawOptions []byte + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.OneofDescriptorProto_Name: + od.L0.FullName = appendFullName(sb, pd.FullName(), v) + case fieldnum.OneofDescriptorProto_Options: + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + od.L1.Options = pf.builder.optionsUnmarshaler(&descopts.Oneof, rawOptions) +} + +func (xd *Extension) unmarshalFull(b []byte, sb *strs.Builder) { + var rawTypeName []byte + var rawOptions []byte + xd.L2 = new(ExtensionL2) + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.FieldDescriptorProto_Proto3Optional: + xd.L2.IsProto3Optional = protowire.DecodeBool(v) + } + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.FieldDescriptorProto_JsonName: + xd.L2.JSONName.Init(sb.MakeString(v)) + case fieldnum.FieldDescriptorProto_DefaultValue: + xd.L2.Default.val = pref.ValueOfBytes(v) // temporarily store as bytes; later resolved in resolveExtensions + case fieldnum.FieldDescriptorProto_TypeName: + rawTypeName = v + case fieldnum.FieldDescriptorProto_Options: + xd.unmarshalOptions(v) + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + if rawTypeName != nil { + name := makeFullName(sb, rawTypeName) + switch xd.L1.Kind { + case pref.EnumKind: + xd.L2.Enum = PlaceholderEnum(name) + case pref.MessageKind, pref.GroupKind: + xd.L2.Message = PlaceholderMessage(name) + } + } + xd.L2.Options = xd.L0.ParentFile.builder.optionsUnmarshaler(&descopts.Field, rawOptions) +} + +func (xd *Extension) unmarshalOptions(b []byte) { + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.FieldOptions_Packed: + xd.L2.IsPacked = protowire.DecodeBool(v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } +} + +func (sd *Service) unmarshalFull(b []byte, sb *strs.Builder) { + var rawMethods [][]byte + var rawOptions []byte + sd.L2 = new(ServiceL2) + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.ServiceDescriptorProto_Method: + rawMethods = append(rawMethods, v) + case fieldnum.ServiceDescriptorProto_Options: + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + if len(rawMethods) > 0 { + sd.L2.Methods.List = make([]Method, len(rawMethods)) + for i, b := range rawMethods { + sd.L2.Methods.List[i].unmarshalFull(b, sb, sd.L0.ParentFile, sd, i) + } + } + sd.L2.Options = sd.L0.ParentFile.builder.optionsUnmarshaler(&descopts.Service, rawOptions) +} + +func (md *Method) unmarshalFull(b []byte, sb *strs.Builder, pf *File, pd pref.Descriptor, i int) { + md.L0.ParentFile = pf + md.L0.Parent = pd + md.L0.Index = i + + var rawOptions []byte + for len(b) > 0 { + num, typ, n := protowire.ConsumeTag(b) + b = b[n:] + switch typ { + case protowire.VarintType: + v, m := protowire.ConsumeVarint(b) + b = b[m:] + switch num { + case fieldnum.MethodDescriptorProto_ClientStreaming: + md.L1.IsStreamingClient = protowire.DecodeBool(v) + case fieldnum.MethodDescriptorProto_ServerStreaming: + md.L1.IsStreamingServer = protowire.DecodeBool(v) + } + case protowire.BytesType: + v, m := protowire.ConsumeBytes(b) + b = b[m:] + switch num { + case fieldnum.MethodDescriptorProto_Name: + md.L0.FullName = appendFullName(sb, pd.FullName(), v) + case fieldnum.MethodDescriptorProto_InputType: + md.L1.Input = PlaceholderMessage(makeFullName(sb, v)) + case fieldnum.MethodDescriptorProto_OutputType: + md.L1.Output = PlaceholderMessage(makeFullName(sb, v)) + case fieldnum.MethodDescriptorProto_Options: + rawOptions = appendOptions(rawOptions, v) + } + default: + m := protowire.ConsumeFieldValue(num, typ, b) + b = b[m:] + } + } + md.L1.Options = pf.builder.optionsUnmarshaler(&descopts.Method, rawOptions) +} + +// appendOptions appends src to dst, where the returned slice is never nil. +// This is necessary to distinguish between empty and unpopulated options. +func appendOptions(dst, src []byte) []byte { + if dst == nil { + dst = []byte{} + } + return append(dst, src...) +} + +// optionsUnmarshaler constructs a lazy unmarshal function for an options message. +// +// The type of message to unmarshal to is passed as a pointer since the +// vars in descopts may not yet be populated at the time this function is called. +func (db *Builder) optionsUnmarshaler(p *pref.ProtoMessage, b []byte) func() pref.ProtoMessage { + if b == nil { + return nil + } + var opts pref.ProtoMessage + var once sync.Once + return func() pref.ProtoMessage { + once.Do(func() { + if *p == nil { + panic("Descriptor.Options called without importing the descriptor package") + } + opts = reflect.New(reflect.TypeOf(*p).Elem()).Interface().(pref.ProtoMessage) + if err := (proto.UnmarshalOptions{ + AllowPartial: true, + Resolver: db.TypeResolver, + }).Unmarshal(b, opts); err != nil { + panic(err) + } + }) + return opts + } +} diff --git a/vendor/google.golang.org/protobuf/internal/filedesc/desc_list.go b/vendor/google.golang.org/protobuf/internal/filedesc/desc_list.go new file mode 100644 index 00000000..1b7089b6 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filedesc/desc_list.go @@ -0,0 +1,286 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package filedesc + +import ( + "fmt" + "math" + "sort" + "sync" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/descfmt" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/reflect/protoreflect" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type FileImports []pref.FileImport + +func (p *FileImports) Len() int { return len(*p) } +func (p *FileImports) Get(i int) pref.FileImport { return (*p)[i] } +func (p *FileImports) Format(s fmt.State, r rune) { descfmt.FormatList(s, r, p) } +func (p *FileImports) ProtoInternal(pragma.DoNotImplement) {} + +type Names struct { + List []pref.Name + once sync.Once + has map[pref.Name]int // protected by once +} + +func (p *Names) Len() int { return len(p.List) } +func (p *Names) Get(i int) pref.Name { return p.List[i] } +func (p *Names) Has(s pref.Name) bool { return p.lazyInit().has[s] > 0 } +func (p *Names) Format(s fmt.State, r rune) { descfmt.FormatList(s, r, p) } +func (p *Names) ProtoInternal(pragma.DoNotImplement) {} +func (p *Names) lazyInit() *Names { + p.once.Do(func() { + if len(p.List) > 0 { + p.has = make(map[pref.Name]int, len(p.List)) + for _, s := range p.List { + p.has[s] = p.has[s] + 1 + } + } + }) + return p +} + +// CheckValid reports any errors with the set of names with an error message +// that completes the sentence: "ranges is invalid because it has ..." +func (p *Names) CheckValid() error { + for s, n := range p.lazyInit().has { + switch { + case n > 1: + return errors.New("duplicate name: %q", s) + case false && !s.IsValid(): + // NOTE: The C++ implementation does not validate the identifier. + // See https://github.com/protocolbuffers/protobuf/issues/6335. + return errors.New("invalid name: %q", s) + } + } + return nil +} + +type EnumRanges struct { + List [][2]pref.EnumNumber // start inclusive; end inclusive + once sync.Once + sorted [][2]pref.EnumNumber // protected by once +} + +func (p *EnumRanges) Len() int { return len(p.List) } +func (p *EnumRanges) Get(i int) [2]pref.EnumNumber { return p.List[i] } +func (p *EnumRanges) Has(n pref.EnumNumber) bool { + for ls := p.lazyInit().sorted; len(ls) > 0; { + i := len(ls) / 2 + switch r := enumRange(ls[i]); { + case n < r.Start(): + ls = ls[:i] // search lower + case n > r.End(): + ls = ls[i+1:] // search upper + default: + return true + } + } + return false +} +func (p *EnumRanges) Format(s fmt.State, r rune) { descfmt.FormatList(s, r, p) } +func (p *EnumRanges) ProtoInternal(pragma.DoNotImplement) {} +func (p *EnumRanges) lazyInit() *EnumRanges { + p.once.Do(func() { + p.sorted = append(p.sorted, p.List...) + sort.Slice(p.sorted, func(i, j int) bool { + return p.sorted[i][0] < p.sorted[j][0] + }) + }) + return p +} + +// CheckValid reports any errors with the set of names with an error message +// that completes the sentence: "ranges is invalid because it has ..." +func (p *EnumRanges) CheckValid() error { + var rp enumRange + for i, r := range p.lazyInit().sorted { + r := enumRange(r) + switch { + case !(r.Start() <= r.End()): + return errors.New("invalid range: %v", r) + case !(rp.End() < r.Start()) && i > 0: + return errors.New("overlapping ranges: %v with %v", rp, r) + } + rp = r + } + return nil +} + +type enumRange [2]protoreflect.EnumNumber + +func (r enumRange) Start() protoreflect.EnumNumber { return r[0] } // inclusive +func (r enumRange) End() protoreflect.EnumNumber { return r[1] } // inclusive +func (r enumRange) String() string { + if r.Start() == r.End() { + return fmt.Sprintf("%d", r.Start()) + } + return fmt.Sprintf("%d to %d", r.Start(), r.End()) +} + +type FieldRanges struct { + List [][2]pref.FieldNumber // start inclusive; end exclusive + once sync.Once + sorted [][2]pref.FieldNumber // protected by once +} + +func (p *FieldRanges) Len() int { return len(p.List) } +func (p *FieldRanges) Get(i int) [2]pref.FieldNumber { return p.List[i] } +func (p *FieldRanges) Has(n pref.FieldNumber) bool { + for ls := p.lazyInit().sorted; len(ls) > 0; { + i := len(ls) / 2 + switch r := fieldRange(ls[i]); { + case n < r.Start(): + ls = ls[:i] // search lower + case n > r.End(): + ls = ls[i+1:] // search upper + default: + return true + } + } + return false +} +func (p *FieldRanges) Format(s fmt.State, r rune) { descfmt.FormatList(s, r, p) } +func (p *FieldRanges) ProtoInternal(pragma.DoNotImplement) {} +func (p *FieldRanges) lazyInit() *FieldRanges { + p.once.Do(func() { + p.sorted = append(p.sorted, p.List...) + sort.Slice(p.sorted, func(i, j int) bool { + return p.sorted[i][0] < p.sorted[j][0] + }) + }) + return p +} + +// CheckValid reports any errors with the set of ranges with an error message +// that completes the sentence: "ranges is invalid because it has ..." +func (p *FieldRanges) CheckValid(isMessageSet bool) error { + var rp fieldRange + for i, r := range p.lazyInit().sorted { + r := fieldRange(r) + switch { + case !isValidFieldNumber(r.Start(), isMessageSet): + return errors.New("invalid field number: %d", r.Start()) + case !isValidFieldNumber(r.End(), isMessageSet): + return errors.New("invalid field number: %d", r.End()) + case !(r.Start() <= r.End()): + return errors.New("invalid range: %v", r) + case !(rp.End() < r.Start()) && i > 0: + return errors.New("overlapping ranges: %v with %v", rp, r) + } + rp = r + } + return nil +} + +// isValidFieldNumber reports whether the field number is valid. +// Unlike the FieldNumber.IsValid method, it allows ranges that cover the +// reserved number range. +func isValidFieldNumber(n protoreflect.FieldNumber, isMessageSet bool) bool { + if isMessageSet { + return protowire.MinValidNumber <= n && n <= math.MaxInt32 + } + return protowire.MinValidNumber <= n && n <= protowire.MaxValidNumber +} + +// CheckOverlap reports an error if p and q overlap. +func (p *FieldRanges) CheckOverlap(q *FieldRanges) error { + rps := p.lazyInit().sorted + rqs := q.lazyInit().sorted + for pi, qi := 0, 0; pi < len(rps) && qi < len(rqs); { + rp := fieldRange(rps[pi]) + rq := fieldRange(rqs[qi]) + if !(rp.End() < rq.Start() || rq.End() < rp.Start()) { + return errors.New("overlapping ranges: %v with %v", rp, rq) + } + if rp.Start() < rq.Start() { + pi++ + } else { + qi++ + } + } + return nil +} + +type fieldRange [2]protoreflect.FieldNumber + +func (r fieldRange) Start() protoreflect.FieldNumber { return r[0] } // inclusive +func (r fieldRange) End() protoreflect.FieldNumber { return r[1] - 1 } // inclusive +func (r fieldRange) String() string { + if r.Start() == r.End() { + return fmt.Sprintf("%d", r.Start()) + } + return fmt.Sprintf("%d to %d", r.Start(), r.End()) +} + +type FieldNumbers struct { + List []pref.FieldNumber + once sync.Once + has map[pref.FieldNumber]struct{} // protected by once +} + +func (p *FieldNumbers) Len() int { return len(p.List) } +func (p *FieldNumbers) Get(i int) pref.FieldNumber { return p.List[i] } +func (p *FieldNumbers) Has(n pref.FieldNumber) bool { + p.once.Do(func() { + if len(p.List) > 0 { + p.has = make(map[pref.FieldNumber]struct{}, len(p.List)) + for _, n := range p.List { + p.has[n] = struct{}{} + } + } + }) + _, ok := p.has[n] + return ok +} +func (p *FieldNumbers) Format(s fmt.State, r rune) { descfmt.FormatList(s, r, p) } +func (p *FieldNumbers) ProtoInternal(pragma.DoNotImplement) {} + +type OneofFields struct { + List []pref.FieldDescriptor + once sync.Once + byName map[pref.Name]pref.FieldDescriptor // protected by once + byJSON map[string]pref.FieldDescriptor // protected by once + byNum map[pref.FieldNumber]pref.FieldDescriptor // protected by once +} + +func (p *OneofFields) Len() int { return len(p.List) } +func (p *OneofFields) Get(i int) pref.FieldDescriptor { return p.List[i] } +func (p *OneofFields) ByName(s pref.Name) pref.FieldDescriptor { return p.lazyInit().byName[s] } +func (p *OneofFields) ByJSONName(s string) pref.FieldDescriptor { return p.lazyInit().byJSON[s] } +func (p *OneofFields) ByNumber(n pref.FieldNumber) pref.FieldDescriptor { return p.lazyInit().byNum[n] } +func (p *OneofFields) Format(s fmt.State, r rune) { descfmt.FormatList(s, r, p) } +func (p *OneofFields) ProtoInternal(pragma.DoNotImplement) {} + +func (p *OneofFields) lazyInit() *OneofFields { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[pref.Name]pref.FieldDescriptor, len(p.List)) + p.byJSON = make(map[string]pref.FieldDescriptor, len(p.List)) + p.byNum = make(map[pref.FieldNumber]pref.FieldDescriptor, len(p.List)) + for _, f := range p.List { + // Field names and numbers are guaranteed to be unique. + p.byName[f.Name()] = f + p.byJSON[f.JSONName()] = f + p.byNum[f.Number()] = f + } + } + }) + return p +} + +type SourceLocations struct { + List []pref.SourceLocation +} + +func (p *SourceLocations) Len() int { return len(p.List) } +func (p *SourceLocations) Get(i int) pref.SourceLocation { return p.List[i] } +func (p *SourceLocations) ProtoInternal(pragma.DoNotImplement) {} diff --git a/vendor/google.golang.org/protobuf/internal/filedesc/desc_list_gen.go b/vendor/google.golang.org/protobuf/internal/filedesc/desc_list_gen.go new file mode 100644 index 00000000..6a8825e8 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filedesc/desc_list_gen.go @@ -0,0 +1,345 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-types. DO NOT EDIT. + +package filedesc + +import ( + "fmt" + "sync" + + "google.golang.org/protobuf/internal/descfmt" + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/reflect/protoreflect" +) + +type Enums struct { + List []Enum + once sync.Once + byName map[protoreflect.Name]*Enum // protected by once +} + +func (p *Enums) Len() int { + return len(p.List) +} +func (p *Enums) Get(i int) protoreflect.EnumDescriptor { + return &p.List[i] +} +func (p *Enums) ByName(s protoreflect.Name) protoreflect.EnumDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *Enums) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *Enums) ProtoInternal(pragma.DoNotImplement) {} +func (p *Enums) lazyInit() *Enums { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*Enum, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + } + } + }) + return p +} + +type EnumValues struct { + List []EnumValue + once sync.Once + byName map[protoreflect.Name]*EnumValue // protected by once + byNum map[protoreflect.EnumNumber]*EnumValue // protected by once +} + +func (p *EnumValues) Len() int { + return len(p.List) +} +func (p *EnumValues) Get(i int) protoreflect.EnumValueDescriptor { + return &p.List[i] +} +func (p *EnumValues) ByName(s protoreflect.Name) protoreflect.EnumValueDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *EnumValues) ByNumber(n protoreflect.EnumNumber) protoreflect.EnumValueDescriptor { + if d := p.lazyInit().byNum[n]; d != nil { + return d + } + return nil +} +func (p *EnumValues) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *EnumValues) ProtoInternal(pragma.DoNotImplement) {} +func (p *EnumValues) lazyInit() *EnumValues { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*EnumValue, len(p.List)) + p.byNum = make(map[protoreflect.EnumNumber]*EnumValue, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + if _, ok := p.byNum[d.Number()]; !ok { + p.byNum[d.Number()] = d + } + } + } + }) + return p +} + +type Messages struct { + List []Message + once sync.Once + byName map[protoreflect.Name]*Message // protected by once +} + +func (p *Messages) Len() int { + return len(p.List) +} +func (p *Messages) Get(i int) protoreflect.MessageDescriptor { + return &p.List[i] +} +func (p *Messages) ByName(s protoreflect.Name) protoreflect.MessageDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *Messages) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *Messages) ProtoInternal(pragma.DoNotImplement) {} +func (p *Messages) lazyInit() *Messages { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*Message, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + } + } + }) + return p +} + +type Fields struct { + List []Field + once sync.Once + byName map[protoreflect.Name]*Field // protected by once + byJSON map[string]*Field // protected by once + byNum map[protoreflect.FieldNumber]*Field // protected by once +} + +func (p *Fields) Len() int { + return len(p.List) +} +func (p *Fields) Get(i int) protoreflect.FieldDescriptor { + return &p.List[i] +} +func (p *Fields) ByName(s protoreflect.Name) protoreflect.FieldDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *Fields) ByJSONName(s string) protoreflect.FieldDescriptor { + if d := p.lazyInit().byJSON[s]; d != nil { + return d + } + return nil +} +func (p *Fields) ByNumber(n protoreflect.FieldNumber) protoreflect.FieldDescriptor { + if d := p.lazyInit().byNum[n]; d != nil { + return d + } + return nil +} +func (p *Fields) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *Fields) ProtoInternal(pragma.DoNotImplement) {} +func (p *Fields) lazyInit() *Fields { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*Field, len(p.List)) + p.byJSON = make(map[string]*Field, len(p.List)) + p.byNum = make(map[protoreflect.FieldNumber]*Field, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + if _, ok := p.byJSON[d.JSONName()]; !ok { + p.byJSON[d.JSONName()] = d + } + if _, ok := p.byNum[d.Number()]; !ok { + p.byNum[d.Number()] = d + } + } + } + }) + return p +} + +type Oneofs struct { + List []Oneof + once sync.Once + byName map[protoreflect.Name]*Oneof // protected by once +} + +func (p *Oneofs) Len() int { + return len(p.List) +} +func (p *Oneofs) Get(i int) protoreflect.OneofDescriptor { + return &p.List[i] +} +func (p *Oneofs) ByName(s protoreflect.Name) protoreflect.OneofDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *Oneofs) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *Oneofs) ProtoInternal(pragma.DoNotImplement) {} +func (p *Oneofs) lazyInit() *Oneofs { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*Oneof, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + } + } + }) + return p +} + +type Extensions struct { + List []Extension + once sync.Once + byName map[protoreflect.Name]*Extension // protected by once +} + +func (p *Extensions) Len() int { + return len(p.List) +} +func (p *Extensions) Get(i int) protoreflect.ExtensionDescriptor { + return &p.List[i] +} +func (p *Extensions) ByName(s protoreflect.Name) protoreflect.ExtensionDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *Extensions) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *Extensions) ProtoInternal(pragma.DoNotImplement) {} +func (p *Extensions) lazyInit() *Extensions { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*Extension, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + } + } + }) + return p +} + +type Services struct { + List []Service + once sync.Once + byName map[protoreflect.Name]*Service // protected by once +} + +func (p *Services) Len() int { + return len(p.List) +} +func (p *Services) Get(i int) protoreflect.ServiceDescriptor { + return &p.List[i] +} +func (p *Services) ByName(s protoreflect.Name) protoreflect.ServiceDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *Services) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *Services) ProtoInternal(pragma.DoNotImplement) {} +func (p *Services) lazyInit() *Services { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*Service, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + } + } + }) + return p +} + +type Methods struct { + List []Method + once sync.Once + byName map[protoreflect.Name]*Method // protected by once +} + +func (p *Methods) Len() int { + return len(p.List) +} +func (p *Methods) Get(i int) protoreflect.MethodDescriptor { + return &p.List[i] +} +func (p *Methods) ByName(s protoreflect.Name) protoreflect.MethodDescriptor { + if d := p.lazyInit().byName[s]; d != nil { + return d + } + return nil +} +func (p *Methods) Format(s fmt.State, r rune) { + descfmt.FormatList(s, r, p) +} +func (p *Methods) ProtoInternal(pragma.DoNotImplement) {} +func (p *Methods) lazyInit() *Methods { + p.once.Do(func() { + if len(p.List) > 0 { + p.byName = make(map[protoreflect.Name]*Method, len(p.List)) + for i := range p.List { + d := &p.List[i] + if _, ok := p.byName[d.Name()]; !ok { + p.byName[d.Name()] = d + } + } + } + }) + return p +} diff --git a/vendor/google.golang.org/protobuf/internal/filedesc/placeholder.go b/vendor/google.golang.org/protobuf/internal/filedesc/placeholder.go new file mode 100644 index 00000000..dbf2c605 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filedesc/placeholder.go @@ -0,0 +1,107 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package filedesc + +import ( + "google.golang.org/protobuf/internal/descopts" + "google.golang.org/protobuf/internal/pragma" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +var ( + emptyNames = new(Names) + emptyEnumRanges = new(EnumRanges) + emptyFieldRanges = new(FieldRanges) + emptyFieldNumbers = new(FieldNumbers) + emptySourceLocations = new(SourceLocations) + + emptyFiles = new(FileImports) + emptyMessages = new(Messages) + emptyFields = new(Fields) + emptyOneofs = new(Oneofs) + emptyEnums = new(Enums) + emptyEnumValues = new(EnumValues) + emptyExtensions = new(Extensions) + emptyServices = new(Services) +) + +// PlaceholderFile is a placeholder, representing only the file path. +type PlaceholderFile string + +func (f PlaceholderFile) ParentFile() pref.FileDescriptor { return f } +func (f PlaceholderFile) Parent() pref.Descriptor { return nil } +func (f PlaceholderFile) Index() int { return 0 } +func (f PlaceholderFile) Syntax() pref.Syntax { return 0 } +func (f PlaceholderFile) Name() pref.Name { return "" } +func (f PlaceholderFile) FullName() pref.FullName { return "" } +func (f PlaceholderFile) IsPlaceholder() bool { return true } +func (f PlaceholderFile) Options() pref.ProtoMessage { return descopts.File } +func (f PlaceholderFile) Path() string { return string(f) } +func (f PlaceholderFile) Package() pref.FullName { return "" } +func (f PlaceholderFile) Imports() pref.FileImports { return emptyFiles } +func (f PlaceholderFile) Messages() pref.MessageDescriptors { return emptyMessages } +func (f PlaceholderFile) Enums() pref.EnumDescriptors { return emptyEnums } +func (f PlaceholderFile) Extensions() pref.ExtensionDescriptors { return emptyExtensions } +func (f PlaceholderFile) Services() pref.ServiceDescriptors { return emptyServices } +func (f PlaceholderFile) SourceLocations() pref.SourceLocations { return emptySourceLocations } +func (f PlaceholderFile) ProtoType(pref.FileDescriptor) { return } +func (f PlaceholderFile) ProtoInternal(pragma.DoNotImplement) { return } + +// PlaceholderEnum is a placeholder, representing only the full name. +type PlaceholderEnum pref.FullName + +func (e PlaceholderEnum) ParentFile() pref.FileDescriptor { return nil } +func (e PlaceholderEnum) Parent() pref.Descriptor { return nil } +func (e PlaceholderEnum) Index() int { return 0 } +func (e PlaceholderEnum) Syntax() pref.Syntax { return 0 } +func (e PlaceholderEnum) Name() pref.Name { return pref.FullName(e).Name() } +func (e PlaceholderEnum) FullName() pref.FullName { return pref.FullName(e) } +func (e PlaceholderEnum) IsPlaceholder() bool { return true } +func (e PlaceholderEnum) Options() pref.ProtoMessage { return descopts.Enum } +func (e PlaceholderEnum) Values() pref.EnumValueDescriptors { return emptyEnumValues } +func (e PlaceholderEnum) ReservedNames() pref.Names { return emptyNames } +func (e PlaceholderEnum) ReservedRanges() pref.EnumRanges { return emptyEnumRanges } +func (e PlaceholderEnum) ProtoType(pref.EnumDescriptor) { return } +func (e PlaceholderEnum) ProtoInternal(pragma.DoNotImplement) { return } + +// PlaceholderEnumValue is a placeholder, representing only the full name. +type PlaceholderEnumValue pref.FullName + +func (e PlaceholderEnumValue) ParentFile() pref.FileDescriptor { return nil } +func (e PlaceholderEnumValue) Parent() pref.Descriptor { return nil } +func (e PlaceholderEnumValue) Index() int { return 0 } +func (e PlaceholderEnumValue) Syntax() pref.Syntax { return 0 } +func (e PlaceholderEnumValue) Name() pref.Name { return pref.FullName(e).Name() } +func (e PlaceholderEnumValue) FullName() pref.FullName { return pref.FullName(e) } +func (e PlaceholderEnumValue) IsPlaceholder() bool { return true } +func (e PlaceholderEnumValue) Options() pref.ProtoMessage { return descopts.EnumValue } +func (e PlaceholderEnumValue) Number() pref.EnumNumber { return 0 } +func (e PlaceholderEnumValue) ProtoType(pref.EnumValueDescriptor) { return } +func (e PlaceholderEnumValue) ProtoInternal(pragma.DoNotImplement) { return } + +// PlaceholderMessage is a placeholder, representing only the full name. +type PlaceholderMessage pref.FullName + +func (m PlaceholderMessage) ParentFile() pref.FileDescriptor { return nil } +func (m PlaceholderMessage) Parent() pref.Descriptor { return nil } +func (m PlaceholderMessage) Index() int { return 0 } +func (m PlaceholderMessage) Syntax() pref.Syntax { return 0 } +func (m PlaceholderMessage) Name() pref.Name { return pref.FullName(m).Name() } +func (m PlaceholderMessage) FullName() pref.FullName { return pref.FullName(m) } +func (m PlaceholderMessage) IsPlaceholder() bool { return true } +func (m PlaceholderMessage) Options() pref.ProtoMessage { return descopts.Message } +func (m PlaceholderMessage) IsMapEntry() bool { return false } +func (m PlaceholderMessage) Fields() pref.FieldDescriptors { return emptyFields } +func (m PlaceholderMessage) Oneofs() pref.OneofDescriptors { return emptyOneofs } +func (m PlaceholderMessage) ReservedNames() pref.Names { return emptyNames } +func (m PlaceholderMessage) ReservedRanges() pref.FieldRanges { return emptyFieldRanges } +func (m PlaceholderMessage) RequiredNumbers() pref.FieldNumbers { return emptyFieldNumbers } +func (m PlaceholderMessage) ExtensionRanges() pref.FieldRanges { return emptyFieldRanges } +func (m PlaceholderMessage) ExtensionRangeOptions(int) pref.ProtoMessage { panic("index out of range") } +func (m PlaceholderMessage) Messages() pref.MessageDescriptors { return emptyMessages } +func (m PlaceholderMessage) Enums() pref.EnumDescriptors { return emptyEnums } +func (m PlaceholderMessage) Extensions() pref.ExtensionDescriptors { return emptyExtensions } +func (m PlaceholderMessage) ProtoType(pref.MessageDescriptor) { return } +func (m PlaceholderMessage) ProtoInternal(pragma.DoNotImplement) { return } diff --git a/vendor/google.golang.org/protobuf/internal/filetype/build.go b/vendor/google.golang.org/protobuf/internal/filetype/build.go new file mode 100644 index 00000000..0a0dd35d --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/filetype/build.go @@ -0,0 +1,297 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package filetype provides functionality for wrapping descriptors +// with Go type information. +package filetype + +import ( + "reflect" + + "google.golang.org/protobuf/internal/descopts" + fdesc "google.golang.org/protobuf/internal/filedesc" + pimpl "google.golang.org/protobuf/internal/impl" + pref "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" +) + +// Builder constructs type descriptors from a raw file descriptor +// and associated Go types for each enum and message declaration. +// +// +// Flattened Ordering +// +// The protobuf type system represents declarations as a tree. Certain nodes in +// the tree require us to either associate it with a concrete Go type or to +// resolve a dependency, which is information that must be provided separately +// since it cannot be derived from the file descriptor alone. +// +// However, representing a tree as Go literals is difficult to simply do in a +// space and time efficient way. Thus, we store them as a flattened list of +// objects where the serialization order from the tree-based form is important. +// +// The "flattened ordering" is defined as a tree traversal of all enum, message, +// extension, and service declarations using the following algorithm: +// +// def VisitFileDecls(fd): +// for e in fd.Enums: yield e +// for m in fd.Messages: yield m +// for x in fd.Extensions: yield x +// for s in fd.Services: yield s +// for m in fd.Messages: yield from VisitMessageDecls(m) +// +// def VisitMessageDecls(md): +// for e in md.Enums: yield e +// for m in md.Messages: yield m +// for x in md.Extensions: yield x +// for m in md.Messages: yield from VisitMessageDecls(m) +// +// The traversal starts at the root file descriptor and yields each direct +// declaration within each node before traversing into sub-declarations +// that children themselves may have. +type Builder struct { + // File is the underlying file descriptor builder. + File fdesc.Builder + + // GoTypes is a unique set of the Go types for all declarations and + // dependencies. Each type is represented as a zero value of the Go type. + // + // Declarations are Go types generated for enums and messages directly + // declared (not publicly imported) in the proto source file. + // Messages for map entries are accounted for, but represented by nil. + // Enum declarations in "flattened ordering" come first, followed by + // message declarations in "flattened ordering". + // + // Dependencies are Go types for enums or messages referenced by + // message fields (excluding weak fields), for parent extended messages of + // extension fields, for enums or messages referenced by extension fields, + // and for input and output messages referenced by service methods. + // Dependencies must come after declarations, but the ordering of + // dependencies themselves is unspecified. + GoTypes []interface{} + + // DependencyIndexes is an ordered list of indexes into GoTypes for the + // dependencies of messages, extensions, or services. + // + // There are 5 sub-lists in "flattened ordering" concatenated back-to-back: + // 0. Message field dependencies: list of the enum or message type + // referred to by every message field. + // 1. Extension field targets: list of the extended parent message of + // every extension. + // 2. Extension field dependencies: list of the enum or message type + // referred to by every extension field. + // 3. Service method inputs: list of the input message type + // referred to by every service method. + // 4. Service method outputs: list of the output message type + // referred to by every service method. + // + // The offset into DependencyIndexes for the start of each sub-list + // is appended to the end in reverse order. + DependencyIndexes []int32 + + // EnumInfos is a list of enum infos in "flattened ordering". + EnumInfos []pimpl.EnumInfo + + // MessageInfos is a list of message infos in "flattened ordering". + // If provided, the GoType and PBType for each element is populated. + // + // Requirement: len(MessageInfos) == len(Build.Messages) + MessageInfos []pimpl.MessageInfo + + // ExtensionInfos is a list of extension infos in "flattened ordering". + // Each element is initialized and registered with the protoregistry package. + // + // Requirement: len(LegacyExtensions) == len(Build.Extensions) + ExtensionInfos []pimpl.ExtensionInfo + + // TypeRegistry is the registry to register each type descriptor. + // If nil, it uses protoregistry.GlobalTypes. + TypeRegistry interface { + RegisterMessage(pref.MessageType) error + RegisterEnum(pref.EnumType) error + RegisterExtension(pref.ExtensionType) error + } +} + +// Out is the output of the builder. +type Out struct { + File pref.FileDescriptor +} + +func (tb Builder) Build() (out Out) { + // Replace the resolver with one that resolves dependencies by index, + // which is faster and more reliable than relying on the global registry. + if tb.File.FileRegistry == nil { + tb.File.FileRegistry = preg.GlobalFiles + } + tb.File.FileRegistry = &resolverByIndex{ + goTypes: tb.GoTypes, + depIdxs: tb.DependencyIndexes, + fileRegistry: tb.File.FileRegistry, + } + + // Initialize registry if unpopulated. + if tb.TypeRegistry == nil { + tb.TypeRegistry = preg.GlobalTypes + } + + fbOut := tb.File.Build() + out.File = fbOut.File + + // Process enums. + enumGoTypes := tb.GoTypes[:len(fbOut.Enums)] + if len(tb.EnumInfos) != len(fbOut.Enums) { + panic("mismatching enum lengths") + } + if len(fbOut.Enums) > 0 { + for i := range fbOut.Enums { + tb.EnumInfos[i] = pimpl.EnumInfo{ + GoReflectType: reflect.TypeOf(enumGoTypes[i]), + Desc: &fbOut.Enums[i], + } + // Register enum types. + if err := tb.TypeRegistry.RegisterEnum(&tb.EnumInfos[i]); err != nil { + panic(err) + } + } + } + + // Process messages. + messageGoTypes := tb.GoTypes[len(fbOut.Enums):][:len(fbOut.Messages)] + if len(tb.MessageInfos) != len(fbOut.Messages) { + panic("mismatching message lengths") + } + if len(fbOut.Messages) > 0 { + for i := range fbOut.Messages { + if messageGoTypes[i] == nil { + continue // skip map entry + } + + tb.MessageInfos[i].GoReflectType = reflect.TypeOf(messageGoTypes[i]) + tb.MessageInfos[i].Desc = &fbOut.Messages[i] + + // Register message types. + if err := tb.TypeRegistry.RegisterMessage(&tb.MessageInfos[i]); err != nil { + panic(err) + } + } + + // As a special-case for descriptor.proto, + // locally register concrete message type for the options. + if out.File.Path() == "google/protobuf/descriptor.proto" && out.File.Package() == "google.protobuf" { + for i := range fbOut.Messages { + switch fbOut.Messages[i].Name() { + case "FileOptions": + descopts.File = messageGoTypes[i].(pref.ProtoMessage) + case "EnumOptions": + descopts.Enum = messageGoTypes[i].(pref.ProtoMessage) + case "EnumValueOptions": + descopts.EnumValue = messageGoTypes[i].(pref.ProtoMessage) + case "MessageOptions": + descopts.Message = messageGoTypes[i].(pref.ProtoMessage) + case "FieldOptions": + descopts.Field = messageGoTypes[i].(pref.ProtoMessage) + case "OneofOptions": + descopts.Oneof = messageGoTypes[i].(pref.ProtoMessage) + case "ExtensionRangeOptions": + descopts.ExtensionRange = messageGoTypes[i].(pref.ProtoMessage) + case "ServiceOptions": + descopts.Service = messageGoTypes[i].(pref.ProtoMessage) + case "MethodOptions": + descopts.Method = messageGoTypes[i].(pref.ProtoMessage) + } + } + } + } + + // Process extensions. + if len(tb.ExtensionInfos) != len(fbOut.Extensions) { + panic("mismatching extension lengths") + } + var depIdx int32 + for i := range fbOut.Extensions { + // For enum and message kinds, determine the referent Go type so + // that we can construct their constructors. + const listExtDeps = 2 + var goType reflect.Type + switch fbOut.Extensions[i].L1.Kind { + case pref.EnumKind: + j := depIdxs.Get(tb.DependencyIndexes, listExtDeps, depIdx) + goType = reflect.TypeOf(tb.GoTypes[j]) + depIdx++ + case pref.MessageKind, pref.GroupKind: + j := depIdxs.Get(tb.DependencyIndexes, listExtDeps, depIdx) + goType = reflect.TypeOf(tb.GoTypes[j]) + depIdx++ + default: + goType = goTypeForPBKind[fbOut.Extensions[i].L1.Kind] + } + if fbOut.Extensions[i].IsList() { + goType = reflect.SliceOf(goType) + } + + pimpl.InitExtensionInfo(&tb.ExtensionInfos[i], &fbOut.Extensions[i], goType) + + // Register extension types. + if err := tb.TypeRegistry.RegisterExtension(&tb.ExtensionInfos[i]); err != nil { + panic(err) + } + } + + return out +} + +var goTypeForPBKind = map[pref.Kind]reflect.Type{ + pref.BoolKind: reflect.TypeOf(bool(false)), + pref.Int32Kind: reflect.TypeOf(int32(0)), + pref.Sint32Kind: reflect.TypeOf(int32(0)), + pref.Sfixed32Kind: reflect.TypeOf(int32(0)), + pref.Int64Kind: reflect.TypeOf(int64(0)), + pref.Sint64Kind: reflect.TypeOf(int64(0)), + pref.Sfixed64Kind: reflect.TypeOf(int64(0)), + pref.Uint32Kind: reflect.TypeOf(uint32(0)), + pref.Fixed32Kind: reflect.TypeOf(uint32(0)), + pref.Uint64Kind: reflect.TypeOf(uint64(0)), + pref.Fixed64Kind: reflect.TypeOf(uint64(0)), + pref.FloatKind: reflect.TypeOf(float32(0)), + pref.DoubleKind: reflect.TypeOf(float64(0)), + pref.StringKind: reflect.TypeOf(string("")), + pref.BytesKind: reflect.TypeOf([]byte(nil)), +} + +type depIdxs []int32 + +// Get retrieves the jth element of the ith sub-list. +func (x depIdxs) Get(i, j int32) int32 { + return x[x[int32(len(x))-i-1]+j] +} + +type ( + resolverByIndex struct { + goTypes []interface{} + depIdxs depIdxs + fileRegistry + } + fileRegistry interface { + FindFileByPath(string) (pref.FileDescriptor, error) + FindDescriptorByName(pref.FullName) (pref.Descriptor, error) + RegisterFile(pref.FileDescriptor) error + } +) + +func (r *resolverByIndex) FindEnumByIndex(i, j int32, es []fdesc.Enum, ms []fdesc.Message) pref.EnumDescriptor { + if depIdx := int(r.depIdxs.Get(i, j)); int(depIdx) < len(es)+len(ms) { + return &es[depIdx] + } else { + return pimpl.Export{}.EnumDescriptorOf(r.goTypes[depIdx]) + } +} + +func (r *resolverByIndex) FindMessageByIndex(i, j int32, es []fdesc.Enum, ms []fdesc.Message) pref.MessageDescriptor { + if depIdx := int(r.depIdxs.Get(i, j)); depIdx < len(es)+len(ms) { + return &ms[depIdx-len(es)] + } else { + return pimpl.Export{}.MessageDescriptorOf(r.goTypes[depIdx]) + } +} diff --git a/vendor/google.golang.org/protobuf/internal/flags/flags.go b/vendor/google.golang.org/protobuf/internal/flags/flags.go new file mode 100644 index 00000000..58372dd3 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/flags/flags.go @@ -0,0 +1,24 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package flags provides a set of flags controlled by build tags. +package flags + +// ProtoLegacy specifies whether to enable support for legacy functionality +// such as MessageSets, weak fields, and various other obscure behavior +// that is necessary to maintain backwards compatibility with proto1 or +// the pre-release variants of proto2 and proto3. +// +// This is disabled by default unless built with the "protolegacy" tag. +// +// WARNING: The compatibility agreement covers nothing provided by this flag. +// As such, functionality may suddenly be removed or changed at our discretion. +const ProtoLegacy = protoLegacy + +// LazyUnmarshalExtensions specifies whether to lazily unmarshal extensions. +// +// Lazy extension unmarshaling validates the contents of message-valued +// extension fields at unmarshal time, but defers creating the message +// structure until the extension is first accessed. +const LazyUnmarshalExtensions = ProtoLegacy diff --git a/vendor/google.golang.org/protobuf/internal/flags/proto_legacy_disable.go b/vendor/google.golang.org/protobuf/internal/flags/proto_legacy_disable.go new file mode 100644 index 00000000..a72995f0 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/flags/proto_legacy_disable.go @@ -0,0 +1,9 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !protolegacy + +package flags + +const protoLegacy = false diff --git a/vendor/google.golang.org/protobuf/internal/flags/proto_legacy_enable.go b/vendor/google.golang.org/protobuf/internal/flags/proto_legacy_enable.go new file mode 100644 index 00000000..772e2f0e --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/flags/proto_legacy_enable.go @@ -0,0 +1,9 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build protolegacy + +package flags + +const protoLegacy = true diff --git a/vendor/google.golang.org/protobuf/internal/genname/name.go b/vendor/google.golang.org/protobuf/internal/genname/name.go new file mode 100644 index 00000000..f45509fb --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/genname/name.go @@ -0,0 +1,25 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package genname contains constants for generated names. +package genname + +const ( + State = "state" + + SizeCache = "sizeCache" + SizeCacheA = "XXX_sizecache" + + WeakFields = "weakFields" + WeakFieldsA = "XXX_weak" + + UnknownFields = "unknownFields" + UnknownFieldsA = "XXX_unrecognized" + + ExtensionFields = "extensionFields" + ExtensionFieldsA = "XXX_InternalExtensions" + ExtensionFieldsB = "XXX_extensions" + + WeakFieldPrefix = "XXX_weak_" +) diff --git a/vendor/google.golang.org/protobuf/internal/impl/api_export.go b/vendor/google.golang.org/protobuf/internal/impl/api_export.go new file mode 100644 index 00000000..4d22c960 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/api_export.go @@ -0,0 +1,170 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + "strconv" + + "google.golang.org/protobuf/encoding/prototext" + "google.golang.org/protobuf/proto" + pref "google.golang.org/protobuf/reflect/protoreflect" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +// Export is a zero-length named type that exists only to export a set of +// functions that we do not want to appear in godoc. +type Export struct{} + +// enum is any enum type generated by protoc-gen-go +// and must be a named int32 type. +type enum = interface{} + +// EnumOf returns the protoreflect.Enum interface over e. +// It returns nil if e is nil. +func (Export) EnumOf(e enum) pref.Enum { + switch e := e.(type) { + case nil: + return nil + case pref.Enum: + return e + default: + return legacyWrapEnum(reflect.ValueOf(e)) + } +} + +// EnumDescriptorOf returns the protoreflect.EnumDescriptor for e. +// It returns nil if e is nil. +func (Export) EnumDescriptorOf(e enum) pref.EnumDescriptor { + switch e := e.(type) { + case nil: + return nil + case pref.Enum: + return e.Descriptor() + default: + return LegacyLoadEnumDesc(reflect.TypeOf(e)) + } +} + +// EnumTypeOf returns the protoreflect.EnumType for e. +// It returns nil if e is nil. +func (Export) EnumTypeOf(e enum) pref.EnumType { + switch e := e.(type) { + case nil: + return nil + case pref.Enum: + return e.Type() + default: + return legacyLoadEnumType(reflect.TypeOf(e)) + } +} + +// EnumStringOf returns the enum value as a string, either as the name if +// the number is resolvable, or the number formatted as a string. +func (Export) EnumStringOf(ed pref.EnumDescriptor, n pref.EnumNumber) string { + ev := ed.Values().ByNumber(n) + if ev != nil { + return string(ev.Name()) + } + return strconv.Itoa(int(n)) +} + +// message is any message type generated by protoc-gen-go +// and must be a pointer to a named struct type. +type message = interface{} + +// legacyMessageWrapper wraps a v2 message as a v1 message. +type legacyMessageWrapper struct{ m pref.ProtoMessage } + +func (m legacyMessageWrapper) Reset() { proto.Reset(m.m) } +func (m legacyMessageWrapper) String() string { return Export{}.MessageStringOf(m.m) } +func (m legacyMessageWrapper) ProtoMessage() {} + +// ProtoMessageV1Of converts either a v1 or v2 message to a v1 message. +// It returns nil if m is nil. +func (Export) ProtoMessageV1Of(m message) piface.MessageV1 { + switch mv := m.(type) { + case nil: + return nil + case piface.MessageV1: + return mv + case unwrapper: + return Export{}.ProtoMessageV1Of(mv.protoUnwrap()) + case pref.ProtoMessage: + return legacyMessageWrapper{mv} + default: + panic(fmt.Sprintf("message %T is neither a v1 or v2 Message", m)) + } +} + +func (Export) protoMessageV2Of(m message) pref.ProtoMessage { + switch mv := m.(type) { + case nil: + return nil + case pref.ProtoMessage: + return mv + case legacyMessageWrapper: + return mv.m + case piface.MessageV1: + return nil + default: + panic(fmt.Sprintf("message %T is neither a v1 or v2 Message", m)) + } +} + +// ProtoMessageV2Of converts either a v1 or v2 message to a v2 message. +// It returns nil if m is nil. +func (Export) ProtoMessageV2Of(m message) pref.ProtoMessage { + if m == nil { + return nil + } + if mv := (Export{}).protoMessageV2Of(m); mv != nil { + return mv + } + return legacyWrapMessage(reflect.ValueOf(m)).Interface() +} + +// MessageOf returns the protoreflect.Message interface over m. +// It returns nil if m is nil. +func (Export) MessageOf(m message) pref.Message { + if m == nil { + return nil + } + if mv := (Export{}).protoMessageV2Of(m); mv != nil { + return mv.ProtoReflect() + } + return legacyWrapMessage(reflect.ValueOf(m)) +} + +// MessageDescriptorOf returns the protoreflect.MessageDescriptor for m. +// It returns nil if m is nil. +func (Export) MessageDescriptorOf(m message) pref.MessageDescriptor { + if m == nil { + return nil + } + if mv := (Export{}).protoMessageV2Of(m); mv != nil { + return mv.ProtoReflect().Descriptor() + } + return LegacyLoadMessageDesc(reflect.TypeOf(m)) +} + +// MessageTypeOf returns the protoreflect.MessageType for m. +// It returns nil if m is nil. +func (Export) MessageTypeOf(m message) pref.MessageType { + if m == nil { + return nil + } + if mv := (Export{}).protoMessageV2Of(m); mv != nil { + return mv.ProtoReflect().Type() + } + return legacyLoadMessageInfo(reflect.TypeOf(m), "") +} + +// MessageStringOf returns the message value as a string, +// which is the message serialized in the protobuf text format. +func (Export) MessageStringOf(m pref.ProtoMessage) string { + return prototext.MarshalOptions{Multiline: false}.Format(m) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/checkinit.go b/vendor/google.golang.org/protobuf/internal/impl/checkinit.go new file mode 100644 index 00000000..b82341e5 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/checkinit.go @@ -0,0 +1,141 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "sync" + + "google.golang.org/protobuf/internal/errors" + pref "google.golang.org/protobuf/reflect/protoreflect" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +func (mi *MessageInfo) checkInitialized(in piface.CheckInitializedInput) (piface.CheckInitializedOutput, error) { + var p pointer + if ms, ok := in.Message.(*messageState); ok { + p = ms.pointer() + } else { + p = in.Message.(*messageReflectWrapper).pointer() + } + return piface.CheckInitializedOutput{}, mi.checkInitializedPointer(p) +} + +func (mi *MessageInfo) checkInitializedPointer(p pointer) error { + mi.init() + if !mi.needsInitCheck { + return nil + } + if p.IsNil() { + for _, f := range mi.orderedCoderFields { + if f.isRequired { + return errors.RequiredNotSet(string(mi.Desc.Fields().ByNumber(f.num).FullName())) + } + } + return nil + } + if mi.extensionOffset.IsValid() { + e := p.Apply(mi.extensionOffset).Extensions() + if err := mi.isInitExtensions(e); err != nil { + return err + } + } + for _, f := range mi.orderedCoderFields { + if !f.isRequired && f.funcs.isInit == nil { + continue + } + fptr := p.Apply(f.offset) + if f.isPointer && fptr.Elem().IsNil() { + if f.isRequired { + return errors.RequiredNotSet(string(mi.Desc.Fields().ByNumber(f.num).FullName())) + } + continue + } + if f.funcs.isInit == nil { + continue + } + if err := f.funcs.isInit(fptr, f); err != nil { + return err + } + } + return nil +} + +func (mi *MessageInfo) isInitExtensions(ext *map[int32]ExtensionField) error { + if ext == nil { + return nil + } + for _, x := range *ext { + ei := getExtensionFieldInfo(x.Type()) + if ei.funcs.isInit == nil { + continue + } + v := x.Value() + if !v.IsValid() { + continue + } + if err := ei.funcs.isInit(v); err != nil { + return err + } + } + return nil +} + +var ( + needsInitCheckMu sync.Mutex + needsInitCheckMap sync.Map +) + +// needsInitCheck reports whether a message needs to be checked for partial initialization. +// +// It returns true if the message transitively includes any required or extension fields. +func needsInitCheck(md pref.MessageDescriptor) bool { + if v, ok := needsInitCheckMap.Load(md); ok { + if has, ok := v.(bool); ok { + return has + } + } + needsInitCheckMu.Lock() + defer needsInitCheckMu.Unlock() + return needsInitCheckLocked(md) +} + +func needsInitCheckLocked(md pref.MessageDescriptor) (has bool) { + if v, ok := needsInitCheckMap.Load(md); ok { + // If has is true, we've previously determined that this message + // needs init checks. + // + // If has is false, we've previously determined that it can never + // be uninitialized. + // + // If has is not a bool, we've just encountered a cycle in the + // message graph. In this case, it is safe to return false: If + // the message does have required fields, we'll detect them later + // in the graph traversal. + has, ok := v.(bool) + return ok && has + } + needsInitCheckMap.Store(md, struct{}{}) // avoid cycles while descending into this message + defer func() { + needsInitCheckMap.Store(md, has) + }() + if md.RequiredNumbers().Len() > 0 { + return true + } + if md.ExtensionRanges().Len() > 0 { + return true + } + for i := 0; i < md.Fields().Len(); i++ { + fd := md.Fields().Get(i) + // Map keys are never messages, so just consider the map value. + if fd.IsMap() { + fd = fd.MapValue() + } + fmd := fd.Message() + if fmd != nil && needsInitCheckLocked(fmd) { + return true + } + } + return false +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_extension.go b/vendor/google.golang.org/protobuf/internal/impl/codec_extension.go new file mode 100644 index 00000000..08d35170 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_extension.go @@ -0,0 +1,223 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "sync" + "sync/atomic" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/errors" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type extensionFieldInfo struct { + wiretag uint64 + tagsize int + unmarshalNeedsValue bool + funcs valueCoderFuncs + validation validationInfo +} + +var legacyExtensionFieldInfoCache sync.Map // map[protoreflect.ExtensionType]*extensionFieldInfo + +func getExtensionFieldInfo(xt pref.ExtensionType) *extensionFieldInfo { + if xi, ok := xt.(*ExtensionInfo); ok { + xi.lazyInit() + return xi.info + } + return legacyLoadExtensionFieldInfo(xt) +} + +// legacyLoadExtensionFieldInfo dynamically loads a *ExtensionInfo for xt. +func legacyLoadExtensionFieldInfo(xt pref.ExtensionType) *extensionFieldInfo { + if xi, ok := legacyExtensionFieldInfoCache.Load(xt); ok { + return xi.(*extensionFieldInfo) + } + e := makeExtensionFieldInfo(xt.TypeDescriptor()) + if e, ok := legacyMessageTypeCache.LoadOrStore(xt, e); ok { + return e.(*extensionFieldInfo) + } + return e +} + +func makeExtensionFieldInfo(xd pref.ExtensionDescriptor) *extensionFieldInfo { + var wiretag uint64 + if !xd.IsPacked() { + wiretag = protowire.EncodeTag(xd.Number(), wireTypes[xd.Kind()]) + } else { + wiretag = protowire.EncodeTag(xd.Number(), protowire.BytesType) + } + e := &extensionFieldInfo{ + wiretag: wiretag, + tagsize: protowire.SizeVarint(wiretag), + funcs: encoderFuncsForValue(xd), + } + // Does the unmarshal function need a value passed to it? + // This is true for composite types, where we pass in a message, list, or map to fill in, + // and for enums, where we pass in a prototype value to specify the concrete enum type. + switch xd.Kind() { + case pref.MessageKind, pref.GroupKind, pref.EnumKind: + e.unmarshalNeedsValue = true + default: + if xd.Cardinality() == pref.Repeated { + e.unmarshalNeedsValue = true + } + } + return e +} + +type lazyExtensionValue struct { + atomicOnce uint32 // atomically set if value is valid + mu sync.Mutex + xi *extensionFieldInfo + value pref.Value + b []byte + fn func() pref.Value +} + +type ExtensionField struct { + typ pref.ExtensionType + + // value is either the value of GetValue, + // or a *lazyExtensionValue that then returns the value of GetValue. + value pref.Value + lazy *lazyExtensionValue +} + +func (f *ExtensionField) appendLazyBytes(xt pref.ExtensionType, xi *extensionFieldInfo, num protowire.Number, wtyp protowire.Type, b []byte) { + if f.lazy == nil { + f.lazy = &lazyExtensionValue{xi: xi} + } + f.typ = xt + f.lazy.xi = xi + f.lazy.b = protowire.AppendTag(f.lazy.b, num, wtyp) + f.lazy.b = append(f.lazy.b, b...) +} + +func (f *ExtensionField) canLazy(xt pref.ExtensionType) bool { + if f.typ == nil { + return true + } + if f.typ == xt && f.lazy != nil && atomic.LoadUint32(&f.lazy.atomicOnce) == 0 { + return true + } + return false +} + +func (f *ExtensionField) lazyInit() { + f.lazy.mu.Lock() + defer f.lazy.mu.Unlock() + if atomic.LoadUint32(&f.lazy.atomicOnce) == 1 { + return + } + if f.lazy.xi != nil { + b := f.lazy.b + val := f.typ.New() + for len(b) > 0 { + var tag uint64 + if b[0] < 0x80 { + tag = uint64(b[0]) + b = b[1:] + } else if len(b) >= 2 && b[1] < 128 { + tag = uint64(b[0]&0x7f) + uint64(b[1])<<7 + b = b[2:] + } else { + var n int + tag, n = protowire.ConsumeVarint(b) + if n < 0 { + panic(errors.New("bad tag in lazy extension decoding")) + } + b = b[n:] + } + num := protowire.Number(tag >> 3) + wtyp := protowire.Type(tag & 7) + var out unmarshalOutput + var err error + val, out, err = f.lazy.xi.funcs.unmarshal(b, val, num, wtyp, lazyUnmarshalOptions) + if err != nil { + panic(errors.New("decode failure in lazy extension decoding: %v", err)) + } + b = b[out.n:] + } + f.lazy.value = val + } else { + f.lazy.value = f.lazy.fn() + } + f.lazy.xi = nil + f.lazy.fn = nil + f.lazy.b = nil + atomic.StoreUint32(&f.lazy.atomicOnce, 1) +} + +// Set sets the type and value of the extension field. +// This must not be called concurrently. +func (f *ExtensionField) Set(t pref.ExtensionType, v pref.Value) { + f.typ = t + f.value = v + f.lazy = nil +} + +// SetLazy sets the type and a value that is to be lazily evaluated upon first use. +// This must not be called concurrently. +func (f *ExtensionField) SetLazy(t pref.ExtensionType, fn func() pref.Value) { + f.typ = t + f.lazy = &lazyExtensionValue{fn: fn} +} + +// Value returns the value of the extension field. +// This may be called concurrently. +func (f *ExtensionField) Value() pref.Value { + if f.lazy != nil { + if atomic.LoadUint32(&f.lazy.atomicOnce) == 0 { + f.lazyInit() + } + return f.lazy.value + } + return f.value +} + +// Type returns the type of the extension field. +// This may be called concurrently. +func (f ExtensionField) Type() pref.ExtensionType { + return f.typ +} + +// IsSet returns whether the extension field is set. +// This may be called concurrently. +func (f ExtensionField) IsSet() bool { + return f.typ != nil +} + +// IsLazy reports whether a field is lazily encoded. +// It is exported for testing. +func IsLazy(m pref.Message, fd pref.FieldDescriptor) bool { + var mi *MessageInfo + var p pointer + switch m := m.(type) { + case *messageState: + mi = m.messageInfo() + p = m.pointer() + case *messageReflectWrapper: + mi = m.messageInfo() + p = m.pointer() + default: + return false + } + xd, ok := fd.(pref.ExtensionTypeDescriptor) + if !ok { + return false + } + xt := xd.Type() + ext := mi.extensionMap(p) + if ext == nil { + return false + } + f, ok := (*ext)[int32(fd.Number())] + if !ok { + return false + } + return f.typ == xt && f.lazy != nil && atomic.LoadUint32(&f.lazy.atomicOnce) == 0 +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_field.go b/vendor/google.golang.org/protobuf/internal/impl/codec_field.go new file mode 100644 index 00000000..c00744d3 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_field.go @@ -0,0 +1,828 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + "sync" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/proto" + pref "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +type errInvalidUTF8 struct{} + +func (errInvalidUTF8) Error() string { return "string field contains invalid UTF-8" } +func (errInvalidUTF8) InvalidUTF8() bool { return true } + +// initOneofFieldCoders initializes the fast-path functions for the fields in a oneof. +// +// For size, marshal, and isInit operations, functions are set only on the first field +// in the oneof. The functions are called when the oneof is non-nil, and will dispatch +// to the appropriate field-specific function as necessary. +// +// The unmarshal function is set on each field individually as usual. +func (mi *MessageInfo) initOneofFieldCoders(od pref.OneofDescriptor, si structInfo) { + fs := si.oneofsByName[od.Name()] + ft := fs.Type + oneofFields := make(map[reflect.Type]*coderFieldInfo) + needIsInit := false + fields := od.Fields() + for i, lim := 0, fields.Len(); i < lim; i++ { + fd := od.Fields().Get(i) + num := fd.Number() + // Make a copy of the original coderFieldInfo for use in unmarshaling. + // + // oneofFields[oneofType].funcs.marshal is the field-specific marshal function. + // + // mi.coderFields[num].marshal is set on only the first field in the oneof, + // and dispatches to the field-specific marshaler in oneofFields. + cf := *mi.coderFields[num] + ot := si.oneofWrappersByNumber[num] + cf.ft = ot.Field(0).Type + cf.mi, cf.funcs = fieldCoder(fd, cf.ft) + oneofFields[ot] = &cf + if cf.funcs.isInit != nil { + needIsInit = true + } + mi.coderFields[num].funcs.unmarshal = func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + var vw reflect.Value // pointer to wrapper type + vi := p.AsValueOf(ft).Elem() // oneof field value of interface kind + if !vi.IsNil() && !vi.Elem().IsNil() && vi.Elem().Elem().Type() == ot { + vw = vi.Elem() + } else { + vw = reflect.New(ot) + } + out, err := cf.funcs.unmarshal(b, pointerOfValue(vw).Apply(zeroOffset), wtyp, &cf, opts) + if err != nil { + return out, err + } + vi.Set(vw) + return out, nil + } + } + getInfo := func(p pointer) (pointer, *coderFieldInfo) { + v := p.AsValueOf(ft).Elem() + if v.IsNil() { + return pointer{}, nil + } + v = v.Elem() // interface -> *struct + if v.IsNil() { + return pointer{}, nil + } + return pointerOfValue(v).Apply(zeroOffset), oneofFields[v.Elem().Type()] + } + first := mi.coderFields[od.Fields().Get(0).Number()] + first.funcs.size = func(p pointer, _ *coderFieldInfo, opts marshalOptions) int { + p, info := getInfo(p) + if info == nil || info.funcs.size == nil { + return 0 + } + return info.funcs.size(p, info, opts) + } + first.funcs.marshal = func(b []byte, p pointer, _ *coderFieldInfo, opts marshalOptions) ([]byte, error) { + p, info := getInfo(p) + if info == nil || info.funcs.marshal == nil { + return b, nil + } + return info.funcs.marshal(b, p, info, opts) + } + first.funcs.merge = func(dst, src pointer, _ *coderFieldInfo, opts mergeOptions) { + srcp, srcinfo := getInfo(src) + if srcinfo == nil || srcinfo.funcs.merge == nil { + return + } + dstp, dstinfo := getInfo(dst) + if dstinfo != srcinfo { + dst.AsValueOf(ft).Elem().Set(reflect.New(src.AsValueOf(ft).Elem().Elem().Elem().Type())) + dstp = pointerOfValue(dst.AsValueOf(ft).Elem().Elem()).Apply(zeroOffset) + } + srcinfo.funcs.merge(dstp, srcp, srcinfo, opts) + } + if needIsInit { + first.funcs.isInit = func(p pointer, _ *coderFieldInfo) error { + p, info := getInfo(p) + if info == nil || info.funcs.isInit == nil { + return nil + } + return info.funcs.isInit(p, info) + } + } +} + +func makeWeakMessageFieldCoder(fd pref.FieldDescriptor) pointerCoderFuncs { + var once sync.Once + var messageType pref.MessageType + lazyInit := func() { + once.Do(func() { + messageName := fd.Message().FullName() + messageType, _ = preg.GlobalTypes.FindMessageByName(messageName) + }) + } + + return pointerCoderFuncs{ + size: func(p pointer, f *coderFieldInfo, opts marshalOptions) int { + m, ok := p.WeakFields().get(f.num) + if !ok { + return 0 + } + lazyInit() + if messageType == nil { + panic(fmt.Sprintf("weak message %v is not linked in", fd.Message().FullName())) + } + return sizeMessage(m, f.tagsize, opts) + }, + marshal: func(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + m, ok := p.WeakFields().get(f.num) + if !ok { + return b, nil + } + lazyInit() + if messageType == nil { + panic(fmt.Sprintf("weak message %v is not linked in", fd.Message().FullName())) + } + return appendMessage(b, m, f.wiretag, opts) + }, + unmarshal: func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + fs := p.WeakFields() + m, ok := fs.get(f.num) + if !ok { + lazyInit() + if messageType == nil { + return unmarshalOutput{}, errUnknown + } + m = messageType.New().Interface() + fs.set(f.num, m) + } + return consumeMessage(b, m, wtyp, opts) + }, + isInit: func(p pointer, f *coderFieldInfo) error { + m, ok := p.WeakFields().get(f.num) + if !ok { + return nil + } + return proto.CheckInitialized(m) + }, + merge: func(dst, src pointer, f *coderFieldInfo, opts mergeOptions) { + sm, ok := src.WeakFields().get(f.num) + if !ok { + return + } + dm, ok := dst.WeakFields().get(f.num) + if !ok { + lazyInit() + if messageType == nil { + panic(fmt.Sprintf("weak message %v is not linked in", fd.Message().FullName())) + } + dm = messageType.New().Interface() + dst.WeakFields().set(f.num, dm) + } + opts.Merge(dm, sm) + }, + } +} + +func makeMessageFieldCoder(fd pref.FieldDescriptor, ft reflect.Type) pointerCoderFuncs { + if mi := getMessageInfo(ft); mi != nil { + funcs := pointerCoderFuncs{ + size: sizeMessageInfo, + marshal: appendMessageInfo, + unmarshal: consumeMessageInfo, + merge: mergeMessage, + } + if needsInitCheck(mi.Desc) { + funcs.isInit = isInitMessageInfo + } + return funcs + } else { + return pointerCoderFuncs{ + size: func(p pointer, f *coderFieldInfo, opts marshalOptions) int { + m := asMessage(p.AsValueOf(ft).Elem()) + return sizeMessage(m, f.tagsize, opts) + }, + marshal: func(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + m := asMessage(p.AsValueOf(ft).Elem()) + return appendMessage(b, m, f.wiretag, opts) + }, + unmarshal: func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + mp := p.AsValueOf(ft).Elem() + if mp.IsNil() { + mp.Set(reflect.New(ft.Elem())) + } + return consumeMessage(b, asMessage(mp), wtyp, opts) + }, + isInit: func(p pointer, f *coderFieldInfo) error { + m := asMessage(p.AsValueOf(ft).Elem()) + return proto.CheckInitialized(m) + }, + merge: mergeMessage, + } + } +} + +func sizeMessageInfo(p pointer, f *coderFieldInfo, opts marshalOptions) int { + return protowire.SizeBytes(f.mi.sizePointer(p.Elem(), opts)) + f.tagsize +} + +func appendMessageInfo(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(f.mi.sizePointer(p.Elem(), opts))) + return f.mi.marshalAppendPointer(b, p.Elem(), opts) +} + +func consumeMessageInfo(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if p.Elem().IsNil() { + p.SetPointer(pointerOfValue(reflect.New(f.mi.GoReflectType.Elem()))) + } + o, err := f.mi.unmarshalPointer(v, p.Elem(), 0, opts) + if err != nil { + return out, err + } + out.n = n + out.initialized = o.initialized + return out, nil +} + +func isInitMessageInfo(p pointer, f *coderFieldInfo) error { + return f.mi.checkInitializedPointer(p.Elem()) +} + +func sizeMessage(m proto.Message, tagsize int, _ marshalOptions) int { + return protowire.SizeBytes(proto.Size(m)) + tagsize +} + +func appendMessage(b []byte, m proto.Message, wiretag uint64, opts marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(proto.Size(m))) + return opts.Options().MarshalAppend(b, m) +} + +func consumeMessage(b []byte, m proto.Message, wtyp protowire.Type, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + o, err := opts.Options().UnmarshalState(piface.UnmarshalInput{ + Buf: v, + Message: m.ProtoReflect(), + }) + if err != nil { + return out, err + } + out.n = n + out.initialized = o.Flags&piface.UnmarshalInitialized != 0 + return out, nil +} + +func sizeMessageValue(v pref.Value, tagsize int, opts marshalOptions) int { + m := v.Message().Interface() + return sizeMessage(m, tagsize, opts) +} + +func appendMessageValue(b []byte, v pref.Value, wiretag uint64, opts marshalOptions) ([]byte, error) { + m := v.Message().Interface() + return appendMessage(b, m, wiretag, opts) +} + +func consumeMessageValue(b []byte, v pref.Value, _ protowire.Number, wtyp protowire.Type, opts unmarshalOptions) (pref.Value, unmarshalOutput, error) { + m := v.Message().Interface() + out, err := consumeMessage(b, m, wtyp, opts) + return v, out, err +} + +func isInitMessageValue(v pref.Value) error { + m := v.Message().Interface() + return proto.CheckInitialized(m) +} + +var coderMessageValue = valueCoderFuncs{ + size: sizeMessageValue, + marshal: appendMessageValue, + unmarshal: consumeMessageValue, + isInit: isInitMessageValue, + merge: mergeMessageValue, +} + +func sizeGroupValue(v pref.Value, tagsize int, opts marshalOptions) int { + m := v.Message().Interface() + return sizeGroup(m, tagsize, opts) +} + +func appendGroupValue(b []byte, v pref.Value, wiretag uint64, opts marshalOptions) ([]byte, error) { + m := v.Message().Interface() + return appendGroup(b, m, wiretag, opts) +} + +func consumeGroupValue(b []byte, v pref.Value, num protowire.Number, wtyp protowire.Type, opts unmarshalOptions) (pref.Value, unmarshalOutput, error) { + m := v.Message().Interface() + out, err := consumeGroup(b, m, num, wtyp, opts) + return v, out, err +} + +var coderGroupValue = valueCoderFuncs{ + size: sizeGroupValue, + marshal: appendGroupValue, + unmarshal: consumeGroupValue, + isInit: isInitMessageValue, + merge: mergeMessageValue, +} + +func makeGroupFieldCoder(fd pref.FieldDescriptor, ft reflect.Type) pointerCoderFuncs { + num := fd.Number() + if mi := getMessageInfo(ft); mi != nil { + funcs := pointerCoderFuncs{ + size: sizeGroupType, + marshal: appendGroupType, + unmarshal: consumeGroupType, + merge: mergeMessage, + } + if needsInitCheck(mi.Desc) { + funcs.isInit = isInitMessageInfo + } + return funcs + } else { + return pointerCoderFuncs{ + size: func(p pointer, f *coderFieldInfo, opts marshalOptions) int { + m := asMessage(p.AsValueOf(ft).Elem()) + return sizeGroup(m, f.tagsize, opts) + }, + marshal: func(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + m := asMessage(p.AsValueOf(ft).Elem()) + return appendGroup(b, m, f.wiretag, opts) + }, + unmarshal: func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + mp := p.AsValueOf(ft).Elem() + if mp.IsNil() { + mp.Set(reflect.New(ft.Elem())) + } + return consumeGroup(b, asMessage(mp), num, wtyp, opts) + }, + isInit: func(p pointer, f *coderFieldInfo) error { + m := asMessage(p.AsValueOf(ft).Elem()) + return proto.CheckInitialized(m) + }, + merge: mergeMessage, + } + } +} + +func sizeGroupType(p pointer, f *coderFieldInfo, opts marshalOptions) int { + return 2*f.tagsize + f.mi.sizePointer(p.Elem(), opts) +} + +func appendGroupType(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, f.wiretag) // start group + b, err := f.mi.marshalAppendPointer(b, p.Elem(), opts) + b = protowire.AppendVarint(b, f.wiretag+1) // end group + return b, err +} + +func consumeGroupType(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.StartGroupType { + return out, errUnknown + } + if p.Elem().IsNil() { + p.SetPointer(pointerOfValue(reflect.New(f.mi.GoReflectType.Elem()))) + } + return f.mi.unmarshalPointer(b, p.Elem(), f.num, opts) +} + +func sizeGroup(m proto.Message, tagsize int, _ marshalOptions) int { + return 2*tagsize + proto.Size(m) +} + +func appendGroup(b []byte, m proto.Message, wiretag uint64, opts marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) // start group + b, err := opts.Options().MarshalAppend(b, m) + b = protowire.AppendVarint(b, wiretag+1) // end group + return b, err +} + +func consumeGroup(b []byte, m proto.Message, num protowire.Number, wtyp protowire.Type, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.StartGroupType { + return out, errUnknown + } + b, n := protowire.ConsumeGroup(num, b) + if n < 0 { + return out, protowire.ParseError(n) + } + o, err := opts.Options().UnmarshalState(piface.UnmarshalInput{ + Buf: b, + Message: m.ProtoReflect(), + }) + if err != nil { + return out, err + } + out.n = n + out.initialized = o.Flags&piface.UnmarshalInitialized != 0 + return out, nil +} + +func makeMessageSliceFieldCoder(fd pref.FieldDescriptor, ft reflect.Type) pointerCoderFuncs { + if mi := getMessageInfo(ft); mi != nil { + funcs := pointerCoderFuncs{ + size: sizeMessageSliceInfo, + marshal: appendMessageSliceInfo, + unmarshal: consumeMessageSliceInfo, + merge: mergeMessageSlice, + } + if needsInitCheck(mi.Desc) { + funcs.isInit = isInitMessageSliceInfo + } + return funcs + } + return pointerCoderFuncs{ + size: func(p pointer, f *coderFieldInfo, opts marshalOptions) int { + return sizeMessageSlice(p, ft, f.tagsize, opts) + }, + marshal: func(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + return appendMessageSlice(b, p, f.wiretag, ft, opts) + }, + unmarshal: func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + return consumeMessageSlice(b, p, ft, wtyp, opts) + }, + isInit: func(p pointer, f *coderFieldInfo) error { + return isInitMessageSlice(p, ft) + }, + merge: mergeMessageSlice, + } +} + +func sizeMessageSliceInfo(p pointer, f *coderFieldInfo, opts marshalOptions) int { + s := p.PointerSlice() + n := 0 + for _, v := range s { + n += protowire.SizeBytes(f.mi.sizePointer(v, opts)) + f.tagsize + } + return n +} + +func appendMessageSliceInfo(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + s := p.PointerSlice() + var err error + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + siz := f.mi.sizePointer(v, opts) + b = protowire.AppendVarint(b, uint64(siz)) + b, err = f.mi.marshalAppendPointer(b, v, opts) + if err != nil { + return b, err + } + } + return b, nil +} + +func consumeMessageSliceInfo(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + m := reflect.New(f.mi.GoReflectType.Elem()).Interface() + mp := pointerOfIface(m) + o, err := f.mi.unmarshalPointer(v, mp, 0, opts) + if err != nil { + return out, err + } + p.AppendPointerSlice(mp) + out.n = n + out.initialized = o.initialized + return out, nil +} + +func isInitMessageSliceInfo(p pointer, f *coderFieldInfo) error { + s := p.PointerSlice() + for _, v := range s { + if err := f.mi.checkInitializedPointer(v); err != nil { + return err + } + } + return nil +} + +func sizeMessageSlice(p pointer, goType reflect.Type, tagsize int, _ marshalOptions) int { + s := p.PointerSlice() + n := 0 + for _, v := range s { + m := asMessage(v.AsValueOf(goType.Elem())) + n += protowire.SizeBytes(proto.Size(m)) + tagsize + } + return n +} + +func appendMessageSlice(b []byte, p pointer, wiretag uint64, goType reflect.Type, opts marshalOptions) ([]byte, error) { + s := p.PointerSlice() + var err error + for _, v := range s { + m := asMessage(v.AsValueOf(goType.Elem())) + b = protowire.AppendVarint(b, wiretag) + siz := proto.Size(m) + b = protowire.AppendVarint(b, uint64(siz)) + b, err = opts.Options().MarshalAppend(b, m) + if err != nil { + return b, err + } + } + return b, nil +} + +func consumeMessageSlice(b []byte, p pointer, goType reflect.Type, wtyp protowire.Type, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + mp := reflect.New(goType.Elem()) + o, err := opts.Options().UnmarshalState(piface.UnmarshalInput{ + Buf: v, + Message: asMessage(mp).ProtoReflect(), + }) + if err != nil { + return out, err + } + p.AppendPointerSlice(pointerOfValue(mp)) + out.n = n + out.initialized = o.Flags&piface.UnmarshalInitialized != 0 + return out, nil +} + +func isInitMessageSlice(p pointer, goType reflect.Type) error { + s := p.PointerSlice() + for _, v := range s { + m := asMessage(v.AsValueOf(goType.Elem())) + if err := proto.CheckInitialized(m); err != nil { + return err + } + } + return nil +} + +// Slices of messages + +func sizeMessageSliceValue(listv pref.Value, tagsize int, opts marshalOptions) int { + list := listv.List() + n := 0 + for i, llen := 0, list.Len(); i < llen; i++ { + m := list.Get(i).Message().Interface() + n += protowire.SizeBytes(proto.Size(m)) + tagsize + } + return n +} + +func appendMessageSliceValue(b []byte, listv pref.Value, wiretag uint64, opts marshalOptions) ([]byte, error) { + list := listv.List() + mopts := opts.Options() + for i, llen := 0, list.Len(); i < llen; i++ { + m := list.Get(i).Message().Interface() + b = protowire.AppendVarint(b, wiretag) + siz := proto.Size(m) + b = protowire.AppendVarint(b, uint64(siz)) + var err error + b, err = mopts.MarshalAppend(b, m) + if err != nil { + return b, err + } + } + return b, nil +} + +func consumeMessageSliceValue(b []byte, listv pref.Value, _ protowire.Number, wtyp protowire.Type, opts unmarshalOptions) (_ pref.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp != protowire.BytesType { + return pref.Value{}, out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return pref.Value{}, out, protowire.ParseError(n) + } + m := list.NewElement() + o, err := opts.Options().UnmarshalState(piface.UnmarshalInput{ + Buf: v, + Message: m.Message(), + }) + if err != nil { + return pref.Value{}, out, err + } + list.Append(m) + out.n = n + out.initialized = o.Flags&piface.UnmarshalInitialized != 0 + return listv, out, nil +} + +func isInitMessageSliceValue(listv pref.Value) error { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + m := list.Get(i).Message().Interface() + if err := proto.CheckInitialized(m); err != nil { + return err + } + } + return nil +} + +var coderMessageSliceValue = valueCoderFuncs{ + size: sizeMessageSliceValue, + marshal: appendMessageSliceValue, + unmarshal: consumeMessageSliceValue, + isInit: isInitMessageSliceValue, + merge: mergeMessageListValue, +} + +func sizeGroupSliceValue(listv pref.Value, tagsize int, opts marshalOptions) int { + list := listv.List() + n := 0 + for i, llen := 0, list.Len(); i < llen; i++ { + m := list.Get(i).Message().Interface() + n += 2*tagsize + proto.Size(m) + } + return n +} + +func appendGroupSliceValue(b []byte, listv pref.Value, wiretag uint64, opts marshalOptions) ([]byte, error) { + list := listv.List() + mopts := opts.Options() + for i, llen := 0, list.Len(); i < llen; i++ { + m := list.Get(i).Message().Interface() + b = protowire.AppendVarint(b, wiretag) // start group + var err error + b, err = mopts.MarshalAppend(b, m) + if err != nil { + return b, err + } + b = protowire.AppendVarint(b, wiretag+1) // end group + } + return b, nil +} + +func consumeGroupSliceValue(b []byte, listv pref.Value, num protowire.Number, wtyp protowire.Type, opts unmarshalOptions) (_ pref.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp != protowire.StartGroupType { + return pref.Value{}, out, errUnknown + } + b, n := protowire.ConsumeGroup(num, b) + if n < 0 { + return pref.Value{}, out, protowire.ParseError(n) + } + m := list.NewElement() + o, err := opts.Options().UnmarshalState(piface.UnmarshalInput{ + Buf: b, + Message: m.Message(), + }) + if err != nil { + return pref.Value{}, out, err + } + list.Append(m) + out.n = n + out.initialized = o.Flags&piface.UnmarshalInitialized != 0 + return listv, out, nil +} + +var coderGroupSliceValue = valueCoderFuncs{ + size: sizeGroupSliceValue, + marshal: appendGroupSliceValue, + unmarshal: consumeGroupSliceValue, + isInit: isInitMessageSliceValue, + merge: mergeMessageListValue, +} + +func makeGroupSliceFieldCoder(fd pref.FieldDescriptor, ft reflect.Type) pointerCoderFuncs { + num := fd.Number() + if mi := getMessageInfo(ft); mi != nil { + funcs := pointerCoderFuncs{ + size: sizeGroupSliceInfo, + marshal: appendGroupSliceInfo, + unmarshal: consumeGroupSliceInfo, + merge: mergeMessageSlice, + } + if needsInitCheck(mi.Desc) { + funcs.isInit = isInitMessageSliceInfo + } + return funcs + } + return pointerCoderFuncs{ + size: func(p pointer, f *coderFieldInfo, opts marshalOptions) int { + return sizeGroupSlice(p, ft, f.tagsize, opts) + }, + marshal: func(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + return appendGroupSlice(b, p, f.wiretag, ft, opts) + }, + unmarshal: func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + return consumeGroupSlice(b, p, num, wtyp, ft, opts) + }, + isInit: func(p pointer, f *coderFieldInfo) error { + return isInitMessageSlice(p, ft) + }, + merge: mergeMessageSlice, + } +} + +func sizeGroupSlice(p pointer, messageType reflect.Type, tagsize int, _ marshalOptions) int { + s := p.PointerSlice() + n := 0 + for _, v := range s { + m := asMessage(v.AsValueOf(messageType.Elem())) + n += 2*tagsize + proto.Size(m) + } + return n +} + +func appendGroupSlice(b []byte, p pointer, wiretag uint64, messageType reflect.Type, opts marshalOptions) ([]byte, error) { + s := p.PointerSlice() + var err error + for _, v := range s { + m := asMessage(v.AsValueOf(messageType.Elem())) + b = protowire.AppendVarint(b, wiretag) // start group + b, err = opts.Options().MarshalAppend(b, m) + if err != nil { + return b, err + } + b = protowire.AppendVarint(b, wiretag+1) // end group + } + return b, nil +} + +func consumeGroupSlice(b []byte, p pointer, num protowire.Number, wtyp protowire.Type, goType reflect.Type, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.StartGroupType { + return out, errUnknown + } + b, n := protowire.ConsumeGroup(num, b) + if n < 0 { + return out, protowire.ParseError(n) + } + mp := reflect.New(goType.Elem()) + o, err := opts.Options().UnmarshalState(piface.UnmarshalInput{ + Buf: b, + Message: asMessage(mp).ProtoReflect(), + }) + if err != nil { + return out, err + } + p.AppendPointerSlice(pointerOfValue(mp)) + out.n = n + out.initialized = o.Flags&piface.UnmarshalInitialized != 0 + return out, nil +} + +func sizeGroupSliceInfo(p pointer, f *coderFieldInfo, opts marshalOptions) int { + s := p.PointerSlice() + n := 0 + for _, v := range s { + n += 2*f.tagsize + f.mi.sizePointer(v, opts) + } + return n +} + +func appendGroupSliceInfo(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + s := p.PointerSlice() + var err error + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) // start group + b, err = f.mi.marshalAppendPointer(b, v, opts) + if err != nil { + return b, err + } + b = protowire.AppendVarint(b, f.wiretag+1) // end group + } + return b, nil +} + +func consumeGroupSliceInfo(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + if wtyp != protowire.StartGroupType { + return unmarshalOutput{}, errUnknown + } + m := reflect.New(f.mi.GoReflectType.Elem()).Interface() + mp := pointerOfIface(m) + out, err := f.mi.unmarshalPointer(b, mp, f.num, opts) + if err != nil { + return out, err + } + p.AppendPointerSlice(mp) + return out, nil +} + +func asMessage(v reflect.Value) pref.ProtoMessage { + if m, ok := v.Interface().(pref.ProtoMessage); ok { + return m + } + return legacyWrapMessage(v).Interface() +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_gen.go b/vendor/google.golang.org/protobuf/internal/impl/codec_gen.go new file mode 100644 index 00000000..ff198d0a --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_gen.go @@ -0,0 +1,5637 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-types. DO NOT EDIT. + +package impl + +import ( + "math" + "unicode/utf8" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/reflect/protoreflect" +) + +// sizeBool returns the size of wire encoding a bool pointer as a Bool. +func sizeBool(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Bool() + return f.tagsize + protowire.SizeVarint(protowire.EncodeBool(v)) +} + +// appendBool wire encodes a bool pointer as a Bool. +func appendBool(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Bool() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeBool(v)) + return b, nil +} + +// consumeBool wire decodes a bool pointer as a Bool. +func consumeBool(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Bool() = protowire.DecodeBool(v) + out.n = n + return out, nil +} + +var coderBool = pointerCoderFuncs{ + size: sizeBool, + marshal: appendBool, + unmarshal: consumeBool, + merge: mergeBool, +} + +// sizeBoolNoZero returns the size of wire encoding a bool pointer as a Bool. +// The zero value is not encoded. +func sizeBoolNoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Bool() + if v == false { + return 0 + } + return f.tagsize + protowire.SizeVarint(protowire.EncodeBool(v)) +} + +// appendBoolNoZero wire encodes a bool pointer as a Bool. +// The zero value is not encoded. +func appendBoolNoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Bool() + if v == false { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeBool(v)) + return b, nil +} + +var coderBoolNoZero = pointerCoderFuncs{ + size: sizeBoolNoZero, + marshal: appendBoolNoZero, + unmarshal: consumeBool, + merge: mergeBoolNoZero, +} + +// sizeBoolPtr returns the size of wire encoding a *bool pointer as a Bool. +// It panics if the pointer is nil. +func sizeBoolPtr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.BoolPtr() + return f.tagsize + protowire.SizeVarint(protowire.EncodeBool(v)) +} + +// appendBoolPtr wire encodes a *bool pointer as a Bool. +// It panics if the pointer is nil. +func appendBoolPtr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.BoolPtr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeBool(v)) + return b, nil +} + +// consumeBoolPtr wire decodes a *bool pointer as a Bool. +func consumeBoolPtr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.BoolPtr() + if *vp == nil { + *vp = new(bool) + } + **vp = protowire.DecodeBool(v) + out.n = n + return out, nil +} + +var coderBoolPtr = pointerCoderFuncs{ + size: sizeBoolPtr, + marshal: appendBoolPtr, + unmarshal: consumeBoolPtr, + merge: mergeBoolPtr, +} + +// sizeBoolSlice returns the size of wire encoding a []bool pointer as a repeated Bool. +func sizeBoolSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.BoolSlice() + for _, v := range s { + size += f.tagsize + protowire.SizeVarint(protowire.EncodeBool(v)) + } + return size +} + +// appendBoolSlice encodes a []bool pointer as a repeated Bool. +func appendBoolSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.BoolSlice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeBool(v)) + } + return b, nil +} + +// consumeBoolSlice wire decodes a []bool pointer as a repeated Bool. +func consumeBoolSlice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.BoolSlice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, protowire.DecodeBool(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, protowire.DecodeBool(v)) + out.n = n + return out, nil +} + +var coderBoolSlice = pointerCoderFuncs{ + size: sizeBoolSlice, + marshal: appendBoolSlice, + unmarshal: consumeBoolSlice, + merge: mergeBoolSlice, +} + +// sizeBoolPackedSlice returns the size of wire encoding a []bool pointer as a packed repeated Bool. +func sizeBoolPackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.BoolSlice() + if len(s) == 0 { + return 0 + } + n := 0 + for _, v := range s { + n += protowire.SizeVarint(protowire.EncodeBool(v)) + } + return f.tagsize + protowire.SizeBytes(n) +} + +// appendBoolPackedSlice encodes a []bool pointer as a packed repeated Bool. +func appendBoolPackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.BoolSlice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for _, v := range s { + n += protowire.SizeVarint(protowire.EncodeBool(v)) + } + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendVarint(b, protowire.EncodeBool(v)) + } + return b, nil +} + +var coderBoolPackedSlice = pointerCoderFuncs{ + size: sizeBoolPackedSlice, + marshal: appendBoolPackedSlice, + unmarshal: consumeBoolSlice, + merge: mergeBoolSlice, +} + +// sizeBoolValue returns the size of wire encoding a bool value as a Bool. +func sizeBoolValue(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(protowire.EncodeBool(v.Bool())) +} + +// appendBoolValue encodes a bool value as a Bool. +func appendBoolValue(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, protowire.EncodeBool(v.Bool())) + return b, nil +} + +// consumeBoolValue decodes a bool value as a Bool. +func consumeBoolValue(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfBool(protowire.DecodeBool(v)), out, nil +} + +var coderBoolValue = valueCoderFuncs{ + size: sizeBoolValue, + marshal: appendBoolValue, + unmarshal: consumeBoolValue, + merge: mergeScalarValue, +} + +// sizeBoolSliceValue returns the size of wire encoding a []bool value as a repeated Bool. +func sizeBoolSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(protowire.EncodeBool(v.Bool())) + } + return size +} + +// appendBoolSliceValue encodes a []bool value as a repeated Bool. +func appendBoolSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, protowire.EncodeBool(v.Bool())) + } + return b, nil +} + +// consumeBoolSliceValue wire decodes a []bool value as a repeated Bool. +func consumeBoolSliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfBool(protowire.DecodeBool(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfBool(protowire.DecodeBool(v))) + out.n = n + return listv, out, nil +} + +var coderBoolSliceValue = valueCoderFuncs{ + size: sizeBoolSliceValue, + marshal: appendBoolSliceValue, + unmarshal: consumeBoolSliceValue, + merge: mergeListValue, +} + +// sizeBoolPackedSliceValue returns the size of wire encoding a []bool value as a packed repeated Bool. +func sizeBoolPackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(protowire.EncodeBool(v.Bool())) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendBoolPackedSliceValue encodes a []bool value as a packed repeated Bool. +func appendBoolPackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(protowire.EncodeBool(v.Bool())) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, protowire.EncodeBool(v.Bool())) + } + return b, nil +} + +var coderBoolPackedSliceValue = valueCoderFuncs{ + size: sizeBoolPackedSliceValue, + marshal: appendBoolPackedSliceValue, + unmarshal: consumeBoolSliceValue, + merge: mergeListValue, +} + +// sizeEnumValue returns the size of wire encoding a value as a Enum. +func sizeEnumValue(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(uint64(v.Enum())) +} + +// appendEnumValue encodes a value as a Enum. +func appendEnumValue(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(v.Enum())) + return b, nil +} + +// consumeEnumValue decodes a value as a Enum. +func consumeEnumValue(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfEnum(protoreflect.EnumNumber(v)), out, nil +} + +var coderEnumValue = valueCoderFuncs{ + size: sizeEnumValue, + marshal: appendEnumValue, + unmarshal: consumeEnumValue, + merge: mergeScalarValue, +} + +// sizeEnumSliceValue returns the size of wire encoding a [] value as a repeated Enum. +func sizeEnumSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(uint64(v.Enum())) + } + return size +} + +// appendEnumSliceValue encodes a [] value as a repeated Enum. +func appendEnumSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(v.Enum())) + } + return b, nil +} + +// consumeEnumSliceValue wire decodes a [] value as a repeated Enum. +func consumeEnumSliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfEnum(protoreflect.EnumNumber(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfEnum(protoreflect.EnumNumber(v))) + out.n = n + return listv, out, nil +} + +var coderEnumSliceValue = valueCoderFuncs{ + size: sizeEnumSliceValue, + marshal: appendEnumSliceValue, + unmarshal: consumeEnumSliceValue, + merge: mergeListValue, +} + +// sizeEnumPackedSliceValue returns the size of wire encoding a [] value as a packed repeated Enum. +func sizeEnumPackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(v.Enum())) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendEnumPackedSliceValue encodes a [] value as a packed repeated Enum. +func appendEnumPackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(v.Enum())) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, uint64(v.Enum())) + } + return b, nil +} + +var coderEnumPackedSliceValue = valueCoderFuncs{ + size: sizeEnumPackedSliceValue, + marshal: appendEnumPackedSliceValue, + unmarshal: consumeEnumSliceValue, + merge: mergeListValue, +} + +// sizeInt32 returns the size of wire encoding a int32 pointer as a Int32. +func sizeInt32(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int32() + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendInt32 wire encodes a int32 pointer as a Int32. +func appendInt32(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int32() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +// consumeInt32 wire decodes a int32 pointer as a Int32. +func consumeInt32(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Int32() = int32(v) + out.n = n + return out, nil +} + +var coderInt32 = pointerCoderFuncs{ + size: sizeInt32, + marshal: appendInt32, + unmarshal: consumeInt32, + merge: mergeInt32, +} + +// sizeInt32NoZero returns the size of wire encoding a int32 pointer as a Int32. +// The zero value is not encoded. +func sizeInt32NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int32() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendInt32NoZero wire encodes a int32 pointer as a Int32. +// The zero value is not encoded. +func appendInt32NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int32() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +var coderInt32NoZero = pointerCoderFuncs{ + size: sizeInt32NoZero, + marshal: appendInt32NoZero, + unmarshal: consumeInt32, + merge: mergeInt32NoZero, +} + +// sizeInt32Ptr returns the size of wire encoding a *int32 pointer as a Int32. +// It panics if the pointer is nil. +func sizeInt32Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.Int32Ptr() + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendInt32Ptr wire encodes a *int32 pointer as a Int32. +// It panics if the pointer is nil. +func appendInt32Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Int32Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +// consumeInt32Ptr wire decodes a *int32 pointer as a Int32. +func consumeInt32Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Int32Ptr() + if *vp == nil { + *vp = new(int32) + } + **vp = int32(v) + out.n = n + return out, nil +} + +var coderInt32Ptr = pointerCoderFuncs{ + size: sizeInt32Ptr, + marshal: appendInt32Ptr, + unmarshal: consumeInt32Ptr, + merge: mergeInt32Ptr, +} + +// sizeInt32Slice returns the size of wire encoding a []int32 pointer as a repeated Int32. +func sizeInt32Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int32Slice() + for _, v := range s { + size += f.tagsize + protowire.SizeVarint(uint64(v)) + } + return size +} + +// appendInt32Slice encodes a []int32 pointer as a repeated Int32. +func appendInt32Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int32Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + } + return b, nil +} + +// consumeInt32Slice wire decodes a []int32 pointer as a repeated Int32. +func consumeInt32Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Int32Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, int32(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, int32(v)) + out.n = n + return out, nil +} + +var coderInt32Slice = pointerCoderFuncs{ + size: sizeInt32Slice, + marshal: appendInt32Slice, + unmarshal: consumeInt32Slice, + merge: mergeInt32Slice, +} + +// sizeInt32PackedSlice returns the size of wire encoding a []int32 pointer as a packed repeated Int32. +func sizeInt32PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int32Slice() + if len(s) == 0 { + return 0 + } + n := 0 + for _, v := range s { + n += protowire.SizeVarint(uint64(v)) + } + return f.tagsize + protowire.SizeBytes(n) +} + +// appendInt32PackedSlice encodes a []int32 pointer as a packed repeated Int32. +func appendInt32PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int32Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for _, v := range s { + n += protowire.SizeVarint(uint64(v)) + } + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendVarint(b, uint64(v)) + } + return b, nil +} + +var coderInt32PackedSlice = pointerCoderFuncs{ + size: sizeInt32PackedSlice, + marshal: appendInt32PackedSlice, + unmarshal: consumeInt32Slice, + merge: mergeInt32Slice, +} + +// sizeInt32Value returns the size of wire encoding a int32 value as a Int32. +func sizeInt32Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(uint64(int32(v.Int()))) +} + +// appendInt32Value encodes a int32 value as a Int32. +func appendInt32Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(int32(v.Int()))) + return b, nil +} + +// consumeInt32Value decodes a int32 value as a Int32. +func consumeInt32Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfInt32(int32(v)), out, nil +} + +var coderInt32Value = valueCoderFuncs{ + size: sizeInt32Value, + marshal: appendInt32Value, + unmarshal: consumeInt32Value, + merge: mergeScalarValue, +} + +// sizeInt32SliceValue returns the size of wire encoding a []int32 value as a repeated Int32. +func sizeInt32SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(uint64(int32(v.Int()))) + } + return size +} + +// appendInt32SliceValue encodes a []int32 value as a repeated Int32. +func appendInt32SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(int32(v.Int()))) + } + return b, nil +} + +// consumeInt32SliceValue wire decodes a []int32 value as a repeated Int32. +func consumeInt32SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(v))) + out.n = n + return listv, out, nil +} + +var coderInt32SliceValue = valueCoderFuncs{ + size: sizeInt32SliceValue, + marshal: appendInt32SliceValue, + unmarshal: consumeInt32SliceValue, + merge: mergeListValue, +} + +// sizeInt32PackedSliceValue returns the size of wire encoding a []int32 value as a packed repeated Int32. +func sizeInt32PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(int32(v.Int()))) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendInt32PackedSliceValue encodes a []int32 value as a packed repeated Int32. +func appendInt32PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(int32(v.Int()))) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, uint64(int32(v.Int()))) + } + return b, nil +} + +var coderInt32PackedSliceValue = valueCoderFuncs{ + size: sizeInt32PackedSliceValue, + marshal: appendInt32PackedSliceValue, + unmarshal: consumeInt32SliceValue, + merge: mergeListValue, +} + +// sizeSint32 returns the size of wire encoding a int32 pointer as a Sint32. +func sizeSint32(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int32() + return f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(int64(v))) +} + +// appendSint32 wire encodes a int32 pointer as a Sint32. +func appendSint32(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int32() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(v))) + return b, nil +} + +// consumeSint32 wire decodes a int32 pointer as a Sint32. +func consumeSint32(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Int32() = int32(protowire.DecodeZigZag(v & math.MaxUint32)) + out.n = n + return out, nil +} + +var coderSint32 = pointerCoderFuncs{ + size: sizeSint32, + marshal: appendSint32, + unmarshal: consumeSint32, + merge: mergeInt32, +} + +// sizeSint32NoZero returns the size of wire encoding a int32 pointer as a Sint32. +// The zero value is not encoded. +func sizeSint32NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int32() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(int64(v))) +} + +// appendSint32NoZero wire encodes a int32 pointer as a Sint32. +// The zero value is not encoded. +func appendSint32NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int32() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(v))) + return b, nil +} + +var coderSint32NoZero = pointerCoderFuncs{ + size: sizeSint32NoZero, + marshal: appendSint32NoZero, + unmarshal: consumeSint32, + merge: mergeInt32NoZero, +} + +// sizeSint32Ptr returns the size of wire encoding a *int32 pointer as a Sint32. +// It panics if the pointer is nil. +func sizeSint32Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.Int32Ptr() + return f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(int64(v))) +} + +// appendSint32Ptr wire encodes a *int32 pointer as a Sint32. +// It panics if the pointer is nil. +func appendSint32Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Int32Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(v))) + return b, nil +} + +// consumeSint32Ptr wire decodes a *int32 pointer as a Sint32. +func consumeSint32Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Int32Ptr() + if *vp == nil { + *vp = new(int32) + } + **vp = int32(protowire.DecodeZigZag(v & math.MaxUint32)) + out.n = n + return out, nil +} + +var coderSint32Ptr = pointerCoderFuncs{ + size: sizeSint32Ptr, + marshal: appendSint32Ptr, + unmarshal: consumeSint32Ptr, + merge: mergeInt32Ptr, +} + +// sizeSint32Slice returns the size of wire encoding a []int32 pointer as a repeated Sint32. +func sizeSint32Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int32Slice() + for _, v := range s { + size += f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(int64(v))) + } + return size +} + +// appendSint32Slice encodes a []int32 pointer as a repeated Sint32. +func appendSint32Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int32Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(v))) + } + return b, nil +} + +// consumeSint32Slice wire decodes a []int32 pointer as a repeated Sint32. +func consumeSint32Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Int32Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, int32(protowire.DecodeZigZag(v&math.MaxUint32))) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, int32(protowire.DecodeZigZag(v&math.MaxUint32))) + out.n = n + return out, nil +} + +var coderSint32Slice = pointerCoderFuncs{ + size: sizeSint32Slice, + marshal: appendSint32Slice, + unmarshal: consumeSint32Slice, + merge: mergeInt32Slice, +} + +// sizeSint32PackedSlice returns the size of wire encoding a []int32 pointer as a packed repeated Sint32. +func sizeSint32PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int32Slice() + if len(s) == 0 { + return 0 + } + n := 0 + for _, v := range s { + n += protowire.SizeVarint(protowire.EncodeZigZag(int64(v))) + } + return f.tagsize + protowire.SizeBytes(n) +} + +// appendSint32PackedSlice encodes a []int32 pointer as a packed repeated Sint32. +func appendSint32PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int32Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for _, v := range s { + n += protowire.SizeVarint(protowire.EncodeZigZag(int64(v))) + } + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(v))) + } + return b, nil +} + +var coderSint32PackedSlice = pointerCoderFuncs{ + size: sizeSint32PackedSlice, + marshal: appendSint32PackedSlice, + unmarshal: consumeSint32Slice, + merge: mergeInt32Slice, +} + +// sizeSint32Value returns the size of wire encoding a int32 value as a Sint32. +func sizeSint32Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(protowire.EncodeZigZag(int64(int32(v.Int())))) +} + +// appendSint32Value encodes a int32 value as a Sint32. +func appendSint32Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(int32(v.Int())))) + return b, nil +} + +// consumeSint32Value decodes a int32 value as a Sint32. +func consumeSint32Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfInt32(int32(protowire.DecodeZigZag(v & math.MaxUint32))), out, nil +} + +var coderSint32Value = valueCoderFuncs{ + size: sizeSint32Value, + marshal: appendSint32Value, + unmarshal: consumeSint32Value, + merge: mergeScalarValue, +} + +// sizeSint32SliceValue returns the size of wire encoding a []int32 value as a repeated Sint32. +func sizeSint32SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(protowire.EncodeZigZag(int64(int32(v.Int())))) + } + return size +} + +// appendSint32SliceValue encodes a []int32 value as a repeated Sint32. +func appendSint32SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(int32(v.Int())))) + } + return b, nil +} + +// consumeSint32SliceValue wire decodes a []int32 value as a repeated Sint32. +func consumeSint32SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(protowire.DecodeZigZag(v & math.MaxUint32)))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(protowire.DecodeZigZag(v & math.MaxUint32)))) + out.n = n + return listv, out, nil +} + +var coderSint32SliceValue = valueCoderFuncs{ + size: sizeSint32SliceValue, + marshal: appendSint32SliceValue, + unmarshal: consumeSint32SliceValue, + merge: mergeListValue, +} + +// sizeSint32PackedSliceValue returns the size of wire encoding a []int32 value as a packed repeated Sint32. +func sizeSint32PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(protowire.EncodeZigZag(int64(int32(v.Int())))) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendSint32PackedSliceValue encodes a []int32 value as a packed repeated Sint32. +func appendSint32PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(protowire.EncodeZigZag(int64(int32(v.Int())))) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(int32(v.Int())))) + } + return b, nil +} + +var coderSint32PackedSliceValue = valueCoderFuncs{ + size: sizeSint32PackedSliceValue, + marshal: appendSint32PackedSliceValue, + unmarshal: consumeSint32SliceValue, + merge: mergeListValue, +} + +// sizeUint32 returns the size of wire encoding a uint32 pointer as a Uint32. +func sizeUint32(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Uint32() + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendUint32 wire encodes a uint32 pointer as a Uint32. +func appendUint32(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint32() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +// consumeUint32 wire decodes a uint32 pointer as a Uint32. +func consumeUint32(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Uint32() = uint32(v) + out.n = n + return out, nil +} + +var coderUint32 = pointerCoderFuncs{ + size: sizeUint32, + marshal: appendUint32, + unmarshal: consumeUint32, + merge: mergeUint32, +} + +// sizeUint32NoZero returns the size of wire encoding a uint32 pointer as a Uint32. +// The zero value is not encoded. +func sizeUint32NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Uint32() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendUint32NoZero wire encodes a uint32 pointer as a Uint32. +// The zero value is not encoded. +func appendUint32NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint32() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +var coderUint32NoZero = pointerCoderFuncs{ + size: sizeUint32NoZero, + marshal: appendUint32NoZero, + unmarshal: consumeUint32, + merge: mergeUint32NoZero, +} + +// sizeUint32Ptr returns the size of wire encoding a *uint32 pointer as a Uint32. +// It panics if the pointer is nil. +func sizeUint32Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.Uint32Ptr() + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendUint32Ptr wire encodes a *uint32 pointer as a Uint32. +// It panics if the pointer is nil. +func appendUint32Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Uint32Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +// consumeUint32Ptr wire decodes a *uint32 pointer as a Uint32. +func consumeUint32Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Uint32Ptr() + if *vp == nil { + *vp = new(uint32) + } + **vp = uint32(v) + out.n = n + return out, nil +} + +var coderUint32Ptr = pointerCoderFuncs{ + size: sizeUint32Ptr, + marshal: appendUint32Ptr, + unmarshal: consumeUint32Ptr, + merge: mergeUint32Ptr, +} + +// sizeUint32Slice returns the size of wire encoding a []uint32 pointer as a repeated Uint32. +func sizeUint32Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint32Slice() + for _, v := range s { + size += f.tagsize + protowire.SizeVarint(uint64(v)) + } + return size +} + +// appendUint32Slice encodes a []uint32 pointer as a repeated Uint32. +func appendUint32Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint32Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + } + return b, nil +} + +// consumeUint32Slice wire decodes a []uint32 pointer as a repeated Uint32. +func consumeUint32Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Uint32Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, uint32(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, uint32(v)) + out.n = n + return out, nil +} + +var coderUint32Slice = pointerCoderFuncs{ + size: sizeUint32Slice, + marshal: appendUint32Slice, + unmarshal: consumeUint32Slice, + merge: mergeUint32Slice, +} + +// sizeUint32PackedSlice returns the size of wire encoding a []uint32 pointer as a packed repeated Uint32. +func sizeUint32PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint32Slice() + if len(s) == 0 { + return 0 + } + n := 0 + for _, v := range s { + n += protowire.SizeVarint(uint64(v)) + } + return f.tagsize + protowire.SizeBytes(n) +} + +// appendUint32PackedSlice encodes a []uint32 pointer as a packed repeated Uint32. +func appendUint32PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint32Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for _, v := range s { + n += protowire.SizeVarint(uint64(v)) + } + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendVarint(b, uint64(v)) + } + return b, nil +} + +var coderUint32PackedSlice = pointerCoderFuncs{ + size: sizeUint32PackedSlice, + marshal: appendUint32PackedSlice, + unmarshal: consumeUint32Slice, + merge: mergeUint32Slice, +} + +// sizeUint32Value returns the size of wire encoding a uint32 value as a Uint32. +func sizeUint32Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(uint64(uint32(v.Uint()))) +} + +// appendUint32Value encodes a uint32 value as a Uint32. +func appendUint32Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(uint32(v.Uint()))) + return b, nil +} + +// consumeUint32Value decodes a uint32 value as a Uint32. +func consumeUint32Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfUint32(uint32(v)), out, nil +} + +var coderUint32Value = valueCoderFuncs{ + size: sizeUint32Value, + marshal: appendUint32Value, + unmarshal: consumeUint32Value, + merge: mergeScalarValue, +} + +// sizeUint32SliceValue returns the size of wire encoding a []uint32 value as a repeated Uint32. +func sizeUint32SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(uint64(uint32(v.Uint()))) + } + return size +} + +// appendUint32SliceValue encodes a []uint32 value as a repeated Uint32. +func appendUint32SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(uint32(v.Uint()))) + } + return b, nil +} + +// consumeUint32SliceValue wire decodes a []uint32 value as a repeated Uint32. +func consumeUint32SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint32(uint32(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint32(uint32(v))) + out.n = n + return listv, out, nil +} + +var coderUint32SliceValue = valueCoderFuncs{ + size: sizeUint32SliceValue, + marshal: appendUint32SliceValue, + unmarshal: consumeUint32SliceValue, + merge: mergeListValue, +} + +// sizeUint32PackedSliceValue returns the size of wire encoding a []uint32 value as a packed repeated Uint32. +func sizeUint32PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(uint32(v.Uint()))) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendUint32PackedSliceValue encodes a []uint32 value as a packed repeated Uint32. +func appendUint32PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(uint32(v.Uint()))) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, uint64(uint32(v.Uint()))) + } + return b, nil +} + +var coderUint32PackedSliceValue = valueCoderFuncs{ + size: sizeUint32PackedSliceValue, + marshal: appendUint32PackedSliceValue, + unmarshal: consumeUint32SliceValue, + merge: mergeListValue, +} + +// sizeInt64 returns the size of wire encoding a int64 pointer as a Int64. +func sizeInt64(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int64() + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendInt64 wire encodes a int64 pointer as a Int64. +func appendInt64(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int64() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +// consumeInt64 wire decodes a int64 pointer as a Int64. +func consumeInt64(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Int64() = int64(v) + out.n = n + return out, nil +} + +var coderInt64 = pointerCoderFuncs{ + size: sizeInt64, + marshal: appendInt64, + unmarshal: consumeInt64, + merge: mergeInt64, +} + +// sizeInt64NoZero returns the size of wire encoding a int64 pointer as a Int64. +// The zero value is not encoded. +func sizeInt64NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int64() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendInt64NoZero wire encodes a int64 pointer as a Int64. +// The zero value is not encoded. +func appendInt64NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int64() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +var coderInt64NoZero = pointerCoderFuncs{ + size: sizeInt64NoZero, + marshal: appendInt64NoZero, + unmarshal: consumeInt64, + merge: mergeInt64NoZero, +} + +// sizeInt64Ptr returns the size of wire encoding a *int64 pointer as a Int64. +// It panics if the pointer is nil. +func sizeInt64Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.Int64Ptr() + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +// appendInt64Ptr wire encodes a *int64 pointer as a Int64. +// It panics if the pointer is nil. +func appendInt64Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Int64Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +// consumeInt64Ptr wire decodes a *int64 pointer as a Int64. +func consumeInt64Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Int64Ptr() + if *vp == nil { + *vp = new(int64) + } + **vp = int64(v) + out.n = n + return out, nil +} + +var coderInt64Ptr = pointerCoderFuncs{ + size: sizeInt64Ptr, + marshal: appendInt64Ptr, + unmarshal: consumeInt64Ptr, + merge: mergeInt64Ptr, +} + +// sizeInt64Slice returns the size of wire encoding a []int64 pointer as a repeated Int64. +func sizeInt64Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int64Slice() + for _, v := range s { + size += f.tagsize + protowire.SizeVarint(uint64(v)) + } + return size +} + +// appendInt64Slice encodes a []int64 pointer as a repeated Int64. +func appendInt64Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int64Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + } + return b, nil +} + +// consumeInt64Slice wire decodes a []int64 pointer as a repeated Int64. +func consumeInt64Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Int64Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, int64(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, int64(v)) + out.n = n + return out, nil +} + +var coderInt64Slice = pointerCoderFuncs{ + size: sizeInt64Slice, + marshal: appendInt64Slice, + unmarshal: consumeInt64Slice, + merge: mergeInt64Slice, +} + +// sizeInt64PackedSlice returns the size of wire encoding a []int64 pointer as a packed repeated Int64. +func sizeInt64PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int64Slice() + if len(s) == 0 { + return 0 + } + n := 0 + for _, v := range s { + n += protowire.SizeVarint(uint64(v)) + } + return f.tagsize + protowire.SizeBytes(n) +} + +// appendInt64PackedSlice encodes a []int64 pointer as a packed repeated Int64. +func appendInt64PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int64Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for _, v := range s { + n += protowire.SizeVarint(uint64(v)) + } + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendVarint(b, uint64(v)) + } + return b, nil +} + +var coderInt64PackedSlice = pointerCoderFuncs{ + size: sizeInt64PackedSlice, + marshal: appendInt64PackedSlice, + unmarshal: consumeInt64Slice, + merge: mergeInt64Slice, +} + +// sizeInt64Value returns the size of wire encoding a int64 value as a Int64. +func sizeInt64Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(uint64(v.Int())) +} + +// appendInt64Value encodes a int64 value as a Int64. +func appendInt64Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(v.Int())) + return b, nil +} + +// consumeInt64Value decodes a int64 value as a Int64. +func consumeInt64Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfInt64(int64(v)), out, nil +} + +var coderInt64Value = valueCoderFuncs{ + size: sizeInt64Value, + marshal: appendInt64Value, + unmarshal: consumeInt64Value, + merge: mergeScalarValue, +} + +// sizeInt64SliceValue returns the size of wire encoding a []int64 value as a repeated Int64. +func sizeInt64SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(uint64(v.Int())) + } + return size +} + +// appendInt64SliceValue encodes a []int64 value as a repeated Int64. +func appendInt64SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, uint64(v.Int())) + } + return b, nil +} + +// consumeInt64SliceValue wire decodes a []int64 value as a repeated Int64. +func consumeInt64SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(int64(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(int64(v))) + out.n = n + return listv, out, nil +} + +var coderInt64SliceValue = valueCoderFuncs{ + size: sizeInt64SliceValue, + marshal: appendInt64SliceValue, + unmarshal: consumeInt64SliceValue, + merge: mergeListValue, +} + +// sizeInt64PackedSliceValue returns the size of wire encoding a []int64 value as a packed repeated Int64. +func sizeInt64PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(v.Int())) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendInt64PackedSliceValue encodes a []int64 value as a packed repeated Int64. +func appendInt64PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(uint64(v.Int())) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, uint64(v.Int())) + } + return b, nil +} + +var coderInt64PackedSliceValue = valueCoderFuncs{ + size: sizeInt64PackedSliceValue, + marshal: appendInt64PackedSliceValue, + unmarshal: consumeInt64SliceValue, + merge: mergeListValue, +} + +// sizeSint64 returns the size of wire encoding a int64 pointer as a Sint64. +func sizeSint64(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int64() + return f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(v)) +} + +// appendSint64 wire encodes a int64 pointer as a Sint64. +func appendSint64(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int64() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v)) + return b, nil +} + +// consumeSint64 wire decodes a int64 pointer as a Sint64. +func consumeSint64(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Int64() = protowire.DecodeZigZag(v) + out.n = n + return out, nil +} + +var coderSint64 = pointerCoderFuncs{ + size: sizeSint64, + marshal: appendSint64, + unmarshal: consumeSint64, + merge: mergeInt64, +} + +// sizeSint64NoZero returns the size of wire encoding a int64 pointer as a Sint64. +// The zero value is not encoded. +func sizeSint64NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int64() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(v)) +} + +// appendSint64NoZero wire encodes a int64 pointer as a Sint64. +// The zero value is not encoded. +func appendSint64NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int64() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v)) + return b, nil +} + +var coderSint64NoZero = pointerCoderFuncs{ + size: sizeSint64NoZero, + marshal: appendSint64NoZero, + unmarshal: consumeSint64, + merge: mergeInt64NoZero, +} + +// sizeSint64Ptr returns the size of wire encoding a *int64 pointer as a Sint64. +// It panics if the pointer is nil. +func sizeSint64Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.Int64Ptr() + return f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(v)) +} + +// appendSint64Ptr wire encodes a *int64 pointer as a Sint64. +// It panics if the pointer is nil. +func appendSint64Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Int64Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v)) + return b, nil +} + +// consumeSint64Ptr wire decodes a *int64 pointer as a Sint64. +func consumeSint64Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Int64Ptr() + if *vp == nil { + *vp = new(int64) + } + **vp = protowire.DecodeZigZag(v) + out.n = n + return out, nil +} + +var coderSint64Ptr = pointerCoderFuncs{ + size: sizeSint64Ptr, + marshal: appendSint64Ptr, + unmarshal: consumeSint64Ptr, + merge: mergeInt64Ptr, +} + +// sizeSint64Slice returns the size of wire encoding a []int64 pointer as a repeated Sint64. +func sizeSint64Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int64Slice() + for _, v := range s { + size += f.tagsize + protowire.SizeVarint(protowire.EncodeZigZag(v)) + } + return size +} + +// appendSint64Slice encodes a []int64 pointer as a repeated Sint64. +func appendSint64Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int64Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v)) + } + return b, nil +} + +// consumeSint64Slice wire decodes a []int64 pointer as a repeated Sint64. +func consumeSint64Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Int64Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, protowire.DecodeZigZag(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, protowire.DecodeZigZag(v)) + out.n = n + return out, nil +} + +var coderSint64Slice = pointerCoderFuncs{ + size: sizeSint64Slice, + marshal: appendSint64Slice, + unmarshal: consumeSint64Slice, + merge: mergeInt64Slice, +} + +// sizeSint64PackedSlice returns the size of wire encoding a []int64 pointer as a packed repeated Sint64. +func sizeSint64PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int64Slice() + if len(s) == 0 { + return 0 + } + n := 0 + for _, v := range s { + n += protowire.SizeVarint(protowire.EncodeZigZag(v)) + } + return f.tagsize + protowire.SizeBytes(n) +} + +// appendSint64PackedSlice encodes a []int64 pointer as a packed repeated Sint64. +func appendSint64PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int64Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for _, v := range s { + n += protowire.SizeVarint(protowire.EncodeZigZag(v)) + } + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v)) + } + return b, nil +} + +var coderSint64PackedSlice = pointerCoderFuncs{ + size: sizeSint64PackedSlice, + marshal: appendSint64PackedSlice, + unmarshal: consumeSint64Slice, + merge: mergeInt64Slice, +} + +// sizeSint64Value returns the size of wire encoding a int64 value as a Sint64. +func sizeSint64Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(protowire.EncodeZigZag(v.Int())) +} + +// appendSint64Value encodes a int64 value as a Sint64. +func appendSint64Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v.Int())) + return b, nil +} + +// consumeSint64Value decodes a int64 value as a Sint64. +func consumeSint64Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfInt64(protowire.DecodeZigZag(v)), out, nil +} + +var coderSint64Value = valueCoderFuncs{ + size: sizeSint64Value, + marshal: appendSint64Value, + unmarshal: consumeSint64Value, + merge: mergeScalarValue, +} + +// sizeSint64SliceValue returns the size of wire encoding a []int64 value as a repeated Sint64. +func sizeSint64SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(protowire.EncodeZigZag(v.Int())) + } + return size +} + +// appendSint64SliceValue encodes a []int64 value as a repeated Sint64. +func appendSint64SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v.Int())) + } + return b, nil +} + +// consumeSint64SliceValue wire decodes a []int64 value as a repeated Sint64. +func consumeSint64SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(protowire.DecodeZigZag(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(protowire.DecodeZigZag(v))) + out.n = n + return listv, out, nil +} + +var coderSint64SliceValue = valueCoderFuncs{ + size: sizeSint64SliceValue, + marshal: appendSint64SliceValue, + unmarshal: consumeSint64SliceValue, + merge: mergeListValue, +} + +// sizeSint64PackedSliceValue returns the size of wire encoding a []int64 value as a packed repeated Sint64. +func sizeSint64PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(protowire.EncodeZigZag(v.Int())) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendSint64PackedSliceValue encodes a []int64 value as a packed repeated Sint64. +func appendSint64PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(protowire.EncodeZigZag(v.Int())) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v.Int())) + } + return b, nil +} + +var coderSint64PackedSliceValue = valueCoderFuncs{ + size: sizeSint64PackedSliceValue, + marshal: appendSint64PackedSliceValue, + unmarshal: consumeSint64SliceValue, + merge: mergeListValue, +} + +// sizeUint64 returns the size of wire encoding a uint64 pointer as a Uint64. +func sizeUint64(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Uint64() + return f.tagsize + protowire.SizeVarint(v) +} + +// appendUint64 wire encodes a uint64 pointer as a Uint64. +func appendUint64(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint64() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, v) + return b, nil +} + +// consumeUint64 wire decodes a uint64 pointer as a Uint64. +func consumeUint64(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Uint64() = v + out.n = n + return out, nil +} + +var coderUint64 = pointerCoderFuncs{ + size: sizeUint64, + marshal: appendUint64, + unmarshal: consumeUint64, + merge: mergeUint64, +} + +// sizeUint64NoZero returns the size of wire encoding a uint64 pointer as a Uint64. +// The zero value is not encoded. +func sizeUint64NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Uint64() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeVarint(v) +} + +// appendUint64NoZero wire encodes a uint64 pointer as a Uint64. +// The zero value is not encoded. +func appendUint64NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint64() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, v) + return b, nil +} + +var coderUint64NoZero = pointerCoderFuncs{ + size: sizeUint64NoZero, + marshal: appendUint64NoZero, + unmarshal: consumeUint64, + merge: mergeUint64NoZero, +} + +// sizeUint64Ptr returns the size of wire encoding a *uint64 pointer as a Uint64. +// It panics if the pointer is nil. +func sizeUint64Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.Uint64Ptr() + return f.tagsize + protowire.SizeVarint(v) +} + +// appendUint64Ptr wire encodes a *uint64 pointer as a Uint64. +// It panics if the pointer is nil. +func appendUint64Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Uint64Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, v) + return b, nil +} + +// consumeUint64Ptr wire decodes a *uint64 pointer as a Uint64. +func consumeUint64Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Uint64Ptr() + if *vp == nil { + *vp = new(uint64) + } + **vp = v + out.n = n + return out, nil +} + +var coderUint64Ptr = pointerCoderFuncs{ + size: sizeUint64Ptr, + marshal: appendUint64Ptr, + unmarshal: consumeUint64Ptr, + merge: mergeUint64Ptr, +} + +// sizeUint64Slice returns the size of wire encoding a []uint64 pointer as a repeated Uint64. +func sizeUint64Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint64Slice() + for _, v := range s { + size += f.tagsize + protowire.SizeVarint(v) + } + return size +} + +// appendUint64Slice encodes a []uint64 pointer as a repeated Uint64. +func appendUint64Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint64Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, v) + } + return b, nil +} + +// consumeUint64Slice wire decodes a []uint64 pointer as a repeated Uint64. +func consumeUint64Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Uint64Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, v) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, v) + out.n = n + return out, nil +} + +var coderUint64Slice = pointerCoderFuncs{ + size: sizeUint64Slice, + marshal: appendUint64Slice, + unmarshal: consumeUint64Slice, + merge: mergeUint64Slice, +} + +// sizeUint64PackedSlice returns the size of wire encoding a []uint64 pointer as a packed repeated Uint64. +func sizeUint64PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint64Slice() + if len(s) == 0 { + return 0 + } + n := 0 + for _, v := range s { + n += protowire.SizeVarint(v) + } + return f.tagsize + protowire.SizeBytes(n) +} + +// appendUint64PackedSlice encodes a []uint64 pointer as a packed repeated Uint64. +func appendUint64PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint64Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for _, v := range s { + n += protowire.SizeVarint(v) + } + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendVarint(b, v) + } + return b, nil +} + +var coderUint64PackedSlice = pointerCoderFuncs{ + size: sizeUint64PackedSlice, + marshal: appendUint64PackedSlice, + unmarshal: consumeUint64Slice, + merge: mergeUint64Slice, +} + +// sizeUint64Value returns the size of wire encoding a uint64 value as a Uint64. +func sizeUint64Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeVarint(v.Uint()) +} + +// appendUint64Value encodes a uint64 value as a Uint64. +func appendUint64Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, v.Uint()) + return b, nil +} + +// consumeUint64Value decodes a uint64 value as a Uint64. +func consumeUint64Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfUint64(v), out, nil +} + +var coderUint64Value = valueCoderFuncs{ + size: sizeUint64Value, + marshal: appendUint64Value, + unmarshal: consumeUint64Value, + merge: mergeScalarValue, +} + +// sizeUint64SliceValue returns the size of wire encoding a []uint64 value as a repeated Uint64. +func sizeUint64SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeVarint(v.Uint()) + } + return size +} + +// appendUint64SliceValue encodes a []uint64 value as a repeated Uint64. +func appendUint64SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendVarint(b, v.Uint()) + } + return b, nil +} + +// consumeUint64SliceValue wire decodes a []uint64 value as a repeated Uint64. +func consumeUint64SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint64(v)) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.VarintType { + return protoreflect.Value{}, out, errUnknown + } + var v uint64 + var n int + if len(b) >= 1 && b[0] < 0x80 { + v = uint64(b[0]) + n = 1 + } else if len(b) >= 2 && b[1] < 128 { + v = uint64(b[0]&0x7f) + uint64(b[1])<<7 + n = 2 + } else { + v, n = protowire.ConsumeVarint(b) + } + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint64(v)) + out.n = n + return listv, out, nil +} + +var coderUint64SliceValue = valueCoderFuncs{ + size: sizeUint64SliceValue, + marshal: appendUint64SliceValue, + unmarshal: consumeUint64SliceValue, + merge: mergeListValue, +} + +// sizeUint64PackedSliceValue returns the size of wire encoding a []uint64 value as a packed repeated Uint64. +func sizeUint64PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := 0 + for i, llen := 0, llen; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(v.Uint()) + } + return tagsize + protowire.SizeBytes(n) +} + +// appendUint64PackedSliceValue encodes a []uint64 value as a packed repeated Uint64. +func appendUint64PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := 0 + for i := 0; i < llen; i++ { + v := list.Get(i) + n += protowire.SizeVarint(v.Uint()) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, v.Uint()) + } + return b, nil +} + +var coderUint64PackedSliceValue = valueCoderFuncs{ + size: sizeUint64PackedSliceValue, + marshal: appendUint64PackedSliceValue, + unmarshal: consumeUint64SliceValue, + merge: mergeListValue, +} + +// sizeSfixed32 returns the size of wire encoding a int32 pointer as a Sfixed32. +func sizeSfixed32(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + + return f.tagsize + protowire.SizeFixed32() +} + +// appendSfixed32 wire encodes a int32 pointer as a Sfixed32. +func appendSfixed32(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int32() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, uint32(v)) + return b, nil +} + +// consumeSfixed32 wire decodes a int32 pointer as a Sfixed32. +func consumeSfixed32(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Int32() = int32(v) + out.n = n + return out, nil +} + +var coderSfixed32 = pointerCoderFuncs{ + size: sizeSfixed32, + marshal: appendSfixed32, + unmarshal: consumeSfixed32, + merge: mergeInt32, +} + +// sizeSfixed32NoZero returns the size of wire encoding a int32 pointer as a Sfixed32. +// The zero value is not encoded. +func sizeSfixed32NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int32() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeFixed32() +} + +// appendSfixed32NoZero wire encodes a int32 pointer as a Sfixed32. +// The zero value is not encoded. +func appendSfixed32NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int32() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, uint32(v)) + return b, nil +} + +var coderSfixed32NoZero = pointerCoderFuncs{ + size: sizeSfixed32NoZero, + marshal: appendSfixed32NoZero, + unmarshal: consumeSfixed32, + merge: mergeInt32NoZero, +} + +// sizeSfixed32Ptr returns the size of wire encoding a *int32 pointer as a Sfixed32. +// It panics if the pointer is nil. +func sizeSfixed32Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + return f.tagsize + protowire.SizeFixed32() +} + +// appendSfixed32Ptr wire encodes a *int32 pointer as a Sfixed32. +// It panics if the pointer is nil. +func appendSfixed32Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Int32Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, uint32(v)) + return b, nil +} + +// consumeSfixed32Ptr wire decodes a *int32 pointer as a Sfixed32. +func consumeSfixed32Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Int32Ptr() + if *vp == nil { + *vp = new(int32) + } + **vp = int32(v) + out.n = n + return out, nil +} + +var coderSfixed32Ptr = pointerCoderFuncs{ + size: sizeSfixed32Ptr, + marshal: appendSfixed32Ptr, + unmarshal: consumeSfixed32Ptr, + merge: mergeInt32Ptr, +} + +// sizeSfixed32Slice returns the size of wire encoding a []int32 pointer as a repeated Sfixed32. +func sizeSfixed32Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int32Slice() + size = len(s) * (f.tagsize + protowire.SizeFixed32()) + return size +} + +// appendSfixed32Slice encodes a []int32 pointer as a repeated Sfixed32. +func appendSfixed32Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int32Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, uint32(v)) + } + return b, nil +} + +// consumeSfixed32Slice wire decodes a []int32 pointer as a repeated Sfixed32. +func consumeSfixed32Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Int32Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, int32(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, int32(v)) + out.n = n + return out, nil +} + +var coderSfixed32Slice = pointerCoderFuncs{ + size: sizeSfixed32Slice, + marshal: appendSfixed32Slice, + unmarshal: consumeSfixed32Slice, + merge: mergeInt32Slice, +} + +// sizeSfixed32PackedSlice returns the size of wire encoding a []int32 pointer as a packed repeated Sfixed32. +func sizeSfixed32PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int32Slice() + if len(s) == 0 { + return 0 + } + n := len(s) * protowire.SizeFixed32() + return f.tagsize + protowire.SizeBytes(n) +} + +// appendSfixed32PackedSlice encodes a []int32 pointer as a packed repeated Sfixed32. +func appendSfixed32PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int32Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := len(s) * protowire.SizeFixed32() + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendFixed32(b, uint32(v)) + } + return b, nil +} + +var coderSfixed32PackedSlice = pointerCoderFuncs{ + size: sizeSfixed32PackedSlice, + marshal: appendSfixed32PackedSlice, + unmarshal: consumeSfixed32Slice, + merge: mergeInt32Slice, +} + +// sizeSfixed32Value returns the size of wire encoding a int32 value as a Sfixed32. +func sizeSfixed32Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeFixed32() +} + +// appendSfixed32Value encodes a int32 value as a Sfixed32. +func appendSfixed32Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed32(b, uint32(v.Int())) + return b, nil +} + +// consumeSfixed32Value decodes a int32 value as a Sfixed32. +func consumeSfixed32Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfInt32(int32(v)), out, nil +} + +var coderSfixed32Value = valueCoderFuncs{ + size: sizeSfixed32Value, + marshal: appendSfixed32Value, + unmarshal: consumeSfixed32Value, + merge: mergeScalarValue, +} + +// sizeSfixed32SliceValue returns the size of wire encoding a []int32 value as a repeated Sfixed32. +func sizeSfixed32SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + size = list.Len() * (tagsize + protowire.SizeFixed32()) + return size +} + +// appendSfixed32SliceValue encodes a []int32 value as a repeated Sfixed32. +func appendSfixed32SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed32(b, uint32(v.Int())) + } + return b, nil +} + +// consumeSfixed32SliceValue wire decodes a []int32 value as a repeated Sfixed32. +func consumeSfixed32SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.Fixed32Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(v))) + out.n = n + return listv, out, nil +} + +var coderSfixed32SliceValue = valueCoderFuncs{ + size: sizeSfixed32SliceValue, + marshal: appendSfixed32SliceValue, + unmarshal: consumeSfixed32SliceValue, + merge: mergeListValue, +} + +// sizeSfixed32PackedSliceValue returns the size of wire encoding a []int32 value as a packed repeated Sfixed32. +func sizeSfixed32PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := llen * protowire.SizeFixed32() + return tagsize + protowire.SizeBytes(n) +} + +// appendSfixed32PackedSliceValue encodes a []int32 value as a packed repeated Sfixed32. +func appendSfixed32PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := llen * protowire.SizeFixed32() + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendFixed32(b, uint32(v.Int())) + } + return b, nil +} + +var coderSfixed32PackedSliceValue = valueCoderFuncs{ + size: sizeSfixed32PackedSliceValue, + marshal: appendSfixed32PackedSliceValue, + unmarshal: consumeSfixed32SliceValue, + merge: mergeListValue, +} + +// sizeFixed32 returns the size of wire encoding a uint32 pointer as a Fixed32. +func sizeFixed32(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + + return f.tagsize + protowire.SizeFixed32() +} + +// appendFixed32 wire encodes a uint32 pointer as a Fixed32. +func appendFixed32(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint32() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, v) + return b, nil +} + +// consumeFixed32 wire decodes a uint32 pointer as a Fixed32. +func consumeFixed32(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Uint32() = v + out.n = n + return out, nil +} + +var coderFixed32 = pointerCoderFuncs{ + size: sizeFixed32, + marshal: appendFixed32, + unmarshal: consumeFixed32, + merge: mergeUint32, +} + +// sizeFixed32NoZero returns the size of wire encoding a uint32 pointer as a Fixed32. +// The zero value is not encoded. +func sizeFixed32NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Uint32() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeFixed32() +} + +// appendFixed32NoZero wire encodes a uint32 pointer as a Fixed32. +// The zero value is not encoded. +func appendFixed32NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint32() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, v) + return b, nil +} + +var coderFixed32NoZero = pointerCoderFuncs{ + size: sizeFixed32NoZero, + marshal: appendFixed32NoZero, + unmarshal: consumeFixed32, + merge: mergeUint32NoZero, +} + +// sizeFixed32Ptr returns the size of wire encoding a *uint32 pointer as a Fixed32. +// It panics if the pointer is nil. +func sizeFixed32Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + return f.tagsize + protowire.SizeFixed32() +} + +// appendFixed32Ptr wire encodes a *uint32 pointer as a Fixed32. +// It panics if the pointer is nil. +func appendFixed32Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Uint32Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, v) + return b, nil +} + +// consumeFixed32Ptr wire decodes a *uint32 pointer as a Fixed32. +func consumeFixed32Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Uint32Ptr() + if *vp == nil { + *vp = new(uint32) + } + **vp = v + out.n = n + return out, nil +} + +var coderFixed32Ptr = pointerCoderFuncs{ + size: sizeFixed32Ptr, + marshal: appendFixed32Ptr, + unmarshal: consumeFixed32Ptr, + merge: mergeUint32Ptr, +} + +// sizeFixed32Slice returns the size of wire encoding a []uint32 pointer as a repeated Fixed32. +func sizeFixed32Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint32Slice() + size = len(s) * (f.tagsize + protowire.SizeFixed32()) + return size +} + +// appendFixed32Slice encodes a []uint32 pointer as a repeated Fixed32. +func appendFixed32Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint32Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, v) + } + return b, nil +} + +// consumeFixed32Slice wire decodes a []uint32 pointer as a repeated Fixed32. +func consumeFixed32Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Uint32Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, v) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, v) + out.n = n + return out, nil +} + +var coderFixed32Slice = pointerCoderFuncs{ + size: sizeFixed32Slice, + marshal: appendFixed32Slice, + unmarshal: consumeFixed32Slice, + merge: mergeUint32Slice, +} + +// sizeFixed32PackedSlice returns the size of wire encoding a []uint32 pointer as a packed repeated Fixed32. +func sizeFixed32PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint32Slice() + if len(s) == 0 { + return 0 + } + n := len(s) * protowire.SizeFixed32() + return f.tagsize + protowire.SizeBytes(n) +} + +// appendFixed32PackedSlice encodes a []uint32 pointer as a packed repeated Fixed32. +func appendFixed32PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint32Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := len(s) * protowire.SizeFixed32() + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendFixed32(b, v) + } + return b, nil +} + +var coderFixed32PackedSlice = pointerCoderFuncs{ + size: sizeFixed32PackedSlice, + marshal: appendFixed32PackedSlice, + unmarshal: consumeFixed32Slice, + merge: mergeUint32Slice, +} + +// sizeFixed32Value returns the size of wire encoding a uint32 value as a Fixed32. +func sizeFixed32Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeFixed32() +} + +// appendFixed32Value encodes a uint32 value as a Fixed32. +func appendFixed32Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed32(b, uint32(v.Uint())) + return b, nil +} + +// consumeFixed32Value decodes a uint32 value as a Fixed32. +func consumeFixed32Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfUint32(uint32(v)), out, nil +} + +var coderFixed32Value = valueCoderFuncs{ + size: sizeFixed32Value, + marshal: appendFixed32Value, + unmarshal: consumeFixed32Value, + merge: mergeScalarValue, +} + +// sizeFixed32SliceValue returns the size of wire encoding a []uint32 value as a repeated Fixed32. +func sizeFixed32SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + size = list.Len() * (tagsize + protowire.SizeFixed32()) + return size +} + +// appendFixed32SliceValue encodes a []uint32 value as a repeated Fixed32. +func appendFixed32SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed32(b, uint32(v.Uint())) + } + return b, nil +} + +// consumeFixed32SliceValue wire decodes a []uint32 value as a repeated Fixed32. +func consumeFixed32SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint32(uint32(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.Fixed32Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint32(uint32(v))) + out.n = n + return listv, out, nil +} + +var coderFixed32SliceValue = valueCoderFuncs{ + size: sizeFixed32SliceValue, + marshal: appendFixed32SliceValue, + unmarshal: consumeFixed32SliceValue, + merge: mergeListValue, +} + +// sizeFixed32PackedSliceValue returns the size of wire encoding a []uint32 value as a packed repeated Fixed32. +func sizeFixed32PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := llen * protowire.SizeFixed32() + return tagsize + protowire.SizeBytes(n) +} + +// appendFixed32PackedSliceValue encodes a []uint32 value as a packed repeated Fixed32. +func appendFixed32PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := llen * protowire.SizeFixed32() + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendFixed32(b, uint32(v.Uint())) + } + return b, nil +} + +var coderFixed32PackedSliceValue = valueCoderFuncs{ + size: sizeFixed32PackedSliceValue, + marshal: appendFixed32PackedSliceValue, + unmarshal: consumeFixed32SliceValue, + merge: mergeListValue, +} + +// sizeFloat returns the size of wire encoding a float32 pointer as a Float. +func sizeFloat(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + + return f.tagsize + protowire.SizeFixed32() +} + +// appendFloat wire encodes a float32 pointer as a Float. +func appendFloat(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Float32() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, math.Float32bits(v)) + return b, nil +} + +// consumeFloat wire decodes a float32 pointer as a Float. +func consumeFloat(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Float32() = math.Float32frombits(v) + out.n = n + return out, nil +} + +var coderFloat = pointerCoderFuncs{ + size: sizeFloat, + marshal: appendFloat, + unmarshal: consumeFloat, + merge: mergeFloat32, +} + +// sizeFloatNoZero returns the size of wire encoding a float32 pointer as a Float. +// The zero value is not encoded. +func sizeFloatNoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Float32() + if v == 0 && !math.Signbit(float64(v)) { + return 0 + } + return f.tagsize + protowire.SizeFixed32() +} + +// appendFloatNoZero wire encodes a float32 pointer as a Float. +// The zero value is not encoded. +func appendFloatNoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Float32() + if v == 0 && !math.Signbit(float64(v)) { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, math.Float32bits(v)) + return b, nil +} + +var coderFloatNoZero = pointerCoderFuncs{ + size: sizeFloatNoZero, + marshal: appendFloatNoZero, + unmarshal: consumeFloat, + merge: mergeFloat32NoZero, +} + +// sizeFloatPtr returns the size of wire encoding a *float32 pointer as a Float. +// It panics if the pointer is nil. +func sizeFloatPtr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + return f.tagsize + protowire.SizeFixed32() +} + +// appendFloatPtr wire encodes a *float32 pointer as a Float. +// It panics if the pointer is nil. +func appendFloatPtr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Float32Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, math.Float32bits(v)) + return b, nil +} + +// consumeFloatPtr wire decodes a *float32 pointer as a Float. +func consumeFloatPtr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Float32Ptr() + if *vp == nil { + *vp = new(float32) + } + **vp = math.Float32frombits(v) + out.n = n + return out, nil +} + +var coderFloatPtr = pointerCoderFuncs{ + size: sizeFloatPtr, + marshal: appendFloatPtr, + unmarshal: consumeFloatPtr, + merge: mergeFloat32Ptr, +} + +// sizeFloatSlice returns the size of wire encoding a []float32 pointer as a repeated Float. +func sizeFloatSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Float32Slice() + size = len(s) * (f.tagsize + protowire.SizeFixed32()) + return size +} + +// appendFloatSlice encodes a []float32 pointer as a repeated Float. +func appendFloatSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Float32Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed32(b, math.Float32bits(v)) + } + return b, nil +} + +// consumeFloatSlice wire decodes a []float32 pointer as a repeated Float. +func consumeFloatSlice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Float32Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, math.Float32frombits(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.Fixed32Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, math.Float32frombits(v)) + out.n = n + return out, nil +} + +var coderFloatSlice = pointerCoderFuncs{ + size: sizeFloatSlice, + marshal: appendFloatSlice, + unmarshal: consumeFloatSlice, + merge: mergeFloat32Slice, +} + +// sizeFloatPackedSlice returns the size of wire encoding a []float32 pointer as a packed repeated Float. +func sizeFloatPackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Float32Slice() + if len(s) == 0 { + return 0 + } + n := len(s) * protowire.SizeFixed32() + return f.tagsize + protowire.SizeBytes(n) +} + +// appendFloatPackedSlice encodes a []float32 pointer as a packed repeated Float. +func appendFloatPackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Float32Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := len(s) * protowire.SizeFixed32() + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendFixed32(b, math.Float32bits(v)) + } + return b, nil +} + +var coderFloatPackedSlice = pointerCoderFuncs{ + size: sizeFloatPackedSlice, + marshal: appendFloatPackedSlice, + unmarshal: consumeFloatSlice, + merge: mergeFloat32Slice, +} + +// sizeFloatValue returns the size of wire encoding a float32 value as a Float. +func sizeFloatValue(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeFixed32() +} + +// appendFloatValue encodes a float32 value as a Float. +func appendFloatValue(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed32(b, math.Float32bits(float32(v.Float()))) + return b, nil +} + +// consumeFloatValue decodes a float32 value as a Float. +func consumeFloatValue(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.Fixed32Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfFloat32(math.Float32frombits(uint32(v))), out, nil +} + +var coderFloatValue = valueCoderFuncs{ + size: sizeFloatValue, + marshal: appendFloatValue, + unmarshal: consumeFloatValue, + merge: mergeScalarValue, +} + +// sizeFloatSliceValue returns the size of wire encoding a []float32 value as a repeated Float. +func sizeFloatSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + size = list.Len() * (tagsize + protowire.SizeFixed32()) + return size +} + +// appendFloatSliceValue encodes a []float32 value as a repeated Float. +func appendFloatSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed32(b, math.Float32bits(float32(v.Float()))) + } + return b, nil +} + +// consumeFloatSliceValue wire decodes a []float32 value as a repeated Float. +func consumeFloatSliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfFloat32(math.Float32frombits(uint32(v)))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.Fixed32Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfFloat32(math.Float32frombits(uint32(v)))) + out.n = n + return listv, out, nil +} + +var coderFloatSliceValue = valueCoderFuncs{ + size: sizeFloatSliceValue, + marshal: appendFloatSliceValue, + unmarshal: consumeFloatSliceValue, + merge: mergeListValue, +} + +// sizeFloatPackedSliceValue returns the size of wire encoding a []float32 value as a packed repeated Float. +func sizeFloatPackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := llen * protowire.SizeFixed32() + return tagsize + protowire.SizeBytes(n) +} + +// appendFloatPackedSliceValue encodes a []float32 value as a packed repeated Float. +func appendFloatPackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := llen * protowire.SizeFixed32() + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendFixed32(b, math.Float32bits(float32(v.Float()))) + } + return b, nil +} + +var coderFloatPackedSliceValue = valueCoderFuncs{ + size: sizeFloatPackedSliceValue, + marshal: appendFloatPackedSliceValue, + unmarshal: consumeFloatSliceValue, + merge: mergeListValue, +} + +// sizeSfixed64 returns the size of wire encoding a int64 pointer as a Sfixed64. +func sizeSfixed64(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + + return f.tagsize + protowire.SizeFixed64() +} + +// appendSfixed64 wire encodes a int64 pointer as a Sfixed64. +func appendSfixed64(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int64() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, uint64(v)) + return b, nil +} + +// consumeSfixed64 wire decodes a int64 pointer as a Sfixed64. +func consumeSfixed64(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Int64() = int64(v) + out.n = n + return out, nil +} + +var coderSfixed64 = pointerCoderFuncs{ + size: sizeSfixed64, + marshal: appendSfixed64, + unmarshal: consumeSfixed64, + merge: mergeInt64, +} + +// sizeSfixed64NoZero returns the size of wire encoding a int64 pointer as a Sfixed64. +// The zero value is not encoded. +func sizeSfixed64NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Int64() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeFixed64() +} + +// appendSfixed64NoZero wire encodes a int64 pointer as a Sfixed64. +// The zero value is not encoded. +func appendSfixed64NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Int64() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, uint64(v)) + return b, nil +} + +var coderSfixed64NoZero = pointerCoderFuncs{ + size: sizeSfixed64NoZero, + marshal: appendSfixed64NoZero, + unmarshal: consumeSfixed64, + merge: mergeInt64NoZero, +} + +// sizeSfixed64Ptr returns the size of wire encoding a *int64 pointer as a Sfixed64. +// It panics if the pointer is nil. +func sizeSfixed64Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + return f.tagsize + protowire.SizeFixed64() +} + +// appendSfixed64Ptr wire encodes a *int64 pointer as a Sfixed64. +// It panics if the pointer is nil. +func appendSfixed64Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Int64Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, uint64(v)) + return b, nil +} + +// consumeSfixed64Ptr wire decodes a *int64 pointer as a Sfixed64. +func consumeSfixed64Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Int64Ptr() + if *vp == nil { + *vp = new(int64) + } + **vp = int64(v) + out.n = n + return out, nil +} + +var coderSfixed64Ptr = pointerCoderFuncs{ + size: sizeSfixed64Ptr, + marshal: appendSfixed64Ptr, + unmarshal: consumeSfixed64Ptr, + merge: mergeInt64Ptr, +} + +// sizeSfixed64Slice returns the size of wire encoding a []int64 pointer as a repeated Sfixed64. +func sizeSfixed64Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int64Slice() + size = len(s) * (f.tagsize + protowire.SizeFixed64()) + return size +} + +// appendSfixed64Slice encodes a []int64 pointer as a repeated Sfixed64. +func appendSfixed64Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int64Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, uint64(v)) + } + return b, nil +} + +// consumeSfixed64Slice wire decodes a []int64 pointer as a repeated Sfixed64. +func consumeSfixed64Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Int64Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, int64(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, int64(v)) + out.n = n + return out, nil +} + +var coderSfixed64Slice = pointerCoderFuncs{ + size: sizeSfixed64Slice, + marshal: appendSfixed64Slice, + unmarshal: consumeSfixed64Slice, + merge: mergeInt64Slice, +} + +// sizeSfixed64PackedSlice returns the size of wire encoding a []int64 pointer as a packed repeated Sfixed64. +func sizeSfixed64PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Int64Slice() + if len(s) == 0 { + return 0 + } + n := len(s) * protowire.SizeFixed64() + return f.tagsize + protowire.SizeBytes(n) +} + +// appendSfixed64PackedSlice encodes a []int64 pointer as a packed repeated Sfixed64. +func appendSfixed64PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Int64Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := len(s) * protowire.SizeFixed64() + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendFixed64(b, uint64(v)) + } + return b, nil +} + +var coderSfixed64PackedSlice = pointerCoderFuncs{ + size: sizeSfixed64PackedSlice, + marshal: appendSfixed64PackedSlice, + unmarshal: consumeSfixed64Slice, + merge: mergeInt64Slice, +} + +// sizeSfixed64Value returns the size of wire encoding a int64 value as a Sfixed64. +func sizeSfixed64Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeFixed64() +} + +// appendSfixed64Value encodes a int64 value as a Sfixed64. +func appendSfixed64Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed64(b, uint64(v.Int())) + return b, nil +} + +// consumeSfixed64Value decodes a int64 value as a Sfixed64. +func consumeSfixed64Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfInt64(int64(v)), out, nil +} + +var coderSfixed64Value = valueCoderFuncs{ + size: sizeSfixed64Value, + marshal: appendSfixed64Value, + unmarshal: consumeSfixed64Value, + merge: mergeScalarValue, +} + +// sizeSfixed64SliceValue returns the size of wire encoding a []int64 value as a repeated Sfixed64. +func sizeSfixed64SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + size = list.Len() * (tagsize + protowire.SizeFixed64()) + return size +} + +// appendSfixed64SliceValue encodes a []int64 value as a repeated Sfixed64. +func appendSfixed64SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed64(b, uint64(v.Int())) + } + return b, nil +} + +// consumeSfixed64SliceValue wire decodes a []int64 value as a repeated Sfixed64. +func consumeSfixed64SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(int64(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.Fixed64Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(int64(v))) + out.n = n + return listv, out, nil +} + +var coderSfixed64SliceValue = valueCoderFuncs{ + size: sizeSfixed64SliceValue, + marshal: appendSfixed64SliceValue, + unmarshal: consumeSfixed64SliceValue, + merge: mergeListValue, +} + +// sizeSfixed64PackedSliceValue returns the size of wire encoding a []int64 value as a packed repeated Sfixed64. +func sizeSfixed64PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := llen * protowire.SizeFixed64() + return tagsize + protowire.SizeBytes(n) +} + +// appendSfixed64PackedSliceValue encodes a []int64 value as a packed repeated Sfixed64. +func appendSfixed64PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := llen * protowire.SizeFixed64() + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendFixed64(b, uint64(v.Int())) + } + return b, nil +} + +var coderSfixed64PackedSliceValue = valueCoderFuncs{ + size: sizeSfixed64PackedSliceValue, + marshal: appendSfixed64PackedSliceValue, + unmarshal: consumeSfixed64SliceValue, + merge: mergeListValue, +} + +// sizeFixed64 returns the size of wire encoding a uint64 pointer as a Fixed64. +func sizeFixed64(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + + return f.tagsize + protowire.SizeFixed64() +} + +// appendFixed64 wire encodes a uint64 pointer as a Fixed64. +func appendFixed64(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint64() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, v) + return b, nil +} + +// consumeFixed64 wire decodes a uint64 pointer as a Fixed64. +func consumeFixed64(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Uint64() = v + out.n = n + return out, nil +} + +var coderFixed64 = pointerCoderFuncs{ + size: sizeFixed64, + marshal: appendFixed64, + unmarshal: consumeFixed64, + merge: mergeUint64, +} + +// sizeFixed64NoZero returns the size of wire encoding a uint64 pointer as a Fixed64. +// The zero value is not encoded. +func sizeFixed64NoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Uint64() + if v == 0 { + return 0 + } + return f.tagsize + protowire.SizeFixed64() +} + +// appendFixed64NoZero wire encodes a uint64 pointer as a Fixed64. +// The zero value is not encoded. +func appendFixed64NoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Uint64() + if v == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, v) + return b, nil +} + +var coderFixed64NoZero = pointerCoderFuncs{ + size: sizeFixed64NoZero, + marshal: appendFixed64NoZero, + unmarshal: consumeFixed64, + merge: mergeUint64NoZero, +} + +// sizeFixed64Ptr returns the size of wire encoding a *uint64 pointer as a Fixed64. +// It panics if the pointer is nil. +func sizeFixed64Ptr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + return f.tagsize + protowire.SizeFixed64() +} + +// appendFixed64Ptr wire encodes a *uint64 pointer as a Fixed64. +// It panics if the pointer is nil. +func appendFixed64Ptr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Uint64Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, v) + return b, nil +} + +// consumeFixed64Ptr wire decodes a *uint64 pointer as a Fixed64. +func consumeFixed64Ptr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Uint64Ptr() + if *vp == nil { + *vp = new(uint64) + } + **vp = v + out.n = n + return out, nil +} + +var coderFixed64Ptr = pointerCoderFuncs{ + size: sizeFixed64Ptr, + marshal: appendFixed64Ptr, + unmarshal: consumeFixed64Ptr, + merge: mergeUint64Ptr, +} + +// sizeFixed64Slice returns the size of wire encoding a []uint64 pointer as a repeated Fixed64. +func sizeFixed64Slice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint64Slice() + size = len(s) * (f.tagsize + protowire.SizeFixed64()) + return size +} + +// appendFixed64Slice encodes a []uint64 pointer as a repeated Fixed64. +func appendFixed64Slice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint64Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, v) + } + return b, nil +} + +// consumeFixed64Slice wire decodes a []uint64 pointer as a repeated Fixed64. +func consumeFixed64Slice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Uint64Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, v) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, v) + out.n = n + return out, nil +} + +var coderFixed64Slice = pointerCoderFuncs{ + size: sizeFixed64Slice, + marshal: appendFixed64Slice, + unmarshal: consumeFixed64Slice, + merge: mergeUint64Slice, +} + +// sizeFixed64PackedSlice returns the size of wire encoding a []uint64 pointer as a packed repeated Fixed64. +func sizeFixed64PackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Uint64Slice() + if len(s) == 0 { + return 0 + } + n := len(s) * protowire.SizeFixed64() + return f.tagsize + protowire.SizeBytes(n) +} + +// appendFixed64PackedSlice encodes a []uint64 pointer as a packed repeated Fixed64. +func appendFixed64PackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Uint64Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := len(s) * protowire.SizeFixed64() + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendFixed64(b, v) + } + return b, nil +} + +var coderFixed64PackedSlice = pointerCoderFuncs{ + size: sizeFixed64PackedSlice, + marshal: appendFixed64PackedSlice, + unmarshal: consumeFixed64Slice, + merge: mergeUint64Slice, +} + +// sizeFixed64Value returns the size of wire encoding a uint64 value as a Fixed64. +func sizeFixed64Value(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeFixed64() +} + +// appendFixed64Value encodes a uint64 value as a Fixed64. +func appendFixed64Value(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed64(b, v.Uint()) + return b, nil +} + +// consumeFixed64Value decodes a uint64 value as a Fixed64. +func consumeFixed64Value(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfUint64(v), out, nil +} + +var coderFixed64Value = valueCoderFuncs{ + size: sizeFixed64Value, + marshal: appendFixed64Value, + unmarshal: consumeFixed64Value, + merge: mergeScalarValue, +} + +// sizeFixed64SliceValue returns the size of wire encoding a []uint64 value as a repeated Fixed64. +func sizeFixed64SliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + size = list.Len() * (tagsize + protowire.SizeFixed64()) + return size +} + +// appendFixed64SliceValue encodes a []uint64 value as a repeated Fixed64. +func appendFixed64SliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed64(b, v.Uint()) + } + return b, nil +} + +// consumeFixed64SliceValue wire decodes a []uint64 value as a repeated Fixed64. +func consumeFixed64SliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint64(v)) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.Fixed64Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint64(v)) + out.n = n + return listv, out, nil +} + +var coderFixed64SliceValue = valueCoderFuncs{ + size: sizeFixed64SliceValue, + marshal: appendFixed64SliceValue, + unmarshal: consumeFixed64SliceValue, + merge: mergeListValue, +} + +// sizeFixed64PackedSliceValue returns the size of wire encoding a []uint64 value as a packed repeated Fixed64. +func sizeFixed64PackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := llen * protowire.SizeFixed64() + return tagsize + protowire.SizeBytes(n) +} + +// appendFixed64PackedSliceValue encodes a []uint64 value as a packed repeated Fixed64. +func appendFixed64PackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := llen * protowire.SizeFixed64() + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendFixed64(b, v.Uint()) + } + return b, nil +} + +var coderFixed64PackedSliceValue = valueCoderFuncs{ + size: sizeFixed64PackedSliceValue, + marshal: appendFixed64PackedSliceValue, + unmarshal: consumeFixed64SliceValue, + merge: mergeListValue, +} + +// sizeDouble returns the size of wire encoding a float64 pointer as a Double. +func sizeDouble(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + + return f.tagsize + protowire.SizeFixed64() +} + +// appendDouble wire encodes a float64 pointer as a Double. +func appendDouble(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Float64() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, math.Float64bits(v)) + return b, nil +} + +// consumeDouble wire decodes a float64 pointer as a Double. +func consumeDouble(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Float64() = math.Float64frombits(v) + out.n = n + return out, nil +} + +var coderDouble = pointerCoderFuncs{ + size: sizeDouble, + marshal: appendDouble, + unmarshal: consumeDouble, + merge: mergeFloat64, +} + +// sizeDoubleNoZero returns the size of wire encoding a float64 pointer as a Double. +// The zero value is not encoded. +func sizeDoubleNoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Float64() + if v == 0 && !math.Signbit(float64(v)) { + return 0 + } + return f.tagsize + protowire.SizeFixed64() +} + +// appendDoubleNoZero wire encodes a float64 pointer as a Double. +// The zero value is not encoded. +func appendDoubleNoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Float64() + if v == 0 && !math.Signbit(float64(v)) { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, math.Float64bits(v)) + return b, nil +} + +var coderDoubleNoZero = pointerCoderFuncs{ + size: sizeDoubleNoZero, + marshal: appendDoubleNoZero, + unmarshal: consumeDouble, + merge: mergeFloat64NoZero, +} + +// sizeDoublePtr returns the size of wire encoding a *float64 pointer as a Double. +// It panics if the pointer is nil. +func sizeDoublePtr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + return f.tagsize + protowire.SizeFixed64() +} + +// appendDoublePtr wire encodes a *float64 pointer as a Double. +// It panics if the pointer is nil. +func appendDoublePtr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.Float64Ptr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, math.Float64bits(v)) + return b, nil +} + +// consumeDoublePtr wire decodes a *float64 pointer as a Double. +func consumeDoublePtr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.Float64Ptr() + if *vp == nil { + *vp = new(float64) + } + **vp = math.Float64frombits(v) + out.n = n + return out, nil +} + +var coderDoublePtr = pointerCoderFuncs{ + size: sizeDoublePtr, + marshal: appendDoublePtr, + unmarshal: consumeDoublePtr, + merge: mergeFloat64Ptr, +} + +// sizeDoubleSlice returns the size of wire encoding a []float64 pointer as a repeated Double. +func sizeDoubleSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Float64Slice() + size = len(s) * (f.tagsize + protowire.SizeFixed64()) + return size +} + +// appendDoubleSlice encodes a []float64 pointer as a repeated Double. +func appendDoubleSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Float64Slice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendFixed64(b, math.Float64bits(v)) + } + return b, nil +} + +// consumeDoubleSlice wire decodes a []float64 pointer as a repeated Double. +func consumeDoubleSlice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.Float64Slice() + if wtyp == protowire.BytesType { + s := *sp + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + s = append(s, math.Float64frombits(v)) + b = b[n:] + } + *sp = s + out.n = n + return out, nil + } + if wtyp != protowire.Fixed64Type { + return out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, math.Float64frombits(v)) + out.n = n + return out, nil +} + +var coderDoubleSlice = pointerCoderFuncs{ + size: sizeDoubleSlice, + marshal: appendDoubleSlice, + unmarshal: consumeDoubleSlice, + merge: mergeFloat64Slice, +} + +// sizeDoublePackedSlice returns the size of wire encoding a []float64 pointer as a packed repeated Double. +func sizeDoublePackedSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.Float64Slice() + if len(s) == 0 { + return 0 + } + n := len(s) * protowire.SizeFixed64() + return f.tagsize + protowire.SizeBytes(n) +} + +// appendDoublePackedSlice encodes a []float64 pointer as a packed repeated Double. +func appendDoublePackedSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.Float64Slice() + if len(s) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := len(s) * protowire.SizeFixed64() + b = protowire.AppendVarint(b, uint64(n)) + for _, v := range s { + b = protowire.AppendFixed64(b, math.Float64bits(v)) + } + return b, nil +} + +var coderDoublePackedSlice = pointerCoderFuncs{ + size: sizeDoublePackedSlice, + marshal: appendDoublePackedSlice, + unmarshal: consumeDoubleSlice, + merge: mergeFloat64Slice, +} + +// sizeDoubleValue returns the size of wire encoding a float64 value as a Double. +func sizeDoubleValue(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeFixed64() +} + +// appendDoubleValue encodes a float64 value as a Double. +func appendDoubleValue(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed64(b, math.Float64bits(v.Float())) + return b, nil +} + +// consumeDoubleValue decodes a float64 value as a Double. +func consumeDoubleValue(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.Fixed64Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfFloat64(math.Float64frombits(v)), out, nil +} + +var coderDoubleValue = valueCoderFuncs{ + size: sizeDoubleValue, + marshal: appendDoubleValue, + unmarshal: consumeDoubleValue, + merge: mergeScalarValue, +} + +// sizeDoubleSliceValue returns the size of wire encoding a []float64 value as a repeated Double. +func sizeDoubleSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + size = list.Len() * (tagsize + protowire.SizeFixed64()) + return size +} + +// appendDoubleSliceValue encodes a []float64 value as a repeated Double. +func appendDoubleSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendFixed64(b, math.Float64bits(v.Float())) + } + return b, nil +} + +// consumeDoubleSliceValue wire decodes a []float64 value as a repeated Double. +func consumeDoubleSliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfFloat64(math.Float64frombits(v))) + b = b[n:] + } + out.n = n + return listv, out, nil + } + if wtyp != protowire.Fixed64Type { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfFloat64(math.Float64frombits(v))) + out.n = n + return listv, out, nil +} + +var coderDoubleSliceValue = valueCoderFuncs{ + size: sizeDoubleSliceValue, + marshal: appendDoubleSliceValue, + unmarshal: consumeDoubleSliceValue, + merge: mergeListValue, +} + +// sizeDoublePackedSliceValue returns the size of wire encoding a []float64 value as a packed repeated Double. +func sizeDoublePackedSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return 0 + } + n := llen * protowire.SizeFixed64() + return tagsize + protowire.SizeBytes(n) +} + +// appendDoublePackedSliceValue encodes a []float64 value as a packed repeated Double. +func appendDoublePackedSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + llen := list.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, wiretag) + n := llen * protowire.SizeFixed64() + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + v := list.Get(i) + b = protowire.AppendFixed64(b, math.Float64bits(v.Float())) + } + return b, nil +} + +var coderDoublePackedSliceValue = valueCoderFuncs{ + size: sizeDoublePackedSliceValue, + marshal: appendDoublePackedSliceValue, + unmarshal: consumeDoubleSliceValue, + merge: mergeListValue, +} + +// sizeString returns the size of wire encoding a string pointer as a String. +func sizeString(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.String() + return f.tagsize + protowire.SizeBytes(len(v)) +} + +// appendString wire encodes a string pointer as a String. +func appendString(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.String() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + return b, nil +} + +// consumeString wire decodes a string pointer as a String. +func consumeString(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.String() = v + out.n = n + return out, nil +} + +var coderString = pointerCoderFuncs{ + size: sizeString, + marshal: appendString, + unmarshal: consumeString, + merge: mergeString, +} + +// appendStringValidateUTF8 wire encodes a string pointer as a String. +func appendStringValidateUTF8(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.String() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + if !utf8.ValidString(v) { + return b, errInvalidUTF8{} + } + return b, nil +} + +// consumeStringValidateUTF8 wire decodes a string pointer as a String. +func consumeStringValidateUTF8(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if !utf8.ValidString(v) { + return out, errInvalidUTF8{} + } + *p.String() = v + out.n = n + return out, nil +} + +var coderStringValidateUTF8 = pointerCoderFuncs{ + size: sizeString, + marshal: appendStringValidateUTF8, + unmarshal: consumeStringValidateUTF8, + merge: mergeString, +} + +// sizeStringNoZero returns the size of wire encoding a string pointer as a String. +// The zero value is not encoded. +func sizeStringNoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.String() + if len(v) == 0 { + return 0 + } + return f.tagsize + protowire.SizeBytes(len(v)) +} + +// appendStringNoZero wire encodes a string pointer as a String. +// The zero value is not encoded. +func appendStringNoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.String() + if len(v) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + return b, nil +} + +var coderStringNoZero = pointerCoderFuncs{ + size: sizeStringNoZero, + marshal: appendStringNoZero, + unmarshal: consumeString, + merge: mergeStringNoZero, +} + +// appendStringNoZeroValidateUTF8 wire encodes a string pointer as a String. +// The zero value is not encoded. +func appendStringNoZeroValidateUTF8(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.String() + if len(v) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + if !utf8.ValidString(v) { + return b, errInvalidUTF8{} + } + return b, nil +} + +var coderStringNoZeroValidateUTF8 = pointerCoderFuncs{ + size: sizeStringNoZero, + marshal: appendStringNoZeroValidateUTF8, + unmarshal: consumeStringValidateUTF8, + merge: mergeStringNoZero, +} + +// sizeStringPtr returns the size of wire encoding a *string pointer as a String. +// It panics if the pointer is nil. +func sizeStringPtr(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := **p.StringPtr() + return f.tagsize + protowire.SizeBytes(len(v)) +} + +// appendStringPtr wire encodes a *string pointer as a String. +// It panics if the pointer is nil. +func appendStringPtr(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.StringPtr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + return b, nil +} + +// consumeStringPtr wire decodes a *string pointer as a String. +func consumeStringPtr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return out, protowire.ParseError(n) + } + vp := p.StringPtr() + if *vp == nil { + *vp = new(string) + } + **vp = v + out.n = n + return out, nil +} + +var coderStringPtr = pointerCoderFuncs{ + size: sizeStringPtr, + marshal: appendStringPtr, + unmarshal: consumeStringPtr, + merge: mergeStringPtr, +} + +// appendStringPtrValidateUTF8 wire encodes a *string pointer as a String. +// It panics if the pointer is nil. +func appendStringPtrValidateUTF8(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := **p.StringPtr() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + if !utf8.ValidString(v) { + return b, errInvalidUTF8{} + } + return b, nil +} + +// consumeStringPtrValidateUTF8 wire decodes a *string pointer as a String. +func consumeStringPtrValidateUTF8(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if !utf8.ValidString(v) { + return out, errInvalidUTF8{} + } + vp := p.StringPtr() + if *vp == nil { + *vp = new(string) + } + **vp = v + out.n = n + return out, nil +} + +var coderStringPtrValidateUTF8 = pointerCoderFuncs{ + size: sizeStringPtr, + marshal: appendStringPtrValidateUTF8, + unmarshal: consumeStringPtrValidateUTF8, + merge: mergeStringPtr, +} + +// sizeStringSlice returns the size of wire encoding a []string pointer as a repeated String. +func sizeStringSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.StringSlice() + for _, v := range s { + size += f.tagsize + protowire.SizeBytes(len(v)) + } + return size +} + +// appendStringSlice encodes a []string pointer as a repeated String. +func appendStringSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.StringSlice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + } + return b, nil +} + +// consumeStringSlice wire decodes a []string pointer as a repeated String. +func consumeStringSlice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.StringSlice() + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, v) + out.n = n + return out, nil +} + +var coderStringSlice = pointerCoderFuncs{ + size: sizeStringSlice, + marshal: appendStringSlice, + unmarshal: consumeStringSlice, + merge: mergeStringSlice, +} + +// appendStringSliceValidateUTF8 encodes a []string pointer as a repeated String. +func appendStringSliceValidateUTF8(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.StringSlice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendString(b, v) + if !utf8.ValidString(v) { + return b, errInvalidUTF8{} + } + } + return b, nil +} + +// consumeStringSliceValidateUTF8 wire decodes a []string pointer as a repeated String. +func consumeStringSliceValidateUTF8(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.StringSlice() + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if !utf8.ValidString(v) { + return out, errInvalidUTF8{} + } + *sp = append(*sp, v) + out.n = n + return out, nil +} + +var coderStringSliceValidateUTF8 = pointerCoderFuncs{ + size: sizeStringSlice, + marshal: appendStringSliceValidateUTF8, + unmarshal: consumeStringSliceValidateUTF8, + merge: mergeStringSlice, +} + +// sizeStringValue returns the size of wire encoding a string value as a String. +func sizeStringValue(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeBytes(len(v.String())) +} + +// appendStringValue encodes a string value as a String. +func appendStringValue(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendString(b, v.String()) + return b, nil +} + +// consumeStringValue decodes a string value as a String. +func consumeStringValue(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfString(string(v)), out, nil +} + +var coderStringValue = valueCoderFuncs{ + size: sizeStringValue, + marshal: appendStringValue, + unmarshal: consumeStringValue, + merge: mergeScalarValue, +} + +// appendStringValueValidateUTF8 encodes a string value as a String. +func appendStringValueValidateUTF8(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendString(b, v.String()) + if !utf8.ValidString(v.String()) { + return b, errInvalidUTF8{} + } + return b, nil +} + +// consumeStringValueValidateUTF8 decodes a string value as a String. +func consumeStringValueValidateUTF8(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + if !utf8.ValidString(v) { + return protoreflect.Value{}, out, errInvalidUTF8{} + } + out.n = n + return protoreflect.ValueOfString(string(v)), out, nil +} + +var coderStringValueValidateUTF8 = valueCoderFuncs{ + size: sizeStringValue, + marshal: appendStringValueValidateUTF8, + unmarshal: consumeStringValueValidateUTF8, + merge: mergeScalarValue, +} + +// sizeStringSliceValue returns the size of wire encoding a []string value as a repeated String. +func sizeStringSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeBytes(len(v.String())) + } + return size +} + +// appendStringSliceValue encodes a []string value as a repeated String. +func appendStringSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendString(b, v.String()) + } + return b, nil +} + +// consumeStringSliceValue wire decodes a []string value as a repeated String. +func consumeStringSliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp != protowire.BytesType { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeString(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfString(string(v))) + out.n = n + return listv, out, nil +} + +var coderStringSliceValue = valueCoderFuncs{ + size: sizeStringSliceValue, + marshal: appendStringSliceValue, + unmarshal: consumeStringSliceValue, + merge: mergeListValue, +} + +// sizeBytes returns the size of wire encoding a []byte pointer as a Bytes. +func sizeBytes(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Bytes() + return f.tagsize + protowire.SizeBytes(len(v)) +} + +// appendBytes wire encodes a []byte pointer as a Bytes. +func appendBytes(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Bytes() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendBytes(b, v) + return b, nil +} + +// consumeBytes wire decodes a []byte pointer as a Bytes. +func consumeBytes(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Bytes() = append(emptyBuf[:], v...) + out.n = n + return out, nil +} + +var coderBytes = pointerCoderFuncs{ + size: sizeBytes, + marshal: appendBytes, + unmarshal: consumeBytes, + merge: mergeBytes, +} + +// appendBytesValidateUTF8 wire encodes a []byte pointer as a Bytes. +func appendBytesValidateUTF8(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Bytes() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendBytes(b, v) + if !utf8.Valid(v) { + return b, errInvalidUTF8{} + } + return b, nil +} + +// consumeBytesValidateUTF8 wire decodes a []byte pointer as a Bytes. +func consumeBytesValidateUTF8(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if !utf8.Valid(v) { + return out, errInvalidUTF8{} + } + *p.Bytes() = append(emptyBuf[:], v...) + out.n = n + return out, nil +} + +var coderBytesValidateUTF8 = pointerCoderFuncs{ + size: sizeBytes, + marshal: appendBytesValidateUTF8, + unmarshal: consumeBytesValidateUTF8, + merge: mergeBytes, +} + +// sizeBytesNoZero returns the size of wire encoding a []byte pointer as a Bytes. +// The zero value is not encoded. +func sizeBytesNoZero(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := *p.Bytes() + if len(v) == 0 { + return 0 + } + return f.tagsize + protowire.SizeBytes(len(v)) +} + +// appendBytesNoZero wire encodes a []byte pointer as a Bytes. +// The zero value is not encoded. +func appendBytesNoZero(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Bytes() + if len(v) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendBytes(b, v) + return b, nil +} + +// consumeBytesNoZero wire decodes a []byte pointer as a Bytes. +// The zero value is not decoded. +func consumeBytesNoZero(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *p.Bytes() = append(([]byte)(nil), v...) + out.n = n + return out, nil +} + +var coderBytesNoZero = pointerCoderFuncs{ + size: sizeBytesNoZero, + marshal: appendBytesNoZero, + unmarshal: consumeBytesNoZero, + merge: mergeBytesNoZero, +} + +// appendBytesNoZeroValidateUTF8 wire encodes a []byte pointer as a Bytes. +// The zero value is not encoded. +func appendBytesNoZeroValidateUTF8(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + v := *p.Bytes() + if len(v) == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendBytes(b, v) + if !utf8.Valid(v) { + return b, errInvalidUTF8{} + } + return b, nil +} + +// consumeBytesNoZeroValidateUTF8 wire decodes a []byte pointer as a Bytes. +func consumeBytesNoZeroValidateUTF8(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if !utf8.Valid(v) { + return out, errInvalidUTF8{} + } + *p.Bytes() = append(([]byte)(nil), v...) + out.n = n + return out, nil +} + +var coderBytesNoZeroValidateUTF8 = pointerCoderFuncs{ + size: sizeBytesNoZero, + marshal: appendBytesNoZeroValidateUTF8, + unmarshal: consumeBytesNoZeroValidateUTF8, + merge: mergeBytesNoZero, +} + +// sizeBytesSlice returns the size of wire encoding a [][]byte pointer as a repeated Bytes. +func sizeBytesSlice(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + s := *p.BytesSlice() + for _, v := range s { + size += f.tagsize + protowire.SizeBytes(len(v)) + } + return size +} + +// appendBytesSlice encodes a [][]byte pointer as a repeated Bytes. +func appendBytesSlice(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.BytesSlice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendBytes(b, v) + } + return b, nil +} + +// consumeBytesSlice wire decodes a [][]byte pointer as a repeated Bytes. +func consumeBytesSlice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.BytesSlice() + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + *sp = append(*sp, append(emptyBuf[:], v...)) + out.n = n + return out, nil +} + +var coderBytesSlice = pointerCoderFuncs{ + size: sizeBytesSlice, + marshal: appendBytesSlice, + unmarshal: consumeBytesSlice, + merge: mergeBytesSlice, +} + +// appendBytesSliceValidateUTF8 encodes a [][]byte pointer as a repeated Bytes. +func appendBytesSliceValidateUTF8(b []byte, p pointer, f *coderFieldInfo, _ marshalOptions) ([]byte, error) { + s := *p.BytesSlice() + for _, v := range s { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendBytes(b, v) + if !utf8.Valid(v) { + return b, errInvalidUTF8{} + } + } + return b, nil +} + +// consumeBytesSliceValidateUTF8 wire decodes a [][]byte pointer as a repeated Bytes. +func consumeBytesSliceValidateUTF8(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + sp := p.BytesSlice() + if wtyp != protowire.BytesType { + return out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if !utf8.Valid(v) { + return out, errInvalidUTF8{} + } + *sp = append(*sp, append(emptyBuf[:], v...)) + out.n = n + return out, nil +} + +var coderBytesSliceValidateUTF8 = pointerCoderFuncs{ + size: sizeBytesSlice, + marshal: appendBytesSliceValidateUTF8, + unmarshal: consumeBytesSliceValidateUTF8, + merge: mergeBytesSlice, +} + +// sizeBytesValue returns the size of wire encoding a []byte value as a Bytes. +func sizeBytesValue(v protoreflect.Value, tagsize int, _ marshalOptions) int { + return tagsize + protowire.SizeBytes(len(v.Bytes())) +} + +// appendBytesValue encodes a []byte value as a Bytes. +func appendBytesValue(b []byte, v protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendBytes(b, v.Bytes()) + return b, nil +} + +// consumeBytesValue decodes a []byte value as a Bytes. +func consumeBytesValue(b []byte, _ protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + out.n = n + return protoreflect.ValueOfBytes(append(emptyBuf[:], v...)), out, nil +} + +var coderBytesValue = valueCoderFuncs{ + size: sizeBytesValue, + marshal: appendBytesValue, + unmarshal: consumeBytesValue, + merge: mergeBytesValue, +} + +// sizeBytesSliceValue returns the size of wire encoding a [][]byte value as a repeated Bytes. +func sizeBytesSliceValue(listv protoreflect.Value, tagsize int, _ marshalOptions) (size int) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + size += tagsize + protowire.SizeBytes(len(v.Bytes())) + } + return size +} + +// appendBytesSliceValue encodes a [][]byte value as a repeated Bytes. +func appendBytesSliceValue(b []byte, listv protoreflect.Value, wiretag uint64, _ marshalOptions) ([]byte, error) { + list := listv.List() + for i, llen := 0, list.Len(); i < llen; i++ { + v := list.Get(i) + b = protowire.AppendVarint(b, wiretag) + b = protowire.AppendBytes(b, v.Bytes()) + } + return b, nil +} + +// consumeBytesSliceValue wire decodes a [][]byte value as a repeated Bytes. +func consumeBytesSliceValue(b []byte, listv protoreflect.Value, _ protowire.Number, wtyp protowire.Type, _ unmarshalOptions) (_ protoreflect.Value, out unmarshalOutput, err error) { + list := listv.List() + if wtyp != protowire.BytesType { + return protoreflect.Value{}, out, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return protoreflect.Value{}, out, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfBytes(append(emptyBuf[:], v...))) + out.n = n + return listv, out, nil +} + +var coderBytesSliceValue = valueCoderFuncs{ + size: sizeBytesSliceValue, + marshal: appendBytesSliceValue, + unmarshal: consumeBytesSliceValue, + merge: mergeBytesListValue, +} + +// We append to an empty array rather than a nil []byte to get non-nil zero-length byte slices. +var emptyBuf [0]byte + +var wireTypes = map[protoreflect.Kind]protowire.Type{ + protoreflect.BoolKind: protowire.VarintType, + protoreflect.EnumKind: protowire.VarintType, + protoreflect.Int32Kind: protowire.VarintType, + protoreflect.Sint32Kind: protowire.VarintType, + protoreflect.Uint32Kind: protowire.VarintType, + protoreflect.Int64Kind: protowire.VarintType, + protoreflect.Sint64Kind: protowire.VarintType, + protoreflect.Uint64Kind: protowire.VarintType, + protoreflect.Sfixed32Kind: protowire.Fixed32Type, + protoreflect.Fixed32Kind: protowire.Fixed32Type, + protoreflect.FloatKind: protowire.Fixed32Type, + protoreflect.Sfixed64Kind: protowire.Fixed64Type, + protoreflect.Fixed64Kind: protowire.Fixed64Type, + protoreflect.DoubleKind: protowire.Fixed64Type, + protoreflect.StringKind: protowire.BytesType, + protoreflect.BytesKind: protowire.BytesType, + protoreflect.MessageKind: protowire.BytesType, + protoreflect.GroupKind: protowire.StartGroupType, +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_map.go b/vendor/google.golang.org/protobuf/internal/impl/codec_map.go new file mode 100644 index 00000000..35a67c25 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_map.go @@ -0,0 +1,388 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "errors" + "reflect" + "sort" + + "google.golang.org/protobuf/encoding/protowire" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type mapInfo struct { + goType reflect.Type + keyWiretag uint64 + valWiretag uint64 + keyFuncs valueCoderFuncs + valFuncs valueCoderFuncs + keyZero pref.Value + keyKind pref.Kind + conv *mapConverter +} + +func encoderFuncsForMap(fd pref.FieldDescriptor, ft reflect.Type) (valueMessage *MessageInfo, funcs pointerCoderFuncs) { + // TODO: Consider generating specialized map coders. + keyField := fd.MapKey() + valField := fd.MapValue() + keyWiretag := protowire.EncodeTag(1, wireTypes[keyField.Kind()]) + valWiretag := protowire.EncodeTag(2, wireTypes[valField.Kind()]) + keyFuncs := encoderFuncsForValue(keyField) + valFuncs := encoderFuncsForValue(valField) + conv := newMapConverter(ft, fd) + + mapi := &mapInfo{ + goType: ft, + keyWiretag: keyWiretag, + valWiretag: valWiretag, + keyFuncs: keyFuncs, + valFuncs: valFuncs, + keyZero: keyField.Default(), + keyKind: keyField.Kind(), + conv: conv, + } + if valField.Kind() == pref.MessageKind { + valueMessage = getMessageInfo(ft.Elem()) + } + + funcs = pointerCoderFuncs{ + size: func(p pointer, f *coderFieldInfo, opts marshalOptions) int { + return sizeMap(p.AsValueOf(ft).Elem(), mapi, f, opts) + }, + marshal: func(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + return appendMap(b, p.AsValueOf(ft).Elem(), mapi, f, opts) + }, + unmarshal: func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) { + mp := p.AsValueOf(ft) + if mp.Elem().IsNil() { + mp.Elem().Set(reflect.MakeMap(mapi.goType)) + } + if f.mi == nil { + return consumeMap(b, mp.Elem(), wtyp, mapi, f, opts) + } else { + return consumeMapOfMessage(b, mp.Elem(), wtyp, mapi, f, opts) + } + }, + } + switch valField.Kind() { + case pref.MessageKind: + funcs.merge = mergeMapOfMessage + case pref.BytesKind: + funcs.merge = mergeMapOfBytes + default: + funcs.merge = mergeMap + } + if valFuncs.isInit != nil { + funcs.isInit = func(p pointer, f *coderFieldInfo) error { + return isInitMap(p.AsValueOf(ft).Elem(), mapi, f) + } + } + return valueMessage, funcs +} + +const ( + mapKeyTagSize = 1 // field 1, tag size 1. + mapValTagSize = 1 // field 2, tag size 2. +) + +func sizeMap(mapv reflect.Value, mapi *mapInfo, f *coderFieldInfo, opts marshalOptions) int { + if mapv.Len() == 0 { + return 0 + } + n := 0 + iter := mapRange(mapv) + for iter.Next() { + key := mapi.conv.keyConv.PBValueOf(iter.Key()).MapKey() + keySize := mapi.keyFuncs.size(key.Value(), mapKeyTagSize, opts) + var valSize int + value := mapi.conv.valConv.PBValueOf(iter.Value()) + if f.mi == nil { + valSize = mapi.valFuncs.size(value, mapValTagSize, opts) + } else { + p := pointerOfValue(iter.Value()) + valSize += mapValTagSize + valSize += protowire.SizeBytes(f.mi.sizePointer(p, opts)) + } + n += f.tagsize + protowire.SizeBytes(keySize+valSize) + } + return n +} + +func consumeMap(b []byte, mapv reflect.Value, wtyp protowire.Type, mapi *mapInfo, f *coderFieldInfo, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + var ( + key = mapi.keyZero + val = mapi.conv.valConv.New() + ) + for len(b) > 0 { + num, wtyp, n := protowire.ConsumeTag(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if num > protowire.MaxValidNumber { + return out, errors.New("invalid field number") + } + b = b[n:] + err := errUnknown + switch num { + case 1: + var v pref.Value + var o unmarshalOutput + v, o, err = mapi.keyFuncs.unmarshal(b, key, num, wtyp, opts) + if err != nil { + break + } + key = v + n = o.n + case 2: + var v pref.Value + var o unmarshalOutput + v, o, err = mapi.valFuncs.unmarshal(b, val, num, wtyp, opts) + if err != nil { + break + } + val = v + n = o.n + } + if err == errUnknown { + n = protowire.ConsumeFieldValue(num, wtyp, b) + if n < 0 { + return out, protowire.ParseError(n) + } + } else if err != nil { + return out, err + } + b = b[n:] + } + mapv.SetMapIndex(mapi.conv.keyConv.GoValueOf(key), mapi.conv.valConv.GoValueOf(val)) + out.n = n + return out, nil +} + +func consumeMapOfMessage(b []byte, mapv reflect.Value, wtyp protowire.Type, mapi *mapInfo, f *coderFieldInfo, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.BytesType { + return out, errUnknown + } + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + var ( + key = mapi.keyZero + val = reflect.New(f.mi.GoReflectType.Elem()) + ) + for len(b) > 0 { + num, wtyp, n := protowire.ConsumeTag(b) + if n < 0 { + return out, protowire.ParseError(n) + } + if num > protowire.MaxValidNumber { + return out, errors.New("invalid field number") + } + b = b[n:] + err := errUnknown + switch num { + case 1: + var v pref.Value + var o unmarshalOutput + v, o, err = mapi.keyFuncs.unmarshal(b, key, num, wtyp, opts) + if err != nil { + break + } + key = v + n = o.n + case 2: + if wtyp != protowire.BytesType { + break + } + var v []byte + v, n = protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + var o unmarshalOutput + o, err = f.mi.unmarshalPointer(v, pointerOfValue(val), 0, opts) + if o.initialized { + // Consider this map item initialized so long as we see + // an initialized value. + out.initialized = true + } + } + if err == errUnknown { + n = protowire.ConsumeFieldValue(num, wtyp, b) + if n < 0 { + return out, protowire.ParseError(n) + } + } else if err != nil { + return out, err + } + b = b[n:] + } + mapv.SetMapIndex(mapi.conv.keyConv.GoValueOf(key), val) + out.n = n + return out, nil +} + +func appendMapItem(b []byte, keyrv, valrv reflect.Value, mapi *mapInfo, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + if f.mi == nil { + key := mapi.conv.keyConv.PBValueOf(keyrv).MapKey() + val := mapi.conv.valConv.PBValueOf(valrv) + size := 0 + size += mapi.keyFuncs.size(key.Value(), mapKeyTagSize, opts) + size += mapi.valFuncs.size(val, mapValTagSize, opts) + b = protowire.AppendVarint(b, uint64(size)) + b, err := mapi.keyFuncs.marshal(b, key.Value(), mapi.keyWiretag, opts) + if err != nil { + return nil, err + } + return mapi.valFuncs.marshal(b, val, mapi.valWiretag, opts) + } else { + key := mapi.conv.keyConv.PBValueOf(keyrv).MapKey() + val := pointerOfValue(valrv) + valSize := f.mi.sizePointer(val, opts) + size := 0 + size += mapi.keyFuncs.size(key.Value(), mapKeyTagSize, opts) + size += mapValTagSize + protowire.SizeBytes(valSize) + b = protowire.AppendVarint(b, uint64(size)) + b, err := mapi.keyFuncs.marshal(b, key.Value(), mapi.keyWiretag, opts) + if err != nil { + return nil, err + } + b = protowire.AppendVarint(b, mapi.valWiretag) + b = protowire.AppendVarint(b, uint64(valSize)) + return f.mi.marshalAppendPointer(b, val, opts) + } +} + +func appendMap(b []byte, mapv reflect.Value, mapi *mapInfo, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + if mapv.Len() == 0 { + return b, nil + } + if opts.Deterministic() { + return appendMapDeterministic(b, mapv, mapi, f, opts) + } + iter := mapRange(mapv) + for iter.Next() { + var err error + b = protowire.AppendVarint(b, f.wiretag) + b, err = appendMapItem(b, iter.Key(), iter.Value(), mapi, f, opts) + if err != nil { + return b, err + } + } + return b, nil +} + +func appendMapDeterministic(b []byte, mapv reflect.Value, mapi *mapInfo, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + keys := mapv.MapKeys() + sort.Slice(keys, func(i, j int) bool { + switch keys[i].Kind() { + case reflect.Bool: + return !keys[i].Bool() && keys[j].Bool() + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + return keys[i].Int() < keys[j].Int() + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: + return keys[i].Uint() < keys[j].Uint() + case reflect.Float32, reflect.Float64: + return keys[i].Float() < keys[j].Float() + case reflect.String: + return keys[i].String() < keys[j].String() + default: + panic("invalid kind: " + keys[i].Kind().String()) + } + }) + for _, key := range keys { + var err error + b = protowire.AppendVarint(b, f.wiretag) + b, err = appendMapItem(b, key, mapv.MapIndex(key), mapi, f, opts) + if err != nil { + return b, err + } + } + return b, nil +} + +func isInitMap(mapv reflect.Value, mapi *mapInfo, f *coderFieldInfo) error { + if mi := f.mi; mi != nil { + mi.init() + if !mi.needsInitCheck { + return nil + } + iter := mapRange(mapv) + for iter.Next() { + val := pointerOfValue(iter.Value()) + if err := mi.checkInitializedPointer(val); err != nil { + return err + } + } + } else { + iter := mapRange(mapv) + for iter.Next() { + val := mapi.conv.valConv.PBValueOf(iter.Value()) + if err := mapi.valFuncs.isInit(val); err != nil { + return err + } + } + } + return nil +} + +func mergeMap(dst, src pointer, f *coderFieldInfo, opts mergeOptions) { + dstm := dst.AsValueOf(f.ft).Elem() + srcm := src.AsValueOf(f.ft).Elem() + if srcm.Len() == 0 { + return + } + if dstm.IsNil() { + dstm.Set(reflect.MakeMap(f.ft)) + } + iter := mapRange(srcm) + for iter.Next() { + dstm.SetMapIndex(iter.Key(), iter.Value()) + } +} + +func mergeMapOfBytes(dst, src pointer, f *coderFieldInfo, opts mergeOptions) { + dstm := dst.AsValueOf(f.ft).Elem() + srcm := src.AsValueOf(f.ft).Elem() + if srcm.Len() == 0 { + return + } + if dstm.IsNil() { + dstm.Set(reflect.MakeMap(f.ft)) + } + iter := mapRange(srcm) + for iter.Next() { + dstm.SetMapIndex(iter.Key(), reflect.ValueOf(append(emptyBuf[:], iter.Value().Bytes()...))) + } +} + +func mergeMapOfMessage(dst, src pointer, f *coderFieldInfo, opts mergeOptions) { + dstm := dst.AsValueOf(f.ft).Elem() + srcm := src.AsValueOf(f.ft).Elem() + if srcm.Len() == 0 { + return + } + if dstm.IsNil() { + dstm.Set(reflect.MakeMap(f.ft)) + } + iter := mapRange(srcm) + for iter.Next() { + val := reflect.New(f.ft.Elem().Elem()) + if f.mi != nil { + f.mi.mergePointer(pointerOfValue(val), pointerOfValue(iter.Value()), opts) + } else { + opts.Merge(asMessage(val), asMessage(iter.Value())) + } + dstm.SetMapIndex(iter.Key(), val) + } +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_map_go111.go b/vendor/google.golang.org/protobuf/internal/impl/codec_map_go111.go new file mode 100644 index 00000000..2706bb67 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_map_go111.go @@ -0,0 +1,37 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !go1.12 + +package impl + +import "reflect" + +type mapIter struct { + v reflect.Value + keys []reflect.Value +} + +// mapRange provides a less-efficient equivalent to +// the Go 1.12 reflect.Value.MapRange method. +func mapRange(v reflect.Value) *mapIter { + return &mapIter{v: v} +} + +func (i *mapIter) Next() bool { + if i.keys == nil { + i.keys = i.v.MapKeys() + } else { + i.keys = i.keys[1:] + } + return len(i.keys) > 0 +} + +func (i *mapIter) Key() reflect.Value { + return i.keys[0] +} + +func (i *mapIter) Value() reflect.Value { + return i.v.MapIndex(i.keys[0]) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_map_go112.go b/vendor/google.golang.org/protobuf/internal/impl/codec_map_go112.go new file mode 100644 index 00000000..1533ef60 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_map_go112.go @@ -0,0 +1,11 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build go1.12 + +package impl + +import "reflect" + +func mapRange(v reflect.Value) *reflect.MapIter { return v.MapRange() } diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_message.go b/vendor/google.golang.org/protobuf/internal/impl/codec_message.go new file mode 100644 index 00000000..0e176d56 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_message.go @@ -0,0 +1,159 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + "sort" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/fieldsort" + pref "google.golang.org/protobuf/reflect/protoreflect" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +// coderMessageInfo contains per-message information used by the fast-path functions. +// This is a different type from MessageInfo to keep MessageInfo as general-purpose as +// possible. +type coderMessageInfo struct { + methods piface.Methods + + orderedCoderFields []*coderFieldInfo + denseCoderFields []*coderFieldInfo + coderFields map[protowire.Number]*coderFieldInfo + sizecacheOffset offset + unknownOffset offset + extensionOffset offset + needsInitCheck bool + isMessageSet bool + numRequiredFields uint8 +} + +type coderFieldInfo struct { + funcs pointerCoderFuncs // fast-path per-field functions + mi *MessageInfo // field's message + ft reflect.Type + validation validationInfo // information used by message validation + num pref.FieldNumber // field number + offset offset // struct field offset + wiretag uint64 // field tag (number + wire type) + tagsize int // size of the varint-encoded tag + isPointer bool // true if IsNil may be called on the struct field + isRequired bool // true if field is required +} + +func (mi *MessageInfo) makeCoderMethods(t reflect.Type, si structInfo) { + mi.sizecacheOffset = si.sizecacheOffset + mi.unknownOffset = si.unknownOffset + mi.extensionOffset = si.extensionOffset + + mi.coderFields = make(map[protowire.Number]*coderFieldInfo) + fields := mi.Desc.Fields() + preallocFields := make([]coderFieldInfo, fields.Len()) + for i := 0; i < fields.Len(); i++ { + fd := fields.Get(i) + + fs := si.fieldsByNumber[fd.Number()] + isOneof := fd.ContainingOneof() != nil && !fd.ContainingOneof().IsSynthetic() + if isOneof { + fs = si.oneofsByName[fd.ContainingOneof().Name()] + } + ft := fs.Type + var wiretag uint64 + if !fd.IsPacked() { + wiretag = protowire.EncodeTag(fd.Number(), wireTypes[fd.Kind()]) + } else { + wiretag = protowire.EncodeTag(fd.Number(), protowire.BytesType) + } + var fieldOffset offset + var funcs pointerCoderFuncs + var childMessage *MessageInfo + switch { + case isOneof: + fieldOffset = offsetOf(fs, mi.Exporter) + case fd.IsWeak(): + fieldOffset = si.weakOffset + funcs = makeWeakMessageFieldCoder(fd) + default: + fieldOffset = offsetOf(fs, mi.Exporter) + childMessage, funcs = fieldCoder(fd, ft) + } + cf := &preallocFields[i] + *cf = coderFieldInfo{ + num: fd.Number(), + offset: fieldOffset, + wiretag: wiretag, + ft: ft, + tagsize: protowire.SizeVarint(wiretag), + funcs: funcs, + mi: childMessage, + validation: newFieldValidationInfo(mi, si, fd, ft), + isPointer: fd.Cardinality() == pref.Repeated || fd.HasPresence(), + isRequired: fd.Cardinality() == pref.Required, + } + mi.orderedCoderFields = append(mi.orderedCoderFields, cf) + mi.coderFields[cf.num] = cf + } + for i, oneofs := 0, mi.Desc.Oneofs(); i < oneofs.Len(); i++ { + if od := oneofs.Get(i); !od.IsSynthetic() { + mi.initOneofFieldCoders(od, si) + } + } + if messageset.IsMessageSet(mi.Desc) { + if !mi.extensionOffset.IsValid() { + panic(fmt.Sprintf("%v: MessageSet with no extensions field", mi.Desc.FullName())) + } + if !mi.unknownOffset.IsValid() { + panic(fmt.Sprintf("%v: MessageSet with no unknown field", mi.Desc.FullName())) + } + mi.isMessageSet = true + } + sort.Slice(mi.orderedCoderFields, func(i, j int) bool { + return mi.orderedCoderFields[i].num < mi.orderedCoderFields[j].num + }) + + var maxDense pref.FieldNumber + for _, cf := range mi.orderedCoderFields { + if cf.num >= 16 && cf.num >= 2*maxDense { + break + } + maxDense = cf.num + } + mi.denseCoderFields = make([]*coderFieldInfo, maxDense+1) + for _, cf := range mi.orderedCoderFields { + if int(cf.num) >= len(mi.denseCoderFields) { + break + } + mi.denseCoderFields[cf.num] = cf + } + + // To preserve compatibility with historic wire output, marshal oneofs last. + if mi.Desc.Oneofs().Len() > 0 { + sort.Slice(mi.orderedCoderFields, func(i, j int) bool { + fi := fields.ByNumber(mi.orderedCoderFields[i].num) + fj := fields.ByNumber(mi.orderedCoderFields[j].num) + return fieldsort.Less(fi, fj) + }) + } + + mi.needsInitCheck = needsInitCheck(mi.Desc) + if mi.methods.Marshal == nil && mi.methods.Size == nil { + mi.methods.Flags |= piface.SupportMarshalDeterministic + mi.methods.Marshal = mi.marshal + mi.methods.Size = mi.size + } + if mi.methods.Unmarshal == nil { + mi.methods.Flags |= piface.SupportUnmarshalDiscardUnknown + mi.methods.Unmarshal = mi.unmarshal + } + if mi.methods.CheckInitialized == nil { + mi.methods.CheckInitialized = mi.checkInitialized + } + if mi.methods.Merge == nil { + mi.methods.Merge = mi.merge + } +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_messageset.go b/vendor/google.golang.org/protobuf/internal/impl/codec_messageset.go new file mode 100644 index 00000000..cfb68e12 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_messageset.go @@ -0,0 +1,120 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "sort" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/flags" +) + +func sizeMessageSet(mi *MessageInfo, p pointer, opts marshalOptions) (size int) { + if !flags.ProtoLegacy { + return 0 + } + + ext := *p.Apply(mi.extensionOffset).Extensions() + for _, x := range ext { + xi := getExtensionFieldInfo(x.Type()) + if xi.funcs.size == nil { + continue + } + num, _ := protowire.DecodeTag(xi.wiretag) + size += messageset.SizeField(num) + size += xi.funcs.size(x.Value(), protowire.SizeTag(messageset.FieldMessage), opts) + } + + unknown := *p.Apply(mi.unknownOffset).Bytes() + size += messageset.SizeUnknown(unknown) + + return size +} + +func marshalMessageSet(mi *MessageInfo, b []byte, p pointer, opts marshalOptions) ([]byte, error) { + if !flags.ProtoLegacy { + return b, errors.New("no support for message_set_wire_format") + } + + ext := *p.Apply(mi.extensionOffset).Extensions() + switch len(ext) { + case 0: + case 1: + // Fast-path for one extension: Don't bother sorting the keys. + for _, x := range ext { + var err error + b, err = marshalMessageSetField(mi, b, x, opts) + if err != nil { + return b, err + } + } + default: + // Sort the keys to provide a deterministic encoding. + // Not sure this is required, but the old code does it. + keys := make([]int, 0, len(ext)) + for k := range ext { + keys = append(keys, int(k)) + } + sort.Ints(keys) + for _, k := range keys { + var err error + b, err = marshalMessageSetField(mi, b, ext[int32(k)], opts) + if err != nil { + return b, err + } + } + } + + unknown := *p.Apply(mi.unknownOffset).Bytes() + b, err := messageset.AppendUnknown(b, unknown) + if err != nil { + return b, err + } + + return b, nil +} + +func marshalMessageSetField(mi *MessageInfo, b []byte, x ExtensionField, opts marshalOptions) ([]byte, error) { + xi := getExtensionFieldInfo(x.Type()) + num, _ := protowire.DecodeTag(xi.wiretag) + b = messageset.AppendFieldStart(b, num) + b, err := xi.funcs.marshal(b, x.Value(), protowire.EncodeTag(messageset.FieldMessage, protowire.BytesType), opts) + if err != nil { + return b, err + } + b = messageset.AppendFieldEnd(b) + return b, nil +} + +func unmarshalMessageSet(mi *MessageInfo, b []byte, p pointer, opts unmarshalOptions) (out unmarshalOutput, err error) { + if !flags.ProtoLegacy { + return out, errors.New("no support for message_set_wire_format") + } + + ep := p.Apply(mi.extensionOffset).Extensions() + if *ep == nil { + *ep = make(map[int32]ExtensionField) + } + ext := *ep + unknown := p.Apply(mi.unknownOffset).Bytes() + initialized := true + err = messageset.Unmarshal(b, true, func(num protowire.Number, v []byte) error { + o, err := mi.unmarshalExtension(v, num, protowire.BytesType, ext, opts) + if err == errUnknown { + *unknown = protowire.AppendTag(*unknown, num, protowire.BytesType) + *unknown = append(*unknown, v...) + return nil + } + if !o.initialized { + initialized = false + } + return err + }) + out.n = len(b) + out.initialized = initialized + return out, err +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_reflect.go b/vendor/google.golang.org/protobuf/internal/impl/codec_reflect.go new file mode 100644 index 00000000..86f7dc3c --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_reflect.go @@ -0,0 +1,209 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build purego appengine + +package impl + +import ( + "reflect" + + "google.golang.org/protobuf/encoding/protowire" +) + +func sizeEnum(p pointer, f *coderFieldInfo, _ marshalOptions) (size int) { + v := p.v.Elem().Int() + return f.tagsize + protowire.SizeVarint(uint64(v)) +} + +func appendEnum(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + v := p.v.Elem().Int() + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(v)) + return b, nil +} + +func consumeEnum(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, _ unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return out, protowire.ParseError(n) + } + p.v.Elem().SetInt(int64(v)) + out.n = n + return out, nil +} + +func mergeEnum(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + dst.v.Elem().Set(src.v.Elem()) +} + +var coderEnum = pointerCoderFuncs{ + size: sizeEnum, + marshal: appendEnum, + unmarshal: consumeEnum, + merge: mergeEnum, +} + +func sizeEnumNoZero(p pointer, f *coderFieldInfo, opts marshalOptions) (size int) { + if p.v.Elem().Int() == 0 { + return 0 + } + return sizeEnum(p, f, opts) +} + +func appendEnumNoZero(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + if p.v.Elem().Int() == 0 { + return b, nil + } + return appendEnum(b, p, f, opts) +} + +func mergeEnumNoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + if src.v.Elem().Int() != 0 { + dst.v.Elem().Set(src.v.Elem()) + } +} + +var coderEnumNoZero = pointerCoderFuncs{ + size: sizeEnumNoZero, + marshal: appendEnumNoZero, + unmarshal: consumeEnum, + merge: mergeEnumNoZero, +} + +func sizeEnumPtr(p pointer, f *coderFieldInfo, opts marshalOptions) (size int) { + return sizeEnum(pointer{p.v.Elem()}, f, opts) +} + +func appendEnumPtr(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + return appendEnum(b, pointer{p.v.Elem()}, f, opts) +} + +func consumeEnumPtr(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (out unmarshalOutput, err error) { + if wtyp != protowire.VarintType { + return out, errUnknown + } + if p.v.Elem().IsNil() { + p.v.Elem().Set(reflect.New(p.v.Elem().Type().Elem())) + } + return consumeEnum(b, pointer{p.v.Elem()}, wtyp, f, opts) +} + +func mergeEnumPtr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + if !src.v.Elem().IsNil() { + v := reflect.New(dst.v.Type().Elem().Elem()) + v.Elem().Set(src.v.Elem().Elem()) + dst.v.Elem().Set(v) + } +} + +var coderEnumPtr = pointerCoderFuncs{ + size: sizeEnumPtr, + marshal: appendEnumPtr, + unmarshal: consumeEnumPtr, + merge: mergeEnumPtr, +} + +func sizeEnumSlice(p pointer, f *coderFieldInfo, opts marshalOptions) (size int) { + s := p.v.Elem() + for i, llen := 0, s.Len(); i < llen; i++ { + size += protowire.SizeVarint(uint64(s.Index(i).Int())) + f.tagsize + } + return size +} + +func appendEnumSlice(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + s := p.v.Elem() + for i, llen := 0, s.Len(); i < llen; i++ { + b = protowire.AppendVarint(b, f.wiretag) + b = protowire.AppendVarint(b, uint64(s.Index(i).Int())) + } + return b, nil +} + +func consumeEnumSlice(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (out unmarshalOutput, err error) { + s := p.v.Elem() + if wtyp == protowire.BytesType { + b, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, protowire.ParseError(n) + } + for len(b) > 0 { + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return out, protowire.ParseError(n) + } + rv := reflect.New(s.Type().Elem()).Elem() + rv.SetInt(int64(v)) + s.Set(reflect.Append(s, rv)) + b = b[n:] + } + out.n = n + return out, nil + } + if wtyp != protowire.VarintType { + return out, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return out, protowire.ParseError(n) + } + rv := reflect.New(s.Type().Elem()).Elem() + rv.SetInt(int64(v)) + s.Set(reflect.Append(s, rv)) + out.n = n + return out, nil +} + +func mergeEnumSlice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + dst.v.Elem().Set(reflect.AppendSlice(dst.v.Elem(), src.v.Elem())) +} + +var coderEnumSlice = pointerCoderFuncs{ + size: sizeEnumSlice, + marshal: appendEnumSlice, + unmarshal: consumeEnumSlice, + merge: mergeEnumSlice, +} + +func sizeEnumPackedSlice(p pointer, f *coderFieldInfo, opts marshalOptions) (size int) { + s := p.v.Elem() + llen := s.Len() + if llen == 0 { + return 0 + } + n := 0 + for i := 0; i < llen; i++ { + n += protowire.SizeVarint(uint64(s.Index(i).Int())) + } + return f.tagsize + protowire.SizeBytes(n) +} + +func appendEnumPackedSlice(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) { + s := p.v.Elem() + llen := s.Len() + if llen == 0 { + return b, nil + } + b = protowire.AppendVarint(b, f.wiretag) + n := 0 + for i := 0; i < llen; i++ { + n += protowire.SizeVarint(uint64(s.Index(i).Int())) + } + b = protowire.AppendVarint(b, uint64(n)) + for i := 0; i < llen; i++ { + b = protowire.AppendVarint(b, uint64(s.Index(i).Int())) + } + return b, nil +} + +var coderEnumPackedSlice = pointerCoderFuncs{ + size: sizeEnumPackedSlice, + marshal: appendEnumPackedSlice, + unmarshal: consumeEnumSlice, + merge: mergeEnumSlice, +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_tables.go b/vendor/google.golang.org/protobuf/internal/impl/codec_tables.go new file mode 100644 index 00000000..e8997123 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_tables.go @@ -0,0 +1,557 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/strs" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +// pointerCoderFuncs is a set of pointer encoding functions. +type pointerCoderFuncs struct { + mi *MessageInfo + size func(p pointer, f *coderFieldInfo, opts marshalOptions) int + marshal func(b []byte, p pointer, f *coderFieldInfo, opts marshalOptions) ([]byte, error) + unmarshal func(b []byte, p pointer, wtyp protowire.Type, f *coderFieldInfo, opts unmarshalOptions) (unmarshalOutput, error) + isInit func(p pointer, f *coderFieldInfo) error + merge func(dst, src pointer, f *coderFieldInfo, opts mergeOptions) +} + +// valueCoderFuncs is a set of protoreflect.Value encoding functions. +type valueCoderFuncs struct { + size func(v pref.Value, tagsize int, opts marshalOptions) int + marshal func(b []byte, v pref.Value, wiretag uint64, opts marshalOptions) ([]byte, error) + unmarshal func(b []byte, v pref.Value, num protowire.Number, wtyp protowire.Type, opts unmarshalOptions) (pref.Value, unmarshalOutput, error) + isInit func(v pref.Value) error + merge func(dst, src pref.Value, opts mergeOptions) pref.Value +} + +// fieldCoder returns pointer functions for a field, used for operating on +// struct fields. +func fieldCoder(fd pref.FieldDescriptor, ft reflect.Type) (*MessageInfo, pointerCoderFuncs) { + switch { + case fd.IsMap(): + return encoderFuncsForMap(fd, ft) + case fd.Cardinality() == pref.Repeated && !fd.IsPacked(): + // Repeated fields (not packed). + if ft.Kind() != reflect.Slice { + break + } + ft := ft.Elem() + switch fd.Kind() { + case pref.BoolKind: + if ft.Kind() == reflect.Bool { + return nil, coderBoolSlice + } + case pref.EnumKind: + if ft.Kind() == reflect.Int32 { + return nil, coderEnumSlice + } + case pref.Int32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderInt32Slice + } + case pref.Sint32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSint32Slice + } + case pref.Uint32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderUint32Slice + } + case pref.Int64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderInt64Slice + } + case pref.Sint64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSint64Slice + } + case pref.Uint64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderUint64Slice + } + case pref.Sfixed32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSfixed32Slice + } + case pref.Fixed32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderFixed32Slice + } + case pref.FloatKind: + if ft.Kind() == reflect.Float32 { + return nil, coderFloatSlice + } + case pref.Sfixed64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSfixed64Slice + } + case pref.Fixed64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderFixed64Slice + } + case pref.DoubleKind: + if ft.Kind() == reflect.Float64 { + return nil, coderDoubleSlice + } + case pref.StringKind: + if ft.Kind() == reflect.String && strs.EnforceUTF8(fd) { + return nil, coderStringSliceValidateUTF8 + } + if ft.Kind() == reflect.String { + return nil, coderStringSlice + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 && strs.EnforceUTF8(fd) { + return nil, coderBytesSliceValidateUTF8 + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 { + return nil, coderBytesSlice + } + case pref.BytesKind: + if ft.Kind() == reflect.String { + return nil, coderStringSlice + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 { + return nil, coderBytesSlice + } + case pref.MessageKind: + return getMessageInfo(ft), makeMessageSliceFieldCoder(fd, ft) + case pref.GroupKind: + return getMessageInfo(ft), makeGroupSliceFieldCoder(fd, ft) + } + case fd.Cardinality() == pref.Repeated && fd.IsPacked(): + // Packed repeated fields. + // + // Only repeated fields of primitive numeric types + // (Varint, Fixed32, or Fixed64 wire type) can be packed. + if ft.Kind() != reflect.Slice { + break + } + ft := ft.Elem() + switch fd.Kind() { + case pref.BoolKind: + if ft.Kind() == reflect.Bool { + return nil, coderBoolPackedSlice + } + case pref.EnumKind: + if ft.Kind() == reflect.Int32 { + return nil, coderEnumPackedSlice + } + case pref.Int32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderInt32PackedSlice + } + case pref.Sint32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSint32PackedSlice + } + case pref.Uint32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderUint32PackedSlice + } + case pref.Int64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderInt64PackedSlice + } + case pref.Sint64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSint64PackedSlice + } + case pref.Uint64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderUint64PackedSlice + } + case pref.Sfixed32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSfixed32PackedSlice + } + case pref.Fixed32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderFixed32PackedSlice + } + case pref.FloatKind: + if ft.Kind() == reflect.Float32 { + return nil, coderFloatPackedSlice + } + case pref.Sfixed64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSfixed64PackedSlice + } + case pref.Fixed64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderFixed64PackedSlice + } + case pref.DoubleKind: + if ft.Kind() == reflect.Float64 { + return nil, coderDoublePackedSlice + } + } + case fd.Kind() == pref.MessageKind: + return getMessageInfo(ft), makeMessageFieldCoder(fd, ft) + case fd.Kind() == pref.GroupKind: + return getMessageInfo(ft), makeGroupFieldCoder(fd, ft) + case fd.Syntax() == pref.Proto3 && fd.ContainingOneof() == nil: + // Populated oneof fields always encode even if set to the zero value, + // which normally are not encoded in proto3. + switch fd.Kind() { + case pref.BoolKind: + if ft.Kind() == reflect.Bool { + return nil, coderBoolNoZero + } + case pref.EnumKind: + if ft.Kind() == reflect.Int32 { + return nil, coderEnumNoZero + } + case pref.Int32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderInt32NoZero + } + case pref.Sint32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSint32NoZero + } + case pref.Uint32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderUint32NoZero + } + case pref.Int64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderInt64NoZero + } + case pref.Sint64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSint64NoZero + } + case pref.Uint64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderUint64NoZero + } + case pref.Sfixed32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSfixed32NoZero + } + case pref.Fixed32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderFixed32NoZero + } + case pref.FloatKind: + if ft.Kind() == reflect.Float32 { + return nil, coderFloatNoZero + } + case pref.Sfixed64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSfixed64NoZero + } + case pref.Fixed64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderFixed64NoZero + } + case pref.DoubleKind: + if ft.Kind() == reflect.Float64 { + return nil, coderDoubleNoZero + } + case pref.StringKind: + if ft.Kind() == reflect.String && strs.EnforceUTF8(fd) { + return nil, coderStringNoZeroValidateUTF8 + } + if ft.Kind() == reflect.String { + return nil, coderStringNoZero + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 && strs.EnforceUTF8(fd) { + return nil, coderBytesNoZeroValidateUTF8 + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 { + return nil, coderBytesNoZero + } + case pref.BytesKind: + if ft.Kind() == reflect.String { + return nil, coderStringNoZero + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 { + return nil, coderBytesNoZero + } + } + case ft.Kind() == reflect.Ptr: + ft := ft.Elem() + switch fd.Kind() { + case pref.BoolKind: + if ft.Kind() == reflect.Bool { + return nil, coderBoolPtr + } + case pref.EnumKind: + if ft.Kind() == reflect.Int32 { + return nil, coderEnumPtr + } + case pref.Int32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderInt32Ptr + } + case pref.Sint32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSint32Ptr + } + case pref.Uint32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderUint32Ptr + } + case pref.Int64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderInt64Ptr + } + case pref.Sint64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSint64Ptr + } + case pref.Uint64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderUint64Ptr + } + case pref.Sfixed32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSfixed32Ptr + } + case pref.Fixed32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderFixed32Ptr + } + case pref.FloatKind: + if ft.Kind() == reflect.Float32 { + return nil, coderFloatPtr + } + case pref.Sfixed64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSfixed64Ptr + } + case pref.Fixed64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderFixed64Ptr + } + case pref.DoubleKind: + if ft.Kind() == reflect.Float64 { + return nil, coderDoublePtr + } + case pref.StringKind: + if ft.Kind() == reflect.String && strs.EnforceUTF8(fd) { + return nil, coderStringPtrValidateUTF8 + } + if ft.Kind() == reflect.String { + return nil, coderStringPtr + } + case pref.BytesKind: + if ft.Kind() == reflect.String { + return nil, coderStringPtr + } + } + default: + switch fd.Kind() { + case pref.BoolKind: + if ft.Kind() == reflect.Bool { + return nil, coderBool + } + case pref.EnumKind: + if ft.Kind() == reflect.Int32 { + return nil, coderEnum + } + case pref.Int32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderInt32 + } + case pref.Sint32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSint32 + } + case pref.Uint32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderUint32 + } + case pref.Int64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderInt64 + } + case pref.Sint64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSint64 + } + case pref.Uint64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderUint64 + } + case pref.Sfixed32Kind: + if ft.Kind() == reflect.Int32 { + return nil, coderSfixed32 + } + case pref.Fixed32Kind: + if ft.Kind() == reflect.Uint32 { + return nil, coderFixed32 + } + case pref.FloatKind: + if ft.Kind() == reflect.Float32 { + return nil, coderFloat + } + case pref.Sfixed64Kind: + if ft.Kind() == reflect.Int64 { + return nil, coderSfixed64 + } + case pref.Fixed64Kind: + if ft.Kind() == reflect.Uint64 { + return nil, coderFixed64 + } + case pref.DoubleKind: + if ft.Kind() == reflect.Float64 { + return nil, coderDouble + } + case pref.StringKind: + if ft.Kind() == reflect.String && strs.EnforceUTF8(fd) { + return nil, coderStringValidateUTF8 + } + if ft.Kind() == reflect.String { + return nil, coderString + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 && strs.EnforceUTF8(fd) { + return nil, coderBytesValidateUTF8 + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 { + return nil, coderBytes + } + case pref.BytesKind: + if ft.Kind() == reflect.String { + return nil, coderString + } + if ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 { + return nil, coderBytes + } + } + } + panic(fmt.Sprintf("invalid type: no encoder for %v %v %v/%v", fd.FullName(), fd.Cardinality(), fd.Kind(), ft)) +} + +// encoderFuncsForValue returns value functions for a field, used for +// extension values and map encoding. +func encoderFuncsForValue(fd pref.FieldDescriptor) valueCoderFuncs { + switch { + case fd.Cardinality() == pref.Repeated && !fd.IsPacked(): + switch fd.Kind() { + case pref.BoolKind: + return coderBoolSliceValue + case pref.EnumKind: + return coderEnumSliceValue + case pref.Int32Kind: + return coderInt32SliceValue + case pref.Sint32Kind: + return coderSint32SliceValue + case pref.Uint32Kind: + return coderUint32SliceValue + case pref.Int64Kind: + return coderInt64SliceValue + case pref.Sint64Kind: + return coderSint64SliceValue + case pref.Uint64Kind: + return coderUint64SliceValue + case pref.Sfixed32Kind: + return coderSfixed32SliceValue + case pref.Fixed32Kind: + return coderFixed32SliceValue + case pref.FloatKind: + return coderFloatSliceValue + case pref.Sfixed64Kind: + return coderSfixed64SliceValue + case pref.Fixed64Kind: + return coderFixed64SliceValue + case pref.DoubleKind: + return coderDoubleSliceValue + case pref.StringKind: + // We don't have a UTF-8 validating coder for repeated string fields. + // Value coders are used for extensions and maps. + // Extensions are never proto3, and maps never contain lists. + return coderStringSliceValue + case pref.BytesKind: + return coderBytesSliceValue + case pref.MessageKind: + return coderMessageSliceValue + case pref.GroupKind: + return coderGroupSliceValue + } + case fd.Cardinality() == pref.Repeated && fd.IsPacked(): + switch fd.Kind() { + case pref.BoolKind: + return coderBoolPackedSliceValue + case pref.EnumKind: + return coderEnumPackedSliceValue + case pref.Int32Kind: + return coderInt32PackedSliceValue + case pref.Sint32Kind: + return coderSint32PackedSliceValue + case pref.Uint32Kind: + return coderUint32PackedSliceValue + case pref.Int64Kind: + return coderInt64PackedSliceValue + case pref.Sint64Kind: + return coderSint64PackedSliceValue + case pref.Uint64Kind: + return coderUint64PackedSliceValue + case pref.Sfixed32Kind: + return coderSfixed32PackedSliceValue + case pref.Fixed32Kind: + return coderFixed32PackedSliceValue + case pref.FloatKind: + return coderFloatPackedSliceValue + case pref.Sfixed64Kind: + return coderSfixed64PackedSliceValue + case pref.Fixed64Kind: + return coderFixed64PackedSliceValue + case pref.DoubleKind: + return coderDoublePackedSliceValue + } + default: + switch fd.Kind() { + default: + case pref.BoolKind: + return coderBoolValue + case pref.EnumKind: + return coderEnumValue + case pref.Int32Kind: + return coderInt32Value + case pref.Sint32Kind: + return coderSint32Value + case pref.Uint32Kind: + return coderUint32Value + case pref.Int64Kind: + return coderInt64Value + case pref.Sint64Kind: + return coderSint64Value + case pref.Uint64Kind: + return coderUint64Value + case pref.Sfixed32Kind: + return coderSfixed32Value + case pref.Fixed32Kind: + return coderFixed32Value + case pref.FloatKind: + return coderFloatValue + case pref.Sfixed64Kind: + return coderSfixed64Value + case pref.Fixed64Kind: + return coderFixed64Value + case pref.DoubleKind: + return coderDoubleValue + case pref.StringKind: + if strs.EnforceUTF8(fd) { + return coderStringValueValidateUTF8 + } + return coderStringValue + case pref.BytesKind: + return coderBytesValue + case pref.MessageKind: + return coderMessageValue + case pref.GroupKind: + return coderGroupValue + } + } + panic(fmt.Sprintf("invalid field: no encoder for %v %v %v", fd.FullName(), fd.Cardinality(), fd.Kind())) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/codec_unsafe.go b/vendor/google.golang.org/protobuf/internal/impl/codec_unsafe.go new file mode 100644 index 00000000..e118af1e --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/codec_unsafe.go @@ -0,0 +1,17 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !purego,!appengine + +package impl + +// When using unsafe pointers, we can just treat enum values as int32s. + +var ( + coderEnumNoZero = coderInt32NoZero + coderEnum = coderInt32 + coderEnumPtr = coderInt32Ptr + coderEnumSlice = coderInt32Slice + coderEnumPackedSlice = coderInt32PackedSlice +) diff --git a/vendor/google.golang.org/protobuf/internal/impl/convert.go b/vendor/google.golang.org/protobuf/internal/impl/convert.go new file mode 100644 index 00000000..36a90dff --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/convert.go @@ -0,0 +1,467 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +// unwrapper unwraps the value to the underlying value. +// This is implemented by List and Map. +type unwrapper interface { + protoUnwrap() interface{} +} + +// A Converter coverts to/from Go reflect.Value types and protobuf protoreflect.Value types. +type Converter interface { + // PBValueOf converts a reflect.Value to a protoreflect.Value. + PBValueOf(reflect.Value) pref.Value + + // GoValueOf converts a protoreflect.Value to a reflect.Value. + GoValueOf(pref.Value) reflect.Value + + // IsValidPB returns whether a protoreflect.Value is compatible with this type. + IsValidPB(pref.Value) bool + + // IsValidGo returns whether a reflect.Value is compatible with this type. + IsValidGo(reflect.Value) bool + + // New returns a new field value. + // For scalars, it returns the default value of the field. + // For composite types, it returns a new mutable value. + New() pref.Value + + // Zero returns a new field value. + // For scalars, it returns the default value of the field. + // For composite types, it returns an immutable, empty value. + Zero() pref.Value +} + +// NewConverter matches a Go type with a protobuf field and returns a Converter +// that converts between the two. Enums must be a named int32 kind that +// implements protoreflect.Enum, and messages must be pointer to a named +// struct type that implements protoreflect.ProtoMessage. +// +// This matcher deliberately supports a wider range of Go types than what +// protoc-gen-go historically generated to be able to automatically wrap some +// v1 messages generated by other forks of protoc-gen-go. +func NewConverter(t reflect.Type, fd pref.FieldDescriptor) Converter { + switch { + case fd.IsList(): + return newListConverter(t, fd) + case fd.IsMap(): + return newMapConverter(t, fd) + default: + return newSingularConverter(t, fd) + } + panic(fmt.Sprintf("invalid Go type %v for field %v", t, fd.FullName())) +} + +var ( + boolType = reflect.TypeOf(bool(false)) + int32Type = reflect.TypeOf(int32(0)) + int64Type = reflect.TypeOf(int64(0)) + uint32Type = reflect.TypeOf(uint32(0)) + uint64Type = reflect.TypeOf(uint64(0)) + float32Type = reflect.TypeOf(float32(0)) + float64Type = reflect.TypeOf(float64(0)) + stringType = reflect.TypeOf(string("")) + bytesType = reflect.TypeOf([]byte(nil)) + byteType = reflect.TypeOf(byte(0)) +) + +var ( + boolZero = pref.ValueOfBool(false) + int32Zero = pref.ValueOfInt32(0) + int64Zero = pref.ValueOfInt64(0) + uint32Zero = pref.ValueOfUint32(0) + uint64Zero = pref.ValueOfUint64(0) + float32Zero = pref.ValueOfFloat32(0) + float64Zero = pref.ValueOfFloat64(0) + stringZero = pref.ValueOfString("") + bytesZero = pref.ValueOfBytes(nil) +) + +func newSingularConverter(t reflect.Type, fd pref.FieldDescriptor) Converter { + defVal := func(fd pref.FieldDescriptor, zero pref.Value) pref.Value { + if fd.Cardinality() == pref.Repeated { + // Default isn't defined for repeated fields. + return zero + } + return fd.Default() + } + switch fd.Kind() { + case pref.BoolKind: + if t.Kind() == reflect.Bool { + return &boolConverter{t, defVal(fd, boolZero)} + } + case pref.Int32Kind, pref.Sint32Kind, pref.Sfixed32Kind: + if t.Kind() == reflect.Int32 { + return &int32Converter{t, defVal(fd, int32Zero)} + } + case pref.Int64Kind, pref.Sint64Kind, pref.Sfixed64Kind: + if t.Kind() == reflect.Int64 { + return &int64Converter{t, defVal(fd, int64Zero)} + } + case pref.Uint32Kind, pref.Fixed32Kind: + if t.Kind() == reflect.Uint32 { + return &uint32Converter{t, defVal(fd, uint32Zero)} + } + case pref.Uint64Kind, pref.Fixed64Kind: + if t.Kind() == reflect.Uint64 { + return &uint64Converter{t, defVal(fd, uint64Zero)} + } + case pref.FloatKind: + if t.Kind() == reflect.Float32 { + return &float32Converter{t, defVal(fd, float32Zero)} + } + case pref.DoubleKind: + if t.Kind() == reflect.Float64 { + return &float64Converter{t, defVal(fd, float64Zero)} + } + case pref.StringKind: + if t.Kind() == reflect.String || (t.Kind() == reflect.Slice && t.Elem() == byteType) { + return &stringConverter{t, defVal(fd, stringZero)} + } + case pref.BytesKind: + if t.Kind() == reflect.String || (t.Kind() == reflect.Slice && t.Elem() == byteType) { + return &bytesConverter{t, defVal(fd, bytesZero)} + } + case pref.EnumKind: + // Handle enums, which must be a named int32 type. + if t.Kind() == reflect.Int32 { + return newEnumConverter(t, fd) + } + case pref.MessageKind, pref.GroupKind: + return newMessageConverter(t) + } + panic(fmt.Sprintf("invalid Go type %v for field %v", t, fd.FullName())) +} + +type boolConverter struct { + goType reflect.Type + def pref.Value +} + +func (c *boolConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfBool(v.Bool()) +} +func (c *boolConverter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(v.Bool()).Convert(c.goType) +} +func (c *boolConverter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(bool) + return ok +} +func (c *boolConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *boolConverter) New() pref.Value { return c.def } +func (c *boolConverter) Zero() pref.Value { return c.def } + +type int32Converter struct { + goType reflect.Type + def pref.Value +} + +func (c *int32Converter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfInt32(int32(v.Int())) +} +func (c *int32Converter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(int32(v.Int())).Convert(c.goType) +} +func (c *int32Converter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(int32) + return ok +} +func (c *int32Converter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *int32Converter) New() pref.Value { return c.def } +func (c *int32Converter) Zero() pref.Value { return c.def } + +type int64Converter struct { + goType reflect.Type + def pref.Value +} + +func (c *int64Converter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfInt64(int64(v.Int())) +} +func (c *int64Converter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(int64(v.Int())).Convert(c.goType) +} +func (c *int64Converter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(int64) + return ok +} +func (c *int64Converter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *int64Converter) New() pref.Value { return c.def } +func (c *int64Converter) Zero() pref.Value { return c.def } + +type uint32Converter struct { + goType reflect.Type + def pref.Value +} + +func (c *uint32Converter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfUint32(uint32(v.Uint())) +} +func (c *uint32Converter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(uint32(v.Uint())).Convert(c.goType) +} +func (c *uint32Converter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(uint32) + return ok +} +func (c *uint32Converter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *uint32Converter) New() pref.Value { return c.def } +func (c *uint32Converter) Zero() pref.Value { return c.def } + +type uint64Converter struct { + goType reflect.Type + def pref.Value +} + +func (c *uint64Converter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfUint64(uint64(v.Uint())) +} +func (c *uint64Converter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(uint64(v.Uint())).Convert(c.goType) +} +func (c *uint64Converter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(uint64) + return ok +} +func (c *uint64Converter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *uint64Converter) New() pref.Value { return c.def } +func (c *uint64Converter) Zero() pref.Value { return c.def } + +type float32Converter struct { + goType reflect.Type + def pref.Value +} + +func (c *float32Converter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfFloat32(float32(v.Float())) +} +func (c *float32Converter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(float32(v.Float())).Convert(c.goType) +} +func (c *float32Converter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(float32) + return ok +} +func (c *float32Converter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *float32Converter) New() pref.Value { return c.def } +func (c *float32Converter) Zero() pref.Value { return c.def } + +type float64Converter struct { + goType reflect.Type + def pref.Value +} + +func (c *float64Converter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfFloat64(float64(v.Float())) +} +func (c *float64Converter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(float64(v.Float())).Convert(c.goType) +} +func (c *float64Converter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(float64) + return ok +} +func (c *float64Converter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *float64Converter) New() pref.Value { return c.def } +func (c *float64Converter) Zero() pref.Value { return c.def } + +type stringConverter struct { + goType reflect.Type + def pref.Value +} + +func (c *stringConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfString(v.Convert(stringType).String()) +} +func (c *stringConverter) GoValueOf(v pref.Value) reflect.Value { + // pref.Value.String never panics, so we go through an interface + // conversion here to check the type. + s := v.Interface().(string) + if c.goType.Kind() == reflect.Slice && s == "" { + return reflect.Zero(c.goType) // ensure empty string is []byte(nil) + } + return reflect.ValueOf(s).Convert(c.goType) +} +func (c *stringConverter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(string) + return ok +} +func (c *stringConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *stringConverter) New() pref.Value { return c.def } +func (c *stringConverter) Zero() pref.Value { return c.def } + +type bytesConverter struct { + goType reflect.Type + def pref.Value +} + +func (c *bytesConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + if c.goType.Kind() == reflect.String && v.Len() == 0 { + return pref.ValueOfBytes(nil) // ensure empty string is []byte(nil) + } + return pref.ValueOfBytes(v.Convert(bytesType).Bytes()) +} +func (c *bytesConverter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(v.Bytes()).Convert(c.goType) +} +func (c *bytesConverter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().([]byte) + return ok +} +func (c *bytesConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} +func (c *bytesConverter) New() pref.Value { return c.def } +func (c *bytesConverter) Zero() pref.Value { return c.def } + +type enumConverter struct { + goType reflect.Type + def pref.Value +} + +func newEnumConverter(goType reflect.Type, fd pref.FieldDescriptor) Converter { + var def pref.Value + if fd.Cardinality() == pref.Repeated { + def = pref.ValueOfEnum(fd.Enum().Values().Get(0).Number()) + } else { + def = fd.Default() + } + return &enumConverter{goType, def} +} + +func (c *enumConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfEnum(pref.EnumNumber(v.Int())) +} + +func (c *enumConverter) GoValueOf(v pref.Value) reflect.Value { + return reflect.ValueOf(v.Enum()).Convert(c.goType) +} + +func (c *enumConverter) IsValidPB(v pref.Value) bool { + _, ok := v.Interface().(pref.EnumNumber) + return ok +} + +func (c *enumConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} + +func (c *enumConverter) New() pref.Value { + return c.def +} + +func (c *enumConverter) Zero() pref.Value { + return c.def +} + +type messageConverter struct { + goType reflect.Type +} + +func newMessageConverter(goType reflect.Type) Converter { + return &messageConverter{goType} +} + +func (c *messageConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + if m, ok := v.Interface().(pref.ProtoMessage); ok { + return pref.ValueOfMessage(m.ProtoReflect()) + } + return pref.ValueOfMessage(legacyWrapMessage(v)) +} + +func (c *messageConverter) GoValueOf(v pref.Value) reflect.Value { + m := v.Message() + var rv reflect.Value + if u, ok := m.(unwrapper); ok { + rv = reflect.ValueOf(u.protoUnwrap()) + } else { + rv = reflect.ValueOf(m.Interface()) + } + if rv.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", rv.Type(), c.goType)) + } + return rv +} + +func (c *messageConverter) IsValidPB(v pref.Value) bool { + m := v.Message() + var rv reflect.Value + if u, ok := m.(unwrapper); ok { + rv = reflect.ValueOf(u.protoUnwrap()) + } else { + rv = reflect.ValueOf(m.Interface()) + } + return rv.Type() == c.goType +} + +func (c *messageConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} + +func (c *messageConverter) New() pref.Value { + return c.PBValueOf(reflect.New(c.goType.Elem())) +} + +func (c *messageConverter) Zero() pref.Value { + return c.PBValueOf(reflect.Zero(c.goType)) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/convert_list.go b/vendor/google.golang.org/protobuf/internal/impl/convert_list.go new file mode 100644 index 00000000..6fccab52 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/convert_list.go @@ -0,0 +1,141 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +func newListConverter(t reflect.Type, fd pref.FieldDescriptor) Converter { + switch { + case t.Kind() == reflect.Ptr && t.Elem().Kind() == reflect.Slice: + return &listPtrConverter{t, newSingularConverter(t.Elem().Elem(), fd)} + case t.Kind() == reflect.Slice: + return &listConverter{t, newSingularConverter(t.Elem(), fd)} + } + panic(fmt.Sprintf("invalid Go type %v for field %v", t, fd.FullName())) +} + +type listConverter struct { + goType reflect.Type // []T + c Converter +} + +func (c *listConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + pv := reflect.New(c.goType) + pv.Elem().Set(v) + return pref.ValueOfList(&listReflect{pv, c.c}) +} + +func (c *listConverter) GoValueOf(v pref.Value) reflect.Value { + rv := v.List().(*listReflect).v + if rv.IsNil() { + return reflect.Zero(c.goType) + } + return rv.Elem() +} + +func (c *listConverter) IsValidPB(v pref.Value) bool { + list, ok := v.Interface().(*listReflect) + if !ok { + return false + } + return list.v.Type().Elem() == c.goType +} + +func (c *listConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} + +func (c *listConverter) New() pref.Value { + return pref.ValueOfList(&listReflect{reflect.New(c.goType), c.c}) +} + +func (c *listConverter) Zero() pref.Value { + return pref.ValueOfList(&listReflect{reflect.Zero(reflect.PtrTo(c.goType)), c.c}) +} + +type listPtrConverter struct { + goType reflect.Type // *[]T + c Converter +} + +func (c *listPtrConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfList(&listReflect{v, c.c}) +} + +func (c *listPtrConverter) GoValueOf(v pref.Value) reflect.Value { + return v.List().(*listReflect).v +} + +func (c *listPtrConverter) IsValidPB(v pref.Value) bool { + list, ok := v.Interface().(*listReflect) + if !ok { + return false + } + return list.v.Type() == c.goType +} + +func (c *listPtrConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} + +func (c *listPtrConverter) New() pref.Value { + return c.PBValueOf(reflect.New(c.goType.Elem())) +} + +func (c *listPtrConverter) Zero() pref.Value { + return c.PBValueOf(reflect.Zero(c.goType)) +} + +type listReflect struct { + v reflect.Value // *[]T + conv Converter +} + +func (ls *listReflect) Len() int { + if ls.v.IsNil() { + return 0 + } + return ls.v.Elem().Len() +} +func (ls *listReflect) Get(i int) pref.Value { + return ls.conv.PBValueOf(ls.v.Elem().Index(i)) +} +func (ls *listReflect) Set(i int, v pref.Value) { + ls.v.Elem().Index(i).Set(ls.conv.GoValueOf(v)) +} +func (ls *listReflect) Append(v pref.Value) { + ls.v.Elem().Set(reflect.Append(ls.v.Elem(), ls.conv.GoValueOf(v))) +} +func (ls *listReflect) AppendMutable() pref.Value { + if _, ok := ls.conv.(*messageConverter); !ok { + panic("invalid AppendMutable on list with non-message type") + } + v := ls.NewElement() + ls.Append(v) + return v +} +func (ls *listReflect) Truncate(i int) { + ls.v.Elem().Set(ls.v.Elem().Slice(0, i)) +} +func (ls *listReflect) NewElement() pref.Value { + return ls.conv.New() +} +func (ls *listReflect) IsValid() bool { + return !ls.v.IsNil() +} +func (ls *listReflect) protoUnwrap() interface{} { + return ls.v.Interface() +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/convert_map.go b/vendor/google.golang.org/protobuf/internal/impl/convert_map.go new file mode 100644 index 00000000..de06b259 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/convert_map.go @@ -0,0 +1,121 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type mapConverter struct { + goType reflect.Type // map[K]V + keyConv, valConv Converter +} + +func newMapConverter(t reflect.Type, fd pref.FieldDescriptor) *mapConverter { + if t.Kind() != reflect.Map { + panic(fmt.Sprintf("invalid Go type %v for field %v", t, fd.FullName())) + } + return &mapConverter{ + goType: t, + keyConv: newSingularConverter(t.Key(), fd.MapKey()), + valConv: newSingularConverter(t.Elem(), fd.MapValue()), + } +} + +func (c *mapConverter) PBValueOf(v reflect.Value) pref.Value { + if v.Type() != c.goType { + panic(fmt.Sprintf("invalid type: got %v, want %v", v.Type(), c.goType)) + } + return pref.ValueOfMap(&mapReflect{v, c.keyConv, c.valConv}) +} + +func (c *mapConverter) GoValueOf(v pref.Value) reflect.Value { + return v.Map().(*mapReflect).v +} + +func (c *mapConverter) IsValidPB(v pref.Value) bool { + mapv, ok := v.Interface().(*mapReflect) + if !ok { + return false + } + return mapv.v.Type() == c.goType +} + +func (c *mapConverter) IsValidGo(v reflect.Value) bool { + return v.IsValid() && v.Type() == c.goType +} + +func (c *mapConverter) New() pref.Value { + return c.PBValueOf(reflect.MakeMap(c.goType)) +} + +func (c *mapConverter) Zero() pref.Value { + return c.PBValueOf(reflect.Zero(c.goType)) +} + +type mapReflect struct { + v reflect.Value // map[K]V + keyConv Converter + valConv Converter +} + +func (ms *mapReflect) Len() int { + return ms.v.Len() +} +func (ms *mapReflect) Has(k pref.MapKey) bool { + rk := ms.keyConv.GoValueOf(k.Value()) + rv := ms.v.MapIndex(rk) + return rv.IsValid() +} +func (ms *mapReflect) Get(k pref.MapKey) pref.Value { + rk := ms.keyConv.GoValueOf(k.Value()) + rv := ms.v.MapIndex(rk) + if !rv.IsValid() { + return pref.Value{} + } + return ms.valConv.PBValueOf(rv) +} +func (ms *mapReflect) Set(k pref.MapKey, v pref.Value) { + rk := ms.keyConv.GoValueOf(k.Value()) + rv := ms.valConv.GoValueOf(v) + ms.v.SetMapIndex(rk, rv) +} +func (ms *mapReflect) Clear(k pref.MapKey) { + rk := ms.keyConv.GoValueOf(k.Value()) + ms.v.SetMapIndex(rk, reflect.Value{}) +} +func (ms *mapReflect) Mutable(k pref.MapKey) pref.Value { + if _, ok := ms.valConv.(*messageConverter); !ok { + panic("invalid Mutable on map with non-message value type") + } + v := ms.Get(k) + if !v.IsValid() { + v = ms.NewValue() + ms.Set(k, v) + } + return v +} +func (ms *mapReflect) Range(f func(pref.MapKey, pref.Value) bool) { + iter := mapRange(ms.v) + for iter.Next() { + k := ms.keyConv.PBValueOf(iter.Key()).MapKey() + v := ms.valConv.PBValueOf(iter.Value()) + if !f(k, v) { + return + } + } +} +func (ms *mapReflect) NewValue() pref.Value { + return ms.valConv.New() +} +func (ms *mapReflect) IsValid() bool { + return !ms.v.IsNil() +} +func (ms *mapReflect) protoUnwrap() interface{} { + return ms.v.Interface() +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/decode.go b/vendor/google.golang.org/protobuf/internal/impl/decode.go new file mode 100644 index 00000000..85ba1d3b --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/decode.go @@ -0,0 +1,274 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "math/bits" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/flags" + "google.golang.org/protobuf/proto" + "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" + "google.golang.org/protobuf/runtime/protoiface" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +type unmarshalOptions struct { + flags protoiface.UnmarshalInputFlags + resolver interface { + FindExtensionByName(field protoreflect.FullName) (protoreflect.ExtensionType, error) + FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) + } +} + +func (o unmarshalOptions) Options() proto.UnmarshalOptions { + return proto.UnmarshalOptions{ + Merge: true, + AllowPartial: true, + DiscardUnknown: o.DiscardUnknown(), + Resolver: o.resolver, + } +} + +func (o unmarshalOptions) DiscardUnknown() bool { return o.flags&piface.UnmarshalDiscardUnknown != 0 } + +func (o unmarshalOptions) IsDefault() bool { + return o.flags == 0 && o.resolver == preg.GlobalTypes +} + +var lazyUnmarshalOptions = unmarshalOptions{ + resolver: preg.GlobalTypes, +} + +type unmarshalOutput struct { + n int // number of bytes consumed + initialized bool +} + +// unmarshal is protoreflect.Methods.Unmarshal. +func (mi *MessageInfo) unmarshal(in piface.UnmarshalInput) (piface.UnmarshalOutput, error) { + var p pointer + if ms, ok := in.Message.(*messageState); ok { + p = ms.pointer() + } else { + p = in.Message.(*messageReflectWrapper).pointer() + } + out, err := mi.unmarshalPointer(in.Buf, p, 0, unmarshalOptions{ + flags: in.Flags, + resolver: in.Resolver, + }) + var flags piface.UnmarshalOutputFlags + if out.initialized { + flags |= piface.UnmarshalInitialized + } + return piface.UnmarshalOutput{ + Flags: flags, + }, err +} + +// errUnknown is returned during unmarshaling to indicate a parse error that +// should result in a field being placed in the unknown fields section (for example, +// when the wire type doesn't match) as opposed to the entire unmarshal operation +// failing (for example, when a field extends past the available input). +// +// This is a sentinel error which should never be visible to the user. +var errUnknown = errors.New("unknown") + +func (mi *MessageInfo) unmarshalPointer(b []byte, p pointer, groupTag protowire.Number, opts unmarshalOptions) (out unmarshalOutput, err error) { + mi.init() + if flags.ProtoLegacy && mi.isMessageSet { + return unmarshalMessageSet(mi, b, p, opts) + } + initialized := true + var requiredMask uint64 + var exts *map[int32]ExtensionField + start := len(b) + for len(b) > 0 { + // Parse the tag (field number and wire type). + var tag uint64 + if b[0] < 0x80 { + tag = uint64(b[0]) + b = b[1:] + } else if len(b) >= 2 && b[1] < 128 { + tag = uint64(b[0]&0x7f) + uint64(b[1])<<7 + b = b[2:] + } else { + var n int + tag, n = protowire.ConsumeVarint(b) + if n < 0 { + return out, protowire.ParseError(n) + } + b = b[n:] + } + var num protowire.Number + if n := tag >> 3; n < uint64(protowire.MinValidNumber) || n > uint64(protowire.MaxValidNumber) { + return out, errors.New("invalid field number") + } else { + num = protowire.Number(n) + } + wtyp := protowire.Type(tag & 7) + + if wtyp == protowire.EndGroupType { + if num != groupTag { + return out, errors.New("mismatching end group marker") + } + groupTag = 0 + break + } + + var f *coderFieldInfo + if int(num) < len(mi.denseCoderFields) { + f = mi.denseCoderFields[num] + } else { + f = mi.coderFields[num] + } + var n int + err := errUnknown + switch { + case f != nil: + if f.funcs.unmarshal == nil { + break + } + var o unmarshalOutput + o, err = f.funcs.unmarshal(b, p.Apply(f.offset), wtyp, f, opts) + n = o.n + if err != nil { + break + } + requiredMask |= f.validation.requiredBit + if f.funcs.isInit != nil && !o.initialized { + initialized = false + } + default: + // Possible extension. + if exts == nil && mi.extensionOffset.IsValid() { + exts = p.Apply(mi.extensionOffset).Extensions() + if *exts == nil { + *exts = make(map[int32]ExtensionField) + } + } + if exts == nil { + break + } + var o unmarshalOutput + o, err = mi.unmarshalExtension(b, num, wtyp, *exts, opts) + if err != nil { + break + } + n = o.n + if !o.initialized { + initialized = false + } + } + if err != nil { + if err != errUnknown { + return out, err + } + n = protowire.ConsumeFieldValue(num, wtyp, b) + if n < 0 { + return out, protowire.ParseError(n) + } + if !opts.DiscardUnknown() && mi.unknownOffset.IsValid() { + u := p.Apply(mi.unknownOffset).Bytes() + *u = protowire.AppendTag(*u, num, wtyp) + *u = append(*u, b[:n]...) + } + } + b = b[n:] + } + if groupTag != 0 { + return out, errors.New("missing end group marker") + } + if mi.numRequiredFields > 0 && bits.OnesCount64(requiredMask) != int(mi.numRequiredFields) { + initialized = false + } + if initialized { + out.initialized = true + } + out.n = start - len(b) + return out, nil +} + +func (mi *MessageInfo) unmarshalExtension(b []byte, num protowire.Number, wtyp protowire.Type, exts map[int32]ExtensionField, opts unmarshalOptions) (out unmarshalOutput, err error) { + x := exts[int32(num)] + xt := x.Type() + if xt == nil { + var err error + xt, err = opts.resolver.FindExtensionByNumber(mi.Desc.FullName(), num) + if err != nil { + if err == preg.NotFound { + return out, errUnknown + } + return out, errors.New("%v: unable to resolve extension %v: %v", mi.Desc.FullName(), num, err) + } + } + xi := getExtensionFieldInfo(xt) + if xi.funcs.unmarshal == nil { + return out, errUnknown + } + if flags.LazyUnmarshalExtensions { + if opts.IsDefault() && x.canLazy(xt) { + out, valid := skipExtension(b, xi, num, wtyp, opts) + switch valid { + case ValidationValid: + if out.initialized { + x.appendLazyBytes(xt, xi, num, wtyp, b[:out.n]) + exts[int32(num)] = x + return out, nil + } + case ValidationInvalid: + return out, errors.New("invalid wire format") + case ValidationUnknown: + } + } + } + ival := x.Value() + if !ival.IsValid() && xi.unmarshalNeedsValue { + // Create a new message, list, or map value to fill in. + // For enums, create a prototype value to let the unmarshal func know the + // concrete type. + ival = xt.New() + } + v, out, err := xi.funcs.unmarshal(b, ival, num, wtyp, opts) + if err != nil { + return out, err + } + if xi.funcs.isInit == nil { + out.initialized = true + } + x.Set(xt, v) + exts[int32(num)] = x + return out, nil +} + +func skipExtension(b []byte, xi *extensionFieldInfo, num protowire.Number, wtyp protowire.Type, opts unmarshalOptions) (out unmarshalOutput, _ ValidationStatus) { + if xi.validation.mi == nil { + return out, ValidationUnknown + } + xi.validation.mi.init() + switch xi.validation.typ { + case validationTypeMessage: + if wtyp != protowire.BytesType { + return out, ValidationUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return out, ValidationUnknown + } + out, st := xi.validation.mi.validate(v, 0, opts) + out.n = n + return out, st + case validationTypeGroup: + if wtyp != protowire.StartGroupType { + return out, ValidationUnknown + } + out, st := xi.validation.mi.validate(b, num, opts) + return out, st + default: + return out, ValidationUnknown + } +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/encode.go b/vendor/google.golang.org/protobuf/internal/impl/encode.go new file mode 100644 index 00000000..8c8a794c --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/encode.go @@ -0,0 +1,199 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "math" + "sort" + "sync/atomic" + + "google.golang.org/protobuf/internal/flags" + proto "google.golang.org/protobuf/proto" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +type marshalOptions struct { + flags piface.MarshalInputFlags +} + +func (o marshalOptions) Options() proto.MarshalOptions { + return proto.MarshalOptions{ + AllowPartial: true, + Deterministic: o.Deterministic(), + UseCachedSize: o.UseCachedSize(), + } +} + +func (o marshalOptions) Deterministic() bool { return o.flags&piface.MarshalDeterministic != 0 } +func (o marshalOptions) UseCachedSize() bool { return o.flags&piface.MarshalUseCachedSize != 0 } + +// size is protoreflect.Methods.Size. +func (mi *MessageInfo) size(in piface.SizeInput) piface.SizeOutput { + var p pointer + if ms, ok := in.Message.(*messageState); ok { + p = ms.pointer() + } else { + p = in.Message.(*messageReflectWrapper).pointer() + } + size := mi.sizePointer(p, marshalOptions{ + flags: in.Flags, + }) + return piface.SizeOutput{Size: size} +} + +func (mi *MessageInfo) sizePointer(p pointer, opts marshalOptions) (size int) { + mi.init() + if p.IsNil() { + return 0 + } + if opts.UseCachedSize() && mi.sizecacheOffset.IsValid() { + if size := atomic.LoadInt32(p.Apply(mi.sizecacheOffset).Int32()); size >= 0 { + return int(size) + } + } + return mi.sizePointerSlow(p, opts) +} + +func (mi *MessageInfo) sizePointerSlow(p pointer, opts marshalOptions) (size int) { + if flags.ProtoLegacy && mi.isMessageSet { + size = sizeMessageSet(mi, p, opts) + if mi.sizecacheOffset.IsValid() { + atomic.StoreInt32(p.Apply(mi.sizecacheOffset).Int32(), int32(size)) + } + return size + } + if mi.extensionOffset.IsValid() { + e := p.Apply(mi.extensionOffset).Extensions() + size += mi.sizeExtensions(e, opts) + } + for _, f := range mi.orderedCoderFields { + if f.funcs.size == nil { + continue + } + fptr := p.Apply(f.offset) + if f.isPointer && fptr.Elem().IsNil() { + continue + } + size += f.funcs.size(fptr, f, opts) + } + if mi.unknownOffset.IsValid() { + u := *p.Apply(mi.unknownOffset).Bytes() + size += len(u) + } + if mi.sizecacheOffset.IsValid() { + if size > math.MaxInt32 { + // The size is too large for the int32 sizecache field. + // We will need to recompute the size when encoding; + // unfortunately expensive, but better than invalid output. + atomic.StoreInt32(p.Apply(mi.sizecacheOffset).Int32(), -1) + } else { + atomic.StoreInt32(p.Apply(mi.sizecacheOffset).Int32(), int32(size)) + } + } + return size +} + +// marshal is protoreflect.Methods.Marshal. +func (mi *MessageInfo) marshal(in piface.MarshalInput) (out piface.MarshalOutput, err error) { + var p pointer + if ms, ok := in.Message.(*messageState); ok { + p = ms.pointer() + } else { + p = in.Message.(*messageReflectWrapper).pointer() + } + b, err := mi.marshalAppendPointer(in.Buf, p, marshalOptions{ + flags: in.Flags, + }) + return piface.MarshalOutput{Buf: b}, err +} + +func (mi *MessageInfo) marshalAppendPointer(b []byte, p pointer, opts marshalOptions) ([]byte, error) { + mi.init() + if p.IsNil() { + return b, nil + } + if flags.ProtoLegacy && mi.isMessageSet { + return marshalMessageSet(mi, b, p, opts) + } + var err error + // The old marshaler encodes extensions at beginning. + if mi.extensionOffset.IsValid() { + e := p.Apply(mi.extensionOffset).Extensions() + // TODO: Special handling for MessageSet? + b, err = mi.appendExtensions(b, e, opts) + if err != nil { + return b, err + } + } + for _, f := range mi.orderedCoderFields { + if f.funcs.marshal == nil { + continue + } + fptr := p.Apply(f.offset) + if f.isPointer && fptr.Elem().IsNil() { + continue + } + b, err = f.funcs.marshal(b, fptr, f, opts) + if err != nil { + return b, err + } + } + if mi.unknownOffset.IsValid() && !mi.isMessageSet { + u := *p.Apply(mi.unknownOffset).Bytes() + b = append(b, u...) + } + return b, nil +} + +func (mi *MessageInfo) sizeExtensions(ext *map[int32]ExtensionField, opts marshalOptions) (n int) { + if ext == nil { + return 0 + } + for _, x := range *ext { + xi := getExtensionFieldInfo(x.Type()) + if xi.funcs.size == nil { + continue + } + n += xi.funcs.size(x.Value(), xi.tagsize, opts) + } + return n +} + +func (mi *MessageInfo) appendExtensions(b []byte, ext *map[int32]ExtensionField, opts marshalOptions) ([]byte, error) { + if ext == nil { + return b, nil + } + + switch len(*ext) { + case 0: + return b, nil + case 1: + // Fast-path for one extension: Don't bother sorting the keys. + var err error + for _, x := range *ext { + xi := getExtensionFieldInfo(x.Type()) + b, err = xi.funcs.marshal(b, x.Value(), xi.wiretag, opts) + } + return b, err + default: + // Sort the keys to provide a deterministic encoding. + // Not sure this is required, but the old code does it. + keys := make([]int, 0, len(*ext)) + for k := range *ext { + keys = append(keys, int(k)) + } + sort.Ints(keys) + var err error + for _, k := range keys { + x := (*ext)[int32(k)] + xi := getExtensionFieldInfo(x.Type()) + b, err = xi.funcs.marshal(b, x.Value(), xi.wiretag, opts) + if err != nil { + return b, err + } + } + return b, nil + } +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/enum.go b/vendor/google.golang.org/protobuf/internal/impl/enum.go new file mode 100644 index 00000000..8c1eab4b --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/enum.go @@ -0,0 +1,21 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "reflect" + + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type EnumInfo struct { + GoReflectType reflect.Type // int32 kind + Desc pref.EnumDescriptor +} + +func (t *EnumInfo) New(n pref.EnumNumber) pref.Enum { + return reflect.ValueOf(n).Convert(t.GoReflectType).Interface().(pref.Enum) +} +func (t *EnumInfo) Descriptor() pref.EnumDescriptor { return t.Desc } diff --git a/vendor/google.golang.org/protobuf/internal/impl/extension.go b/vendor/google.golang.org/protobuf/internal/impl/extension.go new file mode 100644 index 00000000..e904fd99 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/extension.go @@ -0,0 +1,156 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "reflect" + "sync" + "sync/atomic" + + pref "google.golang.org/protobuf/reflect/protoreflect" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +// ExtensionInfo implements ExtensionType. +// +// This type contains a number of exported fields for legacy compatibility. +// The only non-deprecated use of this type is through the methods of the +// ExtensionType interface. +type ExtensionInfo struct { + // An ExtensionInfo may exist in several stages of initialization. + // + // extensionInfoUninitialized: Some or all of the legacy exported + // fields may be set, but none of the unexported fields have been + // initialized. This is the starting state for an ExtensionInfo + // in legacy generated code. + // + // extensionInfoDescInit: The desc field is set, but other unexported fields + // may not be initialized. Legacy exported fields may or may not be set. + // This is the starting state for an ExtensionInfo in newly generated code. + // + // extensionInfoFullInit: The ExtensionInfo is fully initialized. + // This state is only entered after lazy initialization is complete. + init uint32 + mu sync.Mutex + + goType reflect.Type + desc extensionTypeDescriptor + conv Converter + info *extensionFieldInfo // for fast-path method implementations + + // ExtendedType is a typed nil-pointer to the parent message type that + // is being extended. It is possible for this to be unpopulated in v2 + // since the message may no longer implement the MessageV1 interface. + // + // Deprecated: Use the ExtendedType method instead. + ExtendedType piface.MessageV1 + + // ExtensionType is the zero value of the extension type. + // + // For historical reasons, reflect.TypeOf(ExtensionType) and the + // type returned by InterfaceOf may not be identical. + // + // Deprecated: Use InterfaceOf(xt.Zero()) instead. + ExtensionType interface{} + + // Field is the field number of the extension. + // + // Deprecated: Use the Descriptor().Number method instead. + Field int32 + + // Name is the fully qualified name of extension. + // + // Deprecated: Use the Descriptor().FullName method instead. + Name string + + // Tag is the protobuf struct tag used in the v1 API. + // + // Deprecated: Do not use. + Tag string + + // Filename is the proto filename in which the extension is defined. + // + // Deprecated: Use Descriptor().ParentFile().Path() instead. + Filename string +} + +// Stages of initialization: See the ExtensionInfo.init field. +const ( + extensionInfoUninitialized = 0 + extensionInfoDescInit = 1 + extensionInfoFullInit = 2 +) + +func InitExtensionInfo(xi *ExtensionInfo, xd pref.ExtensionDescriptor, goType reflect.Type) { + xi.goType = goType + xi.desc = extensionTypeDescriptor{xd, xi} + xi.init = extensionInfoDescInit +} + +func (xi *ExtensionInfo) New() pref.Value { + return xi.lazyInit().New() +} +func (xi *ExtensionInfo) Zero() pref.Value { + return xi.lazyInit().Zero() +} +func (xi *ExtensionInfo) ValueOf(v interface{}) pref.Value { + return xi.lazyInit().PBValueOf(reflect.ValueOf(v)) +} +func (xi *ExtensionInfo) InterfaceOf(v pref.Value) interface{} { + return xi.lazyInit().GoValueOf(v).Interface() +} +func (xi *ExtensionInfo) IsValidValue(v pref.Value) bool { + return xi.lazyInit().IsValidPB(v) +} +func (xi *ExtensionInfo) IsValidInterface(v interface{}) bool { + return xi.lazyInit().IsValidGo(reflect.ValueOf(v)) +} +func (xi *ExtensionInfo) TypeDescriptor() pref.ExtensionTypeDescriptor { + if atomic.LoadUint32(&xi.init) < extensionInfoDescInit { + xi.lazyInitSlow() + } + return &xi.desc +} + +func (xi *ExtensionInfo) lazyInit() Converter { + if atomic.LoadUint32(&xi.init) < extensionInfoFullInit { + xi.lazyInitSlow() + } + return xi.conv +} + +func (xi *ExtensionInfo) lazyInitSlow() { + xi.mu.Lock() + defer xi.mu.Unlock() + + if xi.init == extensionInfoFullInit { + return + } + defer atomic.StoreUint32(&xi.init, extensionInfoFullInit) + + if xi.desc.ExtensionDescriptor == nil { + xi.initFromLegacy() + } + if !xi.desc.ExtensionDescriptor.IsPlaceholder() { + if xi.ExtensionType == nil { + xi.initToLegacy() + } + xi.conv = NewConverter(xi.goType, xi.desc.ExtensionDescriptor) + xi.info = makeExtensionFieldInfo(xi.desc.ExtensionDescriptor) + xi.info.validation = newValidationInfo(xi.desc.ExtensionDescriptor, xi.goType) + } +} + +type extensionTypeDescriptor struct { + pref.ExtensionDescriptor + xi *ExtensionInfo +} + +func (xtd *extensionTypeDescriptor) Type() pref.ExtensionType { + return xtd.xi +} +func (xtd *extensionTypeDescriptor) Descriptor() pref.ExtensionDescriptor { + return xtd.ExtensionDescriptor +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/legacy_enum.go b/vendor/google.golang.org/protobuf/internal/impl/legacy_enum.go new file mode 100644 index 00000000..f7d7ffb5 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/legacy_enum.go @@ -0,0 +1,219 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + "strings" + "sync" + + "google.golang.org/protobuf/internal/filedesc" + "google.golang.org/protobuf/internal/strs" + "google.golang.org/protobuf/reflect/protoreflect" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +// legacyEnumName returns the name of enums used in legacy code. +// It is neither the protobuf full name nor the qualified Go name, +// but rather an odd hybrid of both. +func legacyEnumName(ed pref.EnumDescriptor) string { + var protoPkg string + enumName := string(ed.FullName()) + if fd := ed.ParentFile(); fd != nil { + protoPkg = string(fd.Package()) + enumName = strings.TrimPrefix(enumName, protoPkg+".") + } + if protoPkg == "" { + return strs.GoCamelCase(enumName) + } + return protoPkg + "." + strs.GoCamelCase(enumName) +} + +// legacyWrapEnum wraps v as a protoreflect.Enum, +// where v must be a int32 kind and not implement the v2 API already. +func legacyWrapEnum(v reflect.Value) pref.Enum { + et := legacyLoadEnumType(v.Type()) + return et.New(pref.EnumNumber(v.Int())) +} + +var legacyEnumTypeCache sync.Map // map[reflect.Type]protoreflect.EnumType + +// legacyLoadEnumType dynamically loads a protoreflect.EnumType for t, +// where t must be an int32 kind and not implement the v2 API already. +func legacyLoadEnumType(t reflect.Type) pref.EnumType { + // Fast-path: check if a EnumType is cached for this concrete type. + if et, ok := legacyEnumTypeCache.Load(t); ok { + return et.(pref.EnumType) + } + + // Slow-path: derive enum descriptor and initialize EnumType. + var et pref.EnumType + ed := LegacyLoadEnumDesc(t) + et = &legacyEnumType{ + desc: ed, + goType: t, + } + if et, ok := legacyEnumTypeCache.LoadOrStore(t, et); ok { + return et.(pref.EnumType) + } + return et +} + +type legacyEnumType struct { + desc pref.EnumDescriptor + goType reflect.Type + m sync.Map // map[protoreflect.EnumNumber]proto.Enum +} + +func (t *legacyEnumType) New(n pref.EnumNumber) pref.Enum { + if e, ok := t.m.Load(n); ok { + return e.(pref.Enum) + } + e := &legacyEnumWrapper{num: n, pbTyp: t, goTyp: t.goType} + t.m.Store(n, e) + return e +} +func (t *legacyEnumType) Descriptor() pref.EnumDescriptor { + return t.desc +} + +type legacyEnumWrapper struct { + num pref.EnumNumber + pbTyp pref.EnumType + goTyp reflect.Type +} + +func (e *legacyEnumWrapper) Descriptor() pref.EnumDescriptor { + return e.pbTyp.Descriptor() +} +func (e *legacyEnumWrapper) Type() pref.EnumType { + return e.pbTyp +} +func (e *legacyEnumWrapper) Number() pref.EnumNumber { + return e.num +} +func (e *legacyEnumWrapper) ProtoReflect() pref.Enum { + return e +} +func (e *legacyEnumWrapper) protoUnwrap() interface{} { + v := reflect.New(e.goTyp).Elem() + v.SetInt(int64(e.num)) + return v.Interface() +} + +var ( + _ pref.Enum = (*legacyEnumWrapper)(nil) + _ unwrapper = (*legacyEnumWrapper)(nil) +) + +var legacyEnumDescCache sync.Map // map[reflect.Type]protoreflect.EnumDescriptor + +// LegacyLoadEnumDesc returns an EnumDescriptor derived from the Go type, +// which must be an int32 kind and not implement the v2 API already. +// +// This is exported for testing purposes. +func LegacyLoadEnumDesc(t reflect.Type) pref.EnumDescriptor { + // Fast-path: check if an EnumDescriptor is cached for this concrete type. + if ed, ok := legacyEnumDescCache.Load(t); ok { + return ed.(pref.EnumDescriptor) + } + + // Slow-path: initialize EnumDescriptor from the raw descriptor. + ev := reflect.Zero(t).Interface() + if _, ok := ev.(pref.Enum); ok { + panic(fmt.Sprintf("%v already implements proto.Enum", t)) + } + edV1, ok := ev.(enumV1) + if !ok { + return aberrantLoadEnumDesc(t) + } + b, idxs := edV1.EnumDescriptor() + + var ed pref.EnumDescriptor + if len(idxs) == 1 { + ed = legacyLoadFileDesc(b).Enums().Get(idxs[0]) + } else { + md := legacyLoadFileDesc(b).Messages().Get(idxs[0]) + for _, i := range idxs[1 : len(idxs)-1] { + md = md.Messages().Get(i) + } + ed = md.Enums().Get(idxs[len(idxs)-1]) + } + if ed, ok := legacyEnumDescCache.LoadOrStore(t, ed); ok { + return ed.(protoreflect.EnumDescriptor) + } + return ed +} + +var aberrantEnumDescCache sync.Map // map[reflect.Type]protoreflect.EnumDescriptor + +// aberrantLoadEnumDesc returns an EnumDescriptor derived from the Go type, +// which must not implement protoreflect.Enum or enumV1. +// +// If the type does not implement enumV1, then there is no reliable +// way to derive the original protobuf type information. +// We are unable to use the global enum registry since it is +// unfortunately keyed by the protobuf full name, which we also do not know. +// Thus, this produces some bogus enum descriptor based on the Go type name. +func aberrantLoadEnumDesc(t reflect.Type) pref.EnumDescriptor { + // Fast-path: check if an EnumDescriptor is cached for this concrete type. + if ed, ok := aberrantEnumDescCache.Load(t); ok { + return ed.(pref.EnumDescriptor) + } + + // Slow-path: construct a bogus, but unique EnumDescriptor. + ed := &filedesc.Enum{L2: new(filedesc.EnumL2)} + ed.L0.FullName = AberrantDeriveFullName(t) // e.g., github_com.user.repo.MyEnum + ed.L0.ParentFile = filedesc.SurrogateProto3 + ed.L2.Values.List = append(ed.L2.Values.List, filedesc.EnumValue{}) + + // TODO: Use the presence of a UnmarshalJSON method to determine proto2? + + vd := &ed.L2.Values.List[0] + vd.L0.FullName = ed.L0.FullName + "_UNKNOWN" // e.g., github_com.user.repo.MyEnum_UNKNOWN + vd.L0.ParentFile = ed.L0.ParentFile + vd.L0.Parent = ed + + // TODO: We could use the String method to obtain some enum value names by + // starting at 0 and print the enum until it produces invalid identifiers. + // An exhaustive query is clearly impractical, but can be best-effort. + + if ed, ok := aberrantEnumDescCache.LoadOrStore(t, ed); ok { + return ed.(pref.EnumDescriptor) + } + return ed +} + +// AberrantDeriveFullName derives a fully qualified protobuf name for the given Go type +// The provided name is not guaranteed to be stable nor universally unique. +// It should be sufficiently unique within a program. +// +// This is exported for testing purposes. +func AberrantDeriveFullName(t reflect.Type) pref.FullName { + sanitize := func(r rune) rune { + switch { + case r == '/': + return '.' + case 'a' <= r && r <= 'z', 'A' <= r && r <= 'Z', '0' <= r && r <= '9': + return r + default: + return '_' + } + } + prefix := strings.Map(sanitize, t.PkgPath()) + suffix := strings.Map(sanitize, t.Name()) + if suffix == "" { + suffix = fmt.Sprintf("UnknownX%X", reflect.ValueOf(t).Pointer()) + } + + ss := append(strings.Split(prefix, "."), suffix) + for i, s := range ss { + if s == "" || ('0' <= s[0] && s[0] <= '9') { + ss[i] = "x" + s + } + } + return pref.FullName(strings.Join(ss, ".")) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/legacy_export.go b/vendor/google.golang.org/protobuf/internal/impl/legacy_export.go new file mode 100644 index 00000000..c3d741c2 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/legacy_export.go @@ -0,0 +1,92 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "encoding/binary" + "encoding/json" + "hash/crc32" + "math" + "reflect" + + "google.golang.org/protobuf/internal/errors" + pref "google.golang.org/protobuf/reflect/protoreflect" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +// These functions exist to support exported APIs in generated protobufs. +// While these are deprecated, they cannot be removed for compatibility reasons. + +// LegacyEnumName returns the name of enums used in legacy code. +func (Export) LegacyEnumName(ed pref.EnumDescriptor) string { + return legacyEnumName(ed) +} + +// LegacyMessageTypeOf returns the protoreflect.MessageType for m, +// with name used as the message name if necessary. +func (Export) LegacyMessageTypeOf(m piface.MessageV1, name pref.FullName) pref.MessageType { + if mv := (Export{}).protoMessageV2Of(m); mv != nil { + return mv.ProtoReflect().Type() + } + return legacyLoadMessageInfo(reflect.TypeOf(m), name) +} + +// UnmarshalJSONEnum unmarshals an enum from a JSON-encoded input. +// The input can either be a string representing the enum value by name, +// or a number representing the enum number itself. +func (Export) UnmarshalJSONEnum(ed pref.EnumDescriptor, b []byte) (pref.EnumNumber, error) { + if b[0] == '"' { + var name pref.Name + if err := json.Unmarshal(b, &name); err != nil { + return 0, errors.New("invalid input for enum %v: %s", ed.FullName(), b) + } + ev := ed.Values().ByName(name) + if ev == nil { + return 0, errors.New("invalid value for enum %v: %s", ed.FullName(), name) + } + return ev.Number(), nil + } else { + var num pref.EnumNumber + if err := json.Unmarshal(b, &num); err != nil { + return 0, errors.New("invalid input for enum %v: %s", ed.FullName(), b) + } + return num, nil + } +} + +// CompressGZIP compresses the input as a GZIP-encoded file. +// The current implementation does no compression. +func (Export) CompressGZIP(in []byte) (out []byte) { + // RFC 1952, section 2.3.1. + var gzipHeader = [10]byte{0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xff} + + // RFC 1951, section 3.2.4. + var blockHeader [5]byte + const maxBlockSize = math.MaxUint16 + numBlocks := 1 + len(in)/maxBlockSize + + // RFC 1952, section 2.3.1. + var gzipFooter [8]byte + binary.LittleEndian.PutUint32(gzipFooter[0:4], crc32.ChecksumIEEE(in)) + binary.LittleEndian.PutUint32(gzipFooter[4:8], uint32(len(in))) + + // Encode the input without compression using raw DEFLATE blocks. + out = make([]byte, 0, len(gzipHeader)+len(blockHeader)*numBlocks+len(in)+len(gzipFooter)) + out = append(out, gzipHeader[:]...) + for blockHeader[0] == 0 { + blockSize := maxBlockSize + if blockSize > len(in) { + blockHeader[0] = 0x01 // final bit per RFC 1951, section 3.2.3. + blockSize = len(in) + } + binary.LittleEndian.PutUint16(blockHeader[1:3], uint16(blockSize)^0x0000) + binary.LittleEndian.PutUint16(blockHeader[3:5], uint16(blockSize)^0xffff) + out = append(out, blockHeader[:]...) + out = append(out, in[:blockSize]...) + in = in[blockSize:] + } + out = append(out, gzipFooter[:]...) + return out +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/legacy_extension.go b/vendor/google.golang.org/protobuf/internal/impl/legacy_extension.go new file mode 100644 index 00000000..61757ce5 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/legacy_extension.go @@ -0,0 +1,175 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "reflect" + + "google.golang.org/protobuf/internal/descopts" + "google.golang.org/protobuf/internal/encoding/messageset" + ptag "google.golang.org/protobuf/internal/encoding/tag" + "google.golang.org/protobuf/internal/filedesc" + "google.golang.org/protobuf/internal/pragma" + pref "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +func (xi *ExtensionInfo) initToLegacy() { + xd := xi.desc + var parent piface.MessageV1 + messageName := xd.ContainingMessage().FullName() + if mt, _ := preg.GlobalTypes.FindMessageByName(messageName); mt != nil { + // Create a new parent message and unwrap it if possible. + mv := mt.New().Interface() + t := reflect.TypeOf(mv) + if mv, ok := mv.(unwrapper); ok { + t = reflect.TypeOf(mv.protoUnwrap()) + } + + // Check whether the message implements the legacy v1 Message interface. + mz := reflect.Zero(t).Interface() + if mz, ok := mz.(piface.MessageV1); ok { + parent = mz + } + } + + // Determine the v1 extension type, which is unfortunately not the same as + // the v2 ExtensionType.GoType. + extType := xi.goType + switch extType.Kind() { + case reflect.Bool, reflect.Int32, reflect.Int64, reflect.Uint32, reflect.Uint64, reflect.Float32, reflect.Float64, reflect.String: + extType = reflect.PtrTo(extType) // T -> *T for singular scalar fields + } + + // Reconstruct the legacy enum full name. + var enumName string + if xd.Kind() == pref.EnumKind { + enumName = legacyEnumName(xd.Enum()) + } + + // Derive the proto file that the extension was declared within. + var filename string + if fd := xd.ParentFile(); fd != nil { + filename = fd.Path() + } + + // For MessageSet extensions, the name used is the parent message. + name := xd.FullName() + if messageset.IsMessageSetExtension(xd) { + name = name.Parent() + } + + xi.ExtendedType = parent + xi.ExtensionType = reflect.Zero(extType).Interface() + xi.Field = int32(xd.Number()) + xi.Name = string(name) + xi.Tag = ptag.Marshal(xd, enumName) + xi.Filename = filename +} + +// initFromLegacy initializes an ExtensionInfo from +// the contents of the deprecated exported fields of the type. +func (xi *ExtensionInfo) initFromLegacy() { + // The v1 API returns "type incomplete" descriptors where only the + // field number is specified. In such a case, use a placeholder. + if xi.ExtendedType == nil || xi.ExtensionType == nil { + xd := placeholderExtension{ + name: pref.FullName(xi.Name), + number: pref.FieldNumber(xi.Field), + } + xi.desc = extensionTypeDescriptor{xd, xi} + return + } + + // Resolve enum or message dependencies. + var ed pref.EnumDescriptor + var md pref.MessageDescriptor + t := reflect.TypeOf(xi.ExtensionType) + isOptional := t.Kind() == reflect.Ptr && t.Elem().Kind() != reflect.Struct + isRepeated := t.Kind() == reflect.Slice && t.Elem().Kind() != reflect.Uint8 + if isOptional || isRepeated { + t = t.Elem() + } + switch v := reflect.Zero(t).Interface().(type) { + case pref.Enum: + ed = v.Descriptor() + case enumV1: + ed = LegacyLoadEnumDesc(t) + case pref.ProtoMessage: + md = v.ProtoReflect().Descriptor() + case messageV1: + md = LegacyLoadMessageDesc(t) + } + + // Derive basic field information from the struct tag. + var evs pref.EnumValueDescriptors + if ed != nil { + evs = ed.Values() + } + fd := ptag.Unmarshal(xi.Tag, t, evs).(*filedesc.Field) + + // Construct a v2 ExtensionType. + xd := &filedesc.Extension{L2: new(filedesc.ExtensionL2)} + xd.L0.ParentFile = filedesc.SurrogateProto2 + xd.L0.FullName = pref.FullName(xi.Name) + xd.L1.Number = pref.FieldNumber(xi.Field) + xd.L1.Cardinality = fd.L1.Cardinality + xd.L1.Kind = fd.L1.Kind + xd.L2.IsPacked = fd.L1.IsPacked + xd.L2.Default = fd.L1.Default + xd.L1.Extendee = Export{}.MessageDescriptorOf(xi.ExtendedType) + xd.L2.Enum = ed + xd.L2.Message = md + + // Derive real extension field name for MessageSets. + if messageset.IsMessageSet(xd.L1.Extendee) && md.FullName() == xd.L0.FullName { + xd.L0.FullName = xd.L0.FullName.Append(messageset.ExtensionName) + } + + tt := reflect.TypeOf(xi.ExtensionType) + if isOptional { + tt = tt.Elem() + } + xi.goType = tt + xi.desc = extensionTypeDescriptor{xd, xi} +} + +type placeholderExtension struct { + name pref.FullName + number pref.FieldNumber +} + +func (x placeholderExtension) ParentFile() pref.FileDescriptor { return nil } +func (x placeholderExtension) Parent() pref.Descriptor { return nil } +func (x placeholderExtension) Index() int { return 0 } +func (x placeholderExtension) Syntax() pref.Syntax { return 0 } +func (x placeholderExtension) Name() pref.Name { return x.name.Name() } +func (x placeholderExtension) FullName() pref.FullName { return x.name } +func (x placeholderExtension) IsPlaceholder() bool { return true } +func (x placeholderExtension) Options() pref.ProtoMessage { return descopts.Field } +func (x placeholderExtension) Number() pref.FieldNumber { return x.number } +func (x placeholderExtension) Cardinality() pref.Cardinality { return 0 } +func (x placeholderExtension) Kind() pref.Kind { return 0 } +func (x placeholderExtension) HasJSONName() bool { return false } +func (x placeholderExtension) JSONName() string { return "" } +func (x placeholderExtension) HasPresence() bool { return false } +func (x placeholderExtension) HasOptionalKeyword() bool { return false } +func (x placeholderExtension) IsExtension() bool { return true } +func (x placeholderExtension) IsWeak() bool { return false } +func (x placeholderExtension) IsPacked() bool { return false } +func (x placeholderExtension) IsList() bool { return false } +func (x placeholderExtension) IsMap() bool { return false } +func (x placeholderExtension) MapKey() pref.FieldDescriptor { return nil } +func (x placeholderExtension) MapValue() pref.FieldDescriptor { return nil } +func (x placeholderExtension) HasDefault() bool { return false } +func (x placeholderExtension) Default() pref.Value { return pref.Value{} } +func (x placeholderExtension) DefaultEnumValue() pref.EnumValueDescriptor { return nil } +func (x placeholderExtension) ContainingOneof() pref.OneofDescriptor { return nil } +func (x placeholderExtension) ContainingMessage() pref.MessageDescriptor { return nil } +func (x placeholderExtension) Enum() pref.EnumDescriptor { return nil } +func (x placeholderExtension) Message() pref.MessageDescriptor { return nil } +func (x placeholderExtension) ProtoType(pref.FieldDescriptor) { return } +func (x placeholderExtension) ProtoInternal(pragma.DoNotImplement) { return } diff --git a/vendor/google.golang.org/protobuf/internal/impl/legacy_file.go b/vendor/google.golang.org/protobuf/internal/impl/legacy_file.go new file mode 100644 index 00000000..9ab09108 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/legacy_file.go @@ -0,0 +1,81 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "bytes" + "compress/gzip" + "io/ioutil" + "sync" + + "google.golang.org/protobuf/internal/filedesc" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +// Every enum and message type generated by protoc-gen-go since commit 2fc053c5 +// on February 25th, 2016 has had a method to get the raw descriptor. +// Types that were not generated by protoc-gen-go or were generated prior +// to that version are not supported. +// +// The []byte returned is the encoded form of a FileDescriptorProto message +// compressed using GZIP. The []int is the path from the top-level file +// to the specific message or enum declaration. +type ( + enumV1 interface { + EnumDescriptor() ([]byte, []int) + } + messageV1 interface { + Descriptor() ([]byte, []int) + } +) + +var legacyFileDescCache sync.Map // map[*byte]protoreflect.FileDescriptor + +// legacyLoadFileDesc unmarshals b as a compressed FileDescriptorProto message. +// +// This assumes that b is immutable and that b does not refer to part of a +// concatenated series of GZIP files (which would require shenanigans that +// rely on the concatenation properties of both protobufs and GZIP). +// File descriptors generated by protoc-gen-go do not rely on that property. +func legacyLoadFileDesc(b []byte) protoreflect.FileDescriptor { + // Fast-path: check whether we already have a cached file descriptor. + if fd, ok := legacyFileDescCache.Load(&b[0]); ok { + return fd.(protoreflect.FileDescriptor) + } + + // Slow-path: decompress and unmarshal the file descriptor proto. + zr, err := gzip.NewReader(bytes.NewReader(b)) + if err != nil { + panic(err) + } + b2, err := ioutil.ReadAll(zr) + if err != nil { + panic(err) + } + + fd := filedesc.Builder{ + RawDescriptor: b2, + FileRegistry: resolverOnly{protoregistry.GlobalFiles}, // do not register back to global registry + }.Build().File + if fd, ok := legacyFileDescCache.LoadOrStore(&b[0], fd); ok { + return fd.(protoreflect.FileDescriptor) + } + return fd +} + +type resolverOnly struct { + reg *protoregistry.Files +} + +func (r resolverOnly) FindFileByPath(path string) (protoreflect.FileDescriptor, error) { + return r.reg.FindFileByPath(path) +} +func (r resolverOnly) FindDescriptorByName(name protoreflect.FullName) (protoreflect.Descriptor, error) { + return r.reg.FindDescriptorByName(name) +} +func (resolverOnly) RegisterFile(protoreflect.FileDescriptor) error { + return nil +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/legacy_message.go b/vendor/google.golang.org/protobuf/internal/impl/legacy_message.go new file mode 100644 index 00000000..06c68e11 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/legacy_message.go @@ -0,0 +1,502 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + "strings" + "sync" + + "google.golang.org/protobuf/internal/descopts" + ptag "google.golang.org/protobuf/internal/encoding/tag" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/filedesc" + "google.golang.org/protobuf/internal/strs" + "google.golang.org/protobuf/reflect/protoreflect" + pref "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +// legacyWrapMessage wraps v as a protoreflect.Message, +// where v must be a *struct kind and not implement the v2 API already. +func legacyWrapMessage(v reflect.Value) pref.Message { + typ := v.Type() + if typ.Kind() != reflect.Ptr || typ.Elem().Kind() != reflect.Struct { + return aberrantMessage{v: v} + } + mt := legacyLoadMessageInfo(typ, "") + return mt.MessageOf(v.Interface()) +} + +var legacyMessageTypeCache sync.Map // map[reflect.Type]*MessageInfo + +// legacyLoadMessageInfo dynamically loads a *MessageInfo for t, +// where t must be a *struct kind and not implement the v2 API already. +// The provided name is used if it cannot be determined from the message. +func legacyLoadMessageInfo(t reflect.Type, name pref.FullName) *MessageInfo { + // Fast-path: check if a MessageInfo is cached for this concrete type. + if mt, ok := legacyMessageTypeCache.Load(t); ok { + return mt.(*MessageInfo) + } + + // Slow-path: derive message descriptor and initialize MessageInfo. + mi := &MessageInfo{ + Desc: legacyLoadMessageDesc(t, name), + GoReflectType: t, + } + + v := reflect.Zero(t).Interface() + if _, ok := v.(legacyMarshaler); ok { + mi.methods.Marshal = legacyMarshal + + // We have no way to tell whether the type's Marshal method + // supports deterministic serialization or not, but this + // preserves the v1 implementation's behavior of always + // calling Marshal methods when present. + mi.methods.Flags |= piface.SupportMarshalDeterministic + } + if _, ok := v.(legacyUnmarshaler); ok { + mi.methods.Unmarshal = legacyUnmarshal + } + if _, ok := v.(legacyMerger); ok { + mi.methods.Merge = legacyMerge + } + + if mi, ok := legacyMessageTypeCache.LoadOrStore(t, mi); ok { + return mi.(*MessageInfo) + } + return mi +} + +var legacyMessageDescCache sync.Map // map[reflect.Type]protoreflect.MessageDescriptor + +// LegacyLoadMessageDesc returns an MessageDescriptor derived from the Go type, +// which must be a *struct kind and not implement the v2 API already. +// +// This is exported for testing purposes. +func LegacyLoadMessageDesc(t reflect.Type) pref.MessageDescriptor { + return legacyLoadMessageDesc(t, "") +} +func legacyLoadMessageDesc(t reflect.Type, name pref.FullName) pref.MessageDescriptor { + // Fast-path: check if a MessageDescriptor is cached for this concrete type. + if mi, ok := legacyMessageDescCache.Load(t); ok { + return mi.(pref.MessageDescriptor) + } + + // Slow-path: initialize MessageDescriptor from the raw descriptor. + mv := reflect.Zero(t).Interface() + if _, ok := mv.(pref.ProtoMessage); ok { + panic(fmt.Sprintf("%v already implements proto.Message", t)) + } + mdV1, ok := mv.(messageV1) + if !ok { + return aberrantLoadMessageDesc(t, name) + } + + // If this is a dynamic message type where there isn't a 1-1 mapping between + // Go and protobuf types, calling the Descriptor method on the zero value of + // the message type isn't likely to work. If it panics, swallow the panic and + // continue as if the Descriptor method wasn't present. + b, idxs := func() ([]byte, []int) { + defer func() { + recover() + }() + return mdV1.Descriptor() + }() + if b == nil { + return aberrantLoadMessageDesc(t, name) + } + + // If the Go type has no fields, then this might be a proto3 empty message + // from before the size cache was added. If there are any fields, check to + // see that at least one of them looks like something we generated. + if nfield := t.Elem().NumField(); nfield > 0 { + hasProtoField := false + for i := 0; i < nfield; i++ { + f := t.Elem().Field(i) + if f.Tag.Get("protobuf") != "" || f.Tag.Get("protobuf_oneof") != "" || strings.HasPrefix(f.Name, "XXX_") { + hasProtoField = true + break + } + } + if !hasProtoField { + return aberrantLoadMessageDesc(t, name) + } + } + + md := legacyLoadFileDesc(b).Messages().Get(idxs[0]) + for _, i := range idxs[1:] { + md = md.Messages().Get(i) + } + if name != "" && md.FullName() != name { + panic(fmt.Sprintf("mismatching message name: got %v, want %v", md.FullName(), name)) + } + if md, ok := legacyMessageDescCache.LoadOrStore(t, md); ok { + return md.(protoreflect.MessageDescriptor) + } + return md +} + +var ( + aberrantMessageDescLock sync.Mutex + aberrantMessageDescCache map[reflect.Type]protoreflect.MessageDescriptor +) + +// aberrantLoadMessageDesc returns an MessageDescriptor derived from the Go type, +// which must not implement protoreflect.ProtoMessage or messageV1. +// +// This is a best-effort derivation of the message descriptor using the protobuf +// tags on the struct fields. +func aberrantLoadMessageDesc(t reflect.Type, name pref.FullName) pref.MessageDescriptor { + aberrantMessageDescLock.Lock() + defer aberrantMessageDescLock.Unlock() + if aberrantMessageDescCache == nil { + aberrantMessageDescCache = make(map[reflect.Type]protoreflect.MessageDescriptor) + } + return aberrantLoadMessageDescReentrant(t, name) +} +func aberrantLoadMessageDescReentrant(t reflect.Type, name pref.FullName) pref.MessageDescriptor { + // Fast-path: check if an MessageDescriptor is cached for this concrete type. + if md, ok := aberrantMessageDescCache[t]; ok { + return md + } + + // Slow-path: construct a descriptor from the Go struct type (best-effort). + // Cache the MessageDescriptor early on so that we can resolve internal + // cyclic references. + md := &filedesc.Message{L2: new(filedesc.MessageL2)} + md.L0.FullName = aberrantDeriveMessageName(t, name) + md.L0.ParentFile = filedesc.SurrogateProto2 + aberrantMessageDescCache[t] = md + + if t.Kind() != reflect.Ptr || t.Elem().Kind() != reflect.Struct { + return md + } + + // Try to determine if the message is using proto3 by checking scalars. + for i := 0; i < t.Elem().NumField(); i++ { + f := t.Elem().Field(i) + if tag := f.Tag.Get("protobuf"); tag != "" { + switch f.Type.Kind() { + case reflect.Bool, reflect.Int32, reflect.Int64, reflect.Uint32, reflect.Uint64, reflect.Float32, reflect.Float64, reflect.String: + md.L0.ParentFile = filedesc.SurrogateProto3 + } + for _, s := range strings.Split(tag, ",") { + if s == "proto3" { + md.L0.ParentFile = filedesc.SurrogateProto3 + } + } + } + } + + // Obtain a list of oneof wrapper types. + var oneofWrappers []reflect.Type + for _, method := range []string{"XXX_OneofFuncs", "XXX_OneofWrappers"} { + if fn, ok := t.MethodByName(method); ok { + for _, v := range fn.Func.Call([]reflect.Value{reflect.Zero(fn.Type.In(0))}) { + if vs, ok := v.Interface().([]interface{}); ok { + for _, v := range vs { + oneofWrappers = append(oneofWrappers, reflect.TypeOf(v)) + } + } + } + } + } + + // Obtain a list of the extension ranges. + if fn, ok := t.MethodByName("ExtensionRangeArray"); ok { + vs := fn.Func.Call([]reflect.Value{reflect.Zero(fn.Type.In(0))})[0] + for i := 0; i < vs.Len(); i++ { + v := vs.Index(i) + md.L2.ExtensionRanges.List = append(md.L2.ExtensionRanges.List, [2]pref.FieldNumber{ + pref.FieldNumber(v.FieldByName("Start").Int()), + pref.FieldNumber(v.FieldByName("End").Int() + 1), + }) + md.L2.ExtensionRangeOptions = append(md.L2.ExtensionRangeOptions, nil) + } + } + + // Derive the message fields by inspecting the struct fields. + for i := 0; i < t.Elem().NumField(); i++ { + f := t.Elem().Field(i) + if tag := f.Tag.Get("protobuf"); tag != "" { + tagKey := f.Tag.Get("protobuf_key") + tagVal := f.Tag.Get("protobuf_val") + aberrantAppendField(md, f.Type, tag, tagKey, tagVal) + } + if tag := f.Tag.Get("protobuf_oneof"); tag != "" { + n := len(md.L2.Oneofs.List) + md.L2.Oneofs.List = append(md.L2.Oneofs.List, filedesc.Oneof{}) + od := &md.L2.Oneofs.List[n] + od.L0.FullName = md.FullName().Append(pref.Name(tag)) + od.L0.ParentFile = md.L0.ParentFile + od.L0.Parent = md + od.L0.Index = n + + for _, t := range oneofWrappers { + if t.Implements(f.Type) { + f := t.Elem().Field(0) + if tag := f.Tag.Get("protobuf"); tag != "" { + aberrantAppendField(md, f.Type, tag, "", "") + fd := &md.L2.Fields.List[len(md.L2.Fields.List)-1] + fd.L1.ContainingOneof = od + od.L1.Fields.List = append(od.L1.Fields.List, fd) + } + } + } + } + } + + return md +} + +func aberrantDeriveMessageName(t reflect.Type, name pref.FullName) pref.FullName { + if name.IsValid() { + return name + } + func() { + defer func() { recover() }() // swallow possible nil panics + if m, ok := reflect.Zero(t).Interface().(interface{ XXX_MessageName() string }); ok { + name = pref.FullName(m.XXX_MessageName()) + } + }() + if name.IsValid() { + return name + } + if t.Kind() == reflect.Ptr { + t = t.Elem() + } + return AberrantDeriveFullName(t) +} + +func aberrantAppendField(md *filedesc.Message, goType reflect.Type, tag, tagKey, tagVal string) { + t := goType + isOptional := t.Kind() == reflect.Ptr && t.Elem().Kind() != reflect.Struct + isRepeated := t.Kind() == reflect.Slice && t.Elem().Kind() != reflect.Uint8 + if isOptional || isRepeated { + t = t.Elem() + } + fd := ptag.Unmarshal(tag, t, placeholderEnumValues{}).(*filedesc.Field) + + // Append field descriptor to the message. + n := len(md.L2.Fields.List) + md.L2.Fields.List = append(md.L2.Fields.List, *fd) + fd = &md.L2.Fields.List[n] + fd.L0.FullName = md.FullName().Append(fd.Name()) + fd.L0.ParentFile = md.L0.ParentFile + fd.L0.Parent = md + fd.L0.Index = n + + if fd.L1.IsWeak || fd.L1.HasPacked { + fd.L1.Options = func() pref.ProtoMessage { + opts := descopts.Field.ProtoReflect().New() + if fd.L1.IsWeak { + opts.Set(opts.Descriptor().Fields().ByName("weak"), protoreflect.ValueOfBool(true)) + } + if fd.L1.HasPacked { + opts.Set(opts.Descriptor().Fields().ByName("packed"), protoreflect.ValueOfBool(fd.L1.IsPacked)) + } + return opts.Interface() + } + } + + // Populate Enum and Message. + if fd.Enum() == nil && fd.Kind() == pref.EnumKind { + switch v := reflect.Zero(t).Interface().(type) { + case pref.Enum: + fd.L1.Enum = v.Descriptor() + default: + fd.L1.Enum = LegacyLoadEnumDesc(t) + } + } + if fd.Message() == nil && (fd.Kind() == pref.MessageKind || fd.Kind() == pref.GroupKind) { + switch v := reflect.Zero(t).Interface().(type) { + case pref.ProtoMessage: + fd.L1.Message = v.ProtoReflect().Descriptor() + case messageV1: + fd.L1.Message = LegacyLoadMessageDesc(t) + default: + if t.Kind() == reflect.Map { + n := len(md.L1.Messages.List) + md.L1.Messages.List = append(md.L1.Messages.List, filedesc.Message{L2: new(filedesc.MessageL2)}) + md2 := &md.L1.Messages.List[n] + md2.L0.FullName = md.FullName().Append(pref.Name(strs.MapEntryName(string(fd.Name())))) + md2.L0.ParentFile = md.L0.ParentFile + md2.L0.Parent = md + md2.L0.Index = n + + md2.L1.IsMapEntry = true + md2.L2.Options = func() pref.ProtoMessage { + opts := descopts.Message.ProtoReflect().New() + opts.Set(opts.Descriptor().Fields().ByName("map_entry"), protoreflect.ValueOfBool(true)) + return opts.Interface() + } + + aberrantAppendField(md2, t.Key(), tagKey, "", "") + aberrantAppendField(md2, t.Elem(), tagVal, "", "") + + fd.L1.Message = md2 + break + } + fd.L1.Message = aberrantLoadMessageDescReentrant(t, "") + } + } +} + +type placeholderEnumValues struct { + protoreflect.EnumValueDescriptors +} + +func (placeholderEnumValues) ByNumber(n pref.EnumNumber) pref.EnumValueDescriptor { + return filedesc.PlaceholderEnumValue(pref.FullName(fmt.Sprintf("UNKNOWN_%d", n))) +} + +// legacyMarshaler is the proto.Marshaler interface superseded by protoiface.Methoder. +type legacyMarshaler interface { + Marshal() ([]byte, error) +} + +// legacyUnmarshaler is the proto.Unmarshaler interface superseded by protoiface.Methoder. +type legacyUnmarshaler interface { + Unmarshal([]byte) error +} + +// legacyMerger is the proto.Merger interface superseded by protoiface.Methoder. +type legacyMerger interface { + Merge(protoiface.MessageV1) +} + +var legacyProtoMethods = &piface.Methods{ + Marshal: legacyMarshal, + Unmarshal: legacyUnmarshal, + Merge: legacyMerge, + + // We have no way to tell whether the type's Marshal method + // supports deterministic serialization or not, but this + // preserves the v1 implementation's behavior of always + // calling Marshal methods when present. + Flags: piface.SupportMarshalDeterministic, +} + +func legacyMarshal(in piface.MarshalInput) (piface.MarshalOutput, error) { + v := in.Message.(unwrapper).protoUnwrap() + marshaler, ok := v.(legacyMarshaler) + if !ok { + return piface.MarshalOutput{}, errors.New("%T does not implement Marshal", v) + } + out, err := marshaler.Marshal() + if in.Buf != nil { + out = append(in.Buf, out...) + } + return piface.MarshalOutput{ + Buf: out, + }, err +} + +func legacyUnmarshal(in piface.UnmarshalInput) (piface.UnmarshalOutput, error) { + v := in.Message.(unwrapper).protoUnwrap() + unmarshaler, ok := v.(legacyUnmarshaler) + if !ok { + return piface.UnmarshalOutput{}, errors.New("%T does not implement Marshal", v) + } + return piface.UnmarshalOutput{}, unmarshaler.Unmarshal(in.Buf) +} + +func legacyMerge(in piface.MergeInput) piface.MergeOutput { + dstv := in.Destination.(unwrapper).protoUnwrap() + merger, ok := dstv.(legacyMerger) + if !ok { + return piface.MergeOutput{} + } + merger.Merge(Export{}.ProtoMessageV1Of(in.Source)) + return piface.MergeOutput{Flags: piface.MergeComplete} +} + +// aberrantMessageType implements MessageType for all types other than pointer-to-struct. +type aberrantMessageType struct { + t reflect.Type +} + +func (mt aberrantMessageType) New() pref.Message { + return aberrantMessage{reflect.Zero(mt.t)} +} +func (mt aberrantMessageType) Zero() pref.Message { + return aberrantMessage{reflect.Zero(mt.t)} +} +func (mt aberrantMessageType) GoType() reflect.Type { + return mt.t +} +func (mt aberrantMessageType) Descriptor() pref.MessageDescriptor { + return LegacyLoadMessageDesc(mt.t) +} + +// aberrantMessage implements Message for all types other than pointer-to-struct. +// +// When the underlying type implements legacyMarshaler or legacyUnmarshaler, +// the aberrant Message can be marshaled or unmarshaled. Otherwise, there is +// not much that can be done with values of this type. +type aberrantMessage struct { + v reflect.Value +} + +func (m aberrantMessage) ProtoReflect() pref.Message { + return m +} + +func (m aberrantMessage) Descriptor() pref.MessageDescriptor { + return LegacyLoadMessageDesc(m.v.Type()) +} +func (m aberrantMessage) Type() pref.MessageType { + return aberrantMessageType{m.v.Type()} +} +func (m aberrantMessage) New() pref.Message { + return aberrantMessage{reflect.Zero(m.v.Type())} +} +func (m aberrantMessage) Interface() pref.ProtoMessage { + return m +} +func (m aberrantMessage) Range(f func(pref.FieldDescriptor, pref.Value) bool) { +} +func (m aberrantMessage) Has(pref.FieldDescriptor) bool { + panic("invalid field descriptor") +} +func (m aberrantMessage) Clear(pref.FieldDescriptor) { + panic("invalid field descriptor") +} +func (m aberrantMessage) Get(pref.FieldDescriptor) pref.Value { + panic("invalid field descriptor") +} +func (m aberrantMessage) Set(pref.FieldDescriptor, pref.Value) { + panic("invalid field descriptor") +} +func (m aberrantMessage) Mutable(pref.FieldDescriptor) pref.Value { + panic("invalid field descriptor") +} +func (m aberrantMessage) NewField(pref.FieldDescriptor) pref.Value { + panic("invalid field descriptor") +} +func (m aberrantMessage) WhichOneof(pref.OneofDescriptor) pref.FieldDescriptor { + panic("invalid oneof descriptor") +} +func (m aberrantMessage) GetUnknown() pref.RawFields { + return nil +} +func (m aberrantMessage) SetUnknown(pref.RawFields) { + // SetUnknown discards its input on messages which don't support unknown field storage. +} +func (m aberrantMessage) IsValid() bool { + // An invalid message is a read-only, empty message. Since we don't know anything + // about the alleged contents of this message, we can't say with confidence that + // it is invalid in this sense. Therefore, report it as valid. + return true +} +func (m aberrantMessage) ProtoMethods() *piface.Methods { + return legacyProtoMethods +} +func (m aberrantMessage) protoUnwrap() interface{} { + return m.v.Interface() +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/merge.go b/vendor/google.golang.org/protobuf/internal/impl/merge.go new file mode 100644 index 00000000..cdc4267d --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/merge.go @@ -0,0 +1,176 @@ +// Copyright 2020 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + + "google.golang.org/protobuf/proto" + pref "google.golang.org/protobuf/reflect/protoreflect" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +type mergeOptions struct{} + +func (o mergeOptions) Merge(dst, src proto.Message) { + proto.Merge(dst, src) +} + +// merge is protoreflect.Methods.Merge. +func (mi *MessageInfo) merge(in piface.MergeInput) piface.MergeOutput { + dp, ok := mi.getPointer(in.Destination) + if !ok { + return piface.MergeOutput{} + } + sp, ok := mi.getPointer(in.Source) + if !ok { + return piface.MergeOutput{} + } + mi.mergePointer(dp, sp, mergeOptions{}) + return piface.MergeOutput{Flags: piface.MergeComplete} +} + +func (mi *MessageInfo) mergePointer(dst, src pointer, opts mergeOptions) { + mi.init() + if dst.IsNil() { + panic(fmt.Sprintf("invalid value: merging into nil message")) + } + if src.IsNil() { + return + } + for _, f := range mi.orderedCoderFields { + if f.funcs.merge == nil { + continue + } + sfptr := src.Apply(f.offset) + if f.isPointer && sfptr.Elem().IsNil() { + continue + } + f.funcs.merge(dst.Apply(f.offset), sfptr, f, opts) + } + if mi.extensionOffset.IsValid() { + sext := src.Apply(mi.extensionOffset).Extensions() + dext := dst.Apply(mi.extensionOffset).Extensions() + if *dext == nil { + *dext = make(map[int32]ExtensionField) + } + for num, sx := range *sext { + xt := sx.Type() + xi := getExtensionFieldInfo(xt) + if xi.funcs.merge == nil { + continue + } + dx := (*dext)[num] + var dv pref.Value + if dx.Type() == sx.Type() { + dv = dx.Value() + } + if !dv.IsValid() && xi.unmarshalNeedsValue { + dv = xt.New() + } + dv = xi.funcs.merge(dv, sx.Value(), opts) + dx.Set(sx.Type(), dv) + (*dext)[num] = dx + } + } + if mi.unknownOffset.IsValid() { + du := dst.Apply(mi.unknownOffset).Bytes() + su := src.Apply(mi.unknownOffset).Bytes() + if len(*su) > 0 { + *du = append(*du, *su...) + } + } +} + +func mergeScalarValue(dst, src pref.Value, opts mergeOptions) pref.Value { + return src +} + +func mergeBytesValue(dst, src pref.Value, opts mergeOptions) pref.Value { + return pref.ValueOfBytes(append(emptyBuf[:], src.Bytes()...)) +} + +func mergeListValue(dst, src pref.Value, opts mergeOptions) pref.Value { + dstl := dst.List() + srcl := src.List() + for i, llen := 0, srcl.Len(); i < llen; i++ { + dstl.Append(srcl.Get(i)) + } + return dst +} + +func mergeBytesListValue(dst, src pref.Value, opts mergeOptions) pref.Value { + dstl := dst.List() + srcl := src.List() + for i, llen := 0, srcl.Len(); i < llen; i++ { + sb := srcl.Get(i).Bytes() + db := append(emptyBuf[:], sb...) + dstl.Append(pref.ValueOfBytes(db)) + } + return dst +} + +func mergeMessageListValue(dst, src pref.Value, opts mergeOptions) pref.Value { + dstl := dst.List() + srcl := src.List() + for i, llen := 0, srcl.Len(); i < llen; i++ { + sm := srcl.Get(i).Message() + dm := proto.Clone(sm.Interface()).ProtoReflect() + dstl.Append(pref.ValueOfMessage(dm)) + } + return dst +} + +func mergeMessageValue(dst, src pref.Value, opts mergeOptions) pref.Value { + opts.Merge(dst.Message().Interface(), src.Message().Interface()) + return dst +} + +func mergeMessage(dst, src pointer, f *coderFieldInfo, opts mergeOptions) { + if f.mi != nil { + if dst.Elem().IsNil() { + dst.SetPointer(pointerOfValue(reflect.New(f.mi.GoReflectType.Elem()))) + } + f.mi.mergePointer(dst.Elem(), src.Elem(), opts) + } else { + dm := dst.AsValueOf(f.ft).Elem() + sm := src.AsValueOf(f.ft).Elem() + if dm.IsNil() { + dm.Set(reflect.New(f.ft.Elem())) + } + opts.Merge(asMessage(dm), asMessage(sm)) + } +} + +func mergeMessageSlice(dst, src pointer, f *coderFieldInfo, opts mergeOptions) { + for _, sp := range src.PointerSlice() { + dm := reflect.New(f.ft.Elem().Elem()) + if f.mi != nil { + f.mi.mergePointer(pointerOfValue(dm), sp, opts) + } else { + opts.Merge(asMessage(dm), asMessage(sp.AsValueOf(f.ft.Elem().Elem()))) + } + dst.AppendPointerSlice(pointerOfValue(dm)) + } +} + +func mergeBytes(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Bytes() = append(emptyBuf[:], *src.Bytes()...) +} + +func mergeBytesNoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Bytes() + if len(v) > 0 { + *dst.Bytes() = append(emptyBuf[:], v...) + } +} + +func mergeBytesSlice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.BytesSlice() + for _, v := range *src.BytesSlice() { + *ds = append(*ds, append(emptyBuf[:], v...)) + } +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/merge_gen.go b/vendor/google.golang.org/protobuf/internal/impl/merge_gen.go new file mode 100644 index 00000000..8816c274 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/merge_gen.go @@ -0,0 +1,209 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-types. DO NOT EDIT. + +package impl + +import () + +func mergeBool(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Bool() = *src.Bool() +} + +func mergeBoolNoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Bool() + if v != false { + *dst.Bool() = v + } +} + +func mergeBoolPtr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.BoolPtr() + if p != nil { + v := *p + *dst.BoolPtr() = &v + } +} + +func mergeBoolSlice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.BoolSlice() + ss := src.BoolSlice() + *ds = append(*ds, *ss...) +} + +func mergeInt32(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Int32() = *src.Int32() +} + +func mergeInt32NoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Int32() + if v != 0 { + *dst.Int32() = v + } +} + +func mergeInt32Ptr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.Int32Ptr() + if p != nil { + v := *p + *dst.Int32Ptr() = &v + } +} + +func mergeInt32Slice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.Int32Slice() + ss := src.Int32Slice() + *ds = append(*ds, *ss...) +} + +func mergeUint32(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Uint32() = *src.Uint32() +} + +func mergeUint32NoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Uint32() + if v != 0 { + *dst.Uint32() = v + } +} + +func mergeUint32Ptr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.Uint32Ptr() + if p != nil { + v := *p + *dst.Uint32Ptr() = &v + } +} + +func mergeUint32Slice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.Uint32Slice() + ss := src.Uint32Slice() + *ds = append(*ds, *ss...) +} + +func mergeInt64(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Int64() = *src.Int64() +} + +func mergeInt64NoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Int64() + if v != 0 { + *dst.Int64() = v + } +} + +func mergeInt64Ptr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.Int64Ptr() + if p != nil { + v := *p + *dst.Int64Ptr() = &v + } +} + +func mergeInt64Slice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.Int64Slice() + ss := src.Int64Slice() + *ds = append(*ds, *ss...) +} + +func mergeUint64(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Uint64() = *src.Uint64() +} + +func mergeUint64NoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Uint64() + if v != 0 { + *dst.Uint64() = v + } +} + +func mergeUint64Ptr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.Uint64Ptr() + if p != nil { + v := *p + *dst.Uint64Ptr() = &v + } +} + +func mergeUint64Slice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.Uint64Slice() + ss := src.Uint64Slice() + *ds = append(*ds, *ss...) +} + +func mergeFloat32(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Float32() = *src.Float32() +} + +func mergeFloat32NoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Float32() + if v != 0 { + *dst.Float32() = v + } +} + +func mergeFloat32Ptr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.Float32Ptr() + if p != nil { + v := *p + *dst.Float32Ptr() = &v + } +} + +func mergeFloat32Slice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.Float32Slice() + ss := src.Float32Slice() + *ds = append(*ds, *ss...) +} + +func mergeFloat64(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.Float64() = *src.Float64() +} + +func mergeFloat64NoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.Float64() + if v != 0 { + *dst.Float64() = v + } +} + +func mergeFloat64Ptr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.Float64Ptr() + if p != nil { + v := *p + *dst.Float64Ptr() = &v + } +} + +func mergeFloat64Slice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.Float64Slice() + ss := src.Float64Slice() + *ds = append(*ds, *ss...) +} + +func mergeString(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + *dst.String() = *src.String() +} + +func mergeStringNoZero(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + v := *src.String() + if v != "" { + *dst.String() = v + } +} + +func mergeStringPtr(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + p := *src.StringPtr() + if p != nil { + v := *p + *dst.StringPtr() = &v + } +} + +func mergeStringSlice(dst, src pointer, _ *coderFieldInfo, _ mergeOptions) { + ds := dst.StringSlice() + ss := src.StringSlice() + *ds = append(*ds, *ss...) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/message.go b/vendor/google.golang.org/protobuf/internal/impl/message.go new file mode 100644 index 00000000..7dd994bd --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/message.go @@ -0,0 +1,215 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + "strconv" + "strings" + "sync" + "sync/atomic" + + "google.golang.org/protobuf/internal/genname" + "google.golang.org/protobuf/reflect/protoreflect" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +// MessageInfo provides protobuf related functionality for a given Go type +// that represents a message. A given instance of MessageInfo is tied to +// exactly one Go type, which must be a pointer to a struct type. +// +// The exported fields must be populated before any methods are called +// and cannot be mutated after set. +type MessageInfo struct { + // GoReflectType is the underlying message Go type and must be populated. + GoReflectType reflect.Type // pointer to struct + + // Desc is the underlying message descriptor type and must be populated. + Desc pref.MessageDescriptor + + // Exporter must be provided in a purego environment in order to provide + // access to unexported fields. + Exporter exporter + + // OneofWrappers is list of pointers to oneof wrapper struct types. + OneofWrappers []interface{} + + initMu sync.Mutex // protects all unexported fields + initDone uint32 + + reflectMessageInfo // for reflection implementation + coderMessageInfo // for fast-path method implementations +} + +// exporter is a function that returns a reference to the ith field of v, +// where v is a pointer to a struct. It returns nil if it does not support +// exporting the requested field (e.g., already exported). +type exporter func(v interface{}, i int) interface{} + +// getMessageInfo returns the MessageInfo for any message type that +// is generated by our implementation of protoc-gen-go (for v2 and on). +// If it is unable to obtain a MessageInfo, it returns nil. +func getMessageInfo(mt reflect.Type) *MessageInfo { + m, ok := reflect.Zero(mt).Interface().(pref.ProtoMessage) + if !ok { + return nil + } + mr, ok := m.ProtoReflect().(interface{ ProtoMessageInfo() *MessageInfo }) + if !ok { + return nil + } + return mr.ProtoMessageInfo() +} + +func (mi *MessageInfo) init() { + // This function is called in the hot path. Inline the sync.Once logic, + // since allocating a closure for Once.Do is expensive. + // Keep init small to ensure that it can be inlined. + if atomic.LoadUint32(&mi.initDone) == 0 { + mi.initOnce() + } +} + +func (mi *MessageInfo) initOnce() { + mi.initMu.Lock() + defer mi.initMu.Unlock() + if mi.initDone == 1 { + return + } + + t := mi.GoReflectType + if t.Kind() != reflect.Ptr && t.Elem().Kind() != reflect.Struct { + panic(fmt.Sprintf("got %v, want *struct kind", t)) + } + t = t.Elem() + + si := mi.makeStructInfo(t) + mi.makeReflectFuncs(t, si) + mi.makeCoderMethods(t, si) + + atomic.StoreUint32(&mi.initDone, 1) +} + +// getPointer returns the pointer for a message, which should be of +// the type of the MessageInfo. If the message is of a different type, +// it returns ok==false. +func (mi *MessageInfo) getPointer(m pref.Message) (p pointer, ok bool) { + switch m := m.(type) { + case *messageState: + return m.pointer(), m.messageInfo() == mi + case *messageReflectWrapper: + return m.pointer(), m.messageInfo() == mi + } + return pointer{}, false +} + +type ( + SizeCache = int32 + WeakFields = map[int32]protoreflect.ProtoMessage + UnknownFields = []byte + ExtensionFields = map[int32]ExtensionField +) + +var ( + sizecacheType = reflect.TypeOf(SizeCache(0)) + weakFieldsType = reflect.TypeOf(WeakFields(nil)) + unknownFieldsType = reflect.TypeOf(UnknownFields(nil)) + extensionFieldsType = reflect.TypeOf(ExtensionFields(nil)) +) + +type structInfo struct { + sizecacheOffset offset + weakOffset offset + unknownOffset offset + extensionOffset offset + + fieldsByNumber map[pref.FieldNumber]reflect.StructField + oneofsByName map[pref.Name]reflect.StructField + oneofWrappersByType map[reflect.Type]pref.FieldNumber + oneofWrappersByNumber map[pref.FieldNumber]reflect.Type +} + +func (mi *MessageInfo) makeStructInfo(t reflect.Type) structInfo { + si := structInfo{ + sizecacheOffset: invalidOffset, + weakOffset: invalidOffset, + unknownOffset: invalidOffset, + extensionOffset: invalidOffset, + + fieldsByNumber: map[pref.FieldNumber]reflect.StructField{}, + oneofsByName: map[pref.Name]reflect.StructField{}, + oneofWrappersByType: map[reflect.Type]pref.FieldNumber{}, + oneofWrappersByNumber: map[pref.FieldNumber]reflect.Type{}, + } + +fieldLoop: + for i := 0; i < t.NumField(); i++ { + switch f := t.Field(i); f.Name { + case genname.SizeCache, genname.SizeCacheA: + if f.Type == sizecacheType { + si.sizecacheOffset = offsetOf(f, mi.Exporter) + } + case genname.WeakFields, genname.WeakFieldsA: + if f.Type == weakFieldsType { + si.weakOffset = offsetOf(f, mi.Exporter) + } + case genname.UnknownFields, genname.UnknownFieldsA: + if f.Type == unknownFieldsType { + si.unknownOffset = offsetOf(f, mi.Exporter) + } + case genname.ExtensionFields, genname.ExtensionFieldsA, genname.ExtensionFieldsB: + if f.Type == extensionFieldsType { + si.extensionOffset = offsetOf(f, mi.Exporter) + } + default: + for _, s := range strings.Split(f.Tag.Get("protobuf"), ",") { + if len(s) > 0 && strings.Trim(s, "0123456789") == "" { + n, _ := strconv.ParseUint(s, 10, 64) + si.fieldsByNumber[pref.FieldNumber(n)] = f + continue fieldLoop + } + } + if s := f.Tag.Get("protobuf_oneof"); len(s) > 0 { + si.oneofsByName[pref.Name(s)] = f + continue fieldLoop + } + } + } + + // Derive a mapping of oneof wrappers to fields. + oneofWrappers := mi.OneofWrappers + for _, method := range []string{"XXX_OneofFuncs", "XXX_OneofWrappers"} { + if fn, ok := reflect.PtrTo(t).MethodByName(method); ok { + for _, v := range fn.Func.Call([]reflect.Value{reflect.Zero(fn.Type.In(0))}) { + if vs, ok := v.Interface().([]interface{}); ok { + oneofWrappers = vs + } + } + } + } + for _, v := range oneofWrappers { + tf := reflect.TypeOf(v).Elem() + f := tf.Field(0) + for _, s := range strings.Split(f.Tag.Get("protobuf"), ",") { + if len(s) > 0 && strings.Trim(s, "0123456789") == "" { + n, _ := strconv.ParseUint(s, 10, 64) + si.oneofWrappersByType[tf] = pref.FieldNumber(n) + si.oneofWrappersByNumber[pref.FieldNumber(n)] = tf + break + } + } + } + + return si +} + +func (mi *MessageInfo) New() protoreflect.Message { + return mi.MessageOf(reflect.New(mi.GoReflectType.Elem()).Interface()) +} +func (mi *MessageInfo) Zero() protoreflect.Message { + return mi.MessageOf(reflect.Zero(mi.GoReflectType).Interface()) +} +func (mi *MessageInfo) Descriptor() protoreflect.MessageDescriptor { return mi.Desc } diff --git a/vendor/google.golang.org/protobuf/internal/impl/message_reflect.go b/vendor/google.golang.org/protobuf/internal/impl/message_reflect.go new file mode 100644 index 00000000..0f4b8db7 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/message_reflect.go @@ -0,0 +1,364 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "reflect" + + "google.golang.org/protobuf/internal/pragma" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type reflectMessageInfo struct { + fields map[pref.FieldNumber]*fieldInfo + oneofs map[pref.Name]*oneofInfo + + // denseFields is a subset of fields where: + // 0 < fieldDesc.Number() < len(denseFields) + // It provides faster access to the fieldInfo, but may be incomplete. + denseFields []*fieldInfo + + // rangeInfos is a list of all fields (not belonging to a oneof) and oneofs. + rangeInfos []interface{} // either *fieldInfo or *oneofInfo + + getUnknown func(pointer) pref.RawFields + setUnknown func(pointer, pref.RawFields) + extensionMap func(pointer) *extensionMap + + nilMessage atomicNilMessage +} + +// makeReflectFuncs generates the set of functions to support reflection. +func (mi *MessageInfo) makeReflectFuncs(t reflect.Type, si structInfo) { + mi.makeKnownFieldsFunc(si) + mi.makeUnknownFieldsFunc(t, si) + mi.makeExtensionFieldsFunc(t, si) +} + +// makeKnownFieldsFunc generates functions for operations that can be performed +// on each protobuf message field. It takes in a reflect.Type representing the +// Go struct and matches message fields with struct fields. +// +// This code assumes that the struct is well-formed and panics if there are +// any discrepancies. +func (mi *MessageInfo) makeKnownFieldsFunc(si structInfo) { + mi.fields = map[pref.FieldNumber]*fieldInfo{} + md := mi.Desc + fds := md.Fields() + for i := 0; i < fds.Len(); i++ { + fd := fds.Get(i) + fs := si.fieldsByNumber[fd.Number()] + var fi fieldInfo + switch { + case fd.ContainingOneof() != nil && !fd.ContainingOneof().IsSynthetic(): + fi = fieldInfoForOneof(fd, si.oneofsByName[fd.ContainingOneof().Name()], mi.Exporter, si.oneofWrappersByNumber[fd.Number()]) + case fd.IsMap(): + fi = fieldInfoForMap(fd, fs, mi.Exporter) + case fd.IsList(): + fi = fieldInfoForList(fd, fs, mi.Exporter) + case fd.IsWeak(): + fi = fieldInfoForWeakMessage(fd, si.weakOffset) + case fd.Kind() == pref.MessageKind || fd.Kind() == pref.GroupKind: + fi = fieldInfoForMessage(fd, fs, mi.Exporter) + default: + fi = fieldInfoForScalar(fd, fs, mi.Exporter) + } + mi.fields[fd.Number()] = &fi + } + + mi.oneofs = map[pref.Name]*oneofInfo{} + for i := 0; i < md.Oneofs().Len(); i++ { + od := md.Oneofs().Get(i) + mi.oneofs[od.Name()] = makeOneofInfo(od, si, mi.Exporter) + } + + mi.denseFields = make([]*fieldInfo, fds.Len()*2) + for i := 0; i < fds.Len(); i++ { + if fd := fds.Get(i); int(fd.Number()) < len(mi.denseFields) { + mi.denseFields[fd.Number()] = mi.fields[fd.Number()] + } + } + + for i := 0; i < fds.Len(); { + fd := fds.Get(i) + if od := fd.ContainingOneof(); od != nil && !od.IsSynthetic() { + mi.rangeInfos = append(mi.rangeInfos, mi.oneofs[od.Name()]) + i += od.Fields().Len() + } else { + mi.rangeInfos = append(mi.rangeInfos, mi.fields[fd.Number()]) + i++ + } + } +} + +func (mi *MessageInfo) makeUnknownFieldsFunc(t reflect.Type, si structInfo) { + mi.getUnknown = func(pointer) pref.RawFields { return nil } + mi.setUnknown = func(pointer, pref.RawFields) { return } + if si.unknownOffset.IsValid() { + mi.getUnknown = func(p pointer) pref.RawFields { + if p.IsNil() { + return nil + } + rv := p.Apply(si.unknownOffset).AsValueOf(unknownFieldsType) + return pref.RawFields(*rv.Interface().(*[]byte)) + } + mi.setUnknown = func(p pointer, b pref.RawFields) { + if p.IsNil() { + panic("invalid SetUnknown on nil Message") + } + rv := p.Apply(si.unknownOffset).AsValueOf(unknownFieldsType) + *rv.Interface().(*[]byte) = []byte(b) + } + } else { + mi.getUnknown = func(pointer) pref.RawFields { + return nil + } + mi.setUnknown = func(p pointer, _ pref.RawFields) { + if p.IsNil() { + panic("invalid SetUnknown on nil Message") + } + } + } +} + +func (mi *MessageInfo) makeExtensionFieldsFunc(t reflect.Type, si structInfo) { + if si.extensionOffset.IsValid() { + mi.extensionMap = func(p pointer) *extensionMap { + if p.IsNil() { + return (*extensionMap)(nil) + } + v := p.Apply(si.extensionOffset).AsValueOf(extensionFieldsType) + return (*extensionMap)(v.Interface().(*map[int32]ExtensionField)) + } + } else { + mi.extensionMap = func(pointer) *extensionMap { + return (*extensionMap)(nil) + } + } +} + +type extensionMap map[int32]ExtensionField + +func (m *extensionMap) Range(f func(pref.FieldDescriptor, pref.Value) bool) { + if m != nil { + for _, x := range *m { + xd := x.Type().TypeDescriptor() + v := x.Value() + if xd.IsList() && v.List().Len() == 0 { + continue + } + if !f(xd, v) { + return + } + } + } +} +func (m *extensionMap) Has(xt pref.ExtensionType) (ok bool) { + if m == nil { + return false + } + xd := xt.TypeDescriptor() + x, ok := (*m)[int32(xd.Number())] + if !ok { + return false + } + switch { + case xd.IsList(): + return x.Value().List().Len() > 0 + case xd.IsMap(): + return x.Value().Map().Len() > 0 + case xd.Message() != nil: + return x.Value().Message().IsValid() + } + return true +} +func (m *extensionMap) Clear(xt pref.ExtensionType) { + delete(*m, int32(xt.TypeDescriptor().Number())) +} +func (m *extensionMap) Get(xt pref.ExtensionType) pref.Value { + xd := xt.TypeDescriptor() + if m != nil { + if x, ok := (*m)[int32(xd.Number())]; ok { + return x.Value() + } + } + return xt.Zero() +} +func (m *extensionMap) Set(xt pref.ExtensionType, v pref.Value) { + xd := xt.TypeDescriptor() + isValid := true + switch { + case !xt.IsValidValue(v): + isValid = false + case xd.IsList(): + isValid = v.List().IsValid() + case xd.IsMap(): + isValid = v.Map().IsValid() + case xd.Message() != nil: + isValid = v.Message().IsValid() + } + if !isValid { + panic(fmt.Sprintf("%v: assigning invalid value", xt.TypeDescriptor().FullName())) + } + + if *m == nil { + *m = make(map[int32]ExtensionField) + } + var x ExtensionField + x.Set(xt, v) + (*m)[int32(xd.Number())] = x +} +func (m *extensionMap) Mutable(xt pref.ExtensionType) pref.Value { + xd := xt.TypeDescriptor() + if xd.Kind() != pref.MessageKind && xd.Kind() != pref.GroupKind && !xd.IsList() && !xd.IsMap() { + panic("invalid Mutable on field with non-composite type") + } + if x, ok := (*m)[int32(xd.Number())]; ok { + return x.Value() + } + v := xt.New() + m.Set(xt, v) + return v +} + +// MessageState is a data structure that is nested as the first field in a +// concrete message. It provides a way to implement the ProtoReflect method +// in an allocation-free way without needing to have a shadow Go type generated +// for every message type. This technique only works using unsafe. +// +// +// Example generated code: +// +// type M struct { +// state protoimpl.MessageState +// +// Field1 int32 +// Field2 string +// Field3 *BarMessage +// ... +// } +// +// func (m *M) ProtoReflect() protoreflect.Message { +// mi := &file_fizz_buzz_proto_msgInfos[5] +// if protoimpl.UnsafeEnabled && m != nil { +// ms := protoimpl.X.MessageStateOf(Pointer(m)) +// if ms.LoadMessageInfo() == nil { +// ms.StoreMessageInfo(mi) +// } +// return ms +// } +// return mi.MessageOf(m) +// } +// +// The MessageState type holds a *MessageInfo, which must be atomically set to +// the message info associated with a given message instance. +// By unsafely converting a *M into a *MessageState, the MessageState object +// has access to all the information needed to implement protobuf reflection. +// It has access to the message info as its first field, and a pointer to the +// MessageState is identical to a pointer to the concrete message value. +// +// +// Requirements: +// • The type M must implement protoreflect.ProtoMessage. +// • The address of m must not be nil. +// • The address of m and the address of m.state must be equal, +// even though they are different Go types. +type MessageState struct { + pragma.NoUnkeyedLiterals + pragma.DoNotCompare + pragma.DoNotCopy + + atomicMessageInfo *MessageInfo +} + +type messageState MessageState + +var ( + _ pref.Message = (*messageState)(nil) + _ unwrapper = (*messageState)(nil) +) + +// messageDataType is a tuple of a pointer to the message data and +// a pointer to the message type. It is a generalized way of providing a +// reflective view over a message instance. The disadvantage of this approach +// is the need to allocate this tuple of 16B. +type messageDataType struct { + p pointer + mi *MessageInfo +} + +type ( + messageReflectWrapper messageDataType + messageIfaceWrapper messageDataType +) + +var ( + _ pref.Message = (*messageReflectWrapper)(nil) + _ unwrapper = (*messageReflectWrapper)(nil) + _ pref.ProtoMessage = (*messageIfaceWrapper)(nil) + _ unwrapper = (*messageIfaceWrapper)(nil) +) + +// MessageOf returns a reflective view over a message. The input must be a +// pointer to a named Go struct. If the provided type has a ProtoReflect method, +// it must be implemented by calling this method. +func (mi *MessageInfo) MessageOf(m interface{}) pref.Message { + // TODO: Switch the input to be an opaque Pointer. + if reflect.TypeOf(m) != mi.GoReflectType { + panic(fmt.Sprintf("type mismatch: got %T, want %v", m, mi.GoReflectType)) + } + p := pointerOfIface(m) + if p.IsNil() { + return mi.nilMessage.Init(mi) + } + return &messageReflectWrapper{p, mi} +} + +func (m *messageReflectWrapper) pointer() pointer { return m.p } +func (m *messageReflectWrapper) messageInfo() *MessageInfo { return m.mi } + +func (m *messageIfaceWrapper) ProtoReflect() pref.Message { + return (*messageReflectWrapper)(m) +} +func (m *messageIfaceWrapper) protoUnwrap() interface{} { + return m.p.AsIfaceOf(m.mi.GoReflectType.Elem()) +} + +// checkField verifies that the provided field descriptor is valid. +// Exactly one of the returned values is populated. +func (mi *MessageInfo) checkField(fd pref.FieldDescriptor) (*fieldInfo, pref.ExtensionType) { + var fi *fieldInfo + if n := fd.Number(); 0 < n && int(n) < len(mi.denseFields) { + fi = mi.denseFields[n] + } else { + fi = mi.fields[n] + } + if fi != nil { + if fi.fieldDesc != fd { + if got, want := fd.FullName(), fi.fieldDesc.FullName(); got != want { + panic(fmt.Sprintf("mismatching field: got %v, want %v", got, want)) + } + panic(fmt.Sprintf("mismatching field: %v", fd.FullName())) + } + return fi, nil + } + + if fd.IsExtension() { + if got, want := fd.ContainingMessage().FullName(), mi.Desc.FullName(); got != want { + // TODO: Should this be exact containing message descriptor match? + panic(fmt.Sprintf("extension %v has mismatching containing message: got %v, want %v", fd.FullName(), got, want)) + } + if !mi.Desc.ExtensionRanges().Has(fd.Number()) { + panic(fmt.Sprintf("extension %v extends %v outside the extension range", fd.FullName(), mi.Desc.FullName())) + } + xtd, ok := fd.(pref.ExtensionTypeDescriptor) + if !ok { + panic(fmt.Sprintf("extension %v does not implement protoreflect.ExtensionTypeDescriptor", fd.FullName())) + } + return nil, xtd.Type() + } + panic(fmt.Sprintf("field %v is invalid", fd.FullName())) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/message_reflect_field.go b/vendor/google.golang.org/protobuf/internal/impl/message_reflect_field.go new file mode 100644 index 00000000..23124a86 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/message_reflect_field.go @@ -0,0 +1,466 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "math" + "reflect" + "sync" + + "google.golang.org/protobuf/internal/flags" + pref "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" +) + +type fieldInfo struct { + fieldDesc pref.FieldDescriptor + + // These fields are used for protobuf reflection support. + has func(pointer) bool + clear func(pointer) + get func(pointer) pref.Value + set func(pointer, pref.Value) + mutable func(pointer) pref.Value + newMessage func() pref.Message + newField func() pref.Value +} + +func fieldInfoForOneof(fd pref.FieldDescriptor, fs reflect.StructField, x exporter, ot reflect.Type) fieldInfo { + ft := fs.Type + if ft.Kind() != reflect.Interface { + panic(fmt.Sprintf("field %v has invalid type: got %v, want interface kind", fd.FullName(), ft)) + } + if ot.Kind() != reflect.Struct { + panic(fmt.Sprintf("field %v has invalid type: got %v, want struct kind", fd.FullName(), ot)) + } + if !reflect.PtrTo(ot).Implements(ft) { + panic(fmt.Sprintf("field %v has invalid type: %v does not implement %v", fd.FullName(), ot, ft)) + } + conv := NewConverter(ot.Field(0).Type, fd) + isMessage := fd.Message() != nil + + // TODO: Implement unsafe fast path? + fieldOffset := offsetOf(fs, x) + return fieldInfo{ + // NOTE: The logic below intentionally assumes that oneof fields are + // well-formatted. That is, the oneof interface never contains a + // typed nil pointer to one of the wrapper structs. + + fieldDesc: fd, + has: func(p pointer) bool { + if p.IsNil() { + return false + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() || rv.Elem().Type().Elem() != ot || rv.Elem().IsNil() { + return false + } + return true + }, + clear: func(p pointer) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() || rv.Elem().Type().Elem() != ot { + // NOTE: We intentionally don't check for rv.Elem().IsNil() + // so that (*OneofWrapperType)(nil) gets cleared to nil. + return + } + rv.Set(reflect.Zero(rv.Type())) + }, + get: func(p pointer) pref.Value { + if p.IsNil() { + return conv.Zero() + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() || rv.Elem().Type().Elem() != ot || rv.Elem().IsNil() { + return conv.Zero() + } + rv = rv.Elem().Elem().Field(0) + return conv.PBValueOf(rv) + }, + set: func(p pointer, v pref.Value) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() || rv.Elem().Type().Elem() != ot || rv.Elem().IsNil() { + rv.Set(reflect.New(ot)) + } + rv = rv.Elem().Elem().Field(0) + rv.Set(conv.GoValueOf(v)) + }, + mutable: func(p pointer) pref.Value { + if !isMessage { + panic(fmt.Sprintf("field %v with invalid Mutable call on field with non-composite type", fd.FullName())) + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() || rv.Elem().Type().Elem() != ot || rv.Elem().IsNil() { + rv.Set(reflect.New(ot)) + } + rv = rv.Elem().Elem().Field(0) + if rv.IsNil() { + rv.Set(conv.GoValueOf(pref.ValueOfMessage(conv.New().Message()))) + } + return conv.PBValueOf(rv) + }, + newMessage: func() pref.Message { + return conv.New().Message() + }, + newField: func() pref.Value { + return conv.New() + }, + } +} + +func fieldInfoForMap(fd pref.FieldDescriptor, fs reflect.StructField, x exporter) fieldInfo { + ft := fs.Type + if ft.Kind() != reflect.Map { + panic(fmt.Sprintf("field %v has invalid type: got %v, want map kind", fd.FullName(), ft)) + } + conv := NewConverter(ft, fd) + + // TODO: Implement unsafe fast path? + fieldOffset := offsetOf(fs, x) + return fieldInfo{ + fieldDesc: fd, + has: func(p pointer) bool { + if p.IsNil() { + return false + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + return rv.Len() > 0 + }, + clear: func(p pointer) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + rv.Set(reflect.Zero(rv.Type())) + }, + get: func(p pointer) pref.Value { + if p.IsNil() { + return conv.Zero() + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.Len() == 0 { + return conv.Zero() + } + return conv.PBValueOf(rv) + }, + set: func(p pointer, v pref.Value) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + pv := conv.GoValueOf(v) + if pv.IsNil() { + panic(fmt.Sprintf("map field %v cannot be set with read-only value", fd.FullName())) + } + rv.Set(pv) + }, + mutable: func(p pointer) pref.Value { + v := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if v.IsNil() { + v.Set(reflect.MakeMap(fs.Type)) + } + return conv.PBValueOf(v) + }, + newField: func() pref.Value { + return conv.New() + }, + } +} + +func fieldInfoForList(fd pref.FieldDescriptor, fs reflect.StructField, x exporter) fieldInfo { + ft := fs.Type + if ft.Kind() != reflect.Slice { + panic(fmt.Sprintf("field %v has invalid type: got %v, want slice kind", fd.FullName(), ft)) + } + conv := NewConverter(reflect.PtrTo(ft), fd) + + // TODO: Implement unsafe fast path? + fieldOffset := offsetOf(fs, x) + return fieldInfo{ + fieldDesc: fd, + has: func(p pointer) bool { + if p.IsNil() { + return false + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + return rv.Len() > 0 + }, + clear: func(p pointer) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + rv.Set(reflect.Zero(rv.Type())) + }, + get: func(p pointer) pref.Value { + if p.IsNil() { + return conv.Zero() + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type) + if rv.Elem().Len() == 0 { + return conv.Zero() + } + return conv.PBValueOf(rv) + }, + set: func(p pointer, v pref.Value) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + pv := conv.GoValueOf(v) + if pv.IsNil() { + panic(fmt.Sprintf("list field %v cannot be set with read-only value", fd.FullName())) + } + rv.Set(pv.Elem()) + }, + mutable: func(p pointer) pref.Value { + v := p.Apply(fieldOffset).AsValueOf(fs.Type) + return conv.PBValueOf(v) + }, + newField: func() pref.Value { + return conv.New() + }, + } +} + +var ( + nilBytes = reflect.ValueOf([]byte(nil)) + emptyBytes = reflect.ValueOf([]byte{}) +) + +func fieldInfoForScalar(fd pref.FieldDescriptor, fs reflect.StructField, x exporter) fieldInfo { + ft := fs.Type + nullable := fd.HasPresence() + isBytes := ft.Kind() == reflect.Slice && ft.Elem().Kind() == reflect.Uint8 + if nullable { + if ft.Kind() != reflect.Ptr && ft.Kind() != reflect.Slice { + panic(fmt.Sprintf("field %v has invalid type: got %v, want pointer", fd.FullName(), ft)) + } + if ft.Kind() == reflect.Ptr { + ft = ft.Elem() + } + } + conv := NewConverter(ft, fd) + + // TODO: Implement unsafe fast path? + fieldOffset := offsetOf(fs, x) + return fieldInfo{ + fieldDesc: fd, + has: func(p pointer) bool { + if p.IsNil() { + return false + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if nullable { + return !rv.IsNil() + } + switch rv.Kind() { + case reflect.Bool: + return rv.Bool() + case reflect.Int32, reflect.Int64: + return rv.Int() != 0 + case reflect.Uint32, reflect.Uint64: + return rv.Uint() != 0 + case reflect.Float32, reflect.Float64: + return rv.Float() != 0 || math.Signbit(rv.Float()) + case reflect.String, reflect.Slice: + return rv.Len() > 0 + default: + panic(fmt.Sprintf("field %v has invalid type: %v", fd.FullName(), rv.Type())) // should never happen + } + }, + clear: func(p pointer) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + rv.Set(reflect.Zero(rv.Type())) + }, + get: func(p pointer) pref.Value { + if p.IsNil() { + return conv.Zero() + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if nullable { + if rv.IsNil() { + return conv.Zero() + } + if rv.Kind() == reflect.Ptr { + rv = rv.Elem() + } + } + return conv.PBValueOf(rv) + }, + set: func(p pointer, v pref.Value) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if nullable && rv.Kind() == reflect.Ptr { + if rv.IsNil() { + rv.Set(reflect.New(ft)) + } + rv = rv.Elem() + } + rv.Set(conv.GoValueOf(v)) + if isBytes && rv.Len() == 0 { + if nullable { + rv.Set(emptyBytes) // preserve presence + } else { + rv.Set(nilBytes) // do not preserve presence + } + } + }, + newField: func() pref.Value { + return conv.New() + }, + } +} + +func fieldInfoForWeakMessage(fd pref.FieldDescriptor, weakOffset offset) fieldInfo { + if !flags.ProtoLegacy { + panic("no support for proto1 weak fields") + } + + var once sync.Once + var messageType pref.MessageType + lazyInit := func() { + once.Do(func() { + messageName := fd.Message().FullName() + messageType, _ = preg.GlobalTypes.FindMessageByName(messageName) + if messageType == nil { + panic(fmt.Sprintf("weak message %v for field %v is not linked in", messageName, fd.FullName())) + } + }) + } + + num := fd.Number() + return fieldInfo{ + fieldDesc: fd, + has: func(p pointer) bool { + if p.IsNil() { + return false + } + _, ok := p.Apply(weakOffset).WeakFields().get(num) + return ok + }, + clear: func(p pointer) { + p.Apply(weakOffset).WeakFields().clear(num) + }, + get: func(p pointer) pref.Value { + lazyInit() + if p.IsNil() { + return pref.ValueOfMessage(messageType.Zero()) + } + m, ok := p.Apply(weakOffset).WeakFields().get(num) + if !ok { + return pref.ValueOfMessage(messageType.Zero()) + } + return pref.ValueOfMessage(m.ProtoReflect()) + }, + set: func(p pointer, v pref.Value) { + lazyInit() + m := v.Message() + if m.Descriptor() != messageType.Descriptor() { + if got, want := m.Descriptor().FullName(), messageType.Descriptor().FullName(); got != want { + panic(fmt.Sprintf("field %v has mismatching message descriptor: got %v, want %v", fd.FullName(), got, want)) + } + panic(fmt.Sprintf("field %v has mismatching message descriptor: %v", fd.FullName(), m.Descriptor().FullName())) + } + p.Apply(weakOffset).WeakFields().set(num, m.Interface()) + }, + mutable: func(p pointer) pref.Value { + lazyInit() + fs := p.Apply(weakOffset).WeakFields() + m, ok := fs.get(num) + if !ok { + m = messageType.New().Interface() + fs.set(num, m) + } + return pref.ValueOfMessage(m.ProtoReflect()) + }, + newMessage: func() pref.Message { + lazyInit() + return messageType.New() + }, + newField: func() pref.Value { + lazyInit() + return pref.ValueOfMessage(messageType.New()) + }, + } +} + +func fieldInfoForMessage(fd pref.FieldDescriptor, fs reflect.StructField, x exporter) fieldInfo { + ft := fs.Type + conv := NewConverter(ft, fd) + + // TODO: Implement unsafe fast path? + fieldOffset := offsetOf(fs, x) + return fieldInfo{ + fieldDesc: fd, + has: func(p pointer) bool { + if p.IsNil() { + return false + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + return !rv.IsNil() + }, + clear: func(p pointer) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + rv.Set(reflect.Zero(rv.Type())) + }, + get: func(p pointer) pref.Value { + if p.IsNil() { + return conv.Zero() + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + return conv.PBValueOf(rv) + }, + set: func(p pointer, v pref.Value) { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + rv.Set(conv.GoValueOf(v)) + if rv.IsNil() { + panic(fmt.Sprintf("field %v has invalid nil pointer", fd.FullName())) + } + }, + mutable: func(p pointer) pref.Value { + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() { + rv.Set(conv.GoValueOf(conv.New())) + } + return conv.PBValueOf(rv) + }, + newMessage: func() pref.Message { + return conv.New().Message() + }, + newField: func() pref.Value { + return conv.New() + }, + } +} + +type oneofInfo struct { + oneofDesc pref.OneofDescriptor + which func(pointer) pref.FieldNumber +} + +func makeOneofInfo(od pref.OneofDescriptor, si structInfo, x exporter) *oneofInfo { + oi := &oneofInfo{oneofDesc: od} + if od.IsSynthetic() { + fs := si.fieldsByNumber[od.Fields().Get(0).Number()] + fieldOffset := offsetOf(fs, x) + oi.which = func(p pointer) pref.FieldNumber { + if p.IsNil() { + return 0 + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() { // valid on either *T or []byte + return 0 + } + return od.Fields().Get(0).Number() + } + } else { + fs := si.oneofsByName[od.Name()] + fieldOffset := offsetOf(fs, x) + oi.which = func(p pointer) pref.FieldNumber { + if p.IsNil() { + return 0 + } + rv := p.Apply(fieldOffset).AsValueOf(fs.Type).Elem() + if rv.IsNil() { + return 0 + } + rv = rv.Elem() + if rv.IsNil() { + return 0 + } + return si.oneofWrappersByType[rv.Type().Elem()] + } + } + return oi +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/message_reflect_gen.go b/vendor/google.golang.org/protobuf/internal/impl/message_reflect_gen.go new file mode 100644 index 00000000..741d6e5b --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/message_reflect_gen.go @@ -0,0 +1,249 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-types. DO NOT EDIT. + +package impl + +import ( + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" +) + +func (m *messageState) Descriptor() protoreflect.MessageDescriptor { + return m.messageInfo().Desc +} +func (m *messageState) Type() protoreflect.MessageType { + return m.messageInfo() +} +func (m *messageState) New() protoreflect.Message { + return m.messageInfo().New() +} +func (m *messageState) Interface() protoreflect.ProtoMessage { + return m.protoUnwrap().(protoreflect.ProtoMessage) +} +func (m *messageState) protoUnwrap() interface{} { + return m.pointer().AsIfaceOf(m.messageInfo().GoReflectType.Elem()) +} +func (m *messageState) ProtoMethods() *protoiface.Methods { + m.messageInfo().init() + return &m.messageInfo().methods +} + +// ProtoMessageInfo is a pseudo-internal API for allowing the v1 code +// to be able to retrieve a v2 MessageInfo struct. +// +// WARNING: This method is exempt from the compatibility promise and +// may be removed in the future without warning. +func (m *messageState) ProtoMessageInfo() *MessageInfo { + return m.messageInfo() +} + +func (m *messageState) Range(f func(protoreflect.FieldDescriptor, protoreflect.Value) bool) { + m.messageInfo().init() + for _, ri := range m.messageInfo().rangeInfos { + switch ri := ri.(type) { + case *fieldInfo: + if ri.has(m.pointer()) { + if !f(ri.fieldDesc, ri.get(m.pointer())) { + return + } + } + case *oneofInfo: + if n := ri.which(m.pointer()); n > 0 { + fi := m.messageInfo().fields[n] + if !f(fi.fieldDesc, fi.get(m.pointer())) { + return + } + } + } + } + m.messageInfo().extensionMap(m.pointer()).Range(f) +} +func (m *messageState) Has(fd protoreflect.FieldDescriptor) bool { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.has(m.pointer()) + } else { + return m.messageInfo().extensionMap(m.pointer()).Has(xt) + } +} +func (m *messageState) Clear(fd protoreflect.FieldDescriptor) { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + fi.clear(m.pointer()) + } else { + m.messageInfo().extensionMap(m.pointer()).Clear(xt) + } +} +func (m *messageState) Get(fd protoreflect.FieldDescriptor) protoreflect.Value { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.get(m.pointer()) + } else { + return m.messageInfo().extensionMap(m.pointer()).Get(xt) + } +} +func (m *messageState) Set(fd protoreflect.FieldDescriptor, v protoreflect.Value) { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + fi.set(m.pointer(), v) + } else { + m.messageInfo().extensionMap(m.pointer()).Set(xt, v) + } +} +func (m *messageState) Mutable(fd protoreflect.FieldDescriptor) protoreflect.Value { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.mutable(m.pointer()) + } else { + return m.messageInfo().extensionMap(m.pointer()).Mutable(xt) + } +} +func (m *messageState) NewField(fd protoreflect.FieldDescriptor) protoreflect.Value { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.newField() + } else { + return xt.New() + } +} +func (m *messageState) WhichOneof(od protoreflect.OneofDescriptor) protoreflect.FieldDescriptor { + m.messageInfo().init() + if oi := m.messageInfo().oneofs[od.Name()]; oi != nil && oi.oneofDesc == od { + return od.Fields().ByNumber(oi.which(m.pointer())) + } + panic("invalid oneof descriptor " + string(od.FullName()) + " for message " + string(m.Descriptor().FullName())) +} +func (m *messageState) GetUnknown() protoreflect.RawFields { + m.messageInfo().init() + return m.messageInfo().getUnknown(m.pointer()) +} +func (m *messageState) SetUnknown(b protoreflect.RawFields) { + m.messageInfo().init() + m.messageInfo().setUnknown(m.pointer(), b) +} +func (m *messageState) IsValid() bool { + return !m.pointer().IsNil() +} + +func (m *messageReflectWrapper) Descriptor() protoreflect.MessageDescriptor { + return m.messageInfo().Desc +} +func (m *messageReflectWrapper) Type() protoreflect.MessageType { + return m.messageInfo() +} +func (m *messageReflectWrapper) New() protoreflect.Message { + return m.messageInfo().New() +} +func (m *messageReflectWrapper) Interface() protoreflect.ProtoMessage { + if m, ok := m.protoUnwrap().(protoreflect.ProtoMessage); ok { + return m + } + return (*messageIfaceWrapper)(m) +} +func (m *messageReflectWrapper) protoUnwrap() interface{} { + return m.pointer().AsIfaceOf(m.messageInfo().GoReflectType.Elem()) +} +func (m *messageReflectWrapper) ProtoMethods() *protoiface.Methods { + m.messageInfo().init() + return &m.messageInfo().methods +} + +// ProtoMessageInfo is a pseudo-internal API for allowing the v1 code +// to be able to retrieve a v2 MessageInfo struct. +// +// WARNING: This method is exempt from the compatibility promise and +// may be removed in the future without warning. +func (m *messageReflectWrapper) ProtoMessageInfo() *MessageInfo { + return m.messageInfo() +} + +func (m *messageReflectWrapper) Range(f func(protoreflect.FieldDescriptor, protoreflect.Value) bool) { + m.messageInfo().init() + for _, ri := range m.messageInfo().rangeInfos { + switch ri := ri.(type) { + case *fieldInfo: + if ri.has(m.pointer()) { + if !f(ri.fieldDesc, ri.get(m.pointer())) { + return + } + } + case *oneofInfo: + if n := ri.which(m.pointer()); n > 0 { + fi := m.messageInfo().fields[n] + if !f(fi.fieldDesc, fi.get(m.pointer())) { + return + } + } + } + } + m.messageInfo().extensionMap(m.pointer()).Range(f) +} +func (m *messageReflectWrapper) Has(fd protoreflect.FieldDescriptor) bool { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.has(m.pointer()) + } else { + return m.messageInfo().extensionMap(m.pointer()).Has(xt) + } +} +func (m *messageReflectWrapper) Clear(fd protoreflect.FieldDescriptor) { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + fi.clear(m.pointer()) + } else { + m.messageInfo().extensionMap(m.pointer()).Clear(xt) + } +} +func (m *messageReflectWrapper) Get(fd protoreflect.FieldDescriptor) protoreflect.Value { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.get(m.pointer()) + } else { + return m.messageInfo().extensionMap(m.pointer()).Get(xt) + } +} +func (m *messageReflectWrapper) Set(fd protoreflect.FieldDescriptor, v protoreflect.Value) { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + fi.set(m.pointer(), v) + } else { + m.messageInfo().extensionMap(m.pointer()).Set(xt, v) + } +} +func (m *messageReflectWrapper) Mutable(fd protoreflect.FieldDescriptor) protoreflect.Value { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.mutable(m.pointer()) + } else { + return m.messageInfo().extensionMap(m.pointer()).Mutable(xt) + } +} +func (m *messageReflectWrapper) NewField(fd protoreflect.FieldDescriptor) protoreflect.Value { + m.messageInfo().init() + if fi, xt := m.messageInfo().checkField(fd); fi != nil { + return fi.newField() + } else { + return xt.New() + } +} +func (m *messageReflectWrapper) WhichOneof(od protoreflect.OneofDescriptor) protoreflect.FieldDescriptor { + m.messageInfo().init() + if oi := m.messageInfo().oneofs[od.Name()]; oi != nil && oi.oneofDesc == od { + return od.Fields().ByNumber(oi.which(m.pointer())) + } + panic("invalid oneof descriptor " + string(od.FullName()) + " for message " + string(m.Descriptor().FullName())) +} +func (m *messageReflectWrapper) GetUnknown() protoreflect.RawFields { + m.messageInfo().init() + return m.messageInfo().getUnknown(m.pointer()) +} +func (m *messageReflectWrapper) SetUnknown(b protoreflect.RawFields) { + m.messageInfo().init() + m.messageInfo().setUnknown(m.pointer(), b) +} +func (m *messageReflectWrapper) IsValid() bool { + return !m.pointer().IsNil() +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/pointer_reflect.go b/vendor/google.golang.org/protobuf/internal/impl/pointer_reflect.go new file mode 100644 index 00000000..67b4ede6 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/pointer_reflect.go @@ -0,0 +1,177 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build purego appengine + +package impl + +import ( + "fmt" + "reflect" + "sync" +) + +const UnsafeEnabled = false + +// Pointer is an opaque pointer type. +type Pointer interface{} + +// offset represents the offset to a struct field, accessible from a pointer. +// The offset is the field index into a struct. +type offset struct { + index int + export exporter +} + +// offsetOf returns a field offset for the struct field. +func offsetOf(f reflect.StructField, x exporter) offset { + if len(f.Index) != 1 { + panic("embedded structs are not supported") + } + if f.PkgPath == "" { + return offset{index: f.Index[0]} // field is already exported + } + if x == nil { + panic("exporter must be provided for unexported field") + } + return offset{index: f.Index[0], export: x} +} + +// IsValid reports whether the offset is valid. +func (f offset) IsValid() bool { return f.index >= 0 } + +// invalidOffset is an invalid field offset. +var invalidOffset = offset{index: -1} + +// zeroOffset is a noop when calling pointer.Apply. +var zeroOffset = offset{index: 0} + +// pointer is an abstract representation of a pointer to a struct or field. +type pointer struct{ v reflect.Value } + +// pointerOf returns p as a pointer. +func pointerOf(p Pointer) pointer { + return pointerOfIface(p) +} + +// pointerOfValue returns v as a pointer. +func pointerOfValue(v reflect.Value) pointer { + return pointer{v: v} +} + +// pointerOfIface returns the pointer portion of an interface. +func pointerOfIface(v interface{}) pointer { + return pointer{v: reflect.ValueOf(v)} +} + +// IsNil reports whether the pointer is nil. +func (p pointer) IsNil() bool { + return p.v.IsNil() +} + +// Apply adds an offset to the pointer to derive a new pointer +// to a specified field. The current pointer must be pointing at a struct. +func (p pointer) Apply(f offset) pointer { + if f.export != nil { + if v := reflect.ValueOf(f.export(p.v.Interface(), f.index)); v.IsValid() { + return pointer{v: v} + } + } + return pointer{v: p.v.Elem().Field(f.index).Addr()} +} + +// AsValueOf treats p as a pointer to an object of type t and returns the value. +// It is equivalent to reflect.ValueOf(p.AsIfaceOf(t)) +func (p pointer) AsValueOf(t reflect.Type) reflect.Value { + if got := p.v.Type().Elem(); got != t { + panic(fmt.Sprintf("invalid type: got %v, want %v", got, t)) + } + return p.v +} + +// AsIfaceOf treats p as a pointer to an object of type t and returns the value. +// It is equivalent to p.AsValueOf(t).Interface() +func (p pointer) AsIfaceOf(t reflect.Type) interface{} { + return p.AsValueOf(t).Interface() +} + +func (p pointer) Bool() *bool { return p.v.Interface().(*bool) } +func (p pointer) BoolPtr() **bool { return p.v.Interface().(**bool) } +func (p pointer) BoolSlice() *[]bool { return p.v.Interface().(*[]bool) } +func (p pointer) Int32() *int32 { return p.v.Interface().(*int32) } +func (p pointer) Int32Ptr() **int32 { return p.v.Interface().(**int32) } +func (p pointer) Int32Slice() *[]int32 { return p.v.Interface().(*[]int32) } +func (p pointer) Int64() *int64 { return p.v.Interface().(*int64) } +func (p pointer) Int64Ptr() **int64 { return p.v.Interface().(**int64) } +func (p pointer) Int64Slice() *[]int64 { return p.v.Interface().(*[]int64) } +func (p pointer) Uint32() *uint32 { return p.v.Interface().(*uint32) } +func (p pointer) Uint32Ptr() **uint32 { return p.v.Interface().(**uint32) } +func (p pointer) Uint32Slice() *[]uint32 { return p.v.Interface().(*[]uint32) } +func (p pointer) Uint64() *uint64 { return p.v.Interface().(*uint64) } +func (p pointer) Uint64Ptr() **uint64 { return p.v.Interface().(**uint64) } +func (p pointer) Uint64Slice() *[]uint64 { return p.v.Interface().(*[]uint64) } +func (p pointer) Float32() *float32 { return p.v.Interface().(*float32) } +func (p pointer) Float32Ptr() **float32 { return p.v.Interface().(**float32) } +func (p pointer) Float32Slice() *[]float32 { return p.v.Interface().(*[]float32) } +func (p pointer) Float64() *float64 { return p.v.Interface().(*float64) } +func (p pointer) Float64Ptr() **float64 { return p.v.Interface().(**float64) } +func (p pointer) Float64Slice() *[]float64 { return p.v.Interface().(*[]float64) } +func (p pointer) String() *string { return p.v.Interface().(*string) } +func (p pointer) StringPtr() **string { return p.v.Interface().(**string) } +func (p pointer) StringSlice() *[]string { return p.v.Interface().(*[]string) } +func (p pointer) Bytes() *[]byte { return p.v.Interface().(*[]byte) } +func (p pointer) BytesSlice() *[][]byte { return p.v.Interface().(*[][]byte) } +func (p pointer) WeakFields() *weakFields { return (*weakFields)(p.v.Interface().(*WeakFields)) } +func (p pointer) Extensions() *map[int32]ExtensionField { + return p.v.Interface().(*map[int32]ExtensionField) +} + +func (p pointer) Elem() pointer { + return pointer{v: p.v.Elem()} +} + +// PointerSlice copies []*T from p as a new []pointer. +// This behavior differs from the implementation in pointer_unsafe.go. +func (p pointer) PointerSlice() []pointer { + // TODO: reconsider this + if p.v.IsNil() { + return nil + } + n := p.v.Elem().Len() + s := make([]pointer, n) + for i := 0; i < n; i++ { + s[i] = pointer{v: p.v.Elem().Index(i)} + } + return s +} + +// AppendPointerSlice appends v to p, which must be a []*T. +func (p pointer) AppendPointerSlice(v pointer) { + sp := p.v.Elem() + sp.Set(reflect.Append(sp, v.v)) +} + +// SetPointer sets *p to v. +func (p pointer) SetPointer(v pointer) { + p.v.Elem().Set(v.v) +} + +func (Export) MessageStateOf(p Pointer) *messageState { panic("not supported") } +func (ms *messageState) pointer() pointer { panic("not supported") } +func (ms *messageState) messageInfo() *MessageInfo { panic("not supported") } +func (ms *messageState) LoadMessageInfo() *MessageInfo { panic("not supported") } +func (ms *messageState) StoreMessageInfo(mi *MessageInfo) { panic("not supported") } + +type atomicNilMessage struct { + once sync.Once + m messageReflectWrapper +} + +func (m *atomicNilMessage) Init(mi *MessageInfo) *messageReflectWrapper { + m.once.Do(func() { + m.m.p = pointerOfIface(reflect.Zero(mi.GoReflectType).Interface()) + m.m.mi = mi + }) + return &m.m +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/pointer_unsafe.go b/vendor/google.golang.org/protobuf/internal/impl/pointer_unsafe.go new file mode 100644 index 00000000..088aa85d --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/pointer_unsafe.go @@ -0,0 +1,173 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !purego,!appengine + +package impl + +import ( + "reflect" + "sync/atomic" + "unsafe" +) + +const UnsafeEnabled = true + +// Pointer is an opaque pointer type. +type Pointer unsafe.Pointer + +// offset represents the offset to a struct field, accessible from a pointer. +// The offset is the byte offset to the field from the start of the struct. +type offset uintptr + +// offsetOf returns a field offset for the struct field. +func offsetOf(f reflect.StructField, x exporter) offset { + return offset(f.Offset) +} + +// IsValid reports whether the offset is valid. +func (f offset) IsValid() bool { return f != invalidOffset } + +// invalidOffset is an invalid field offset. +var invalidOffset = ^offset(0) + +// zeroOffset is a noop when calling pointer.Apply. +var zeroOffset = offset(0) + +// pointer is a pointer to a message struct or field. +type pointer struct{ p unsafe.Pointer } + +// pointerOf returns p as a pointer. +func pointerOf(p Pointer) pointer { + return pointer{p: unsafe.Pointer(p)} +} + +// pointerOfValue returns v as a pointer. +func pointerOfValue(v reflect.Value) pointer { + return pointer{p: unsafe.Pointer(v.Pointer())} +} + +// pointerOfIface returns the pointer portion of an interface. +func pointerOfIface(v interface{}) pointer { + type ifaceHeader struct { + Type unsafe.Pointer + Data unsafe.Pointer + } + return pointer{p: (*ifaceHeader)(unsafe.Pointer(&v)).Data} +} + +// IsNil reports whether the pointer is nil. +func (p pointer) IsNil() bool { + return p.p == nil +} + +// Apply adds an offset to the pointer to derive a new pointer +// to a specified field. The pointer must be valid and pointing at a struct. +func (p pointer) Apply(f offset) pointer { + if p.IsNil() { + panic("invalid nil pointer") + } + return pointer{p: unsafe.Pointer(uintptr(p.p) + uintptr(f))} +} + +// AsValueOf treats p as a pointer to an object of type t and returns the value. +// It is equivalent to reflect.ValueOf(p.AsIfaceOf(t)) +func (p pointer) AsValueOf(t reflect.Type) reflect.Value { + return reflect.NewAt(t, p.p) +} + +// AsIfaceOf treats p as a pointer to an object of type t and returns the value. +// It is equivalent to p.AsValueOf(t).Interface() +func (p pointer) AsIfaceOf(t reflect.Type) interface{} { + // TODO: Use tricky unsafe magic to directly create ifaceHeader. + return p.AsValueOf(t).Interface() +} + +func (p pointer) Bool() *bool { return (*bool)(p.p) } +func (p pointer) BoolPtr() **bool { return (**bool)(p.p) } +func (p pointer) BoolSlice() *[]bool { return (*[]bool)(p.p) } +func (p pointer) Int32() *int32 { return (*int32)(p.p) } +func (p pointer) Int32Ptr() **int32 { return (**int32)(p.p) } +func (p pointer) Int32Slice() *[]int32 { return (*[]int32)(p.p) } +func (p pointer) Int64() *int64 { return (*int64)(p.p) } +func (p pointer) Int64Ptr() **int64 { return (**int64)(p.p) } +func (p pointer) Int64Slice() *[]int64 { return (*[]int64)(p.p) } +func (p pointer) Uint32() *uint32 { return (*uint32)(p.p) } +func (p pointer) Uint32Ptr() **uint32 { return (**uint32)(p.p) } +func (p pointer) Uint32Slice() *[]uint32 { return (*[]uint32)(p.p) } +func (p pointer) Uint64() *uint64 { return (*uint64)(p.p) } +func (p pointer) Uint64Ptr() **uint64 { return (**uint64)(p.p) } +func (p pointer) Uint64Slice() *[]uint64 { return (*[]uint64)(p.p) } +func (p pointer) Float32() *float32 { return (*float32)(p.p) } +func (p pointer) Float32Ptr() **float32 { return (**float32)(p.p) } +func (p pointer) Float32Slice() *[]float32 { return (*[]float32)(p.p) } +func (p pointer) Float64() *float64 { return (*float64)(p.p) } +func (p pointer) Float64Ptr() **float64 { return (**float64)(p.p) } +func (p pointer) Float64Slice() *[]float64 { return (*[]float64)(p.p) } +func (p pointer) String() *string { return (*string)(p.p) } +func (p pointer) StringPtr() **string { return (**string)(p.p) } +func (p pointer) StringSlice() *[]string { return (*[]string)(p.p) } +func (p pointer) Bytes() *[]byte { return (*[]byte)(p.p) } +func (p pointer) BytesSlice() *[][]byte { return (*[][]byte)(p.p) } +func (p pointer) WeakFields() *weakFields { return (*weakFields)(p.p) } +func (p pointer) Extensions() *map[int32]ExtensionField { return (*map[int32]ExtensionField)(p.p) } + +func (p pointer) Elem() pointer { + return pointer{p: *(*unsafe.Pointer)(p.p)} +} + +// PointerSlice loads []*T from p as a []pointer. +// The value returned is aliased with the original slice. +// This behavior differs from the implementation in pointer_reflect.go. +func (p pointer) PointerSlice() []pointer { + // Super-tricky - p should point to a []*T where T is a + // message type. We load it as []pointer. + return *(*[]pointer)(p.p) +} + +// AppendPointerSlice appends v to p, which must be a []*T. +func (p pointer) AppendPointerSlice(v pointer) { + *(*[]pointer)(p.p) = append(*(*[]pointer)(p.p), v) +} + +// SetPointer sets *p to v. +func (p pointer) SetPointer(v pointer) { + *(*unsafe.Pointer)(p.p) = (unsafe.Pointer)(v.p) +} + +// Static check that MessageState does not exceed the size of a pointer. +const _ = uint(unsafe.Sizeof(unsafe.Pointer(nil)) - unsafe.Sizeof(MessageState{})) + +func (Export) MessageStateOf(p Pointer) *messageState { + // Super-tricky - see documentation on MessageState. + return (*messageState)(unsafe.Pointer(p)) +} +func (ms *messageState) pointer() pointer { + // Super-tricky - see documentation on MessageState. + return pointer{p: unsafe.Pointer(ms)} +} +func (ms *messageState) messageInfo() *MessageInfo { + mi := ms.LoadMessageInfo() + if mi == nil { + panic("invalid nil message info; this suggests memory corruption due to a race or shallow copy on the message struct") + } + return mi +} +func (ms *messageState) LoadMessageInfo() *MessageInfo { + return (*MessageInfo)(atomic.LoadPointer((*unsafe.Pointer)(unsafe.Pointer(&ms.atomicMessageInfo)))) +} +func (ms *messageState) StoreMessageInfo(mi *MessageInfo) { + atomic.StorePointer((*unsafe.Pointer)(unsafe.Pointer(&ms.atomicMessageInfo)), unsafe.Pointer(mi)) +} + +type atomicNilMessage struct{ p unsafe.Pointer } // p is a *messageReflectWrapper + +func (m *atomicNilMessage) Init(mi *MessageInfo) *messageReflectWrapper { + if p := atomic.LoadPointer(&m.p); p != nil { + return (*messageReflectWrapper)(p) + } + w := &messageReflectWrapper{mi: mi} + atomic.CompareAndSwapPointer(&m.p, nil, (unsafe.Pointer)(w)) + return (*messageReflectWrapper)(atomic.LoadPointer(&m.p)) +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/validate.go b/vendor/google.golang.org/protobuf/internal/impl/validate.go new file mode 100644 index 00000000..57de9cc8 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/validate.go @@ -0,0 +1,575 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + "math" + "math/bits" + "reflect" + "unicode/utf8" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/flags" + "google.golang.org/protobuf/internal/strs" + pref "google.golang.org/protobuf/reflect/protoreflect" + preg "google.golang.org/protobuf/reflect/protoregistry" + piface "google.golang.org/protobuf/runtime/protoiface" +) + +// ValidationStatus is the result of validating the wire-format encoding of a message. +type ValidationStatus int + +const ( + // ValidationUnknown indicates that unmarshaling the message might succeed or fail. + // The validator was unable to render a judgement. + // + // The only causes of this status are an aberrant message type appearing somewhere + // in the message or a failure in the extension resolver. + ValidationUnknown ValidationStatus = iota + 1 + + // ValidationInvalid indicates that unmarshaling the message will fail. + ValidationInvalid + + // ValidationValid indicates that unmarshaling the message will succeed. + ValidationValid +) + +func (v ValidationStatus) String() string { + switch v { + case ValidationUnknown: + return "ValidationUnknown" + case ValidationInvalid: + return "ValidationInvalid" + case ValidationValid: + return "ValidationValid" + default: + return fmt.Sprintf("ValidationStatus(%d)", int(v)) + } +} + +// Validate determines whether the contents of the buffer are a valid wire encoding +// of the message type. +// +// This function is exposed for testing. +func Validate(mt pref.MessageType, in piface.UnmarshalInput) (out piface.UnmarshalOutput, _ ValidationStatus) { + mi, ok := mt.(*MessageInfo) + if !ok { + return out, ValidationUnknown + } + if in.Resolver == nil { + in.Resolver = preg.GlobalTypes + } + o, st := mi.validate(in.Buf, 0, unmarshalOptions{ + flags: in.Flags, + resolver: in.Resolver, + }) + if o.initialized { + out.Flags |= piface.UnmarshalInitialized + } + return out, st +} + +type validationInfo struct { + mi *MessageInfo + typ validationType + keyType, valType validationType + + // For non-required fields, requiredBit is 0. + // + // For required fields, requiredBit's nth bit is set, where n is a + // unique index in the range [0, MessageInfo.numRequiredFields). + // + // If there are more than 64 required fields, requiredBit is 0. + requiredBit uint64 +} + +type validationType uint8 + +const ( + validationTypeOther validationType = iota + validationTypeMessage + validationTypeGroup + validationTypeMap + validationTypeRepeatedVarint + validationTypeRepeatedFixed32 + validationTypeRepeatedFixed64 + validationTypeVarint + validationTypeFixed32 + validationTypeFixed64 + validationTypeBytes + validationTypeUTF8String + validationTypeMessageSetItem +) + +func newFieldValidationInfo(mi *MessageInfo, si structInfo, fd pref.FieldDescriptor, ft reflect.Type) validationInfo { + var vi validationInfo + switch { + case fd.ContainingOneof() != nil && !fd.ContainingOneof().IsSynthetic(): + switch fd.Kind() { + case pref.MessageKind: + vi.typ = validationTypeMessage + if ot, ok := si.oneofWrappersByNumber[fd.Number()]; ok { + vi.mi = getMessageInfo(ot.Field(0).Type) + } + case pref.GroupKind: + vi.typ = validationTypeGroup + if ot, ok := si.oneofWrappersByNumber[fd.Number()]; ok { + vi.mi = getMessageInfo(ot.Field(0).Type) + } + case pref.StringKind: + if strs.EnforceUTF8(fd) { + vi.typ = validationTypeUTF8String + } + } + default: + vi = newValidationInfo(fd, ft) + } + if fd.Cardinality() == pref.Required { + // Avoid overflow. The required field check is done with a 64-bit mask, with + // any message containing more than 64 required fields always reported as + // potentially uninitialized, so it is not important to get a precise count + // of the required fields past 64. + if mi.numRequiredFields < math.MaxUint8 { + mi.numRequiredFields++ + vi.requiredBit = 1 << (mi.numRequiredFields - 1) + } + } + return vi +} + +func newValidationInfo(fd pref.FieldDescriptor, ft reflect.Type) validationInfo { + var vi validationInfo + switch { + case fd.IsList(): + switch fd.Kind() { + case pref.MessageKind: + vi.typ = validationTypeMessage + if ft.Kind() == reflect.Slice { + vi.mi = getMessageInfo(ft.Elem()) + } + case pref.GroupKind: + vi.typ = validationTypeGroup + if ft.Kind() == reflect.Slice { + vi.mi = getMessageInfo(ft.Elem()) + } + case pref.StringKind: + vi.typ = validationTypeBytes + if strs.EnforceUTF8(fd) { + vi.typ = validationTypeUTF8String + } + default: + switch wireTypes[fd.Kind()] { + case protowire.VarintType: + vi.typ = validationTypeRepeatedVarint + case protowire.Fixed32Type: + vi.typ = validationTypeRepeatedFixed32 + case protowire.Fixed64Type: + vi.typ = validationTypeRepeatedFixed64 + } + } + case fd.IsMap(): + vi.typ = validationTypeMap + switch fd.MapKey().Kind() { + case pref.StringKind: + if strs.EnforceUTF8(fd) { + vi.keyType = validationTypeUTF8String + } + } + switch fd.MapValue().Kind() { + case pref.MessageKind: + vi.valType = validationTypeMessage + if ft.Kind() == reflect.Map { + vi.mi = getMessageInfo(ft.Elem()) + } + case pref.StringKind: + if strs.EnforceUTF8(fd) { + vi.valType = validationTypeUTF8String + } + } + default: + switch fd.Kind() { + case pref.MessageKind: + vi.typ = validationTypeMessage + if !fd.IsWeak() { + vi.mi = getMessageInfo(ft) + } + case pref.GroupKind: + vi.typ = validationTypeGroup + vi.mi = getMessageInfo(ft) + case pref.StringKind: + vi.typ = validationTypeBytes + if strs.EnforceUTF8(fd) { + vi.typ = validationTypeUTF8String + } + default: + switch wireTypes[fd.Kind()] { + case protowire.VarintType: + vi.typ = validationTypeVarint + case protowire.Fixed32Type: + vi.typ = validationTypeFixed32 + case protowire.Fixed64Type: + vi.typ = validationTypeFixed64 + case protowire.BytesType: + vi.typ = validationTypeBytes + } + } + } + return vi +} + +func (mi *MessageInfo) validate(b []byte, groupTag protowire.Number, opts unmarshalOptions) (out unmarshalOutput, result ValidationStatus) { + mi.init() + type validationState struct { + typ validationType + keyType, valType validationType + endGroup protowire.Number + mi *MessageInfo + tail []byte + requiredMask uint64 + } + + // Pre-allocate some slots to avoid repeated slice reallocation. + states := make([]validationState, 0, 16) + states = append(states, validationState{ + typ: validationTypeMessage, + mi: mi, + }) + if groupTag > 0 { + states[0].typ = validationTypeGroup + states[0].endGroup = groupTag + } + initialized := true + start := len(b) +State: + for len(states) > 0 { + st := &states[len(states)-1] + for len(b) > 0 { + // Parse the tag (field number and wire type). + var tag uint64 + if b[0] < 0x80 { + tag = uint64(b[0]) + b = b[1:] + } else if len(b) >= 2 && b[1] < 128 { + tag = uint64(b[0]&0x7f) + uint64(b[1])<<7 + b = b[2:] + } else { + var n int + tag, n = protowire.ConsumeVarint(b) + if n < 0 { + return out, ValidationInvalid + } + b = b[n:] + } + var num protowire.Number + if n := tag >> 3; n < uint64(protowire.MinValidNumber) || n > uint64(protowire.MaxValidNumber) { + return out, ValidationInvalid + } else { + num = protowire.Number(n) + } + wtyp := protowire.Type(tag & 7) + + if wtyp == protowire.EndGroupType { + if st.endGroup == num { + goto PopState + } + return out, ValidationInvalid + } + var vi validationInfo + switch { + case st.typ == validationTypeMap: + switch num { + case 1: + vi.typ = st.keyType + case 2: + vi.typ = st.valType + vi.mi = st.mi + vi.requiredBit = 1 + } + case flags.ProtoLegacy && st.mi.isMessageSet: + switch num { + case messageset.FieldItem: + vi.typ = validationTypeMessageSetItem + } + default: + var f *coderFieldInfo + if int(num) < len(st.mi.denseCoderFields) { + f = st.mi.denseCoderFields[num] + } else { + f = st.mi.coderFields[num] + } + if f != nil { + vi = f.validation + if vi.typ == validationTypeMessage && vi.mi == nil { + // Probable weak field. + // + // TODO: Consider storing the results of this lookup somewhere + // rather than recomputing it on every validation. + fd := st.mi.Desc.Fields().ByNumber(num) + if fd == nil || !fd.IsWeak() { + break + } + messageName := fd.Message().FullName() + messageType, err := preg.GlobalTypes.FindMessageByName(messageName) + switch err { + case nil: + vi.mi, _ = messageType.(*MessageInfo) + case preg.NotFound: + vi.typ = validationTypeBytes + default: + return out, ValidationUnknown + } + } + break + } + // Possible extension field. + // + // TODO: We should return ValidationUnknown when: + // 1. The resolver is not frozen. (More extensions may be added to it.) + // 2. The resolver returns preg.NotFound. + // In this case, a type added to the resolver in the future could cause + // unmarshaling to begin failing. Supporting this requires some way to + // determine if the resolver is frozen. + xt, err := opts.resolver.FindExtensionByNumber(st.mi.Desc.FullName(), num) + if err != nil && err != preg.NotFound { + return out, ValidationUnknown + } + if err == nil { + vi = getExtensionFieldInfo(xt).validation + } + } + if vi.requiredBit != 0 { + // Check that the field has a compatible wire type. + // We only need to consider non-repeated field types, + // since repeated fields (and maps) can never be required. + ok := false + switch vi.typ { + case validationTypeVarint: + ok = wtyp == protowire.VarintType + case validationTypeFixed32: + ok = wtyp == protowire.Fixed32Type + case validationTypeFixed64: + ok = wtyp == protowire.Fixed64Type + case validationTypeBytes, validationTypeUTF8String, validationTypeMessage: + ok = wtyp == protowire.BytesType + case validationTypeGroup: + ok = wtyp == protowire.StartGroupType + } + if ok { + st.requiredMask |= vi.requiredBit + } + } + + switch wtyp { + case protowire.VarintType: + if len(b) >= 10 { + switch { + case b[0] < 0x80: + b = b[1:] + case b[1] < 0x80: + b = b[2:] + case b[2] < 0x80: + b = b[3:] + case b[3] < 0x80: + b = b[4:] + case b[4] < 0x80: + b = b[5:] + case b[5] < 0x80: + b = b[6:] + case b[6] < 0x80: + b = b[7:] + case b[7] < 0x80: + b = b[8:] + case b[8] < 0x80: + b = b[9:] + case b[9] < 0x80 && b[9] < 2: + b = b[10:] + default: + return out, ValidationInvalid + } + } else { + switch { + case len(b) > 0 && b[0] < 0x80: + b = b[1:] + case len(b) > 1 && b[1] < 0x80: + b = b[2:] + case len(b) > 2 && b[2] < 0x80: + b = b[3:] + case len(b) > 3 && b[3] < 0x80: + b = b[4:] + case len(b) > 4 && b[4] < 0x80: + b = b[5:] + case len(b) > 5 && b[5] < 0x80: + b = b[6:] + case len(b) > 6 && b[6] < 0x80: + b = b[7:] + case len(b) > 7 && b[7] < 0x80: + b = b[8:] + case len(b) > 8 && b[8] < 0x80: + b = b[9:] + case len(b) > 9 && b[9] < 2: + b = b[10:] + default: + return out, ValidationInvalid + } + } + continue State + case protowire.BytesType: + var size uint64 + if len(b) >= 1 && b[0] < 0x80 { + size = uint64(b[0]) + b = b[1:] + } else if len(b) >= 2 && b[1] < 128 { + size = uint64(b[0]&0x7f) + uint64(b[1])<<7 + b = b[2:] + } else { + var n int + size, n = protowire.ConsumeVarint(b) + if n < 0 { + return out, ValidationInvalid + } + b = b[n:] + } + if size > uint64(len(b)) { + return out, ValidationInvalid + } + v := b[:size] + b = b[size:] + switch vi.typ { + case validationTypeMessage: + if vi.mi == nil { + return out, ValidationUnknown + } + vi.mi.init() + fallthrough + case validationTypeMap: + if vi.mi != nil { + vi.mi.init() + } + states = append(states, validationState{ + typ: vi.typ, + keyType: vi.keyType, + valType: vi.valType, + mi: vi.mi, + tail: b, + }) + b = v + continue State + case validationTypeRepeatedVarint: + // Packed field. + for len(v) > 0 { + _, n := protowire.ConsumeVarint(v) + if n < 0 { + return out, ValidationInvalid + } + v = v[n:] + } + case validationTypeRepeatedFixed32: + // Packed field. + if len(v)%4 != 0 { + return out, ValidationInvalid + } + case validationTypeRepeatedFixed64: + // Packed field. + if len(v)%8 != 0 { + return out, ValidationInvalid + } + case validationTypeUTF8String: + if !utf8.Valid(v) { + return out, ValidationInvalid + } + } + case protowire.Fixed32Type: + if len(b) < 4 { + return out, ValidationInvalid + } + b = b[4:] + case protowire.Fixed64Type: + if len(b) < 8 { + return out, ValidationInvalid + } + b = b[8:] + case protowire.StartGroupType: + switch { + case vi.typ == validationTypeGroup: + if vi.mi == nil { + return out, ValidationUnknown + } + vi.mi.init() + states = append(states, validationState{ + typ: validationTypeGroup, + mi: vi.mi, + endGroup: num, + }) + continue State + case flags.ProtoLegacy && vi.typ == validationTypeMessageSetItem: + typeid, v, n, err := messageset.ConsumeFieldValue(b, false) + if err != nil { + return out, ValidationInvalid + } + xt, err := opts.resolver.FindExtensionByNumber(st.mi.Desc.FullName(), typeid) + switch { + case err == preg.NotFound: + b = b[n:] + case err != nil: + return out, ValidationUnknown + default: + xvi := getExtensionFieldInfo(xt).validation + if xvi.mi != nil { + xvi.mi.init() + } + states = append(states, validationState{ + typ: xvi.typ, + mi: xvi.mi, + tail: b[n:], + }) + b = v + continue State + } + default: + n := protowire.ConsumeFieldValue(num, wtyp, b) + if n < 0 { + return out, ValidationInvalid + } + b = b[n:] + } + default: + return out, ValidationInvalid + } + } + if st.endGroup != 0 { + return out, ValidationInvalid + } + if len(b) != 0 { + return out, ValidationInvalid + } + b = st.tail + PopState: + numRequiredFields := 0 + switch st.typ { + case validationTypeMessage, validationTypeGroup: + numRequiredFields = int(st.mi.numRequiredFields) + case validationTypeMap: + // If this is a map field with a message value that contains + // required fields, require that the value be present. + if st.mi != nil && st.mi.numRequiredFields > 0 { + numRequiredFields = 1 + } + } + // If there are more than 64 required fields, this check will + // always fail and we will report that the message is potentially + // uninitialized. + if numRequiredFields > 0 && bits.OnesCount64(st.requiredMask) != numRequiredFields { + initialized = false + } + states = states[:len(states)-1] + } + out.n = start - len(b) + if initialized { + out.initialized = true + } + return out, ValidationValid +} diff --git a/vendor/google.golang.org/protobuf/internal/impl/weak.go b/vendor/google.golang.org/protobuf/internal/impl/weak.go new file mode 100644 index 00000000..009cbefd --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/impl/weak.go @@ -0,0 +1,74 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package impl + +import ( + "fmt" + + pref "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +// weakFields adds methods to the exported WeakFields type for internal use. +// +// The exported type is an alias to an unnamed type, so methods can't be +// defined directly on it. +type weakFields WeakFields + +func (w weakFields) get(num pref.FieldNumber) (pref.ProtoMessage, bool) { + m, ok := w[int32(num)] + return m, ok +} + +func (w *weakFields) set(num pref.FieldNumber, m pref.ProtoMessage) { + if *w == nil { + *w = make(weakFields) + } + (*w)[int32(num)] = m +} + +func (w *weakFields) clear(num pref.FieldNumber) { + delete(*w, int32(num)) +} + +func (Export) HasWeak(w WeakFields, num pref.FieldNumber) bool { + _, ok := w[int32(num)] + return ok +} + +func (Export) ClearWeak(w *WeakFields, num pref.FieldNumber) { + delete(*w, int32(num)) +} + +func (Export) GetWeak(w WeakFields, num pref.FieldNumber, name pref.FullName) pref.ProtoMessage { + if m, ok := w[int32(num)]; ok { + return m + } + mt, _ := protoregistry.GlobalTypes.FindMessageByName(name) + if mt == nil { + panic(fmt.Sprintf("message %v for weak field is not linked in", name)) + } + return mt.Zero().Interface() +} + +func (Export) SetWeak(w *WeakFields, num pref.FieldNumber, name pref.FullName, m pref.ProtoMessage) { + if m != nil { + mt, _ := protoregistry.GlobalTypes.FindMessageByName(name) + if mt == nil { + panic(fmt.Sprintf("message %v for weak field is not linked in", name)) + } + if mt != m.ProtoReflect().Type() { + panic(fmt.Sprintf("invalid message type for weak field: got %T, want %T", m, mt.Zero().Interface())) + } + } + if m == nil || !m.ProtoReflect().IsValid() { + delete(*w, int32(num)) + return + } + if *w == nil { + *w = make(weakFields) + } + (*w)[int32(num)] = m +} diff --git a/vendor/google.golang.org/protobuf/internal/mapsort/mapsort.go b/vendor/google.golang.org/protobuf/internal/mapsort/mapsort.go new file mode 100644 index 00000000..a3de1cf3 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/mapsort/mapsort.go @@ -0,0 +1,43 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package mapsort provides sorted access to maps. +package mapsort + +import ( + "sort" + + "google.golang.org/protobuf/reflect/protoreflect" +) + +// Range iterates over every map entry in sorted key order, +// calling f for each key and value encountered. +func Range(mapv protoreflect.Map, keyKind protoreflect.Kind, f func(protoreflect.MapKey, protoreflect.Value) bool) { + var keys []protoreflect.MapKey + mapv.Range(func(key protoreflect.MapKey, _ protoreflect.Value) bool { + keys = append(keys, key) + return true + }) + sort.Slice(keys, func(i, j int) bool { + switch keyKind { + case protoreflect.BoolKind: + return !keys[i].Bool() && keys[j].Bool() + case protoreflect.Int32Kind, protoreflect.Sint32Kind, protoreflect.Sfixed32Kind, + protoreflect.Int64Kind, protoreflect.Sint64Kind, protoreflect.Sfixed64Kind: + return keys[i].Int() < keys[j].Int() + case protoreflect.Uint32Kind, protoreflect.Fixed32Kind, + protoreflect.Uint64Kind, protoreflect.Fixed64Kind: + return keys[i].Uint() < keys[j].Uint() + case protoreflect.StringKind: + return keys[i].String() < keys[j].String() + default: + panic("invalid kind: " + keyKind.String()) + } + }) + for _, key := range keys { + if !f(key, mapv.Get(key)) { + break + } + } +} diff --git a/vendor/google.golang.org/protobuf/internal/pragma/pragma.go b/vendor/google.golang.org/protobuf/internal/pragma/pragma.go new file mode 100644 index 00000000..49dc4fcd --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/pragma/pragma.go @@ -0,0 +1,29 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package pragma provides types that can be embedded into a struct to +// statically enforce or prevent certain language properties. +package pragma + +import "sync" + +// NoUnkeyedLiterals can be embedded in a struct to prevent unkeyed literals. +type NoUnkeyedLiterals struct{} + +// DoNotImplement can be embedded in an interface to prevent trivial +// implementations of the interface. +// +// This is useful to prevent unauthorized implementations of an interface +// so that it can be extended in the future for any protobuf language changes. +type DoNotImplement interface{ ProtoInternal(DoNotImplement) } + +// DoNotCompare can be embedded in a struct to prevent comparability. +type DoNotCompare [0]func() + +// DoNotCopy can be embedded in a struct to help prevent shallow copies. +// This does not rely on a Go language feature, but rather a special case +// within the vet checker. +// +// See https://golang.org/issues/8005. +type DoNotCopy [0]sync.Mutex diff --git a/vendor/google.golang.org/protobuf/internal/set/ints.go b/vendor/google.golang.org/protobuf/internal/set/ints.go new file mode 100644 index 00000000..d3d7f89a --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/set/ints.go @@ -0,0 +1,58 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package set provides simple set data structures for uint64s. +package set + +import "math/bits" + +// int64s represents a set of integers within the range of 0..63. +type int64s uint64 + +func (bs *int64s) Len() int { + return bits.OnesCount64(uint64(*bs)) +} +func (bs *int64s) Has(n uint64) bool { + return uint64(*bs)&(uint64(1)< 0 +} +func (bs *int64s) Set(n uint64) { + *(*uint64)(bs) |= uint64(1) << n +} +func (bs *int64s) Clear(n uint64) { + *(*uint64)(bs) &^= uint64(1) << n +} + +// Ints represents a set of integers within the range of 0..math.MaxUint64. +type Ints struct { + lo int64s + hi map[uint64]struct{} +} + +func (bs *Ints) Len() int { + return bs.lo.Len() + len(bs.hi) +} +func (bs *Ints) Has(n uint64) bool { + if n < 64 { + return bs.lo.Has(n) + } + _, ok := bs.hi[n] + return ok +} +func (bs *Ints) Set(n uint64) { + if n < 64 { + bs.lo.Set(n) + return + } + if bs.hi == nil { + bs.hi = make(map[uint64]struct{}) + } + bs.hi[n] = struct{}{} +} +func (bs *Ints) Clear(n uint64) { + if n < 64 { + bs.lo.Clear(n) + return + } + delete(bs.hi, n) +} diff --git a/vendor/google.golang.org/protobuf/internal/strs/strings.go b/vendor/google.golang.org/protobuf/internal/strs/strings.go new file mode 100644 index 00000000..0b74e765 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/strs/strings.go @@ -0,0 +1,196 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package strs provides string manipulation functionality specific to protobuf. +package strs + +import ( + "go/token" + "strings" + "unicode" + "unicode/utf8" + + "google.golang.org/protobuf/internal/flags" + "google.golang.org/protobuf/reflect/protoreflect" +) + +// EnforceUTF8 reports whether to enforce strict UTF-8 validation. +func EnforceUTF8(fd protoreflect.FieldDescriptor) bool { + if flags.ProtoLegacy { + if fd, ok := fd.(interface{ EnforceUTF8() bool }); ok { + return fd.EnforceUTF8() + } + } + return fd.Syntax() == protoreflect.Proto3 +} + +// GoCamelCase camel-cases a protobuf name for use as a Go identifier. +// +// If there is an interior underscore followed by a lower case letter, +// drop the underscore and convert the letter to upper case. +func GoCamelCase(s string) string { + // Invariant: if the next letter is lower case, it must be converted + // to upper case. + // That is, we process a word at a time, where words are marked by _ or + // upper case letter. Digits are treated as words. + var b []byte + for i := 0; i < len(s); i++ { + c := s[i] + switch { + case c == '.' && i+1 < len(s) && isASCIILower(s[i+1]): + // Skip over '.' in ".{{lowercase}}". + case c == '.': + b = append(b, '_') // convert '.' to '_' + case c == '_' && (i == 0 || s[i-1] == '.'): + // Convert initial '_' to ensure we start with a capital letter. + // Do the same for '_' after '.' to match historic behavior. + b = append(b, 'X') // convert '_' to 'X' + case c == '_' && i+1 < len(s) && isASCIILower(s[i+1]): + // Skip over '_' in "_{{lowercase}}". + case isASCIIDigit(c): + b = append(b, c) + default: + // Assume we have a letter now - if not, it's a bogus identifier. + // The next word is a sequence of characters that must start upper case. + if isASCIILower(c) { + c -= 'a' - 'A' // convert lowercase to uppercase + } + b = append(b, c) + + // Accept lower case sequence that follows. + for ; i+1 < len(s) && isASCIILower(s[i+1]); i++ { + b = append(b, s[i+1]) + } + } + } + return string(b) +} + +// GoSanitized converts a string to a valid Go identifier. +func GoSanitized(s string) string { + // Sanitize the input to the set of valid characters, + // which must be '_' or be in the Unicode L or N categories. + s = strings.Map(func(r rune) rune { + if unicode.IsLetter(r) || unicode.IsDigit(r) { + return r + } + return '_' + }, s) + + // Prepend '_' in the event of a Go keyword conflict or if + // the identifier is invalid (does not start in the Unicode L category). + r, _ := utf8.DecodeRuneInString(s) + if token.Lookup(s).IsKeyword() || !unicode.IsLetter(r) { + return "_" + s + } + return s +} + +// JSONCamelCase converts a snake_case identifier to a camelCase identifier, +// according to the protobuf JSON specification. +func JSONCamelCase(s string) string { + var b []byte + var wasUnderscore bool + for i := 0; i < len(s); i++ { // proto identifiers are always ASCII + c := s[i] + if c != '_' { + if wasUnderscore && isASCIILower(c) { + c -= 'a' - 'A' // convert to uppercase + } + b = append(b, c) + } + wasUnderscore = c == '_' + } + return string(b) +} + +// JSONSnakeCase converts a camelCase identifier to a snake_case identifier, +// according to the protobuf JSON specification. +func JSONSnakeCase(s string) string { + var b []byte + for i := 0; i < len(s); i++ { // proto identifiers are always ASCII + c := s[i] + if isASCIIUpper(c) { + b = append(b, '_') + c += 'a' - 'A' // convert to lowercase + } + b = append(b, c) + } + return string(b) +} + +// MapEntryName derives the name of the map entry message given the field name. +// See protoc v3.8.0: src/google/protobuf/descriptor.cc:254-276,6057 +func MapEntryName(s string) string { + var b []byte + upperNext := true + for _, c := range s { + switch { + case c == '_': + upperNext = true + case upperNext: + b = append(b, byte(unicode.ToUpper(c))) + upperNext = false + default: + b = append(b, byte(c)) + } + } + b = append(b, "Entry"...) + return string(b) +} + +// EnumValueName derives the camel-cased enum value name. +// See protoc v3.8.0: src/google/protobuf/descriptor.cc:297-313 +func EnumValueName(s string) string { + var b []byte + upperNext := true + for _, c := range s { + switch { + case c == '_': + upperNext = true + case upperNext: + b = append(b, byte(unicode.ToUpper(c))) + upperNext = false + default: + b = append(b, byte(unicode.ToLower(c))) + upperNext = false + } + } + return string(b) +} + +// TrimEnumPrefix trims the enum name prefix from an enum value name, +// where the prefix is all lowercase without underscores. +// See protoc v3.8.0: src/google/protobuf/descriptor.cc:330-375 +func TrimEnumPrefix(s, prefix string) string { + s0 := s // original input + for len(s) > 0 && len(prefix) > 0 { + if s[0] == '_' { + s = s[1:] + continue + } + if unicode.ToLower(rune(s[0])) != rune(prefix[0]) { + return s0 // no prefix match + } + s, prefix = s[1:], prefix[1:] + } + if len(prefix) > 0 { + return s0 // no prefix match + } + s = strings.TrimLeft(s, "_") + if len(s) == 0 { + return s0 // avoid returning empty string + } + return s +} + +func isASCIILower(c byte) bool { + return 'a' <= c && c <= 'z' +} +func isASCIIUpper(c byte) bool { + return 'A' <= c && c <= 'Z' +} +func isASCIIDigit(c byte) bool { + return '0' <= c && c <= '9' +} diff --git a/vendor/google.golang.org/protobuf/internal/strs/strings_pure.go b/vendor/google.golang.org/protobuf/internal/strs/strings_pure.go new file mode 100644 index 00000000..85e074c9 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/strs/strings_pure.go @@ -0,0 +1,27 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build purego appengine + +package strs + +import pref "google.golang.org/protobuf/reflect/protoreflect" + +func UnsafeString(b []byte) string { + return string(b) +} + +func UnsafeBytes(s string) []byte { + return []byte(s) +} + +type Builder struct{} + +func (*Builder) AppendFullName(prefix pref.FullName, name pref.Name) pref.FullName { + return prefix.Append(name) +} + +func (*Builder) MakeString(b []byte) string { + return string(b) +} diff --git a/vendor/google.golang.org/protobuf/internal/strs/strings_unsafe.go b/vendor/google.golang.org/protobuf/internal/strs/strings_unsafe.go new file mode 100644 index 00000000..2160c701 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/strs/strings_unsafe.go @@ -0,0 +1,94 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !purego,!appengine + +package strs + +import ( + "unsafe" + + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +type ( + stringHeader struct { + Data unsafe.Pointer + Len int + } + sliceHeader struct { + Data unsafe.Pointer + Len int + Cap int + } +) + +// UnsafeString returns an unsafe string reference of b. +// The caller must treat the input slice as immutable. +// +// WARNING: Use carefully. The returned result must not leak to the end user +// unless the input slice is provably immutable. +func UnsafeString(b []byte) (s string) { + src := (*sliceHeader)(unsafe.Pointer(&b)) + dst := (*stringHeader)(unsafe.Pointer(&s)) + dst.Data = src.Data + dst.Len = src.Len + return s +} + +// UnsafeBytes returns an unsafe bytes slice reference of s. +// The caller must treat returned slice as immutable. +// +// WARNING: Use carefully. The returned result must not leak to the end user. +func UnsafeBytes(s string) (b []byte) { + src := (*stringHeader)(unsafe.Pointer(&s)) + dst := (*sliceHeader)(unsafe.Pointer(&b)) + dst.Data = src.Data + dst.Len = src.Len + dst.Cap = src.Len + return b +} + +// Builder builds a set of strings with shared lifetime. +// This differs from strings.Builder, which is for building a single string. +type Builder struct { + buf []byte +} + +// AppendFullName is equivalent to protoreflect.FullName.Append, +// but optimized for large batches where each name has a shared lifetime. +func (sb *Builder) AppendFullName(prefix pref.FullName, name pref.Name) pref.FullName { + n := len(prefix) + len(".") + len(name) + if len(prefix) == 0 { + n -= len(".") + } + sb.grow(n) + sb.buf = append(sb.buf, prefix...) + sb.buf = append(sb.buf, '.') + sb.buf = append(sb.buf, name...) + return pref.FullName(sb.last(n)) +} + +// MakeString is equivalent to string(b), but optimized for large batches +// with a shared lifetime. +func (sb *Builder) MakeString(b []byte) string { + sb.grow(len(b)) + sb.buf = append(sb.buf, b...) + return sb.last(len(b)) +} + +func (sb *Builder) grow(n int) { + if cap(sb.buf)-len(sb.buf) >= n { + return + } + + // Unlike strings.Builder, we do not need to copy over the contents + // of the old buffer since our builder provides no API for + // retrieving previously created strings. + sb.buf = make([]byte, 2*(cap(sb.buf)+n)) +} + +func (sb *Builder) last(n int) string { + return UnsafeString(sb.buf[len(sb.buf)-n:]) +} diff --git a/vendor/google.golang.org/protobuf/internal/version/version.go b/vendor/google.golang.org/protobuf/internal/version/version.go new file mode 100644 index 00000000..6b3001c6 --- /dev/null +++ b/vendor/google.golang.org/protobuf/internal/version/version.go @@ -0,0 +1,79 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package version records versioning information about this module. +package version + +import ( + "fmt" + "strings" +) + +// These constants determine the current version of this module. +// +// +// For our release process, we enforce the following rules: +// * Tagged releases use a tag that is identical to String. +// * Tagged releases never reference a commit where the String +// contains "devel". +// * The set of all commits in this repository where String +// does not contain "devel" must have a unique String. +// +// +// Steps for tagging a new release: +// 1. Create a new CL. +// +// 2. Update Minor, Patch, and/or PreRelease as necessary. +// PreRelease must not contain the string "devel". +// +// 3. Since the last released minor version, have there been any changes to +// generator that relies on new functionality in the runtime? +// If yes, then increment RequiredGenerated. +// +// 4. Since the last released minor version, have there been any changes to +// the runtime that removes support for old .pb.go source code? +// If yes, then increment SupportMinimum. +// +// 5. Send out the CL for review and submit it. +// Note that the next CL in step 8 must be submitted after this CL +// without any other CLs in-between. +// +// 6. Tag a new version, where the tag is is the current String. +// +// 7. Write release notes for all notable changes +// between this release and the last release. +// +// 8. Create a new CL. +// +// 9. Update PreRelease to include the string "devel". +// For example: "" -> "devel" or "rc.1" -> "rc.1.devel" +// +// 10. Send out the CL for review and submit it. +const ( + Major = 1 + Minor = 24 + Patch = 0 + PreRelease = "" +) + +// String formats the version string for this module in semver format. +// +// Examples: +// v1.20.1 +// v1.21.0-rc.1 +func String() string { + v := fmt.Sprintf("v%d.%d.%d", Major, Minor, Patch) + if PreRelease != "" { + v += "-" + PreRelease + + // TODO: Add metadata about the commit or build hash. + // See https://golang.org/issue/29814 + // See https://golang.org/issue/33533 + var metadata string + if strings.Contains(PreRelease, "devel") && metadata != "" { + v += "+" + metadata + } + } + return v +} diff --git a/vendor/google.golang.org/protobuf/proto/checkinit.go b/vendor/google.golang.org/protobuf/proto/checkinit.go new file mode 100644 index 00000000..3e9a6a2f --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/checkinit.go @@ -0,0 +1,71 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" +) + +// CheckInitialized returns an error if any required fields in m are not set. +func CheckInitialized(m Message) error { + // Treat a nil message interface as an "untyped" empty message, + // which we assume to have no required fields. + if m == nil { + return nil + } + + return checkInitialized(m.ProtoReflect()) +} + +// CheckInitialized returns an error if any required fields in m are not set. +func checkInitialized(m protoreflect.Message) error { + if methods := protoMethods(m); methods != nil && methods.CheckInitialized != nil { + _, err := methods.CheckInitialized(protoiface.CheckInitializedInput{ + Message: m, + }) + return err + } + return checkInitializedSlow(m) +} + +func checkInitializedSlow(m protoreflect.Message) error { + md := m.Descriptor() + fds := md.Fields() + for i, nums := 0, md.RequiredNumbers(); i < nums.Len(); i++ { + fd := fds.ByNumber(nums.Get(i)) + if !m.Has(fd) { + return errors.RequiredNotSet(string(fd.FullName())) + } + } + var err error + m.Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + switch { + case fd.IsList(): + if fd.Message() == nil { + return true + } + for i, list := 0, v.List(); i < list.Len() && err == nil; i++ { + err = checkInitialized(list.Get(i).Message()) + } + case fd.IsMap(): + if fd.MapValue().Message() == nil { + return true + } + v.Map().Range(func(key protoreflect.MapKey, v protoreflect.Value) bool { + err = checkInitialized(v.Message()) + return err == nil + }) + default: + if fd.Message() == nil { + return true + } + err = checkInitialized(v.Message()) + } + return err == nil + }) + return err +} diff --git a/vendor/google.golang.org/protobuf/proto/decode.go b/vendor/google.golang.org/protobuf/proto/decode.go new file mode 100644 index 00000000..4974b16d --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/decode.go @@ -0,0 +1,273 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/flags" + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" + "google.golang.org/protobuf/runtime/protoiface" +) + +// UnmarshalOptions configures the unmarshaler. +// +// Example usage: +// err := UnmarshalOptions{DiscardUnknown: true}.Unmarshal(b, m) +type UnmarshalOptions struct { + pragma.NoUnkeyedLiterals + + // Merge merges the input into the destination message. + // The default behavior is to always reset the message before unmarshaling, + // unless Merge is specified. + Merge bool + + // AllowPartial accepts input for messages that will result in missing + // required fields. If AllowPartial is false (the default), Unmarshal will + // return an error if there are any missing required fields. + AllowPartial bool + + // If DiscardUnknown is set, unknown fields are ignored. + DiscardUnknown bool + + // Resolver is used for looking up types when unmarshaling extension fields. + // If nil, this defaults to using protoregistry.GlobalTypes. + Resolver interface { + FindExtensionByName(field protoreflect.FullName) (protoreflect.ExtensionType, error) + FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) + } +} + +// Unmarshal parses the wire-format message in b and places the result in m. +func Unmarshal(b []byte, m Message) error { + _, err := UnmarshalOptions{}.unmarshal(b, m.ProtoReflect()) + return err +} + +// Unmarshal parses the wire-format message in b and places the result in m. +func (o UnmarshalOptions) Unmarshal(b []byte, m Message) error { + _, err := o.unmarshal(b, m.ProtoReflect()) + return err +} + +// UnmarshalState parses a wire-format message and places the result in m. +// +// This method permits fine-grained control over the unmarshaler. +// Most users should use Unmarshal instead. +func (o UnmarshalOptions) UnmarshalState(in protoiface.UnmarshalInput) (protoiface.UnmarshalOutput, error) { + return o.unmarshal(in.Buf, in.Message) +} + +// unmarshal is a centralized function that all unmarshal operations go through. +// For profiling purposes, avoid changing the name of this function or +// introducing other code paths for unmarshal that do not go through this. +func (o UnmarshalOptions) unmarshal(b []byte, m protoreflect.Message) (out protoiface.UnmarshalOutput, err error) { + if o.Resolver == nil { + o.Resolver = protoregistry.GlobalTypes + } + if !o.Merge { + Reset(m.Interface()) + } + allowPartial := o.AllowPartial + o.Merge = true + o.AllowPartial = true + methods := protoMethods(m) + if methods != nil && methods.Unmarshal != nil && + !(o.DiscardUnknown && methods.Flags&protoiface.SupportUnmarshalDiscardUnknown == 0) { + in := protoiface.UnmarshalInput{ + Message: m, + Buf: b, + Resolver: o.Resolver, + } + if o.DiscardUnknown { + in.Flags |= protoiface.UnmarshalDiscardUnknown + } + out, err = methods.Unmarshal(in) + } else { + err = o.unmarshalMessageSlow(b, m) + } + if err != nil { + return out, err + } + if allowPartial || (out.Flags&protoiface.UnmarshalInitialized != 0) { + return out, nil + } + return out, checkInitialized(m) +} + +func (o UnmarshalOptions) unmarshalMessage(b []byte, m protoreflect.Message) error { + _, err := o.unmarshal(b, m) + return err +} + +func (o UnmarshalOptions) unmarshalMessageSlow(b []byte, m protoreflect.Message) error { + md := m.Descriptor() + if messageset.IsMessageSet(md) { + return o.unmarshalMessageSet(b, m) + } + fields := md.Fields() + for len(b) > 0 { + // Parse the tag (field number and wire type). + num, wtyp, tagLen := protowire.ConsumeTag(b) + if tagLen < 0 { + return protowire.ParseError(tagLen) + } + if num > protowire.MaxValidNumber { + return errors.New("invalid field number") + } + + // Find the field descriptor for this field number. + fd := fields.ByNumber(num) + if fd == nil && md.ExtensionRanges().Has(num) { + extType, err := o.Resolver.FindExtensionByNumber(md.FullName(), num) + if err != nil && err != protoregistry.NotFound { + return errors.New("%v: unable to resolve extension %v: %v", md.FullName(), num, err) + } + if extType != nil { + fd = extType.TypeDescriptor() + } + } + var err error + if fd == nil { + err = errUnknown + } else if flags.ProtoLegacy { + if fd.IsWeak() && fd.Message().IsPlaceholder() { + err = errUnknown // weak referent is not linked in + } + } + + // Parse the field value. + var valLen int + switch { + case err != nil: + case fd.IsList(): + valLen, err = o.unmarshalList(b[tagLen:], wtyp, m.Mutable(fd).List(), fd) + case fd.IsMap(): + valLen, err = o.unmarshalMap(b[tagLen:], wtyp, m.Mutable(fd).Map(), fd) + default: + valLen, err = o.unmarshalSingular(b[tagLen:], wtyp, m, fd) + } + if err != nil { + if err != errUnknown { + return err + } + valLen = protowire.ConsumeFieldValue(num, wtyp, b[tagLen:]) + if valLen < 0 { + return protowire.ParseError(valLen) + } + if !o.DiscardUnknown { + m.SetUnknown(append(m.GetUnknown(), b[:tagLen+valLen]...)) + } + } + b = b[tagLen+valLen:] + } + return nil +} + +func (o UnmarshalOptions) unmarshalSingular(b []byte, wtyp protowire.Type, m protoreflect.Message, fd protoreflect.FieldDescriptor) (n int, err error) { + v, n, err := o.unmarshalScalar(b, wtyp, fd) + if err != nil { + return 0, err + } + switch fd.Kind() { + case protoreflect.GroupKind, protoreflect.MessageKind: + m2 := m.Mutable(fd).Message() + if err := o.unmarshalMessage(v.Bytes(), m2); err != nil { + return n, err + } + default: + // Non-message scalars replace the previous value. + m.Set(fd, v) + } + return n, nil +} + +func (o UnmarshalOptions) unmarshalMap(b []byte, wtyp protowire.Type, mapv protoreflect.Map, fd protoreflect.FieldDescriptor) (n int, err error) { + if wtyp != protowire.BytesType { + return 0, errUnknown + } + b, n = protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + var ( + keyField = fd.MapKey() + valField = fd.MapValue() + key protoreflect.Value + val protoreflect.Value + haveKey bool + haveVal bool + ) + switch valField.Kind() { + case protoreflect.GroupKind, protoreflect.MessageKind: + val = mapv.NewValue() + } + // Map entries are represented as a two-element message with fields + // containing the key and value. + for len(b) > 0 { + num, wtyp, n := protowire.ConsumeTag(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + if num > protowire.MaxValidNumber { + return 0, errors.New("invalid field number") + } + b = b[n:] + err = errUnknown + switch num { + case 1: + key, n, err = o.unmarshalScalar(b, wtyp, keyField) + if err != nil { + break + } + haveKey = true + case 2: + var v protoreflect.Value + v, n, err = o.unmarshalScalar(b, wtyp, valField) + if err != nil { + break + } + switch valField.Kind() { + case protoreflect.GroupKind, protoreflect.MessageKind: + if err := o.unmarshalMessage(v.Bytes(), val.Message()); err != nil { + return 0, err + } + default: + val = v + } + haveVal = true + } + if err == errUnknown { + n = protowire.ConsumeFieldValue(num, wtyp, b) + if n < 0 { + return 0, protowire.ParseError(n) + } + } else if err != nil { + return 0, err + } + b = b[n:] + } + // Every map entry should have entries for key and value, but this is not strictly required. + if !haveKey { + key = keyField.Default() + } + if !haveVal { + switch valField.Kind() { + case protoreflect.GroupKind, protoreflect.MessageKind: + default: + val = valField.Default() + } + } + mapv.Set(key.MapKey(), val) + return n, nil +} + +// errUnknown is used internally to indicate fields which should be added +// to the unknown field set of a message. It is never returned from an exported +// function. +var errUnknown = errors.New("BUG: internal error (unknown)") diff --git a/vendor/google.golang.org/protobuf/proto/decode_gen.go b/vendor/google.golang.org/protobuf/proto/decode_gen.go new file mode 100644 index 00000000..d6dc904d --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/decode_gen.go @@ -0,0 +1,603 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-types. DO NOT EDIT. + +package proto + +import ( + "math" + "unicode/utf8" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/strs" + "google.golang.org/protobuf/reflect/protoreflect" +) + +// unmarshalScalar decodes a value of the given kind. +// +// Message values are decoded into a []byte which aliases the input data. +func (o UnmarshalOptions) unmarshalScalar(b []byte, wtyp protowire.Type, fd protoreflect.FieldDescriptor) (val protoreflect.Value, n int, err error) { + switch fd.Kind() { + case protoreflect.BoolKind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfBool(protowire.DecodeBool(v)), n, nil + case protoreflect.EnumKind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfEnum(protoreflect.EnumNumber(v)), n, nil + case protoreflect.Int32Kind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfInt32(int32(v)), n, nil + case protoreflect.Sint32Kind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfInt32(int32(protowire.DecodeZigZag(v & math.MaxUint32))), n, nil + case protoreflect.Uint32Kind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfUint32(uint32(v)), n, nil + case protoreflect.Int64Kind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfInt64(int64(v)), n, nil + case protoreflect.Sint64Kind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfInt64(protowire.DecodeZigZag(v)), n, nil + case protoreflect.Uint64Kind: + if wtyp != protowire.VarintType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfUint64(v), n, nil + case protoreflect.Sfixed32Kind: + if wtyp != protowire.Fixed32Type { + return val, 0, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfInt32(int32(v)), n, nil + case protoreflect.Fixed32Kind: + if wtyp != protowire.Fixed32Type { + return val, 0, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfUint32(uint32(v)), n, nil + case protoreflect.FloatKind: + if wtyp != protowire.Fixed32Type { + return val, 0, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfFloat32(math.Float32frombits(uint32(v))), n, nil + case protoreflect.Sfixed64Kind: + if wtyp != protowire.Fixed64Type { + return val, 0, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfInt64(int64(v)), n, nil + case protoreflect.Fixed64Kind: + if wtyp != protowire.Fixed64Type { + return val, 0, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfUint64(v), n, nil + case protoreflect.DoubleKind: + if wtyp != protowire.Fixed64Type { + return val, 0, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfFloat64(math.Float64frombits(v)), n, nil + case protoreflect.StringKind: + if wtyp != protowire.BytesType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + if strs.EnforceUTF8(fd) && !utf8.Valid(v) { + return protoreflect.Value{}, 0, errors.InvalidUTF8(string(fd.FullName())) + } + return protoreflect.ValueOfString(string(v)), n, nil + case protoreflect.BytesKind: + if wtyp != protowire.BytesType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfBytes(append(emptyBuf[:], v...)), n, nil + case protoreflect.MessageKind: + if wtyp != protowire.BytesType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfBytes(v), n, nil + case protoreflect.GroupKind: + if wtyp != protowire.StartGroupType { + return val, 0, errUnknown + } + v, n := protowire.ConsumeGroup(fd.Number(), b) + if n < 0 { + return val, 0, protowire.ParseError(n) + } + return protoreflect.ValueOfBytes(v), n, nil + default: + return val, 0, errUnknown + } +} + +func (o UnmarshalOptions) unmarshalList(b []byte, wtyp protowire.Type, list protoreflect.List, fd protoreflect.FieldDescriptor) (n int, err error) { + switch fd.Kind() { + case protoreflect.BoolKind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfBool(protowire.DecodeBool(v))) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfBool(protowire.DecodeBool(v))) + return n, nil + case protoreflect.EnumKind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfEnum(protoreflect.EnumNumber(v))) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfEnum(protoreflect.EnumNumber(v))) + return n, nil + case protoreflect.Int32Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfInt32(int32(v))) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(v))) + return n, nil + case protoreflect.Sint32Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfInt32(int32(protowire.DecodeZigZag(v & math.MaxUint32)))) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(protowire.DecodeZigZag(v & math.MaxUint32)))) + return n, nil + case protoreflect.Uint32Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfUint32(uint32(v))) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint32(uint32(v))) + return n, nil + case protoreflect.Int64Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfInt64(int64(v))) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(int64(v))) + return n, nil + case protoreflect.Sint64Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfInt64(protowire.DecodeZigZag(v))) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(protowire.DecodeZigZag(v))) + return n, nil + case protoreflect.Uint64Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeVarint(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfUint64(v)) + } + return n, nil + } + if wtyp != protowire.VarintType { + return 0, errUnknown + } + v, n := protowire.ConsumeVarint(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint64(v)) + return n, nil + case protoreflect.Sfixed32Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeFixed32(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfInt32(int32(v))) + } + return n, nil + } + if wtyp != protowire.Fixed32Type { + return 0, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt32(int32(v))) + return n, nil + case protoreflect.Fixed32Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeFixed32(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfUint32(uint32(v))) + } + return n, nil + } + if wtyp != protowire.Fixed32Type { + return 0, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint32(uint32(v))) + return n, nil + case protoreflect.FloatKind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeFixed32(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfFloat32(math.Float32frombits(uint32(v)))) + } + return n, nil + } + if wtyp != protowire.Fixed32Type { + return 0, errUnknown + } + v, n := protowire.ConsumeFixed32(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfFloat32(math.Float32frombits(uint32(v)))) + return n, nil + case protoreflect.Sfixed64Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeFixed64(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfInt64(int64(v))) + } + return n, nil + } + if wtyp != protowire.Fixed64Type { + return 0, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfInt64(int64(v))) + return n, nil + case protoreflect.Fixed64Kind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeFixed64(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfUint64(v)) + } + return n, nil + } + if wtyp != protowire.Fixed64Type { + return 0, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfUint64(v)) + return n, nil + case protoreflect.DoubleKind: + if wtyp == protowire.BytesType { + buf, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + for len(buf) > 0 { + v, n := protowire.ConsumeFixed64(buf) + if n < 0 { + return 0, protowire.ParseError(n) + } + buf = buf[n:] + list.Append(protoreflect.ValueOfFloat64(math.Float64frombits(v))) + } + return n, nil + } + if wtyp != protowire.Fixed64Type { + return 0, errUnknown + } + v, n := protowire.ConsumeFixed64(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfFloat64(math.Float64frombits(v))) + return n, nil + case protoreflect.StringKind: + if wtyp != protowire.BytesType { + return 0, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + if strs.EnforceUTF8(fd) && !utf8.Valid(v) { + return 0, errors.InvalidUTF8(string(fd.FullName())) + } + list.Append(protoreflect.ValueOfString(string(v))) + return n, nil + case protoreflect.BytesKind: + if wtyp != protowire.BytesType { + return 0, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + list.Append(protoreflect.ValueOfBytes(append(emptyBuf[:], v...))) + return n, nil + case protoreflect.MessageKind: + if wtyp != protowire.BytesType { + return 0, errUnknown + } + v, n := protowire.ConsumeBytes(b) + if n < 0 { + return 0, protowire.ParseError(n) + } + m := list.NewElement() + if err := o.unmarshalMessage(v, m.Message()); err != nil { + return 0, err + } + list.Append(m) + return n, nil + case protoreflect.GroupKind: + if wtyp != protowire.StartGroupType { + return 0, errUnknown + } + v, n := protowire.ConsumeGroup(fd.Number(), b) + if n < 0 { + return 0, protowire.ParseError(n) + } + m := list.NewElement() + if err := o.unmarshalMessage(v, m.Message()); err != nil { + return 0, err + } + list.Append(m) + return n, nil + default: + return 0, errUnknown + } +} + +// We append to an empty array rather than a nil []byte to get non-nil zero-length byte slices. +var emptyBuf [0]byte diff --git a/vendor/google.golang.org/protobuf/proto/doc.go b/vendor/google.golang.org/protobuf/proto/doc.go new file mode 100644 index 00000000..c52d8c4a --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/doc.go @@ -0,0 +1,94 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package proto provides functions operating on protocol buffer messages. +// +// For documentation on protocol buffers in general, see: +// +// https://developers.google.com/protocol-buffers +// +// For a tutorial on using protocol buffers with Go, see: +// +// https://developers.google.com/protocol-buffers/docs/gotutorial +// +// For a guide to generated Go protocol buffer code, see: +// +// https://developers.google.com/protocol-buffers/docs/reference/go-generated +// +// +// Binary serialization +// +// This package contains functions to convert to and from the wire format, +// an efficient binary serialization of protocol buffers. +// +// • Size reports the size of a message in the wire format. +// +// • Marshal converts a message to the wire format. +// The MarshalOptions type provides more control over wire marshaling. +// +// • Unmarshal converts a message from the wire format. +// The UnmarshalOptions type provides more control over wire unmarshaling. +// +// +// Basic message operations +// +// • Clone makes a deep copy of a message. +// +// • Merge merges the content of a message into another. +// +// • Equal compares two messages. For more control over comparisons +// and detailed reporting of differences, see package +// "google.golang.org/protobuf/testing/protocmp". +// +// • Reset clears the content of a message. +// +// • CheckInitialized reports whether all required fields in a message are set. +// +// +// Optional scalar constructors +// +// The API for some generated messages represents optional scalar fields +// as pointers to a value. For example, an optional string field has the +// Go type *string. +// +// • Bool, Int32, Int64, Uint32, Uint64, Float32, Float64, and String +// take a value and return a pointer to a new instance of it, +// to simplify construction of optional field values. +// +// Generated enum types usually have an Enum method which performs the +// same operation. +// +// Optional scalar fields are only supported in proto2. +// +// +// Extension accessors +// +// • HasExtension, GetExtension, SetExtension, and ClearExtension +// access extension field values in a protocol buffer message. +// +// Extension fields are only supported in proto2. +// +// +// Related packages +// +// • Package "google.golang.org/protobuf/encoding/protojson" converts messages to +// and from JSON. +// +// • Package "google.golang.org/protobuf/encoding/prototext" converts messages to +// and from the text format. +// +// • Package "google.golang.org/protobuf/reflect/protoreflect" provides a +// reflection interface for protocol buffer data types. +// +// • Package "google.golang.org/protobuf/testing/protocmp" provides features +// to compare protocol buffer messages with the "github.com/google/go-cmp/cmp" +// package. +// +// • Package "google.golang.org/protobuf/types/dynamicpb" provides a dynamic +// message type, suitable for working with messages where the protocol buffer +// type is only known at runtime. +// +// This module contains additional packages for more specialized use cases. +// Consult the individual package documentation for details. +package proto diff --git a/vendor/google.golang.org/protobuf/proto/encode.go b/vendor/google.golang.org/protobuf/proto/encode.go new file mode 100644 index 00000000..7b47a118 --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/encode.go @@ -0,0 +1,346 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "sort" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/fieldsort" + "google.golang.org/protobuf/internal/mapsort" + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" +) + +// MarshalOptions configures the marshaler. +// +// Example usage: +// b, err := MarshalOptions{Deterministic: true}.Marshal(m) +type MarshalOptions struct { + pragma.NoUnkeyedLiterals + + // AllowPartial allows messages that have missing required fields to marshal + // without returning an error. If AllowPartial is false (the default), + // Marshal will return an error if there are any missing required fields. + AllowPartial bool + + // Deterministic controls whether the same message will always be + // serialized to the same bytes within the same binary. + // + // Setting this option guarantees that repeated serialization of + // the same message will return the same bytes, and that different + // processes of the same binary (which may be executing on different + // machines) will serialize equal messages to the same bytes. + // It has no effect on the resulting size of the encoded message compared + // to a non-deterministic marshal. + // + // Note that the deterministic serialization is NOT canonical across + // languages. It is not guaranteed to remain stable over time. It is + // unstable across different builds with schema changes due to unknown + // fields. Users who need canonical serialization (e.g., persistent + // storage in a canonical form, fingerprinting, etc.) must define + // their own canonicalization specification and implement their own + // serializer rather than relying on this API. + // + // If deterministic serialization is requested, map entries will be + // sorted by keys in lexographical order. This is an implementation + // detail and subject to change. + Deterministic bool + + // UseCachedSize indicates that the result of a previous Size call + // may be reused. + // + // Setting this option asserts that: + // + // 1. Size has previously been called on this message with identical + // options (except for UseCachedSize itself). + // + // 2. The message and all its submessages have not changed in any + // way since the Size call. + // + // If either of these invariants is violated, + // the results are undefined and may include panics or corrupted output. + // + // Implementations MAY take this option into account to provide + // better performance, but there is no guarantee that they will do so. + // There is absolutely no guarantee that Size followed by Marshal with + // UseCachedSize set will perform equivalently to Marshal alone. + UseCachedSize bool +} + +// Marshal returns the wire-format encoding of m. +func Marshal(m Message) ([]byte, error) { + // Treat nil message interface as an empty message; nothing to output. + if m == nil { + return nil, nil + } + + out, err := MarshalOptions{}.marshal(nil, m.ProtoReflect()) + if len(out.Buf) == 0 && err == nil { + out.Buf = emptyBytesForMessage(m) + } + return out.Buf, err +} + +// Marshal returns the wire-format encoding of m. +func (o MarshalOptions) Marshal(m Message) ([]byte, error) { + // Treat nil message interface as an empty message; nothing to output. + if m == nil { + return nil, nil + } + + out, err := o.marshal(nil, m.ProtoReflect()) + if len(out.Buf) == 0 && err == nil { + out.Buf = emptyBytesForMessage(m) + } + return out.Buf, err +} + +// emptyBytesForMessage returns a nil buffer if and only if m is invalid, +// otherwise it returns a non-nil empty buffer. +// +// This is to assist the edge-case where user-code does the following: +// m1.OptionalBytes, _ = proto.Marshal(m2) +// where they expect the proto2 "optional_bytes" field to be populated +// if any only if m2 is a valid message. +func emptyBytesForMessage(m Message) []byte { + if m == nil || !m.ProtoReflect().IsValid() { + return nil + } + return emptyBuf[:] +} + +// MarshalAppend appends the wire-format encoding of m to b, +// returning the result. +func (o MarshalOptions) MarshalAppend(b []byte, m Message) ([]byte, error) { + // Treat nil message interface as an empty message; nothing to append. + if m == nil { + return b, nil + } + + out, err := o.marshal(b, m.ProtoReflect()) + return out.Buf, err +} + +// MarshalState returns the wire-format encoding of a message. +// +// This method permits fine-grained control over the marshaler. +// Most users should use Marshal instead. +func (o MarshalOptions) MarshalState(in protoiface.MarshalInput) (protoiface.MarshalOutput, error) { + return o.marshal(in.Buf, in.Message) +} + +// marshal is a centralized function that all marshal operations go through. +// For profiling purposes, avoid changing the name of this function or +// introducing other code paths for marshal that do not go through this. +func (o MarshalOptions) marshal(b []byte, m protoreflect.Message) (out protoiface.MarshalOutput, err error) { + allowPartial := o.AllowPartial + o.AllowPartial = true + if methods := protoMethods(m); methods != nil && methods.Marshal != nil && + !(o.Deterministic && methods.Flags&protoiface.SupportMarshalDeterministic == 0) { + in := protoiface.MarshalInput{ + Message: m, + Buf: b, + } + if o.Deterministic { + in.Flags |= protoiface.MarshalDeterministic + } + if o.UseCachedSize { + in.Flags |= protoiface.MarshalUseCachedSize + } + if methods.Size != nil { + sout := methods.Size(protoiface.SizeInput{ + Message: m, + Flags: in.Flags, + }) + if cap(b) < len(b)+sout.Size { + in.Buf = make([]byte, len(b), growcap(cap(b), len(b)+sout.Size)) + copy(in.Buf, b) + } + in.Flags |= protoiface.MarshalUseCachedSize + } + out, err = methods.Marshal(in) + } else { + out.Buf, err = o.marshalMessageSlow(b, m) + } + if err != nil { + return out, err + } + if allowPartial { + return out, nil + } + return out, checkInitialized(m) +} + +func (o MarshalOptions) marshalMessage(b []byte, m protoreflect.Message) ([]byte, error) { + out, err := o.marshal(b, m) + return out.Buf, err +} + +// growcap scales up the capacity of a slice. +// +// Given a slice with a current capacity of oldcap and a desired +// capacity of wantcap, growcap returns a new capacity >= wantcap. +// +// The algorithm is mostly identical to the one used by append as of Go 1.14. +func growcap(oldcap, wantcap int) (newcap int) { + if wantcap > oldcap*2 { + newcap = wantcap + } else if oldcap < 1024 { + // The Go 1.14 runtime takes this case when len(s) < 1024, + // not when cap(s) < 1024. The difference doesn't seem + // significant here. + newcap = oldcap * 2 + } else { + newcap = oldcap + for 0 < newcap && newcap < wantcap { + newcap += newcap / 4 + } + if newcap <= 0 { + newcap = wantcap + } + } + return newcap +} + +func (o MarshalOptions) marshalMessageSlow(b []byte, m protoreflect.Message) ([]byte, error) { + if messageset.IsMessageSet(m.Descriptor()) { + return o.marshalMessageSet(b, m) + } + // There are many choices for what order we visit fields in. The default one here + // is chosen for reasonable efficiency and simplicity given the protoreflect API. + // It is not deterministic, since Message.Range does not return fields in any + // defined order. + // + // When using deterministic serialization, we sort the known fields. + var err error + o.rangeFields(m, func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + b, err = o.marshalField(b, fd, v) + return err == nil + }) + if err != nil { + return b, err + } + b = append(b, m.GetUnknown()...) + return b, nil +} + +// rangeFields visits fields in a defined order when deterministic serialization is enabled. +func (o MarshalOptions) rangeFields(m protoreflect.Message, f func(protoreflect.FieldDescriptor, protoreflect.Value) bool) { + if !o.Deterministic { + m.Range(f) + return + } + var fds []protoreflect.FieldDescriptor + m.Range(func(fd protoreflect.FieldDescriptor, _ protoreflect.Value) bool { + fds = append(fds, fd) + return true + }) + sort.Slice(fds, func(a, b int) bool { + return fieldsort.Less(fds[a], fds[b]) + }) + for _, fd := range fds { + if !f(fd, m.Get(fd)) { + break + } + } +} + +func (o MarshalOptions) marshalField(b []byte, fd protoreflect.FieldDescriptor, value protoreflect.Value) ([]byte, error) { + switch { + case fd.IsList(): + return o.marshalList(b, fd, value.List()) + case fd.IsMap(): + return o.marshalMap(b, fd, value.Map()) + default: + b = protowire.AppendTag(b, fd.Number(), wireTypes[fd.Kind()]) + return o.marshalSingular(b, fd, value) + } +} + +func (o MarshalOptions) marshalList(b []byte, fd protoreflect.FieldDescriptor, list protoreflect.List) ([]byte, error) { + if fd.IsPacked() && list.Len() > 0 { + b = protowire.AppendTag(b, fd.Number(), protowire.BytesType) + b, pos := appendSpeculativeLength(b) + for i, llen := 0, list.Len(); i < llen; i++ { + var err error + b, err = o.marshalSingular(b, fd, list.Get(i)) + if err != nil { + return b, err + } + } + b = finishSpeculativeLength(b, pos) + return b, nil + } + + kind := fd.Kind() + for i, llen := 0, list.Len(); i < llen; i++ { + var err error + b = protowire.AppendTag(b, fd.Number(), wireTypes[kind]) + b, err = o.marshalSingular(b, fd, list.Get(i)) + if err != nil { + return b, err + } + } + return b, nil +} + +func (o MarshalOptions) marshalMap(b []byte, fd protoreflect.FieldDescriptor, mapv protoreflect.Map) ([]byte, error) { + keyf := fd.MapKey() + valf := fd.MapValue() + var err error + o.rangeMap(mapv, keyf.Kind(), func(key protoreflect.MapKey, value protoreflect.Value) bool { + b = protowire.AppendTag(b, fd.Number(), protowire.BytesType) + var pos int + b, pos = appendSpeculativeLength(b) + + b, err = o.marshalField(b, keyf, key.Value()) + if err != nil { + return false + } + b, err = o.marshalField(b, valf, value) + if err != nil { + return false + } + b = finishSpeculativeLength(b, pos) + return true + }) + return b, err +} + +func (o MarshalOptions) rangeMap(mapv protoreflect.Map, kind protoreflect.Kind, f func(protoreflect.MapKey, protoreflect.Value) bool) { + if !o.Deterministic { + mapv.Range(f) + return + } + mapsort.Range(mapv, kind, f) +} + +// When encoding length-prefixed fields, we speculatively set aside some number of bytes +// for the length, encode the data, and then encode the length (shifting the data if necessary +// to make room). +const speculativeLength = 1 + +func appendSpeculativeLength(b []byte) ([]byte, int) { + pos := len(b) + b = append(b, "\x00\x00\x00\x00"[:speculativeLength]...) + return b, pos +} + +func finishSpeculativeLength(b []byte, pos int) []byte { + mlen := len(b) - pos - speculativeLength + msiz := protowire.SizeVarint(uint64(mlen)) + if msiz != speculativeLength { + for i := 0; i < msiz-speculativeLength; i++ { + b = append(b, 0) + } + copy(b[pos+msiz:], b[pos+speculativeLength:]) + b = b[:pos+msiz+mlen] + } + protowire.AppendVarint(b[:pos], uint64(mlen)) + return b +} diff --git a/vendor/google.golang.org/protobuf/proto/encode_gen.go b/vendor/google.golang.org/protobuf/proto/encode_gen.go new file mode 100644 index 00000000..185dacfb --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/encode_gen.go @@ -0,0 +1,97 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-types. DO NOT EDIT. + +package proto + +import ( + "math" + "unicode/utf8" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/strs" + "google.golang.org/protobuf/reflect/protoreflect" +) + +var wireTypes = map[protoreflect.Kind]protowire.Type{ + protoreflect.BoolKind: protowire.VarintType, + protoreflect.EnumKind: protowire.VarintType, + protoreflect.Int32Kind: protowire.VarintType, + protoreflect.Sint32Kind: protowire.VarintType, + protoreflect.Uint32Kind: protowire.VarintType, + protoreflect.Int64Kind: protowire.VarintType, + protoreflect.Sint64Kind: protowire.VarintType, + protoreflect.Uint64Kind: protowire.VarintType, + protoreflect.Sfixed32Kind: protowire.Fixed32Type, + protoreflect.Fixed32Kind: protowire.Fixed32Type, + protoreflect.FloatKind: protowire.Fixed32Type, + protoreflect.Sfixed64Kind: protowire.Fixed64Type, + protoreflect.Fixed64Kind: protowire.Fixed64Type, + protoreflect.DoubleKind: protowire.Fixed64Type, + protoreflect.StringKind: protowire.BytesType, + protoreflect.BytesKind: protowire.BytesType, + protoreflect.MessageKind: protowire.BytesType, + protoreflect.GroupKind: protowire.StartGroupType, +} + +func (o MarshalOptions) marshalSingular(b []byte, fd protoreflect.FieldDescriptor, v protoreflect.Value) ([]byte, error) { + switch fd.Kind() { + case protoreflect.BoolKind: + b = protowire.AppendVarint(b, protowire.EncodeBool(v.Bool())) + case protoreflect.EnumKind: + b = protowire.AppendVarint(b, uint64(v.Enum())) + case protoreflect.Int32Kind: + b = protowire.AppendVarint(b, uint64(int32(v.Int()))) + case protoreflect.Sint32Kind: + b = protowire.AppendVarint(b, protowire.EncodeZigZag(int64(int32(v.Int())))) + case protoreflect.Uint32Kind: + b = protowire.AppendVarint(b, uint64(uint32(v.Uint()))) + case protoreflect.Int64Kind: + b = protowire.AppendVarint(b, uint64(v.Int())) + case protoreflect.Sint64Kind: + b = protowire.AppendVarint(b, protowire.EncodeZigZag(v.Int())) + case protoreflect.Uint64Kind: + b = protowire.AppendVarint(b, v.Uint()) + case protoreflect.Sfixed32Kind: + b = protowire.AppendFixed32(b, uint32(v.Int())) + case protoreflect.Fixed32Kind: + b = protowire.AppendFixed32(b, uint32(v.Uint())) + case protoreflect.FloatKind: + b = protowire.AppendFixed32(b, math.Float32bits(float32(v.Float()))) + case protoreflect.Sfixed64Kind: + b = protowire.AppendFixed64(b, uint64(v.Int())) + case protoreflect.Fixed64Kind: + b = protowire.AppendFixed64(b, v.Uint()) + case protoreflect.DoubleKind: + b = protowire.AppendFixed64(b, math.Float64bits(v.Float())) + case protoreflect.StringKind: + if strs.EnforceUTF8(fd) && !utf8.ValidString(v.String()) { + return b, errors.InvalidUTF8(string(fd.FullName())) + } + b = protowire.AppendString(b, v.String()) + case protoreflect.BytesKind: + b = protowire.AppendBytes(b, v.Bytes()) + case protoreflect.MessageKind: + var pos int + var err error + b, pos = appendSpeculativeLength(b) + b, err = o.marshalMessage(b, v.Message()) + if err != nil { + return b, err + } + b = finishSpeculativeLength(b, pos) + case protoreflect.GroupKind: + var err error + b, err = o.marshalMessage(b, v.Message()) + if err != nil { + return b, err + } + b = protowire.AppendVarint(b, protowire.EncodeTag(fd.Number(), protowire.EndGroupType)) + default: + return b, errors.New("invalid kind %v", fd.Kind()) + } + return b, nil +} diff --git a/vendor/google.golang.org/protobuf/proto/equal.go b/vendor/google.golang.org/protobuf/proto/equal.go new file mode 100644 index 00000000..10902bd8 --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/equal.go @@ -0,0 +1,154 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "bytes" + "math" + "reflect" + + "google.golang.org/protobuf/encoding/protowire" + pref "google.golang.org/protobuf/reflect/protoreflect" +) + +// Equal reports whether two messages are equal. +// If two messages marshal to the same bytes under deterministic serialization, +// then Equal is guaranteed to report true. +// +// Two messages are equal if they belong to the same message descriptor, +// have the same set of populated known and extension field values, +// and the same set of unknown fields values. If either of the top-level +// messages are invalid, then Equal reports true only if both are invalid. +// +// Scalar values are compared with the equivalent of the == operator in Go, +// except bytes values which are compared using bytes.Equal and +// floating point values which specially treat NaNs as equal. +// Message values are compared by recursively calling Equal. +// Lists are equal if each element value is also equal. +// Maps are equal if they have the same set of keys, where the pair of values +// for each key is also equal. +func Equal(x, y Message) bool { + if x == nil || y == nil { + return x == nil && y == nil + } + mx := x.ProtoReflect() + my := y.ProtoReflect() + if mx.IsValid() != my.IsValid() { + return false + } + return equalMessage(mx, my) +} + +// equalMessage compares two messages. +func equalMessage(mx, my pref.Message) bool { + if mx.Descriptor() != my.Descriptor() { + return false + } + + nx := 0 + equal := true + mx.Range(func(fd pref.FieldDescriptor, vx pref.Value) bool { + nx++ + vy := my.Get(fd) + equal = my.Has(fd) && equalField(fd, vx, vy) + return equal + }) + if !equal { + return false + } + ny := 0 + my.Range(func(fd pref.FieldDescriptor, vx pref.Value) bool { + ny++ + return true + }) + if nx != ny { + return false + } + + return equalUnknown(mx.GetUnknown(), my.GetUnknown()) +} + +// equalField compares two fields. +func equalField(fd pref.FieldDescriptor, x, y pref.Value) bool { + switch { + case fd.IsList(): + return equalList(fd, x.List(), y.List()) + case fd.IsMap(): + return equalMap(fd, x.Map(), y.Map()) + default: + return equalValue(fd, x, y) + } +} + +// equalMap compares two maps. +func equalMap(fd pref.FieldDescriptor, x, y pref.Map) bool { + if x.Len() != y.Len() { + return false + } + equal := true + x.Range(func(k pref.MapKey, vx pref.Value) bool { + vy := y.Get(k) + equal = y.Has(k) && equalValue(fd.MapValue(), vx, vy) + return equal + }) + return equal +} + +// equalList compares two lists. +func equalList(fd pref.FieldDescriptor, x, y pref.List) bool { + if x.Len() != y.Len() { + return false + } + for i := x.Len() - 1; i >= 0; i-- { + if !equalValue(fd, x.Get(i), y.Get(i)) { + return false + } + } + return true +} + +// equalValue compares two singular values. +func equalValue(fd pref.FieldDescriptor, x, y pref.Value) bool { + switch { + case fd.Message() != nil: + return equalMessage(x.Message(), y.Message()) + case fd.Kind() == pref.BytesKind: + return bytes.Equal(x.Bytes(), y.Bytes()) + case fd.Kind() == pref.FloatKind, fd.Kind() == pref.DoubleKind: + fx := x.Float() + fy := y.Float() + if math.IsNaN(fx) || math.IsNaN(fy) { + return math.IsNaN(fx) && math.IsNaN(fy) + } + return fx == fy + default: + return x.Interface() == y.Interface() + } +} + +// equalUnknown compares unknown fields by direct comparison on the raw bytes +// of each individual field number. +func equalUnknown(x, y pref.RawFields) bool { + if len(x) != len(y) { + return false + } + if bytes.Equal([]byte(x), []byte(y)) { + return true + } + + mx := make(map[pref.FieldNumber]pref.RawFields) + my := make(map[pref.FieldNumber]pref.RawFields) + for len(x) > 0 { + fnum, _, n := protowire.ConsumeField(x) + mx[fnum] = append(mx[fnum], x[:n]...) + x = x[n:] + } + for len(y) > 0 { + fnum, _, n := protowire.ConsumeField(y) + my[fnum] = append(my[fnum], y[:n]...) + y = y[n:] + } + return reflect.DeepEqual(mx, my) +} diff --git a/vendor/google.golang.org/protobuf/proto/extension.go b/vendor/google.golang.org/protobuf/proto/extension.go new file mode 100644 index 00000000..5f293cda --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/extension.go @@ -0,0 +1,92 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "google.golang.org/protobuf/reflect/protoreflect" +) + +// HasExtension reports whether an extension field is populated. +// It returns false if m is invalid or if xt does not extend m. +func HasExtension(m Message, xt protoreflect.ExtensionType) bool { + // Treat nil message interface as an empty message; no populated fields. + if m == nil { + return false + } + + // As a special-case, we reports invalid or mismatching descriptors + // as always not being populated (since they aren't). + if xt == nil || m.ProtoReflect().Descriptor() != xt.TypeDescriptor().ContainingMessage() { + return false + } + + return m.ProtoReflect().Has(xt.TypeDescriptor()) +} + +// ClearExtension clears an extension field such that subsequent +// HasExtension calls return false. +// It panics if m is invalid or if xt does not extend m. +func ClearExtension(m Message, xt protoreflect.ExtensionType) { + m.ProtoReflect().Clear(xt.TypeDescriptor()) +} + +// GetExtension retrieves the value for an extension field. +// If the field is unpopulated, it returns the default value for +// scalars and an immutable, empty value for lists or messages. +// It panics if xt does not extend m. +func GetExtension(m Message, xt protoreflect.ExtensionType) interface{} { + // Treat nil message interface as an empty message; return the default. + if m == nil { + return xt.InterfaceOf(xt.Zero()) + } + + return xt.InterfaceOf(m.ProtoReflect().Get(xt.TypeDescriptor())) +} + +// SetExtension stores the value of an extension field. +// It panics if m is invalid, xt does not extend m, or if type of v +// is invalid for the specified extension field. +func SetExtension(m Message, xt protoreflect.ExtensionType, v interface{}) { + xd := xt.TypeDescriptor() + pv := xt.ValueOf(v) + + // Specially treat an invalid list, map, or message as clear. + isValid := true + switch { + case xd.IsList(): + isValid = pv.List().IsValid() + case xd.IsMap(): + isValid = pv.Map().IsValid() + case xd.Message() != nil: + isValid = pv.Message().IsValid() + } + if !isValid { + m.ProtoReflect().Clear(xd) + return + } + + m.ProtoReflect().Set(xd, pv) +} + +// RangeExtensions iterates over every populated extension field in m in an +// undefined order, calling f for each extension type and value encountered. +// It returns immediately if f returns false. +// While iterating, mutating operations may only be performed +// on the current extension field. +func RangeExtensions(m Message, f func(protoreflect.ExtensionType, interface{}) bool) { + // Treat nil message interface as an empty message; nothing to range over. + if m == nil { + return + } + + m.ProtoReflect().Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + if fd.IsExtension() { + xt := fd.(protoreflect.ExtensionTypeDescriptor).Type() + vi := xt.InterfaceOf(v) + return f(xt, vi) + } + return true + }) +} diff --git a/vendor/google.golang.org/protobuf/proto/merge.go b/vendor/google.golang.org/protobuf/proto/merge.go new file mode 100644 index 00000000..d761ab33 --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/merge.go @@ -0,0 +1,139 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "fmt" + + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" +) + +// Merge merges src into dst, which must be a message with the same descriptor. +// +// Populated scalar fields in src are copied to dst, while populated +// singular messages in src are merged into dst by recursively calling Merge. +// The elements of every list field in src is appended to the corresponded +// list fields in dst. The entries of every map field in src is copied into +// the corresponding map field in dst, possibly replacing existing entries. +// The unknown fields of src are appended to the unknown fields of dst. +// +// It is semantically equivalent to unmarshaling the encoded form of src +// into dst with the UnmarshalOptions.Merge option specified. +func Merge(dst, src Message) { + // TODO: Should nil src be treated as semantically equivalent to a + // untyped, read-only, empty message? What about a nil dst? + + dstMsg, srcMsg := dst.ProtoReflect(), src.ProtoReflect() + if dstMsg.Descriptor() != srcMsg.Descriptor() { + if got, want := dstMsg.Descriptor().FullName(), srcMsg.Descriptor().FullName(); got != want { + panic(fmt.Sprintf("descriptor mismatch: %v != %v", got, want)) + } + panic("descriptor mismatch") + } + mergeOptions{}.mergeMessage(dstMsg, srcMsg) +} + +// Clone returns a deep copy of m. +// If the top-level message is invalid, it returns an invalid message as well. +func Clone(m Message) Message { + // NOTE: Most usages of Clone assume the following properties: + // t := reflect.TypeOf(m) + // t == reflect.TypeOf(m.ProtoReflect().New().Interface()) + // t == reflect.TypeOf(m.ProtoReflect().Type().Zero().Interface()) + // + // Embedding protobuf messages breaks this since the parent type will have + // a forwarded ProtoReflect method, but the Interface method will return + // the underlying embedded message type. + if m == nil { + return nil + } + src := m.ProtoReflect() + if !src.IsValid() { + return src.Type().Zero().Interface() + } + dst := src.New() + mergeOptions{}.mergeMessage(dst, src) + return dst.Interface() +} + +// mergeOptions provides a namespace for merge functions, and can be +// exported in the future if we add user-visible merge options. +type mergeOptions struct{} + +func (o mergeOptions) mergeMessage(dst, src protoreflect.Message) { + methods := protoMethods(dst) + if methods != nil && methods.Merge != nil { + in := protoiface.MergeInput{ + Destination: dst, + Source: src, + } + out := methods.Merge(in) + if out.Flags&protoiface.MergeComplete != 0 { + return + } + } + + if !dst.IsValid() { + panic(fmt.Sprintf("cannot merge into invalid %v message", dst.Descriptor().FullName())) + } + + src.Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + switch { + case fd.IsList(): + o.mergeList(dst.Mutable(fd).List(), v.List(), fd) + case fd.IsMap(): + o.mergeMap(dst.Mutable(fd).Map(), v.Map(), fd.MapValue()) + case fd.Message() != nil: + o.mergeMessage(dst.Mutable(fd).Message(), v.Message()) + case fd.Kind() == protoreflect.BytesKind: + dst.Set(fd, o.cloneBytes(v)) + default: + dst.Set(fd, v) + } + return true + }) + + if len(src.GetUnknown()) > 0 { + dst.SetUnknown(append(dst.GetUnknown(), src.GetUnknown()...)) + } +} + +func (o mergeOptions) mergeList(dst, src protoreflect.List, fd protoreflect.FieldDescriptor) { + // Merge semantics appends to the end of the existing list. + for i, n := 0, src.Len(); i < n; i++ { + switch v := src.Get(i); { + case fd.Message() != nil: + dstv := dst.NewElement() + o.mergeMessage(dstv.Message(), v.Message()) + dst.Append(dstv) + case fd.Kind() == protoreflect.BytesKind: + dst.Append(o.cloneBytes(v)) + default: + dst.Append(v) + } + } +} + +func (o mergeOptions) mergeMap(dst, src protoreflect.Map, fd protoreflect.FieldDescriptor) { + // Merge semantics replaces, rather than merges into existing entries. + src.Range(func(k protoreflect.MapKey, v protoreflect.Value) bool { + switch { + case fd.Message() != nil: + dstv := dst.NewValue() + o.mergeMessage(dstv.Message(), v.Message()) + dst.Set(k, dstv) + case fd.Kind() == protoreflect.BytesKind: + dst.Set(k, o.cloneBytes(v)) + default: + dst.Set(k, v) + } + return true + }) +} + +func (o mergeOptions) cloneBytes(v protoreflect.Value) protoreflect.Value { + return protoreflect.ValueOfBytes(append([]byte{}, v.Bytes()...)) +} diff --git a/vendor/google.golang.org/protobuf/proto/messageset.go b/vendor/google.golang.org/protobuf/proto/messageset.go new file mode 100644 index 00000000..1d692c3a --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/messageset.go @@ -0,0 +1,88 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/internal/flags" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" +) + +func (o MarshalOptions) sizeMessageSet(m protoreflect.Message) (size int) { + m.Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + size += messageset.SizeField(fd.Number()) + size += protowire.SizeTag(messageset.FieldMessage) + size += protowire.SizeBytes(o.size(v.Message())) + return true + }) + size += messageset.SizeUnknown(m.GetUnknown()) + return size +} + +func (o MarshalOptions) marshalMessageSet(b []byte, m protoreflect.Message) ([]byte, error) { + if !flags.ProtoLegacy { + return b, errors.New("no support for message_set_wire_format") + } + var err error + o.rangeFields(m, func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + b, err = o.marshalMessageSetField(b, fd, v) + return err == nil + }) + if err != nil { + return b, err + } + return messageset.AppendUnknown(b, m.GetUnknown()) +} + +func (o MarshalOptions) marshalMessageSetField(b []byte, fd protoreflect.FieldDescriptor, value protoreflect.Value) ([]byte, error) { + b = messageset.AppendFieldStart(b, fd.Number()) + b = protowire.AppendTag(b, messageset.FieldMessage, protowire.BytesType) + b = protowire.AppendVarint(b, uint64(o.Size(value.Message().Interface()))) + b, err := o.marshalMessage(b, value.Message()) + if err != nil { + return b, err + } + b = messageset.AppendFieldEnd(b) + return b, nil +} + +func (o UnmarshalOptions) unmarshalMessageSet(b []byte, m protoreflect.Message) error { + if !flags.ProtoLegacy { + return errors.New("no support for message_set_wire_format") + } + return messageset.Unmarshal(b, false, func(num protowire.Number, v []byte) error { + err := o.unmarshalMessageSetField(m, num, v) + if err == errUnknown { + unknown := m.GetUnknown() + unknown = protowire.AppendTag(unknown, num, protowire.BytesType) + unknown = protowire.AppendBytes(unknown, v) + m.SetUnknown(unknown) + return nil + } + return err + }) +} + +func (o UnmarshalOptions) unmarshalMessageSetField(m protoreflect.Message, num protowire.Number, v []byte) error { + md := m.Descriptor() + if !md.ExtensionRanges().Has(num) { + return errUnknown + } + xt, err := o.Resolver.FindExtensionByNumber(md.FullName(), num) + if err == protoregistry.NotFound { + return errUnknown + } + if err != nil { + return errors.New("%v: unable to resolve extension %v: %v", md.FullName(), num, err) + } + xd := xt.TypeDescriptor() + if err := o.unmarshalMessage(v, m.Mutable(xd).Message()); err != nil { + return err + } + return nil +} diff --git a/vendor/google.golang.org/protobuf/proto/proto.go b/vendor/google.golang.org/protobuf/proto/proto.go new file mode 100644 index 00000000..ca14b09c --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/proto.go @@ -0,0 +1,34 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/reflect/protoreflect" +) + +// Message is the top-level interface that all messages must implement. +// It provides access to a reflective view of a message. +// Any implementation of this interface may be used with all functions in the +// protobuf module that accept a Message, except where otherwise specified. +// +// This is the v2 interface definition for protobuf messages. +// The v1 interface definition is "github.com/golang/protobuf/proto".Message. +// +// To convert a v1 message to a v2 message, +// use "github.com/golang/protobuf/proto".MessageV2. +// To convert a v2 message to a v1 message, +// use "github.com/golang/protobuf/proto".MessageV1. +type Message = protoreflect.ProtoMessage + +// Error matches all errors produced by packages in the protobuf module. +// +// That is, errors.Is(err, Error) reports whether an error is produced +// by this module. +var Error error + +func init() { + Error = errors.Error +} diff --git a/vendor/google.golang.org/protobuf/proto/proto_methods.go b/vendor/google.golang.org/protobuf/proto/proto_methods.go new file mode 100644 index 00000000..d8dd604f --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/proto_methods.go @@ -0,0 +1,19 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// The protoreflect build tag disables use of fast-path methods. +// +build !protoreflect + +package proto + +import ( + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" +) + +const hasProtoMethods = true + +func protoMethods(m protoreflect.Message) *protoiface.Methods { + return m.ProtoMethods() +} diff --git a/vendor/google.golang.org/protobuf/proto/proto_reflect.go b/vendor/google.golang.org/protobuf/proto/proto_reflect.go new file mode 100644 index 00000000..b103d432 --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/proto_reflect.go @@ -0,0 +1,19 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// The protoreflect build tag disables use of fast-path methods. +// +build protoreflect + +package proto + +import ( + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" +) + +const hasProtoMethods = false + +func protoMethods(m protoreflect.Message) *protoiface.Methods { + return nil +} diff --git a/vendor/google.golang.org/protobuf/proto/reset.go b/vendor/google.golang.org/protobuf/proto/reset.go new file mode 100644 index 00000000..3d7f8943 --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/reset.go @@ -0,0 +1,43 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "fmt" + + "google.golang.org/protobuf/reflect/protoreflect" +) + +// Reset clears every field in the message. +// The resulting message shares no observable memory with its previous state +// other than the memory for the message itself. +func Reset(m Message) { + if mr, ok := m.(interface{ Reset() }); ok && hasProtoMethods { + mr.Reset() + return + } + resetMessage(m.ProtoReflect()) +} + +func resetMessage(m protoreflect.Message) { + if !m.IsValid() { + panic(fmt.Sprintf("cannot reset invalid %v message", m.Descriptor().FullName())) + } + + // Clear all known fields. + fds := m.Descriptor().Fields() + for i := 0; i < fds.Len(); i++ { + m.Clear(fds.Get(i)) + } + + // Clear extension fields. + m.Range(func(fd protoreflect.FieldDescriptor, _ protoreflect.Value) bool { + m.Clear(fd) + return true + }) + + // Clear unknown fields. + m.SetUnknown(nil) +} diff --git a/vendor/google.golang.org/protobuf/proto/size.go b/vendor/google.golang.org/protobuf/proto/size.go new file mode 100644 index 00000000..554b9c6c --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/size.go @@ -0,0 +1,97 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +import ( + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/encoding/messageset" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/runtime/protoiface" +) + +// Size returns the size in bytes of the wire-format encoding of m. +func Size(m Message) int { + return MarshalOptions{}.Size(m) +} + +// Size returns the size in bytes of the wire-format encoding of m. +func (o MarshalOptions) Size(m Message) int { + // Treat a nil message interface as an empty message; nothing to output. + if m == nil { + return 0 + } + + return o.size(m.ProtoReflect()) +} + +// size is a centralized function that all size operations go through. +// For profiling purposes, avoid changing the name of this function or +// introducing other code paths for size that do not go through this. +func (o MarshalOptions) size(m protoreflect.Message) (size int) { + methods := protoMethods(m) + if methods != nil && methods.Size != nil { + out := methods.Size(protoiface.SizeInput{ + Message: m, + }) + return out.Size + } + if methods != nil && methods.Marshal != nil { + // This is not efficient, but we don't have any choice. + // This case is mainly used for legacy types with a Marshal method. + out, _ := methods.Marshal(protoiface.MarshalInput{ + Message: m, + }) + return len(out.Buf) + } + return o.sizeMessageSlow(m) +} + +func (o MarshalOptions) sizeMessageSlow(m protoreflect.Message) (size int) { + if messageset.IsMessageSet(m.Descriptor()) { + return o.sizeMessageSet(m) + } + m.Range(func(fd protoreflect.FieldDescriptor, v protoreflect.Value) bool { + size += o.sizeField(fd, v) + return true + }) + size += len(m.GetUnknown()) + return size +} + +func (o MarshalOptions) sizeField(fd protoreflect.FieldDescriptor, value protoreflect.Value) (size int) { + num := fd.Number() + switch { + case fd.IsList(): + return o.sizeList(num, fd, value.List()) + case fd.IsMap(): + return o.sizeMap(num, fd, value.Map()) + default: + return protowire.SizeTag(num) + o.sizeSingular(num, fd.Kind(), value) + } +} + +func (o MarshalOptions) sizeList(num protowire.Number, fd protoreflect.FieldDescriptor, list protoreflect.List) (size int) { + if fd.IsPacked() && list.Len() > 0 { + content := 0 + for i, llen := 0, list.Len(); i < llen; i++ { + content += o.sizeSingular(num, fd.Kind(), list.Get(i)) + } + return protowire.SizeTag(num) + protowire.SizeBytes(content) + } + + for i, llen := 0, list.Len(); i < llen; i++ { + size += protowire.SizeTag(num) + o.sizeSingular(num, fd.Kind(), list.Get(i)) + } + return size +} + +func (o MarshalOptions) sizeMap(num protowire.Number, fd protoreflect.FieldDescriptor, mapv protoreflect.Map) (size int) { + mapv.Range(func(key protoreflect.MapKey, value protoreflect.Value) bool { + size += protowire.SizeTag(num) + size += protowire.SizeBytes(o.sizeField(fd.MapKey(), key.Value()) + o.sizeField(fd.MapValue(), value)) + return true + }) + return size +} diff --git a/vendor/google.golang.org/protobuf/proto/size_gen.go b/vendor/google.golang.org/protobuf/proto/size_gen.go new file mode 100644 index 00000000..3cf61a82 --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/size_gen.go @@ -0,0 +1,55 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Code generated by generate-types. DO NOT EDIT. + +package proto + +import ( + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/reflect/protoreflect" +) + +func (o MarshalOptions) sizeSingular(num protowire.Number, kind protoreflect.Kind, v protoreflect.Value) int { + switch kind { + case protoreflect.BoolKind: + return protowire.SizeVarint(protowire.EncodeBool(v.Bool())) + case protoreflect.EnumKind: + return protowire.SizeVarint(uint64(v.Enum())) + case protoreflect.Int32Kind: + return protowire.SizeVarint(uint64(int32(v.Int()))) + case protoreflect.Sint32Kind: + return protowire.SizeVarint(protowire.EncodeZigZag(int64(int32(v.Int())))) + case protoreflect.Uint32Kind: + return protowire.SizeVarint(uint64(uint32(v.Uint()))) + case protoreflect.Int64Kind: + return protowire.SizeVarint(uint64(v.Int())) + case protoreflect.Sint64Kind: + return protowire.SizeVarint(protowire.EncodeZigZag(v.Int())) + case protoreflect.Uint64Kind: + return protowire.SizeVarint(v.Uint()) + case protoreflect.Sfixed32Kind: + return protowire.SizeFixed32() + case protoreflect.Fixed32Kind: + return protowire.SizeFixed32() + case protoreflect.FloatKind: + return protowire.SizeFixed32() + case protoreflect.Sfixed64Kind: + return protowire.SizeFixed64() + case protoreflect.Fixed64Kind: + return protowire.SizeFixed64() + case protoreflect.DoubleKind: + return protowire.SizeFixed64() + case protoreflect.StringKind: + return protowire.SizeBytes(len(v.String())) + case protoreflect.BytesKind: + return protowire.SizeBytes(len(v.Bytes())) + case protoreflect.MessageKind: + return protowire.SizeBytes(o.size(v.Message())) + case protoreflect.GroupKind: + return protowire.SizeGroup(num, o.size(v.Message())) + default: + return 0 + } +} diff --git a/vendor/google.golang.org/protobuf/proto/wrappers.go b/vendor/google.golang.org/protobuf/proto/wrappers.go new file mode 100644 index 00000000..653b12c3 --- /dev/null +++ b/vendor/google.golang.org/protobuf/proto/wrappers.go @@ -0,0 +1,29 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package proto + +// Bool stores v in a new bool value and returns a pointer to it. +func Bool(v bool) *bool { return &v } + +// Int32 stores v in a new int32 value and returns a pointer to it. +func Int32(v int32) *int32 { return &v } + +// Int64 stores v in a new int64 value and returns a pointer to it. +func Int64(v int64) *int64 { return &v } + +// Float32 stores v in a new float32 value and returns a pointer to it. +func Float32(v float32) *float32 { return &v } + +// Float64 stores v in a new float64 value and returns a pointer to it. +func Float64(v float64) *float64 { return &v } + +// Uint32 stores v in a new uint32 value and returns a pointer to it. +func Uint32(v uint32) *uint32 { return &v } + +// Uint64 stores v in a new uint64 value and returns a pointer to it. +func Uint64(v uint64) *uint64 { return &v } + +// String stores v in a new string value and returns a pointer to it. +func String(v string) *string { return &v } diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/methods.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/methods.go new file mode 100644 index 00000000..6be5d16e --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/methods.go @@ -0,0 +1,77 @@ +// Copyright 2020 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package protoreflect + +import ( + "google.golang.org/protobuf/internal/pragma" +) + +// The following types are used by the fast-path Message.ProtoMethods method. +// +// To avoid polluting the public protoreflect API with types used only by +// low-level implementations, the canonical definitions of these types are +// in the runtime/protoiface package. The definitions here and in protoiface +// must be kept in sync. +type ( + methods = struct { + pragma.NoUnkeyedLiterals + Flags supportFlags + Size func(sizeInput) sizeOutput + Marshal func(marshalInput) (marshalOutput, error) + Unmarshal func(unmarshalInput) (unmarshalOutput, error) + Merge func(mergeInput) mergeOutput + CheckInitialized func(checkInitializedInput) (checkInitializedOutput, error) + } + supportFlags = uint64 + sizeInput = struct { + pragma.NoUnkeyedLiterals + Message Message + Flags uint8 + } + sizeOutput = struct { + pragma.NoUnkeyedLiterals + Size int + } + marshalInput = struct { + pragma.NoUnkeyedLiterals + Message Message + Buf []byte + Flags uint8 + } + marshalOutput = struct { + pragma.NoUnkeyedLiterals + Buf []byte + } + unmarshalInput = struct { + pragma.NoUnkeyedLiterals + Message Message + Buf []byte + Flags uint8 + Resolver interface { + FindExtensionByName(field FullName) (ExtensionType, error) + FindExtensionByNumber(message FullName, field FieldNumber) (ExtensionType, error) + } + } + unmarshalOutput = struct { + pragma.NoUnkeyedLiterals + Flags uint8 + } + mergeInput = struct { + pragma.NoUnkeyedLiterals + Source Message + Destination Message + } + mergeOutput = struct { + pragma.NoUnkeyedLiterals + Flags uint8 + } + checkInitializedInput = struct { + pragma.NoUnkeyedLiterals + Message Message + } + checkInitializedOutput = struct { + pragma.NoUnkeyedLiterals + } +) diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/proto.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/proto.go new file mode 100644 index 00000000..b669a4e7 --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/proto.go @@ -0,0 +1,478 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package protoreflect provides interfaces to dynamically manipulate messages. +// +// This package includes type descriptors which describe the structure of types +// defined in proto source files and value interfaces which provide the +// ability to examine and manipulate the contents of messages. +// +// +// Protocol Buffer Descriptors +// +// Protobuf descriptors (e.g., EnumDescriptor or MessageDescriptor) +// are immutable objects that represent protobuf type information. +// They are wrappers around the messages declared in descriptor.proto. +// Protobuf descriptors alone lack any information regarding Go types. +// +// Enums and messages generated by this module implement Enum and ProtoMessage, +// where the Descriptor and ProtoReflect.Descriptor accessors respectively +// return the protobuf descriptor for the values. +// +// The protobuf descriptor interfaces are not meant to be implemented by +// user code since they might need to be extended in the future to support +// additions to the protobuf language. +// The "google.golang.org/protobuf/reflect/protodesc" package converts between +// google.protobuf.DescriptorProto messages and protobuf descriptors. +// +// +// Go Type Descriptors +// +// A type descriptor (e.g., EnumType or MessageType) is a constructor for +// a concrete Go type that represents the associated protobuf descriptor. +// There is commonly a one-to-one relationship between protobuf descriptors and +// Go type descriptors, but it can potentially be a one-to-many relationship. +// +// Enums and messages generated by this module implement Enum and ProtoMessage, +// where the Type and ProtoReflect.Type accessors respectively +// return the protobuf descriptor for the values. +// +// The "google.golang.org/protobuf/types/dynamicpb" package can be used to +// create Go type descriptors from protobuf descriptors. +// +// +// Value Interfaces +// +// The Enum and Message interfaces provide a reflective view over an +// enum or message instance. For enums, it provides the ability to retrieve +// the enum value number for any concrete enum type. For messages, it provides +// the ability to access or manipulate fields of the message. +// +// To convert a proto.Message to a protoreflect.Message, use the +// former's ProtoReflect method. Since the ProtoReflect method is new to the +// v2 message interface, it may not be present on older message implementations. +// The "github.com/golang/protobuf/proto".MessageReflect function can be used +// to obtain a reflective view on older messages. +// +// +// Relationships +// +// The following diagrams demonstrate the relationships between +// various types declared in this package. +// +// +// ┌───────────────────────────────────┐ +// V │ +// ┌────────────── New(n) ─────────────┐ │ +// │ │ │ +// │ ┌──── Descriptor() ──┐ │ ┌── Number() ──┐ │ +// │ │ V V │ V │ +// ╔════════════╗ ╔════════════════╗ ╔════════╗ ╔════════════╗ +// ║ EnumType ║ ║ EnumDescriptor ║ ║ Enum ║ ║ EnumNumber ║ +// ╚════════════╝ ╚════════════════╝ ╚════════╝ ╚════════════╝ +// Λ Λ │ │ +// │ └─── Descriptor() ──┘ │ +// │ │ +// └────────────────── Type() ───────┘ +// +// • An EnumType describes a concrete Go enum type. +// It has an EnumDescriptor and can construct an Enum instance. +// +// • An EnumDescriptor describes an abstract protobuf enum type. +// +// • An Enum is a concrete enum instance. Generated enums implement Enum. +// +// +// ┌──────────────── New() ─────────────────┐ +// │ │ +// │ ┌─── Descriptor() ─────┐ │ ┌── Interface() ───┐ +// │ │ V V │ V +// ╔═════════════╗ ╔═══════════════════╗ ╔═════════╗ ╔══════════════╗ +// ║ MessageType ║ ║ MessageDescriptor ║ ║ Message ║ ║ ProtoMessage ║ +// ╚═════════════╝ ╚═══════════════════╝ ╚═════════╝ ╚══════════════╝ +// Λ Λ │ │ Λ │ +// │ └──── Descriptor() ────┘ │ └─ ProtoReflect() ─┘ +// │ │ +// └─────────────────── Type() ─────────┘ +// +// • A MessageType describes a concrete Go message type. +// It has a MessageDescriptor and can construct a Message instance. +// +// • A MessageDescriptor describes an abstract protobuf message type. +// +// • A Message is a concrete message instance. Generated messages implement +// ProtoMessage, which can convert to/from a Message. +// +// +// ┌── TypeDescriptor() ──┐ ┌───── Descriptor() ─────┐ +// │ V │ V +// ╔═══════════════╗ ╔═════════════════════════╗ ╔═════════════════════╗ +// ║ ExtensionType ║ ║ ExtensionTypeDescriptor ║ ║ ExtensionDescriptor ║ +// ╚═══════════════╝ ╚═════════════════════════╝ ╚═════════════════════╝ +// Λ │ │ Λ │ Λ +// └─────── Type() ───────┘ │ └─── may implement ────┘ │ +// │ │ +// └────── implements ────────┘ +// +// • An ExtensionType describes a concrete Go implementation of an extension. +// It has an ExtensionTypeDescriptor and can convert to/from +// abstract Values and Go values. +// +// • An ExtensionTypeDescriptor is an ExtensionDescriptor +// which also has an ExtensionType. +// +// • An ExtensionDescriptor describes an abstract protobuf extension field and +// may not always be an ExtensionTypeDescriptor. +package protoreflect + +import ( + "fmt" + "regexp" + "strings" + + "google.golang.org/protobuf/encoding/protowire" + "google.golang.org/protobuf/internal/pragma" +) + +type doNotImplement pragma.DoNotImplement + +// ProtoMessage is the top-level interface that all proto messages implement. +// This is declared in the protoreflect package to avoid a cyclic dependency; +// use the proto.Message type instead, which aliases this type. +type ProtoMessage interface{ ProtoReflect() Message } + +// Syntax is the language version of the proto file. +type Syntax syntax + +type syntax int8 // keep exact type opaque as the int type may change + +const ( + Proto2 Syntax = 2 + Proto3 Syntax = 3 +) + +// IsValid reports whether the syntax is valid. +func (s Syntax) IsValid() bool { + switch s { + case Proto2, Proto3: + return true + default: + return false + } +} + +// String returns s as a proto source identifier (e.g., "proto2"). +func (s Syntax) String() string { + switch s { + case Proto2: + return "proto2" + case Proto3: + return "proto3" + default: + return fmt.Sprintf("", s) + } +} + +// GoString returns s as a Go source identifier (e.g., "Proto2"). +func (s Syntax) GoString() string { + switch s { + case Proto2: + return "Proto2" + case Proto3: + return "Proto3" + default: + return fmt.Sprintf("Syntax(%d)", s) + } +} + +// Cardinality determines whether a field is optional, required, or repeated. +type Cardinality cardinality + +type cardinality int8 // keep exact type opaque as the int type may change + +// Constants as defined by the google.protobuf.Cardinality enumeration. +const ( + Optional Cardinality = 1 // appears zero or one times + Required Cardinality = 2 // appears exactly one time; invalid with Proto3 + Repeated Cardinality = 3 // appears zero or more times +) + +// IsValid reports whether the cardinality is valid. +func (c Cardinality) IsValid() bool { + switch c { + case Optional, Required, Repeated: + return true + default: + return false + } +} + +// String returns c as a proto source identifier (e.g., "optional"). +func (c Cardinality) String() string { + switch c { + case Optional: + return "optional" + case Required: + return "required" + case Repeated: + return "repeated" + default: + return fmt.Sprintf("", c) + } +} + +// GoString returns c as a Go source identifier (e.g., "Optional"). +func (c Cardinality) GoString() string { + switch c { + case Optional: + return "Optional" + case Required: + return "Required" + case Repeated: + return "Repeated" + default: + return fmt.Sprintf("Cardinality(%d)", c) + } +} + +// Kind indicates the basic proto kind of a field. +type Kind kind + +type kind int8 // keep exact type opaque as the int type may change + +// Constants as defined by the google.protobuf.Field.Kind enumeration. +const ( + BoolKind Kind = 8 + EnumKind Kind = 14 + Int32Kind Kind = 5 + Sint32Kind Kind = 17 + Uint32Kind Kind = 13 + Int64Kind Kind = 3 + Sint64Kind Kind = 18 + Uint64Kind Kind = 4 + Sfixed32Kind Kind = 15 + Fixed32Kind Kind = 7 + FloatKind Kind = 2 + Sfixed64Kind Kind = 16 + Fixed64Kind Kind = 6 + DoubleKind Kind = 1 + StringKind Kind = 9 + BytesKind Kind = 12 + MessageKind Kind = 11 + GroupKind Kind = 10 +) + +// IsValid reports whether the kind is valid. +func (k Kind) IsValid() bool { + switch k { + case BoolKind, EnumKind, + Int32Kind, Sint32Kind, Uint32Kind, + Int64Kind, Sint64Kind, Uint64Kind, + Sfixed32Kind, Fixed32Kind, FloatKind, + Sfixed64Kind, Fixed64Kind, DoubleKind, + StringKind, BytesKind, MessageKind, GroupKind: + return true + default: + return false + } +} + +// String returns k as a proto source identifier (e.g., "bool"). +func (k Kind) String() string { + switch k { + case BoolKind: + return "bool" + case EnumKind: + return "enum" + case Int32Kind: + return "int32" + case Sint32Kind: + return "sint32" + case Uint32Kind: + return "uint32" + case Int64Kind: + return "int64" + case Sint64Kind: + return "sint64" + case Uint64Kind: + return "uint64" + case Sfixed32Kind: + return "sfixed32" + case Fixed32Kind: + return "fixed32" + case FloatKind: + return "float" + case Sfixed64Kind: + return "sfixed64" + case Fixed64Kind: + return "fixed64" + case DoubleKind: + return "double" + case StringKind: + return "string" + case BytesKind: + return "bytes" + case MessageKind: + return "message" + case GroupKind: + return "group" + default: + return fmt.Sprintf("", k) + } +} + +// GoString returns k as a Go source identifier (e.g., "BoolKind"). +func (k Kind) GoString() string { + switch k { + case BoolKind: + return "BoolKind" + case EnumKind: + return "EnumKind" + case Int32Kind: + return "Int32Kind" + case Sint32Kind: + return "Sint32Kind" + case Uint32Kind: + return "Uint32Kind" + case Int64Kind: + return "Int64Kind" + case Sint64Kind: + return "Sint64Kind" + case Uint64Kind: + return "Uint64Kind" + case Sfixed32Kind: + return "Sfixed32Kind" + case Fixed32Kind: + return "Fixed32Kind" + case FloatKind: + return "FloatKind" + case Sfixed64Kind: + return "Sfixed64Kind" + case Fixed64Kind: + return "Fixed64Kind" + case DoubleKind: + return "DoubleKind" + case StringKind: + return "StringKind" + case BytesKind: + return "BytesKind" + case MessageKind: + return "MessageKind" + case GroupKind: + return "GroupKind" + default: + return fmt.Sprintf("Kind(%d)", k) + } +} + +// FieldNumber is the field number in a message. +type FieldNumber = protowire.Number + +// FieldNumbers represent a list of field numbers. +type FieldNumbers interface { + // Len reports the number of fields in the list. + Len() int + // Get returns the ith field number. It panics if out of bounds. + Get(i int) FieldNumber + // Has reports whether n is within the list of fields. + Has(n FieldNumber) bool + + doNotImplement +} + +// FieldRanges represent a list of field number ranges. +type FieldRanges interface { + // Len reports the number of ranges in the list. + Len() int + // Get returns the ith range. It panics if out of bounds. + Get(i int) [2]FieldNumber // start inclusive; end exclusive + // Has reports whether n is within any of the ranges. + Has(n FieldNumber) bool + + doNotImplement +} + +// EnumNumber is the numeric value for an enum. +type EnumNumber int32 + +// EnumRanges represent a list of enum number ranges. +type EnumRanges interface { + // Len reports the number of ranges in the list. + Len() int + // Get returns the ith range. It panics if out of bounds. + Get(i int) [2]EnumNumber // start inclusive; end inclusive + // Has reports whether n is within any of the ranges. + Has(n EnumNumber) bool + + doNotImplement +} + +var ( + regexName = regexp.MustCompile(`^[_a-zA-Z][_a-zA-Z0-9]*$`) + regexFullName = regexp.MustCompile(`^[_a-zA-Z][_a-zA-Z0-9]*(\.[_a-zA-Z][_a-zA-Z0-9]*)*$`) +) + +// Name is the short name for a proto declaration. This is not the name +// as used in Go source code, which might not be identical to the proto name. +type Name string // e.g., "Kind" + +// IsValid reports whether n is a syntactically valid name. +// An empty name is invalid. +func (n Name) IsValid() bool { + return regexName.MatchString(string(n)) +} + +// Names represent a list of names. +type Names interface { + // Len reports the number of names in the list. + Len() int + // Get returns the ith name. It panics if out of bounds. + Get(i int) Name + // Has reports whether s matches any names in the list. + Has(s Name) bool + + doNotImplement +} + +// FullName is a qualified name that uniquely identifies a proto declaration. +// A qualified name is the concatenation of the proto package along with the +// fully-declared name (i.e., name of parent preceding the name of the child), +// with a '.' delimiter placed between each Name. +// +// This should not have any leading or trailing dots. +type FullName string // e.g., "google.protobuf.Field.Kind" + +// IsValid reports whether n is a syntactically valid full name. +// An empty full name is invalid. +func (n FullName) IsValid() bool { + return regexFullName.MatchString(string(n)) +} + +// Name returns the short name, which is the last identifier segment. +// A single segment FullName is the Name itself. +func (n FullName) Name() Name { + if i := strings.LastIndexByte(string(n), '.'); i >= 0 { + return Name(n[i+1:]) + } + return Name(n) +} + +// Parent returns the full name with the trailing identifier removed. +// A single segment FullName has no parent. +func (n FullName) Parent() FullName { + if i := strings.LastIndexByte(string(n), '.'); i >= 0 { + return n[:i] + } + return "" +} + +// Append returns the qualified name appended with the provided short name. +// +// Invariant: n == n.Parent().Append(n.Name()) // assuming n is valid +func (n FullName) Append(s Name) FullName { + if n == "" { + return FullName(s) + } + return n + "." + FullName(s) +} diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/source.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/source.go new file mode 100644 index 00000000..32ea3d98 --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/source.go @@ -0,0 +1,52 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package protoreflect + +// SourceLocations is a list of source locations. +type SourceLocations interface { + // Len reports the number of source locations in the proto file. + Len() int + // Get returns the ith SourceLocation. It panics if out of bounds. + Get(int) SourceLocation + + doNotImplement + + // TODO: Add ByPath and ByDescriptor helper methods. +} + +// SourceLocation describes a source location and +// corresponds with the google.protobuf.SourceCodeInfo.Location message. +type SourceLocation struct { + // Path is the path to the declaration from the root file descriptor. + // The contents of this slice must not be mutated. + Path SourcePath + + // StartLine and StartColumn are the zero-indexed starting location + // in the source file for the declaration. + StartLine, StartColumn int + // EndLine and EndColumn are the zero-indexed ending location + // in the source file for the declaration. + // In the descriptor.proto, the end line may be omitted if it is identical + // to the start line. Here, it is always populated. + EndLine, EndColumn int + + // LeadingDetachedComments are the leading detached comments + // for the declaration. The contents of this slice must not be mutated. + LeadingDetachedComments []string + // LeadingComments is the leading attached comment for the declaration. + LeadingComments string + // TrailingComments is the trailing attached comment for the declaration. + TrailingComments string +} + +// SourcePath identifies part of a file descriptor for a source location. +// The SourcePath is a sequence of either field numbers or indexes into +// a repeated field that form a path starting from the root file descriptor. +// +// See google.protobuf.SourceCodeInfo.Location.path. +type SourcePath []int32 + +// TODO: Add SourcePath.String method to pretty-print the path. For example: +// ".message_type[6].nested_type[15].field[3]" diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/type.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/type.go new file mode 100644 index 00000000..5be14a72 --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/type.go @@ -0,0 +1,631 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package protoreflect + +// Descriptor provides a set of accessors that are common to every descriptor. +// Each descriptor type wraps the equivalent google.protobuf.XXXDescriptorProto, +// but provides efficient lookup and immutability. +// +// Each descriptor is comparable. Equality implies that the two types are +// exactly identical. However, it is possible for the same semantically +// identical proto type to be represented by multiple type descriptors. +// +// For example, suppose we have t1 and t2 which are both MessageDescriptors. +// If t1 == t2, then the types are definitely equal and all accessors return +// the same information. However, if t1 != t2, then it is still possible that +// they still represent the same proto type (e.g., t1.FullName == t2.FullName). +// This can occur if a descriptor type is created dynamically, or multiple +// versions of the same proto type are accidentally linked into the Go binary. +type Descriptor interface { + // ParentFile returns the parent file descriptor that this descriptor + // is declared within. The parent file for the file descriptor is itself. + // + // Support for this functionality is optional and may return nil. + ParentFile() FileDescriptor + + // Parent returns the parent containing this descriptor declaration. + // The following shows the mapping from child type to possible parent types: + // + // ╔═════════════════════╤═══════════════════════════════════╗ + // ║ Child type │ Possible parent types ║ + // ╠═════════════════════╪═══════════════════════════════════╣ + // ║ FileDescriptor │ nil ║ + // ║ MessageDescriptor │ FileDescriptor, MessageDescriptor ║ + // ║ FieldDescriptor │ FileDescriptor, MessageDescriptor ║ + // ║ OneofDescriptor │ MessageDescriptor ║ + // ║ EnumDescriptor │ FileDescriptor, MessageDescriptor ║ + // ║ EnumValueDescriptor │ EnumDescriptor ║ + // ║ ServiceDescriptor │ FileDescriptor ║ + // ║ MethodDescriptor │ ServiceDescriptor ║ + // ╚═════════════════════╧═══════════════════════════════════╝ + // + // Support for this functionality is optional and may return nil. + Parent() Descriptor + + // Index returns the index of this descriptor within its parent. + // It returns 0 if the descriptor does not have a parent or if the parent + // is unknown. + Index() int + + // Syntax is the protobuf syntax. + Syntax() Syntax // e.g., Proto2 or Proto3 + + // Name is the short name of the declaration (i.e., FullName.Name). + Name() Name // e.g., "Any" + + // FullName is the fully-qualified name of the declaration. + // + // The FullName is a concatenation of the full name of the type that this + // type is declared within and the declaration name. For example, + // field "foo_field" in message "proto.package.MyMessage" is + // uniquely identified as "proto.package.MyMessage.foo_field". + // Enum values are an exception to the rule (see EnumValueDescriptor). + FullName() FullName // e.g., "google.protobuf.Any" + + // IsPlaceholder reports whether type information is missing since a + // dependency is not resolved, in which case only name information is known. + // + // Placeholder types may only be returned by the following accessors + // as a result of unresolved dependencies or weak imports: + // + // ╔═══════════════════════════════════╤═════════════════════╗ + // ║ Accessor │ Descriptor ║ + // ╠═══════════════════════════════════╪═════════════════════╣ + // ║ FileImports.FileDescriptor │ FileDescriptor ║ + // ║ FieldDescriptor.Enum │ EnumDescriptor ║ + // ║ FieldDescriptor.Message │ MessageDescriptor ║ + // ║ FieldDescriptor.DefaultEnumValue │ EnumValueDescriptor ║ + // ║ FieldDescriptor.ContainingMessage │ MessageDescriptor ║ + // ║ MethodDescriptor.Input │ MessageDescriptor ║ + // ║ MethodDescriptor.Output │ MessageDescriptor ║ + // ╚═══════════════════════════════════╧═════════════════════╝ + // + // If true, only Name and FullName are valid. + // For FileDescriptor, the Path is also valid. + IsPlaceholder() bool + + // Options returns the descriptor options. The caller must not modify + // the returned value. + // + // To avoid a dependency cycle, this function returns a proto.Message value. + // The proto message type returned for each descriptor type is as follows: + // ╔═════════════════════╤══════════════════════════════════════════╗ + // ║ Go type │ Protobuf message type ║ + // ╠═════════════════════╪══════════════════════════════════════════╣ + // ║ FileDescriptor │ google.protobuf.FileOptions ║ + // ║ EnumDescriptor │ google.protobuf.EnumOptions ║ + // ║ EnumValueDescriptor │ google.protobuf.EnumValueOptions ║ + // ║ MessageDescriptor │ google.protobuf.MessageOptions ║ + // ║ FieldDescriptor │ google.protobuf.FieldOptions ║ + // ║ OneofDescriptor │ google.protobuf.OneofOptions ║ + // ║ ServiceDescriptor │ google.protobuf.ServiceOptions ║ + // ║ MethodDescriptor │ google.protobuf.MethodOptions ║ + // ╚═════════════════════╧══════════════════════════════════════════╝ + // + // This method returns a typed nil-pointer if no options are present. + // The caller must import the descriptorpb package to use this. + Options() ProtoMessage + + doNotImplement +} + +// FileDescriptor describes the types in a complete proto file and +// corresponds with the google.protobuf.FileDescriptorProto message. +// +// Top-level declarations: +// EnumDescriptor, MessageDescriptor, FieldDescriptor, and/or ServiceDescriptor. +type FileDescriptor interface { + Descriptor // Descriptor.FullName is identical to Package + + // Path returns the file name, relative to the source tree root. + Path() string // e.g., "path/to/file.proto" + // Package returns the protobuf package namespace. + Package() FullName // e.g., "google.protobuf" + + // Imports is a list of imported proto files. + Imports() FileImports + + // Enums is a list of the top-level enum declarations. + Enums() EnumDescriptors + // Messages is a list of the top-level message declarations. + Messages() MessageDescriptors + // Extensions is a list of the top-level extension declarations. + Extensions() ExtensionDescriptors + // Services is a list of the top-level service declarations. + Services() ServiceDescriptors + + // SourceLocations is a list of source locations. + SourceLocations() SourceLocations + + isFileDescriptor +} +type isFileDescriptor interface{ ProtoType(FileDescriptor) } + +// FileImports is a list of file imports. +type FileImports interface { + // Len reports the number of files imported by this proto file. + Len() int + // Get returns the ith FileImport. It panics if out of bounds. + Get(i int) FileImport + + doNotImplement +} + +// FileImport is the declaration for a proto file import. +type FileImport struct { + // FileDescriptor is the file type for the given import. + // It is a placeholder descriptor if IsWeak is set or if a dependency has + // not been regenerated to implement the new reflection APIs. + FileDescriptor + + // IsPublic reports whether this is a public import, which causes this file + // to alias declarations within the imported file. The intended use cases + // for this feature is the ability to move proto files without breaking + // existing dependencies. + // + // The current file and the imported file must be within proto package. + IsPublic bool + + // IsWeak reports whether this is a weak import, which does not impose + // a direct dependency on the target file. + // + // Weak imports are a legacy proto1 feature. Equivalent behavior is + // achieved using proto2 extension fields or proto3 Any messages. + IsWeak bool +} + +// MessageDescriptor describes a message and +// corresponds with the google.protobuf.DescriptorProto message. +// +// Nested declarations: +// FieldDescriptor, OneofDescriptor, FieldDescriptor, EnumDescriptor, +// and/or MessageDescriptor. +type MessageDescriptor interface { + Descriptor + + // IsMapEntry indicates that this is an auto-generated message type to + // represent the entry type for a map field. + // + // Map entry messages have only two fields: + // • a "key" field with a field number of 1 + // • a "value" field with a field number of 2 + // The key and value types are determined by these two fields. + // + // If IsMapEntry is true, it implies that FieldDescriptor.IsMap is true + // for some field with this message type. + IsMapEntry() bool + + // Fields is a list of nested field declarations. + Fields() FieldDescriptors + // Oneofs is a list of nested oneof declarations. + Oneofs() OneofDescriptors + + // ReservedNames is a list of reserved field names. + ReservedNames() Names + // ReservedRanges is a list of reserved ranges of field numbers. + ReservedRanges() FieldRanges + // RequiredNumbers is a list of required field numbers. + // In Proto3, it is always an empty list. + RequiredNumbers() FieldNumbers + // ExtensionRanges is the field ranges used for extension fields. + // In Proto3, it is always an empty ranges. + ExtensionRanges() FieldRanges + // ExtensionRangeOptions returns the ith extension range options. + // + // To avoid a dependency cycle, this method returns a proto.Message value, + // which always contains a google.protobuf.ExtensionRangeOptions message. + // This method returns a typed nil-pointer if no options are present. + // The caller must import the descriptorpb package to use this. + ExtensionRangeOptions(i int) ProtoMessage + + // Enums is a list of nested enum declarations. + Enums() EnumDescriptors + // Messages is a list of nested message declarations. + Messages() MessageDescriptors + // Extensions is a list of nested extension declarations. + Extensions() ExtensionDescriptors + + isMessageDescriptor +} +type isMessageDescriptor interface{ ProtoType(MessageDescriptor) } + +// MessageType encapsulates a MessageDescriptor with a concrete Go implementation. +type MessageType interface { + // New returns a newly allocated empty message. + New() Message + + // Zero returns an empty, read-only message. + Zero() Message + + // Descriptor returns the message descriptor. + // + // Invariant: t.Descriptor() == t.New().Descriptor() + Descriptor() MessageDescriptor +} + +// MessageDescriptors is a list of message declarations. +type MessageDescriptors interface { + // Len reports the number of messages. + Len() int + // Get returns the ith MessageDescriptor. It panics if out of bounds. + Get(i int) MessageDescriptor + // ByName returns the MessageDescriptor for a message named s. + // It returns nil if not found. + ByName(s Name) MessageDescriptor + + doNotImplement +} + +// FieldDescriptor describes a field within a message and +// corresponds with the google.protobuf.FieldDescriptorProto message. +// +// It is used for both normal fields defined within the parent message +// (e.g., MessageDescriptor.Fields) and fields that extend some remote message +// (e.g., FileDescriptor.Extensions or MessageDescriptor.Extensions). +type FieldDescriptor interface { + Descriptor + + // Number reports the unique number for this field. + Number() FieldNumber + // Cardinality reports the cardinality for this field. + Cardinality() Cardinality + // Kind reports the basic kind for this field. + Kind() Kind + + // HasJSONName reports whether this field has an explicitly set JSON name. + HasJSONName() bool + + // JSONName reports the name used for JSON serialization. + // It is usually the camel-cased form of the field name. + JSONName() string + + // HasPresence reports whether the field distinguishes between unpopulated + // and default values. + HasPresence() bool + + // IsExtension reports whether this is an extension field. If false, + // then Parent and ContainingMessage refer to the same message. + // Otherwise, ContainingMessage and Parent likely differ. + IsExtension() bool + + // HasOptionalKeyword reports whether the "optional" keyword was explicitly + // specified in the source .proto file. + HasOptionalKeyword() bool + + // IsWeak reports whether this is a weak field, which does not impose a + // direct dependency on the target type. + // If true, then Message returns a placeholder type. + IsWeak() bool + + // IsPacked reports whether repeated primitive numeric kinds should be + // serialized using a packed encoding. + // If true, then it implies Cardinality is Repeated. + IsPacked() bool + + // IsList reports whether this field represents a list, + // where the value type for the associated field is a List. + // It is equivalent to checking whether Cardinality is Repeated and + // that IsMap reports false. + IsList() bool + + // IsMap reports whether this field represents a map, + // where the value type for the associated field is a Map. + // It is equivalent to checking whether Cardinality is Repeated, + // that the Kind is MessageKind, and that Message.IsMapEntry reports true. + IsMap() bool + + // MapKey returns the field descriptor for the key in the map entry. + // It returns nil if IsMap reports false. + MapKey() FieldDescriptor + + // MapValue returns the field descriptor for the value in the map entry. + // It returns nil if IsMap reports false. + MapValue() FieldDescriptor + + // HasDefault reports whether this field has a default value. + HasDefault() bool + + // Default returns the default value for scalar fields. + // For proto2, it is the default value as specified in the proto file, + // or the zero value if unspecified. + // For proto3, it is always the zero value of the scalar. + // The Value type is determined by the Kind. + Default() Value + + // DefaultEnumValue returns the enum value descriptor for the default value + // of an enum field, and is nil for any other kind of field. + DefaultEnumValue() EnumValueDescriptor + + // ContainingOneof is the containing oneof that this field belongs to, + // and is nil if this field is not part of a oneof. + ContainingOneof() OneofDescriptor + + // ContainingMessage is the containing message that this field belongs to. + // For extension fields, this may not necessarily be the parent message + // that the field is declared within. + ContainingMessage() MessageDescriptor + + // Enum is the enum descriptor if Kind is EnumKind. + // It returns nil for any other Kind. + Enum() EnumDescriptor + + // Message is the message descriptor if Kind is + // MessageKind or GroupKind. It returns nil for any other Kind. + Message() MessageDescriptor + + isFieldDescriptor +} +type isFieldDescriptor interface{ ProtoType(FieldDescriptor) } + +// FieldDescriptors is a list of field declarations. +type FieldDescriptors interface { + // Len reports the number of fields. + Len() int + // Get returns the ith FieldDescriptor. It panics if out of bounds. + Get(i int) FieldDescriptor + // ByName returns the FieldDescriptor for a field named s. + // It returns nil if not found. + ByName(s Name) FieldDescriptor + // ByJSONName returns the FieldDescriptor for a field with s as the JSON name. + // It returns nil if not found. + ByJSONName(s string) FieldDescriptor + // ByNumber returns the FieldDescriptor for a field numbered n. + // It returns nil if not found. + ByNumber(n FieldNumber) FieldDescriptor + + doNotImplement +} + +// OneofDescriptor describes a oneof field set within a given message and +// corresponds with the google.protobuf.OneofDescriptorProto message. +type OneofDescriptor interface { + Descriptor + + // IsSynthetic reports whether this is a synthetic oneof created to support + // proto3 optional semantics. If true, Fields contains exactly one field + // with HasOptionalKeyword specified. + IsSynthetic() bool + + // Fields is a list of fields belonging to this oneof. + Fields() FieldDescriptors + + isOneofDescriptor +} +type isOneofDescriptor interface{ ProtoType(OneofDescriptor) } + +// OneofDescriptors is a list of oneof declarations. +type OneofDescriptors interface { + // Len reports the number of oneof fields. + Len() int + // Get returns the ith OneofDescriptor. It panics if out of bounds. + Get(i int) OneofDescriptor + // ByName returns the OneofDescriptor for a oneof named s. + // It returns nil if not found. + ByName(s Name) OneofDescriptor + + doNotImplement +} + +// ExtensionDescriptor is an alias of FieldDescriptor for documentation. +type ExtensionDescriptor = FieldDescriptor + +// ExtensionTypeDescriptor is an ExtensionDescriptor with an associated ExtensionType. +type ExtensionTypeDescriptor interface { + ExtensionDescriptor + + // Type returns the associated ExtensionType. + Type() ExtensionType + + // Descriptor returns the plain ExtensionDescriptor without the + // associated ExtensionType. + Descriptor() ExtensionDescriptor +} + +// ExtensionDescriptors is a list of field declarations. +type ExtensionDescriptors interface { + // Len reports the number of fields. + Len() int + // Get returns the ith ExtensionDescriptor. It panics if out of bounds. + Get(i int) ExtensionDescriptor + // ByName returns the ExtensionDescriptor for a field named s. + // It returns nil if not found. + ByName(s Name) ExtensionDescriptor + + doNotImplement +} + +// ExtensionType encapsulates an ExtensionDescriptor with a concrete +// Go implementation. The nested field descriptor must be for a extension field. +// +// While a normal field is a member of the parent message that it is declared +// within (see Descriptor.Parent), an extension field is a member of some other +// target message (see ExtensionDescriptor.Extendee) and may have no +// relationship with the parent. However, the full name of an extension field is +// relative to the parent that it is declared within. +// +// For example: +// syntax = "proto2"; +// package example; +// message FooMessage { +// extensions 100 to max; +// } +// message BarMessage { +// extends FooMessage { optional BarMessage bar_field = 100; } +// } +// +// Field "bar_field" is an extension of FooMessage, but its full name is +// "example.BarMessage.bar_field" instead of "example.FooMessage.bar_field". +type ExtensionType interface { + // New returns a new value for the field. + // For scalars, this returns the default value in native Go form. + New() Value + + // Zero returns a new value for the field. + // For scalars, this returns the default value in native Go form. + // For composite types, this returns an empty, read-only message, list, or map. + Zero() Value + + // TypeDescriptor returns the extension type descriptor. + TypeDescriptor() ExtensionTypeDescriptor + + // ValueOf wraps the input and returns it as a Value. + // ValueOf panics if the input value is invalid or not the appropriate type. + // + // ValueOf is more extensive than protoreflect.ValueOf for a given field's + // value as it has more type information available. + ValueOf(interface{}) Value + + // InterfaceOf completely unwraps the Value to the underlying Go type. + // InterfaceOf panics if the input is nil or does not represent the + // appropriate underlying Go type. For composite types, it panics if the + // value is not mutable. + // + // InterfaceOf is able to unwrap the Value further than Value.Interface + // as it has more type information available. + InterfaceOf(Value) interface{} + + // IsValidValue reports whether the Value is valid to assign to the field. + IsValidValue(Value) bool + + // IsValidInterface reports whether the input is valid to assign to the field. + IsValidInterface(interface{}) bool +} + +// EnumDescriptor describes an enum and +// corresponds with the google.protobuf.EnumDescriptorProto message. +// +// Nested declarations: +// EnumValueDescriptor. +type EnumDescriptor interface { + Descriptor + + // Values is a list of nested enum value declarations. + Values() EnumValueDescriptors + + // ReservedNames is a list of reserved enum names. + ReservedNames() Names + // ReservedRanges is a list of reserved ranges of enum numbers. + ReservedRanges() EnumRanges + + isEnumDescriptor +} +type isEnumDescriptor interface{ ProtoType(EnumDescriptor) } + +// EnumType encapsulates an EnumDescriptor with a concrete Go implementation. +type EnumType interface { + // New returns an instance of this enum type with its value set to n. + New(n EnumNumber) Enum + + // Descriptor returns the enum descriptor. + // + // Invariant: t.Descriptor() == t.New(0).Descriptor() + Descriptor() EnumDescriptor +} + +// EnumDescriptors is a list of enum declarations. +type EnumDescriptors interface { + // Len reports the number of enum types. + Len() int + // Get returns the ith EnumDescriptor. It panics if out of bounds. + Get(i int) EnumDescriptor + // ByName returns the EnumDescriptor for an enum named s. + // It returns nil if not found. + ByName(s Name) EnumDescriptor + + doNotImplement +} + +// EnumValueDescriptor describes an enum value and +// corresponds with the google.protobuf.EnumValueDescriptorProto message. +// +// All other proto declarations are in the namespace of the parent. +// However, enum values do not follow this rule and are within the namespace +// of the parent's parent (i.e., they are a sibling of the containing enum). +// Thus, a value named "FOO_VALUE" declared within an enum uniquely identified +// as "proto.package.MyEnum" has a full name of "proto.package.FOO_VALUE". +type EnumValueDescriptor interface { + Descriptor + + // Number returns the enum value as an integer. + Number() EnumNumber + + isEnumValueDescriptor +} +type isEnumValueDescriptor interface{ ProtoType(EnumValueDescriptor) } + +// EnumValueDescriptors is a list of enum value declarations. +type EnumValueDescriptors interface { + // Len reports the number of enum values. + Len() int + // Get returns the ith EnumValueDescriptor. It panics if out of bounds. + Get(i int) EnumValueDescriptor + // ByName returns the EnumValueDescriptor for the enum value named s. + // It returns nil if not found. + ByName(s Name) EnumValueDescriptor + // ByNumber returns the EnumValueDescriptor for the enum value numbered n. + // If multiple have the same number, the first one defined is returned + // It returns nil if not found. + ByNumber(n EnumNumber) EnumValueDescriptor + + doNotImplement +} + +// ServiceDescriptor describes a service and +// corresponds with the google.protobuf.ServiceDescriptorProto message. +// +// Nested declarations: MethodDescriptor. +type ServiceDescriptor interface { + Descriptor + + // Methods is a list of nested message declarations. + Methods() MethodDescriptors + + isServiceDescriptor +} +type isServiceDescriptor interface{ ProtoType(ServiceDescriptor) } + +// ServiceDescriptors is a list of service declarations. +type ServiceDescriptors interface { + // Len reports the number of services. + Len() int + // Get returns the ith ServiceDescriptor. It panics if out of bounds. + Get(i int) ServiceDescriptor + // ByName returns the ServiceDescriptor for a service named s. + // It returns nil if not found. + ByName(s Name) ServiceDescriptor + + doNotImplement +} + +// MethodDescriptor describes a method and +// corresponds with the google.protobuf.MethodDescriptorProto message. +type MethodDescriptor interface { + Descriptor + + // Input is the input message descriptor. + Input() MessageDescriptor + // Output is the output message descriptor. + Output() MessageDescriptor + // IsStreamingClient reports whether the client streams multiple messages. + IsStreamingClient() bool + // IsStreamingServer reports whether the server streams multiple messages. + IsStreamingServer() bool + + isMethodDescriptor +} +type isMethodDescriptor interface{ ProtoType(MethodDescriptor) } + +// MethodDescriptors is a list of method declarations. +type MethodDescriptors interface { + // Len reports the number of methods. + Len() int + // Get returns the ith MethodDescriptor. It panics if out of bounds. + Get(i int) MethodDescriptor + // ByName returns the MethodDescriptor for a service method named s. + // It returns nil if not found. + ByName(s Name) MethodDescriptor + + doNotImplement +} diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/value.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/value.go new file mode 100644 index 00000000..f3198107 --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/value.go @@ -0,0 +1,285 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package protoreflect + +import "google.golang.org/protobuf/encoding/protowire" + +// Enum is a reflection interface for a concrete enum value, +// which provides type information and a getter for the enum number. +// Enum does not provide a mutable API since enums are commonly backed by +// Go constants, which are not addressable. +type Enum interface { + // Descriptor returns enum descriptor, which contains only the protobuf + // type information for the enum. + Descriptor() EnumDescriptor + + // Type returns the enum type, which encapsulates both Go and protobuf + // type information. If the Go type information is not needed, + // it is recommended that the enum descriptor be used instead. + Type() EnumType + + // Number returns the enum value as an integer. + Number() EnumNumber +} + +// Message is a reflective interface for a concrete message value, +// encapsulating both type and value information for the message. +// +// Accessor/mutators for individual fields are keyed by FieldDescriptor. +// For non-extension fields, the descriptor must exactly match the +// field known by the parent message. +// For extension fields, the descriptor must implement ExtensionTypeDescriptor, +// extend the parent message (i.e., have the same message FullName), and +// be within the parent's extension range. +// +// Each field Value can be a scalar or a composite type (Message, List, or Map). +// See Value for the Go types associated with a FieldDescriptor. +// Providing a Value that is invalid or of an incorrect type panics. +type Message interface { + // Descriptor returns message descriptor, which contains only the protobuf + // type information for the message. + Descriptor() MessageDescriptor + + // Type returns the message type, which encapsulates both Go and protobuf + // type information. If the Go type information is not needed, + // it is recommended that the message descriptor be used instead. + Type() MessageType + + // New returns a newly allocated and mutable empty message. + New() Message + + // Interface unwraps the message reflection interface and + // returns the underlying ProtoMessage interface. + Interface() ProtoMessage + + // Range iterates over every populated field in an undefined order, + // calling f for each field descriptor and value encountered. + // Range returns immediately if f returns false. + // While iterating, mutating operations may only be performed + // on the current field descriptor. + Range(f func(FieldDescriptor, Value) bool) + + // Has reports whether a field is populated. + // + // Some fields have the property of nullability where it is possible to + // distinguish between the default value of a field and whether the field + // was explicitly populated with the default value. Singular message fields, + // member fields of a oneof, and proto2 scalar fields are nullable. Such + // fields are populated only if explicitly set. + // + // In other cases (aside from the nullable cases above), + // a proto3 scalar field is populated if it contains a non-zero value, and + // a repeated field is populated if it is non-empty. + Has(FieldDescriptor) bool + + // Clear clears the field such that a subsequent Has call reports false. + // + // Clearing an extension field clears both the extension type and value + // associated with the given field number. + // + // Clear is a mutating operation and unsafe for concurrent use. + Clear(FieldDescriptor) + + // Get retrieves the value for a field. + // + // For unpopulated scalars, it returns the default value, where + // the default value of a bytes scalar is guaranteed to be a copy. + // For unpopulated composite types, it returns an empty, read-only view + // of the value; to obtain a mutable reference, use Mutable. + Get(FieldDescriptor) Value + + // Set stores the value for a field. + // + // For a field belonging to a oneof, it implicitly clears any other field + // that may be currently set within the same oneof. + // For extension fields, it implicitly stores the provided ExtensionType. + // When setting a composite type, it is unspecified whether the stored value + // aliases the source's memory in any way. If the composite value is an + // empty, read-only value, then it panics. + // + // Set is a mutating operation and unsafe for concurrent use. + Set(FieldDescriptor, Value) + + // Mutable returns a mutable reference to a composite type. + // + // If the field is unpopulated, it may allocate a composite value. + // For a field belonging to a oneof, it implicitly clears any other field + // that may be currently set within the same oneof. + // For extension fields, it implicitly stores the provided ExtensionType + // if not already stored. + // It panics if the field does not contain a composite type. + // + // Mutable is a mutating operation and unsafe for concurrent use. + Mutable(FieldDescriptor) Value + + // NewField returns a new value that is assignable to the field + // for the given descriptor. For scalars, this returns the default value. + // For lists, maps, and messages, this returns a new, empty, mutable value. + NewField(FieldDescriptor) Value + + // WhichOneof reports which field within the oneof is populated, + // returning nil if none are populated. + // It panics if the oneof descriptor does not belong to this message. + WhichOneof(OneofDescriptor) FieldDescriptor + + // GetUnknown retrieves the entire list of unknown fields. + // The caller may only mutate the contents of the RawFields + // if the mutated bytes are stored back into the message with SetUnknown. + GetUnknown() RawFields + + // SetUnknown stores an entire list of unknown fields. + // The raw fields must be syntactically valid according to the wire format. + // An implementation may panic if this is not the case. + // Once stored, the caller must not mutate the content of the RawFields. + // An empty RawFields may be passed to clear the fields. + // + // SetUnknown is a mutating operation and unsafe for concurrent use. + SetUnknown(RawFields) + + // IsValid reports whether the message is valid. + // + // An invalid message is an empty, read-only value. + // + // An invalid message often corresponds to a nil pointer of the concrete + // message type, but the details are implementation dependent. + // Validity is not part of the protobuf data model, and may not + // be preserved in marshaling or other operations. + IsValid() bool + + // ProtoMethods returns optional fast-path implementions of various operations. + // This method may return nil. + // + // The returned methods type is identical to + // "google.golang.org/protobuf/runtime/protoiface".Methods. + // Consult the protoiface package documentation for details. + ProtoMethods() *methods +} + +// RawFields is the raw bytes for an ordered sequence of fields. +// Each field contains both the tag (representing field number and wire type), +// and also the wire data itself. +type RawFields []byte + +// IsValid reports whether b is syntactically correct wire format. +func (b RawFields) IsValid() bool { + for len(b) > 0 { + _, _, n := protowire.ConsumeField(b) + if n < 0 { + return false + } + b = b[n:] + } + return true +} + +// List is a zero-indexed, ordered list. +// The element Value type is determined by FieldDescriptor.Kind. +// Providing a Value that is invalid or of an incorrect type panics. +type List interface { + // Len reports the number of entries in the List. + // Get, Set, and Truncate panic with out of bound indexes. + Len() int + + // Get retrieves the value at the given index. + // It never returns an invalid value. + Get(int) Value + + // Set stores a value for the given index. + // When setting a composite type, it is unspecified whether the set + // value aliases the source's memory in any way. + // + // Set is a mutating operation and unsafe for concurrent use. + Set(int, Value) + + // Append appends the provided value to the end of the list. + // When appending a composite type, it is unspecified whether the appended + // value aliases the source's memory in any way. + // + // Append is a mutating operation and unsafe for concurrent use. + Append(Value) + + // AppendMutable appends a new, empty, mutable message value to the end + // of the list and returns it. + // It panics if the list does not contain a message type. + AppendMutable() Value + + // Truncate truncates the list to a smaller length. + // + // Truncate is a mutating operation and unsafe for concurrent use. + Truncate(int) + + // NewElement returns a new value for a list element. + // For enums, this returns the first enum value. + // For other scalars, this returns the zero value. + // For messages, this returns a new, empty, mutable value. + NewElement() Value + + // IsValid reports whether the list is valid. + // + // An invalid list is an empty, read-only value. + // + // Validity is not part of the protobuf data model, and may not + // be preserved in marshaling or other operations. + IsValid() bool +} + +// Map is an unordered, associative map. +// The entry MapKey type is determined by FieldDescriptor.MapKey.Kind. +// The entry Value type is determined by FieldDescriptor.MapValue.Kind. +// Providing a MapKey or Value that is invalid or of an incorrect type panics. +type Map interface { + // Len reports the number of elements in the map. + Len() int + + // Range iterates over every map entry in an undefined order, + // calling f for each key and value encountered. + // Range calls f Len times unless f returns false, which stops iteration. + // While iterating, mutating operations may only be performed + // on the current map key. + Range(f func(MapKey, Value) bool) + + // Has reports whether an entry with the given key is in the map. + Has(MapKey) bool + + // Clear clears the entry associated with they given key. + // The operation does nothing if there is no entry associated with the key. + // + // Clear is a mutating operation and unsafe for concurrent use. + Clear(MapKey) + + // Get retrieves the value for an entry with the given key. + // It returns an invalid value for non-existent entries. + Get(MapKey) Value + + // Set stores the value for an entry with the given key. + // It panics when given a key or value that is invalid or the wrong type. + // When setting a composite type, it is unspecified whether the set + // value aliases the source's memory in any way. + // + // Set is a mutating operation and unsafe for concurrent use. + Set(MapKey, Value) + + // Mutable retrieves a mutable reference to the entry for the given key. + // If no entry exists for the key, it creates a new, empty, mutable value + // and stores it as the entry for the key. + // It panics if the map value is not a message. + Mutable(MapKey) Value + + // NewValue returns a new value assignable as a map value. + // For enums, this returns the first enum value. + // For other scalars, this returns the zero value. + // For messages, this returns a new, empty, mutable value. + NewValue() Value + + // IsValid reports whether the map is valid. + // + // An invalid map is an empty, read-only value. + // + // An invalid message often corresponds to a nil Go map value, + // but the details are implementation dependent. + // Validity is not part of the protobuf data model, and may not + // be preserved in marshaling or other operations. + IsValid() bool +} diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/value_pure.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/value_pure.go new file mode 100644 index 00000000..918e685e --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/value_pure.go @@ -0,0 +1,59 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build purego appengine + +package protoreflect + +import "google.golang.org/protobuf/internal/pragma" + +type valueType int + +const ( + nilType valueType = iota + boolType + int32Type + int64Type + uint32Type + uint64Type + float32Type + float64Type + stringType + bytesType + enumType + ifaceType +) + +// value is a union where only one type can be represented at a time. +// This uses a distinct field for each type. This is type safe in Go, but +// occupies more memory than necessary (72B). +type value struct { + pragma.DoNotCompare // 0B + + typ valueType // 8B + num uint64 // 8B + str string // 16B + bin []byte // 24B + iface interface{} // 16B +} + +func valueOfString(v string) Value { + return Value{typ: stringType, str: v} +} +func valueOfBytes(v []byte) Value { + return Value{typ: bytesType, bin: v} +} +func valueOfIface(v interface{}) Value { + return Value{typ: ifaceType, iface: v} +} + +func (v Value) getString() string { + return v.str +} +func (v Value) getBytes() []byte { + return v.bin +} +func (v Value) getIface() interface{} { + return v.iface +} diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/value_union.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/value_union.go new file mode 100644 index 00000000..5a341472 --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/value_union.go @@ -0,0 +1,411 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package protoreflect + +import ( + "fmt" + "math" +) + +// Value is a union where only one Go type may be set at a time. +// The Value is used to represent all possible values a field may take. +// The following shows which Go type is used to represent each proto Kind: +// +// ╔════════════╤═════════════════════════════════════╗ +// ║ Go type │ Protobuf kind ║ +// ╠════════════╪═════════════════════════════════════╣ +// ║ bool │ BoolKind ║ +// ║ int32 │ Int32Kind, Sint32Kind, Sfixed32Kind ║ +// ║ int64 │ Int64Kind, Sint64Kind, Sfixed64Kind ║ +// ║ uint32 │ Uint32Kind, Fixed32Kind ║ +// ║ uint64 │ Uint64Kind, Fixed64Kind ║ +// ║ float32 │ FloatKind ║ +// ║ float64 │ DoubleKind ║ +// ║ string │ StringKind ║ +// ║ []byte │ BytesKind ║ +// ║ EnumNumber │ EnumKind ║ +// ║ Message │ MessageKind, GroupKind ║ +// ╚════════════╧═════════════════════════════════════╝ +// +// Multiple protobuf Kinds may be represented by a single Go type if the type +// can losslessly represent the information for the proto kind. For example, +// Int64Kind, Sint64Kind, and Sfixed64Kind are all represented by int64, +// but use different integer encoding methods. +// +// The List or Map types are used if the field cardinality is repeated. +// A field is a List if FieldDescriptor.IsList reports true. +// A field is a Map if FieldDescriptor.IsMap reports true. +// +// Converting to/from a Value and a concrete Go value panics on type mismatch. +// For example, ValueOf("hello").Int() panics because this attempts to +// retrieve an int64 from a string. +type Value value + +// The protoreflect API uses a custom Value union type instead of interface{} +// to keep the future open for performance optimizations. Using an interface{} +// always incurs an allocation for primitives (e.g., int64) since it needs to +// be boxed on the heap (as interfaces can only contain pointers natively). +// Instead, we represent the Value union as a flat struct that internally keeps +// track of which type is set. Using unsafe, the Value union can be reduced +// down to 24B, which is identical in size to a slice. +// +// The latest compiler (Go1.11) currently suffers from some limitations: +// • With inlining, the compiler should be able to statically prove that +// only one of these switch cases are taken and inline one specific case. +// See https://golang.org/issue/22310. + +// ValueOf returns a Value initialized with the concrete value stored in v. +// This panics if the type does not match one of the allowed types in the +// Value union. +func ValueOf(v interface{}) Value { + switch v := v.(type) { + case nil: + return Value{} + case bool: + return ValueOfBool(v) + case int32: + return ValueOfInt32(v) + case int64: + return ValueOfInt64(v) + case uint32: + return ValueOfUint32(v) + case uint64: + return ValueOfUint64(v) + case float32: + return ValueOfFloat32(v) + case float64: + return ValueOfFloat64(v) + case string: + return ValueOfString(v) + case []byte: + return ValueOfBytes(v) + case EnumNumber: + return ValueOfEnum(v) + case Message, List, Map: + return valueOfIface(v) + case ProtoMessage: + panic(fmt.Sprintf("invalid proto.Message(%T) type, expected a protoreflect.Message type", v)) + default: + panic(fmt.Sprintf("invalid type: %T", v)) + } +} + +// ValueOfBool returns a new boolean value. +func ValueOfBool(v bool) Value { + if v { + return Value{typ: boolType, num: 1} + } else { + return Value{typ: boolType, num: 0} + } +} + +// ValueOfInt32 returns a new int32 value. +func ValueOfInt32(v int32) Value { + return Value{typ: int32Type, num: uint64(v)} +} + +// ValueOfInt64 returns a new int64 value. +func ValueOfInt64(v int64) Value { + return Value{typ: int64Type, num: uint64(v)} +} + +// ValueOfUint32 returns a new uint32 value. +func ValueOfUint32(v uint32) Value { + return Value{typ: uint32Type, num: uint64(v)} +} + +// ValueOfUint64 returns a new uint64 value. +func ValueOfUint64(v uint64) Value { + return Value{typ: uint64Type, num: v} +} + +// ValueOfFloat32 returns a new float32 value. +func ValueOfFloat32(v float32) Value { + return Value{typ: float32Type, num: uint64(math.Float64bits(float64(v)))} +} + +// ValueOfFloat64 returns a new float64 value. +func ValueOfFloat64(v float64) Value { + return Value{typ: float64Type, num: uint64(math.Float64bits(float64(v)))} +} + +// ValueOfString returns a new string value. +func ValueOfString(v string) Value { + return valueOfString(v) +} + +// ValueOfBytes returns a new bytes value. +func ValueOfBytes(v []byte) Value { + return valueOfBytes(v[:len(v):len(v)]) +} + +// ValueOfEnum returns a new enum value. +func ValueOfEnum(v EnumNumber) Value { + return Value{typ: enumType, num: uint64(v)} +} + +// ValueOfMessage returns a new Message value. +func ValueOfMessage(v Message) Value { + return valueOfIface(v) +} + +// ValueOfList returns a new List value. +func ValueOfList(v List) Value { + return valueOfIface(v) +} + +// ValueOfMap returns a new Map value. +func ValueOfMap(v Map) Value { + return valueOfIface(v) +} + +// IsValid reports whether v is populated with a value. +func (v Value) IsValid() bool { + return v.typ != nilType +} + +// Interface returns v as an interface{}. +// +// Invariant: v == ValueOf(v).Interface() +func (v Value) Interface() interface{} { + switch v.typ { + case nilType: + return nil + case boolType: + return v.Bool() + case int32Type: + return int32(v.Int()) + case int64Type: + return int64(v.Int()) + case uint32Type: + return uint32(v.Uint()) + case uint64Type: + return uint64(v.Uint()) + case float32Type: + return float32(v.Float()) + case float64Type: + return float64(v.Float()) + case stringType: + return v.String() + case bytesType: + return v.Bytes() + case enumType: + return v.Enum() + default: + return v.getIface() + } +} + +func (v Value) typeName() string { + switch v.typ { + case nilType: + return "nil" + case boolType: + return "bool" + case int32Type: + return "int32" + case int64Type: + return "int64" + case uint32Type: + return "uint32" + case uint64Type: + return "uint64" + case float32Type: + return "float32" + case float64Type: + return "float64" + case stringType: + return "string" + case bytesType: + return "bytes" + case enumType: + return "enum" + default: + switch v := v.getIface().(type) { + case Message: + return "message" + case List: + return "list" + case Map: + return "map" + default: + return fmt.Sprintf("", v) + } + } +} + +func (v Value) panicMessage(what string) string { + return fmt.Sprintf("type mismatch: cannot convert %v to %s", v.typeName(), what) +} + +// Bool returns v as a bool and panics if the type is not a bool. +func (v Value) Bool() bool { + switch v.typ { + case boolType: + return v.num > 0 + default: + panic(v.panicMessage("bool")) + } +} + +// Int returns v as a int64 and panics if the type is not a int32 or int64. +func (v Value) Int() int64 { + switch v.typ { + case int32Type, int64Type: + return int64(v.num) + default: + panic(v.panicMessage("int")) + } +} + +// Uint returns v as a uint64 and panics if the type is not a uint32 or uint64. +func (v Value) Uint() uint64 { + switch v.typ { + case uint32Type, uint64Type: + return uint64(v.num) + default: + panic(v.panicMessage("uint")) + } +} + +// Float returns v as a float64 and panics if the type is not a float32 or float64. +func (v Value) Float() float64 { + switch v.typ { + case float32Type, float64Type: + return math.Float64frombits(uint64(v.num)) + default: + panic(v.panicMessage("float")) + } +} + +// String returns v as a string. Since this method implements fmt.Stringer, +// this returns the formatted string value for any non-string type. +func (v Value) String() string { + switch v.typ { + case stringType: + return v.getString() + default: + return fmt.Sprint(v.Interface()) + } +} + +// Bytes returns v as a []byte and panics if the type is not a []byte. +func (v Value) Bytes() []byte { + switch v.typ { + case bytesType: + return v.getBytes() + default: + panic(v.panicMessage("bytes")) + } +} + +// Enum returns v as a EnumNumber and panics if the type is not a EnumNumber. +func (v Value) Enum() EnumNumber { + switch v.typ { + case enumType: + return EnumNumber(v.num) + default: + panic(v.panicMessage("enum")) + } +} + +// Message returns v as a Message and panics if the type is not a Message. +func (v Value) Message() Message { + switch vi := v.getIface().(type) { + case Message: + return vi + default: + panic(v.panicMessage("message")) + } +} + +// List returns v as a List and panics if the type is not a List. +func (v Value) List() List { + switch vi := v.getIface().(type) { + case List: + return vi + default: + panic(v.panicMessage("list")) + } +} + +// Map returns v as a Map and panics if the type is not a Map. +func (v Value) Map() Map { + switch vi := v.getIface().(type) { + case Map: + return vi + default: + panic(v.panicMessage("map")) + } +} + +// MapKey returns v as a MapKey and panics for invalid MapKey types. +func (v Value) MapKey() MapKey { + switch v.typ { + case boolType, int32Type, int64Type, uint32Type, uint64Type, stringType: + return MapKey(v) + default: + panic(v.panicMessage("map key")) + } +} + +// MapKey is used to index maps, where the Go type of the MapKey must match +// the specified key Kind (see MessageDescriptor.IsMapEntry). +// The following shows what Go type is used to represent each proto Kind: +// +// ╔═════════╤═════════════════════════════════════╗ +// ║ Go type │ Protobuf kind ║ +// ╠═════════╪═════════════════════════════════════╣ +// ║ bool │ BoolKind ║ +// ║ int32 │ Int32Kind, Sint32Kind, Sfixed32Kind ║ +// ║ int64 │ Int64Kind, Sint64Kind, Sfixed64Kind ║ +// ║ uint32 │ Uint32Kind, Fixed32Kind ║ +// ║ uint64 │ Uint64Kind, Fixed64Kind ║ +// ║ string │ StringKind ║ +// ╚═════════╧═════════════════════════════════════╝ +// +// A MapKey is constructed and accessed through a Value: +// k := ValueOf("hash").MapKey() // convert string to MapKey +// s := k.String() // convert MapKey to string +// +// The MapKey is a strict subset of valid types used in Value; +// converting a Value to a MapKey with an invalid type panics. +type MapKey value + +// IsValid reports whether k is populated with a value. +func (k MapKey) IsValid() bool { + return Value(k).IsValid() +} + +// Interface returns k as an interface{}. +func (k MapKey) Interface() interface{} { + return Value(k).Interface() +} + +// Bool returns k as a bool and panics if the type is not a bool. +func (k MapKey) Bool() bool { + return Value(k).Bool() +} + +// Int returns k as a int64 and panics if the type is not a int32 or int64. +func (k MapKey) Int() int64 { + return Value(k).Int() +} + +// Uint returns k as a uint64 and panics if the type is not a uint32 or uint64. +func (k MapKey) Uint() uint64 { + return Value(k).Uint() +} + +// String returns k as a string. Since this method implements fmt.Stringer, +// this returns the formatted string value for any non-string type. +func (k MapKey) String() string { + return Value(k).String() +} + +// Value returns k as a Value. +func (k MapKey) Value() Value { + return Value(k) +} diff --git a/vendor/google.golang.org/protobuf/reflect/protoreflect/value_unsafe.go b/vendor/google.golang.org/protobuf/reflect/protoreflect/value_unsafe.go new file mode 100644 index 00000000..c45debdc --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoreflect/value_unsafe.go @@ -0,0 +1,98 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !purego,!appengine + +package protoreflect + +import ( + "unsafe" + + "google.golang.org/protobuf/internal/pragma" +) + +type ( + stringHeader struct { + Data unsafe.Pointer + Len int + } + sliceHeader struct { + Data unsafe.Pointer + Len int + Cap int + } + ifaceHeader struct { + Type unsafe.Pointer + Data unsafe.Pointer + } +) + +var ( + nilType = typeOf(nil) + boolType = typeOf(*new(bool)) + int32Type = typeOf(*new(int32)) + int64Type = typeOf(*new(int64)) + uint32Type = typeOf(*new(uint32)) + uint64Type = typeOf(*new(uint64)) + float32Type = typeOf(*new(float32)) + float64Type = typeOf(*new(float64)) + stringType = typeOf(*new(string)) + bytesType = typeOf(*new([]byte)) + enumType = typeOf(*new(EnumNumber)) +) + +// typeOf returns a pointer to the Go type information. +// The pointer is comparable and equal if and only if the types are identical. +func typeOf(t interface{}) unsafe.Pointer { + return (*ifaceHeader)(unsafe.Pointer(&t)).Type +} + +// value is a union where only one type can be represented at a time. +// The struct is 24B large on 64-bit systems and requires the minimum storage +// necessary to represent each possible type. +// +// The Go GC needs to be able to scan variables containing pointers. +// As such, pointers and non-pointers cannot be intermixed. +type value struct { + pragma.DoNotCompare // 0B + + // typ stores the type of the value as a pointer to the Go type. + typ unsafe.Pointer // 8B + + // ptr stores the data pointer for a String, Bytes, or interface value. + ptr unsafe.Pointer // 8B + + // num stores a Bool, Int32, Int64, Uint32, Uint64, Float32, Float64, or + // Enum value as a raw uint64. + // + // It is also used to store the length of a String or Bytes value; + // the capacity is ignored. + num uint64 // 8B +} + +func valueOfString(v string) Value { + p := (*stringHeader)(unsafe.Pointer(&v)) + return Value{typ: stringType, ptr: p.Data, num: uint64(len(v))} +} +func valueOfBytes(v []byte) Value { + p := (*sliceHeader)(unsafe.Pointer(&v)) + return Value{typ: bytesType, ptr: p.Data, num: uint64(len(v))} +} +func valueOfIface(v interface{}) Value { + p := (*ifaceHeader)(unsafe.Pointer(&v)) + return Value{typ: p.Type, ptr: p.Data} +} + +func (v Value) getString() (x string) { + *(*stringHeader)(unsafe.Pointer(&x)) = stringHeader{Data: v.ptr, Len: int(v.num)} + return x +} +func (v Value) getBytes() (x []byte) { + *(*sliceHeader)(unsafe.Pointer(&x)) = sliceHeader{Data: v.ptr, Len: int(v.num), Cap: int(v.num)} + return x +} +func (v Value) getIface() (x interface{}) { + *(*ifaceHeader)(unsafe.Pointer(&x)) = ifaceHeader{Type: v.typ, Data: v.ptr} + return x +} diff --git a/vendor/google.golang.org/protobuf/reflect/protoregistry/registry.go b/vendor/google.golang.org/protobuf/reflect/protoregistry/registry.go new file mode 100644 index 00000000..5e5f9671 --- /dev/null +++ b/vendor/google.golang.org/protobuf/reflect/protoregistry/registry.go @@ -0,0 +1,800 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package protoregistry provides data structures to register and lookup +// protobuf descriptor types. +// +// The Files registry contains file descriptors and provides the ability +// to iterate over the files or lookup a specific descriptor within the files. +// Files only contains protobuf descriptors and has no understanding of Go +// type information that may be associated with each descriptor. +// +// The Types registry contains descriptor types for which there is a known +// Go type associated with that descriptor. It provides the ability to iterate +// over the registered types or lookup a type by name. +package protoregistry + +import ( + "fmt" + "log" + "strings" + "sync" + + "google.golang.org/protobuf/internal/errors" + "google.golang.org/protobuf/reflect/protoreflect" +) + +// ignoreConflict reports whether to ignore a registration conflict +// given the descriptor being registered and the error. +// It is a variable so that the behavior is easily overridden in another file. +var ignoreConflict = func(d protoreflect.Descriptor, err error) bool { + log.Printf(""+ + "WARNING: %v\n"+ + "A future release will panic on registration conflicts. See:\n"+ + "https://developers.google.com/protocol-buffers/docs/reference/go/faq#namespace-conflict\n"+ + "\n", err) + return true +} + +var globalMutex sync.RWMutex + +// GlobalFiles is a global registry of file descriptors. +var GlobalFiles *Files = new(Files) + +// GlobalTypes is the registry used by default for type lookups +// unless a local registry is provided by the user. +var GlobalTypes *Types = new(Types) + +// NotFound is a sentinel error value to indicate that the type was not found. +// +// Since registry lookup can happen in the critical performance path, resolvers +// must return this exact error value, not an error wrapping it. +var NotFound = errors.New("not found") + +// Files is a registry for looking up or iterating over files and the +// descriptors contained within them. +// The Find and Range methods are safe for concurrent use. +type Files struct { + // The map of descsByName contains: + // EnumDescriptor + // EnumValueDescriptor + // MessageDescriptor + // ExtensionDescriptor + // ServiceDescriptor + // *packageDescriptor + // + // Note that files are stored as a slice, since a package may contain + // multiple files. Only top-level declarations are registered. + // Note that enum values are in the top-level since that are in the same + // scope as the parent enum. + descsByName map[protoreflect.FullName]interface{} + filesByPath map[string]protoreflect.FileDescriptor +} + +type packageDescriptor struct { + files []protoreflect.FileDescriptor +} + +// RegisterFile registers the provided file descriptor. +// +// If any descriptor within the file conflicts with the descriptor of any +// previously registered file (e.g., two enums with the same full name), +// then the file is not registered and an error is returned. +// +// It is permitted for multiple files to have the same file path. +func (r *Files) RegisterFile(file protoreflect.FileDescriptor) error { + if r == GlobalFiles { + globalMutex.Lock() + defer globalMutex.Unlock() + } + if r.descsByName == nil { + r.descsByName = map[protoreflect.FullName]interface{}{ + "": &packageDescriptor{}, + } + r.filesByPath = make(map[string]protoreflect.FileDescriptor) + } + path := file.Path() + if prev := r.filesByPath[path]; prev != nil { + // TODO: Remove this after some soak-in period after moving these types. + var prevPath string + const prevModule = "google.golang.org/genproto" + const prevVersion = "cb27e3aa (May 26th, 2020)" + switch path { + case "google/protobuf/field_mask.proto": + prevPath = prevModule + "/protobuf/field_mask" + case "google/protobuf/api.proto": + prevPath = prevModule + "/protobuf/api" + case "google/protobuf/type.proto": + prevPath = prevModule + "/protobuf/ptype" + case "google/protobuf/source_context.proto": + prevPath = prevModule + "/protobuf/source_context" + } + if r == GlobalFiles && prevPath != "" { + pkgName := strings.TrimSuffix(strings.TrimPrefix(path, "google/protobuf/"), ".proto") + pkgName = strings.Replace(pkgName, "_", "", -1) + "pb" + currPath := "google.golang.org/protobuf/types/known/" + pkgName + panic(fmt.Sprintf(""+ + "duplicate registration of %q\n"+ + "\n"+ + "The generated definition for this file has moved:\n"+ + "\tfrom: %q\n"+ + "\tto: %q\n"+ + "A dependency on the %q module must\n"+ + "be at version %v or higher.\n"+ + "\n"+ + "Upgrade the dependency by running:\n"+ + "\tgo get -u %v\n", + path, prevPath, currPath, prevModule, prevVersion, prevPath)) + } + + err := errors.New("file %q is already registered", file.Path()) + err = amendErrorWithCaller(err, prev, file) + if r == GlobalFiles && ignoreConflict(file, err) { + err = nil + } + return err + } + + for name := file.Package(); name != ""; name = name.Parent() { + switch prev := r.descsByName[name]; prev.(type) { + case nil, *packageDescriptor: + default: + err := errors.New("file %q has a package name conflict over %v", file.Path(), name) + err = amendErrorWithCaller(err, prev, file) + if r == GlobalFiles && ignoreConflict(file, err) { + err = nil + } + return err + } + } + var err error + var hasConflict bool + rangeTopLevelDescriptors(file, func(d protoreflect.Descriptor) { + if prev := r.descsByName[d.FullName()]; prev != nil { + hasConflict = true + err = errors.New("file %q has a name conflict over %v", file.Path(), d.FullName()) + err = amendErrorWithCaller(err, prev, file) + if r == GlobalFiles && ignoreConflict(d, err) { + err = nil + } + } + }) + if hasConflict { + return err + } + + for name := file.Package(); name != ""; name = name.Parent() { + if r.descsByName[name] == nil { + r.descsByName[name] = &packageDescriptor{} + } + } + p := r.descsByName[file.Package()].(*packageDescriptor) + p.files = append(p.files, file) + rangeTopLevelDescriptors(file, func(d protoreflect.Descriptor) { + r.descsByName[d.FullName()] = d + }) + r.filesByPath[path] = file + return nil +} + +// FindDescriptorByName looks up a descriptor by the full name. +// +// This returns (nil, NotFound) if not found. +func (r *Files) FindDescriptorByName(name protoreflect.FullName) (protoreflect.Descriptor, error) { + if r == nil { + return nil, NotFound + } + if r == GlobalFiles { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + prefix := name + suffix := nameSuffix("") + for prefix != "" { + if d, ok := r.descsByName[prefix]; ok { + switch d := d.(type) { + case protoreflect.EnumDescriptor: + if d.FullName() == name { + return d, nil + } + case protoreflect.EnumValueDescriptor: + if d.FullName() == name { + return d, nil + } + case protoreflect.MessageDescriptor: + if d.FullName() == name { + return d, nil + } + if d := findDescriptorInMessage(d, suffix); d != nil && d.FullName() == name { + return d, nil + } + case protoreflect.ExtensionDescriptor: + if d.FullName() == name { + return d, nil + } + case protoreflect.ServiceDescriptor: + if d.FullName() == name { + return d, nil + } + if d := d.Methods().ByName(suffix.Pop()); d != nil && d.FullName() == name { + return d, nil + } + } + return nil, NotFound + } + prefix = prefix.Parent() + suffix = nameSuffix(name[len(prefix)+len("."):]) + } + return nil, NotFound +} + +func findDescriptorInMessage(md protoreflect.MessageDescriptor, suffix nameSuffix) protoreflect.Descriptor { + name := suffix.Pop() + if suffix == "" { + if ed := md.Enums().ByName(name); ed != nil { + return ed + } + for i := md.Enums().Len() - 1; i >= 0; i-- { + if vd := md.Enums().Get(i).Values().ByName(name); vd != nil { + return vd + } + } + if xd := md.Extensions().ByName(name); xd != nil { + return xd + } + if fd := md.Fields().ByName(name); fd != nil { + return fd + } + if od := md.Oneofs().ByName(name); od != nil { + return od + } + } + if md := md.Messages().ByName(name); md != nil { + if suffix == "" { + return md + } + return findDescriptorInMessage(md, suffix) + } + return nil +} + +type nameSuffix string + +func (s *nameSuffix) Pop() (name protoreflect.Name) { + if i := strings.IndexByte(string(*s), '.'); i >= 0 { + name, *s = protoreflect.Name((*s)[:i]), (*s)[i+1:] + } else { + name, *s = protoreflect.Name((*s)), "" + } + return name +} + +// FindFileByPath looks up a file by the path. +// +// This returns (nil, NotFound) if not found. +func (r *Files) FindFileByPath(path string) (protoreflect.FileDescriptor, error) { + if r == nil { + return nil, NotFound + } + if r == GlobalFiles { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + if fd, ok := r.filesByPath[path]; ok { + return fd, nil + } + return nil, NotFound +} + +// NumFiles reports the number of registered files. +func (r *Files) NumFiles() int { + if r == nil { + return 0 + } + if r == GlobalFiles { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + return len(r.filesByPath) +} + +// RangeFiles iterates over all registered files while f returns true. +// The iteration order is undefined. +func (r *Files) RangeFiles(f func(protoreflect.FileDescriptor) bool) { + if r == nil { + return + } + if r == GlobalFiles { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + for _, file := range r.filesByPath { + if !f(file) { + return + } + } +} + +// NumFilesByPackage reports the number of registered files in a proto package. +func (r *Files) NumFilesByPackage(name protoreflect.FullName) int { + if r == nil { + return 0 + } + if r == GlobalFiles { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + p, ok := r.descsByName[name].(*packageDescriptor) + if !ok { + return 0 + } + return len(p.files) +} + +// RangeFilesByPackage iterates over all registered files in a given proto package +// while f returns true. The iteration order is undefined. +func (r *Files) RangeFilesByPackage(name protoreflect.FullName, f func(protoreflect.FileDescriptor) bool) { + if r == nil { + return + } + if r == GlobalFiles { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + p, ok := r.descsByName[name].(*packageDescriptor) + if !ok { + return + } + for _, file := range p.files { + if !f(file) { + return + } + } +} + +// rangeTopLevelDescriptors iterates over all top-level descriptors in a file +// which will be directly entered into the registry. +func rangeTopLevelDescriptors(fd protoreflect.FileDescriptor, f func(protoreflect.Descriptor)) { + eds := fd.Enums() + for i := eds.Len() - 1; i >= 0; i-- { + f(eds.Get(i)) + vds := eds.Get(i).Values() + for i := vds.Len() - 1; i >= 0; i-- { + f(vds.Get(i)) + } + } + mds := fd.Messages() + for i := mds.Len() - 1; i >= 0; i-- { + f(mds.Get(i)) + } + xds := fd.Extensions() + for i := xds.Len() - 1; i >= 0; i-- { + f(xds.Get(i)) + } + sds := fd.Services() + for i := sds.Len() - 1; i >= 0; i-- { + f(sds.Get(i)) + } +} + +// MessageTypeResolver is an interface for looking up messages. +// +// A compliant implementation must deterministically return the same type +// if no error is encountered. +// +// The Types type implements this interface. +type MessageTypeResolver interface { + // FindMessageByName looks up a message by its full name. + // E.g., "google.protobuf.Any" + // + // This return (nil, NotFound) if not found. + FindMessageByName(message protoreflect.FullName) (protoreflect.MessageType, error) + + // FindMessageByURL looks up a message by a URL identifier. + // See documentation on google.protobuf.Any.type_url for the URL format. + // + // This returns (nil, NotFound) if not found. + FindMessageByURL(url string) (protoreflect.MessageType, error) +} + +// ExtensionTypeResolver is an interface for looking up extensions. +// +// A compliant implementation must deterministically return the same type +// if no error is encountered. +// +// The Types type implements this interface. +type ExtensionTypeResolver interface { + // FindExtensionByName looks up a extension field by the field's full name. + // Note that this is the full name of the field as determined by + // where the extension is declared and is unrelated to the full name of the + // message being extended. + // + // This returns (nil, NotFound) if not found. + FindExtensionByName(field protoreflect.FullName) (protoreflect.ExtensionType, error) + + // FindExtensionByNumber looks up a extension field by the field number + // within some parent message, identified by full name. + // + // This returns (nil, NotFound) if not found. + FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) +} + +var ( + _ MessageTypeResolver = (*Types)(nil) + _ ExtensionTypeResolver = (*Types)(nil) +) + +// Types is a registry for looking up or iterating over descriptor types. +// The Find and Range methods are safe for concurrent use. +type Types struct { + typesByName typesByName + extensionsByMessage extensionsByMessage + + numEnums int + numMessages int + numExtensions int +} + +type ( + typesByName map[protoreflect.FullName]interface{} + extensionsByMessage map[protoreflect.FullName]extensionsByNumber + extensionsByNumber map[protoreflect.FieldNumber]protoreflect.ExtensionType +) + +// RegisterMessage registers the provided message type. +// +// If a naming conflict occurs, the type is not registered and an error is returned. +func (r *Types) RegisterMessage(mt protoreflect.MessageType) error { + // Under rare circumstances getting the descriptor might recursively + // examine the registry, so fetch it before locking. + md := mt.Descriptor() + + if r == GlobalTypes { + globalMutex.Lock() + defer globalMutex.Unlock() + } + + if err := r.register("message", md, mt); err != nil { + return err + } + r.numMessages++ + return nil +} + +// RegisterEnum registers the provided enum type. +// +// If a naming conflict occurs, the type is not registered and an error is returned. +func (r *Types) RegisterEnum(et protoreflect.EnumType) error { + // Under rare circumstances getting the descriptor might recursively + // examine the registry, so fetch it before locking. + ed := et.Descriptor() + + if r == GlobalTypes { + globalMutex.Lock() + defer globalMutex.Unlock() + } + + if err := r.register("enum", ed, et); err != nil { + return err + } + r.numEnums++ + return nil +} + +// RegisterExtension registers the provided extension type. +// +// If a naming conflict occurs, the type is not registered and an error is returned. +func (r *Types) RegisterExtension(xt protoreflect.ExtensionType) error { + // Under rare circumstances getting the descriptor might recursively + // examine the registry, so fetch it before locking. + // + // A known case where this can happen: Fetching the TypeDescriptor for a + // legacy ExtensionDesc can consult the global registry. + xd := xt.TypeDescriptor() + + if r == GlobalTypes { + globalMutex.Lock() + defer globalMutex.Unlock() + } + + field := xd.Number() + message := xd.ContainingMessage().FullName() + if prev := r.extensionsByMessage[message][field]; prev != nil { + err := errors.New("extension number %d is already registered on message %v", field, message) + err = amendErrorWithCaller(err, prev, xt) + if !(r == GlobalTypes && ignoreConflict(xd, err)) { + return err + } + } + + if err := r.register("extension", xd, xt); err != nil { + return err + } + if r.extensionsByMessage == nil { + r.extensionsByMessage = make(extensionsByMessage) + } + if r.extensionsByMessage[message] == nil { + r.extensionsByMessage[message] = make(extensionsByNumber) + } + r.extensionsByMessage[message][field] = xt + r.numExtensions++ + return nil +} + +func (r *Types) register(kind string, desc protoreflect.Descriptor, typ interface{}) error { + name := desc.FullName() + prev := r.typesByName[name] + if prev != nil { + err := errors.New("%v %v is already registered", kind, name) + err = amendErrorWithCaller(err, prev, typ) + if !(r == GlobalTypes && ignoreConflict(desc, err)) { + return err + } + } + if r.typesByName == nil { + r.typesByName = make(typesByName) + } + r.typesByName[name] = typ + return nil +} + +// FindEnumByName looks up an enum by its full name. +// E.g., "google.protobuf.Field.Kind". +// +// This returns (nil, NotFound) if not found. +func (r *Types) FindEnumByName(enum protoreflect.FullName) (protoreflect.EnumType, error) { + if r == nil { + return nil, NotFound + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + if v := r.typesByName[enum]; v != nil { + if et, _ := v.(protoreflect.EnumType); et != nil { + return et, nil + } + return nil, errors.New("found wrong type: got %v, want enum", typeName(v)) + } + return nil, NotFound +} + +// FindMessageByName looks up a message by its full name. +// E.g., "google.protobuf.Any" +// +// This return (nil, NotFound) if not found. +func (r *Types) FindMessageByName(message protoreflect.FullName) (protoreflect.MessageType, error) { + // The full name by itself is a valid URL. + return r.FindMessageByURL(string(message)) +} + +// FindMessageByURL looks up a message by a URL identifier. +// See documentation on google.protobuf.Any.type_url for the URL format. +// +// This returns (nil, NotFound) if not found. +func (r *Types) FindMessageByURL(url string) (protoreflect.MessageType, error) { + if r == nil { + return nil, NotFound + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + message := protoreflect.FullName(url) + if i := strings.LastIndexByte(url, '/'); i >= 0 { + message = message[i+len("/"):] + } + + if v := r.typesByName[message]; v != nil { + if mt, _ := v.(protoreflect.MessageType); mt != nil { + return mt, nil + } + return nil, errors.New("found wrong type: got %v, want message", typeName(v)) + } + return nil, NotFound +} + +// FindExtensionByName looks up a extension field by the field's full name. +// Note that this is the full name of the field as determined by +// where the extension is declared and is unrelated to the full name of the +// message being extended. +// +// This returns (nil, NotFound) if not found. +func (r *Types) FindExtensionByName(field protoreflect.FullName) (protoreflect.ExtensionType, error) { + if r == nil { + return nil, NotFound + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + if v := r.typesByName[field]; v != nil { + if xt, _ := v.(protoreflect.ExtensionType); xt != nil { + return xt, nil + } + return nil, errors.New("found wrong type: got %v, want extension", typeName(v)) + } + return nil, NotFound +} + +// FindExtensionByNumber looks up a extension field by the field number +// within some parent message, identified by full name. +// +// This returns (nil, NotFound) if not found. +func (r *Types) FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) { + if r == nil { + return nil, NotFound + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + if xt, ok := r.extensionsByMessage[message][field]; ok { + return xt, nil + } + return nil, NotFound +} + +// NumEnums reports the number of registered enums. +func (r *Types) NumEnums() int { + if r == nil { + return 0 + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + return r.numEnums +} + +// RangeEnums iterates over all registered enums while f returns true. +// Iteration order is undefined. +func (r *Types) RangeEnums(f func(protoreflect.EnumType) bool) { + if r == nil { + return + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + for _, typ := range r.typesByName { + if et, ok := typ.(protoreflect.EnumType); ok { + if !f(et) { + return + } + } + } +} + +// NumMessages reports the number of registered messages. +func (r *Types) NumMessages() int { + if r == nil { + return 0 + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + return r.numMessages +} + +// RangeMessages iterates over all registered messages while f returns true. +// Iteration order is undefined. +func (r *Types) RangeMessages(f func(protoreflect.MessageType) bool) { + if r == nil { + return + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + for _, typ := range r.typesByName { + if mt, ok := typ.(protoreflect.MessageType); ok { + if !f(mt) { + return + } + } + } +} + +// NumExtensions reports the number of registered extensions. +func (r *Types) NumExtensions() int { + if r == nil { + return 0 + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + return r.numExtensions +} + +// RangeExtensions iterates over all registered extensions while f returns true. +// Iteration order is undefined. +func (r *Types) RangeExtensions(f func(protoreflect.ExtensionType) bool) { + if r == nil { + return + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + for _, typ := range r.typesByName { + if xt, ok := typ.(protoreflect.ExtensionType); ok { + if !f(xt) { + return + } + } + } +} + +// NumExtensionsByMessage reports the number of registered extensions for +// a given message type. +func (r *Types) NumExtensionsByMessage(message protoreflect.FullName) int { + if r == nil { + return 0 + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + return len(r.extensionsByMessage[message]) +} + +// RangeExtensionsByMessage iterates over all registered extensions filtered +// by a given message type while f returns true. Iteration order is undefined. +func (r *Types) RangeExtensionsByMessage(message protoreflect.FullName, f func(protoreflect.ExtensionType) bool) { + if r == nil { + return + } + if r == GlobalTypes { + globalMutex.RLock() + defer globalMutex.RUnlock() + } + for _, xt := range r.extensionsByMessage[message] { + if !f(xt) { + return + } + } +} + +func typeName(t interface{}) string { + switch t.(type) { + case protoreflect.EnumType: + return "enum" + case protoreflect.MessageType: + return "message" + case protoreflect.ExtensionType: + return "extension" + default: + return fmt.Sprintf("%T", t) + } +} + +func amendErrorWithCaller(err error, prev, curr interface{}) error { + prevPkg := goPackage(prev) + currPkg := goPackage(curr) + if prevPkg == "" || currPkg == "" || prevPkg == currPkg { + return err + } + return errors.New("%s\n\tpreviously from: %q\n\tcurrently from: %q", err, prevPkg, currPkg) +} + +func goPackage(v interface{}) string { + switch d := v.(type) { + case protoreflect.EnumType: + v = d.Descriptor() + case protoreflect.MessageType: + v = d.Descriptor() + case protoreflect.ExtensionType: + v = d.TypeDescriptor() + } + if d, ok := v.(protoreflect.Descriptor); ok { + v = d.ParentFile() + } + if d, ok := v.(interface{ GoPackagePath() string }); ok { + return d.GoPackagePath() + } + return "" +} diff --git a/vendor/google.golang.org/protobuf/runtime/protoiface/legacy.go b/vendor/google.golang.org/protobuf/runtime/protoiface/legacy.go new file mode 100644 index 00000000..c5872767 --- /dev/null +++ b/vendor/google.golang.org/protobuf/runtime/protoiface/legacy.go @@ -0,0 +1,15 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package protoiface + +type MessageV1 interface { + Reset() + String() string + ProtoMessage() +} + +type ExtensionRangeV1 struct { + Start, End int32 // both inclusive +} diff --git a/vendor/google.golang.org/protobuf/runtime/protoiface/methods.go b/vendor/google.golang.org/protobuf/runtime/protoiface/methods.go new file mode 100644 index 00000000..32c04f67 --- /dev/null +++ b/vendor/google.golang.org/protobuf/runtime/protoiface/methods.go @@ -0,0 +1,167 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package protoiface contains types referenced or implemented by messages. +// +// WARNING: This package should only be imported by message implementations. +// The functionality found in this package should be accessed through +// higher-level abstractions provided by the proto package. +package protoiface + +import ( + "google.golang.org/protobuf/internal/pragma" + "google.golang.org/protobuf/reflect/protoreflect" +) + +// Methods is a set of optional fast-path implementations of various operations. +type Methods = struct { + pragma.NoUnkeyedLiterals + + // Flags indicate support for optional features. + Flags SupportFlags + + // Size returns the size in bytes of the wire-format encoding of a message. + // Marshal must be provided if a custom Size is provided. + Size func(SizeInput) SizeOutput + + // Marshal formats a message in the wire-format encoding to the provided buffer. + // Size should be provided if a custom Marshal is provided. + // It must not return an error for a partial message. + Marshal func(MarshalInput) (MarshalOutput, error) + + // Unmarshal parses the wire-format encoding and merges the result into a message. + // It must not reset the target message or return an error for a partial message. + Unmarshal func(UnmarshalInput) (UnmarshalOutput, error) + + // Merge merges the contents of a source message into a destination message. + Merge func(MergeInput) MergeOutput + + // CheckInitialized returns an error if any required fields in the message are not set. + CheckInitialized func(CheckInitializedInput) (CheckInitializedOutput, error) +} + +// SupportFlags indicate support for optional features. +type SupportFlags = uint64 + +const ( + // SupportMarshalDeterministic reports whether MarshalOptions.Deterministic is supported. + SupportMarshalDeterministic SupportFlags = 1 << iota + + // SupportUnmarshalDiscardUnknown reports whether UnmarshalOptions.DiscardUnknown is supported. + SupportUnmarshalDiscardUnknown +) + +// SizeInput is input to the Size method. +type SizeInput = struct { + pragma.NoUnkeyedLiterals + + Message protoreflect.Message + Flags MarshalInputFlags +} + +// SizeOutput is output from the Size method. +type SizeOutput = struct { + pragma.NoUnkeyedLiterals + + Size int +} + +// MarshalInput is input to the Marshal method. +type MarshalInput = struct { + pragma.NoUnkeyedLiterals + + Message protoreflect.Message + Buf []byte // output is appended to this buffer + Flags MarshalInputFlags +} + +// MarshalOutput is output from the Marshal method. +type MarshalOutput = struct { + pragma.NoUnkeyedLiterals + + Buf []byte // contains marshaled message +} + +// MarshalInputFlags configure the marshaler. +// Most flags correspond to fields in proto.MarshalOptions. +type MarshalInputFlags = uint8 + +const ( + MarshalDeterministic MarshalInputFlags = 1 << iota + MarshalUseCachedSize +) + +// UnmarshalInput is input to the Unmarshal method. +type UnmarshalInput = struct { + pragma.NoUnkeyedLiterals + + Message protoreflect.Message + Buf []byte // input buffer + Flags UnmarshalInputFlags + Resolver interface { + FindExtensionByName(field protoreflect.FullName) (protoreflect.ExtensionType, error) + FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) + } +} + +// UnmarshalOutput is output from the Unmarshal method. +type UnmarshalOutput = struct { + pragma.NoUnkeyedLiterals + + Flags UnmarshalOutputFlags +} + +// UnmarshalInputFlags configure the unmarshaler. +// Most flags correspond to fields in proto.UnmarshalOptions. +type UnmarshalInputFlags = uint8 + +const ( + UnmarshalDiscardUnknown UnmarshalInputFlags = 1 << iota +) + +// UnmarshalOutputFlags are output from the Unmarshal method. +type UnmarshalOutputFlags = uint8 + +const ( + // UnmarshalInitialized may be set on return if all required fields are known to be set. + // If unset, then it does not necessarily indicate that the message is uninitialized, + // only that its status could not be confirmed. + UnmarshalInitialized UnmarshalOutputFlags = 1 << iota +) + +// MergeInput is input to the Merge method. +type MergeInput = struct { + pragma.NoUnkeyedLiterals + + Source protoreflect.Message + Destination protoreflect.Message +} + +// MergeOutput is output from the Merge method. +type MergeOutput = struct { + pragma.NoUnkeyedLiterals + + Flags MergeOutputFlags +} + +// MergeOutputFlags are output from the Merge method. +type MergeOutputFlags = uint8 + +const ( + // MergeComplete reports whether the merge was performed. + // If unset, the merger must have made no changes to the destination. + MergeComplete MergeOutputFlags = 1 << iota +) + +// CheckInitializedInput is input to the CheckInitialized method. +type CheckInitializedInput = struct { + pragma.NoUnkeyedLiterals + + Message protoreflect.Message +} + +// CheckInitializedOutput is output from the CheckInitialized method. +type CheckInitializedOutput = struct { + pragma.NoUnkeyedLiterals +} diff --git a/vendor/google.golang.org/protobuf/runtime/protoimpl/impl.go b/vendor/google.golang.org/protobuf/runtime/protoimpl/impl.go new file mode 100644 index 00000000..4a1ab7fb --- /dev/null +++ b/vendor/google.golang.org/protobuf/runtime/protoimpl/impl.go @@ -0,0 +1,44 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// Package protoimpl contains the default implementation for messages +// generated by protoc-gen-go. +// +// WARNING: This package should only ever be imported by generated messages. +// The compatibility agreement covers nothing except for functionality needed +// to keep existing generated messages operational. Breakages that occur due +// to unauthorized usages of this package are not the author's responsibility. +package protoimpl + +import ( + "google.golang.org/protobuf/internal/filedesc" + "google.golang.org/protobuf/internal/filetype" + "google.golang.org/protobuf/internal/impl" +) + +// UnsafeEnabled specifies whether package unsafe can be used. +const UnsafeEnabled = impl.UnsafeEnabled + +type ( + // Types used by generated code in init functions. + DescBuilder = filedesc.Builder + TypeBuilder = filetype.Builder + + // Types used by generated code to implement EnumType, MessageType, and ExtensionType. + EnumInfo = impl.EnumInfo + MessageInfo = impl.MessageInfo + ExtensionInfo = impl.ExtensionInfo + + // Types embedded in generated messages. + MessageState = impl.MessageState + SizeCache = impl.SizeCache + WeakFields = impl.WeakFields + UnknownFields = impl.UnknownFields + ExtensionFields = impl.ExtensionFields + ExtensionFieldV1 = impl.ExtensionField + + Pointer = impl.Pointer +) + +var X impl.Export diff --git a/vendor/google.golang.org/protobuf/runtime/protoimpl/version.go b/vendor/google.golang.org/protobuf/runtime/protoimpl/version.go new file mode 100644 index 00000000..ff094e1b --- /dev/null +++ b/vendor/google.golang.org/protobuf/runtime/protoimpl/version.go @@ -0,0 +1,56 @@ +// Copyright 2019 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package protoimpl + +import ( + "google.golang.org/protobuf/internal/version" +) + +const ( + // MaxVersion is the maximum supported version for generated .pb.go files. + // It is always the current version of the module. + MaxVersion = version.Minor + + // GenVersion is the runtime version required by generated .pb.go files. + // This is incremented when generated code relies on new functionality + // in the runtime. + GenVersion = 20 + + // MinVersion is the minimum supported version for generated .pb.go files. + // This is incremented when the runtime drops support for old code. + MinVersion = 0 +) + +// EnforceVersion is used by code generated by protoc-gen-go +// to statically enforce minimum and maximum versions of this package. +// A compilation failure implies either that: +// * the runtime package is too old and needs to be updated OR +// * the generated code is too old and needs to be regenerated. +// +// The runtime package can be upgraded by running: +// go get google.golang.org/protobuf +// +// The generated code can be regenerated by running: +// protoc --go_out=${PROTOC_GEN_GO_ARGS} ${PROTO_FILES} +// +// Example usage by generated code: +// const ( +// // Verify that this generated code is sufficiently up-to-date. +// _ = protoimpl.EnforceVersion(genVersion - protoimpl.MinVersion) +// // Verify that runtime/protoimpl is sufficiently up-to-date. +// _ = protoimpl.EnforceVersion(protoimpl.MaxVersion - genVersion) +// ) +// +// The genVersion is the current minor version used to generated the code. +// This compile-time check relies on negative integer overflow of a uint +// being a compilation failure (guaranteed by the Go specification). +type EnforceVersion uint + +// This enforces the following invariant: +// MinVersion ≤ GenVersion ≤ MaxVersion +const ( + _ = EnforceVersion(GenVersion - MinVersion) + _ = EnforceVersion(MaxVersion - GenVersion) +) diff --git a/vendor/google.golang.org/protobuf/types/known/anypb/any.pb.go b/vendor/google.golang.org/protobuf/types/known/anypb/any.pb.go new file mode 100644 index 00000000..5f9498e4 --- /dev/null +++ b/vendor/google.golang.org/protobuf/types/known/anypb/any.pb.go @@ -0,0 +1,287 @@ +// Protocol Buffers - Google's data interchange format +// Copyright 2008 Google Inc. All rights reserved. +// https://developers.google.com/protocol-buffers/ +// +// Redistribution and use in source and binary forms, with or without +// modification, are permitted provided that the following conditions are +// met: +// +// * Redistributions of source code must retain the above copyright +// notice, this list of conditions and the following disclaimer. +// * Redistributions in binary form must reproduce the above +// copyright notice, this list of conditions and the following disclaimer +// in the documentation and/or other materials provided with the +// distribution. +// * Neither the name of Google Inc. nor the names of its +// contributors may be used to endorse or promote products derived from +// this software without specific prior written permission. +// +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/protobuf/any.proto + +package anypb + +import ( + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + reflect "reflect" + sync "sync" +) + +// `Any` contains an arbitrary serialized protocol buffer message along with a +// URL that describes the type of the serialized message. +// +// Protobuf library provides support to pack/unpack Any values in the form +// of utility functions or additional generated methods of the Any type. +// +// Example 1: Pack and unpack a message in C++. +// +// Foo foo = ...; +// Any any; +// any.PackFrom(foo); +// ... +// if (any.UnpackTo(&foo)) { +// ... +// } +// +// Example 2: Pack and unpack a message in Java. +// +// Foo foo = ...; +// Any any = Any.pack(foo); +// ... +// if (any.is(Foo.class)) { +// foo = any.unpack(Foo.class); +// } +// +// Example 3: Pack and unpack a message in Python. +// +// foo = Foo(...) +// any = Any() +// any.Pack(foo) +// ... +// if any.Is(Foo.DESCRIPTOR): +// any.Unpack(foo) +// ... +// +// Example 4: Pack and unpack a message in Go +// +// foo := &pb.Foo{...} +// any, err := ptypes.MarshalAny(foo) +// ... +// foo := &pb.Foo{} +// if err := ptypes.UnmarshalAny(any, foo); err != nil { +// ... +// } +// +// The pack methods provided by protobuf library will by default use +// 'type.googleapis.com/full.type.name' as the type URL and the unpack +// methods only use the fully qualified type name after the last '/' +// in the type URL, for example "foo.bar.com/x/y.z" will yield type +// name "y.z". +// +// +// JSON +// ==== +// The JSON representation of an `Any` value uses the regular +// representation of the deserialized, embedded message, with an +// additional field `@type` which contains the type URL. Example: +// +// package google.profile; +// message Person { +// string first_name = 1; +// string last_name = 2; +// } +// +// { +// "@type": "type.googleapis.com/google.profile.Person", +// "firstName": , +// "lastName": +// } +// +// If the embedded message type is well-known and has a custom JSON +// representation, that representation will be embedded adding a field +// `value` which holds the custom JSON in addition to the `@type` +// field. Example (for message [google.protobuf.Duration][]): +// +// { +// "@type": "type.googleapis.com/google.protobuf.Duration", +// "value": "1.212s" +// } +// +type Any struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // A URL/resource name that uniquely identifies the type of the serialized + // protocol buffer message. This string must contain at least + // one "/" character. The last segment of the URL's path must represent + // the fully qualified name of the type (as in + // `path/google.protobuf.Duration`). The name should be in a canonical form + // (e.g., leading "." is not accepted). + // + // In practice, teams usually precompile into the binary all types that they + // expect it to use in the context of Any. However, for URLs which use the + // scheme `http`, `https`, or no scheme, one can optionally set up a type + // server that maps type URLs to message definitions as follows: + // + // * If no scheme is provided, `https` is assumed. + // * An HTTP GET on the URL must yield a [google.protobuf.Type][] + // value in binary format, or produce an error. + // * Applications are allowed to cache lookup results based on the + // URL, or have them precompiled into a binary to avoid any + // lookup. Therefore, binary compatibility needs to be preserved + // on changes to types. (Use versioned type names to manage + // breaking changes.) + // + // Note: this functionality is not currently available in the official + // protobuf release, and it is not used for type URLs beginning with + // type.googleapis.com. + // + // Schemes other than `http`, `https` (or the empty scheme) might be + // used with implementation specific semantics. + // + TypeUrl string `protobuf:"bytes,1,opt,name=type_url,json=typeUrl,proto3" json:"type_url,omitempty"` + // Must be a valid serialized protocol buffer of the above specified type. + Value []byte `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` +} + +func (x *Any) Reset() { + *x = Any{} + if protoimpl.UnsafeEnabled { + mi := &file_google_protobuf_any_proto_msgTypes[0] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Any) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Any) ProtoMessage() {} + +func (x *Any) ProtoReflect() protoreflect.Message { + mi := &file_google_protobuf_any_proto_msgTypes[0] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Any.ProtoReflect.Descriptor instead. +func (*Any) Descriptor() ([]byte, []int) { + return file_google_protobuf_any_proto_rawDescGZIP(), []int{0} +} + +func (x *Any) GetTypeUrl() string { + if x != nil { + return x.TypeUrl + } + return "" +} + +func (x *Any) GetValue() []byte { + if x != nil { + return x.Value + } + return nil +} + +var File_google_protobuf_any_proto protoreflect.FileDescriptor + +var file_google_protobuf_any_proto_rawDesc = []byte{ + 0x0a, 0x19, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, + 0x66, 0x2f, 0x61, 0x6e, 0x79, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x12, 0x0f, 0x67, 0x6f, 0x6f, + 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x22, 0x36, 0x0a, 0x03, + 0x41, 0x6e, 0x79, 0x12, 0x19, 0x0a, 0x08, 0x74, 0x79, 0x70, 0x65, 0x5f, 0x75, 0x72, 0x6c, 0x18, + 0x01, 0x20, 0x01, 0x28, 0x09, 0x52, 0x07, 0x74, 0x79, 0x70, 0x65, 0x55, 0x72, 0x6c, 0x12, 0x14, + 0x0a, 0x05, 0x76, 0x61, 0x6c, 0x75, 0x65, 0x18, 0x02, 0x20, 0x01, 0x28, 0x0c, 0x52, 0x05, 0x76, + 0x61, 0x6c, 0x75, 0x65, 0x42, 0x6f, 0x0a, 0x13, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, + 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x42, 0x08, 0x41, 0x6e, 0x79, + 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, 0x25, 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, + 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, + 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, 0x70, 0x65, 0x73, 0x2f, 0x61, 0x6e, 0x79, 0xa2, 0x02, + 0x03, 0x47, 0x50, 0x42, 0xaa, 0x02, 0x1e, 0x47, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x50, 0x72, + 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x57, 0x65, 0x6c, 0x6c, 0x4b, 0x6e, 0x6f, 0x77, 0x6e, + 0x54, 0x79, 0x70, 0x65, 0x73, 0x62, 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33, +} + +var ( + file_google_protobuf_any_proto_rawDescOnce sync.Once + file_google_protobuf_any_proto_rawDescData = file_google_protobuf_any_proto_rawDesc +) + +func file_google_protobuf_any_proto_rawDescGZIP() []byte { + file_google_protobuf_any_proto_rawDescOnce.Do(func() { + file_google_protobuf_any_proto_rawDescData = protoimpl.X.CompressGZIP(file_google_protobuf_any_proto_rawDescData) + }) + return file_google_protobuf_any_proto_rawDescData +} + +var file_google_protobuf_any_proto_msgTypes = make([]protoimpl.MessageInfo, 1) +var file_google_protobuf_any_proto_goTypes = []interface{}{ + (*Any)(nil), // 0: google.protobuf.Any +} +var file_google_protobuf_any_proto_depIdxs = []int32{ + 0, // [0:0] is the sub-list for method output_type + 0, // [0:0] is the sub-list for method input_type + 0, // [0:0] is the sub-list for extension type_name + 0, // [0:0] is the sub-list for extension extendee + 0, // [0:0] is the sub-list for field type_name +} + +func init() { file_google_protobuf_any_proto_init() } +func file_google_protobuf_any_proto_init() { + if File_google_protobuf_any_proto != nil { + return + } + if !protoimpl.UnsafeEnabled { + file_google_protobuf_any_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Any); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + } + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_google_protobuf_any_proto_rawDesc, + NumEnums: 0, + NumMessages: 1, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_google_protobuf_any_proto_goTypes, + DependencyIndexes: file_google_protobuf_any_proto_depIdxs, + MessageInfos: file_google_protobuf_any_proto_msgTypes, + }.Build() + File_google_protobuf_any_proto = out.File + file_google_protobuf_any_proto_rawDesc = nil + file_google_protobuf_any_proto_goTypes = nil + file_google_protobuf_any_proto_depIdxs = nil +} diff --git a/vendor/google.golang.org/protobuf/types/known/durationpb/duration.pb.go b/vendor/google.golang.org/protobuf/types/known/durationpb/duration.pb.go new file mode 100644 index 00000000..3997c604 --- /dev/null +++ b/vendor/google.golang.org/protobuf/types/known/durationpb/duration.pb.go @@ -0,0 +1,249 @@ +// Protocol Buffers - Google's data interchange format +// Copyright 2008 Google Inc. All rights reserved. +// https://developers.google.com/protocol-buffers/ +// +// Redistribution and use in source and binary forms, with or without +// modification, are permitted provided that the following conditions are +// met: +// +// * Redistributions of source code must retain the above copyright +// notice, this list of conditions and the following disclaimer. +// * Redistributions in binary form must reproduce the above +// copyright notice, this list of conditions and the following disclaimer +// in the documentation and/or other materials provided with the +// distribution. +// * Neither the name of Google Inc. nor the names of its +// contributors may be used to endorse or promote products derived from +// this software without specific prior written permission. +// +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/protobuf/duration.proto + +package durationpb + +import ( + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + reflect "reflect" + sync "sync" +) + +// A Duration represents a signed, fixed-length span of time represented +// as a count of seconds and fractions of seconds at nanosecond +// resolution. It is independent of any calendar and concepts like "day" +// or "month". It is related to Timestamp in that the difference between +// two Timestamp values is a Duration and it can be added or subtracted +// from a Timestamp. Range is approximately +-10,000 years. +// +// # Examples +// +// Example 1: Compute Duration from two Timestamps in pseudo code. +// +// Timestamp start = ...; +// Timestamp end = ...; +// Duration duration = ...; +// +// duration.seconds = end.seconds - start.seconds; +// duration.nanos = end.nanos - start.nanos; +// +// if (duration.seconds < 0 && duration.nanos > 0) { +// duration.seconds += 1; +// duration.nanos -= 1000000000; +// } else if (duration.seconds > 0 && duration.nanos < 0) { +// duration.seconds -= 1; +// duration.nanos += 1000000000; +// } +// +// Example 2: Compute Timestamp from Timestamp + Duration in pseudo code. +// +// Timestamp start = ...; +// Duration duration = ...; +// Timestamp end = ...; +// +// end.seconds = start.seconds + duration.seconds; +// end.nanos = start.nanos + duration.nanos; +// +// if (end.nanos < 0) { +// end.seconds -= 1; +// end.nanos += 1000000000; +// } else if (end.nanos >= 1000000000) { +// end.seconds += 1; +// end.nanos -= 1000000000; +// } +// +// Example 3: Compute Duration from datetime.timedelta in Python. +// +// td = datetime.timedelta(days=3, minutes=10) +// duration = Duration() +// duration.FromTimedelta(td) +// +// # JSON Mapping +// +// In JSON format, the Duration type is encoded as a string rather than an +// object, where the string ends in the suffix "s" (indicating seconds) and +// is preceded by the number of seconds, with nanoseconds expressed as +// fractional seconds. For example, 3 seconds with 0 nanoseconds should be +// encoded in JSON format as "3s", while 3 seconds and 1 nanosecond should +// be expressed in JSON format as "3.000000001s", and 3 seconds and 1 +// microsecond should be expressed in JSON format as "3.000001s". +// +// +type Duration struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Signed seconds of the span of time. Must be from -315,576,000,000 + // to +315,576,000,000 inclusive. Note: these bounds are computed from: + // 60 sec/min * 60 min/hr * 24 hr/day * 365.25 days/year * 10000 years + Seconds int64 `protobuf:"varint,1,opt,name=seconds,proto3" json:"seconds,omitempty"` + // Signed fractions of a second at nanosecond resolution of the span + // of time. Durations less than one second are represented with a 0 + // `seconds` field and a positive or negative `nanos` field. For durations + // of one second or more, a non-zero value for the `nanos` field must be + // of the same sign as the `seconds` field. Must be from -999,999,999 + // to +999,999,999 inclusive. + Nanos int32 `protobuf:"varint,2,opt,name=nanos,proto3" json:"nanos,omitempty"` +} + +func (x *Duration) Reset() { + *x = Duration{} + if protoimpl.UnsafeEnabled { + mi := &file_google_protobuf_duration_proto_msgTypes[0] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Duration) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Duration) ProtoMessage() {} + +func (x *Duration) ProtoReflect() protoreflect.Message { + mi := &file_google_protobuf_duration_proto_msgTypes[0] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Duration.ProtoReflect.Descriptor instead. +func (*Duration) Descriptor() ([]byte, []int) { + return file_google_protobuf_duration_proto_rawDescGZIP(), []int{0} +} + +func (x *Duration) GetSeconds() int64 { + if x != nil { + return x.Seconds + } + return 0 +} + +func (x *Duration) GetNanos() int32 { + if x != nil { + return x.Nanos + } + return 0 +} + +var File_google_protobuf_duration_proto protoreflect.FileDescriptor + +var file_google_protobuf_duration_proto_rawDesc = []byte{ + 0x0a, 0x1e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, + 0x66, 0x2f, 0x64, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, + 0x12, 0x0f, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, + 0x66, 0x22, 0x3a, 0x0a, 0x08, 0x44, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x18, 0x0a, + 0x07, 0x73, 0x65, 0x63, 0x6f, 0x6e, 0x64, 0x73, 0x18, 0x01, 0x20, 0x01, 0x28, 0x03, 0x52, 0x07, + 0x73, 0x65, 0x63, 0x6f, 0x6e, 0x64, 0x73, 0x12, 0x14, 0x0a, 0x05, 0x6e, 0x61, 0x6e, 0x6f, 0x73, + 0x18, 0x02, 0x20, 0x01, 0x28, 0x05, 0x52, 0x05, 0x6e, 0x61, 0x6e, 0x6f, 0x73, 0x42, 0x7c, 0x0a, + 0x13, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, + 0x6f, 0x62, 0x75, 0x66, 0x42, 0x0d, 0x44, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x50, 0x72, + 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, 0x2a, 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, 0x2e, 0x63, 0x6f, + 0x6d, 0x2f, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, + 0x66, 0x2f, 0x70, 0x74, 0x79, 0x70, 0x65, 0x73, 0x2f, 0x64, 0x75, 0x72, 0x61, 0x74, 0x69, 0x6f, + 0x6e, 0xf8, 0x01, 0x01, 0xa2, 0x02, 0x03, 0x47, 0x50, 0x42, 0xaa, 0x02, 0x1e, 0x47, 0x6f, 0x6f, + 0x67, 0x6c, 0x65, 0x2e, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x2e, 0x57, 0x65, 0x6c, + 0x6c, 0x4b, 0x6e, 0x6f, 0x77, 0x6e, 0x54, 0x79, 0x70, 0x65, 0x73, 0x62, 0x06, 0x70, 0x72, 0x6f, + 0x74, 0x6f, 0x33, +} + +var ( + file_google_protobuf_duration_proto_rawDescOnce sync.Once + file_google_protobuf_duration_proto_rawDescData = file_google_protobuf_duration_proto_rawDesc +) + +func file_google_protobuf_duration_proto_rawDescGZIP() []byte { + file_google_protobuf_duration_proto_rawDescOnce.Do(func() { + file_google_protobuf_duration_proto_rawDescData = protoimpl.X.CompressGZIP(file_google_protobuf_duration_proto_rawDescData) + }) + return file_google_protobuf_duration_proto_rawDescData +} + +var file_google_protobuf_duration_proto_msgTypes = make([]protoimpl.MessageInfo, 1) +var file_google_protobuf_duration_proto_goTypes = []interface{}{ + (*Duration)(nil), // 0: google.protobuf.Duration +} +var file_google_protobuf_duration_proto_depIdxs = []int32{ + 0, // [0:0] is the sub-list for method output_type + 0, // [0:0] is the sub-list for method input_type + 0, // [0:0] is the sub-list for extension type_name + 0, // [0:0] is the sub-list for extension extendee + 0, // [0:0] is the sub-list for field type_name +} + +func init() { file_google_protobuf_duration_proto_init() } +func file_google_protobuf_duration_proto_init() { + if File_google_protobuf_duration_proto != nil { + return + } + if !protoimpl.UnsafeEnabled { + file_google_protobuf_duration_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Duration); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + } + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_google_protobuf_duration_proto_rawDesc, + NumEnums: 0, + NumMessages: 1, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_google_protobuf_duration_proto_goTypes, + DependencyIndexes: file_google_protobuf_duration_proto_depIdxs, + MessageInfos: file_google_protobuf_duration_proto_msgTypes, + }.Build() + File_google_protobuf_duration_proto = out.File + file_google_protobuf_duration_proto_rawDesc = nil + file_google_protobuf_duration_proto_goTypes = nil + file_google_protobuf_duration_proto_depIdxs = nil +} diff --git a/vendor/google.golang.org/protobuf/types/known/timestamppb/timestamp.pb.go b/vendor/google.golang.org/protobuf/types/known/timestamppb/timestamp.pb.go new file mode 100644 index 00000000..6fe6d42f --- /dev/null +++ b/vendor/google.golang.org/protobuf/types/known/timestamppb/timestamp.pb.go @@ -0,0 +1,271 @@ +// Protocol Buffers - Google's data interchange format +// Copyright 2008 Google Inc. All rights reserved. +// https://developers.google.com/protocol-buffers/ +// +// Redistribution and use in source and binary forms, with or without +// modification, are permitted provided that the following conditions are +// met: +// +// * Redistributions of source code must retain the above copyright +// notice, this list of conditions and the following disclaimer. +// * Redistributions in binary form must reproduce the above +// copyright notice, this list of conditions and the following disclaimer +// in the documentation and/or other materials provided with the +// distribution. +// * Neither the name of Google Inc. nor the names of its +// contributors may be used to endorse or promote products derived from +// this software without specific prior written permission. +// +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/protobuf/timestamp.proto + +package timestamppb + +import ( + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" + reflect "reflect" + sync "sync" +) + +// A Timestamp represents a point in time independent of any time zone or local +// calendar, encoded as a count of seconds and fractions of seconds at +// nanosecond resolution. The count is relative to an epoch at UTC midnight on +// January 1, 1970, in the proleptic Gregorian calendar which extends the +// Gregorian calendar backwards to year one. +// +// All minutes are 60 seconds long. Leap seconds are "smeared" so that no leap +// second table is needed for interpretation, using a [24-hour linear +// smear](https://developers.google.com/time/smear). +// +// The range is from 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z. By +// restricting to that range, we ensure that we can convert to and from [RFC +// 3339](https://www.ietf.org/rfc/rfc3339.txt) date strings. +// +// # Examples +// +// Example 1: Compute Timestamp from POSIX `time()`. +// +// Timestamp timestamp; +// timestamp.set_seconds(time(NULL)); +// timestamp.set_nanos(0); +// +// Example 2: Compute Timestamp from POSIX `gettimeofday()`. +// +// struct timeval tv; +// gettimeofday(&tv, NULL); +// +// Timestamp timestamp; +// timestamp.set_seconds(tv.tv_sec); +// timestamp.set_nanos(tv.tv_usec * 1000); +// +// Example 3: Compute Timestamp from Win32 `GetSystemTimeAsFileTime()`. +// +// FILETIME ft; +// GetSystemTimeAsFileTime(&ft); +// UINT64 ticks = (((UINT64)ft.dwHighDateTime) << 32) | ft.dwLowDateTime; +// +// // A Windows tick is 100 nanoseconds. Windows epoch 1601-01-01T00:00:00Z +// // is 11644473600 seconds before Unix epoch 1970-01-01T00:00:00Z. +// Timestamp timestamp; +// timestamp.set_seconds((INT64) ((ticks / 10000000) - 11644473600LL)); +// timestamp.set_nanos((INT32) ((ticks % 10000000) * 100)); +// +// Example 4: Compute Timestamp from Java `System.currentTimeMillis()`. +// +// long millis = System.currentTimeMillis(); +// +// Timestamp timestamp = Timestamp.newBuilder().setSeconds(millis / 1000) +// .setNanos((int) ((millis % 1000) * 1000000)).build(); +// +// +// Example 5: Compute Timestamp from current time in Python. +// +// timestamp = Timestamp() +// timestamp.GetCurrentTime() +// +// # JSON Mapping +// +// In JSON format, the Timestamp type is encoded as a string in the +// [RFC 3339](https://www.ietf.org/rfc/rfc3339.txt) format. That is, the +// format is "{year}-{month}-{day}T{hour}:{min}:{sec}[.{frac_sec}]Z" +// where {year} is always expressed using four digits while {month}, {day}, +// {hour}, {min}, and {sec} are zero-padded to two digits each. The fractional +// seconds, which can go up to 9 digits (i.e. up to 1 nanosecond resolution), +// are optional. The "Z" suffix indicates the timezone ("UTC"); the timezone +// is required. A proto3 JSON serializer should always use UTC (as indicated by +// "Z") when printing the Timestamp type and a proto3 JSON parser should be +// able to accept both UTC and other timezones (as indicated by an offset). +// +// For example, "2017-01-15T01:30:15.01Z" encodes 15.01 seconds past +// 01:30 UTC on January 15, 2017. +// +// In JavaScript, one can convert a Date object to this format using the +// standard +// [toISOString()](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Date/toISOString) +// method. In Python, a standard `datetime.datetime` object can be converted +// to this format using +// [`strftime`](https://docs.python.org/2/library/time.html#time.strftime) with +// the time format spec '%Y-%m-%dT%H:%M:%S.%fZ'. Likewise, in Java, one can use +// the Joda Time's [`ISODateTimeFormat.dateTime()`]( +// http://www.joda.org/joda-time/apidocs/org/joda/time/format/ISODateTimeFormat.html#dateTime%2D%2D +// ) to obtain a formatter capable of generating timestamps in this format. +// +// +type Timestamp struct { + state protoimpl.MessageState + sizeCache protoimpl.SizeCache + unknownFields protoimpl.UnknownFields + + // Represents seconds of UTC time since Unix epoch + // 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to + // 9999-12-31T23:59:59Z inclusive. + Seconds int64 `protobuf:"varint,1,opt,name=seconds,proto3" json:"seconds,omitempty"` + // Non-negative fractions of a second at nanosecond resolution. Negative + // second values with fractions must still have non-negative nanos values + // that count forward in time. Must be from 0 to 999,999,999 + // inclusive. + Nanos int32 `protobuf:"varint,2,opt,name=nanos,proto3" json:"nanos,omitempty"` +} + +func (x *Timestamp) Reset() { + *x = Timestamp{} + if protoimpl.UnsafeEnabled { + mi := &file_google_protobuf_timestamp_proto_msgTypes[0] + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + ms.StoreMessageInfo(mi) + } +} + +func (x *Timestamp) String() string { + return protoimpl.X.MessageStringOf(x) +} + +func (*Timestamp) ProtoMessage() {} + +func (x *Timestamp) ProtoReflect() protoreflect.Message { + mi := &file_google_protobuf_timestamp_proto_msgTypes[0] + if protoimpl.UnsafeEnabled && x != nil { + ms := protoimpl.X.MessageStateOf(protoimpl.Pointer(x)) + if ms.LoadMessageInfo() == nil { + ms.StoreMessageInfo(mi) + } + return ms + } + return mi.MessageOf(x) +} + +// Deprecated: Use Timestamp.ProtoReflect.Descriptor instead. +func (*Timestamp) Descriptor() ([]byte, []int) { + return file_google_protobuf_timestamp_proto_rawDescGZIP(), []int{0} +} + +func (x *Timestamp) GetSeconds() int64 { + if x != nil { + return x.Seconds + } + return 0 +} + +func (x *Timestamp) GetNanos() int32 { + if x != nil { + return x.Nanos + } + return 0 +} + +var File_google_protobuf_timestamp_proto protoreflect.FileDescriptor + +var file_google_protobuf_timestamp_proto_rawDesc = []byte{ + 0x0a, 0x1f, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2f, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, + 0x66, 0x2f, 0x74, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x2e, 0x70, 0x72, 0x6f, 0x74, + 0x6f, 0x12, 0x0f, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x62, + 0x75, 0x66, 0x22, 0x3b, 0x0a, 0x09, 0x54, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, 0x70, 0x12, + 0x18, 0x0a, 0x07, 0x73, 0x65, 0x63, 0x6f, 0x6e, 0x64, 0x73, 0x18, 0x01, 0x20, 0x01, 0x28, 0x03, + 0x52, 0x07, 0x73, 0x65, 0x63, 0x6f, 0x6e, 0x64, 0x73, 0x12, 0x14, 0x0a, 0x05, 0x6e, 0x61, 0x6e, + 0x6f, 0x73, 0x18, 0x02, 0x20, 0x01, 0x28, 0x05, 0x52, 0x05, 0x6e, 0x61, 0x6e, 0x6f, 0x73, 0x42, + 0x7e, 0x0a, 0x13, 0x63, 0x6f, 0x6d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x70, 0x72, + 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, 0x42, 0x0e, 0x54, 0x69, 0x6d, 0x65, 0x73, 0x74, 0x61, 0x6d, + 0x70, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x50, 0x01, 0x5a, 0x2b, 0x67, 0x69, 0x74, 0x68, 0x75, 0x62, + 0x2e, 0x63, 0x6f, 0x6d, 0x2f, 0x67, 0x6f, 0x6c, 0x61, 0x6e, 0x67, 0x2f, 0x70, 0x72, 0x6f, 0x74, + 0x6f, 0x62, 0x75, 0x66, 0x2f, 0x70, 0x74, 0x79, 0x70, 0x65, 0x73, 0x2f, 0x74, 0x69, 0x6d, 0x65, + 0x73, 0x74, 0x61, 0x6d, 0x70, 0xf8, 0x01, 0x01, 0xa2, 0x02, 0x03, 0x47, 0x50, 0x42, 0xaa, 0x02, + 0x1e, 0x47, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x50, 0x72, 0x6f, 0x74, 0x6f, 0x62, 0x75, 0x66, + 0x2e, 0x57, 0x65, 0x6c, 0x6c, 0x4b, 0x6e, 0x6f, 0x77, 0x6e, 0x54, 0x79, 0x70, 0x65, 0x73, 0x62, + 0x06, 0x70, 0x72, 0x6f, 0x74, 0x6f, 0x33, +} + +var ( + file_google_protobuf_timestamp_proto_rawDescOnce sync.Once + file_google_protobuf_timestamp_proto_rawDescData = file_google_protobuf_timestamp_proto_rawDesc +) + +func file_google_protobuf_timestamp_proto_rawDescGZIP() []byte { + file_google_protobuf_timestamp_proto_rawDescOnce.Do(func() { + file_google_protobuf_timestamp_proto_rawDescData = protoimpl.X.CompressGZIP(file_google_protobuf_timestamp_proto_rawDescData) + }) + return file_google_protobuf_timestamp_proto_rawDescData +} + +var file_google_protobuf_timestamp_proto_msgTypes = make([]protoimpl.MessageInfo, 1) +var file_google_protobuf_timestamp_proto_goTypes = []interface{}{ + (*Timestamp)(nil), // 0: google.protobuf.Timestamp +} +var file_google_protobuf_timestamp_proto_depIdxs = []int32{ + 0, // [0:0] is the sub-list for method output_type + 0, // [0:0] is the sub-list for method input_type + 0, // [0:0] is the sub-list for extension type_name + 0, // [0:0] is the sub-list for extension extendee + 0, // [0:0] is the sub-list for field type_name +} + +func init() { file_google_protobuf_timestamp_proto_init() } +func file_google_protobuf_timestamp_proto_init() { + if File_google_protobuf_timestamp_proto != nil { + return + } + if !protoimpl.UnsafeEnabled { + file_google_protobuf_timestamp_proto_msgTypes[0].Exporter = func(v interface{}, i int) interface{} { + switch v := v.(*Timestamp); i { + case 0: + return &v.state + case 1: + return &v.sizeCache + case 2: + return &v.unknownFields + default: + return nil + } + } + } + type x struct{} + out := protoimpl.TypeBuilder{ + File: protoimpl.DescBuilder{ + GoPackagePath: reflect.TypeOf(x{}).PkgPath(), + RawDescriptor: file_google_protobuf_timestamp_proto_rawDesc, + NumEnums: 0, + NumMessages: 1, + NumExtensions: 0, + NumServices: 0, + }, + GoTypes: file_google_protobuf_timestamp_proto_goTypes, + DependencyIndexes: file_google_protobuf_timestamp_proto_depIdxs, + MessageInfos: file_google_protobuf_timestamp_proto_msgTypes, + }.Build() + File_google_protobuf_timestamp_proto = out.File + file_google_protobuf_timestamp_proto_rawDesc = nil + file_google_protobuf_timestamp_proto_goTypes = nil + file_google_protobuf_timestamp_proto_depIdxs = nil +} diff --git a/vendor/gopkg.in/yaml.v3/.travis.yml b/vendor/gopkg.in/yaml.v3/.travis.yml new file mode 100644 index 00000000..a130fe88 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/.travis.yml @@ -0,0 +1,17 @@ +language: go + +go: + - "1.4.x" + - "1.5.x" + - "1.6.x" + - "1.7.x" + - "1.8.x" + - "1.9.x" + - "1.10.x" + - "1.11.x" + - "1.12.x" + - "1.13.x" + - "1.14.x" + - "tip" + +go_import_path: gopkg.in/yaml.v3 diff --git a/vendor/gopkg.in/yaml.v3/LICENSE b/vendor/gopkg.in/yaml.v3/LICENSE new file mode 100644 index 00000000..2683e4bb --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/LICENSE @@ -0,0 +1,50 @@ + +This project is covered by two different licenses: MIT and Apache. + +#### MIT License #### + +The following files were ported to Go from C files of libyaml, and thus +are still covered by their original MIT license, with the additional +copyright staring in 2011 when the project was ported over: + + apic.go emitterc.go parserc.go readerc.go scannerc.go + writerc.go yamlh.go yamlprivateh.go + +Copyright (c) 2006-2010 Kirill Simonov +Copyright (c) 2006-2011 Kirill Simonov + +Permission is hereby granted, free of charge, to any person obtaining a copy of +this software and associated documentation files (the "Software"), to deal in +the Software without restriction, including without limitation the rights to +use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +of the Software, and to permit persons to whom the Software is furnished to do +so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in all +copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +SOFTWARE. + +### Apache License ### + +All the remaining project files are covered by the Apache license: + +Copyright (c) 2011-2019 Canonical Ltd + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. diff --git a/vendor/gopkg.in/yaml.v3/NOTICE b/vendor/gopkg.in/yaml.v3/NOTICE new file mode 100644 index 00000000..866d74a7 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/NOTICE @@ -0,0 +1,13 @@ +Copyright 2011-2016 Canonical Ltd. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. diff --git a/vendor/gopkg.in/yaml.v3/README.md b/vendor/gopkg.in/yaml.v3/README.md new file mode 100644 index 00000000..08eb1bab --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/README.md @@ -0,0 +1,150 @@ +# YAML support for the Go language + +Introduction +------------ + +The yaml package enables Go programs to comfortably encode and decode YAML +values. It was developed within [Canonical](https://www.canonical.com) as +part of the [juju](https://juju.ubuntu.com) project, and is based on a +pure Go port of the well-known [libyaml](http://pyyaml.org/wiki/LibYAML) +C library to parse and generate YAML data quickly and reliably. + +Compatibility +------------- + +The yaml package supports most of YAML 1.2, but preserves some behavior +from 1.1 for backwards compatibility. + +Specifically, as of v3 of the yaml package: + + - YAML 1.1 bools (_yes/no, on/off_) are supported as long as they are being + decoded into a typed bool value. Otherwise they behave as a string. Booleans + in YAML 1.2 are _true/false_ only. + - Octals encode and decode as _0777_ per YAML 1.1, rather than _0o777_ + as specified in YAML 1.2, because most parsers still use the old format. + Octals in the _0o777_ format are supported though, so new files work. + - Does not support base-60 floats. These are gone from YAML 1.2, and were + actually never supported by this package as it's clearly a poor choice. + +and offers backwards +compatibility with YAML 1.1 in some cases. +1.2, including support for +anchors, tags, map merging, etc. Multi-document unmarshalling is not yet +implemented, and base-60 floats from YAML 1.1 are purposefully not +supported since they're a poor design and are gone in YAML 1.2. + +Installation and usage +---------------------- + +The import path for the package is *gopkg.in/yaml.v3*. + +To install it, run: + + go get gopkg.in/yaml.v3 + +API documentation +----------------- + +If opened in a browser, the import path itself leads to the API documentation: + + - [https://gopkg.in/yaml.v3](https://gopkg.in/yaml.v3) + +API stability +------------- + +The package API for yaml v3 will remain stable as described in [gopkg.in](https://gopkg.in). + + +License +------- + +The yaml package is licensed under the MIT and Apache License 2.0 licenses. +Please see the LICENSE file for details. + + +Example +------- + +```Go +package main + +import ( + "fmt" + "log" + + "gopkg.in/yaml.v3" +) + +var data = ` +a: Easy! +b: + c: 2 + d: [3, 4] +` + +// Note: struct fields must be public in order for unmarshal to +// correctly populate the data. +type T struct { + A string + B struct { + RenamedC int `yaml:"c"` + D []int `yaml:",flow"` + } +} + +func main() { + t := T{} + + err := yaml.Unmarshal([]byte(data), &t) + if err != nil { + log.Fatalf("error: %v", err) + } + fmt.Printf("--- t:\n%v\n\n", t) + + d, err := yaml.Marshal(&t) + if err != nil { + log.Fatalf("error: %v", err) + } + fmt.Printf("--- t dump:\n%s\n\n", string(d)) + + m := make(map[interface{}]interface{}) + + err = yaml.Unmarshal([]byte(data), &m) + if err != nil { + log.Fatalf("error: %v", err) + } + fmt.Printf("--- m:\n%v\n\n", m) + + d, err = yaml.Marshal(&m) + if err != nil { + log.Fatalf("error: %v", err) + } + fmt.Printf("--- m dump:\n%s\n\n", string(d)) +} +``` + +This example will generate the following output: + +``` +--- t: +{Easy! {2 [3 4]}} + +--- t dump: +a: Easy! +b: + c: 2 + d: [3, 4] + + +--- m: +map[a:Easy! b:map[c:2 d:[3 4]]] + +--- m dump: +a: Easy! +b: + c: 2 + d: + - 3 + - 4 +``` + diff --git a/vendor/gopkg.in/yaml.v3/apic.go b/vendor/gopkg.in/yaml.v3/apic.go new file mode 100644 index 00000000..ae7d049f --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/apic.go @@ -0,0 +1,747 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +import ( + "io" +) + +func yaml_insert_token(parser *yaml_parser_t, pos int, token *yaml_token_t) { + //fmt.Println("yaml_insert_token", "pos:", pos, "typ:", token.typ, "head:", parser.tokens_head, "len:", len(parser.tokens)) + + // Check if we can move the queue at the beginning of the buffer. + if parser.tokens_head > 0 && len(parser.tokens) == cap(parser.tokens) { + if parser.tokens_head != len(parser.tokens) { + copy(parser.tokens, parser.tokens[parser.tokens_head:]) + } + parser.tokens = parser.tokens[:len(parser.tokens)-parser.tokens_head] + parser.tokens_head = 0 + } + parser.tokens = append(parser.tokens, *token) + if pos < 0 { + return + } + copy(parser.tokens[parser.tokens_head+pos+1:], parser.tokens[parser.tokens_head+pos:]) + parser.tokens[parser.tokens_head+pos] = *token +} + +// Create a new parser object. +func yaml_parser_initialize(parser *yaml_parser_t) bool { + *parser = yaml_parser_t{ + raw_buffer: make([]byte, 0, input_raw_buffer_size), + buffer: make([]byte, 0, input_buffer_size), + } + return true +} + +// Destroy a parser object. +func yaml_parser_delete(parser *yaml_parser_t) { + *parser = yaml_parser_t{} +} + +// String read handler. +func yaml_string_read_handler(parser *yaml_parser_t, buffer []byte) (n int, err error) { + if parser.input_pos == len(parser.input) { + return 0, io.EOF + } + n = copy(buffer, parser.input[parser.input_pos:]) + parser.input_pos += n + return n, nil +} + +// Reader read handler. +func yaml_reader_read_handler(parser *yaml_parser_t, buffer []byte) (n int, err error) { + return parser.input_reader.Read(buffer) +} + +// Set a string input. +func yaml_parser_set_input_string(parser *yaml_parser_t, input []byte) { + if parser.read_handler != nil { + panic("must set the input source only once") + } + parser.read_handler = yaml_string_read_handler + parser.input = input + parser.input_pos = 0 +} + +// Set a file input. +func yaml_parser_set_input_reader(parser *yaml_parser_t, r io.Reader) { + if parser.read_handler != nil { + panic("must set the input source only once") + } + parser.read_handler = yaml_reader_read_handler + parser.input_reader = r +} + +// Set the source encoding. +func yaml_parser_set_encoding(parser *yaml_parser_t, encoding yaml_encoding_t) { + if parser.encoding != yaml_ANY_ENCODING { + panic("must set the encoding only once") + } + parser.encoding = encoding +} + +// Create a new emitter object. +func yaml_emitter_initialize(emitter *yaml_emitter_t) { + *emitter = yaml_emitter_t{ + buffer: make([]byte, output_buffer_size), + raw_buffer: make([]byte, 0, output_raw_buffer_size), + states: make([]yaml_emitter_state_t, 0, initial_stack_size), + events: make([]yaml_event_t, 0, initial_queue_size), + best_width: -1, + } +} + +// Destroy an emitter object. +func yaml_emitter_delete(emitter *yaml_emitter_t) { + *emitter = yaml_emitter_t{} +} + +// String write handler. +func yaml_string_write_handler(emitter *yaml_emitter_t, buffer []byte) error { + *emitter.output_buffer = append(*emitter.output_buffer, buffer...) + return nil +} + +// yaml_writer_write_handler uses emitter.output_writer to write the +// emitted text. +func yaml_writer_write_handler(emitter *yaml_emitter_t, buffer []byte) error { + _, err := emitter.output_writer.Write(buffer) + return err +} + +// Set a string output. +func yaml_emitter_set_output_string(emitter *yaml_emitter_t, output_buffer *[]byte) { + if emitter.write_handler != nil { + panic("must set the output target only once") + } + emitter.write_handler = yaml_string_write_handler + emitter.output_buffer = output_buffer +} + +// Set a file output. +func yaml_emitter_set_output_writer(emitter *yaml_emitter_t, w io.Writer) { + if emitter.write_handler != nil { + panic("must set the output target only once") + } + emitter.write_handler = yaml_writer_write_handler + emitter.output_writer = w +} + +// Set the output encoding. +func yaml_emitter_set_encoding(emitter *yaml_emitter_t, encoding yaml_encoding_t) { + if emitter.encoding != yaml_ANY_ENCODING { + panic("must set the output encoding only once") + } + emitter.encoding = encoding +} + +// Set the canonical output style. +func yaml_emitter_set_canonical(emitter *yaml_emitter_t, canonical bool) { + emitter.canonical = canonical +} + +// Set the indentation increment. +func yaml_emitter_set_indent(emitter *yaml_emitter_t, indent int) { + if indent < 2 || indent > 9 { + indent = 2 + } + emitter.best_indent = indent +} + +// Set the preferred line width. +func yaml_emitter_set_width(emitter *yaml_emitter_t, width int) { + if width < 0 { + width = -1 + } + emitter.best_width = width +} + +// Set if unescaped non-ASCII characters are allowed. +func yaml_emitter_set_unicode(emitter *yaml_emitter_t, unicode bool) { + emitter.unicode = unicode +} + +// Set the preferred line break character. +func yaml_emitter_set_break(emitter *yaml_emitter_t, line_break yaml_break_t) { + emitter.line_break = line_break +} + +///* +// * Destroy a token object. +// */ +// +//YAML_DECLARE(void) +//yaml_token_delete(yaml_token_t *token) +//{ +// assert(token); // Non-NULL token object expected. +// +// switch (token.type) +// { +// case YAML_TAG_DIRECTIVE_TOKEN: +// yaml_free(token.data.tag_directive.handle); +// yaml_free(token.data.tag_directive.prefix); +// break; +// +// case YAML_ALIAS_TOKEN: +// yaml_free(token.data.alias.value); +// break; +// +// case YAML_ANCHOR_TOKEN: +// yaml_free(token.data.anchor.value); +// break; +// +// case YAML_TAG_TOKEN: +// yaml_free(token.data.tag.handle); +// yaml_free(token.data.tag.suffix); +// break; +// +// case YAML_SCALAR_TOKEN: +// yaml_free(token.data.scalar.value); +// break; +// +// default: +// break; +// } +// +// memset(token, 0, sizeof(yaml_token_t)); +//} +// +///* +// * Check if a string is a valid UTF-8 sequence. +// * +// * Check 'reader.c' for more details on UTF-8 encoding. +// */ +// +//static int +//yaml_check_utf8(yaml_char_t *start, size_t length) +//{ +// yaml_char_t *end = start+length; +// yaml_char_t *pointer = start; +// +// while (pointer < end) { +// unsigned char octet; +// unsigned int width; +// unsigned int value; +// size_t k; +// +// octet = pointer[0]; +// width = (octet & 0x80) == 0x00 ? 1 : +// (octet & 0xE0) == 0xC0 ? 2 : +// (octet & 0xF0) == 0xE0 ? 3 : +// (octet & 0xF8) == 0xF0 ? 4 : 0; +// value = (octet & 0x80) == 0x00 ? octet & 0x7F : +// (octet & 0xE0) == 0xC0 ? octet & 0x1F : +// (octet & 0xF0) == 0xE0 ? octet & 0x0F : +// (octet & 0xF8) == 0xF0 ? octet & 0x07 : 0; +// if (!width) return 0; +// if (pointer+width > end) return 0; +// for (k = 1; k < width; k ++) { +// octet = pointer[k]; +// if ((octet & 0xC0) != 0x80) return 0; +// value = (value << 6) + (octet & 0x3F); +// } +// if (!((width == 1) || +// (width == 2 && value >= 0x80) || +// (width == 3 && value >= 0x800) || +// (width == 4 && value >= 0x10000))) return 0; +// +// pointer += width; +// } +// +// return 1; +//} +// + +// Create STREAM-START. +func yaml_stream_start_event_initialize(event *yaml_event_t, encoding yaml_encoding_t) { + *event = yaml_event_t{ + typ: yaml_STREAM_START_EVENT, + encoding: encoding, + } +} + +// Create STREAM-END. +func yaml_stream_end_event_initialize(event *yaml_event_t) { + *event = yaml_event_t{ + typ: yaml_STREAM_END_EVENT, + } +} + +// Create DOCUMENT-START. +func yaml_document_start_event_initialize( + event *yaml_event_t, + version_directive *yaml_version_directive_t, + tag_directives []yaml_tag_directive_t, + implicit bool, +) { + *event = yaml_event_t{ + typ: yaml_DOCUMENT_START_EVENT, + version_directive: version_directive, + tag_directives: tag_directives, + implicit: implicit, + } +} + +// Create DOCUMENT-END. +func yaml_document_end_event_initialize(event *yaml_event_t, implicit bool) { + *event = yaml_event_t{ + typ: yaml_DOCUMENT_END_EVENT, + implicit: implicit, + } +} + +// Create ALIAS. +func yaml_alias_event_initialize(event *yaml_event_t, anchor []byte) bool { + *event = yaml_event_t{ + typ: yaml_ALIAS_EVENT, + anchor: anchor, + } + return true +} + +// Create SCALAR. +func yaml_scalar_event_initialize(event *yaml_event_t, anchor, tag, value []byte, plain_implicit, quoted_implicit bool, style yaml_scalar_style_t) bool { + *event = yaml_event_t{ + typ: yaml_SCALAR_EVENT, + anchor: anchor, + tag: tag, + value: value, + implicit: plain_implicit, + quoted_implicit: quoted_implicit, + style: yaml_style_t(style), + } + return true +} + +// Create SEQUENCE-START. +func yaml_sequence_start_event_initialize(event *yaml_event_t, anchor, tag []byte, implicit bool, style yaml_sequence_style_t) bool { + *event = yaml_event_t{ + typ: yaml_SEQUENCE_START_EVENT, + anchor: anchor, + tag: tag, + implicit: implicit, + style: yaml_style_t(style), + } + return true +} + +// Create SEQUENCE-END. +func yaml_sequence_end_event_initialize(event *yaml_event_t) bool { + *event = yaml_event_t{ + typ: yaml_SEQUENCE_END_EVENT, + } + return true +} + +// Create MAPPING-START. +func yaml_mapping_start_event_initialize(event *yaml_event_t, anchor, tag []byte, implicit bool, style yaml_mapping_style_t) { + *event = yaml_event_t{ + typ: yaml_MAPPING_START_EVENT, + anchor: anchor, + tag: tag, + implicit: implicit, + style: yaml_style_t(style), + } +} + +// Create MAPPING-END. +func yaml_mapping_end_event_initialize(event *yaml_event_t) { + *event = yaml_event_t{ + typ: yaml_MAPPING_END_EVENT, + } +} + +// Destroy an event object. +func yaml_event_delete(event *yaml_event_t) { + *event = yaml_event_t{} +} + +///* +// * Create a document object. +// */ +// +//YAML_DECLARE(int) +//yaml_document_initialize(document *yaml_document_t, +// version_directive *yaml_version_directive_t, +// tag_directives_start *yaml_tag_directive_t, +// tag_directives_end *yaml_tag_directive_t, +// start_implicit int, end_implicit int) +//{ +// struct { +// error yaml_error_type_t +// } context +// struct { +// start *yaml_node_t +// end *yaml_node_t +// top *yaml_node_t +// } nodes = { NULL, NULL, NULL } +// version_directive_copy *yaml_version_directive_t = NULL +// struct { +// start *yaml_tag_directive_t +// end *yaml_tag_directive_t +// top *yaml_tag_directive_t +// } tag_directives_copy = { NULL, NULL, NULL } +// value yaml_tag_directive_t = { NULL, NULL } +// mark yaml_mark_t = { 0, 0, 0 } +// +// assert(document) // Non-NULL document object is expected. +// assert((tag_directives_start && tag_directives_end) || +// (tag_directives_start == tag_directives_end)) +// // Valid tag directives are expected. +// +// if (!STACK_INIT(&context, nodes, INITIAL_STACK_SIZE)) goto error +// +// if (version_directive) { +// version_directive_copy = yaml_malloc(sizeof(yaml_version_directive_t)) +// if (!version_directive_copy) goto error +// version_directive_copy.major = version_directive.major +// version_directive_copy.minor = version_directive.minor +// } +// +// if (tag_directives_start != tag_directives_end) { +// tag_directive *yaml_tag_directive_t +// if (!STACK_INIT(&context, tag_directives_copy, INITIAL_STACK_SIZE)) +// goto error +// for (tag_directive = tag_directives_start +// tag_directive != tag_directives_end; tag_directive ++) { +// assert(tag_directive.handle) +// assert(tag_directive.prefix) +// if (!yaml_check_utf8(tag_directive.handle, +// strlen((char *)tag_directive.handle))) +// goto error +// if (!yaml_check_utf8(tag_directive.prefix, +// strlen((char *)tag_directive.prefix))) +// goto error +// value.handle = yaml_strdup(tag_directive.handle) +// value.prefix = yaml_strdup(tag_directive.prefix) +// if (!value.handle || !value.prefix) goto error +// if (!PUSH(&context, tag_directives_copy, value)) +// goto error +// value.handle = NULL +// value.prefix = NULL +// } +// } +// +// DOCUMENT_INIT(*document, nodes.start, nodes.end, version_directive_copy, +// tag_directives_copy.start, tag_directives_copy.top, +// start_implicit, end_implicit, mark, mark) +// +// return 1 +// +//error: +// STACK_DEL(&context, nodes) +// yaml_free(version_directive_copy) +// while (!STACK_EMPTY(&context, tag_directives_copy)) { +// value yaml_tag_directive_t = POP(&context, tag_directives_copy) +// yaml_free(value.handle) +// yaml_free(value.prefix) +// } +// STACK_DEL(&context, tag_directives_copy) +// yaml_free(value.handle) +// yaml_free(value.prefix) +// +// return 0 +//} +// +///* +// * Destroy a document object. +// */ +// +//YAML_DECLARE(void) +//yaml_document_delete(document *yaml_document_t) +//{ +// struct { +// error yaml_error_type_t +// } context +// tag_directive *yaml_tag_directive_t +// +// context.error = YAML_NO_ERROR // Eliminate a compiler warning. +// +// assert(document) // Non-NULL document object is expected. +// +// while (!STACK_EMPTY(&context, document.nodes)) { +// node yaml_node_t = POP(&context, document.nodes) +// yaml_free(node.tag) +// switch (node.type) { +// case YAML_SCALAR_NODE: +// yaml_free(node.data.scalar.value) +// break +// case YAML_SEQUENCE_NODE: +// STACK_DEL(&context, node.data.sequence.items) +// break +// case YAML_MAPPING_NODE: +// STACK_DEL(&context, node.data.mapping.pairs) +// break +// default: +// assert(0) // Should not happen. +// } +// } +// STACK_DEL(&context, document.nodes) +// +// yaml_free(document.version_directive) +// for (tag_directive = document.tag_directives.start +// tag_directive != document.tag_directives.end +// tag_directive++) { +// yaml_free(tag_directive.handle) +// yaml_free(tag_directive.prefix) +// } +// yaml_free(document.tag_directives.start) +// +// memset(document, 0, sizeof(yaml_document_t)) +//} +// +///** +// * Get a document node. +// */ +// +//YAML_DECLARE(yaml_node_t *) +//yaml_document_get_node(document *yaml_document_t, index int) +//{ +// assert(document) // Non-NULL document object is expected. +// +// if (index > 0 && document.nodes.start + index <= document.nodes.top) { +// return document.nodes.start + index - 1 +// } +// return NULL +//} +// +///** +// * Get the root object. +// */ +// +//YAML_DECLARE(yaml_node_t *) +//yaml_document_get_root_node(document *yaml_document_t) +//{ +// assert(document) // Non-NULL document object is expected. +// +// if (document.nodes.top != document.nodes.start) { +// return document.nodes.start +// } +// return NULL +//} +// +///* +// * Add a scalar node to a document. +// */ +// +//YAML_DECLARE(int) +//yaml_document_add_scalar(document *yaml_document_t, +// tag *yaml_char_t, value *yaml_char_t, length int, +// style yaml_scalar_style_t) +//{ +// struct { +// error yaml_error_type_t +// } context +// mark yaml_mark_t = { 0, 0, 0 } +// tag_copy *yaml_char_t = NULL +// value_copy *yaml_char_t = NULL +// node yaml_node_t +// +// assert(document) // Non-NULL document object is expected. +// assert(value) // Non-NULL value is expected. +// +// if (!tag) { +// tag = (yaml_char_t *)YAML_DEFAULT_SCALAR_TAG +// } +// +// if (!yaml_check_utf8(tag, strlen((char *)tag))) goto error +// tag_copy = yaml_strdup(tag) +// if (!tag_copy) goto error +// +// if (length < 0) { +// length = strlen((char *)value) +// } +// +// if (!yaml_check_utf8(value, length)) goto error +// value_copy = yaml_malloc(length+1) +// if (!value_copy) goto error +// memcpy(value_copy, value, length) +// value_copy[length] = '\0' +// +// SCALAR_NODE_INIT(node, tag_copy, value_copy, length, style, mark, mark) +// if (!PUSH(&context, document.nodes, node)) goto error +// +// return document.nodes.top - document.nodes.start +// +//error: +// yaml_free(tag_copy) +// yaml_free(value_copy) +// +// return 0 +//} +// +///* +// * Add a sequence node to a document. +// */ +// +//YAML_DECLARE(int) +//yaml_document_add_sequence(document *yaml_document_t, +// tag *yaml_char_t, style yaml_sequence_style_t) +//{ +// struct { +// error yaml_error_type_t +// } context +// mark yaml_mark_t = { 0, 0, 0 } +// tag_copy *yaml_char_t = NULL +// struct { +// start *yaml_node_item_t +// end *yaml_node_item_t +// top *yaml_node_item_t +// } items = { NULL, NULL, NULL } +// node yaml_node_t +// +// assert(document) // Non-NULL document object is expected. +// +// if (!tag) { +// tag = (yaml_char_t *)YAML_DEFAULT_SEQUENCE_TAG +// } +// +// if (!yaml_check_utf8(tag, strlen((char *)tag))) goto error +// tag_copy = yaml_strdup(tag) +// if (!tag_copy) goto error +// +// if (!STACK_INIT(&context, items, INITIAL_STACK_SIZE)) goto error +// +// SEQUENCE_NODE_INIT(node, tag_copy, items.start, items.end, +// style, mark, mark) +// if (!PUSH(&context, document.nodes, node)) goto error +// +// return document.nodes.top - document.nodes.start +// +//error: +// STACK_DEL(&context, items) +// yaml_free(tag_copy) +// +// return 0 +//} +// +///* +// * Add a mapping node to a document. +// */ +// +//YAML_DECLARE(int) +//yaml_document_add_mapping(document *yaml_document_t, +// tag *yaml_char_t, style yaml_mapping_style_t) +//{ +// struct { +// error yaml_error_type_t +// } context +// mark yaml_mark_t = { 0, 0, 0 } +// tag_copy *yaml_char_t = NULL +// struct { +// start *yaml_node_pair_t +// end *yaml_node_pair_t +// top *yaml_node_pair_t +// } pairs = { NULL, NULL, NULL } +// node yaml_node_t +// +// assert(document) // Non-NULL document object is expected. +// +// if (!tag) { +// tag = (yaml_char_t *)YAML_DEFAULT_MAPPING_TAG +// } +// +// if (!yaml_check_utf8(tag, strlen((char *)tag))) goto error +// tag_copy = yaml_strdup(tag) +// if (!tag_copy) goto error +// +// if (!STACK_INIT(&context, pairs, INITIAL_STACK_SIZE)) goto error +// +// MAPPING_NODE_INIT(node, tag_copy, pairs.start, pairs.end, +// style, mark, mark) +// if (!PUSH(&context, document.nodes, node)) goto error +// +// return document.nodes.top - document.nodes.start +// +//error: +// STACK_DEL(&context, pairs) +// yaml_free(tag_copy) +// +// return 0 +//} +// +///* +// * Append an item to a sequence node. +// */ +// +//YAML_DECLARE(int) +//yaml_document_append_sequence_item(document *yaml_document_t, +// sequence int, item int) +//{ +// struct { +// error yaml_error_type_t +// } context +// +// assert(document) // Non-NULL document is required. +// assert(sequence > 0 +// && document.nodes.start + sequence <= document.nodes.top) +// // Valid sequence id is required. +// assert(document.nodes.start[sequence-1].type == YAML_SEQUENCE_NODE) +// // A sequence node is required. +// assert(item > 0 && document.nodes.start + item <= document.nodes.top) +// // Valid item id is required. +// +// if (!PUSH(&context, +// document.nodes.start[sequence-1].data.sequence.items, item)) +// return 0 +// +// return 1 +//} +// +///* +// * Append a pair of a key and a value to a mapping node. +// */ +// +//YAML_DECLARE(int) +//yaml_document_append_mapping_pair(document *yaml_document_t, +// mapping int, key int, value int) +//{ +// struct { +// error yaml_error_type_t +// } context +// +// pair yaml_node_pair_t +// +// assert(document) // Non-NULL document is required. +// assert(mapping > 0 +// && document.nodes.start + mapping <= document.nodes.top) +// // Valid mapping id is required. +// assert(document.nodes.start[mapping-1].type == YAML_MAPPING_NODE) +// // A mapping node is required. +// assert(key > 0 && document.nodes.start + key <= document.nodes.top) +// // Valid key id is required. +// assert(value > 0 && document.nodes.start + value <= document.nodes.top) +// // Valid value id is required. +// +// pair.key = key +// pair.value = value +// +// if (!PUSH(&context, +// document.nodes.start[mapping-1].data.mapping.pairs, pair)) +// return 0 +// +// return 1 +//} +// +// diff --git a/vendor/gopkg.in/yaml.v3/decode.go b/vendor/gopkg.in/yaml.v3/decode.go new file mode 100644 index 00000000..21c0dacf --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/decode.go @@ -0,0 +1,948 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package yaml + +import ( + "encoding" + "encoding/base64" + "fmt" + "io" + "math" + "reflect" + "strconv" + "time" +) + +// ---------------------------------------------------------------------------- +// Parser, produces a node tree out of a libyaml event stream. + +type parser struct { + parser yaml_parser_t + event yaml_event_t + doc *Node + anchors map[string]*Node + doneInit bool + textless bool +} + +func newParser(b []byte) *parser { + p := parser{} + if !yaml_parser_initialize(&p.parser) { + panic("failed to initialize YAML emitter") + } + if len(b) == 0 { + b = []byte{'\n'} + } + yaml_parser_set_input_string(&p.parser, b) + return &p +} + +func newParserFromReader(r io.Reader) *parser { + p := parser{} + if !yaml_parser_initialize(&p.parser) { + panic("failed to initialize YAML emitter") + } + yaml_parser_set_input_reader(&p.parser, r) + return &p +} + +func (p *parser) init() { + if p.doneInit { + return + } + p.anchors = make(map[string]*Node) + p.expect(yaml_STREAM_START_EVENT) + p.doneInit = true +} + +func (p *parser) destroy() { + if p.event.typ != yaml_NO_EVENT { + yaml_event_delete(&p.event) + } + yaml_parser_delete(&p.parser) +} + +// expect consumes an event from the event stream and +// checks that it's of the expected type. +func (p *parser) expect(e yaml_event_type_t) { + if p.event.typ == yaml_NO_EVENT { + if !yaml_parser_parse(&p.parser, &p.event) { + p.fail() + } + } + if p.event.typ == yaml_STREAM_END_EVENT { + failf("attempted to go past the end of stream; corrupted value?") + } + if p.event.typ != e { + p.parser.problem = fmt.Sprintf("expected %s event but got %s", e, p.event.typ) + p.fail() + } + yaml_event_delete(&p.event) + p.event.typ = yaml_NO_EVENT +} + +// peek peeks at the next event in the event stream, +// puts the results into p.event and returns the event type. +func (p *parser) peek() yaml_event_type_t { + if p.event.typ != yaml_NO_EVENT { + return p.event.typ + } + if !yaml_parser_parse(&p.parser, &p.event) { + p.fail() + } + return p.event.typ +} + +func (p *parser) fail() { + var where string + var line int + if p.parser.context_mark.line != 0 { + line = p.parser.context_mark.line + // Scanner errors don't iterate line before returning error + if p.parser.error == yaml_SCANNER_ERROR { + line++ + } + } else if p.parser.problem_mark.line != 0 { + line = p.parser.problem_mark.line + // Scanner errors don't iterate line before returning error + if p.parser.error == yaml_SCANNER_ERROR { + line++ + } + } + if line != 0 { + where = "line " + strconv.Itoa(line) + ": " + } + var msg string + if len(p.parser.problem) > 0 { + msg = p.parser.problem + } else { + msg = "unknown problem parsing YAML content" + } + failf("%s%s", where, msg) +} + +func (p *parser) anchor(n *Node, anchor []byte) { + if anchor != nil { + n.Anchor = string(anchor) + p.anchors[n.Anchor] = n + } +} + +func (p *parser) parse() *Node { + p.init() + switch p.peek() { + case yaml_SCALAR_EVENT: + return p.scalar() + case yaml_ALIAS_EVENT: + return p.alias() + case yaml_MAPPING_START_EVENT: + return p.mapping() + case yaml_SEQUENCE_START_EVENT: + return p.sequence() + case yaml_DOCUMENT_START_EVENT: + return p.document() + case yaml_STREAM_END_EVENT: + // Happens when attempting to decode an empty buffer. + return nil + case yaml_TAIL_COMMENT_EVENT: + panic("internal error: unexpected tail comment event (please report)") + default: + panic("internal error: attempted to parse unknown event (please report): " + p.event.typ.String()) + } +} + +func (p *parser) node(kind Kind, defaultTag, tag, value string) *Node { + var style Style + if tag != "" && tag != "!" { + tag = shortTag(tag) + style = TaggedStyle + } else if defaultTag != "" { + tag = defaultTag + } else if kind == ScalarNode { + tag, _ = resolve("", value) + } + n := &Node{ + Kind: kind, + Tag: tag, + Value: value, + Style: style, + } + if !p.textless { + n.Line = p.event.start_mark.line + 1 + n.Column = p.event.start_mark.column + 1 + n.HeadComment = string(p.event.head_comment) + n.LineComment = string(p.event.line_comment) + n.FootComment = string(p.event.foot_comment) + } + return n +} + +func (p *parser) parseChild(parent *Node) *Node { + child := p.parse() + parent.Content = append(parent.Content, child) + return child +} + +func (p *parser) document() *Node { + n := p.node(DocumentNode, "", "", "") + p.doc = n + p.expect(yaml_DOCUMENT_START_EVENT) + p.parseChild(n) + if p.peek() == yaml_DOCUMENT_END_EVENT { + n.FootComment = string(p.event.foot_comment) + } + p.expect(yaml_DOCUMENT_END_EVENT) + return n +} + +func (p *parser) alias() *Node { + n := p.node(AliasNode, "", "", string(p.event.anchor)) + n.Alias = p.anchors[n.Value] + if n.Alias == nil { + failf("unknown anchor '%s' referenced", n.Value) + } + p.expect(yaml_ALIAS_EVENT) + return n +} + +func (p *parser) scalar() *Node { + var parsedStyle = p.event.scalar_style() + var nodeStyle Style + switch { + case parsedStyle&yaml_DOUBLE_QUOTED_SCALAR_STYLE != 0: + nodeStyle = DoubleQuotedStyle + case parsedStyle&yaml_SINGLE_QUOTED_SCALAR_STYLE != 0: + nodeStyle = SingleQuotedStyle + case parsedStyle&yaml_LITERAL_SCALAR_STYLE != 0: + nodeStyle = LiteralStyle + case parsedStyle&yaml_FOLDED_SCALAR_STYLE != 0: + nodeStyle = FoldedStyle + } + var nodeValue = string(p.event.value) + var nodeTag = string(p.event.tag) + var defaultTag string + if nodeStyle == 0 { + if nodeValue == "<<" { + defaultTag = mergeTag + } + } else { + defaultTag = strTag + } + n := p.node(ScalarNode, defaultTag, nodeTag, nodeValue) + n.Style |= nodeStyle + p.anchor(n, p.event.anchor) + p.expect(yaml_SCALAR_EVENT) + return n +} + +func (p *parser) sequence() *Node { + n := p.node(SequenceNode, seqTag, string(p.event.tag), "") + if p.event.sequence_style()&yaml_FLOW_SEQUENCE_STYLE != 0 { + n.Style |= FlowStyle + } + p.anchor(n, p.event.anchor) + p.expect(yaml_SEQUENCE_START_EVENT) + for p.peek() != yaml_SEQUENCE_END_EVENT { + p.parseChild(n) + } + n.LineComment = string(p.event.line_comment) + n.FootComment = string(p.event.foot_comment) + p.expect(yaml_SEQUENCE_END_EVENT) + return n +} + +func (p *parser) mapping() *Node { + n := p.node(MappingNode, mapTag, string(p.event.tag), "") + block := true + if p.event.mapping_style()&yaml_FLOW_MAPPING_STYLE != 0 { + block = false + n.Style |= FlowStyle + } + p.anchor(n, p.event.anchor) + p.expect(yaml_MAPPING_START_EVENT) + for p.peek() != yaml_MAPPING_END_EVENT { + k := p.parseChild(n) + if block && k.FootComment != "" { + // Must be a foot comment for the prior value when being dedented. + if len(n.Content) > 2 { + n.Content[len(n.Content)-3].FootComment = k.FootComment + k.FootComment = "" + } + } + v := p.parseChild(n) + if k.FootComment == "" && v.FootComment != "" { + k.FootComment = v.FootComment + v.FootComment = "" + } + if p.peek() == yaml_TAIL_COMMENT_EVENT { + if k.FootComment == "" { + k.FootComment = string(p.event.foot_comment) + } + p.expect(yaml_TAIL_COMMENT_EVENT) + } + } + n.LineComment = string(p.event.line_comment) + n.FootComment = string(p.event.foot_comment) + if n.Style&FlowStyle == 0 && n.FootComment != "" && len(n.Content) > 1 { + n.Content[len(n.Content)-2].FootComment = n.FootComment + n.FootComment = "" + } + p.expect(yaml_MAPPING_END_EVENT) + return n +} + +// ---------------------------------------------------------------------------- +// Decoder, unmarshals a node into a provided value. + +type decoder struct { + doc *Node + aliases map[*Node]bool + terrors []string + + stringMapType reflect.Type + generalMapType reflect.Type + + knownFields bool + uniqueKeys bool + decodeCount int + aliasCount int + aliasDepth int +} + +var ( + nodeType = reflect.TypeOf(Node{}) + durationType = reflect.TypeOf(time.Duration(0)) + stringMapType = reflect.TypeOf(map[string]interface{}{}) + generalMapType = reflect.TypeOf(map[interface{}]interface{}{}) + ifaceType = generalMapType.Elem() + timeType = reflect.TypeOf(time.Time{}) + ptrTimeType = reflect.TypeOf(&time.Time{}) +) + +func newDecoder() *decoder { + d := &decoder{ + stringMapType: stringMapType, + generalMapType: generalMapType, + uniqueKeys: true, + } + d.aliases = make(map[*Node]bool) + return d +} + +func (d *decoder) terror(n *Node, tag string, out reflect.Value) { + if n.Tag != "" { + tag = n.Tag + } + value := n.Value + if tag != seqTag && tag != mapTag { + if len(value) > 10 { + value = " `" + value[:7] + "...`" + } else { + value = " `" + value + "`" + } + } + d.terrors = append(d.terrors, fmt.Sprintf("line %d: cannot unmarshal %s%s into %s", n.Line, shortTag(tag), value, out.Type())) +} + +func (d *decoder) callUnmarshaler(n *Node, u Unmarshaler) (good bool) { + err := u.UnmarshalYAML(n) + if e, ok := err.(*TypeError); ok { + d.terrors = append(d.terrors, e.Errors...) + return false + } + if err != nil { + fail(err) + } + return true +} + +func (d *decoder) callObsoleteUnmarshaler(n *Node, u obsoleteUnmarshaler) (good bool) { + terrlen := len(d.terrors) + err := u.UnmarshalYAML(func(v interface{}) (err error) { + defer handleErr(&err) + d.unmarshal(n, reflect.ValueOf(v)) + if len(d.terrors) > terrlen { + issues := d.terrors[terrlen:] + d.terrors = d.terrors[:terrlen] + return &TypeError{issues} + } + return nil + }) + if e, ok := err.(*TypeError); ok { + d.terrors = append(d.terrors, e.Errors...) + return false + } + if err != nil { + fail(err) + } + return true +} + +// d.prepare initializes and dereferences pointers and calls UnmarshalYAML +// if a value is found to implement it. +// It returns the initialized and dereferenced out value, whether +// unmarshalling was already done by UnmarshalYAML, and if so whether +// its types unmarshalled appropriately. +// +// If n holds a null value, prepare returns before doing anything. +func (d *decoder) prepare(n *Node, out reflect.Value) (newout reflect.Value, unmarshaled, good bool) { + if n.ShortTag() == nullTag || n.Kind == 0 && n.IsZero() { + return out, false, false + } + again := true + for again { + again = false + if out.Kind() == reflect.Ptr { + if out.IsNil() { + out.Set(reflect.New(out.Type().Elem())) + } + out = out.Elem() + again = true + } + if out.CanAddr() { + outi := out.Addr().Interface() + if u, ok := outi.(Unmarshaler); ok { + good = d.callUnmarshaler(n, u) + return out, true, good + } + if u, ok := outi.(obsoleteUnmarshaler); ok { + good = d.callObsoleteUnmarshaler(n, u) + return out, true, good + } + } + } + return out, false, false +} + +func (d *decoder) fieldByIndex(n *Node, v reflect.Value, index []int) (field reflect.Value) { + if n.ShortTag() == nullTag { + return reflect.Value{} + } + for _, num := range index { + for { + if v.Kind() == reflect.Ptr { + if v.IsNil() { + v.Set(reflect.New(v.Type().Elem())) + } + v = v.Elem() + continue + } + break + } + v = v.Field(num) + } + return v +} + +const ( + // 400,000 decode operations is ~500kb of dense object declarations, or + // ~5kb of dense object declarations with 10000% alias expansion + alias_ratio_range_low = 400000 + + // 4,000,000 decode operations is ~5MB of dense object declarations, or + // ~4.5MB of dense object declarations with 10% alias expansion + alias_ratio_range_high = 4000000 + + // alias_ratio_range is the range over which we scale allowed alias ratios + alias_ratio_range = float64(alias_ratio_range_high - alias_ratio_range_low) +) + +func allowedAliasRatio(decodeCount int) float64 { + switch { + case decodeCount <= alias_ratio_range_low: + // allow 99% to come from alias expansion for small-to-medium documents + return 0.99 + case decodeCount >= alias_ratio_range_high: + // allow 10% to come from alias expansion for very large documents + return 0.10 + default: + // scale smoothly from 99% down to 10% over the range. + // this maps to 396,000 - 400,000 allowed alias-driven decodes over the range. + // 400,000 decode operations is ~100MB of allocations in worst-case scenarios (single-item maps). + return 0.99 - 0.89*(float64(decodeCount-alias_ratio_range_low)/alias_ratio_range) + } +} + +func (d *decoder) unmarshal(n *Node, out reflect.Value) (good bool) { + d.decodeCount++ + if d.aliasDepth > 0 { + d.aliasCount++ + } + if d.aliasCount > 100 && d.decodeCount > 1000 && float64(d.aliasCount)/float64(d.decodeCount) > allowedAliasRatio(d.decodeCount) { + failf("document contains excessive aliasing") + } + if out.Type() == nodeType { + out.Set(reflect.ValueOf(n).Elem()) + return true + } + switch n.Kind { + case DocumentNode: + return d.document(n, out) + case AliasNode: + return d.alias(n, out) + } + out, unmarshaled, good := d.prepare(n, out) + if unmarshaled { + return good + } + switch n.Kind { + case ScalarNode: + good = d.scalar(n, out) + case MappingNode: + good = d.mapping(n, out) + case SequenceNode: + good = d.sequence(n, out) + case 0: + if n.IsZero() { + return d.null(out) + } + fallthrough + default: + failf("cannot decode node with unknown kind %d", n.Kind) + } + return good +} + +func (d *decoder) document(n *Node, out reflect.Value) (good bool) { + if len(n.Content) == 1 { + d.doc = n + d.unmarshal(n.Content[0], out) + return true + } + return false +} + +func (d *decoder) alias(n *Node, out reflect.Value) (good bool) { + if d.aliases[n] { + // TODO this could actually be allowed in some circumstances. + failf("anchor '%s' value contains itself", n.Value) + } + d.aliases[n] = true + d.aliasDepth++ + good = d.unmarshal(n.Alias, out) + d.aliasDepth-- + delete(d.aliases, n) + return good +} + +var zeroValue reflect.Value + +func resetMap(out reflect.Value) { + for _, k := range out.MapKeys() { + out.SetMapIndex(k, zeroValue) + } +} + +func (d *decoder) null(out reflect.Value) bool { + if out.CanAddr() { + switch out.Kind() { + case reflect.Interface, reflect.Ptr, reflect.Map, reflect.Slice: + out.Set(reflect.Zero(out.Type())) + return true + } + } + return false +} + +func (d *decoder) scalar(n *Node, out reflect.Value) bool { + var tag string + var resolved interface{} + if n.indicatedString() { + tag = strTag + resolved = n.Value + } else { + tag, resolved = resolve(n.Tag, n.Value) + if tag == binaryTag { + data, err := base64.StdEncoding.DecodeString(resolved.(string)) + if err != nil { + failf("!!binary value contains invalid base64 data") + } + resolved = string(data) + } + } + if resolved == nil { + return d.null(out) + } + if resolvedv := reflect.ValueOf(resolved); out.Type() == resolvedv.Type() { + // We've resolved to exactly the type we want, so use that. + out.Set(resolvedv) + return true + } + // Perhaps we can use the value as a TextUnmarshaler to + // set its value. + if out.CanAddr() { + u, ok := out.Addr().Interface().(encoding.TextUnmarshaler) + if ok { + var text []byte + if tag == binaryTag { + text = []byte(resolved.(string)) + } else { + // We let any value be unmarshaled into TextUnmarshaler. + // That might be more lax than we'd like, but the + // TextUnmarshaler itself should bowl out any dubious values. + text = []byte(n.Value) + } + err := u.UnmarshalText(text) + if err != nil { + fail(err) + } + return true + } + } + switch out.Kind() { + case reflect.String: + if tag == binaryTag { + out.SetString(resolved.(string)) + return true + } + out.SetString(n.Value) + return true + case reflect.Interface: + out.Set(reflect.ValueOf(resolved)) + return true + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + // This used to work in v2, but it's very unfriendly. + isDuration := out.Type() == durationType + + switch resolved := resolved.(type) { + case int: + if !isDuration && !out.OverflowInt(int64(resolved)) { + out.SetInt(int64(resolved)) + return true + } + case int64: + if !isDuration && !out.OverflowInt(resolved) { + out.SetInt(resolved) + return true + } + case uint64: + if !isDuration && resolved <= math.MaxInt64 && !out.OverflowInt(int64(resolved)) { + out.SetInt(int64(resolved)) + return true + } + case float64: + if !isDuration && resolved <= math.MaxInt64 && !out.OverflowInt(int64(resolved)) { + out.SetInt(int64(resolved)) + return true + } + case string: + if out.Type() == durationType { + d, err := time.ParseDuration(resolved) + if err == nil { + out.SetInt(int64(d)) + return true + } + } + } + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: + switch resolved := resolved.(type) { + case int: + if resolved >= 0 && !out.OverflowUint(uint64(resolved)) { + out.SetUint(uint64(resolved)) + return true + } + case int64: + if resolved >= 0 && !out.OverflowUint(uint64(resolved)) { + out.SetUint(uint64(resolved)) + return true + } + case uint64: + if !out.OverflowUint(uint64(resolved)) { + out.SetUint(uint64(resolved)) + return true + } + case float64: + if resolved <= math.MaxUint64 && !out.OverflowUint(uint64(resolved)) { + out.SetUint(uint64(resolved)) + return true + } + } + case reflect.Bool: + switch resolved := resolved.(type) { + case bool: + out.SetBool(resolved) + return true + case string: + // This offers some compatibility with the 1.1 spec (https://yaml.org/type/bool.html). + // It only works if explicitly attempting to unmarshal into a typed bool value. + switch resolved { + case "y", "Y", "yes", "Yes", "YES", "on", "On", "ON": + out.SetBool(true) + return true + case "n", "N", "no", "No", "NO", "off", "Off", "OFF": + out.SetBool(false) + return true + } + } + case reflect.Float32, reflect.Float64: + switch resolved := resolved.(type) { + case int: + out.SetFloat(float64(resolved)) + return true + case int64: + out.SetFloat(float64(resolved)) + return true + case uint64: + out.SetFloat(float64(resolved)) + return true + case float64: + out.SetFloat(resolved) + return true + } + case reflect.Struct: + if resolvedv := reflect.ValueOf(resolved); out.Type() == resolvedv.Type() { + out.Set(resolvedv) + return true + } + case reflect.Ptr: + panic("yaml internal error: please report the issue") + } + d.terror(n, tag, out) + return false +} + +func settableValueOf(i interface{}) reflect.Value { + v := reflect.ValueOf(i) + sv := reflect.New(v.Type()).Elem() + sv.Set(v) + return sv +} + +func (d *decoder) sequence(n *Node, out reflect.Value) (good bool) { + l := len(n.Content) + + var iface reflect.Value + switch out.Kind() { + case reflect.Slice: + out.Set(reflect.MakeSlice(out.Type(), l, l)) + case reflect.Array: + if l != out.Len() { + failf("invalid array: want %d elements but got %d", out.Len(), l) + } + case reflect.Interface: + // No type hints. Will have to use a generic sequence. + iface = out + out = settableValueOf(make([]interface{}, l)) + default: + d.terror(n, seqTag, out) + return false + } + et := out.Type().Elem() + + j := 0 + for i := 0; i < l; i++ { + e := reflect.New(et).Elem() + if ok := d.unmarshal(n.Content[i], e); ok { + out.Index(j).Set(e) + j++ + } + } + if out.Kind() != reflect.Array { + out.Set(out.Slice(0, j)) + } + if iface.IsValid() { + iface.Set(out) + } + return true +} + +func (d *decoder) mapping(n *Node, out reflect.Value) (good bool) { + l := len(n.Content) + if d.uniqueKeys { + nerrs := len(d.terrors) + for i := 0; i < l; i += 2 { + ni := n.Content[i] + for j := i + 2; j < l; j += 2 { + nj := n.Content[j] + if ni.Kind == nj.Kind && ni.Value == nj.Value { + d.terrors = append(d.terrors, fmt.Sprintf("line %d: mapping key %#v already defined at line %d", nj.Line, nj.Value, ni.Line)) + } + } + } + if len(d.terrors) > nerrs { + return false + } + } + switch out.Kind() { + case reflect.Struct: + return d.mappingStruct(n, out) + case reflect.Map: + // okay + case reflect.Interface: + iface := out + if isStringMap(n) { + out = reflect.MakeMap(d.stringMapType) + } else { + out = reflect.MakeMap(d.generalMapType) + } + iface.Set(out) + default: + d.terror(n, mapTag, out) + return false + } + + outt := out.Type() + kt := outt.Key() + et := outt.Elem() + + stringMapType := d.stringMapType + generalMapType := d.generalMapType + if outt.Elem() == ifaceType { + if outt.Key().Kind() == reflect.String { + d.stringMapType = outt + } else if outt.Key() == ifaceType { + d.generalMapType = outt + } + } + + if out.IsNil() { + out.Set(reflect.MakeMap(outt)) + } + for i := 0; i < l; i += 2 { + if isMerge(n.Content[i]) { + d.merge(n.Content[i+1], out) + continue + } + k := reflect.New(kt).Elem() + if d.unmarshal(n.Content[i], k) { + kkind := k.Kind() + if kkind == reflect.Interface { + kkind = k.Elem().Kind() + } + if kkind == reflect.Map || kkind == reflect.Slice { + failf("invalid map key: %#v", k.Interface()) + } + e := reflect.New(et).Elem() + if d.unmarshal(n.Content[i+1], e) { + out.SetMapIndex(k, e) + } + } + } + d.stringMapType = stringMapType + d.generalMapType = generalMapType + return true +} + +func isStringMap(n *Node) bool { + if n.Kind != MappingNode { + return false + } + l := len(n.Content) + for i := 0; i < l; i += 2 { + if n.Content[i].ShortTag() != strTag { + return false + } + } + return true +} + +func (d *decoder) mappingStruct(n *Node, out reflect.Value) (good bool) { + sinfo, err := getStructInfo(out.Type()) + if err != nil { + panic(err) + } + + var inlineMap reflect.Value + var elemType reflect.Type + if sinfo.InlineMap != -1 { + inlineMap = out.Field(sinfo.InlineMap) + inlineMap.Set(reflect.New(inlineMap.Type()).Elem()) + elemType = inlineMap.Type().Elem() + } + + for _, index := range sinfo.InlineUnmarshalers { + field := d.fieldByIndex(n, out, index) + d.prepare(n, field) + } + + var doneFields []bool + if d.uniqueKeys { + doneFields = make([]bool, len(sinfo.FieldsList)) + } + name := settableValueOf("") + l := len(n.Content) + for i := 0; i < l; i += 2 { + ni := n.Content[i] + if isMerge(ni) { + d.merge(n.Content[i+1], out) + continue + } + if !d.unmarshal(ni, name) { + continue + } + if info, ok := sinfo.FieldsMap[name.String()]; ok { + if d.uniqueKeys { + if doneFields[info.Id] { + d.terrors = append(d.terrors, fmt.Sprintf("line %d: field %s already set in type %s", ni.Line, name.String(), out.Type())) + continue + } + doneFields[info.Id] = true + } + var field reflect.Value + if info.Inline == nil { + field = out.Field(info.Num) + } else { + field = d.fieldByIndex(n, out, info.Inline) + } + d.unmarshal(n.Content[i+1], field) + } else if sinfo.InlineMap != -1 { + if inlineMap.IsNil() { + inlineMap.Set(reflect.MakeMap(inlineMap.Type())) + } + value := reflect.New(elemType).Elem() + d.unmarshal(n.Content[i+1], value) + inlineMap.SetMapIndex(name, value) + } else if d.knownFields { + d.terrors = append(d.terrors, fmt.Sprintf("line %d: field %s not found in type %s", ni.Line, name.String(), out.Type())) + } + } + return true +} + +func failWantMap() { + failf("map merge requires map or sequence of maps as the value") +} + +func (d *decoder) merge(n *Node, out reflect.Value) { + switch n.Kind { + case MappingNode: + d.unmarshal(n, out) + case AliasNode: + if n.Alias != nil && n.Alias.Kind != MappingNode { + failWantMap() + } + d.unmarshal(n, out) + case SequenceNode: + // Step backwards as earlier nodes take precedence. + for i := len(n.Content) - 1; i >= 0; i-- { + ni := n.Content[i] + if ni.Kind == AliasNode { + if ni.Alias != nil && ni.Alias.Kind != MappingNode { + failWantMap() + } + } else if ni.Kind != MappingNode { + failWantMap() + } + d.unmarshal(ni, out) + } + default: + failWantMap() + } +} + +func isMerge(n *Node) bool { + return n.Kind == ScalarNode && n.Value == "<<" && (n.Tag == "" || n.Tag == "!" || shortTag(n.Tag) == mergeTag) +} diff --git a/vendor/gopkg.in/yaml.v3/emitterc.go b/vendor/gopkg.in/yaml.v3/emitterc.go new file mode 100644 index 00000000..c29217ef --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/emitterc.go @@ -0,0 +1,2022 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +import ( + "bytes" + "fmt" +) + +// Flush the buffer if needed. +func flush(emitter *yaml_emitter_t) bool { + if emitter.buffer_pos+5 >= len(emitter.buffer) { + return yaml_emitter_flush(emitter) + } + return true +} + +// Put a character to the output buffer. +func put(emitter *yaml_emitter_t, value byte) bool { + if emitter.buffer_pos+5 >= len(emitter.buffer) && !yaml_emitter_flush(emitter) { + return false + } + emitter.buffer[emitter.buffer_pos] = value + emitter.buffer_pos++ + emitter.column++ + return true +} + +// Put a line break to the output buffer. +func put_break(emitter *yaml_emitter_t) bool { + if emitter.buffer_pos+5 >= len(emitter.buffer) && !yaml_emitter_flush(emitter) { + return false + } + switch emitter.line_break { + case yaml_CR_BREAK: + emitter.buffer[emitter.buffer_pos] = '\r' + emitter.buffer_pos += 1 + case yaml_LN_BREAK: + emitter.buffer[emitter.buffer_pos] = '\n' + emitter.buffer_pos += 1 + case yaml_CRLN_BREAK: + emitter.buffer[emitter.buffer_pos+0] = '\r' + emitter.buffer[emitter.buffer_pos+1] = '\n' + emitter.buffer_pos += 2 + default: + panic("unknown line break setting") + } + if emitter.column == 0 { + emitter.space_above = true + } + emitter.column = 0 + emitter.line++ + // [Go] Do this here and below and drop from everywhere else (see commented lines). + emitter.indention = true + return true +} + +// Copy a character from a string into buffer. +func write(emitter *yaml_emitter_t, s []byte, i *int) bool { + if emitter.buffer_pos+5 >= len(emitter.buffer) && !yaml_emitter_flush(emitter) { + return false + } + p := emitter.buffer_pos + w := width(s[*i]) + switch w { + case 4: + emitter.buffer[p+3] = s[*i+3] + fallthrough + case 3: + emitter.buffer[p+2] = s[*i+2] + fallthrough + case 2: + emitter.buffer[p+1] = s[*i+1] + fallthrough + case 1: + emitter.buffer[p+0] = s[*i+0] + default: + panic("unknown character width") + } + emitter.column++ + emitter.buffer_pos += w + *i += w + return true +} + +// Write a whole string into buffer. +func write_all(emitter *yaml_emitter_t, s []byte) bool { + for i := 0; i < len(s); { + if !write(emitter, s, &i) { + return false + } + } + return true +} + +// Copy a line break character from a string into buffer. +func write_break(emitter *yaml_emitter_t, s []byte, i *int) bool { + if s[*i] == '\n' { + if !put_break(emitter) { + return false + } + *i++ + } else { + if !write(emitter, s, i) { + return false + } + if emitter.column == 0 { + emitter.space_above = true + } + emitter.column = 0 + emitter.line++ + // [Go] Do this here and above and drop from everywhere else (see commented lines). + emitter.indention = true + } + return true +} + +// Set an emitter error and return false. +func yaml_emitter_set_emitter_error(emitter *yaml_emitter_t, problem string) bool { + emitter.error = yaml_EMITTER_ERROR + emitter.problem = problem + return false +} + +// Emit an event. +func yaml_emitter_emit(emitter *yaml_emitter_t, event *yaml_event_t) bool { + emitter.events = append(emitter.events, *event) + for !yaml_emitter_need_more_events(emitter) { + event := &emitter.events[emitter.events_head] + if !yaml_emitter_analyze_event(emitter, event) { + return false + } + if !yaml_emitter_state_machine(emitter, event) { + return false + } + yaml_event_delete(event) + emitter.events_head++ + } + return true +} + +// Check if we need to accumulate more events before emitting. +// +// We accumulate extra +// - 1 event for DOCUMENT-START +// - 2 events for SEQUENCE-START +// - 3 events for MAPPING-START +// +func yaml_emitter_need_more_events(emitter *yaml_emitter_t) bool { + if emitter.events_head == len(emitter.events) { + return true + } + var accumulate int + switch emitter.events[emitter.events_head].typ { + case yaml_DOCUMENT_START_EVENT: + accumulate = 1 + break + case yaml_SEQUENCE_START_EVENT: + accumulate = 2 + break + case yaml_MAPPING_START_EVENT: + accumulate = 3 + break + default: + return false + } + if len(emitter.events)-emitter.events_head > accumulate { + return false + } + var level int + for i := emitter.events_head; i < len(emitter.events); i++ { + switch emitter.events[i].typ { + case yaml_STREAM_START_EVENT, yaml_DOCUMENT_START_EVENT, yaml_SEQUENCE_START_EVENT, yaml_MAPPING_START_EVENT: + level++ + case yaml_STREAM_END_EVENT, yaml_DOCUMENT_END_EVENT, yaml_SEQUENCE_END_EVENT, yaml_MAPPING_END_EVENT: + level-- + } + if level == 0 { + return false + } + } + return true +} + +// Append a directive to the directives stack. +func yaml_emitter_append_tag_directive(emitter *yaml_emitter_t, value *yaml_tag_directive_t, allow_duplicates bool) bool { + for i := 0; i < len(emitter.tag_directives); i++ { + if bytes.Equal(value.handle, emitter.tag_directives[i].handle) { + if allow_duplicates { + return true + } + return yaml_emitter_set_emitter_error(emitter, "duplicate %TAG directive") + } + } + + // [Go] Do we actually need to copy this given garbage collection + // and the lack of deallocating destructors? + tag_copy := yaml_tag_directive_t{ + handle: make([]byte, len(value.handle)), + prefix: make([]byte, len(value.prefix)), + } + copy(tag_copy.handle, value.handle) + copy(tag_copy.prefix, value.prefix) + emitter.tag_directives = append(emitter.tag_directives, tag_copy) + return true +} + +// Increase the indentation level. +func yaml_emitter_increase_indent(emitter *yaml_emitter_t, flow, indentless bool) bool { + emitter.indents = append(emitter.indents, emitter.indent) + if emitter.indent < 0 { + if flow { + emitter.indent = emitter.best_indent + } else { + emitter.indent = 0 + } + } else if !indentless { + // [Go] This was changed so that indentations are more regular. + if emitter.states[len(emitter.states)-1] == yaml_EMIT_BLOCK_SEQUENCE_ITEM_STATE { + // The first indent inside a sequence will just skip the "- " indicator. + emitter.indent += 2 + } else { + // Everything else aligns to the chosen indentation. + emitter.indent = emitter.best_indent*((emitter.indent+emitter.best_indent)/emitter.best_indent) + } + } + return true +} + +// State dispatcher. +func yaml_emitter_state_machine(emitter *yaml_emitter_t, event *yaml_event_t) bool { + switch emitter.state { + default: + case yaml_EMIT_STREAM_START_STATE: + return yaml_emitter_emit_stream_start(emitter, event) + + case yaml_EMIT_FIRST_DOCUMENT_START_STATE: + return yaml_emitter_emit_document_start(emitter, event, true) + + case yaml_EMIT_DOCUMENT_START_STATE: + return yaml_emitter_emit_document_start(emitter, event, false) + + case yaml_EMIT_DOCUMENT_CONTENT_STATE: + return yaml_emitter_emit_document_content(emitter, event) + + case yaml_EMIT_DOCUMENT_END_STATE: + return yaml_emitter_emit_document_end(emitter, event) + + case yaml_EMIT_FLOW_SEQUENCE_FIRST_ITEM_STATE: + return yaml_emitter_emit_flow_sequence_item(emitter, event, true, false) + + case yaml_EMIT_FLOW_SEQUENCE_TRAIL_ITEM_STATE: + return yaml_emitter_emit_flow_sequence_item(emitter, event, false, true) + + case yaml_EMIT_FLOW_SEQUENCE_ITEM_STATE: + return yaml_emitter_emit_flow_sequence_item(emitter, event, false, false) + + case yaml_EMIT_FLOW_MAPPING_FIRST_KEY_STATE: + return yaml_emitter_emit_flow_mapping_key(emitter, event, true, false) + + case yaml_EMIT_FLOW_MAPPING_TRAIL_KEY_STATE: + return yaml_emitter_emit_flow_mapping_key(emitter, event, false, true) + + case yaml_EMIT_FLOW_MAPPING_KEY_STATE: + return yaml_emitter_emit_flow_mapping_key(emitter, event, false, false) + + case yaml_EMIT_FLOW_MAPPING_SIMPLE_VALUE_STATE: + return yaml_emitter_emit_flow_mapping_value(emitter, event, true) + + case yaml_EMIT_FLOW_MAPPING_VALUE_STATE: + return yaml_emitter_emit_flow_mapping_value(emitter, event, false) + + case yaml_EMIT_BLOCK_SEQUENCE_FIRST_ITEM_STATE: + return yaml_emitter_emit_block_sequence_item(emitter, event, true) + + case yaml_EMIT_BLOCK_SEQUENCE_ITEM_STATE: + return yaml_emitter_emit_block_sequence_item(emitter, event, false) + + case yaml_EMIT_BLOCK_MAPPING_FIRST_KEY_STATE: + return yaml_emitter_emit_block_mapping_key(emitter, event, true) + + case yaml_EMIT_BLOCK_MAPPING_KEY_STATE: + return yaml_emitter_emit_block_mapping_key(emitter, event, false) + + case yaml_EMIT_BLOCK_MAPPING_SIMPLE_VALUE_STATE: + return yaml_emitter_emit_block_mapping_value(emitter, event, true) + + case yaml_EMIT_BLOCK_MAPPING_VALUE_STATE: + return yaml_emitter_emit_block_mapping_value(emitter, event, false) + + case yaml_EMIT_END_STATE: + return yaml_emitter_set_emitter_error(emitter, "expected nothing after STREAM-END") + } + panic("invalid emitter state") +} + +// Expect STREAM-START. +func yaml_emitter_emit_stream_start(emitter *yaml_emitter_t, event *yaml_event_t) bool { + if event.typ != yaml_STREAM_START_EVENT { + return yaml_emitter_set_emitter_error(emitter, "expected STREAM-START") + } + if emitter.encoding == yaml_ANY_ENCODING { + emitter.encoding = event.encoding + if emitter.encoding == yaml_ANY_ENCODING { + emitter.encoding = yaml_UTF8_ENCODING + } + } + if emitter.best_indent < 2 || emitter.best_indent > 9 { + emitter.best_indent = 2 + } + if emitter.best_width >= 0 && emitter.best_width <= emitter.best_indent*2 { + emitter.best_width = 80 + } + if emitter.best_width < 0 { + emitter.best_width = 1<<31 - 1 + } + if emitter.line_break == yaml_ANY_BREAK { + emitter.line_break = yaml_LN_BREAK + } + + emitter.indent = -1 + emitter.line = 0 + emitter.column = 0 + emitter.whitespace = true + emitter.indention = true + emitter.space_above = true + emitter.foot_indent = -1 + + if emitter.encoding != yaml_UTF8_ENCODING { + if !yaml_emitter_write_bom(emitter) { + return false + } + } + emitter.state = yaml_EMIT_FIRST_DOCUMENT_START_STATE + return true +} + +// Expect DOCUMENT-START or STREAM-END. +func yaml_emitter_emit_document_start(emitter *yaml_emitter_t, event *yaml_event_t, first bool) bool { + + if event.typ == yaml_DOCUMENT_START_EVENT { + + if event.version_directive != nil { + if !yaml_emitter_analyze_version_directive(emitter, event.version_directive) { + return false + } + } + + for i := 0; i < len(event.tag_directives); i++ { + tag_directive := &event.tag_directives[i] + if !yaml_emitter_analyze_tag_directive(emitter, tag_directive) { + return false + } + if !yaml_emitter_append_tag_directive(emitter, tag_directive, false) { + return false + } + } + + for i := 0; i < len(default_tag_directives); i++ { + tag_directive := &default_tag_directives[i] + if !yaml_emitter_append_tag_directive(emitter, tag_directive, true) { + return false + } + } + + implicit := event.implicit + if !first || emitter.canonical { + implicit = false + } + + if emitter.open_ended && (event.version_directive != nil || len(event.tag_directives) > 0) { + if !yaml_emitter_write_indicator(emitter, []byte("..."), true, false, false) { + return false + } + if !yaml_emitter_write_indent(emitter) { + return false + } + } + + if event.version_directive != nil { + implicit = false + if !yaml_emitter_write_indicator(emitter, []byte("%YAML"), true, false, false) { + return false + } + if !yaml_emitter_write_indicator(emitter, []byte("1.1"), true, false, false) { + return false + } + if !yaml_emitter_write_indent(emitter) { + return false + } + } + + if len(event.tag_directives) > 0 { + implicit = false + for i := 0; i < len(event.tag_directives); i++ { + tag_directive := &event.tag_directives[i] + if !yaml_emitter_write_indicator(emitter, []byte("%TAG"), true, false, false) { + return false + } + if !yaml_emitter_write_tag_handle(emitter, tag_directive.handle) { + return false + } + if !yaml_emitter_write_tag_content(emitter, tag_directive.prefix, true) { + return false + } + if !yaml_emitter_write_indent(emitter) { + return false + } + } + } + + if yaml_emitter_check_empty_document(emitter) { + implicit = false + } + if !implicit { + if !yaml_emitter_write_indent(emitter) { + return false + } + if !yaml_emitter_write_indicator(emitter, []byte("---"), true, false, false) { + return false + } + if emitter.canonical || true { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + } + + if len(emitter.head_comment) > 0 { + if !yaml_emitter_process_head_comment(emitter) { + return false + } + if !put_break(emitter) { + return false + } + } + + emitter.state = yaml_EMIT_DOCUMENT_CONTENT_STATE + return true + } + + if event.typ == yaml_STREAM_END_EVENT { + if emitter.open_ended { + if !yaml_emitter_write_indicator(emitter, []byte("..."), true, false, false) { + return false + } + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if !yaml_emitter_flush(emitter) { + return false + } + emitter.state = yaml_EMIT_END_STATE + return true + } + + return yaml_emitter_set_emitter_error(emitter, "expected DOCUMENT-START or STREAM-END") +} + +// Expect the root node. +func yaml_emitter_emit_document_content(emitter *yaml_emitter_t, event *yaml_event_t) bool { + emitter.states = append(emitter.states, yaml_EMIT_DOCUMENT_END_STATE) + + if !yaml_emitter_process_head_comment(emitter) { + return false + } + if !yaml_emitter_emit_node(emitter, event, true, false, false, false) { + return false + } + if !yaml_emitter_process_line_comment(emitter) { + return false + } + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + return true +} + +// Expect DOCUMENT-END. +func yaml_emitter_emit_document_end(emitter *yaml_emitter_t, event *yaml_event_t) bool { + if event.typ != yaml_DOCUMENT_END_EVENT { + return yaml_emitter_set_emitter_error(emitter, "expected DOCUMENT-END") + } + // [Go] Force document foot separation. + emitter.foot_indent = 0 + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + emitter.foot_indent = -1 + if !yaml_emitter_write_indent(emitter) { + return false + } + if !event.implicit { + // [Go] Allocate the slice elsewhere. + if !yaml_emitter_write_indicator(emitter, []byte("..."), true, false, false) { + return false + } + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if !yaml_emitter_flush(emitter) { + return false + } + emitter.state = yaml_EMIT_DOCUMENT_START_STATE + emitter.tag_directives = emitter.tag_directives[:0] + return true +} + +// Expect a flow item node. +func yaml_emitter_emit_flow_sequence_item(emitter *yaml_emitter_t, event *yaml_event_t, first, trail bool) bool { + if first { + if !yaml_emitter_write_indicator(emitter, []byte{'['}, true, true, false) { + return false + } + if !yaml_emitter_increase_indent(emitter, true, false) { + return false + } + emitter.flow_level++ + } + + if event.typ == yaml_SEQUENCE_END_EVENT { + if emitter.canonical && !first && !trail { + if !yaml_emitter_write_indicator(emitter, []byte{','}, false, false, false) { + return false + } + } + emitter.flow_level-- + emitter.indent = emitter.indents[len(emitter.indents)-1] + emitter.indents = emitter.indents[:len(emitter.indents)-1] + if emitter.column == 0 || emitter.canonical && !first { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if !yaml_emitter_write_indicator(emitter, []byte{']'}, false, false, false) { + return false + } + if !yaml_emitter_process_line_comment(emitter) { + return false + } + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + emitter.state = emitter.states[len(emitter.states)-1] + emitter.states = emitter.states[:len(emitter.states)-1] + + return true + } + + if !first && !trail { + if !yaml_emitter_write_indicator(emitter, []byte{','}, false, false, false) { + return false + } + } + + if !yaml_emitter_process_head_comment(emitter) { + return false + } + if emitter.column == 0 { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + + if emitter.canonical || emitter.column > emitter.best_width { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if len(emitter.line_comment)+len(emitter.foot_comment)+len(emitter.tail_comment) > 0 { + emitter.states = append(emitter.states, yaml_EMIT_FLOW_SEQUENCE_TRAIL_ITEM_STATE) + } else { + emitter.states = append(emitter.states, yaml_EMIT_FLOW_SEQUENCE_ITEM_STATE) + } + if !yaml_emitter_emit_node(emitter, event, false, true, false, false) { + return false + } + if len(emitter.line_comment)+len(emitter.foot_comment)+len(emitter.tail_comment) > 0 { + if !yaml_emitter_write_indicator(emitter, []byte{','}, false, false, false) { + return false + } + } + if !yaml_emitter_process_line_comment(emitter) { + return false + } + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + return true +} + +// Expect a flow key node. +func yaml_emitter_emit_flow_mapping_key(emitter *yaml_emitter_t, event *yaml_event_t, first, trail bool) bool { + if first { + if !yaml_emitter_write_indicator(emitter, []byte{'{'}, true, true, false) { + return false + } + if !yaml_emitter_increase_indent(emitter, true, false) { + return false + } + emitter.flow_level++ + } + + if event.typ == yaml_MAPPING_END_EVENT { + if (emitter.canonical || len(emitter.head_comment)+len(emitter.foot_comment)+len(emitter.tail_comment) > 0) && !first && !trail { + if !yaml_emitter_write_indicator(emitter, []byte{','}, false, false, false) { + return false + } + } + if !yaml_emitter_process_head_comment(emitter) { + return false + } + emitter.flow_level-- + emitter.indent = emitter.indents[len(emitter.indents)-1] + emitter.indents = emitter.indents[:len(emitter.indents)-1] + if emitter.canonical && !first { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if !yaml_emitter_write_indicator(emitter, []byte{'}'}, false, false, false) { + return false + } + if !yaml_emitter_process_line_comment(emitter) { + return false + } + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + emitter.state = emitter.states[len(emitter.states)-1] + emitter.states = emitter.states[:len(emitter.states)-1] + return true + } + + if !first && !trail { + if !yaml_emitter_write_indicator(emitter, []byte{','}, false, false, false) { + return false + } + } + + if !yaml_emitter_process_head_comment(emitter) { + return false + } + + if emitter.column == 0 { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + + if emitter.canonical || emitter.column > emitter.best_width { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + + if !emitter.canonical && yaml_emitter_check_simple_key(emitter) { + emitter.states = append(emitter.states, yaml_EMIT_FLOW_MAPPING_SIMPLE_VALUE_STATE) + return yaml_emitter_emit_node(emitter, event, false, false, true, true) + } + if !yaml_emitter_write_indicator(emitter, []byte{'?'}, true, false, false) { + return false + } + emitter.states = append(emitter.states, yaml_EMIT_FLOW_MAPPING_VALUE_STATE) + return yaml_emitter_emit_node(emitter, event, false, false, true, false) +} + +// Expect a flow value node. +func yaml_emitter_emit_flow_mapping_value(emitter *yaml_emitter_t, event *yaml_event_t, simple bool) bool { + if simple { + if !yaml_emitter_write_indicator(emitter, []byte{':'}, false, false, false) { + return false + } + } else { + if emitter.canonical || emitter.column > emitter.best_width { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if !yaml_emitter_write_indicator(emitter, []byte{':'}, true, false, false) { + return false + } + } + if len(emitter.line_comment)+len(emitter.foot_comment)+len(emitter.tail_comment) > 0 { + emitter.states = append(emitter.states, yaml_EMIT_FLOW_MAPPING_TRAIL_KEY_STATE) + } else { + emitter.states = append(emitter.states, yaml_EMIT_FLOW_MAPPING_KEY_STATE) + } + if !yaml_emitter_emit_node(emitter, event, false, false, true, false) { + return false + } + if len(emitter.line_comment)+len(emitter.foot_comment)+len(emitter.tail_comment) > 0 { + if !yaml_emitter_write_indicator(emitter, []byte{','}, false, false, false) { + return false + } + } + if !yaml_emitter_process_line_comment(emitter) { + return false + } + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + return true +} + +// Expect a block item node. +func yaml_emitter_emit_block_sequence_item(emitter *yaml_emitter_t, event *yaml_event_t, first bool) bool { + if first { + if !yaml_emitter_increase_indent(emitter, false, false) { + return false + } + } + if event.typ == yaml_SEQUENCE_END_EVENT { + emitter.indent = emitter.indents[len(emitter.indents)-1] + emitter.indents = emitter.indents[:len(emitter.indents)-1] + emitter.state = emitter.states[len(emitter.states)-1] + emitter.states = emitter.states[:len(emitter.states)-1] + return true + } + if !yaml_emitter_process_head_comment(emitter) { + return false + } + if !yaml_emitter_write_indent(emitter) { + return false + } + if !yaml_emitter_write_indicator(emitter, []byte{'-'}, true, false, true) { + return false + } + emitter.states = append(emitter.states, yaml_EMIT_BLOCK_SEQUENCE_ITEM_STATE) + if !yaml_emitter_emit_node(emitter, event, false, true, false, false) { + return false + } + if !yaml_emitter_process_line_comment(emitter) { + return false + } + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + return true +} + +// Expect a block key node. +func yaml_emitter_emit_block_mapping_key(emitter *yaml_emitter_t, event *yaml_event_t, first bool) bool { + if first { + if !yaml_emitter_increase_indent(emitter, false, false) { + return false + } + } + if !yaml_emitter_process_head_comment(emitter) { + return false + } + if event.typ == yaml_MAPPING_END_EVENT { + emitter.indent = emitter.indents[len(emitter.indents)-1] + emitter.indents = emitter.indents[:len(emitter.indents)-1] + emitter.state = emitter.states[len(emitter.states)-1] + emitter.states = emitter.states[:len(emitter.states)-1] + return true + } + if !yaml_emitter_write_indent(emitter) { + return false + } + if len(emitter.line_comment) > 0 { + // [Go] A line comment was provided for the key. That's unusual as the + // scanner associates line comments with the value. Either way, + // save the line comment and render it appropriately later. + emitter.key_line_comment = emitter.line_comment + emitter.line_comment = nil + } + if yaml_emitter_check_simple_key(emitter) { + emitter.states = append(emitter.states, yaml_EMIT_BLOCK_MAPPING_SIMPLE_VALUE_STATE) + return yaml_emitter_emit_node(emitter, event, false, false, true, true) + } + if !yaml_emitter_write_indicator(emitter, []byte{'?'}, true, false, true) { + return false + } + emitter.states = append(emitter.states, yaml_EMIT_BLOCK_MAPPING_VALUE_STATE) + return yaml_emitter_emit_node(emitter, event, false, false, true, false) +} + +// Expect a block value node. +func yaml_emitter_emit_block_mapping_value(emitter *yaml_emitter_t, event *yaml_event_t, simple bool) bool { + if simple { + if !yaml_emitter_write_indicator(emitter, []byte{':'}, false, false, false) { + return false + } + } else { + if !yaml_emitter_write_indent(emitter) { + return false + } + if !yaml_emitter_write_indicator(emitter, []byte{':'}, true, false, true) { + return false + } + } + if len(emitter.key_line_comment) > 0 { + // [Go] A line comment was previously provided for the key. Handle it before + // the value so the inline comments are placed correctly. + if yaml_emitter_silent_nil_event(emitter, event) && len(emitter.line_comment) == 0 { + // Nothing other than the line comment will be written on the line. + emitter.line_comment = emitter.key_line_comment + emitter.key_line_comment = nil + } else { + // An actual value is coming, so emit the comment line. + emitter.line_comment, emitter.key_line_comment = emitter.key_line_comment, emitter.line_comment + if !yaml_emitter_process_line_comment(emitter) { + return false + } + emitter.line_comment, emitter.key_line_comment = emitter.key_line_comment, emitter.line_comment + // Indent in unless it's a block that will reindent anyway. + if event.sequence_style() == yaml_FLOW_SEQUENCE_STYLE || (event.typ != yaml_MAPPING_START_EVENT && event.typ != yaml_SEQUENCE_START_EVENT) { + emitter.indent = emitter.best_indent*((emitter.indent+emitter.best_indent)/emitter.best_indent) + if !yaml_emitter_write_indent(emitter) { + return false + } + } + } + } + emitter.states = append(emitter.states, yaml_EMIT_BLOCK_MAPPING_KEY_STATE) + if !yaml_emitter_emit_node(emitter, event, false, false, true, false) { + return false + } + if !yaml_emitter_process_line_comment(emitter) { + return false + } + if !yaml_emitter_process_foot_comment(emitter) { + return false + } + return true +} + +func yaml_emitter_silent_nil_event(emitter *yaml_emitter_t, event *yaml_event_t) bool { + return event.typ == yaml_SCALAR_EVENT && event.implicit && !emitter.canonical && len(emitter.scalar_data.value) == 0 +} + +// Expect a node. +func yaml_emitter_emit_node(emitter *yaml_emitter_t, event *yaml_event_t, + root bool, sequence bool, mapping bool, simple_key bool) bool { + + emitter.root_context = root + emitter.sequence_context = sequence + emitter.mapping_context = mapping + emitter.simple_key_context = simple_key + + switch event.typ { + case yaml_ALIAS_EVENT: + return yaml_emitter_emit_alias(emitter, event) + case yaml_SCALAR_EVENT: + return yaml_emitter_emit_scalar(emitter, event) + case yaml_SEQUENCE_START_EVENT: + return yaml_emitter_emit_sequence_start(emitter, event) + case yaml_MAPPING_START_EVENT: + return yaml_emitter_emit_mapping_start(emitter, event) + default: + return yaml_emitter_set_emitter_error(emitter, + fmt.Sprintf("expected SCALAR, SEQUENCE-START, MAPPING-START, or ALIAS, but got %v", event.typ)) + } +} + +// Expect ALIAS. +func yaml_emitter_emit_alias(emitter *yaml_emitter_t, event *yaml_event_t) bool { + if !yaml_emitter_process_anchor(emitter) { + return false + } + emitter.state = emitter.states[len(emitter.states)-1] + emitter.states = emitter.states[:len(emitter.states)-1] + return true +} + +// Expect SCALAR. +func yaml_emitter_emit_scalar(emitter *yaml_emitter_t, event *yaml_event_t) bool { + if !yaml_emitter_select_scalar_style(emitter, event) { + return false + } + if !yaml_emitter_process_anchor(emitter) { + return false + } + if !yaml_emitter_process_tag(emitter) { + return false + } + if !yaml_emitter_increase_indent(emitter, true, false) { + return false + } + if !yaml_emitter_process_scalar(emitter) { + return false + } + emitter.indent = emitter.indents[len(emitter.indents)-1] + emitter.indents = emitter.indents[:len(emitter.indents)-1] + emitter.state = emitter.states[len(emitter.states)-1] + emitter.states = emitter.states[:len(emitter.states)-1] + return true +} + +// Expect SEQUENCE-START. +func yaml_emitter_emit_sequence_start(emitter *yaml_emitter_t, event *yaml_event_t) bool { + if !yaml_emitter_process_anchor(emitter) { + return false + } + if !yaml_emitter_process_tag(emitter) { + return false + } + if emitter.flow_level > 0 || emitter.canonical || event.sequence_style() == yaml_FLOW_SEQUENCE_STYLE || + yaml_emitter_check_empty_sequence(emitter) { + emitter.state = yaml_EMIT_FLOW_SEQUENCE_FIRST_ITEM_STATE + } else { + emitter.state = yaml_EMIT_BLOCK_SEQUENCE_FIRST_ITEM_STATE + } + return true +} + +// Expect MAPPING-START. +func yaml_emitter_emit_mapping_start(emitter *yaml_emitter_t, event *yaml_event_t) bool { + if !yaml_emitter_process_anchor(emitter) { + return false + } + if !yaml_emitter_process_tag(emitter) { + return false + } + if emitter.flow_level > 0 || emitter.canonical || event.mapping_style() == yaml_FLOW_MAPPING_STYLE || + yaml_emitter_check_empty_mapping(emitter) { + emitter.state = yaml_EMIT_FLOW_MAPPING_FIRST_KEY_STATE + } else { + emitter.state = yaml_EMIT_BLOCK_MAPPING_FIRST_KEY_STATE + } + return true +} + +// Check if the document content is an empty scalar. +func yaml_emitter_check_empty_document(emitter *yaml_emitter_t) bool { + return false // [Go] Huh? +} + +// Check if the next events represent an empty sequence. +func yaml_emitter_check_empty_sequence(emitter *yaml_emitter_t) bool { + if len(emitter.events)-emitter.events_head < 2 { + return false + } + return emitter.events[emitter.events_head].typ == yaml_SEQUENCE_START_EVENT && + emitter.events[emitter.events_head+1].typ == yaml_SEQUENCE_END_EVENT +} + +// Check if the next events represent an empty mapping. +func yaml_emitter_check_empty_mapping(emitter *yaml_emitter_t) bool { + if len(emitter.events)-emitter.events_head < 2 { + return false + } + return emitter.events[emitter.events_head].typ == yaml_MAPPING_START_EVENT && + emitter.events[emitter.events_head+1].typ == yaml_MAPPING_END_EVENT +} + +// Check if the next node can be expressed as a simple key. +func yaml_emitter_check_simple_key(emitter *yaml_emitter_t) bool { + length := 0 + switch emitter.events[emitter.events_head].typ { + case yaml_ALIAS_EVENT: + length += len(emitter.anchor_data.anchor) + case yaml_SCALAR_EVENT: + if emitter.scalar_data.multiline { + return false + } + length += len(emitter.anchor_data.anchor) + + len(emitter.tag_data.handle) + + len(emitter.tag_data.suffix) + + len(emitter.scalar_data.value) + case yaml_SEQUENCE_START_EVENT: + if !yaml_emitter_check_empty_sequence(emitter) { + return false + } + length += len(emitter.anchor_data.anchor) + + len(emitter.tag_data.handle) + + len(emitter.tag_data.suffix) + case yaml_MAPPING_START_EVENT: + if !yaml_emitter_check_empty_mapping(emitter) { + return false + } + length += len(emitter.anchor_data.anchor) + + len(emitter.tag_data.handle) + + len(emitter.tag_data.suffix) + default: + return false + } + return length <= 128 +} + +// Determine an acceptable scalar style. +func yaml_emitter_select_scalar_style(emitter *yaml_emitter_t, event *yaml_event_t) bool { + + no_tag := len(emitter.tag_data.handle) == 0 && len(emitter.tag_data.suffix) == 0 + if no_tag && !event.implicit && !event.quoted_implicit { + return yaml_emitter_set_emitter_error(emitter, "neither tag nor implicit flags are specified") + } + + style := event.scalar_style() + if style == yaml_ANY_SCALAR_STYLE { + style = yaml_PLAIN_SCALAR_STYLE + } + if emitter.canonical { + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } + if emitter.simple_key_context && emitter.scalar_data.multiline { + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } + + if style == yaml_PLAIN_SCALAR_STYLE { + if emitter.flow_level > 0 && !emitter.scalar_data.flow_plain_allowed || + emitter.flow_level == 0 && !emitter.scalar_data.block_plain_allowed { + style = yaml_SINGLE_QUOTED_SCALAR_STYLE + } + if len(emitter.scalar_data.value) == 0 && (emitter.flow_level > 0 || emitter.simple_key_context) { + style = yaml_SINGLE_QUOTED_SCALAR_STYLE + } + if no_tag && !event.implicit { + style = yaml_SINGLE_QUOTED_SCALAR_STYLE + } + } + if style == yaml_SINGLE_QUOTED_SCALAR_STYLE { + if !emitter.scalar_data.single_quoted_allowed { + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } + } + if style == yaml_LITERAL_SCALAR_STYLE || style == yaml_FOLDED_SCALAR_STYLE { + if !emitter.scalar_data.block_allowed || emitter.flow_level > 0 || emitter.simple_key_context { + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } + } + + if no_tag && !event.quoted_implicit && style != yaml_PLAIN_SCALAR_STYLE { + emitter.tag_data.handle = []byte{'!'} + } + emitter.scalar_data.style = style + return true +} + +// Write an anchor. +func yaml_emitter_process_anchor(emitter *yaml_emitter_t) bool { + if emitter.anchor_data.anchor == nil { + return true + } + c := []byte{'&'} + if emitter.anchor_data.alias { + c[0] = '*' + } + if !yaml_emitter_write_indicator(emitter, c, true, false, false) { + return false + } + return yaml_emitter_write_anchor(emitter, emitter.anchor_data.anchor) +} + +// Write a tag. +func yaml_emitter_process_tag(emitter *yaml_emitter_t) bool { + if len(emitter.tag_data.handle) == 0 && len(emitter.tag_data.suffix) == 0 { + return true + } + if len(emitter.tag_data.handle) > 0 { + if !yaml_emitter_write_tag_handle(emitter, emitter.tag_data.handle) { + return false + } + if len(emitter.tag_data.suffix) > 0 { + if !yaml_emitter_write_tag_content(emitter, emitter.tag_data.suffix, false) { + return false + } + } + } else { + // [Go] Allocate these slices elsewhere. + if !yaml_emitter_write_indicator(emitter, []byte("!<"), true, false, false) { + return false + } + if !yaml_emitter_write_tag_content(emitter, emitter.tag_data.suffix, false) { + return false + } + if !yaml_emitter_write_indicator(emitter, []byte{'>'}, false, false, false) { + return false + } + } + return true +} + +// Write a scalar. +func yaml_emitter_process_scalar(emitter *yaml_emitter_t) bool { + switch emitter.scalar_data.style { + case yaml_PLAIN_SCALAR_STYLE: + return yaml_emitter_write_plain_scalar(emitter, emitter.scalar_data.value, !emitter.simple_key_context) + + case yaml_SINGLE_QUOTED_SCALAR_STYLE: + return yaml_emitter_write_single_quoted_scalar(emitter, emitter.scalar_data.value, !emitter.simple_key_context) + + case yaml_DOUBLE_QUOTED_SCALAR_STYLE: + return yaml_emitter_write_double_quoted_scalar(emitter, emitter.scalar_data.value, !emitter.simple_key_context) + + case yaml_LITERAL_SCALAR_STYLE: + return yaml_emitter_write_literal_scalar(emitter, emitter.scalar_data.value) + + case yaml_FOLDED_SCALAR_STYLE: + return yaml_emitter_write_folded_scalar(emitter, emitter.scalar_data.value) + } + panic("unknown scalar style") +} + +// Write a head comment. +func yaml_emitter_process_head_comment(emitter *yaml_emitter_t) bool { + if len(emitter.tail_comment) > 0 { + if !yaml_emitter_write_indent(emitter) { + return false + } + if !yaml_emitter_write_comment(emitter, emitter.tail_comment) { + return false + } + emitter.tail_comment = emitter.tail_comment[:0] + emitter.foot_indent = emitter.indent + if emitter.foot_indent < 0 { + emitter.foot_indent = 0 + } + } + + if len(emitter.head_comment) == 0 { + return true + } + if !yaml_emitter_write_indent(emitter) { + return false + } + if !yaml_emitter_write_comment(emitter, emitter.head_comment) { + return false + } + emitter.head_comment = emitter.head_comment[:0] + return true +} + +// Write an line comment. +func yaml_emitter_process_line_comment(emitter *yaml_emitter_t) bool { + if len(emitter.line_comment) == 0 { + return true + } + if !emitter.whitespace { + if !put(emitter, ' ') { + return false + } + } + if !yaml_emitter_write_comment(emitter, emitter.line_comment) { + return false + } + emitter.line_comment = emitter.line_comment[:0] + return true +} + +// Write a foot comment. +func yaml_emitter_process_foot_comment(emitter *yaml_emitter_t) bool { + if len(emitter.foot_comment) == 0 { + return true + } + if !yaml_emitter_write_indent(emitter) { + return false + } + if !yaml_emitter_write_comment(emitter, emitter.foot_comment) { + return false + } + emitter.foot_comment = emitter.foot_comment[:0] + emitter.foot_indent = emitter.indent + if emitter.foot_indent < 0 { + emitter.foot_indent = 0 + } + return true +} + +// Check if a %YAML directive is valid. +func yaml_emitter_analyze_version_directive(emitter *yaml_emitter_t, version_directive *yaml_version_directive_t) bool { + if version_directive.major != 1 || version_directive.minor != 1 { + return yaml_emitter_set_emitter_error(emitter, "incompatible %YAML directive") + } + return true +} + +// Check if a %TAG directive is valid. +func yaml_emitter_analyze_tag_directive(emitter *yaml_emitter_t, tag_directive *yaml_tag_directive_t) bool { + handle := tag_directive.handle + prefix := tag_directive.prefix + if len(handle) == 0 { + return yaml_emitter_set_emitter_error(emitter, "tag handle must not be empty") + } + if handle[0] != '!' { + return yaml_emitter_set_emitter_error(emitter, "tag handle must start with '!'") + } + if handle[len(handle)-1] != '!' { + return yaml_emitter_set_emitter_error(emitter, "tag handle must end with '!'") + } + for i := 1; i < len(handle)-1; i += width(handle[i]) { + if !is_alpha(handle, i) { + return yaml_emitter_set_emitter_error(emitter, "tag handle must contain alphanumerical characters only") + } + } + if len(prefix) == 0 { + return yaml_emitter_set_emitter_error(emitter, "tag prefix must not be empty") + } + return true +} + +// Check if an anchor is valid. +func yaml_emitter_analyze_anchor(emitter *yaml_emitter_t, anchor []byte, alias bool) bool { + if len(anchor) == 0 { + problem := "anchor value must not be empty" + if alias { + problem = "alias value must not be empty" + } + return yaml_emitter_set_emitter_error(emitter, problem) + } + for i := 0; i < len(anchor); i += width(anchor[i]) { + if !is_alpha(anchor, i) { + problem := "anchor value must contain alphanumerical characters only" + if alias { + problem = "alias value must contain alphanumerical characters only" + } + return yaml_emitter_set_emitter_error(emitter, problem) + } + } + emitter.anchor_data.anchor = anchor + emitter.anchor_data.alias = alias + return true +} + +// Check if a tag is valid. +func yaml_emitter_analyze_tag(emitter *yaml_emitter_t, tag []byte) bool { + if len(tag) == 0 { + return yaml_emitter_set_emitter_error(emitter, "tag value must not be empty") + } + for i := 0; i < len(emitter.tag_directives); i++ { + tag_directive := &emitter.tag_directives[i] + if bytes.HasPrefix(tag, tag_directive.prefix) { + emitter.tag_data.handle = tag_directive.handle + emitter.tag_data.suffix = tag[len(tag_directive.prefix):] + return true + } + } + emitter.tag_data.suffix = tag + return true +} + +// Check if a scalar is valid. +func yaml_emitter_analyze_scalar(emitter *yaml_emitter_t, value []byte) bool { + var ( + block_indicators = false + flow_indicators = false + line_breaks = false + special_characters = false + tab_characters = false + + leading_space = false + leading_break = false + trailing_space = false + trailing_break = false + break_space = false + space_break = false + + preceded_by_whitespace = false + followed_by_whitespace = false + previous_space = false + previous_break = false + ) + + emitter.scalar_data.value = value + + if len(value) == 0 { + emitter.scalar_data.multiline = false + emitter.scalar_data.flow_plain_allowed = false + emitter.scalar_data.block_plain_allowed = true + emitter.scalar_data.single_quoted_allowed = true + emitter.scalar_data.block_allowed = false + return true + } + + if len(value) >= 3 && ((value[0] == '-' && value[1] == '-' && value[2] == '-') || (value[0] == '.' && value[1] == '.' && value[2] == '.')) { + block_indicators = true + flow_indicators = true + } + + preceded_by_whitespace = true + for i, w := 0, 0; i < len(value); i += w { + w = width(value[i]) + followed_by_whitespace = i+w >= len(value) || is_blank(value, i+w) + + if i == 0 { + switch value[i] { + case '#', ',', '[', ']', '{', '}', '&', '*', '!', '|', '>', '\'', '"', '%', '@', '`': + flow_indicators = true + block_indicators = true + case '?', ':': + flow_indicators = true + if followed_by_whitespace { + block_indicators = true + } + case '-': + if followed_by_whitespace { + flow_indicators = true + block_indicators = true + } + } + } else { + switch value[i] { + case ',', '?', '[', ']', '{', '}': + flow_indicators = true + case ':': + flow_indicators = true + if followed_by_whitespace { + block_indicators = true + } + case '#': + if preceded_by_whitespace { + flow_indicators = true + block_indicators = true + } + } + } + + if value[i] == '\t' { + tab_characters = true + } else if !is_printable(value, i) || !is_ascii(value, i) && !emitter.unicode { + special_characters = true + } + if is_space(value, i) { + if i == 0 { + leading_space = true + } + if i+width(value[i]) == len(value) { + trailing_space = true + } + if previous_break { + break_space = true + } + previous_space = true + previous_break = false + } else if is_break(value, i) { + line_breaks = true + if i == 0 { + leading_break = true + } + if i+width(value[i]) == len(value) { + trailing_break = true + } + if previous_space { + space_break = true + } + previous_space = false + previous_break = true + } else { + previous_space = false + previous_break = false + } + + // [Go]: Why 'z'? Couldn't be the end of the string as that's the loop condition. + preceded_by_whitespace = is_blankz(value, i) + } + + emitter.scalar_data.multiline = line_breaks + emitter.scalar_data.flow_plain_allowed = true + emitter.scalar_data.block_plain_allowed = true + emitter.scalar_data.single_quoted_allowed = true + emitter.scalar_data.block_allowed = true + + if leading_space || leading_break || trailing_space || trailing_break { + emitter.scalar_data.flow_plain_allowed = false + emitter.scalar_data.block_plain_allowed = false + } + if trailing_space { + emitter.scalar_data.block_allowed = false + } + if break_space { + emitter.scalar_data.flow_plain_allowed = false + emitter.scalar_data.block_plain_allowed = false + emitter.scalar_data.single_quoted_allowed = false + } + if space_break || tab_characters || special_characters { + emitter.scalar_data.flow_plain_allowed = false + emitter.scalar_data.block_plain_allowed = false + emitter.scalar_data.single_quoted_allowed = false + } + if space_break || special_characters { + emitter.scalar_data.block_allowed = false + } + if line_breaks { + emitter.scalar_data.flow_plain_allowed = false + emitter.scalar_data.block_plain_allowed = false + } + if flow_indicators { + emitter.scalar_data.flow_plain_allowed = false + } + if block_indicators { + emitter.scalar_data.block_plain_allowed = false + } + return true +} + +// Check if the event data is valid. +func yaml_emitter_analyze_event(emitter *yaml_emitter_t, event *yaml_event_t) bool { + + emitter.anchor_data.anchor = nil + emitter.tag_data.handle = nil + emitter.tag_data.suffix = nil + emitter.scalar_data.value = nil + + if len(event.head_comment) > 0 { + emitter.head_comment = event.head_comment + } + if len(event.line_comment) > 0 { + emitter.line_comment = event.line_comment + } + if len(event.foot_comment) > 0 { + emitter.foot_comment = event.foot_comment + } + if len(event.tail_comment) > 0 { + emitter.tail_comment = event.tail_comment + } + + switch event.typ { + case yaml_ALIAS_EVENT: + if !yaml_emitter_analyze_anchor(emitter, event.anchor, true) { + return false + } + + case yaml_SCALAR_EVENT: + if len(event.anchor) > 0 { + if !yaml_emitter_analyze_anchor(emitter, event.anchor, false) { + return false + } + } + if len(event.tag) > 0 && (emitter.canonical || (!event.implicit && !event.quoted_implicit)) { + if !yaml_emitter_analyze_tag(emitter, event.tag) { + return false + } + } + if !yaml_emitter_analyze_scalar(emitter, event.value) { + return false + } + + case yaml_SEQUENCE_START_EVENT: + if len(event.anchor) > 0 { + if !yaml_emitter_analyze_anchor(emitter, event.anchor, false) { + return false + } + } + if len(event.tag) > 0 && (emitter.canonical || !event.implicit) { + if !yaml_emitter_analyze_tag(emitter, event.tag) { + return false + } + } + + case yaml_MAPPING_START_EVENT: + if len(event.anchor) > 0 { + if !yaml_emitter_analyze_anchor(emitter, event.anchor, false) { + return false + } + } + if len(event.tag) > 0 && (emitter.canonical || !event.implicit) { + if !yaml_emitter_analyze_tag(emitter, event.tag) { + return false + } + } + } + return true +} + +// Write the BOM character. +func yaml_emitter_write_bom(emitter *yaml_emitter_t) bool { + if !flush(emitter) { + return false + } + pos := emitter.buffer_pos + emitter.buffer[pos+0] = '\xEF' + emitter.buffer[pos+1] = '\xBB' + emitter.buffer[pos+2] = '\xBF' + emitter.buffer_pos += 3 + return true +} + +func yaml_emitter_write_indent(emitter *yaml_emitter_t) bool { + indent := emitter.indent + if indent < 0 { + indent = 0 + } + if !emitter.indention || emitter.column > indent || (emitter.column == indent && !emitter.whitespace) { + if !put_break(emitter) { + return false + } + } + if emitter.foot_indent == indent { + if !put_break(emitter) { + return false + } + } + for emitter.column < indent { + if !put(emitter, ' ') { + return false + } + } + emitter.whitespace = true + //emitter.indention = true + emitter.space_above = false + emitter.foot_indent = -1 + return true +} + +func yaml_emitter_write_indicator(emitter *yaml_emitter_t, indicator []byte, need_whitespace, is_whitespace, is_indention bool) bool { + if need_whitespace && !emitter.whitespace { + if !put(emitter, ' ') { + return false + } + } + if !write_all(emitter, indicator) { + return false + } + emitter.whitespace = is_whitespace + emitter.indention = (emitter.indention && is_indention) + emitter.open_ended = false + return true +} + +func yaml_emitter_write_anchor(emitter *yaml_emitter_t, value []byte) bool { + if !write_all(emitter, value) { + return false + } + emitter.whitespace = false + emitter.indention = false + return true +} + +func yaml_emitter_write_tag_handle(emitter *yaml_emitter_t, value []byte) bool { + if !emitter.whitespace { + if !put(emitter, ' ') { + return false + } + } + if !write_all(emitter, value) { + return false + } + emitter.whitespace = false + emitter.indention = false + return true +} + +func yaml_emitter_write_tag_content(emitter *yaml_emitter_t, value []byte, need_whitespace bool) bool { + if need_whitespace && !emitter.whitespace { + if !put(emitter, ' ') { + return false + } + } + for i := 0; i < len(value); { + var must_write bool + switch value[i] { + case ';', '/', '?', ':', '@', '&', '=', '+', '$', ',', '_', '.', '~', '*', '\'', '(', ')', '[', ']': + must_write = true + default: + must_write = is_alpha(value, i) + } + if must_write { + if !write(emitter, value, &i) { + return false + } + } else { + w := width(value[i]) + for k := 0; k < w; k++ { + octet := value[i] + i++ + if !put(emitter, '%') { + return false + } + + c := octet >> 4 + if c < 10 { + c += '0' + } else { + c += 'A' - 10 + } + if !put(emitter, c) { + return false + } + + c = octet & 0x0f + if c < 10 { + c += '0' + } else { + c += 'A' - 10 + } + if !put(emitter, c) { + return false + } + } + } + } + emitter.whitespace = false + emitter.indention = false + return true +} + +func yaml_emitter_write_plain_scalar(emitter *yaml_emitter_t, value []byte, allow_breaks bool) bool { + if len(value) > 0 && !emitter.whitespace { + if !put(emitter, ' ') { + return false + } + } + + spaces := false + breaks := false + for i := 0; i < len(value); { + if is_space(value, i) { + if allow_breaks && !spaces && emitter.column > emitter.best_width && !is_space(value, i+1) { + if !yaml_emitter_write_indent(emitter) { + return false + } + i += width(value[i]) + } else { + if !write(emitter, value, &i) { + return false + } + } + spaces = true + } else if is_break(value, i) { + if !breaks && value[i] == '\n' { + if !put_break(emitter) { + return false + } + } + if !write_break(emitter, value, &i) { + return false + } + //emitter.indention = true + breaks = true + } else { + if breaks { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if !write(emitter, value, &i) { + return false + } + emitter.indention = false + spaces = false + breaks = false + } + } + + if len(value) > 0 { + emitter.whitespace = false + } + emitter.indention = false + if emitter.root_context { + emitter.open_ended = true + } + + return true +} + +func yaml_emitter_write_single_quoted_scalar(emitter *yaml_emitter_t, value []byte, allow_breaks bool) bool { + + if !yaml_emitter_write_indicator(emitter, []byte{'\''}, true, false, false) { + return false + } + + spaces := false + breaks := false + for i := 0; i < len(value); { + if is_space(value, i) { + if allow_breaks && !spaces && emitter.column > emitter.best_width && i > 0 && i < len(value)-1 && !is_space(value, i+1) { + if !yaml_emitter_write_indent(emitter) { + return false + } + i += width(value[i]) + } else { + if !write(emitter, value, &i) { + return false + } + } + spaces = true + } else if is_break(value, i) { + if !breaks && value[i] == '\n' { + if !put_break(emitter) { + return false + } + } + if !write_break(emitter, value, &i) { + return false + } + //emitter.indention = true + breaks = true + } else { + if breaks { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if value[i] == '\'' { + if !put(emitter, '\'') { + return false + } + } + if !write(emitter, value, &i) { + return false + } + emitter.indention = false + spaces = false + breaks = false + } + } + if !yaml_emitter_write_indicator(emitter, []byte{'\''}, false, false, false) { + return false + } + emitter.whitespace = false + emitter.indention = false + return true +} + +func yaml_emitter_write_double_quoted_scalar(emitter *yaml_emitter_t, value []byte, allow_breaks bool) bool { + spaces := false + if !yaml_emitter_write_indicator(emitter, []byte{'"'}, true, false, false) { + return false + } + + for i := 0; i < len(value); { + if !is_printable(value, i) || (!emitter.unicode && !is_ascii(value, i)) || + is_bom(value, i) || is_break(value, i) || + value[i] == '"' || value[i] == '\\' { + + octet := value[i] + + var w int + var v rune + switch { + case octet&0x80 == 0x00: + w, v = 1, rune(octet&0x7F) + case octet&0xE0 == 0xC0: + w, v = 2, rune(octet&0x1F) + case octet&0xF0 == 0xE0: + w, v = 3, rune(octet&0x0F) + case octet&0xF8 == 0xF0: + w, v = 4, rune(octet&0x07) + } + for k := 1; k < w; k++ { + octet = value[i+k] + v = (v << 6) + (rune(octet) & 0x3F) + } + i += w + + if !put(emitter, '\\') { + return false + } + + var ok bool + switch v { + case 0x00: + ok = put(emitter, '0') + case 0x07: + ok = put(emitter, 'a') + case 0x08: + ok = put(emitter, 'b') + case 0x09: + ok = put(emitter, 't') + case 0x0A: + ok = put(emitter, 'n') + case 0x0b: + ok = put(emitter, 'v') + case 0x0c: + ok = put(emitter, 'f') + case 0x0d: + ok = put(emitter, 'r') + case 0x1b: + ok = put(emitter, 'e') + case 0x22: + ok = put(emitter, '"') + case 0x5c: + ok = put(emitter, '\\') + case 0x85: + ok = put(emitter, 'N') + case 0xA0: + ok = put(emitter, '_') + case 0x2028: + ok = put(emitter, 'L') + case 0x2029: + ok = put(emitter, 'P') + default: + if v <= 0xFF { + ok = put(emitter, 'x') + w = 2 + } else if v <= 0xFFFF { + ok = put(emitter, 'u') + w = 4 + } else { + ok = put(emitter, 'U') + w = 8 + } + for k := (w - 1) * 4; ok && k >= 0; k -= 4 { + digit := byte((v >> uint(k)) & 0x0F) + if digit < 10 { + ok = put(emitter, digit+'0') + } else { + ok = put(emitter, digit+'A'-10) + } + } + } + if !ok { + return false + } + spaces = false + } else if is_space(value, i) { + if allow_breaks && !spaces && emitter.column > emitter.best_width && i > 0 && i < len(value)-1 { + if !yaml_emitter_write_indent(emitter) { + return false + } + if is_space(value, i+1) { + if !put(emitter, '\\') { + return false + } + } + i += width(value[i]) + } else if !write(emitter, value, &i) { + return false + } + spaces = true + } else { + if !write(emitter, value, &i) { + return false + } + spaces = false + } + } + if !yaml_emitter_write_indicator(emitter, []byte{'"'}, false, false, false) { + return false + } + emitter.whitespace = false + emitter.indention = false + return true +} + +func yaml_emitter_write_block_scalar_hints(emitter *yaml_emitter_t, value []byte) bool { + if is_space(value, 0) || is_break(value, 0) { + indent_hint := []byte{'0' + byte(emitter.best_indent)} + if !yaml_emitter_write_indicator(emitter, indent_hint, false, false, false) { + return false + } + } + + emitter.open_ended = false + + var chomp_hint [1]byte + if len(value) == 0 { + chomp_hint[0] = '-' + } else { + i := len(value) - 1 + for value[i]&0xC0 == 0x80 { + i-- + } + if !is_break(value, i) { + chomp_hint[0] = '-' + } else if i == 0 { + chomp_hint[0] = '+' + emitter.open_ended = true + } else { + i-- + for value[i]&0xC0 == 0x80 { + i-- + } + if is_break(value, i) { + chomp_hint[0] = '+' + emitter.open_ended = true + } + } + } + if chomp_hint[0] != 0 { + if !yaml_emitter_write_indicator(emitter, chomp_hint[:], false, false, false) { + return false + } + } + return true +} + +func yaml_emitter_write_literal_scalar(emitter *yaml_emitter_t, value []byte) bool { + if !yaml_emitter_write_indicator(emitter, []byte{'|'}, true, false, false) { + return false + } + if !yaml_emitter_write_block_scalar_hints(emitter, value) { + return false + } + if !put_break(emitter) { + return false + } + //emitter.indention = true + emitter.whitespace = true + breaks := true + for i := 0; i < len(value); { + if is_break(value, i) { + if !write_break(emitter, value, &i) { + return false + } + //emitter.indention = true + breaks = true + } else { + if breaks { + if !yaml_emitter_write_indent(emitter) { + return false + } + } + if !write(emitter, value, &i) { + return false + } + emitter.indention = false + breaks = false + } + } + + return true +} + +func yaml_emitter_write_folded_scalar(emitter *yaml_emitter_t, value []byte) bool { + if !yaml_emitter_write_indicator(emitter, []byte{'>'}, true, false, false) { + return false + } + if !yaml_emitter_write_block_scalar_hints(emitter, value) { + return false + } + + if !put_break(emitter) { + return false + } + //emitter.indention = true + emitter.whitespace = true + + breaks := true + leading_spaces := true + for i := 0; i < len(value); { + if is_break(value, i) { + if !breaks && !leading_spaces && value[i] == '\n' { + k := 0 + for is_break(value, k) { + k += width(value[k]) + } + if !is_blankz(value, k) { + if !put_break(emitter) { + return false + } + } + } + if !write_break(emitter, value, &i) { + return false + } + //emitter.indention = true + breaks = true + } else { + if breaks { + if !yaml_emitter_write_indent(emitter) { + return false + } + leading_spaces = is_blank(value, i) + } + if !breaks && is_space(value, i) && !is_space(value, i+1) && emitter.column > emitter.best_width { + if !yaml_emitter_write_indent(emitter) { + return false + } + i += width(value[i]) + } else { + if !write(emitter, value, &i) { + return false + } + } + emitter.indention = false + breaks = false + } + } + return true +} + +func yaml_emitter_write_comment(emitter *yaml_emitter_t, comment []byte) bool { + breaks := false + pound := false + for i := 0; i < len(comment); { + if is_break(comment, i) { + if !write_break(emitter, comment, &i) { + return false + } + //emitter.indention = true + breaks = true + pound = false + } else { + if breaks && !yaml_emitter_write_indent(emitter) { + return false + } + if !pound { + if comment[i] != '#' && (!put(emitter, '#') || !put(emitter, ' ')) { + return false + } + pound = true + } + if !write(emitter, comment, &i) { + return false + } + emitter.indention = false + breaks = false + } + } + if !breaks && !put_break(emitter) { + return false + } + + emitter.whitespace = true + //emitter.indention = true + return true +} diff --git a/vendor/gopkg.in/yaml.v3/encode.go b/vendor/gopkg.in/yaml.v3/encode.go new file mode 100644 index 00000000..45e8d1e1 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/encode.go @@ -0,0 +1,572 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package yaml + +import ( + "encoding" + "fmt" + "io" + "reflect" + "regexp" + "sort" + "strconv" + "strings" + "time" + "unicode/utf8" +) + +type encoder struct { + emitter yaml_emitter_t + event yaml_event_t + out []byte + flow bool + indent int + doneInit bool +} + +func newEncoder() *encoder { + e := &encoder{} + yaml_emitter_initialize(&e.emitter) + yaml_emitter_set_output_string(&e.emitter, &e.out) + yaml_emitter_set_unicode(&e.emitter, true) + return e +} + +func newEncoderWithWriter(w io.Writer) *encoder { + e := &encoder{} + yaml_emitter_initialize(&e.emitter) + yaml_emitter_set_output_writer(&e.emitter, w) + yaml_emitter_set_unicode(&e.emitter, true) + return e +} + +func (e *encoder) init() { + if e.doneInit { + return + } + if e.indent == 0 { + e.indent = 4 + } + e.emitter.best_indent = e.indent + yaml_stream_start_event_initialize(&e.event, yaml_UTF8_ENCODING) + e.emit() + e.doneInit = true +} + +func (e *encoder) finish() { + e.emitter.open_ended = false + yaml_stream_end_event_initialize(&e.event) + e.emit() +} + +func (e *encoder) destroy() { + yaml_emitter_delete(&e.emitter) +} + +func (e *encoder) emit() { + // This will internally delete the e.event value. + e.must(yaml_emitter_emit(&e.emitter, &e.event)) +} + +func (e *encoder) must(ok bool) { + if !ok { + msg := e.emitter.problem + if msg == "" { + msg = "unknown problem generating YAML content" + } + failf("%s", msg) + } +} + +func (e *encoder) marshalDoc(tag string, in reflect.Value) { + e.init() + var node *Node + if in.IsValid() { + node, _ = in.Interface().(*Node) + } + if node != nil && node.Kind == DocumentNode { + e.nodev(in) + } else { + yaml_document_start_event_initialize(&e.event, nil, nil, true) + e.emit() + e.marshal(tag, in) + yaml_document_end_event_initialize(&e.event, true) + e.emit() + } +} + +func (e *encoder) marshal(tag string, in reflect.Value) { + tag = shortTag(tag) + if !in.IsValid() || in.Kind() == reflect.Ptr && in.IsNil() { + e.nilv() + return + } + iface := in.Interface() + switch value := iface.(type) { + case *Node: + e.nodev(in) + return + case Node: + e.nodev(in.Addr()) + return + case time.Time: + e.timev(tag, in) + return + case *time.Time: + e.timev(tag, in.Elem()) + return + case time.Duration: + e.stringv(tag, reflect.ValueOf(value.String())) + return + case Marshaler: + v, err := value.MarshalYAML() + if err != nil { + fail(err) + } + if v == nil { + e.nilv() + return + } + e.marshal(tag, reflect.ValueOf(v)) + return + case encoding.TextMarshaler: + text, err := value.MarshalText() + if err != nil { + fail(err) + } + in = reflect.ValueOf(string(text)) + case nil: + e.nilv() + return + } + switch in.Kind() { + case reflect.Interface: + e.marshal(tag, in.Elem()) + case reflect.Map: + e.mapv(tag, in) + case reflect.Ptr: + e.marshal(tag, in.Elem()) + case reflect.Struct: + e.structv(tag, in) + case reflect.Slice, reflect.Array: + e.slicev(tag, in) + case reflect.String: + e.stringv(tag, in) + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + e.intv(tag, in) + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: + e.uintv(tag, in) + case reflect.Float32, reflect.Float64: + e.floatv(tag, in) + case reflect.Bool: + e.boolv(tag, in) + default: + panic("cannot marshal type: " + in.Type().String()) + } +} + +func (e *encoder) mapv(tag string, in reflect.Value) { + e.mappingv(tag, func() { + keys := keyList(in.MapKeys()) + sort.Sort(keys) + for _, k := range keys { + e.marshal("", k) + e.marshal("", in.MapIndex(k)) + } + }) +} + +func (e *encoder) fieldByIndex(v reflect.Value, index []int) (field reflect.Value) { + for _, num := range index { + for { + if v.Kind() == reflect.Ptr { + if v.IsNil() { + return reflect.Value{} + } + v = v.Elem() + continue + } + break + } + v = v.Field(num) + } + return v +} + +func (e *encoder) structv(tag string, in reflect.Value) { + sinfo, err := getStructInfo(in.Type()) + if err != nil { + panic(err) + } + e.mappingv(tag, func() { + for _, info := range sinfo.FieldsList { + var value reflect.Value + if info.Inline == nil { + value = in.Field(info.Num) + } else { + value = e.fieldByIndex(in, info.Inline) + if !value.IsValid() { + continue + } + } + if info.OmitEmpty && isZero(value) { + continue + } + e.marshal("", reflect.ValueOf(info.Key)) + e.flow = info.Flow + e.marshal("", value) + } + if sinfo.InlineMap >= 0 { + m := in.Field(sinfo.InlineMap) + if m.Len() > 0 { + e.flow = false + keys := keyList(m.MapKeys()) + sort.Sort(keys) + for _, k := range keys { + if _, found := sinfo.FieldsMap[k.String()]; found { + panic(fmt.Sprintf("cannot have key %q in inlined map: conflicts with struct field", k.String())) + } + e.marshal("", k) + e.flow = false + e.marshal("", m.MapIndex(k)) + } + } + } + }) +} + +func (e *encoder) mappingv(tag string, f func()) { + implicit := tag == "" + style := yaml_BLOCK_MAPPING_STYLE + if e.flow { + e.flow = false + style = yaml_FLOW_MAPPING_STYLE + } + yaml_mapping_start_event_initialize(&e.event, nil, []byte(tag), implicit, style) + e.emit() + f() + yaml_mapping_end_event_initialize(&e.event) + e.emit() +} + +func (e *encoder) slicev(tag string, in reflect.Value) { + implicit := tag == "" + style := yaml_BLOCK_SEQUENCE_STYLE + if e.flow { + e.flow = false + style = yaml_FLOW_SEQUENCE_STYLE + } + e.must(yaml_sequence_start_event_initialize(&e.event, nil, []byte(tag), implicit, style)) + e.emit() + n := in.Len() + for i := 0; i < n; i++ { + e.marshal("", in.Index(i)) + } + e.must(yaml_sequence_end_event_initialize(&e.event)) + e.emit() +} + +// isBase60 returns whether s is in base 60 notation as defined in YAML 1.1. +// +// The base 60 float notation in YAML 1.1 is a terrible idea and is unsupported +// in YAML 1.2 and by this package, but these should be marshalled quoted for +// the time being for compatibility with other parsers. +func isBase60Float(s string) (result bool) { + // Fast path. + if s == "" { + return false + } + c := s[0] + if !(c == '+' || c == '-' || c >= '0' && c <= '9') || strings.IndexByte(s, ':') < 0 { + return false + } + // Do the full match. + return base60float.MatchString(s) +} + +// From http://yaml.org/type/float.html, except the regular expression there +// is bogus. In practice parsers do not enforce the "\.[0-9_]*" suffix. +var base60float = regexp.MustCompile(`^[-+]?[0-9][0-9_]*(?::[0-5]?[0-9])+(?:\.[0-9_]*)?$`) + +// isOldBool returns whether s is bool notation as defined in YAML 1.1. +// +// We continue to force strings that YAML 1.1 would interpret as booleans to be +// rendered as quotes strings so that the marshalled output valid for YAML 1.1 +// parsing. +func isOldBool(s string) (result bool) { + switch s { + case "y", "Y", "yes", "Yes", "YES", "on", "On", "ON", + "n", "N", "no", "No", "NO", "off", "Off", "OFF": + return true + default: + return false + } +} + +func (e *encoder) stringv(tag string, in reflect.Value) { + var style yaml_scalar_style_t + s := in.String() + canUsePlain := true + switch { + case !utf8.ValidString(s): + if tag == binaryTag { + failf("explicitly tagged !!binary data must be base64-encoded") + } + if tag != "" { + failf("cannot marshal invalid UTF-8 data as %s", shortTag(tag)) + } + // It can't be encoded directly as YAML so use a binary tag + // and encode it as base64. + tag = binaryTag + s = encodeBase64(s) + case tag == "": + // Check to see if it would resolve to a specific + // tag when encoded unquoted. If it doesn't, + // there's no need to quote it. + rtag, _ := resolve("", s) + canUsePlain = rtag == strTag && !(isBase60Float(s) || isOldBool(s)) + } + // Note: it's possible for user code to emit invalid YAML + // if they explicitly specify a tag and a string containing + // text that's incompatible with that tag. + switch { + case strings.Contains(s, "\n"): + if e.flow { + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } else { + style = yaml_LITERAL_SCALAR_STYLE + } + case canUsePlain: + style = yaml_PLAIN_SCALAR_STYLE + default: + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } + e.emitScalar(s, "", tag, style, nil, nil, nil, nil) +} + +func (e *encoder) boolv(tag string, in reflect.Value) { + var s string + if in.Bool() { + s = "true" + } else { + s = "false" + } + e.emitScalar(s, "", tag, yaml_PLAIN_SCALAR_STYLE, nil, nil, nil, nil) +} + +func (e *encoder) intv(tag string, in reflect.Value) { + s := strconv.FormatInt(in.Int(), 10) + e.emitScalar(s, "", tag, yaml_PLAIN_SCALAR_STYLE, nil, nil, nil, nil) +} + +func (e *encoder) uintv(tag string, in reflect.Value) { + s := strconv.FormatUint(in.Uint(), 10) + e.emitScalar(s, "", tag, yaml_PLAIN_SCALAR_STYLE, nil, nil, nil, nil) +} + +func (e *encoder) timev(tag string, in reflect.Value) { + t := in.Interface().(time.Time) + s := t.Format(time.RFC3339Nano) + e.emitScalar(s, "", tag, yaml_PLAIN_SCALAR_STYLE, nil, nil, nil, nil) +} + +func (e *encoder) floatv(tag string, in reflect.Value) { + // Issue #352: When formatting, use the precision of the underlying value + precision := 64 + if in.Kind() == reflect.Float32 { + precision = 32 + } + + s := strconv.FormatFloat(in.Float(), 'g', -1, precision) + switch s { + case "+Inf": + s = ".inf" + case "-Inf": + s = "-.inf" + case "NaN": + s = ".nan" + } + e.emitScalar(s, "", tag, yaml_PLAIN_SCALAR_STYLE, nil, nil, nil, nil) +} + +func (e *encoder) nilv() { + e.emitScalar("null", "", "", yaml_PLAIN_SCALAR_STYLE, nil, nil, nil, nil) +} + +func (e *encoder) emitScalar(value, anchor, tag string, style yaml_scalar_style_t, head, line, foot, tail []byte) { + // TODO Kill this function. Replace all initialize calls by their underlining Go literals. + implicit := tag == "" + if !implicit { + tag = longTag(tag) + } + e.must(yaml_scalar_event_initialize(&e.event, []byte(anchor), []byte(tag), []byte(value), implicit, implicit, style)) + e.event.head_comment = head + e.event.line_comment = line + e.event.foot_comment = foot + e.event.tail_comment = tail + e.emit() +} + +func (e *encoder) nodev(in reflect.Value) { + e.node(in.Interface().(*Node), "") +} + +func (e *encoder) node(node *Node, tail string) { + // Zero nodes behave as nil. + if node.Kind == 0 && node.IsZero() { + e.nilv() + return + } + + // If the tag was not explicitly requested, and dropping it won't change the + // implicit tag of the value, don't include it in the presentation. + var tag = node.Tag + var stag = shortTag(tag) + var forceQuoting bool + if tag != "" && node.Style&TaggedStyle == 0 { + if node.Kind == ScalarNode { + if stag == strTag && node.Style&(SingleQuotedStyle|DoubleQuotedStyle|LiteralStyle|FoldedStyle) != 0 { + tag = "" + } else { + rtag, _ := resolve("", node.Value) + if rtag == stag { + tag = "" + } else if stag == strTag { + tag = "" + forceQuoting = true + } + } + } else { + var rtag string + switch node.Kind { + case MappingNode: + rtag = mapTag + case SequenceNode: + rtag = seqTag + } + if rtag == stag { + tag = "" + } + } + } + + switch node.Kind { + case DocumentNode: + yaml_document_start_event_initialize(&e.event, nil, nil, true) + e.event.head_comment = []byte(node.HeadComment) + e.emit() + for _, node := range node.Content { + e.node(node, "") + } + yaml_document_end_event_initialize(&e.event, true) + e.event.foot_comment = []byte(node.FootComment) + e.emit() + + case SequenceNode: + style := yaml_BLOCK_SEQUENCE_STYLE + if node.Style&FlowStyle != 0 { + style = yaml_FLOW_SEQUENCE_STYLE + } + e.must(yaml_sequence_start_event_initialize(&e.event, []byte(node.Anchor), []byte(longTag(tag)), tag == "", style)) + e.event.head_comment = []byte(node.HeadComment) + e.emit() + for _, node := range node.Content { + e.node(node, "") + } + e.must(yaml_sequence_end_event_initialize(&e.event)) + e.event.line_comment = []byte(node.LineComment) + e.event.foot_comment = []byte(node.FootComment) + e.emit() + + case MappingNode: + style := yaml_BLOCK_MAPPING_STYLE + if node.Style&FlowStyle != 0 { + style = yaml_FLOW_MAPPING_STYLE + } + yaml_mapping_start_event_initialize(&e.event, []byte(node.Anchor), []byte(longTag(tag)), tag == "", style) + e.event.tail_comment = []byte(tail) + e.event.head_comment = []byte(node.HeadComment) + e.emit() + + // The tail logic below moves the foot comment of prior keys to the following key, + // since the value for each key may be a nested structure and the foot needs to be + // processed only the entirety of the value is streamed. The last tail is processed + // with the mapping end event. + var tail string + for i := 0; i+1 < len(node.Content); i += 2 { + k := node.Content[i] + foot := k.FootComment + if foot != "" { + kopy := *k + kopy.FootComment = "" + k = &kopy + } + e.node(k, tail) + tail = foot + + v := node.Content[i+1] + e.node(v, "") + } + + yaml_mapping_end_event_initialize(&e.event) + e.event.tail_comment = []byte(tail) + e.event.line_comment = []byte(node.LineComment) + e.event.foot_comment = []byte(node.FootComment) + e.emit() + + case AliasNode: + yaml_alias_event_initialize(&e.event, []byte(node.Value)) + e.event.head_comment = []byte(node.HeadComment) + e.event.line_comment = []byte(node.LineComment) + e.event.foot_comment = []byte(node.FootComment) + e.emit() + + case ScalarNode: + value := node.Value + if !utf8.ValidString(value) { + if stag == binaryTag { + failf("explicitly tagged !!binary data must be base64-encoded") + } + if stag != "" { + failf("cannot marshal invalid UTF-8 data as %s", stag) + } + // It can't be encoded directly as YAML so use a binary tag + // and encode it as base64. + tag = binaryTag + value = encodeBase64(value) + } + + style := yaml_PLAIN_SCALAR_STYLE + switch { + case node.Style&DoubleQuotedStyle != 0: + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + case node.Style&SingleQuotedStyle != 0: + style = yaml_SINGLE_QUOTED_SCALAR_STYLE + case node.Style&LiteralStyle != 0: + style = yaml_LITERAL_SCALAR_STYLE + case node.Style&FoldedStyle != 0: + style = yaml_FOLDED_SCALAR_STYLE + case strings.Contains(value, "\n"): + style = yaml_LITERAL_SCALAR_STYLE + case forceQuoting: + style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } + + e.emitScalar(value, node.Anchor, tag, style, []byte(node.HeadComment), []byte(node.LineComment), []byte(node.FootComment), []byte(tail)) + default: + failf("cannot encode node with unknown kind %d", node.Kind) + } +} diff --git a/vendor/gopkg.in/yaml.v3/go.mod b/vendor/gopkg.in/yaml.v3/go.mod new file mode 100644 index 00000000..f407ea32 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/go.mod @@ -0,0 +1,5 @@ +module "gopkg.in/yaml.v3" + +require ( + "gopkg.in/check.v1" v0.0.0-20161208181325-20d25e280405 +) diff --git a/vendor/gopkg.in/yaml.v3/parserc.go b/vendor/gopkg.in/yaml.v3/parserc.go new file mode 100644 index 00000000..ac66fccc --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/parserc.go @@ -0,0 +1,1249 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +import ( + "bytes" +) + +// The parser implements the following grammar: +// +// stream ::= STREAM-START implicit_document? explicit_document* STREAM-END +// implicit_document ::= block_node DOCUMENT-END* +// explicit_document ::= DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* +// block_node_or_indentless_sequence ::= +// ALIAS +// | properties (block_content | indentless_block_sequence)? +// | block_content +// | indentless_block_sequence +// block_node ::= ALIAS +// | properties block_content? +// | block_content +// flow_node ::= ALIAS +// | properties flow_content? +// | flow_content +// properties ::= TAG ANCHOR? | ANCHOR TAG? +// block_content ::= block_collection | flow_collection | SCALAR +// flow_content ::= flow_collection | SCALAR +// block_collection ::= block_sequence | block_mapping +// flow_collection ::= flow_sequence | flow_mapping +// block_sequence ::= BLOCK-SEQUENCE-START (BLOCK-ENTRY block_node?)* BLOCK-END +// indentless_sequence ::= (BLOCK-ENTRY block_node?)+ +// block_mapping ::= BLOCK-MAPPING_START +// ((KEY block_node_or_indentless_sequence?)? +// (VALUE block_node_or_indentless_sequence?)?)* +// BLOCK-END +// flow_sequence ::= FLOW-SEQUENCE-START +// (flow_sequence_entry FLOW-ENTRY)* +// flow_sequence_entry? +// FLOW-SEQUENCE-END +// flow_sequence_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? +// flow_mapping ::= FLOW-MAPPING-START +// (flow_mapping_entry FLOW-ENTRY)* +// flow_mapping_entry? +// FLOW-MAPPING-END +// flow_mapping_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? + +// Peek the next token in the token queue. +func peek_token(parser *yaml_parser_t) *yaml_token_t { + if parser.token_available || yaml_parser_fetch_more_tokens(parser) { + token := &parser.tokens[parser.tokens_head] + yaml_parser_unfold_comments(parser, token) + return token + } + return nil +} + +// yaml_parser_unfold_comments walks through the comments queue and joins all +// comments behind the position of the provided token into the respective +// top-level comment slices in the parser. +func yaml_parser_unfold_comments(parser *yaml_parser_t, token *yaml_token_t) { + for parser.comments_head < len(parser.comments) && token.start_mark.index >= parser.comments[parser.comments_head].token_mark.index { + comment := &parser.comments[parser.comments_head] + if len(comment.head) > 0 { + if token.typ == yaml_BLOCK_END_TOKEN { + // No heads on ends, so keep comment.head for a follow up token. + break + } + if len(parser.head_comment) > 0 { + parser.head_comment = append(parser.head_comment, '\n') + } + parser.head_comment = append(parser.head_comment, comment.head...) + } + if len(comment.foot) > 0 { + if len(parser.foot_comment) > 0 { + parser.foot_comment = append(parser.foot_comment, '\n') + } + parser.foot_comment = append(parser.foot_comment, comment.foot...) + } + if len(comment.line) > 0 { + if len(parser.line_comment) > 0 { + parser.line_comment = append(parser.line_comment, '\n') + } + parser.line_comment = append(parser.line_comment, comment.line...) + } + *comment = yaml_comment_t{} + parser.comments_head++ + } +} + +// Remove the next token from the queue (must be called after peek_token). +func skip_token(parser *yaml_parser_t) { + parser.token_available = false + parser.tokens_parsed++ + parser.stream_end_produced = parser.tokens[parser.tokens_head].typ == yaml_STREAM_END_TOKEN + parser.tokens_head++ +} + +// Get the next event. +func yaml_parser_parse(parser *yaml_parser_t, event *yaml_event_t) bool { + // Erase the event object. + *event = yaml_event_t{} + + // No events after the end of the stream or error. + if parser.stream_end_produced || parser.error != yaml_NO_ERROR || parser.state == yaml_PARSE_END_STATE { + return true + } + + // Generate the next event. + return yaml_parser_state_machine(parser, event) +} + +// Set parser error. +func yaml_parser_set_parser_error(parser *yaml_parser_t, problem string, problem_mark yaml_mark_t) bool { + parser.error = yaml_PARSER_ERROR + parser.problem = problem + parser.problem_mark = problem_mark + return false +} + +func yaml_parser_set_parser_error_context(parser *yaml_parser_t, context string, context_mark yaml_mark_t, problem string, problem_mark yaml_mark_t) bool { + parser.error = yaml_PARSER_ERROR + parser.context = context + parser.context_mark = context_mark + parser.problem = problem + parser.problem_mark = problem_mark + return false +} + +// State dispatcher. +func yaml_parser_state_machine(parser *yaml_parser_t, event *yaml_event_t) bool { + //trace("yaml_parser_state_machine", "state:", parser.state.String()) + + switch parser.state { + case yaml_PARSE_STREAM_START_STATE: + return yaml_parser_parse_stream_start(parser, event) + + case yaml_PARSE_IMPLICIT_DOCUMENT_START_STATE: + return yaml_parser_parse_document_start(parser, event, true) + + case yaml_PARSE_DOCUMENT_START_STATE: + return yaml_parser_parse_document_start(parser, event, false) + + case yaml_PARSE_DOCUMENT_CONTENT_STATE: + return yaml_parser_parse_document_content(parser, event) + + case yaml_PARSE_DOCUMENT_END_STATE: + return yaml_parser_parse_document_end(parser, event) + + case yaml_PARSE_BLOCK_NODE_STATE: + return yaml_parser_parse_node(parser, event, true, false) + + case yaml_PARSE_BLOCK_NODE_OR_INDENTLESS_SEQUENCE_STATE: + return yaml_parser_parse_node(parser, event, true, true) + + case yaml_PARSE_FLOW_NODE_STATE: + return yaml_parser_parse_node(parser, event, false, false) + + case yaml_PARSE_BLOCK_SEQUENCE_FIRST_ENTRY_STATE: + return yaml_parser_parse_block_sequence_entry(parser, event, true) + + case yaml_PARSE_BLOCK_SEQUENCE_ENTRY_STATE: + return yaml_parser_parse_block_sequence_entry(parser, event, false) + + case yaml_PARSE_INDENTLESS_SEQUENCE_ENTRY_STATE: + return yaml_parser_parse_indentless_sequence_entry(parser, event) + + case yaml_PARSE_BLOCK_MAPPING_FIRST_KEY_STATE: + return yaml_parser_parse_block_mapping_key(parser, event, true) + + case yaml_PARSE_BLOCK_MAPPING_KEY_STATE: + return yaml_parser_parse_block_mapping_key(parser, event, false) + + case yaml_PARSE_BLOCK_MAPPING_VALUE_STATE: + return yaml_parser_parse_block_mapping_value(parser, event) + + case yaml_PARSE_FLOW_SEQUENCE_FIRST_ENTRY_STATE: + return yaml_parser_parse_flow_sequence_entry(parser, event, true) + + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_STATE: + return yaml_parser_parse_flow_sequence_entry(parser, event, false) + + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_KEY_STATE: + return yaml_parser_parse_flow_sequence_entry_mapping_key(parser, event) + + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_VALUE_STATE: + return yaml_parser_parse_flow_sequence_entry_mapping_value(parser, event) + + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_END_STATE: + return yaml_parser_parse_flow_sequence_entry_mapping_end(parser, event) + + case yaml_PARSE_FLOW_MAPPING_FIRST_KEY_STATE: + return yaml_parser_parse_flow_mapping_key(parser, event, true) + + case yaml_PARSE_FLOW_MAPPING_KEY_STATE: + return yaml_parser_parse_flow_mapping_key(parser, event, false) + + case yaml_PARSE_FLOW_MAPPING_VALUE_STATE: + return yaml_parser_parse_flow_mapping_value(parser, event, false) + + case yaml_PARSE_FLOW_MAPPING_EMPTY_VALUE_STATE: + return yaml_parser_parse_flow_mapping_value(parser, event, true) + + default: + panic("invalid parser state") + } +} + +// Parse the production: +// stream ::= STREAM-START implicit_document? explicit_document* STREAM-END +// ************ +func yaml_parser_parse_stream_start(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_STREAM_START_TOKEN { + return yaml_parser_set_parser_error(parser, "did not find expected ", token.start_mark) + } + parser.state = yaml_PARSE_IMPLICIT_DOCUMENT_START_STATE + *event = yaml_event_t{ + typ: yaml_STREAM_START_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + encoding: token.encoding, + } + skip_token(parser) + return true +} + +// Parse the productions: +// implicit_document ::= block_node DOCUMENT-END* +// * +// explicit_document ::= DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* +// ************************* +func yaml_parser_parse_document_start(parser *yaml_parser_t, event *yaml_event_t, implicit bool) bool { + + token := peek_token(parser) + if token == nil { + return false + } + + // Parse extra document end indicators. + if !implicit { + for token.typ == yaml_DOCUMENT_END_TOKEN { + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + } + } + + if implicit && token.typ != yaml_VERSION_DIRECTIVE_TOKEN && + token.typ != yaml_TAG_DIRECTIVE_TOKEN && + token.typ != yaml_DOCUMENT_START_TOKEN && + token.typ != yaml_STREAM_END_TOKEN { + // Parse an implicit document. + if !yaml_parser_process_directives(parser, nil, nil) { + return false + } + parser.states = append(parser.states, yaml_PARSE_DOCUMENT_END_STATE) + parser.state = yaml_PARSE_BLOCK_NODE_STATE + + var head_comment []byte + if len(parser.head_comment) > 0 { + // [Go] Scan the header comment backwards, and if an empty line is found, break + // the header so the part before the last empty line goes into the + // document header, while the bottom of it goes into a follow up event. + for i := len(parser.head_comment) - 1; i > 0; i-- { + if parser.head_comment[i] == '\n' { + if i == len(parser.head_comment)-1 { + head_comment = parser.head_comment[:i] + parser.head_comment = parser.head_comment[i+1:] + break + } else if parser.head_comment[i-1] == '\n' { + head_comment = parser.head_comment[:i-1] + parser.head_comment = parser.head_comment[i+1:] + break + } + } + } + } + + *event = yaml_event_t{ + typ: yaml_DOCUMENT_START_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + + head_comment: head_comment, + } + + } else if token.typ != yaml_STREAM_END_TOKEN { + // Parse an explicit document. + var version_directive *yaml_version_directive_t + var tag_directives []yaml_tag_directive_t + start_mark := token.start_mark + if !yaml_parser_process_directives(parser, &version_directive, &tag_directives) { + return false + } + token = peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_DOCUMENT_START_TOKEN { + yaml_parser_set_parser_error(parser, + "did not find expected ", token.start_mark) + return false + } + parser.states = append(parser.states, yaml_PARSE_DOCUMENT_END_STATE) + parser.state = yaml_PARSE_DOCUMENT_CONTENT_STATE + end_mark := token.end_mark + + *event = yaml_event_t{ + typ: yaml_DOCUMENT_START_EVENT, + start_mark: start_mark, + end_mark: end_mark, + version_directive: version_directive, + tag_directives: tag_directives, + implicit: false, + } + skip_token(parser) + + } else { + // Parse the stream end. + parser.state = yaml_PARSE_END_STATE + *event = yaml_event_t{ + typ: yaml_STREAM_END_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + } + skip_token(parser) + } + + return true +} + +// Parse the productions: +// explicit_document ::= DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* +// *********** +// +func yaml_parser_parse_document_content(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + + if token.typ == yaml_VERSION_DIRECTIVE_TOKEN || + token.typ == yaml_TAG_DIRECTIVE_TOKEN || + token.typ == yaml_DOCUMENT_START_TOKEN || + token.typ == yaml_DOCUMENT_END_TOKEN || + token.typ == yaml_STREAM_END_TOKEN { + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + return yaml_parser_process_empty_scalar(parser, event, + token.start_mark) + } + return yaml_parser_parse_node(parser, event, true, false) +} + +// Parse the productions: +// implicit_document ::= block_node DOCUMENT-END* +// ************* +// explicit_document ::= DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* +// +func yaml_parser_parse_document_end(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + + start_mark := token.start_mark + end_mark := token.start_mark + + implicit := true + if token.typ == yaml_DOCUMENT_END_TOKEN { + end_mark = token.end_mark + skip_token(parser) + implicit = false + } + + parser.tag_directives = parser.tag_directives[:0] + + parser.state = yaml_PARSE_DOCUMENT_START_STATE + *event = yaml_event_t{ + typ: yaml_DOCUMENT_END_EVENT, + start_mark: start_mark, + end_mark: end_mark, + implicit: implicit, + } + yaml_parser_set_event_comments(parser, event) + if len(event.head_comment) > 0 && len(event.foot_comment) == 0 { + event.foot_comment = event.head_comment + event.head_comment = nil + } + return true +} + +func yaml_parser_set_event_comments(parser *yaml_parser_t, event *yaml_event_t) { + event.head_comment = parser.head_comment + event.line_comment = parser.line_comment + event.foot_comment = parser.foot_comment + parser.head_comment = nil + parser.line_comment = nil + parser.foot_comment = nil + parser.tail_comment = nil + parser.stem_comment = nil +} + +// Parse the productions: +// block_node_or_indentless_sequence ::= +// ALIAS +// ***** +// | properties (block_content | indentless_block_sequence)? +// ********** * +// | block_content | indentless_block_sequence +// * +// block_node ::= ALIAS +// ***** +// | properties block_content? +// ********** * +// | block_content +// * +// flow_node ::= ALIAS +// ***** +// | properties flow_content? +// ********** * +// | flow_content +// * +// properties ::= TAG ANCHOR? | ANCHOR TAG? +// ************************* +// block_content ::= block_collection | flow_collection | SCALAR +// ****** +// flow_content ::= flow_collection | SCALAR +// ****** +func yaml_parser_parse_node(parser *yaml_parser_t, event *yaml_event_t, block, indentless_sequence bool) bool { + //defer trace("yaml_parser_parse_node", "block:", block, "indentless_sequence:", indentless_sequence)() + + token := peek_token(parser) + if token == nil { + return false + } + + if token.typ == yaml_ALIAS_TOKEN { + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + *event = yaml_event_t{ + typ: yaml_ALIAS_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + anchor: token.value, + } + yaml_parser_set_event_comments(parser, event) + skip_token(parser) + return true + } + + start_mark := token.start_mark + end_mark := token.start_mark + + var tag_token bool + var tag_handle, tag_suffix, anchor []byte + var tag_mark yaml_mark_t + if token.typ == yaml_ANCHOR_TOKEN { + anchor = token.value + start_mark = token.start_mark + end_mark = token.end_mark + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ == yaml_TAG_TOKEN { + tag_token = true + tag_handle = token.value + tag_suffix = token.suffix + tag_mark = token.start_mark + end_mark = token.end_mark + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + } + } else if token.typ == yaml_TAG_TOKEN { + tag_token = true + tag_handle = token.value + tag_suffix = token.suffix + start_mark = token.start_mark + tag_mark = token.start_mark + end_mark = token.end_mark + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ == yaml_ANCHOR_TOKEN { + anchor = token.value + end_mark = token.end_mark + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + } + } + + var tag []byte + if tag_token { + if len(tag_handle) == 0 { + tag = tag_suffix + tag_suffix = nil + } else { + for i := range parser.tag_directives { + if bytes.Equal(parser.tag_directives[i].handle, tag_handle) { + tag = append([]byte(nil), parser.tag_directives[i].prefix...) + tag = append(tag, tag_suffix...) + break + } + } + if len(tag) == 0 { + yaml_parser_set_parser_error_context(parser, + "while parsing a node", start_mark, + "found undefined tag handle", tag_mark) + return false + } + } + } + + implicit := len(tag) == 0 + if indentless_sequence && token.typ == yaml_BLOCK_ENTRY_TOKEN { + end_mark = token.end_mark + parser.state = yaml_PARSE_INDENTLESS_SEQUENCE_ENTRY_STATE + *event = yaml_event_t{ + typ: yaml_SEQUENCE_START_EVENT, + start_mark: start_mark, + end_mark: end_mark, + anchor: anchor, + tag: tag, + implicit: implicit, + style: yaml_style_t(yaml_BLOCK_SEQUENCE_STYLE), + } + return true + } + if token.typ == yaml_SCALAR_TOKEN { + var plain_implicit, quoted_implicit bool + end_mark = token.end_mark + if (len(tag) == 0 && token.style == yaml_PLAIN_SCALAR_STYLE) || (len(tag) == 1 && tag[0] == '!') { + plain_implicit = true + } else if len(tag) == 0 { + quoted_implicit = true + } + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + + *event = yaml_event_t{ + typ: yaml_SCALAR_EVENT, + start_mark: start_mark, + end_mark: end_mark, + anchor: anchor, + tag: tag, + value: token.value, + implicit: plain_implicit, + quoted_implicit: quoted_implicit, + style: yaml_style_t(token.style), + } + yaml_parser_set_event_comments(parser, event) + skip_token(parser) + return true + } + if token.typ == yaml_FLOW_SEQUENCE_START_TOKEN { + // [Go] Some of the events below can be merged as they differ only on style. + end_mark = token.end_mark + parser.state = yaml_PARSE_FLOW_SEQUENCE_FIRST_ENTRY_STATE + *event = yaml_event_t{ + typ: yaml_SEQUENCE_START_EVENT, + start_mark: start_mark, + end_mark: end_mark, + anchor: anchor, + tag: tag, + implicit: implicit, + style: yaml_style_t(yaml_FLOW_SEQUENCE_STYLE), + } + yaml_parser_set_event_comments(parser, event) + return true + } + if token.typ == yaml_FLOW_MAPPING_START_TOKEN { + end_mark = token.end_mark + parser.state = yaml_PARSE_FLOW_MAPPING_FIRST_KEY_STATE + *event = yaml_event_t{ + typ: yaml_MAPPING_START_EVENT, + start_mark: start_mark, + end_mark: end_mark, + anchor: anchor, + tag: tag, + implicit: implicit, + style: yaml_style_t(yaml_FLOW_MAPPING_STYLE), + } + yaml_parser_set_event_comments(parser, event) + return true + } + if block && token.typ == yaml_BLOCK_SEQUENCE_START_TOKEN { + end_mark = token.end_mark + parser.state = yaml_PARSE_BLOCK_SEQUENCE_FIRST_ENTRY_STATE + *event = yaml_event_t{ + typ: yaml_SEQUENCE_START_EVENT, + start_mark: start_mark, + end_mark: end_mark, + anchor: anchor, + tag: tag, + implicit: implicit, + style: yaml_style_t(yaml_BLOCK_SEQUENCE_STYLE), + } + if parser.stem_comment != nil { + event.head_comment = parser.stem_comment + parser.stem_comment = nil + } + return true + } + if block && token.typ == yaml_BLOCK_MAPPING_START_TOKEN { + end_mark = token.end_mark + parser.state = yaml_PARSE_BLOCK_MAPPING_FIRST_KEY_STATE + *event = yaml_event_t{ + typ: yaml_MAPPING_START_EVENT, + start_mark: start_mark, + end_mark: end_mark, + anchor: anchor, + tag: tag, + implicit: implicit, + style: yaml_style_t(yaml_BLOCK_MAPPING_STYLE), + } + if parser.stem_comment != nil { + event.head_comment = parser.stem_comment + parser.stem_comment = nil + } + return true + } + if len(anchor) > 0 || len(tag) > 0 { + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + + *event = yaml_event_t{ + typ: yaml_SCALAR_EVENT, + start_mark: start_mark, + end_mark: end_mark, + anchor: anchor, + tag: tag, + implicit: implicit, + quoted_implicit: false, + style: yaml_style_t(yaml_PLAIN_SCALAR_STYLE), + } + return true + } + + context := "while parsing a flow node" + if block { + context = "while parsing a block node" + } + yaml_parser_set_parser_error_context(parser, context, start_mark, + "did not find expected node content", token.start_mark) + return false +} + +// Parse the productions: +// block_sequence ::= BLOCK-SEQUENCE-START (BLOCK-ENTRY block_node?)* BLOCK-END +// ******************** *********** * ********* +// +func yaml_parser_parse_block_sequence_entry(parser *yaml_parser_t, event *yaml_event_t, first bool) bool { + if first { + token := peek_token(parser) + parser.marks = append(parser.marks, token.start_mark) + skip_token(parser) + } + + token := peek_token(parser) + if token == nil { + return false + } + + if token.typ == yaml_BLOCK_ENTRY_TOKEN { + mark := token.end_mark + prior_head_len := len(parser.head_comment) + skip_token(parser) + yaml_parser_split_stem_comment(parser, prior_head_len) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_BLOCK_ENTRY_TOKEN && token.typ != yaml_BLOCK_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_BLOCK_SEQUENCE_ENTRY_STATE) + return yaml_parser_parse_node(parser, event, true, false) + } else { + parser.state = yaml_PARSE_BLOCK_SEQUENCE_ENTRY_STATE + return yaml_parser_process_empty_scalar(parser, event, mark) + } + } + if token.typ == yaml_BLOCK_END_TOKEN { + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + + *event = yaml_event_t{ + typ: yaml_SEQUENCE_END_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + } + + skip_token(parser) + return true + } + + context_mark := parser.marks[len(parser.marks)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + return yaml_parser_set_parser_error_context(parser, + "while parsing a block collection", context_mark, + "did not find expected '-' indicator", token.start_mark) +} + +// Parse the productions: +// indentless_sequence ::= (BLOCK-ENTRY block_node?)+ +// *********** * +func yaml_parser_parse_indentless_sequence_entry(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + + if token.typ == yaml_BLOCK_ENTRY_TOKEN { + mark := token.end_mark + prior_head_len := len(parser.head_comment) + skip_token(parser) + yaml_parser_split_stem_comment(parser, prior_head_len) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_BLOCK_ENTRY_TOKEN && + token.typ != yaml_KEY_TOKEN && + token.typ != yaml_VALUE_TOKEN && + token.typ != yaml_BLOCK_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_INDENTLESS_SEQUENCE_ENTRY_STATE) + return yaml_parser_parse_node(parser, event, true, false) + } + parser.state = yaml_PARSE_INDENTLESS_SEQUENCE_ENTRY_STATE + return yaml_parser_process_empty_scalar(parser, event, mark) + } + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + + *event = yaml_event_t{ + typ: yaml_SEQUENCE_END_EVENT, + start_mark: token.start_mark, + end_mark: token.start_mark, // [Go] Shouldn't this be token.end_mark? + } + return true +} + +// Split stem comment from head comment. +// +// When a sequence or map is found under a sequence entry, the former head comment +// is assigned to the underlying sequence or map as a whole, not the individual +// sequence or map entry as would be expected otherwise. To handle this case the +// previous head comment is moved aside as the stem comment. +func yaml_parser_split_stem_comment(parser *yaml_parser_t, stem_len int) { + if stem_len == 0 { + return + } + + token := peek_token(parser) + if token.typ != yaml_BLOCK_SEQUENCE_START_TOKEN && token.typ != yaml_BLOCK_MAPPING_START_TOKEN { + return + } + + parser.stem_comment = parser.head_comment[:stem_len] + if len(parser.head_comment) == stem_len { + parser.head_comment = nil + } else { + // Copy suffix to prevent very strange bugs if someone ever appends + // further bytes to the prefix in the stem_comment slice above. + parser.head_comment = append([]byte(nil), parser.head_comment[stem_len+1:]...) + } +} + +// Parse the productions: +// block_mapping ::= BLOCK-MAPPING_START +// ******************* +// ((KEY block_node_or_indentless_sequence?)? +// *** * +// (VALUE block_node_or_indentless_sequence?)?)* +// +// BLOCK-END +// ********* +// +func yaml_parser_parse_block_mapping_key(parser *yaml_parser_t, event *yaml_event_t, first bool) bool { + if first { + token := peek_token(parser) + parser.marks = append(parser.marks, token.start_mark) + skip_token(parser) + } + + token := peek_token(parser) + if token == nil { + return false + } + + // [Go] A tail comment was left from the prior mapping value processed. Emit an event + // as it needs to be processed with that value and not the following key. + if len(parser.tail_comment) > 0 { + *event = yaml_event_t{ + typ: yaml_TAIL_COMMENT_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + foot_comment: parser.tail_comment, + } + parser.tail_comment = nil + return true + } + + if token.typ == yaml_KEY_TOKEN { + mark := token.end_mark + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_KEY_TOKEN && + token.typ != yaml_VALUE_TOKEN && + token.typ != yaml_BLOCK_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_BLOCK_MAPPING_VALUE_STATE) + return yaml_parser_parse_node(parser, event, true, true) + } else { + parser.state = yaml_PARSE_BLOCK_MAPPING_VALUE_STATE + return yaml_parser_process_empty_scalar(parser, event, mark) + } + } else if token.typ == yaml_BLOCK_END_TOKEN { + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + *event = yaml_event_t{ + typ: yaml_MAPPING_END_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + } + yaml_parser_set_event_comments(parser, event) + skip_token(parser) + return true + } + + context_mark := parser.marks[len(parser.marks)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + return yaml_parser_set_parser_error_context(parser, + "while parsing a block mapping", context_mark, + "did not find expected key", token.start_mark) +} + +// Parse the productions: +// block_mapping ::= BLOCK-MAPPING_START +// +// ((KEY block_node_or_indentless_sequence?)? +// +// (VALUE block_node_or_indentless_sequence?)?)* +// ***** * +// BLOCK-END +// +// +func yaml_parser_parse_block_mapping_value(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + if token.typ == yaml_VALUE_TOKEN { + mark := token.end_mark + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_KEY_TOKEN && + token.typ != yaml_VALUE_TOKEN && + token.typ != yaml_BLOCK_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_BLOCK_MAPPING_KEY_STATE) + return yaml_parser_parse_node(parser, event, true, true) + } + parser.state = yaml_PARSE_BLOCK_MAPPING_KEY_STATE + return yaml_parser_process_empty_scalar(parser, event, mark) + } + parser.state = yaml_PARSE_BLOCK_MAPPING_KEY_STATE + return yaml_parser_process_empty_scalar(parser, event, token.start_mark) +} + +// Parse the productions: +// flow_sequence ::= FLOW-SEQUENCE-START +// ******************* +// (flow_sequence_entry FLOW-ENTRY)* +// * ********** +// flow_sequence_entry? +// * +// FLOW-SEQUENCE-END +// ***************** +// flow_sequence_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? +// * +// +func yaml_parser_parse_flow_sequence_entry(parser *yaml_parser_t, event *yaml_event_t, first bool) bool { + if first { + token := peek_token(parser) + parser.marks = append(parser.marks, token.start_mark) + skip_token(parser) + } + token := peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_FLOW_SEQUENCE_END_TOKEN { + if !first { + if token.typ == yaml_FLOW_ENTRY_TOKEN { + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + } else { + context_mark := parser.marks[len(parser.marks)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + return yaml_parser_set_parser_error_context(parser, + "while parsing a flow sequence", context_mark, + "did not find expected ',' or ']'", token.start_mark) + } + } + + if token.typ == yaml_KEY_TOKEN { + parser.state = yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_KEY_STATE + *event = yaml_event_t{ + typ: yaml_MAPPING_START_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + implicit: true, + style: yaml_style_t(yaml_FLOW_MAPPING_STYLE), + } + skip_token(parser) + return true + } else if token.typ != yaml_FLOW_SEQUENCE_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_FLOW_SEQUENCE_ENTRY_STATE) + return yaml_parser_parse_node(parser, event, false, false) + } + } + + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + + *event = yaml_event_t{ + typ: yaml_SEQUENCE_END_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + } + yaml_parser_set_event_comments(parser, event) + + skip_token(parser) + return true +} + +// +// Parse the productions: +// flow_sequence_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? +// *** * +// +func yaml_parser_parse_flow_sequence_entry_mapping_key(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_VALUE_TOKEN && + token.typ != yaml_FLOW_ENTRY_TOKEN && + token.typ != yaml_FLOW_SEQUENCE_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_VALUE_STATE) + return yaml_parser_parse_node(parser, event, false, false) + } + mark := token.end_mark + skip_token(parser) + parser.state = yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_VALUE_STATE + return yaml_parser_process_empty_scalar(parser, event, mark) +} + +// Parse the productions: +// flow_sequence_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? +// ***** * +// +func yaml_parser_parse_flow_sequence_entry_mapping_value(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + if token.typ == yaml_VALUE_TOKEN { + skip_token(parser) + token := peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_FLOW_ENTRY_TOKEN && token.typ != yaml_FLOW_SEQUENCE_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_END_STATE) + return yaml_parser_parse_node(parser, event, false, false) + } + } + parser.state = yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_END_STATE + return yaml_parser_process_empty_scalar(parser, event, token.start_mark) +} + +// Parse the productions: +// flow_sequence_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? +// * +// +func yaml_parser_parse_flow_sequence_entry_mapping_end(parser *yaml_parser_t, event *yaml_event_t) bool { + token := peek_token(parser) + if token == nil { + return false + } + parser.state = yaml_PARSE_FLOW_SEQUENCE_ENTRY_STATE + *event = yaml_event_t{ + typ: yaml_MAPPING_END_EVENT, + start_mark: token.start_mark, + end_mark: token.start_mark, // [Go] Shouldn't this be end_mark? + } + return true +} + +// Parse the productions: +// flow_mapping ::= FLOW-MAPPING-START +// ****************** +// (flow_mapping_entry FLOW-ENTRY)* +// * ********** +// flow_mapping_entry? +// ****************** +// FLOW-MAPPING-END +// **************** +// flow_mapping_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? +// * *** * +// +func yaml_parser_parse_flow_mapping_key(parser *yaml_parser_t, event *yaml_event_t, first bool) bool { + if first { + token := peek_token(parser) + parser.marks = append(parser.marks, token.start_mark) + skip_token(parser) + } + + token := peek_token(parser) + if token == nil { + return false + } + + if token.typ != yaml_FLOW_MAPPING_END_TOKEN { + if !first { + if token.typ == yaml_FLOW_ENTRY_TOKEN { + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + } else { + context_mark := parser.marks[len(parser.marks)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + return yaml_parser_set_parser_error_context(parser, + "while parsing a flow mapping", context_mark, + "did not find expected ',' or '}'", token.start_mark) + } + } + + if token.typ == yaml_KEY_TOKEN { + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_VALUE_TOKEN && + token.typ != yaml_FLOW_ENTRY_TOKEN && + token.typ != yaml_FLOW_MAPPING_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_FLOW_MAPPING_VALUE_STATE) + return yaml_parser_parse_node(parser, event, false, false) + } else { + parser.state = yaml_PARSE_FLOW_MAPPING_VALUE_STATE + return yaml_parser_process_empty_scalar(parser, event, token.start_mark) + } + } else if token.typ != yaml_FLOW_MAPPING_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_FLOW_MAPPING_EMPTY_VALUE_STATE) + return yaml_parser_parse_node(parser, event, false, false) + } + } + + parser.state = parser.states[len(parser.states)-1] + parser.states = parser.states[:len(parser.states)-1] + parser.marks = parser.marks[:len(parser.marks)-1] + *event = yaml_event_t{ + typ: yaml_MAPPING_END_EVENT, + start_mark: token.start_mark, + end_mark: token.end_mark, + } + yaml_parser_set_event_comments(parser, event) + skip_token(parser) + return true +} + +// Parse the productions: +// flow_mapping_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? +// * ***** * +// +func yaml_parser_parse_flow_mapping_value(parser *yaml_parser_t, event *yaml_event_t, empty bool) bool { + token := peek_token(parser) + if token == nil { + return false + } + if empty { + parser.state = yaml_PARSE_FLOW_MAPPING_KEY_STATE + return yaml_parser_process_empty_scalar(parser, event, token.start_mark) + } + if token.typ == yaml_VALUE_TOKEN { + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + if token.typ != yaml_FLOW_ENTRY_TOKEN && token.typ != yaml_FLOW_MAPPING_END_TOKEN { + parser.states = append(parser.states, yaml_PARSE_FLOW_MAPPING_KEY_STATE) + return yaml_parser_parse_node(parser, event, false, false) + } + } + parser.state = yaml_PARSE_FLOW_MAPPING_KEY_STATE + return yaml_parser_process_empty_scalar(parser, event, token.start_mark) +} + +// Generate an empty scalar event. +func yaml_parser_process_empty_scalar(parser *yaml_parser_t, event *yaml_event_t, mark yaml_mark_t) bool { + *event = yaml_event_t{ + typ: yaml_SCALAR_EVENT, + start_mark: mark, + end_mark: mark, + value: nil, // Empty + implicit: true, + style: yaml_style_t(yaml_PLAIN_SCALAR_STYLE), + } + return true +} + +var default_tag_directives = []yaml_tag_directive_t{ + {[]byte("!"), []byte("!")}, + {[]byte("!!"), []byte("tag:yaml.org,2002:")}, +} + +// Parse directives. +func yaml_parser_process_directives(parser *yaml_parser_t, + version_directive_ref **yaml_version_directive_t, + tag_directives_ref *[]yaml_tag_directive_t) bool { + + var version_directive *yaml_version_directive_t + var tag_directives []yaml_tag_directive_t + + token := peek_token(parser) + if token == nil { + return false + } + + for token.typ == yaml_VERSION_DIRECTIVE_TOKEN || token.typ == yaml_TAG_DIRECTIVE_TOKEN { + if token.typ == yaml_VERSION_DIRECTIVE_TOKEN { + if version_directive != nil { + yaml_parser_set_parser_error(parser, + "found duplicate %YAML directive", token.start_mark) + return false + } + if token.major != 1 || token.minor != 1 { + yaml_parser_set_parser_error(parser, + "found incompatible YAML document", token.start_mark) + return false + } + version_directive = &yaml_version_directive_t{ + major: token.major, + minor: token.minor, + } + } else if token.typ == yaml_TAG_DIRECTIVE_TOKEN { + value := yaml_tag_directive_t{ + handle: token.value, + prefix: token.prefix, + } + if !yaml_parser_append_tag_directive(parser, value, false, token.start_mark) { + return false + } + tag_directives = append(tag_directives, value) + } + + skip_token(parser) + token = peek_token(parser) + if token == nil { + return false + } + } + + for i := range default_tag_directives { + if !yaml_parser_append_tag_directive(parser, default_tag_directives[i], true, token.start_mark) { + return false + } + } + + if version_directive_ref != nil { + *version_directive_ref = version_directive + } + if tag_directives_ref != nil { + *tag_directives_ref = tag_directives + } + return true +} + +// Append a tag directive to the directives stack. +func yaml_parser_append_tag_directive(parser *yaml_parser_t, value yaml_tag_directive_t, allow_duplicates bool, mark yaml_mark_t) bool { + for i := range parser.tag_directives { + if bytes.Equal(value.handle, parser.tag_directives[i].handle) { + if allow_duplicates { + return true + } + return yaml_parser_set_parser_error(parser, "found duplicate %TAG directive", mark) + } + } + + // [Go] I suspect the copy is unnecessary. This was likely done + // because there was no way to track ownership of the data. + value_copy := yaml_tag_directive_t{ + handle: make([]byte, len(value.handle)), + prefix: make([]byte, len(value.prefix)), + } + copy(value_copy.handle, value.handle) + copy(value_copy.prefix, value.prefix) + parser.tag_directives = append(parser.tag_directives, value_copy) + return true +} diff --git a/vendor/gopkg.in/yaml.v3/readerc.go b/vendor/gopkg.in/yaml.v3/readerc.go new file mode 100644 index 00000000..b7de0a89 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/readerc.go @@ -0,0 +1,434 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +import ( + "io" +) + +// Set the reader error and return 0. +func yaml_parser_set_reader_error(parser *yaml_parser_t, problem string, offset int, value int) bool { + parser.error = yaml_READER_ERROR + parser.problem = problem + parser.problem_offset = offset + parser.problem_value = value + return false +} + +// Byte order marks. +const ( + bom_UTF8 = "\xef\xbb\xbf" + bom_UTF16LE = "\xff\xfe" + bom_UTF16BE = "\xfe\xff" +) + +// Determine the input stream encoding by checking the BOM symbol. If no BOM is +// found, the UTF-8 encoding is assumed. Return 1 on success, 0 on failure. +func yaml_parser_determine_encoding(parser *yaml_parser_t) bool { + // Ensure that we had enough bytes in the raw buffer. + for !parser.eof && len(parser.raw_buffer)-parser.raw_buffer_pos < 3 { + if !yaml_parser_update_raw_buffer(parser) { + return false + } + } + + // Determine the encoding. + buf := parser.raw_buffer + pos := parser.raw_buffer_pos + avail := len(buf) - pos + if avail >= 2 && buf[pos] == bom_UTF16LE[0] && buf[pos+1] == bom_UTF16LE[1] { + parser.encoding = yaml_UTF16LE_ENCODING + parser.raw_buffer_pos += 2 + parser.offset += 2 + } else if avail >= 2 && buf[pos] == bom_UTF16BE[0] && buf[pos+1] == bom_UTF16BE[1] { + parser.encoding = yaml_UTF16BE_ENCODING + parser.raw_buffer_pos += 2 + parser.offset += 2 + } else if avail >= 3 && buf[pos] == bom_UTF8[0] && buf[pos+1] == bom_UTF8[1] && buf[pos+2] == bom_UTF8[2] { + parser.encoding = yaml_UTF8_ENCODING + parser.raw_buffer_pos += 3 + parser.offset += 3 + } else { + parser.encoding = yaml_UTF8_ENCODING + } + return true +} + +// Update the raw buffer. +func yaml_parser_update_raw_buffer(parser *yaml_parser_t) bool { + size_read := 0 + + // Return if the raw buffer is full. + if parser.raw_buffer_pos == 0 && len(parser.raw_buffer) == cap(parser.raw_buffer) { + return true + } + + // Return on EOF. + if parser.eof { + return true + } + + // Move the remaining bytes in the raw buffer to the beginning. + if parser.raw_buffer_pos > 0 && parser.raw_buffer_pos < len(parser.raw_buffer) { + copy(parser.raw_buffer, parser.raw_buffer[parser.raw_buffer_pos:]) + } + parser.raw_buffer = parser.raw_buffer[:len(parser.raw_buffer)-parser.raw_buffer_pos] + parser.raw_buffer_pos = 0 + + // Call the read handler to fill the buffer. + size_read, err := parser.read_handler(parser, parser.raw_buffer[len(parser.raw_buffer):cap(parser.raw_buffer)]) + parser.raw_buffer = parser.raw_buffer[:len(parser.raw_buffer)+size_read] + if err == io.EOF { + parser.eof = true + } else if err != nil { + return yaml_parser_set_reader_error(parser, "input error: "+err.Error(), parser.offset, -1) + } + return true +} + +// Ensure that the buffer contains at least `length` characters. +// Return true on success, false on failure. +// +// The length is supposed to be significantly less that the buffer size. +func yaml_parser_update_buffer(parser *yaml_parser_t, length int) bool { + if parser.read_handler == nil { + panic("read handler must be set") + } + + // [Go] This function was changed to guarantee the requested length size at EOF. + // The fact we need to do this is pretty awful, but the description above implies + // for that to be the case, and there are tests + + // If the EOF flag is set and the raw buffer is empty, do nothing. + if parser.eof && parser.raw_buffer_pos == len(parser.raw_buffer) { + // [Go] ACTUALLY! Read the documentation of this function above. + // This is just broken. To return true, we need to have the + // given length in the buffer. Not doing that means every single + // check that calls this function to make sure the buffer has a + // given length is Go) panicking; or C) accessing invalid memory. + //return true + } + + // Return if the buffer contains enough characters. + if parser.unread >= length { + return true + } + + // Determine the input encoding if it is not known yet. + if parser.encoding == yaml_ANY_ENCODING { + if !yaml_parser_determine_encoding(parser) { + return false + } + } + + // Move the unread characters to the beginning of the buffer. + buffer_len := len(parser.buffer) + if parser.buffer_pos > 0 && parser.buffer_pos < buffer_len { + copy(parser.buffer, parser.buffer[parser.buffer_pos:]) + buffer_len -= parser.buffer_pos + parser.buffer_pos = 0 + } else if parser.buffer_pos == buffer_len { + buffer_len = 0 + parser.buffer_pos = 0 + } + + // Open the whole buffer for writing, and cut it before returning. + parser.buffer = parser.buffer[:cap(parser.buffer)] + + // Fill the buffer until it has enough characters. + first := true + for parser.unread < length { + + // Fill the raw buffer if necessary. + if !first || parser.raw_buffer_pos == len(parser.raw_buffer) { + if !yaml_parser_update_raw_buffer(parser) { + parser.buffer = parser.buffer[:buffer_len] + return false + } + } + first = false + + // Decode the raw buffer. + inner: + for parser.raw_buffer_pos != len(parser.raw_buffer) { + var value rune + var width int + + raw_unread := len(parser.raw_buffer) - parser.raw_buffer_pos + + // Decode the next character. + switch parser.encoding { + case yaml_UTF8_ENCODING: + // Decode a UTF-8 character. Check RFC 3629 + // (http://www.ietf.org/rfc/rfc3629.txt) for more details. + // + // The following table (taken from the RFC) is used for + // decoding. + // + // Char. number range | UTF-8 octet sequence + // (hexadecimal) | (binary) + // --------------------+------------------------------------ + // 0000 0000-0000 007F | 0xxxxxxx + // 0000 0080-0000 07FF | 110xxxxx 10xxxxxx + // 0000 0800-0000 FFFF | 1110xxxx 10xxxxxx 10xxxxxx + // 0001 0000-0010 FFFF | 11110xxx 10xxxxxx 10xxxxxx 10xxxxxx + // + // Additionally, the characters in the range 0xD800-0xDFFF + // are prohibited as they are reserved for use with UTF-16 + // surrogate pairs. + + // Determine the length of the UTF-8 sequence. + octet := parser.raw_buffer[parser.raw_buffer_pos] + switch { + case octet&0x80 == 0x00: + width = 1 + case octet&0xE0 == 0xC0: + width = 2 + case octet&0xF0 == 0xE0: + width = 3 + case octet&0xF8 == 0xF0: + width = 4 + default: + // The leading octet is invalid. + return yaml_parser_set_reader_error(parser, + "invalid leading UTF-8 octet", + parser.offset, int(octet)) + } + + // Check if the raw buffer contains an incomplete character. + if width > raw_unread { + if parser.eof { + return yaml_parser_set_reader_error(parser, + "incomplete UTF-8 octet sequence", + parser.offset, -1) + } + break inner + } + + // Decode the leading octet. + switch { + case octet&0x80 == 0x00: + value = rune(octet & 0x7F) + case octet&0xE0 == 0xC0: + value = rune(octet & 0x1F) + case octet&0xF0 == 0xE0: + value = rune(octet & 0x0F) + case octet&0xF8 == 0xF0: + value = rune(octet & 0x07) + default: + value = 0 + } + + // Check and decode the trailing octets. + for k := 1; k < width; k++ { + octet = parser.raw_buffer[parser.raw_buffer_pos+k] + + // Check if the octet is valid. + if (octet & 0xC0) != 0x80 { + return yaml_parser_set_reader_error(parser, + "invalid trailing UTF-8 octet", + parser.offset+k, int(octet)) + } + + // Decode the octet. + value = (value << 6) + rune(octet&0x3F) + } + + // Check the length of the sequence against the value. + switch { + case width == 1: + case width == 2 && value >= 0x80: + case width == 3 && value >= 0x800: + case width == 4 && value >= 0x10000: + default: + return yaml_parser_set_reader_error(parser, + "invalid length of a UTF-8 sequence", + parser.offset, -1) + } + + // Check the range of the value. + if value >= 0xD800 && value <= 0xDFFF || value > 0x10FFFF { + return yaml_parser_set_reader_error(parser, + "invalid Unicode character", + parser.offset, int(value)) + } + + case yaml_UTF16LE_ENCODING, yaml_UTF16BE_ENCODING: + var low, high int + if parser.encoding == yaml_UTF16LE_ENCODING { + low, high = 0, 1 + } else { + low, high = 1, 0 + } + + // The UTF-16 encoding is not as simple as one might + // naively think. Check RFC 2781 + // (http://www.ietf.org/rfc/rfc2781.txt). + // + // Normally, two subsequent bytes describe a Unicode + // character. However a special technique (called a + // surrogate pair) is used for specifying character + // values larger than 0xFFFF. + // + // A surrogate pair consists of two pseudo-characters: + // high surrogate area (0xD800-0xDBFF) + // low surrogate area (0xDC00-0xDFFF) + // + // The following formulas are used for decoding + // and encoding characters using surrogate pairs: + // + // U = U' + 0x10000 (0x01 00 00 <= U <= 0x10 FF FF) + // U' = yyyyyyyyyyxxxxxxxxxx (0 <= U' <= 0x0F FF FF) + // W1 = 110110yyyyyyyyyy + // W2 = 110111xxxxxxxxxx + // + // where U is the character value, W1 is the high surrogate + // area, W2 is the low surrogate area. + + // Check for incomplete UTF-16 character. + if raw_unread < 2 { + if parser.eof { + return yaml_parser_set_reader_error(parser, + "incomplete UTF-16 character", + parser.offset, -1) + } + break inner + } + + // Get the character. + value = rune(parser.raw_buffer[parser.raw_buffer_pos+low]) + + (rune(parser.raw_buffer[parser.raw_buffer_pos+high]) << 8) + + // Check for unexpected low surrogate area. + if value&0xFC00 == 0xDC00 { + return yaml_parser_set_reader_error(parser, + "unexpected low surrogate area", + parser.offset, int(value)) + } + + // Check for a high surrogate area. + if value&0xFC00 == 0xD800 { + width = 4 + + // Check for incomplete surrogate pair. + if raw_unread < 4 { + if parser.eof { + return yaml_parser_set_reader_error(parser, + "incomplete UTF-16 surrogate pair", + parser.offset, -1) + } + break inner + } + + // Get the next character. + value2 := rune(parser.raw_buffer[parser.raw_buffer_pos+low+2]) + + (rune(parser.raw_buffer[parser.raw_buffer_pos+high+2]) << 8) + + // Check for a low surrogate area. + if value2&0xFC00 != 0xDC00 { + return yaml_parser_set_reader_error(parser, + "expected low surrogate area", + parser.offset+2, int(value2)) + } + + // Generate the value of the surrogate pair. + value = 0x10000 + ((value & 0x3FF) << 10) + (value2 & 0x3FF) + } else { + width = 2 + } + + default: + panic("impossible") + } + + // Check if the character is in the allowed range: + // #x9 | #xA | #xD | [#x20-#x7E] (8 bit) + // | #x85 | [#xA0-#xD7FF] | [#xE000-#xFFFD] (16 bit) + // | [#x10000-#x10FFFF] (32 bit) + switch { + case value == 0x09: + case value == 0x0A: + case value == 0x0D: + case value >= 0x20 && value <= 0x7E: + case value == 0x85: + case value >= 0xA0 && value <= 0xD7FF: + case value >= 0xE000 && value <= 0xFFFD: + case value >= 0x10000 && value <= 0x10FFFF: + default: + return yaml_parser_set_reader_error(parser, + "control characters are not allowed", + parser.offset, int(value)) + } + + // Move the raw pointers. + parser.raw_buffer_pos += width + parser.offset += width + + // Finally put the character into the buffer. + if value <= 0x7F { + // 0000 0000-0000 007F . 0xxxxxxx + parser.buffer[buffer_len+0] = byte(value) + buffer_len += 1 + } else if value <= 0x7FF { + // 0000 0080-0000 07FF . 110xxxxx 10xxxxxx + parser.buffer[buffer_len+0] = byte(0xC0 + (value >> 6)) + parser.buffer[buffer_len+1] = byte(0x80 + (value & 0x3F)) + buffer_len += 2 + } else if value <= 0xFFFF { + // 0000 0800-0000 FFFF . 1110xxxx 10xxxxxx 10xxxxxx + parser.buffer[buffer_len+0] = byte(0xE0 + (value >> 12)) + parser.buffer[buffer_len+1] = byte(0x80 + ((value >> 6) & 0x3F)) + parser.buffer[buffer_len+2] = byte(0x80 + (value & 0x3F)) + buffer_len += 3 + } else { + // 0001 0000-0010 FFFF . 11110xxx 10xxxxxx 10xxxxxx 10xxxxxx + parser.buffer[buffer_len+0] = byte(0xF0 + (value >> 18)) + parser.buffer[buffer_len+1] = byte(0x80 + ((value >> 12) & 0x3F)) + parser.buffer[buffer_len+2] = byte(0x80 + ((value >> 6) & 0x3F)) + parser.buffer[buffer_len+3] = byte(0x80 + (value & 0x3F)) + buffer_len += 4 + } + + parser.unread++ + } + + // On EOF, put NUL into the buffer and return. + if parser.eof { + parser.buffer[buffer_len] = 0 + buffer_len++ + parser.unread++ + break + } + } + // [Go] Read the documentation of this function above. To return true, + // we need to have the given length in the buffer. Not doing that means + // every single check that calls this function to make sure the buffer + // has a given length is Go) panicking; or C) accessing invalid memory. + // This happens here due to the EOF above breaking early. + for buffer_len < length { + parser.buffer[buffer_len] = 0 + buffer_len++ + } + parser.buffer = parser.buffer[:buffer_len] + return true +} diff --git a/vendor/gopkg.in/yaml.v3/resolve.go b/vendor/gopkg.in/yaml.v3/resolve.go new file mode 100644 index 00000000..64ae8880 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/resolve.go @@ -0,0 +1,326 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package yaml + +import ( + "encoding/base64" + "math" + "regexp" + "strconv" + "strings" + "time" +) + +type resolveMapItem struct { + value interface{} + tag string +} + +var resolveTable = make([]byte, 256) +var resolveMap = make(map[string]resolveMapItem) + +func init() { + t := resolveTable + t[int('+')] = 'S' // Sign + t[int('-')] = 'S' + for _, c := range "0123456789" { + t[int(c)] = 'D' // Digit + } + for _, c := range "yYnNtTfFoO~" { + t[int(c)] = 'M' // In map + } + t[int('.')] = '.' // Float (potentially in map) + + var resolveMapList = []struct { + v interface{} + tag string + l []string + }{ + {true, boolTag, []string{"true", "True", "TRUE"}}, + {false, boolTag, []string{"false", "False", "FALSE"}}, + {nil, nullTag, []string{"", "~", "null", "Null", "NULL"}}, + {math.NaN(), floatTag, []string{".nan", ".NaN", ".NAN"}}, + {math.Inf(+1), floatTag, []string{".inf", ".Inf", ".INF"}}, + {math.Inf(+1), floatTag, []string{"+.inf", "+.Inf", "+.INF"}}, + {math.Inf(-1), floatTag, []string{"-.inf", "-.Inf", "-.INF"}}, + {"<<", mergeTag, []string{"<<"}}, + } + + m := resolveMap + for _, item := range resolveMapList { + for _, s := range item.l { + m[s] = resolveMapItem{item.v, item.tag} + } + } +} + +const ( + nullTag = "!!null" + boolTag = "!!bool" + strTag = "!!str" + intTag = "!!int" + floatTag = "!!float" + timestampTag = "!!timestamp" + seqTag = "!!seq" + mapTag = "!!map" + binaryTag = "!!binary" + mergeTag = "!!merge" +) + +var longTags = make(map[string]string) +var shortTags = make(map[string]string) + +func init() { + for _, stag := range []string{nullTag, boolTag, strTag, intTag, floatTag, timestampTag, seqTag, mapTag, binaryTag, mergeTag} { + ltag := longTag(stag) + longTags[stag] = ltag + shortTags[ltag] = stag + } +} + +const longTagPrefix = "tag:yaml.org,2002:" + +func shortTag(tag string) string { + if strings.HasPrefix(tag, longTagPrefix) { + if stag, ok := shortTags[tag]; ok { + return stag + } + return "!!" + tag[len(longTagPrefix):] + } + return tag +} + +func longTag(tag string) string { + if strings.HasPrefix(tag, "!!") { + if ltag, ok := longTags[tag]; ok { + return ltag + } + return longTagPrefix + tag[2:] + } + return tag +} + +func resolvableTag(tag string) bool { + switch tag { + case "", strTag, boolTag, intTag, floatTag, nullTag, timestampTag: + return true + } + return false +} + +var yamlStyleFloat = regexp.MustCompile(`^[-+]?(\.[0-9]+|[0-9]+(\.[0-9]*)?)([eE][-+]?[0-9]+)?$`) + +func resolve(tag string, in string) (rtag string, out interface{}) { + tag = shortTag(tag) + if !resolvableTag(tag) { + return tag, in + } + + defer func() { + switch tag { + case "", rtag, strTag, binaryTag: + return + case floatTag: + if rtag == intTag { + switch v := out.(type) { + case int64: + rtag = floatTag + out = float64(v) + return + case int: + rtag = floatTag + out = float64(v) + return + } + } + } + failf("cannot decode %s `%s` as a %s", shortTag(rtag), in, shortTag(tag)) + }() + + // Any data is accepted as a !!str or !!binary. + // Otherwise, the prefix is enough of a hint about what it might be. + hint := byte('N') + if in != "" { + hint = resolveTable[in[0]] + } + if hint != 0 && tag != strTag && tag != binaryTag { + // Handle things we can lookup in a map. + if item, ok := resolveMap[in]; ok { + return item.tag, item.value + } + + // Base 60 floats are a bad idea, were dropped in YAML 1.2, and + // are purposefully unsupported here. They're still quoted on + // the way out for compatibility with other parser, though. + + switch hint { + case 'M': + // We've already checked the map above. + + case '.': + // Not in the map, so maybe a normal float. + floatv, err := strconv.ParseFloat(in, 64) + if err == nil { + return floatTag, floatv + } + + case 'D', 'S': + // Int, float, or timestamp. + // Only try values as a timestamp if the value is unquoted or there's an explicit + // !!timestamp tag. + if tag == "" || tag == timestampTag { + t, ok := parseTimestamp(in) + if ok { + return timestampTag, t + } + } + + plain := strings.Replace(in, "_", "", -1) + intv, err := strconv.ParseInt(plain, 0, 64) + if err == nil { + if intv == int64(int(intv)) { + return intTag, int(intv) + } else { + return intTag, intv + } + } + uintv, err := strconv.ParseUint(plain, 0, 64) + if err == nil { + return intTag, uintv + } + if yamlStyleFloat.MatchString(plain) { + floatv, err := strconv.ParseFloat(plain, 64) + if err == nil { + return floatTag, floatv + } + } + if strings.HasPrefix(plain, "0b") { + intv, err := strconv.ParseInt(plain[2:], 2, 64) + if err == nil { + if intv == int64(int(intv)) { + return intTag, int(intv) + } else { + return intTag, intv + } + } + uintv, err := strconv.ParseUint(plain[2:], 2, 64) + if err == nil { + return intTag, uintv + } + } else if strings.HasPrefix(plain, "-0b") { + intv, err := strconv.ParseInt("-"+plain[3:], 2, 64) + if err == nil { + if true || intv == int64(int(intv)) { + return intTag, int(intv) + } else { + return intTag, intv + } + } + } + // Octals as introduced in version 1.2 of the spec. + // Octals from the 1.1 spec, spelled as 0777, are still + // decoded by default in v3 as well for compatibility. + // May be dropped in v4 depending on how usage evolves. + if strings.HasPrefix(plain, "0o") { + intv, err := strconv.ParseInt(plain[2:], 8, 64) + if err == nil { + if intv == int64(int(intv)) { + return intTag, int(intv) + } else { + return intTag, intv + } + } + uintv, err := strconv.ParseUint(plain[2:], 8, 64) + if err == nil { + return intTag, uintv + } + } else if strings.HasPrefix(plain, "-0o") { + intv, err := strconv.ParseInt("-"+plain[3:], 8, 64) + if err == nil { + if true || intv == int64(int(intv)) { + return intTag, int(intv) + } else { + return intTag, intv + } + } + } + default: + panic("internal error: missing handler for resolver table: " + string(rune(hint)) + " (with " + in + ")") + } + } + return strTag, in +} + +// encodeBase64 encodes s as base64 that is broken up into multiple lines +// as appropriate for the resulting length. +func encodeBase64(s string) string { + const lineLen = 70 + encLen := base64.StdEncoding.EncodedLen(len(s)) + lines := encLen/lineLen + 1 + buf := make([]byte, encLen*2+lines) + in := buf[0:encLen] + out := buf[encLen:] + base64.StdEncoding.Encode(in, []byte(s)) + k := 0 + for i := 0; i < len(in); i += lineLen { + j := i + lineLen + if j > len(in) { + j = len(in) + } + k += copy(out[k:], in[i:j]) + if lines > 1 { + out[k] = '\n' + k++ + } + } + return string(out[:k]) +} + +// This is a subset of the formats allowed by the regular expression +// defined at http://yaml.org/type/timestamp.html. +var allowedTimestampFormats = []string{ + "2006-1-2T15:4:5.999999999Z07:00", // RCF3339Nano with short date fields. + "2006-1-2t15:4:5.999999999Z07:00", // RFC3339Nano with short date fields and lower-case "t". + "2006-1-2 15:4:5.999999999", // space separated with no time zone + "2006-1-2", // date only + // Notable exception: time.Parse cannot handle: "2001-12-14 21:59:43.10 -5" + // from the set of examples. +} + +// parseTimestamp parses s as a timestamp string and +// returns the timestamp and reports whether it succeeded. +// Timestamp formats are defined at http://yaml.org/type/timestamp.html +func parseTimestamp(s string) (time.Time, bool) { + // TODO write code to check all the formats supported by + // http://yaml.org/type/timestamp.html instead of using time.Parse. + + // Quick check: all date formats start with YYYY-. + i := 0 + for ; i < len(s); i++ { + if c := s[i]; c < '0' || c > '9' { + break + } + } + if i != 4 || i == len(s) || s[i] != '-' { + return time.Time{}, false + } + for _, format := range allowedTimestampFormats { + if t, err := time.Parse(format, s); err == nil { + return t, true + } + } + return time.Time{}, false +} diff --git a/vendor/gopkg.in/yaml.v3/scannerc.go b/vendor/gopkg.in/yaml.v3/scannerc.go new file mode 100644 index 00000000..d9a539c3 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/scannerc.go @@ -0,0 +1,3028 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +import ( + "bytes" + "fmt" +) + +// Introduction +// ************ +// +// The following notes assume that you are familiar with the YAML specification +// (http://yaml.org/spec/1.2/spec.html). We mostly follow it, although in +// some cases we are less restrictive that it requires. +// +// The process of transforming a YAML stream into a sequence of events is +// divided on two steps: Scanning and Parsing. +// +// The Scanner transforms the input stream into a sequence of tokens, while the +// parser transform the sequence of tokens produced by the Scanner into a +// sequence of parsing events. +// +// The Scanner is rather clever and complicated. The Parser, on the contrary, +// is a straightforward implementation of a recursive-descendant parser (or, +// LL(1) parser, as it is usually called). +// +// Actually there are two issues of Scanning that might be called "clever", the +// rest is quite straightforward. The issues are "block collection start" and +// "simple keys". Both issues are explained below in details. +// +// Here the Scanning step is explained and implemented. We start with the list +// of all the tokens produced by the Scanner together with short descriptions. +// +// Now, tokens: +// +// STREAM-START(encoding) # The stream start. +// STREAM-END # The stream end. +// VERSION-DIRECTIVE(major,minor) # The '%YAML' directive. +// TAG-DIRECTIVE(handle,prefix) # The '%TAG' directive. +// DOCUMENT-START # '---' +// DOCUMENT-END # '...' +// BLOCK-SEQUENCE-START # Indentation increase denoting a block +// BLOCK-MAPPING-START # sequence or a block mapping. +// BLOCK-END # Indentation decrease. +// FLOW-SEQUENCE-START # '[' +// FLOW-SEQUENCE-END # ']' +// BLOCK-SEQUENCE-START # '{' +// BLOCK-SEQUENCE-END # '}' +// BLOCK-ENTRY # '-' +// FLOW-ENTRY # ',' +// KEY # '?' or nothing (simple keys). +// VALUE # ':' +// ALIAS(anchor) # '*anchor' +// ANCHOR(anchor) # '&anchor' +// TAG(handle,suffix) # '!handle!suffix' +// SCALAR(value,style) # A scalar. +// +// The following two tokens are "virtual" tokens denoting the beginning and the +// end of the stream: +// +// STREAM-START(encoding) +// STREAM-END +// +// We pass the information about the input stream encoding with the +// STREAM-START token. +// +// The next two tokens are responsible for tags: +// +// VERSION-DIRECTIVE(major,minor) +// TAG-DIRECTIVE(handle,prefix) +// +// Example: +// +// %YAML 1.1 +// %TAG ! !foo +// %TAG !yaml! tag:yaml.org,2002: +// --- +// +// The correspoding sequence of tokens: +// +// STREAM-START(utf-8) +// VERSION-DIRECTIVE(1,1) +// TAG-DIRECTIVE("!","!foo") +// TAG-DIRECTIVE("!yaml","tag:yaml.org,2002:") +// DOCUMENT-START +// STREAM-END +// +// Note that the VERSION-DIRECTIVE and TAG-DIRECTIVE tokens occupy a whole +// line. +// +// The document start and end indicators are represented by: +// +// DOCUMENT-START +// DOCUMENT-END +// +// Note that if a YAML stream contains an implicit document (without '---' +// and '...' indicators), no DOCUMENT-START and DOCUMENT-END tokens will be +// produced. +// +// In the following examples, we present whole documents together with the +// produced tokens. +// +// 1. An implicit document: +// +// 'a scalar' +// +// Tokens: +// +// STREAM-START(utf-8) +// SCALAR("a scalar",single-quoted) +// STREAM-END +// +// 2. An explicit document: +// +// --- +// 'a scalar' +// ... +// +// Tokens: +// +// STREAM-START(utf-8) +// DOCUMENT-START +// SCALAR("a scalar",single-quoted) +// DOCUMENT-END +// STREAM-END +// +// 3. Several documents in a stream: +// +// 'a scalar' +// --- +// 'another scalar' +// --- +// 'yet another scalar' +// +// Tokens: +// +// STREAM-START(utf-8) +// SCALAR("a scalar",single-quoted) +// DOCUMENT-START +// SCALAR("another scalar",single-quoted) +// DOCUMENT-START +// SCALAR("yet another scalar",single-quoted) +// STREAM-END +// +// We have already introduced the SCALAR token above. The following tokens are +// used to describe aliases, anchors, tag, and scalars: +// +// ALIAS(anchor) +// ANCHOR(anchor) +// TAG(handle,suffix) +// SCALAR(value,style) +// +// The following series of examples illustrate the usage of these tokens: +// +// 1. A recursive sequence: +// +// &A [ *A ] +// +// Tokens: +// +// STREAM-START(utf-8) +// ANCHOR("A") +// FLOW-SEQUENCE-START +// ALIAS("A") +// FLOW-SEQUENCE-END +// STREAM-END +// +// 2. A tagged scalar: +// +// !!float "3.14" # A good approximation. +// +// Tokens: +// +// STREAM-START(utf-8) +// TAG("!!","float") +// SCALAR("3.14",double-quoted) +// STREAM-END +// +// 3. Various scalar styles: +// +// --- # Implicit empty plain scalars do not produce tokens. +// --- a plain scalar +// --- 'a single-quoted scalar' +// --- "a double-quoted scalar" +// --- |- +// a literal scalar +// --- >- +// a folded +// scalar +// +// Tokens: +// +// STREAM-START(utf-8) +// DOCUMENT-START +// DOCUMENT-START +// SCALAR("a plain scalar",plain) +// DOCUMENT-START +// SCALAR("a single-quoted scalar",single-quoted) +// DOCUMENT-START +// SCALAR("a double-quoted scalar",double-quoted) +// DOCUMENT-START +// SCALAR("a literal scalar",literal) +// DOCUMENT-START +// SCALAR("a folded scalar",folded) +// STREAM-END +// +// Now it's time to review collection-related tokens. We will start with +// flow collections: +// +// FLOW-SEQUENCE-START +// FLOW-SEQUENCE-END +// FLOW-MAPPING-START +// FLOW-MAPPING-END +// FLOW-ENTRY +// KEY +// VALUE +// +// The tokens FLOW-SEQUENCE-START, FLOW-SEQUENCE-END, FLOW-MAPPING-START, and +// FLOW-MAPPING-END represent the indicators '[', ']', '{', and '}' +// correspondingly. FLOW-ENTRY represent the ',' indicator. Finally the +// indicators '?' and ':', which are used for denoting mapping keys and values, +// are represented by the KEY and VALUE tokens. +// +// The following examples show flow collections: +// +// 1. A flow sequence: +// +// [item 1, item 2, item 3] +// +// Tokens: +// +// STREAM-START(utf-8) +// FLOW-SEQUENCE-START +// SCALAR("item 1",plain) +// FLOW-ENTRY +// SCALAR("item 2",plain) +// FLOW-ENTRY +// SCALAR("item 3",plain) +// FLOW-SEQUENCE-END +// STREAM-END +// +// 2. A flow mapping: +// +// { +// a simple key: a value, # Note that the KEY token is produced. +// ? a complex key: another value, +// } +// +// Tokens: +// +// STREAM-START(utf-8) +// FLOW-MAPPING-START +// KEY +// SCALAR("a simple key",plain) +// VALUE +// SCALAR("a value",plain) +// FLOW-ENTRY +// KEY +// SCALAR("a complex key",plain) +// VALUE +// SCALAR("another value",plain) +// FLOW-ENTRY +// FLOW-MAPPING-END +// STREAM-END +// +// A simple key is a key which is not denoted by the '?' indicator. Note that +// the Scanner still produce the KEY token whenever it encounters a simple key. +// +// For scanning block collections, the following tokens are used (note that we +// repeat KEY and VALUE here): +// +// BLOCK-SEQUENCE-START +// BLOCK-MAPPING-START +// BLOCK-END +// BLOCK-ENTRY +// KEY +// VALUE +// +// The tokens BLOCK-SEQUENCE-START and BLOCK-MAPPING-START denote indentation +// increase that precedes a block collection (cf. the INDENT token in Python). +// The token BLOCK-END denote indentation decrease that ends a block collection +// (cf. the DEDENT token in Python). However YAML has some syntax pecularities +// that makes detections of these tokens more complex. +// +// The tokens BLOCK-ENTRY, KEY, and VALUE are used to represent the indicators +// '-', '?', and ':' correspondingly. +// +// The following examples show how the tokens BLOCK-SEQUENCE-START, +// BLOCK-MAPPING-START, and BLOCK-END are emitted by the Scanner: +// +// 1. Block sequences: +// +// - item 1 +// - item 2 +// - +// - item 3.1 +// - item 3.2 +// - +// key 1: value 1 +// key 2: value 2 +// +// Tokens: +// +// STREAM-START(utf-8) +// BLOCK-SEQUENCE-START +// BLOCK-ENTRY +// SCALAR("item 1",plain) +// BLOCK-ENTRY +// SCALAR("item 2",plain) +// BLOCK-ENTRY +// BLOCK-SEQUENCE-START +// BLOCK-ENTRY +// SCALAR("item 3.1",plain) +// BLOCK-ENTRY +// SCALAR("item 3.2",plain) +// BLOCK-END +// BLOCK-ENTRY +// BLOCK-MAPPING-START +// KEY +// SCALAR("key 1",plain) +// VALUE +// SCALAR("value 1",plain) +// KEY +// SCALAR("key 2",plain) +// VALUE +// SCALAR("value 2",plain) +// BLOCK-END +// BLOCK-END +// STREAM-END +// +// 2. Block mappings: +// +// a simple key: a value # The KEY token is produced here. +// ? a complex key +// : another value +// a mapping: +// key 1: value 1 +// key 2: value 2 +// a sequence: +// - item 1 +// - item 2 +// +// Tokens: +// +// STREAM-START(utf-8) +// BLOCK-MAPPING-START +// KEY +// SCALAR("a simple key",plain) +// VALUE +// SCALAR("a value",plain) +// KEY +// SCALAR("a complex key",plain) +// VALUE +// SCALAR("another value",plain) +// KEY +// SCALAR("a mapping",plain) +// BLOCK-MAPPING-START +// KEY +// SCALAR("key 1",plain) +// VALUE +// SCALAR("value 1",plain) +// KEY +// SCALAR("key 2",plain) +// VALUE +// SCALAR("value 2",plain) +// BLOCK-END +// KEY +// SCALAR("a sequence",plain) +// VALUE +// BLOCK-SEQUENCE-START +// BLOCK-ENTRY +// SCALAR("item 1",plain) +// BLOCK-ENTRY +// SCALAR("item 2",plain) +// BLOCK-END +// BLOCK-END +// STREAM-END +// +// YAML does not always require to start a new block collection from a new +// line. If the current line contains only '-', '?', and ':' indicators, a new +// block collection may start at the current line. The following examples +// illustrate this case: +// +// 1. Collections in a sequence: +// +// - - item 1 +// - item 2 +// - key 1: value 1 +// key 2: value 2 +// - ? complex key +// : complex value +// +// Tokens: +// +// STREAM-START(utf-8) +// BLOCK-SEQUENCE-START +// BLOCK-ENTRY +// BLOCK-SEQUENCE-START +// BLOCK-ENTRY +// SCALAR("item 1",plain) +// BLOCK-ENTRY +// SCALAR("item 2",plain) +// BLOCK-END +// BLOCK-ENTRY +// BLOCK-MAPPING-START +// KEY +// SCALAR("key 1",plain) +// VALUE +// SCALAR("value 1",plain) +// KEY +// SCALAR("key 2",plain) +// VALUE +// SCALAR("value 2",plain) +// BLOCK-END +// BLOCK-ENTRY +// BLOCK-MAPPING-START +// KEY +// SCALAR("complex key") +// VALUE +// SCALAR("complex value") +// BLOCK-END +// BLOCK-END +// STREAM-END +// +// 2. Collections in a mapping: +// +// ? a sequence +// : - item 1 +// - item 2 +// ? a mapping +// : key 1: value 1 +// key 2: value 2 +// +// Tokens: +// +// STREAM-START(utf-8) +// BLOCK-MAPPING-START +// KEY +// SCALAR("a sequence",plain) +// VALUE +// BLOCK-SEQUENCE-START +// BLOCK-ENTRY +// SCALAR("item 1",plain) +// BLOCK-ENTRY +// SCALAR("item 2",plain) +// BLOCK-END +// KEY +// SCALAR("a mapping",plain) +// VALUE +// BLOCK-MAPPING-START +// KEY +// SCALAR("key 1",plain) +// VALUE +// SCALAR("value 1",plain) +// KEY +// SCALAR("key 2",plain) +// VALUE +// SCALAR("value 2",plain) +// BLOCK-END +// BLOCK-END +// STREAM-END +// +// YAML also permits non-indented sequences if they are included into a block +// mapping. In this case, the token BLOCK-SEQUENCE-START is not produced: +// +// key: +// - item 1 # BLOCK-SEQUENCE-START is NOT produced here. +// - item 2 +// +// Tokens: +// +// STREAM-START(utf-8) +// BLOCK-MAPPING-START +// KEY +// SCALAR("key",plain) +// VALUE +// BLOCK-ENTRY +// SCALAR("item 1",plain) +// BLOCK-ENTRY +// SCALAR("item 2",plain) +// BLOCK-END +// + +// Ensure that the buffer contains the required number of characters. +// Return true on success, false on failure (reader error or memory error). +func cache(parser *yaml_parser_t, length int) bool { + // [Go] This was inlined: !cache(A, B) -> unread < B && !update(A, B) + return parser.unread >= length || yaml_parser_update_buffer(parser, length) +} + +// Advance the buffer pointer. +func skip(parser *yaml_parser_t) { + if !is_blank(parser.buffer, parser.buffer_pos) { + parser.newlines = 0 + } + parser.mark.index++ + parser.mark.column++ + parser.unread-- + parser.buffer_pos += width(parser.buffer[parser.buffer_pos]) +} + +func skip_line(parser *yaml_parser_t) { + if is_crlf(parser.buffer, parser.buffer_pos) { + parser.mark.index += 2 + parser.mark.column = 0 + parser.mark.line++ + parser.unread -= 2 + parser.buffer_pos += 2 + parser.newlines++ + } else if is_break(parser.buffer, parser.buffer_pos) { + parser.mark.index++ + parser.mark.column = 0 + parser.mark.line++ + parser.unread-- + parser.buffer_pos += width(parser.buffer[parser.buffer_pos]) + parser.newlines++ + } +} + +// Copy a character to a string buffer and advance pointers. +func read(parser *yaml_parser_t, s []byte) []byte { + if !is_blank(parser.buffer, parser.buffer_pos) { + parser.newlines = 0 + } + w := width(parser.buffer[parser.buffer_pos]) + if w == 0 { + panic("invalid character sequence") + } + if len(s) == 0 { + s = make([]byte, 0, 32) + } + if w == 1 && len(s)+w <= cap(s) { + s = s[:len(s)+1] + s[len(s)-1] = parser.buffer[parser.buffer_pos] + parser.buffer_pos++ + } else { + s = append(s, parser.buffer[parser.buffer_pos:parser.buffer_pos+w]...) + parser.buffer_pos += w + } + parser.mark.index++ + parser.mark.column++ + parser.unread-- + return s +} + +// Copy a line break character to a string buffer and advance pointers. +func read_line(parser *yaml_parser_t, s []byte) []byte { + buf := parser.buffer + pos := parser.buffer_pos + switch { + case buf[pos] == '\r' && buf[pos+1] == '\n': + // CR LF . LF + s = append(s, '\n') + parser.buffer_pos += 2 + parser.mark.index++ + parser.unread-- + case buf[pos] == '\r' || buf[pos] == '\n': + // CR|LF . LF + s = append(s, '\n') + parser.buffer_pos += 1 + case buf[pos] == '\xC2' && buf[pos+1] == '\x85': + // NEL . LF + s = append(s, '\n') + parser.buffer_pos += 2 + case buf[pos] == '\xE2' && buf[pos+1] == '\x80' && (buf[pos+2] == '\xA8' || buf[pos+2] == '\xA9'): + // LS|PS . LS|PS + s = append(s, buf[parser.buffer_pos:pos+3]...) + parser.buffer_pos += 3 + default: + return s + } + parser.mark.index++ + parser.mark.column = 0 + parser.mark.line++ + parser.unread-- + parser.newlines++ + return s +} + +// Get the next token. +func yaml_parser_scan(parser *yaml_parser_t, token *yaml_token_t) bool { + // Erase the token object. + *token = yaml_token_t{} // [Go] Is this necessary? + + // No tokens after STREAM-END or error. + if parser.stream_end_produced || parser.error != yaml_NO_ERROR { + return true + } + + // Ensure that the tokens queue contains enough tokens. + if !parser.token_available { + if !yaml_parser_fetch_more_tokens(parser) { + return false + } + } + + // Fetch the next token from the queue. + *token = parser.tokens[parser.tokens_head] + parser.tokens_head++ + parser.tokens_parsed++ + parser.token_available = false + + if token.typ == yaml_STREAM_END_TOKEN { + parser.stream_end_produced = true + } + return true +} + +// Set the scanner error and return false. +func yaml_parser_set_scanner_error(parser *yaml_parser_t, context string, context_mark yaml_mark_t, problem string) bool { + parser.error = yaml_SCANNER_ERROR + parser.context = context + parser.context_mark = context_mark + parser.problem = problem + parser.problem_mark = parser.mark + return false +} + +func yaml_parser_set_scanner_tag_error(parser *yaml_parser_t, directive bool, context_mark yaml_mark_t, problem string) bool { + context := "while parsing a tag" + if directive { + context = "while parsing a %TAG directive" + } + return yaml_parser_set_scanner_error(parser, context, context_mark, problem) +} + +func trace(args ...interface{}) func() { + pargs := append([]interface{}{"+++"}, args...) + fmt.Println(pargs...) + pargs = append([]interface{}{"---"}, args...) + return func() { fmt.Println(pargs...) } +} + +// Ensure that the tokens queue contains at least one token which can be +// returned to the Parser. +func yaml_parser_fetch_more_tokens(parser *yaml_parser_t) bool { + // While we need more tokens to fetch, do it. + for { + // [Go] The comment parsing logic requires a lookahead of two tokens + // so that foot comments may be parsed in time of associating them + // with the tokens that are parsed before them, and also for line + // comments to be transformed into head comments in some edge cases. + if parser.tokens_head < len(parser.tokens)-2 { + // If a potential simple key is at the head position, we need to fetch + // the next token to disambiguate it. + head_tok_idx, ok := parser.simple_keys_by_tok[parser.tokens_parsed] + if !ok { + break + } else if valid, ok := yaml_simple_key_is_valid(parser, &parser.simple_keys[head_tok_idx]); !ok { + return false + } else if !valid { + break + } + } + // Fetch the next token. + if !yaml_parser_fetch_next_token(parser) { + return false + } + } + + parser.token_available = true + return true +} + +// The dispatcher for token fetchers. +func yaml_parser_fetch_next_token(parser *yaml_parser_t) (ok bool) { + // Ensure that the buffer is initialized. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + // Check if we just started scanning. Fetch STREAM-START then. + if !parser.stream_start_produced { + return yaml_parser_fetch_stream_start(parser) + } + + scan_mark := parser.mark + + // Eat whitespaces and comments until we reach the next token. + if !yaml_parser_scan_to_next_token(parser) { + return false + } + + // [Go] While unrolling indents, transform the head comments of prior + // indentation levels observed after scan_start into foot comments at + // the respective indexes. + + // Check the indentation level against the current column. + if !yaml_parser_unroll_indent(parser, parser.mark.column, scan_mark) { + return false + } + + // Ensure that the buffer contains at least 4 characters. 4 is the length + // of the longest indicators ('--- ' and '... '). + if parser.unread < 4 && !yaml_parser_update_buffer(parser, 4) { + return false + } + + // Is it the end of the stream? + if is_z(parser.buffer, parser.buffer_pos) { + return yaml_parser_fetch_stream_end(parser) + } + + // Is it a directive? + if parser.mark.column == 0 && parser.buffer[parser.buffer_pos] == '%' { + return yaml_parser_fetch_directive(parser) + } + + buf := parser.buffer + pos := parser.buffer_pos + + // Is it the document start indicator? + if parser.mark.column == 0 && buf[pos] == '-' && buf[pos+1] == '-' && buf[pos+2] == '-' && is_blankz(buf, pos+3) { + return yaml_parser_fetch_document_indicator(parser, yaml_DOCUMENT_START_TOKEN) + } + + // Is it the document end indicator? + if parser.mark.column == 0 && buf[pos] == '.' && buf[pos+1] == '.' && buf[pos+2] == '.' && is_blankz(buf, pos+3) { + return yaml_parser_fetch_document_indicator(parser, yaml_DOCUMENT_END_TOKEN) + } + + comment_mark := parser.mark + if len(parser.tokens) > 0 && (parser.flow_level == 0 && buf[pos] == ':' || parser.flow_level > 0 && buf[pos] == ',') { + // Associate any following comments with the prior token. + comment_mark = parser.tokens[len(parser.tokens)-1].start_mark + } + defer func() { + if !ok { + return + } + if len(parser.tokens) > 0 && parser.tokens[len(parser.tokens)-1].typ == yaml_BLOCK_ENTRY_TOKEN { + // Sequence indicators alone have no line comments. It becomes + // a head comment for whatever follows. + return + } + if !yaml_parser_scan_line_comment(parser, comment_mark) { + ok = false + return + } + }() + + // Is it the flow sequence start indicator? + if buf[pos] == '[' { + return yaml_parser_fetch_flow_collection_start(parser, yaml_FLOW_SEQUENCE_START_TOKEN) + } + + // Is it the flow mapping start indicator? + if parser.buffer[parser.buffer_pos] == '{' { + return yaml_parser_fetch_flow_collection_start(parser, yaml_FLOW_MAPPING_START_TOKEN) + } + + // Is it the flow sequence end indicator? + if parser.buffer[parser.buffer_pos] == ']' { + return yaml_parser_fetch_flow_collection_end(parser, + yaml_FLOW_SEQUENCE_END_TOKEN) + } + + // Is it the flow mapping end indicator? + if parser.buffer[parser.buffer_pos] == '}' { + return yaml_parser_fetch_flow_collection_end(parser, + yaml_FLOW_MAPPING_END_TOKEN) + } + + // Is it the flow entry indicator? + if parser.buffer[parser.buffer_pos] == ',' { + return yaml_parser_fetch_flow_entry(parser) + } + + // Is it the block entry indicator? + if parser.buffer[parser.buffer_pos] == '-' && is_blankz(parser.buffer, parser.buffer_pos+1) { + return yaml_parser_fetch_block_entry(parser) + } + + // Is it the key indicator? + if parser.buffer[parser.buffer_pos] == '?' && (parser.flow_level > 0 || is_blankz(parser.buffer, parser.buffer_pos+1)) { + return yaml_parser_fetch_key(parser) + } + + // Is it the value indicator? + if parser.buffer[parser.buffer_pos] == ':' && (parser.flow_level > 0 || is_blankz(parser.buffer, parser.buffer_pos+1)) { + return yaml_parser_fetch_value(parser) + } + + // Is it an alias? + if parser.buffer[parser.buffer_pos] == '*' { + return yaml_parser_fetch_anchor(parser, yaml_ALIAS_TOKEN) + } + + // Is it an anchor? + if parser.buffer[parser.buffer_pos] == '&' { + return yaml_parser_fetch_anchor(parser, yaml_ANCHOR_TOKEN) + } + + // Is it a tag? + if parser.buffer[parser.buffer_pos] == '!' { + return yaml_parser_fetch_tag(parser) + } + + // Is it a literal scalar? + if parser.buffer[parser.buffer_pos] == '|' && parser.flow_level == 0 { + return yaml_parser_fetch_block_scalar(parser, true) + } + + // Is it a folded scalar? + if parser.buffer[parser.buffer_pos] == '>' && parser.flow_level == 0 { + return yaml_parser_fetch_block_scalar(parser, false) + } + + // Is it a single-quoted scalar? + if parser.buffer[parser.buffer_pos] == '\'' { + return yaml_parser_fetch_flow_scalar(parser, true) + } + + // Is it a double-quoted scalar? + if parser.buffer[parser.buffer_pos] == '"' { + return yaml_parser_fetch_flow_scalar(parser, false) + } + + // Is it a plain scalar? + // + // A plain scalar may start with any non-blank characters except + // + // '-', '?', ':', ',', '[', ']', '{', '}', + // '#', '&', '*', '!', '|', '>', '\'', '\"', + // '%', '@', '`'. + // + // In the block context (and, for the '-' indicator, in the flow context + // too), it may also start with the characters + // + // '-', '?', ':' + // + // if it is followed by a non-space character. + // + // The last rule is more restrictive than the specification requires. + // [Go] TODO Make this logic more reasonable. + //switch parser.buffer[parser.buffer_pos] { + //case '-', '?', ':', ',', '?', '-', ',', ':', ']', '[', '}', '{', '&', '#', '!', '*', '>', '|', '"', '\'', '@', '%', '-', '`': + //} + if !(is_blankz(parser.buffer, parser.buffer_pos) || parser.buffer[parser.buffer_pos] == '-' || + parser.buffer[parser.buffer_pos] == '?' || parser.buffer[parser.buffer_pos] == ':' || + parser.buffer[parser.buffer_pos] == ',' || parser.buffer[parser.buffer_pos] == '[' || + parser.buffer[parser.buffer_pos] == ']' || parser.buffer[parser.buffer_pos] == '{' || + parser.buffer[parser.buffer_pos] == '}' || parser.buffer[parser.buffer_pos] == '#' || + parser.buffer[parser.buffer_pos] == '&' || parser.buffer[parser.buffer_pos] == '*' || + parser.buffer[parser.buffer_pos] == '!' || parser.buffer[parser.buffer_pos] == '|' || + parser.buffer[parser.buffer_pos] == '>' || parser.buffer[parser.buffer_pos] == '\'' || + parser.buffer[parser.buffer_pos] == '"' || parser.buffer[parser.buffer_pos] == '%' || + parser.buffer[parser.buffer_pos] == '@' || parser.buffer[parser.buffer_pos] == '`') || + (parser.buffer[parser.buffer_pos] == '-' && !is_blank(parser.buffer, parser.buffer_pos+1)) || + (parser.flow_level == 0 && + (parser.buffer[parser.buffer_pos] == '?' || parser.buffer[parser.buffer_pos] == ':') && + !is_blankz(parser.buffer, parser.buffer_pos+1)) { + return yaml_parser_fetch_plain_scalar(parser) + } + + // If we don't determine the token type so far, it is an error. + return yaml_parser_set_scanner_error(parser, + "while scanning for the next token", parser.mark, + "found character that cannot start any token") +} + +func yaml_simple_key_is_valid(parser *yaml_parser_t, simple_key *yaml_simple_key_t) (valid, ok bool) { + if !simple_key.possible { + return false, true + } + + // The 1.2 specification says: + // + // "If the ? indicator is omitted, parsing needs to see past the + // implicit key to recognize it as such. To limit the amount of + // lookahead required, the “:” indicator must appear at most 1024 + // Unicode characters beyond the start of the key. In addition, the key + // is restricted to a single line." + // + if simple_key.mark.line < parser.mark.line || simple_key.mark.index+1024 < parser.mark.index { + // Check if the potential simple key to be removed is required. + if simple_key.required { + return false, yaml_parser_set_scanner_error(parser, + "while scanning a simple key", simple_key.mark, + "could not find expected ':'") + } + simple_key.possible = false + return false, true + } + return true, true +} + +// Check if a simple key may start at the current position and add it if +// needed. +func yaml_parser_save_simple_key(parser *yaml_parser_t) bool { + // A simple key is required at the current position if the scanner is in + // the block context and the current column coincides with the indentation + // level. + + required := parser.flow_level == 0 && parser.indent == parser.mark.column + + // + // If the current position may start a simple key, save it. + // + if parser.simple_key_allowed { + simple_key := yaml_simple_key_t{ + possible: true, + required: required, + token_number: parser.tokens_parsed + (len(parser.tokens) - parser.tokens_head), + mark: parser.mark, + } + + if !yaml_parser_remove_simple_key(parser) { + return false + } + parser.simple_keys[len(parser.simple_keys)-1] = simple_key + parser.simple_keys_by_tok[simple_key.token_number] = len(parser.simple_keys) - 1 + } + return true +} + +// Remove a potential simple key at the current flow level. +func yaml_parser_remove_simple_key(parser *yaml_parser_t) bool { + i := len(parser.simple_keys) - 1 + if parser.simple_keys[i].possible { + // If the key is required, it is an error. + if parser.simple_keys[i].required { + return yaml_parser_set_scanner_error(parser, + "while scanning a simple key", parser.simple_keys[i].mark, + "could not find expected ':'") + } + // Remove the key from the stack. + parser.simple_keys[i].possible = false + delete(parser.simple_keys_by_tok, parser.simple_keys[i].token_number) + } + return true +} + +// max_flow_level limits the flow_level +const max_flow_level = 10000 + +// Increase the flow level and resize the simple key list if needed. +func yaml_parser_increase_flow_level(parser *yaml_parser_t) bool { + // Reset the simple key on the next level. + parser.simple_keys = append(parser.simple_keys, yaml_simple_key_t{ + possible: false, + required: false, + token_number: parser.tokens_parsed + (len(parser.tokens) - parser.tokens_head), + mark: parser.mark, + }) + + // Increase the flow level. + parser.flow_level++ + if parser.flow_level > max_flow_level { + return yaml_parser_set_scanner_error(parser, + "while increasing flow level", parser.simple_keys[len(parser.simple_keys)-1].mark, + fmt.Sprintf("exceeded max depth of %d", max_flow_level)) + } + return true +} + +// Decrease the flow level. +func yaml_parser_decrease_flow_level(parser *yaml_parser_t) bool { + if parser.flow_level > 0 { + parser.flow_level-- + last := len(parser.simple_keys) - 1 + delete(parser.simple_keys_by_tok, parser.simple_keys[last].token_number) + parser.simple_keys = parser.simple_keys[:last] + } + return true +} + +// max_indents limits the indents stack size +const max_indents = 10000 + +// Push the current indentation level to the stack and set the new level +// the current column is greater than the indentation level. In this case, +// append or insert the specified token into the token queue. +func yaml_parser_roll_indent(parser *yaml_parser_t, column, number int, typ yaml_token_type_t, mark yaml_mark_t) bool { + // In the flow context, do nothing. + if parser.flow_level > 0 { + return true + } + + if parser.indent < column { + // Push the current indentation level to the stack and set the new + // indentation level. + parser.indents = append(parser.indents, parser.indent) + parser.indent = column + if len(parser.indents) > max_indents { + return yaml_parser_set_scanner_error(parser, + "while increasing indent level", parser.simple_keys[len(parser.simple_keys)-1].mark, + fmt.Sprintf("exceeded max depth of %d", max_indents)) + } + + // Create a token and insert it into the queue. + token := yaml_token_t{ + typ: typ, + start_mark: mark, + end_mark: mark, + } + if number > -1 { + number -= parser.tokens_parsed + } + yaml_insert_token(parser, number, &token) + } + return true +} + +// Pop indentation levels from the indents stack until the current level +// becomes less or equal to the column. For each indentation level, append +// the BLOCK-END token. +func yaml_parser_unroll_indent(parser *yaml_parser_t, column int, scan_mark yaml_mark_t) bool { + // In the flow context, do nothing. + if parser.flow_level > 0 { + return true + } + + block_mark := scan_mark + block_mark.index-- + + // Loop through the indentation levels in the stack. + for parser.indent > column { + + // [Go] Reposition the end token before potential following + // foot comments of parent blocks. For that, search + // backwards for recent comments that were at the same + // indent as the block that is ending now. + stop_index := block_mark.index + for i := len(parser.comments) - 1; i >= 0; i-- { + comment := &parser.comments[i] + + if comment.end_mark.index < stop_index { + // Don't go back beyond the start of the comment/whitespace scan, unless column < 0. + // If requested indent column is < 0, then the document is over and everything else + // is a foot anyway. + break + } + if comment.start_mark.column == parser.indent+1 { + // This is a good match. But maybe there's a former comment + // at that same indent level, so keep searching. + block_mark = comment.start_mark + } + + // While the end of the former comment matches with + // the start of the following one, we know there's + // nothing in between and scanning is still safe. + stop_index = comment.scan_mark.index + } + + // Create a token and append it to the queue. + token := yaml_token_t{ + typ: yaml_BLOCK_END_TOKEN, + start_mark: block_mark, + end_mark: block_mark, + } + yaml_insert_token(parser, -1, &token) + + // Pop the indentation level. + parser.indent = parser.indents[len(parser.indents)-1] + parser.indents = parser.indents[:len(parser.indents)-1] + } + return true +} + +// Initialize the scanner and produce the STREAM-START token. +func yaml_parser_fetch_stream_start(parser *yaml_parser_t) bool { + + // Set the initial indentation. + parser.indent = -1 + + // Initialize the simple key stack. + parser.simple_keys = append(parser.simple_keys, yaml_simple_key_t{}) + + parser.simple_keys_by_tok = make(map[int]int) + + // A simple key is allowed at the beginning of the stream. + parser.simple_key_allowed = true + + // We have started. + parser.stream_start_produced = true + + // Create the STREAM-START token and append it to the queue. + token := yaml_token_t{ + typ: yaml_STREAM_START_TOKEN, + start_mark: parser.mark, + end_mark: parser.mark, + encoding: parser.encoding, + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the STREAM-END token and shut down the scanner. +func yaml_parser_fetch_stream_end(parser *yaml_parser_t) bool { + + // Force new line. + if parser.mark.column != 0 { + parser.mark.column = 0 + parser.mark.line++ + } + + // Reset the indentation level. + if !yaml_parser_unroll_indent(parser, -1, parser.mark) { + return false + } + + // Reset simple keys. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + parser.simple_key_allowed = false + + // Create the STREAM-END token and append it to the queue. + token := yaml_token_t{ + typ: yaml_STREAM_END_TOKEN, + start_mark: parser.mark, + end_mark: parser.mark, + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce a VERSION-DIRECTIVE or TAG-DIRECTIVE token. +func yaml_parser_fetch_directive(parser *yaml_parser_t) bool { + // Reset the indentation level. + if !yaml_parser_unroll_indent(parser, -1, parser.mark) { + return false + } + + // Reset simple keys. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + parser.simple_key_allowed = false + + // Create the YAML-DIRECTIVE or TAG-DIRECTIVE token. + token := yaml_token_t{} + if !yaml_parser_scan_directive(parser, &token) { + return false + } + // Append the token to the queue. + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the DOCUMENT-START or DOCUMENT-END token. +func yaml_parser_fetch_document_indicator(parser *yaml_parser_t, typ yaml_token_type_t) bool { + // Reset the indentation level. + if !yaml_parser_unroll_indent(parser, -1, parser.mark) { + return false + } + + // Reset simple keys. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + parser.simple_key_allowed = false + + // Consume the token. + start_mark := parser.mark + + skip(parser) + skip(parser) + skip(parser) + + end_mark := parser.mark + + // Create the DOCUMENT-START or DOCUMENT-END token. + token := yaml_token_t{ + typ: typ, + start_mark: start_mark, + end_mark: end_mark, + } + // Append the token to the queue. + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the FLOW-SEQUENCE-START or FLOW-MAPPING-START token. +func yaml_parser_fetch_flow_collection_start(parser *yaml_parser_t, typ yaml_token_type_t) bool { + + // The indicators '[' and '{' may start a simple key. + if !yaml_parser_save_simple_key(parser) { + return false + } + + // Increase the flow level. + if !yaml_parser_increase_flow_level(parser) { + return false + } + + // A simple key may follow the indicators '[' and '{'. + parser.simple_key_allowed = true + + // Consume the token. + start_mark := parser.mark + skip(parser) + end_mark := parser.mark + + // Create the FLOW-SEQUENCE-START of FLOW-MAPPING-START token. + token := yaml_token_t{ + typ: typ, + start_mark: start_mark, + end_mark: end_mark, + } + // Append the token to the queue. + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the FLOW-SEQUENCE-END or FLOW-MAPPING-END token. +func yaml_parser_fetch_flow_collection_end(parser *yaml_parser_t, typ yaml_token_type_t) bool { + // Reset any potential simple key on the current flow level. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + // Decrease the flow level. + if !yaml_parser_decrease_flow_level(parser) { + return false + } + + // No simple keys after the indicators ']' and '}'. + parser.simple_key_allowed = false + + // Consume the token. + + start_mark := parser.mark + skip(parser) + end_mark := parser.mark + + // Create the FLOW-SEQUENCE-END of FLOW-MAPPING-END token. + token := yaml_token_t{ + typ: typ, + start_mark: start_mark, + end_mark: end_mark, + } + // Append the token to the queue. + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the FLOW-ENTRY token. +func yaml_parser_fetch_flow_entry(parser *yaml_parser_t) bool { + // Reset any potential simple keys on the current flow level. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + // Simple keys are allowed after ','. + parser.simple_key_allowed = true + + // Consume the token. + start_mark := parser.mark + skip(parser) + end_mark := parser.mark + + // Create the FLOW-ENTRY token and append it to the queue. + token := yaml_token_t{ + typ: yaml_FLOW_ENTRY_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the BLOCK-ENTRY token. +func yaml_parser_fetch_block_entry(parser *yaml_parser_t) bool { + // Check if the scanner is in the block context. + if parser.flow_level == 0 { + // Check if we are allowed to start a new entry. + if !parser.simple_key_allowed { + return yaml_parser_set_scanner_error(parser, "", parser.mark, + "block sequence entries are not allowed in this context") + } + // Add the BLOCK-SEQUENCE-START token if needed. + if !yaml_parser_roll_indent(parser, parser.mark.column, -1, yaml_BLOCK_SEQUENCE_START_TOKEN, parser.mark) { + return false + } + } else { + // It is an error for the '-' indicator to occur in the flow context, + // but we let the Parser detect and report about it because the Parser + // is able to point to the context. + } + + // Reset any potential simple keys on the current flow level. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + // Simple keys are allowed after '-'. + parser.simple_key_allowed = true + + // Consume the token. + start_mark := parser.mark + skip(parser) + end_mark := parser.mark + + // Create the BLOCK-ENTRY token and append it to the queue. + token := yaml_token_t{ + typ: yaml_BLOCK_ENTRY_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the KEY token. +func yaml_parser_fetch_key(parser *yaml_parser_t) bool { + + // In the block context, additional checks are required. + if parser.flow_level == 0 { + // Check if we are allowed to start a new key (not nessesary simple). + if !parser.simple_key_allowed { + return yaml_parser_set_scanner_error(parser, "", parser.mark, + "mapping keys are not allowed in this context") + } + // Add the BLOCK-MAPPING-START token if needed. + if !yaml_parser_roll_indent(parser, parser.mark.column, -1, yaml_BLOCK_MAPPING_START_TOKEN, parser.mark) { + return false + } + } + + // Reset any potential simple keys on the current flow level. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + // Simple keys are allowed after '?' in the block context. + parser.simple_key_allowed = parser.flow_level == 0 + + // Consume the token. + start_mark := parser.mark + skip(parser) + end_mark := parser.mark + + // Create the KEY token and append it to the queue. + token := yaml_token_t{ + typ: yaml_KEY_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the VALUE token. +func yaml_parser_fetch_value(parser *yaml_parser_t) bool { + + simple_key := &parser.simple_keys[len(parser.simple_keys)-1] + + // Have we found a simple key? + if valid, ok := yaml_simple_key_is_valid(parser, simple_key); !ok { + return false + + } else if valid { + + // Create the KEY token and insert it into the queue. + token := yaml_token_t{ + typ: yaml_KEY_TOKEN, + start_mark: simple_key.mark, + end_mark: simple_key.mark, + } + yaml_insert_token(parser, simple_key.token_number-parser.tokens_parsed, &token) + + // In the block context, we may need to add the BLOCK-MAPPING-START token. + if !yaml_parser_roll_indent(parser, simple_key.mark.column, + simple_key.token_number, + yaml_BLOCK_MAPPING_START_TOKEN, simple_key.mark) { + return false + } + + // Remove the simple key. + simple_key.possible = false + delete(parser.simple_keys_by_tok, simple_key.token_number) + + // A simple key cannot follow another simple key. + parser.simple_key_allowed = false + + } else { + // The ':' indicator follows a complex key. + + // In the block context, extra checks are required. + if parser.flow_level == 0 { + + // Check if we are allowed to start a complex value. + if !parser.simple_key_allowed { + return yaml_parser_set_scanner_error(parser, "", parser.mark, + "mapping values are not allowed in this context") + } + + // Add the BLOCK-MAPPING-START token if needed. + if !yaml_parser_roll_indent(parser, parser.mark.column, -1, yaml_BLOCK_MAPPING_START_TOKEN, parser.mark) { + return false + } + } + + // Simple keys after ':' are allowed in the block context. + parser.simple_key_allowed = parser.flow_level == 0 + } + + // Consume the token. + start_mark := parser.mark + skip(parser) + end_mark := parser.mark + + // Create the VALUE token and append it to the queue. + token := yaml_token_t{ + typ: yaml_VALUE_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the ALIAS or ANCHOR token. +func yaml_parser_fetch_anchor(parser *yaml_parser_t, typ yaml_token_type_t) bool { + // An anchor or an alias could be a simple key. + if !yaml_parser_save_simple_key(parser) { + return false + } + + // A simple key cannot follow an anchor or an alias. + parser.simple_key_allowed = false + + // Create the ALIAS or ANCHOR token and append it to the queue. + var token yaml_token_t + if !yaml_parser_scan_anchor(parser, &token, typ) { + return false + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the TAG token. +func yaml_parser_fetch_tag(parser *yaml_parser_t) bool { + // A tag could be a simple key. + if !yaml_parser_save_simple_key(parser) { + return false + } + + // A simple key cannot follow a tag. + parser.simple_key_allowed = false + + // Create the TAG token and append it to the queue. + var token yaml_token_t + if !yaml_parser_scan_tag(parser, &token) { + return false + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the SCALAR(...,literal) or SCALAR(...,folded) tokens. +func yaml_parser_fetch_block_scalar(parser *yaml_parser_t, literal bool) bool { + // Remove any potential simple keys. + if !yaml_parser_remove_simple_key(parser) { + return false + } + + // A simple key may follow a block scalar. + parser.simple_key_allowed = true + + // Create the SCALAR token and append it to the queue. + var token yaml_token_t + if !yaml_parser_scan_block_scalar(parser, &token, literal) { + return false + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the SCALAR(...,single-quoted) or SCALAR(...,double-quoted) tokens. +func yaml_parser_fetch_flow_scalar(parser *yaml_parser_t, single bool) bool { + // A plain scalar could be a simple key. + if !yaml_parser_save_simple_key(parser) { + return false + } + + // A simple key cannot follow a flow scalar. + parser.simple_key_allowed = false + + // Create the SCALAR token and append it to the queue. + var token yaml_token_t + if !yaml_parser_scan_flow_scalar(parser, &token, single) { + return false + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Produce the SCALAR(...,plain) token. +func yaml_parser_fetch_plain_scalar(parser *yaml_parser_t) bool { + // A plain scalar could be a simple key. + if !yaml_parser_save_simple_key(parser) { + return false + } + + // A simple key cannot follow a flow scalar. + parser.simple_key_allowed = false + + // Create the SCALAR token and append it to the queue. + var token yaml_token_t + if !yaml_parser_scan_plain_scalar(parser, &token) { + return false + } + yaml_insert_token(parser, -1, &token) + return true +} + +// Eat whitespaces and comments until the next token is found. +func yaml_parser_scan_to_next_token(parser *yaml_parser_t) bool { + + scan_mark := parser.mark + + // Until the next token is not found. + for { + // Allow the BOM mark to start a line. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if parser.mark.column == 0 && is_bom(parser.buffer, parser.buffer_pos) { + skip(parser) + } + + // Eat whitespaces. + // Tabs are allowed: + // - in the flow context + // - in the block context, but not at the beginning of the line or + // after '-', '?', or ':' (complex value). + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + for parser.buffer[parser.buffer_pos] == ' ' || ((parser.flow_level > 0 || !parser.simple_key_allowed) && parser.buffer[parser.buffer_pos] == '\t') { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Check if we just had a line comment under a sequence entry that + // looks more like a header to the following content. Similar to this: + // + // - # The comment + // - Some data + // + // If so, transform the line comment to a head comment and reposition. + if len(parser.comments) > 0 && len(parser.tokens) > 1 { + tokenA := parser.tokens[len(parser.tokens)-2] + tokenB := parser.tokens[len(parser.tokens)-1] + comment := &parser.comments[len(parser.comments)-1] + if tokenA.typ == yaml_BLOCK_SEQUENCE_START_TOKEN && tokenB.typ == yaml_BLOCK_ENTRY_TOKEN && len(comment.line) > 0 && !is_break(parser.buffer, parser.buffer_pos) { + // If it was in the prior line, reposition so it becomes a + // header of the follow up token. Otherwise, keep it in place + // so it becomes a header of the former. + comment.head = comment.line + comment.line = nil + if comment.start_mark.line == parser.mark.line-1 { + comment.token_mark = parser.mark + } + } + } + + // Eat a comment until a line break. + if parser.buffer[parser.buffer_pos] == '#' { + if !yaml_parser_scan_comments(parser, scan_mark) { + return false + } + } + + // If it is a line break, eat it. + if is_break(parser.buffer, parser.buffer_pos) { + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + skip_line(parser) + + // In the block context, a new line may start a simple key. + if parser.flow_level == 0 { + parser.simple_key_allowed = true + } + } else { + break // We have found a token. + } + } + + return true +} + +// Scan a YAML-DIRECTIVE or TAG-DIRECTIVE token. +// +// Scope: +// %YAML 1.1 # a comment \n +// ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +// %TAG !yaml! tag:yaml.org,2002: \n +// ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +// +func yaml_parser_scan_directive(parser *yaml_parser_t, token *yaml_token_t) bool { + // Eat '%'. + start_mark := parser.mark + skip(parser) + + // Scan the directive name. + var name []byte + if !yaml_parser_scan_directive_name(parser, start_mark, &name) { + return false + } + + // Is it a YAML directive? + if bytes.Equal(name, []byte("YAML")) { + // Scan the VERSION directive value. + var major, minor int8 + if !yaml_parser_scan_version_directive_value(parser, start_mark, &major, &minor) { + return false + } + end_mark := parser.mark + + // Create a VERSION-DIRECTIVE token. + *token = yaml_token_t{ + typ: yaml_VERSION_DIRECTIVE_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + major: major, + minor: minor, + } + + // Is it a TAG directive? + } else if bytes.Equal(name, []byte("TAG")) { + // Scan the TAG directive value. + var handle, prefix []byte + if !yaml_parser_scan_tag_directive_value(parser, start_mark, &handle, &prefix) { + return false + } + end_mark := parser.mark + + // Create a TAG-DIRECTIVE token. + *token = yaml_token_t{ + typ: yaml_TAG_DIRECTIVE_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + value: handle, + prefix: prefix, + } + + // Unknown directive. + } else { + yaml_parser_set_scanner_error(parser, "while scanning a directive", + start_mark, "found unknown directive name") + return false + } + + // Eat the rest of the line including any comments. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + for is_blank(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + if parser.buffer[parser.buffer_pos] == '#' { + // [Go] Discard this inline comment for the time being. + //if !yaml_parser_scan_line_comment(parser, start_mark) { + // return false + //} + for !is_breakz(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + } + + // Check if we are at the end of the line. + if !is_breakz(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a directive", + start_mark, "did not find expected comment or line break") + return false + } + + // Eat a line break. + if is_break(parser.buffer, parser.buffer_pos) { + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + skip_line(parser) + } + + return true +} + +// Scan the directive name. +// +// Scope: +// %YAML 1.1 # a comment \n +// ^^^^ +// %TAG !yaml! tag:yaml.org,2002: \n +// ^^^ +// +func yaml_parser_scan_directive_name(parser *yaml_parser_t, start_mark yaml_mark_t, name *[]byte) bool { + // Consume the directive name. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + var s []byte + for is_alpha(parser.buffer, parser.buffer_pos) { + s = read(parser, s) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Check if the name is empty. + if len(s) == 0 { + yaml_parser_set_scanner_error(parser, "while scanning a directive", + start_mark, "could not find expected directive name") + return false + } + + // Check for an blank character after the name. + if !is_blankz(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a directive", + start_mark, "found unexpected non-alphabetical character") + return false + } + *name = s + return true +} + +// Scan the value of VERSION-DIRECTIVE. +// +// Scope: +// %YAML 1.1 # a comment \n +// ^^^^^^ +func yaml_parser_scan_version_directive_value(parser *yaml_parser_t, start_mark yaml_mark_t, major, minor *int8) bool { + // Eat whitespaces. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + for is_blank(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Consume the major version number. + if !yaml_parser_scan_version_directive_number(parser, start_mark, major) { + return false + } + + // Eat '.'. + if parser.buffer[parser.buffer_pos] != '.' { + return yaml_parser_set_scanner_error(parser, "while scanning a %YAML directive", + start_mark, "did not find expected digit or '.' character") + } + + skip(parser) + + // Consume the minor version number. + if !yaml_parser_scan_version_directive_number(parser, start_mark, minor) { + return false + } + return true +} + +const max_number_length = 2 + +// Scan the version number of VERSION-DIRECTIVE. +// +// Scope: +// %YAML 1.1 # a comment \n +// ^ +// %YAML 1.1 # a comment \n +// ^ +func yaml_parser_scan_version_directive_number(parser *yaml_parser_t, start_mark yaml_mark_t, number *int8) bool { + + // Repeat while the next character is digit. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + var value, length int8 + for is_digit(parser.buffer, parser.buffer_pos) { + // Check if the number is too long. + length++ + if length > max_number_length { + return yaml_parser_set_scanner_error(parser, "while scanning a %YAML directive", + start_mark, "found extremely long version number") + } + value = value*10 + int8(as_digit(parser.buffer, parser.buffer_pos)) + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Check if the number was present. + if length == 0 { + return yaml_parser_set_scanner_error(parser, "while scanning a %YAML directive", + start_mark, "did not find expected version number") + } + *number = value + return true +} + +// Scan the value of a TAG-DIRECTIVE token. +// +// Scope: +// %TAG !yaml! tag:yaml.org,2002: \n +// ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +// +func yaml_parser_scan_tag_directive_value(parser *yaml_parser_t, start_mark yaml_mark_t, handle, prefix *[]byte) bool { + var handle_value, prefix_value []byte + + // Eat whitespaces. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + for is_blank(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Scan a handle. + if !yaml_parser_scan_tag_handle(parser, true, start_mark, &handle_value) { + return false + } + + // Expect a whitespace. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if !is_blank(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a %TAG directive", + start_mark, "did not find expected whitespace") + return false + } + + // Eat whitespaces. + for is_blank(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Scan a prefix. + if !yaml_parser_scan_tag_uri(parser, true, nil, start_mark, &prefix_value) { + return false + } + + // Expect a whitespace or line break. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if !is_blankz(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a %TAG directive", + start_mark, "did not find expected whitespace or line break") + return false + } + + *handle = handle_value + *prefix = prefix_value + return true +} + +func yaml_parser_scan_anchor(parser *yaml_parser_t, token *yaml_token_t, typ yaml_token_type_t) bool { + var s []byte + + // Eat the indicator character. + start_mark := parser.mark + skip(parser) + + // Consume the value. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + for is_alpha(parser.buffer, parser.buffer_pos) { + s = read(parser, s) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + end_mark := parser.mark + + /* + * Check if length of the anchor is greater than 0 and it is followed by + * a whitespace character or one of the indicators: + * + * '?', ':', ',', ']', '}', '%', '@', '`'. + */ + + if len(s) == 0 || + !(is_blankz(parser.buffer, parser.buffer_pos) || parser.buffer[parser.buffer_pos] == '?' || + parser.buffer[parser.buffer_pos] == ':' || parser.buffer[parser.buffer_pos] == ',' || + parser.buffer[parser.buffer_pos] == ']' || parser.buffer[parser.buffer_pos] == '}' || + parser.buffer[parser.buffer_pos] == '%' || parser.buffer[parser.buffer_pos] == '@' || + parser.buffer[parser.buffer_pos] == '`') { + context := "while scanning an alias" + if typ == yaml_ANCHOR_TOKEN { + context = "while scanning an anchor" + } + yaml_parser_set_scanner_error(parser, context, start_mark, + "did not find expected alphabetic or numeric character") + return false + } + + // Create a token. + *token = yaml_token_t{ + typ: typ, + start_mark: start_mark, + end_mark: end_mark, + value: s, + } + + return true +} + +/* + * Scan a TAG token. + */ + +func yaml_parser_scan_tag(parser *yaml_parser_t, token *yaml_token_t) bool { + var handle, suffix []byte + + start_mark := parser.mark + + // Check if the tag is in the canonical form. + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + + if parser.buffer[parser.buffer_pos+1] == '<' { + // Keep the handle as '' + + // Eat '!<' + skip(parser) + skip(parser) + + // Consume the tag value. + if !yaml_parser_scan_tag_uri(parser, false, nil, start_mark, &suffix) { + return false + } + + // Check for '>' and eat it. + if parser.buffer[parser.buffer_pos] != '>' { + yaml_parser_set_scanner_error(parser, "while scanning a tag", + start_mark, "did not find the expected '>'") + return false + } + + skip(parser) + } else { + // The tag has either the '!suffix' or the '!handle!suffix' form. + + // First, try to scan a handle. + if !yaml_parser_scan_tag_handle(parser, false, start_mark, &handle) { + return false + } + + // Check if it is, indeed, handle. + if handle[0] == '!' && len(handle) > 1 && handle[len(handle)-1] == '!' { + // Scan the suffix now. + if !yaml_parser_scan_tag_uri(parser, false, nil, start_mark, &suffix) { + return false + } + } else { + // It wasn't a handle after all. Scan the rest of the tag. + if !yaml_parser_scan_tag_uri(parser, false, handle, start_mark, &suffix) { + return false + } + + // Set the handle to '!'. + handle = []byte{'!'} + + // A special case: the '!' tag. Set the handle to '' and the + // suffix to '!'. + if len(suffix) == 0 { + handle, suffix = suffix, handle + } + } + } + + // Check the character which ends the tag. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if !is_blankz(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a tag", + start_mark, "did not find expected whitespace or line break") + return false + } + + end_mark := parser.mark + + // Create a token. + *token = yaml_token_t{ + typ: yaml_TAG_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + value: handle, + suffix: suffix, + } + return true +} + +// Scan a tag handle. +func yaml_parser_scan_tag_handle(parser *yaml_parser_t, directive bool, start_mark yaml_mark_t, handle *[]byte) bool { + // Check the initial '!' character. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if parser.buffer[parser.buffer_pos] != '!' { + yaml_parser_set_scanner_tag_error(parser, directive, + start_mark, "did not find expected '!'") + return false + } + + var s []byte + + // Copy the '!' character. + s = read(parser, s) + + // Copy all subsequent alphabetical and numerical characters. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + for is_alpha(parser.buffer, parser.buffer_pos) { + s = read(parser, s) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Check if the trailing character is '!' and copy it. + if parser.buffer[parser.buffer_pos] == '!' { + s = read(parser, s) + } else { + // It's either the '!' tag or not really a tag handle. If it's a %TAG + // directive, it's an error. If it's a tag token, it must be a part of URI. + if directive && string(s) != "!" { + yaml_parser_set_scanner_tag_error(parser, directive, + start_mark, "did not find expected '!'") + return false + } + } + + *handle = s + return true +} + +// Scan a tag. +func yaml_parser_scan_tag_uri(parser *yaml_parser_t, directive bool, head []byte, start_mark yaml_mark_t, uri *[]byte) bool { + //size_t length = head ? strlen((char *)head) : 0 + var s []byte + hasTag := len(head) > 0 + + // Copy the head if needed. + // + // Note that we don't copy the leading '!' character. + if len(head) > 1 { + s = append(s, head[1:]...) + } + + // Scan the tag. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + // The set of characters that may appear in URI is as follows: + // + // '0'-'9', 'A'-'Z', 'a'-'z', '_', '-', ';', '/', '?', ':', '@', '&', + // '=', '+', '$', ',', '.', '!', '~', '*', '\'', '(', ')', '[', ']', + // '%'. + // [Go] TODO Convert this into more reasonable logic. + for is_alpha(parser.buffer, parser.buffer_pos) || parser.buffer[parser.buffer_pos] == ';' || + parser.buffer[parser.buffer_pos] == '/' || parser.buffer[parser.buffer_pos] == '?' || + parser.buffer[parser.buffer_pos] == ':' || parser.buffer[parser.buffer_pos] == '@' || + parser.buffer[parser.buffer_pos] == '&' || parser.buffer[parser.buffer_pos] == '=' || + parser.buffer[parser.buffer_pos] == '+' || parser.buffer[parser.buffer_pos] == '$' || + parser.buffer[parser.buffer_pos] == ',' || parser.buffer[parser.buffer_pos] == '.' || + parser.buffer[parser.buffer_pos] == '!' || parser.buffer[parser.buffer_pos] == '~' || + parser.buffer[parser.buffer_pos] == '*' || parser.buffer[parser.buffer_pos] == '\'' || + parser.buffer[parser.buffer_pos] == '(' || parser.buffer[parser.buffer_pos] == ')' || + parser.buffer[parser.buffer_pos] == '[' || parser.buffer[parser.buffer_pos] == ']' || + parser.buffer[parser.buffer_pos] == '%' { + // Check if it is a URI-escape sequence. + if parser.buffer[parser.buffer_pos] == '%' { + if !yaml_parser_scan_uri_escapes(parser, directive, start_mark, &s) { + return false + } + } else { + s = read(parser, s) + } + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + hasTag = true + } + + if !hasTag { + yaml_parser_set_scanner_tag_error(parser, directive, + start_mark, "did not find expected tag URI") + return false + } + *uri = s + return true +} + +// Decode an URI-escape sequence corresponding to a single UTF-8 character. +func yaml_parser_scan_uri_escapes(parser *yaml_parser_t, directive bool, start_mark yaml_mark_t, s *[]byte) bool { + + // Decode the required number of characters. + w := 1024 + for w > 0 { + // Check for a URI-escaped octet. + if parser.unread < 3 && !yaml_parser_update_buffer(parser, 3) { + return false + } + + if !(parser.buffer[parser.buffer_pos] == '%' && + is_hex(parser.buffer, parser.buffer_pos+1) && + is_hex(parser.buffer, parser.buffer_pos+2)) { + return yaml_parser_set_scanner_tag_error(parser, directive, + start_mark, "did not find URI escaped octet") + } + + // Get the octet. + octet := byte((as_hex(parser.buffer, parser.buffer_pos+1) << 4) + as_hex(parser.buffer, parser.buffer_pos+2)) + + // If it is the leading octet, determine the length of the UTF-8 sequence. + if w == 1024 { + w = width(octet) + if w == 0 { + return yaml_parser_set_scanner_tag_error(parser, directive, + start_mark, "found an incorrect leading UTF-8 octet") + } + } else { + // Check if the trailing octet is correct. + if octet&0xC0 != 0x80 { + return yaml_parser_set_scanner_tag_error(parser, directive, + start_mark, "found an incorrect trailing UTF-8 octet") + } + } + + // Copy the octet and move the pointers. + *s = append(*s, octet) + skip(parser) + skip(parser) + skip(parser) + w-- + } + return true +} + +// Scan a block scalar. +func yaml_parser_scan_block_scalar(parser *yaml_parser_t, token *yaml_token_t, literal bool) bool { + // Eat the indicator '|' or '>'. + start_mark := parser.mark + skip(parser) + + // Scan the additional block scalar indicators. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + // Check for a chomping indicator. + var chomping, increment int + if parser.buffer[parser.buffer_pos] == '+' || parser.buffer[parser.buffer_pos] == '-' { + // Set the chomping method and eat the indicator. + if parser.buffer[parser.buffer_pos] == '+' { + chomping = +1 + } else { + chomping = -1 + } + skip(parser) + + // Check for an indentation indicator. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if is_digit(parser.buffer, parser.buffer_pos) { + // Check that the indentation is greater than 0. + if parser.buffer[parser.buffer_pos] == '0' { + yaml_parser_set_scanner_error(parser, "while scanning a block scalar", + start_mark, "found an indentation indicator equal to 0") + return false + } + + // Get the indentation level and eat the indicator. + increment = as_digit(parser.buffer, parser.buffer_pos) + skip(parser) + } + + } else if is_digit(parser.buffer, parser.buffer_pos) { + // Do the same as above, but in the opposite order. + + if parser.buffer[parser.buffer_pos] == '0' { + yaml_parser_set_scanner_error(parser, "while scanning a block scalar", + start_mark, "found an indentation indicator equal to 0") + return false + } + increment = as_digit(parser.buffer, parser.buffer_pos) + skip(parser) + + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if parser.buffer[parser.buffer_pos] == '+' || parser.buffer[parser.buffer_pos] == '-' { + if parser.buffer[parser.buffer_pos] == '+' { + chomping = +1 + } else { + chomping = -1 + } + skip(parser) + } + } + + // Eat whitespaces and comments to the end of the line. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + for is_blank(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + if parser.buffer[parser.buffer_pos] == '#' { + // TODO Test this and then re-enable it. + //if !yaml_parser_scan_line_comment(parser, start_mark) { + // return false + //} + for !is_breakz(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + } + + // Check if we are at the end of the line. + if !is_breakz(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a block scalar", + start_mark, "did not find expected comment or line break") + return false + } + + // Eat a line break. + if is_break(parser.buffer, parser.buffer_pos) { + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + skip_line(parser) + } + + end_mark := parser.mark + + // Set the indentation level if it was specified. + var indent int + if increment > 0 { + if parser.indent >= 0 { + indent = parser.indent + increment + } else { + indent = increment + } + } + + // Scan the leading line breaks and determine the indentation level if needed. + var s, leading_break, trailing_breaks []byte + if !yaml_parser_scan_block_scalar_breaks(parser, &indent, &trailing_breaks, start_mark, &end_mark) { + return false + } + + // Scan the block scalar content. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + var leading_blank, trailing_blank bool + for parser.mark.column == indent && !is_z(parser.buffer, parser.buffer_pos) { + // We are at the beginning of a non-empty line. + + // Is it a trailing whitespace? + trailing_blank = is_blank(parser.buffer, parser.buffer_pos) + + // Check if we need to fold the leading line break. + if !literal && !leading_blank && !trailing_blank && len(leading_break) > 0 && leading_break[0] == '\n' { + // Do we need to join the lines by space? + if len(trailing_breaks) == 0 { + s = append(s, ' ') + } + } else { + s = append(s, leading_break...) + } + leading_break = leading_break[:0] + + // Append the remaining line breaks. + s = append(s, trailing_breaks...) + trailing_breaks = trailing_breaks[:0] + + // Is it a leading whitespace? + leading_blank = is_blank(parser.buffer, parser.buffer_pos) + + // Consume the current line. + for !is_breakz(parser.buffer, parser.buffer_pos) { + s = read(parser, s) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Consume the line break. + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + + leading_break = read_line(parser, leading_break) + + // Eat the following indentation spaces and line breaks. + if !yaml_parser_scan_block_scalar_breaks(parser, &indent, &trailing_breaks, start_mark, &end_mark) { + return false + } + } + + // Chomp the tail. + if chomping != -1 { + s = append(s, leading_break...) + } + if chomping == 1 { + s = append(s, trailing_breaks...) + } + + // Create a token. + *token = yaml_token_t{ + typ: yaml_SCALAR_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + value: s, + style: yaml_LITERAL_SCALAR_STYLE, + } + if !literal { + token.style = yaml_FOLDED_SCALAR_STYLE + } + return true +} + +// Scan indentation spaces and line breaks for a block scalar. Determine the +// indentation level if needed. +func yaml_parser_scan_block_scalar_breaks(parser *yaml_parser_t, indent *int, breaks *[]byte, start_mark yaml_mark_t, end_mark *yaml_mark_t) bool { + *end_mark = parser.mark + + // Eat the indentation spaces and line breaks. + max_indent := 0 + for { + // Eat the indentation spaces. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + for (*indent == 0 || parser.mark.column < *indent) && is_space(parser.buffer, parser.buffer_pos) { + skip(parser) + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + if parser.mark.column > max_indent { + max_indent = parser.mark.column + } + + // Check for a tab character messing the indentation. + if (*indent == 0 || parser.mark.column < *indent) && is_tab(parser.buffer, parser.buffer_pos) { + return yaml_parser_set_scanner_error(parser, "while scanning a block scalar", + start_mark, "found a tab character where an indentation space is expected") + } + + // Have we found a non-empty line? + if !is_break(parser.buffer, parser.buffer_pos) { + break + } + + // Consume the line break. + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + // [Go] Should really be returning breaks instead. + *breaks = read_line(parser, *breaks) + *end_mark = parser.mark + } + + // Determine the indentation level if needed. + if *indent == 0 { + *indent = max_indent + if *indent < parser.indent+1 { + *indent = parser.indent + 1 + } + if *indent < 1 { + *indent = 1 + } + } + return true +} + +// Scan a quoted scalar. +func yaml_parser_scan_flow_scalar(parser *yaml_parser_t, token *yaml_token_t, single bool) bool { + // Eat the left quote. + start_mark := parser.mark + skip(parser) + + // Consume the content of the quoted scalar. + var s, leading_break, trailing_breaks, whitespaces []byte + for { + // Check that there are no document indicators at the beginning of the line. + if parser.unread < 4 && !yaml_parser_update_buffer(parser, 4) { + return false + } + + if parser.mark.column == 0 && + ((parser.buffer[parser.buffer_pos+0] == '-' && + parser.buffer[parser.buffer_pos+1] == '-' && + parser.buffer[parser.buffer_pos+2] == '-') || + (parser.buffer[parser.buffer_pos+0] == '.' && + parser.buffer[parser.buffer_pos+1] == '.' && + parser.buffer[parser.buffer_pos+2] == '.')) && + is_blankz(parser.buffer, parser.buffer_pos+3) { + yaml_parser_set_scanner_error(parser, "while scanning a quoted scalar", + start_mark, "found unexpected document indicator") + return false + } + + // Check for EOF. + if is_z(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a quoted scalar", + start_mark, "found unexpected end of stream") + return false + } + + // Consume non-blank characters. + leading_blanks := false + for !is_blankz(parser.buffer, parser.buffer_pos) { + if single && parser.buffer[parser.buffer_pos] == '\'' && parser.buffer[parser.buffer_pos+1] == '\'' { + // Is is an escaped single quote. + s = append(s, '\'') + skip(parser) + skip(parser) + + } else if single && parser.buffer[parser.buffer_pos] == '\'' { + // It is a right single quote. + break + } else if !single && parser.buffer[parser.buffer_pos] == '"' { + // It is a right double quote. + break + + } else if !single && parser.buffer[parser.buffer_pos] == '\\' && is_break(parser.buffer, parser.buffer_pos+1) { + // It is an escaped line break. + if parser.unread < 3 && !yaml_parser_update_buffer(parser, 3) { + return false + } + skip(parser) + skip_line(parser) + leading_blanks = true + break + + } else if !single && parser.buffer[parser.buffer_pos] == '\\' { + // It is an escape sequence. + code_length := 0 + + // Check the escape character. + switch parser.buffer[parser.buffer_pos+1] { + case '0': + s = append(s, 0) + case 'a': + s = append(s, '\x07') + case 'b': + s = append(s, '\x08') + case 't', '\t': + s = append(s, '\x09') + case 'n': + s = append(s, '\x0A') + case 'v': + s = append(s, '\x0B') + case 'f': + s = append(s, '\x0C') + case 'r': + s = append(s, '\x0D') + case 'e': + s = append(s, '\x1B') + case ' ': + s = append(s, '\x20') + case '"': + s = append(s, '"') + case '\'': + s = append(s, '\'') + case '\\': + s = append(s, '\\') + case 'N': // NEL (#x85) + s = append(s, '\xC2') + s = append(s, '\x85') + case '_': // #xA0 + s = append(s, '\xC2') + s = append(s, '\xA0') + case 'L': // LS (#x2028) + s = append(s, '\xE2') + s = append(s, '\x80') + s = append(s, '\xA8') + case 'P': // PS (#x2029) + s = append(s, '\xE2') + s = append(s, '\x80') + s = append(s, '\xA9') + case 'x': + code_length = 2 + case 'u': + code_length = 4 + case 'U': + code_length = 8 + default: + yaml_parser_set_scanner_error(parser, "while parsing a quoted scalar", + start_mark, "found unknown escape character") + return false + } + + skip(parser) + skip(parser) + + // Consume an arbitrary escape code. + if code_length > 0 { + var value int + + // Scan the character value. + if parser.unread < code_length && !yaml_parser_update_buffer(parser, code_length) { + return false + } + for k := 0; k < code_length; k++ { + if !is_hex(parser.buffer, parser.buffer_pos+k) { + yaml_parser_set_scanner_error(parser, "while parsing a quoted scalar", + start_mark, "did not find expected hexdecimal number") + return false + } + value = (value << 4) + as_hex(parser.buffer, parser.buffer_pos+k) + } + + // Check the value and write the character. + if (value >= 0xD800 && value <= 0xDFFF) || value > 0x10FFFF { + yaml_parser_set_scanner_error(parser, "while parsing a quoted scalar", + start_mark, "found invalid Unicode character escape code") + return false + } + if value <= 0x7F { + s = append(s, byte(value)) + } else if value <= 0x7FF { + s = append(s, byte(0xC0+(value>>6))) + s = append(s, byte(0x80+(value&0x3F))) + } else if value <= 0xFFFF { + s = append(s, byte(0xE0+(value>>12))) + s = append(s, byte(0x80+((value>>6)&0x3F))) + s = append(s, byte(0x80+(value&0x3F))) + } else { + s = append(s, byte(0xF0+(value>>18))) + s = append(s, byte(0x80+((value>>12)&0x3F))) + s = append(s, byte(0x80+((value>>6)&0x3F))) + s = append(s, byte(0x80+(value&0x3F))) + } + + // Advance the pointer. + for k := 0; k < code_length; k++ { + skip(parser) + } + } + } else { + // It is a non-escaped non-blank character. + s = read(parser, s) + } + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + } + + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + // Check if we are at the end of the scalar. + if single { + if parser.buffer[parser.buffer_pos] == '\'' { + break + } + } else { + if parser.buffer[parser.buffer_pos] == '"' { + break + } + } + + // Consume blank characters. + for is_blank(parser.buffer, parser.buffer_pos) || is_break(parser.buffer, parser.buffer_pos) { + if is_blank(parser.buffer, parser.buffer_pos) { + // Consume a space or a tab character. + if !leading_blanks { + whitespaces = read(parser, whitespaces) + } else { + skip(parser) + } + } else { + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + + // Check if it is a first line break. + if !leading_blanks { + whitespaces = whitespaces[:0] + leading_break = read_line(parser, leading_break) + leading_blanks = true + } else { + trailing_breaks = read_line(parser, trailing_breaks) + } + } + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Join the whitespaces or fold line breaks. + if leading_blanks { + // Do we need to fold line breaks? + if len(leading_break) > 0 && leading_break[0] == '\n' { + if len(trailing_breaks) == 0 { + s = append(s, ' ') + } else { + s = append(s, trailing_breaks...) + } + } else { + s = append(s, leading_break...) + s = append(s, trailing_breaks...) + } + trailing_breaks = trailing_breaks[:0] + leading_break = leading_break[:0] + } else { + s = append(s, whitespaces...) + whitespaces = whitespaces[:0] + } + } + + // Eat the right quote. + skip(parser) + end_mark := parser.mark + + // Create a token. + *token = yaml_token_t{ + typ: yaml_SCALAR_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + value: s, + style: yaml_SINGLE_QUOTED_SCALAR_STYLE, + } + if !single { + token.style = yaml_DOUBLE_QUOTED_SCALAR_STYLE + } + return true +} + +// Scan a plain scalar. +func yaml_parser_scan_plain_scalar(parser *yaml_parser_t, token *yaml_token_t) bool { + + var s, leading_break, trailing_breaks, whitespaces []byte + var leading_blanks bool + var indent = parser.indent + 1 + + start_mark := parser.mark + end_mark := parser.mark + + // Consume the content of the plain scalar. + for { + // Check for a document indicator. + if parser.unread < 4 && !yaml_parser_update_buffer(parser, 4) { + return false + } + if parser.mark.column == 0 && + ((parser.buffer[parser.buffer_pos+0] == '-' && + parser.buffer[parser.buffer_pos+1] == '-' && + parser.buffer[parser.buffer_pos+2] == '-') || + (parser.buffer[parser.buffer_pos+0] == '.' && + parser.buffer[parser.buffer_pos+1] == '.' && + parser.buffer[parser.buffer_pos+2] == '.')) && + is_blankz(parser.buffer, parser.buffer_pos+3) { + break + } + + // Check for a comment. + if parser.buffer[parser.buffer_pos] == '#' { + break + } + + // Consume non-blank characters. + for !is_blankz(parser.buffer, parser.buffer_pos) { + + // Check for indicators that may end a plain scalar. + if (parser.buffer[parser.buffer_pos] == ':' && is_blankz(parser.buffer, parser.buffer_pos+1)) || + (parser.flow_level > 0 && + (parser.buffer[parser.buffer_pos] == ',' || + parser.buffer[parser.buffer_pos] == '?' || parser.buffer[parser.buffer_pos] == '[' || + parser.buffer[parser.buffer_pos] == ']' || parser.buffer[parser.buffer_pos] == '{' || + parser.buffer[parser.buffer_pos] == '}')) { + break + } + + // Check if we need to join whitespaces and breaks. + if leading_blanks || len(whitespaces) > 0 { + if leading_blanks { + // Do we need to fold line breaks? + if leading_break[0] == '\n' { + if len(trailing_breaks) == 0 { + s = append(s, ' ') + } else { + s = append(s, trailing_breaks...) + } + } else { + s = append(s, leading_break...) + s = append(s, trailing_breaks...) + } + trailing_breaks = trailing_breaks[:0] + leading_break = leading_break[:0] + leading_blanks = false + } else { + s = append(s, whitespaces...) + whitespaces = whitespaces[:0] + } + } + + // Copy the character. + s = read(parser, s) + + end_mark = parser.mark + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + } + + // Is it the end? + if !(is_blank(parser.buffer, parser.buffer_pos) || is_break(parser.buffer, parser.buffer_pos)) { + break + } + + // Consume blank characters. + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + + for is_blank(parser.buffer, parser.buffer_pos) || is_break(parser.buffer, parser.buffer_pos) { + if is_blank(parser.buffer, parser.buffer_pos) { + + // Check for tab characters that abuse indentation. + if leading_blanks && parser.mark.column < indent && is_tab(parser.buffer, parser.buffer_pos) { + yaml_parser_set_scanner_error(parser, "while scanning a plain scalar", + start_mark, "found a tab character that violates indentation") + return false + } + + // Consume a space or a tab character. + if !leading_blanks { + whitespaces = read(parser, whitespaces) + } else { + skip(parser) + } + } else { + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + + // Check if it is a first line break. + if !leading_blanks { + whitespaces = whitespaces[:0] + leading_break = read_line(parser, leading_break) + leading_blanks = true + } else { + trailing_breaks = read_line(parser, trailing_breaks) + } + } + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + } + + // Check indentation level. + if parser.flow_level == 0 && parser.mark.column < indent { + break + } + } + + // Create a token. + *token = yaml_token_t{ + typ: yaml_SCALAR_TOKEN, + start_mark: start_mark, + end_mark: end_mark, + value: s, + style: yaml_PLAIN_SCALAR_STYLE, + } + + // Note that we change the 'simple_key_allowed' flag. + if leading_blanks { + parser.simple_key_allowed = true + } + return true +} + +func yaml_parser_scan_line_comment(parser *yaml_parser_t, token_mark yaml_mark_t) bool { + if parser.newlines > 0 { + return true + } + + var start_mark yaml_mark_t + var text []byte + + for peek := 0; peek < 512; peek++ { + if parser.unread < peek+1 && !yaml_parser_update_buffer(parser, peek+1) { + break + } + if is_blank(parser.buffer, parser.buffer_pos+peek) { + continue + } + if parser.buffer[parser.buffer_pos+peek] == '#' { + seen := parser.mark.index+peek + for { + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if is_breakz(parser.buffer, parser.buffer_pos) { + if parser.mark.index >= seen { + break + } + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + skip_line(parser) + } else if parser.mark.index >= seen { + if len(text) == 0 { + start_mark = parser.mark + } + text = read(parser, text) + } else { + skip(parser) + } + } + } + break + } + if len(text) > 0 { + parser.comments = append(parser.comments, yaml_comment_t{ + token_mark: token_mark, + start_mark: start_mark, + line: text, + }) + } + return true +} + +func yaml_parser_scan_comments(parser *yaml_parser_t, scan_mark yaml_mark_t) bool { + token := parser.tokens[len(parser.tokens)-1] + + if token.typ == yaml_FLOW_ENTRY_TOKEN && len(parser.tokens) > 1 { + token = parser.tokens[len(parser.tokens)-2] + } + + var token_mark = token.start_mark + var start_mark yaml_mark_t + + var recent_empty = false + var first_empty = parser.newlines <= 1 + + var line = parser.mark.line + var column = parser.mark.column + + var text []byte + + // The foot line is the place where a comment must start to + // still be considered as a foot of the prior content. + // If there's some content in the currently parsed line, then + // the foot is the line below it. + var foot_line = -1 + if scan_mark.line > 0 { + foot_line = parser.mark.line-parser.newlines+1 + if parser.newlines == 0 && parser.mark.column > 1 { + foot_line++ + } + } + + var peek = 0 + for ; peek < 512; peek++ { + if parser.unread < peek+1 && !yaml_parser_update_buffer(parser, peek+1) { + break + } + column++ + if is_blank(parser.buffer, parser.buffer_pos+peek) { + continue + } + c := parser.buffer[parser.buffer_pos+peek] + if is_breakz(parser.buffer, parser.buffer_pos+peek) || parser.flow_level > 0 && (c == ']' || c == '}') { + // Got line break or terminator. + if !recent_empty { + if first_empty && (start_mark.line == foot_line || start_mark.column-1 < parser.indent) { + // This is the first empty line and there were no empty lines before, + // so this initial part of the comment is a foot of the prior token + // instead of being a head for the following one. Split it up. + if len(text) > 0 { + if start_mark.column-1 < parser.indent { + // If dedented it's unrelated to the prior token. + token_mark = start_mark + } + parser.comments = append(parser.comments, yaml_comment_t{ + scan_mark: scan_mark, + token_mark: token_mark, + start_mark: start_mark, + end_mark: yaml_mark_t{parser.mark.index + peek, line, column}, + foot: text, + }) + scan_mark = yaml_mark_t{parser.mark.index + peek, line, column} + token_mark = scan_mark + text = nil + } + } else { + if len(text) > 0 && parser.buffer[parser.buffer_pos+peek] != 0 { + text = append(text, '\n') + } + } + } + if !is_break(parser.buffer, parser.buffer_pos+peek) { + break + } + first_empty = false + recent_empty = true + column = 0 + line++ + continue + } + + if len(text) > 0 && column < parser.indent+1 && column != start_mark.column { + // The comment at the different indentation is a foot of the + // preceding data rather than a head of the upcoming one. + parser.comments = append(parser.comments, yaml_comment_t{ + scan_mark: scan_mark, + token_mark: token_mark, + start_mark: start_mark, + end_mark: yaml_mark_t{parser.mark.index + peek, line, column}, + foot: text, + }) + scan_mark = yaml_mark_t{parser.mark.index + peek, line, column} + token_mark = scan_mark + text = nil + } + + if parser.buffer[parser.buffer_pos+peek] != '#' { + break + } + + if len(text) == 0 { + start_mark = yaml_mark_t{parser.mark.index + peek, line, column} + } else { + text = append(text, '\n') + } + + recent_empty = false + + // Consume until after the consumed comment line. + seen := parser.mark.index+peek + for { + if parser.unread < 1 && !yaml_parser_update_buffer(parser, 1) { + return false + } + if is_breakz(parser.buffer, parser.buffer_pos) { + if parser.mark.index >= seen { + break + } + if parser.unread < 2 && !yaml_parser_update_buffer(parser, 2) { + return false + } + skip_line(parser) + } else if parser.mark.index >= seen { + text = read(parser, text) + } else { + skip(parser) + } + } + + peek = 0 + column = 0 + line = parser.mark.line + } + + if len(text) > 0 { + parser.comments = append(parser.comments, yaml_comment_t{ + scan_mark: scan_mark, + token_mark: start_mark, + start_mark: start_mark, + end_mark: yaml_mark_t{parser.mark.index + peek - 1, line, column}, + head: text, + }) + } + return true +} diff --git a/vendor/gopkg.in/yaml.v3/sorter.go b/vendor/gopkg.in/yaml.v3/sorter.go new file mode 100644 index 00000000..9210ece7 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/sorter.go @@ -0,0 +1,134 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package yaml + +import ( + "reflect" + "unicode" +) + +type keyList []reflect.Value + +func (l keyList) Len() int { return len(l) } +func (l keyList) Swap(i, j int) { l[i], l[j] = l[j], l[i] } +func (l keyList) Less(i, j int) bool { + a := l[i] + b := l[j] + ak := a.Kind() + bk := b.Kind() + for (ak == reflect.Interface || ak == reflect.Ptr) && !a.IsNil() { + a = a.Elem() + ak = a.Kind() + } + for (bk == reflect.Interface || bk == reflect.Ptr) && !b.IsNil() { + b = b.Elem() + bk = b.Kind() + } + af, aok := keyFloat(a) + bf, bok := keyFloat(b) + if aok && bok { + if af != bf { + return af < bf + } + if ak != bk { + return ak < bk + } + return numLess(a, b) + } + if ak != reflect.String || bk != reflect.String { + return ak < bk + } + ar, br := []rune(a.String()), []rune(b.String()) + digits := false + for i := 0; i < len(ar) && i < len(br); i++ { + if ar[i] == br[i] { + digits = unicode.IsDigit(ar[i]) + continue + } + al := unicode.IsLetter(ar[i]) + bl := unicode.IsLetter(br[i]) + if al && bl { + return ar[i] < br[i] + } + if al || bl { + if digits { + return al + } else { + return bl + } + } + var ai, bi int + var an, bn int64 + if ar[i] == '0' || br[i] == '0' { + for j := i - 1; j >= 0 && unicode.IsDigit(ar[j]); j-- { + if ar[j] != '0' { + an = 1 + bn = 1 + break + } + } + } + for ai = i; ai < len(ar) && unicode.IsDigit(ar[ai]); ai++ { + an = an*10 + int64(ar[ai]-'0') + } + for bi = i; bi < len(br) && unicode.IsDigit(br[bi]); bi++ { + bn = bn*10 + int64(br[bi]-'0') + } + if an != bn { + return an < bn + } + if ai != bi { + return ai < bi + } + return ar[i] < br[i] + } + return len(ar) < len(br) +} + +// keyFloat returns a float value for v if it is a number/bool +// and whether it is a number/bool or not. +func keyFloat(v reflect.Value) (f float64, ok bool) { + switch v.Kind() { + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + return float64(v.Int()), true + case reflect.Float32, reflect.Float64: + return v.Float(), true + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: + return float64(v.Uint()), true + case reflect.Bool: + if v.Bool() { + return 1, true + } + return 0, true + } + return 0, false +} + +// numLess returns whether a < b. +// a and b must necessarily have the same kind. +func numLess(a, b reflect.Value) bool { + switch a.Kind() { + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + return a.Int() < b.Int() + case reflect.Float32, reflect.Float64: + return a.Float() < b.Float() + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: + return a.Uint() < b.Uint() + case reflect.Bool: + return !a.Bool() && b.Bool() + } + panic("not a number") +} diff --git a/vendor/gopkg.in/yaml.v3/writerc.go b/vendor/gopkg.in/yaml.v3/writerc.go new file mode 100644 index 00000000..b8a116bf --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/writerc.go @@ -0,0 +1,48 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +// Set the writer error and return false. +func yaml_emitter_set_writer_error(emitter *yaml_emitter_t, problem string) bool { + emitter.error = yaml_WRITER_ERROR + emitter.problem = problem + return false +} + +// Flush the output buffer. +func yaml_emitter_flush(emitter *yaml_emitter_t) bool { + if emitter.write_handler == nil { + panic("write handler not set") + } + + // Check if the buffer is empty. + if emitter.buffer_pos == 0 { + return true + } + + if err := emitter.write_handler(emitter, emitter.buffer[:emitter.buffer_pos]); err != nil { + return yaml_emitter_set_writer_error(emitter, "write error: "+err.Error()) + } + emitter.buffer_pos = 0 + return true +} diff --git a/vendor/gopkg.in/yaml.v3/yaml.go b/vendor/gopkg.in/yaml.v3/yaml.go new file mode 100644 index 00000000..56e8a849 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/yaml.go @@ -0,0 +1,693 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Package yaml implements YAML support for the Go language. +// +// Source code and other details for the project are available at GitHub: +// +// https://github.com/go-yaml/yaml +// +package yaml + +import ( + "errors" + "fmt" + "io" + "reflect" + "strings" + "sync" + "unicode/utf8" +) + +// The Unmarshaler interface may be implemented by types to customize their +// behavior when being unmarshaled from a YAML document. +type Unmarshaler interface { + UnmarshalYAML(value *Node) error +} + +type obsoleteUnmarshaler interface { + UnmarshalYAML(unmarshal func(interface{}) error) error +} + +// The Marshaler interface may be implemented by types to customize their +// behavior when being marshaled into a YAML document. The returned value +// is marshaled in place of the original value implementing Marshaler. +// +// If an error is returned by MarshalYAML, the marshaling procedure stops +// and returns with the provided error. +type Marshaler interface { + MarshalYAML() (interface{}, error) +} + +// Unmarshal decodes the first document found within the in byte slice +// and assigns decoded values into the out value. +// +// Maps and pointers (to a struct, string, int, etc) are accepted as out +// values. If an internal pointer within a struct is not initialized, +// the yaml package will initialize it if necessary for unmarshalling +// the provided data. The out parameter must not be nil. +// +// The type of the decoded values should be compatible with the respective +// values in out. If one or more values cannot be decoded due to a type +// mismatches, decoding continues partially until the end of the YAML +// content, and a *yaml.TypeError is returned with details for all +// missed values. +// +// Struct fields are only unmarshalled if they are exported (have an +// upper case first letter), and are unmarshalled using the field name +// lowercased as the default key. Custom keys may be defined via the +// "yaml" name in the field tag: the content preceding the first comma +// is used as the key, and the following comma-separated options are +// used to tweak the marshalling process (see Marshal). +// Conflicting names result in a runtime error. +// +// For example: +// +// type T struct { +// F int `yaml:"a,omitempty"` +// B int +// } +// var t T +// yaml.Unmarshal([]byte("a: 1\nb: 2"), &t) +// +// See the documentation of Marshal for the format of tags and a list of +// supported tag options. +// +func Unmarshal(in []byte, out interface{}) (err error) { + return unmarshal(in, out, false) +} + +// A Decoder reads and decodes YAML values from an input stream. +type Decoder struct { + parser *parser + knownFields bool +} + +// NewDecoder returns a new decoder that reads from r. +// +// The decoder introduces its own buffering and may read +// data from r beyond the YAML values requested. +func NewDecoder(r io.Reader) *Decoder { + return &Decoder{ + parser: newParserFromReader(r), + } +} + +// KnownFields ensures that the keys in decoded mappings to +// exist as fields in the struct being decoded into. +func (dec *Decoder) KnownFields(enable bool) { + dec.knownFields = enable +} + +// Decode reads the next YAML-encoded value from its input +// and stores it in the value pointed to by v. +// +// See the documentation for Unmarshal for details about the +// conversion of YAML into a Go value. +func (dec *Decoder) Decode(v interface{}) (err error) { + d := newDecoder() + d.knownFields = dec.knownFields + defer handleErr(&err) + node := dec.parser.parse() + if node == nil { + return io.EOF + } + out := reflect.ValueOf(v) + if out.Kind() == reflect.Ptr && !out.IsNil() { + out = out.Elem() + } + d.unmarshal(node, out) + if len(d.terrors) > 0 { + return &TypeError{d.terrors} + } + return nil +} + +// Decode decodes the node and stores its data into the value pointed to by v. +// +// See the documentation for Unmarshal for details about the +// conversion of YAML into a Go value. +func (n *Node) Decode(v interface{}) (err error) { + d := newDecoder() + defer handleErr(&err) + out := reflect.ValueOf(v) + if out.Kind() == reflect.Ptr && !out.IsNil() { + out = out.Elem() + } + d.unmarshal(n, out) + if len(d.terrors) > 0 { + return &TypeError{d.terrors} + } + return nil +} + +func unmarshal(in []byte, out interface{}, strict bool) (err error) { + defer handleErr(&err) + d := newDecoder() + p := newParser(in) + defer p.destroy() + node := p.parse() + if node != nil { + v := reflect.ValueOf(out) + if v.Kind() == reflect.Ptr && !v.IsNil() { + v = v.Elem() + } + d.unmarshal(node, v) + } + if len(d.terrors) > 0 { + return &TypeError{d.terrors} + } + return nil +} + +// Marshal serializes the value provided into a YAML document. The structure +// of the generated document will reflect the structure of the value itself. +// Maps and pointers (to struct, string, int, etc) are accepted as the in value. +// +// Struct fields are only marshalled if they are exported (have an upper case +// first letter), and are marshalled using the field name lowercased as the +// default key. Custom keys may be defined via the "yaml" name in the field +// tag: the content preceding the first comma is used as the key, and the +// following comma-separated options are used to tweak the marshalling process. +// Conflicting names result in a runtime error. +// +// The field tag format accepted is: +// +// `(...) yaml:"[][,[,]]" (...)` +// +// The following flags are currently supported: +// +// omitempty Only include the field if it's not set to the zero +// value for the type or to empty slices or maps. +// Zero valued structs will be omitted if all their public +// fields are zero, unless they implement an IsZero +// method (see the IsZeroer interface type), in which +// case the field will be excluded if IsZero returns true. +// +// flow Marshal using a flow style (useful for structs, +// sequences and maps). +// +// inline Inline the field, which must be a struct or a map, +// causing all of its fields or keys to be processed as if +// they were part of the outer struct. For maps, keys must +// not conflict with the yaml keys of other struct fields. +// +// In addition, if the key is "-", the field is ignored. +// +// For example: +// +// type T struct { +// F int `yaml:"a,omitempty"` +// B int +// } +// yaml.Marshal(&T{B: 2}) // Returns "b: 2\n" +// yaml.Marshal(&T{F: 1}} // Returns "a: 1\nb: 0\n" +// +func Marshal(in interface{}) (out []byte, err error) { + defer handleErr(&err) + e := newEncoder() + defer e.destroy() + e.marshalDoc("", reflect.ValueOf(in)) + e.finish() + out = e.out + return +} + +// An Encoder writes YAML values to an output stream. +type Encoder struct { + encoder *encoder +} + +// NewEncoder returns a new encoder that writes to w. +// The Encoder should be closed after use to flush all data +// to w. +func NewEncoder(w io.Writer) *Encoder { + return &Encoder{ + encoder: newEncoderWithWriter(w), + } +} + +// Encode writes the YAML encoding of v to the stream. +// If multiple items are encoded to the stream, the +// second and subsequent document will be preceded +// with a "---" document separator, but the first will not. +// +// See the documentation for Marshal for details about the conversion of Go +// values to YAML. +func (e *Encoder) Encode(v interface{}) (err error) { + defer handleErr(&err) + e.encoder.marshalDoc("", reflect.ValueOf(v)) + return nil +} + +// Encode encodes value v and stores its representation in n. +// +// See the documentation for Marshal for details about the +// conversion of Go values into YAML. +func (n *Node) Encode(v interface{}) (err error) { + defer handleErr(&err) + e := newEncoder() + defer e.destroy() + e.marshalDoc("", reflect.ValueOf(v)) + e.finish() + p := newParser(e.out) + p.textless = true + defer p.destroy() + doc := p.parse() + *n = *doc.Content[0] + return nil +} + +// SetIndent changes the used indentation used when encoding. +func (e *Encoder) SetIndent(spaces int) { + if spaces < 0 { + panic("yaml: cannot indent to a negative number of spaces") + } + e.encoder.indent = spaces +} + +// Close closes the encoder by writing any remaining data. +// It does not write a stream terminating string "...". +func (e *Encoder) Close() (err error) { + defer handleErr(&err) + e.encoder.finish() + return nil +} + +func handleErr(err *error) { + if v := recover(); v != nil { + if e, ok := v.(yamlError); ok { + *err = e.err + } else { + panic(v) + } + } +} + +type yamlError struct { + err error +} + +func fail(err error) { + panic(yamlError{err}) +} + +func failf(format string, args ...interface{}) { + panic(yamlError{fmt.Errorf("yaml: "+format, args...)}) +} + +// A TypeError is returned by Unmarshal when one or more fields in +// the YAML document cannot be properly decoded into the requested +// types. When this error is returned, the value is still +// unmarshaled partially. +type TypeError struct { + Errors []string +} + +func (e *TypeError) Error() string { + return fmt.Sprintf("yaml: unmarshal errors:\n %s", strings.Join(e.Errors, "\n ")) +} + +type Kind uint32 + +const ( + DocumentNode Kind = 1 << iota + SequenceNode + MappingNode + ScalarNode + AliasNode +) + +type Style uint32 + +const ( + TaggedStyle Style = 1 << iota + DoubleQuotedStyle + SingleQuotedStyle + LiteralStyle + FoldedStyle + FlowStyle +) + +// Node represents an element in the YAML document hierarchy. While documents +// are typically encoded and decoded into higher level types, such as structs +// and maps, Node is an intermediate representation that allows detailed +// control over the content being decoded or encoded. +// +// It's worth noting that although Node offers access into details such as +// line numbers, colums, and comments, the content when re-encoded will not +// have its original textual representation preserved. An effort is made to +// render the data plesantly, and to preserve comments near the data they +// describe, though. +// +// Values that make use of the Node type interact with the yaml package in the +// same way any other type would do, by encoding and decoding yaml data +// directly or indirectly into them. +// +// For example: +// +// var person struct { +// Name string +// Address yaml.Node +// } +// err := yaml.Unmarshal(data, &person) +// +// Or by itself: +// +// var person Node +// err := yaml.Unmarshal(data, &person) +// +type Node struct { + // Kind defines whether the node is a document, a mapping, a sequence, + // a scalar value, or an alias to another node. The specific data type of + // scalar nodes may be obtained via the ShortTag and LongTag methods. + Kind Kind + + // Style allows customizing the apperance of the node in the tree. + Style Style + + // Tag holds the YAML tag defining the data type for the value. + // When decoding, this field will always be set to the resolved tag, + // even when it wasn't explicitly provided in the YAML content. + // When encoding, if this field is unset the value type will be + // implied from the node properties, and if it is set, it will only + // be serialized into the representation if TaggedStyle is used or + // the implicit tag diverges from the provided one. + Tag string + + // Value holds the unescaped and unquoted represenation of the value. + Value string + + // Anchor holds the anchor name for this node, which allows aliases to point to it. + Anchor string + + // Alias holds the node that this alias points to. Only valid when Kind is AliasNode. + Alias *Node + + // Content holds contained nodes for documents, mappings, and sequences. + Content []*Node + + // HeadComment holds any comments in the lines preceding the node and + // not separated by an empty line. + HeadComment string + + // LineComment holds any comments at the end of the line where the node is in. + LineComment string + + // FootComment holds any comments following the node and before empty lines. + FootComment string + + // Line and Column hold the node position in the decoded YAML text. + // These fields are not respected when encoding the node. + Line int + Column int +} + +// IsZero returns whether the node has all of its fields unset. +func (n *Node) IsZero() bool { + return n.Kind == 0 && n.Style == 0 && n.Tag == "" && n.Value == "" && n.Anchor == "" && n.Alias == nil && n.Content == nil && + n.HeadComment == "" && n.LineComment == "" && n.FootComment == "" && n.Line == 0 && n.Column == 0 +} + + +// LongTag returns the long form of the tag that indicates the data type for +// the node. If the Tag field isn't explicitly defined, one will be computed +// based on the node properties. +func (n *Node) LongTag() string { + return longTag(n.ShortTag()) +} + +// ShortTag returns the short form of the YAML tag that indicates data type for +// the node. If the Tag field isn't explicitly defined, one will be computed +// based on the node properties. +func (n *Node) ShortTag() string { + if n.indicatedString() { + return strTag + } + if n.Tag == "" || n.Tag == "!" { + switch n.Kind { + case MappingNode: + return mapTag + case SequenceNode: + return seqTag + case AliasNode: + if n.Alias != nil { + return n.Alias.ShortTag() + } + case ScalarNode: + tag, _ := resolve("", n.Value) + return tag + } + return "" + } + return shortTag(n.Tag) +} + +func (n *Node) indicatedString() bool { + return n.Kind == ScalarNode && + (shortTag(n.Tag) == strTag || + (n.Tag == "" || n.Tag == "!") && n.Style&(SingleQuotedStyle|DoubleQuotedStyle|LiteralStyle|FoldedStyle) != 0) +} + +// SetString is a convenience function that sets the node to a string value +// and defines its style in a pleasant way depending on its content. +func (n *Node) SetString(s string) { + n.Kind = ScalarNode + if utf8.ValidString(s) { + n.Value = s + n.Tag = strTag + } else { + n.Value = encodeBase64(s) + n.Tag = binaryTag + } + if strings.Contains(n.Value, "\n") { + n.Style = LiteralStyle + } +} + +// -------------------------------------------------------------------------- +// Maintain a mapping of keys to structure field indexes + +// The code in this section was copied from mgo/bson. + +// structInfo holds details for the serialization of fields of +// a given struct. +type structInfo struct { + FieldsMap map[string]fieldInfo + FieldsList []fieldInfo + + // InlineMap is the number of the field in the struct that + // contains an ,inline map, or -1 if there's none. + InlineMap int + + // InlineUnmarshalers holds indexes to inlined fields that + // contain unmarshaler values. + InlineUnmarshalers [][]int +} + +type fieldInfo struct { + Key string + Num int + OmitEmpty bool + Flow bool + // Id holds the unique field identifier, so we can cheaply + // check for field duplicates without maintaining an extra map. + Id int + + // Inline holds the field index if the field is part of an inlined struct. + Inline []int +} + +var structMap = make(map[reflect.Type]*structInfo) +var fieldMapMutex sync.RWMutex +var unmarshalerType reflect.Type + +func init() { + var v Unmarshaler + unmarshalerType = reflect.ValueOf(&v).Elem().Type() +} + +func getStructInfo(st reflect.Type) (*structInfo, error) { + fieldMapMutex.RLock() + sinfo, found := structMap[st] + fieldMapMutex.RUnlock() + if found { + return sinfo, nil + } + + n := st.NumField() + fieldsMap := make(map[string]fieldInfo) + fieldsList := make([]fieldInfo, 0, n) + inlineMap := -1 + inlineUnmarshalers := [][]int(nil) + for i := 0; i != n; i++ { + field := st.Field(i) + if field.PkgPath != "" && !field.Anonymous { + continue // Private field + } + + info := fieldInfo{Num: i} + + tag := field.Tag.Get("yaml") + if tag == "" && strings.Index(string(field.Tag), ":") < 0 { + tag = string(field.Tag) + } + if tag == "-" { + continue + } + + inline := false + fields := strings.Split(tag, ",") + if len(fields) > 1 { + for _, flag := range fields[1:] { + switch flag { + case "omitempty": + info.OmitEmpty = true + case "flow": + info.Flow = true + case "inline": + inline = true + default: + return nil, errors.New(fmt.Sprintf("unsupported flag %q in tag %q of type %s", flag, tag, st)) + } + } + tag = fields[0] + } + + if inline { + switch field.Type.Kind() { + case reflect.Map: + if inlineMap >= 0 { + return nil, errors.New("multiple ,inline maps in struct " + st.String()) + } + if field.Type.Key() != reflect.TypeOf("") { + return nil, errors.New("option ,inline needs a map with string keys in struct " + st.String()) + } + inlineMap = info.Num + case reflect.Struct, reflect.Ptr: + ftype := field.Type + for ftype.Kind() == reflect.Ptr { + ftype = ftype.Elem() + } + if ftype.Kind() != reflect.Struct { + return nil, errors.New("option ,inline may only be used on a struct or map field") + } + if reflect.PtrTo(ftype).Implements(unmarshalerType) { + inlineUnmarshalers = append(inlineUnmarshalers, []int{i}) + } else { + sinfo, err := getStructInfo(ftype) + if err != nil { + return nil, err + } + for _, index := range sinfo.InlineUnmarshalers { + inlineUnmarshalers = append(inlineUnmarshalers, append([]int{i}, index...)) + } + for _, finfo := range sinfo.FieldsList { + if _, found := fieldsMap[finfo.Key]; found { + msg := "duplicated key '" + finfo.Key + "' in struct " + st.String() + return nil, errors.New(msg) + } + if finfo.Inline == nil { + finfo.Inline = []int{i, finfo.Num} + } else { + finfo.Inline = append([]int{i}, finfo.Inline...) + } + finfo.Id = len(fieldsList) + fieldsMap[finfo.Key] = finfo + fieldsList = append(fieldsList, finfo) + } + } + default: + return nil, errors.New("option ,inline may only be used on a struct or map field") + } + continue + } + + if tag != "" { + info.Key = tag + } else { + info.Key = strings.ToLower(field.Name) + } + + if _, found = fieldsMap[info.Key]; found { + msg := "duplicated key '" + info.Key + "' in struct " + st.String() + return nil, errors.New(msg) + } + + info.Id = len(fieldsList) + fieldsList = append(fieldsList, info) + fieldsMap[info.Key] = info + } + + sinfo = &structInfo{ + FieldsMap: fieldsMap, + FieldsList: fieldsList, + InlineMap: inlineMap, + InlineUnmarshalers: inlineUnmarshalers, + } + + fieldMapMutex.Lock() + structMap[st] = sinfo + fieldMapMutex.Unlock() + return sinfo, nil +} + +// IsZeroer is used to check whether an object is zero to +// determine whether it should be omitted when marshaling +// with the omitempty flag. One notable implementation +// is time.Time. +type IsZeroer interface { + IsZero() bool +} + +func isZero(v reflect.Value) bool { + kind := v.Kind() + if z, ok := v.Interface().(IsZeroer); ok { + if (kind == reflect.Ptr || kind == reflect.Interface) && v.IsNil() { + return true + } + return z.IsZero() + } + switch kind { + case reflect.String: + return len(v.String()) == 0 + case reflect.Interface, reflect.Ptr: + return v.IsNil() + case reflect.Slice: + return v.Len() == 0 + case reflect.Map: + return v.Len() == 0 + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + return v.Int() == 0 + case reflect.Float32, reflect.Float64: + return v.Float() == 0 + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: + return v.Uint() == 0 + case reflect.Bool: + return !v.Bool() + case reflect.Struct: + vt := v.Type() + for i := v.NumField() - 1; i >= 0; i-- { + if vt.Field(i).PkgPath != "" { + continue // Private field + } + if !isZero(v.Field(i)) { + return false + } + } + return true + } + return false +} diff --git a/vendor/gopkg.in/yaml.v3/yamlh.go b/vendor/gopkg.in/yaml.v3/yamlh.go new file mode 100644 index 00000000..7c6d0077 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/yamlh.go @@ -0,0 +1,807 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +import ( + "fmt" + "io" +) + +// The version directive data. +type yaml_version_directive_t struct { + major int8 // The major version number. + minor int8 // The minor version number. +} + +// The tag directive data. +type yaml_tag_directive_t struct { + handle []byte // The tag handle. + prefix []byte // The tag prefix. +} + +type yaml_encoding_t int + +// The stream encoding. +const ( + // Let the parser choose the encoding. + yaml_ANY_ENCODING yaml_encoding_t = iota + + yaml_UTF8_ENCODING // The default UTF-8 encoding. + yaml_UTF16LE_ENCODING // The UTF-16-LE encoding with BOM. + yaml_UTF16BE_ENCODING // The UTF-16-BE encoding with BOM. +) + +type yaml_break_t int + +// Line break types. +const ( + // Let the parser choose the break type. + yaml_ANY_BREAK yaml_break_t = iota + + yaml_CR_BREAK // Use CR for line breaks (Mac style). + yaml_LN_BREAK // Use LN for line breaks (Unix style). + yaml_CRLN_BREAK // Use CR LN for line breaks (DOS style). +) + +type yaml_error_type_t int + +// Many bad things could happen with the parser and emitter. +const ( + // No error is produced. + yaml_NO_ERROR yaml_error_type_t = iota + + yaml_MEMORY_ERROR // Cannot allocate or reallocate a block of memory. + yaml_READER_ERROR // Cannot read or decode the input stream. + yaml_SCANNER_ERROR // Cannot scan the input stream. + yaml_PARSER_ERROR // Cannot parse the input stream. + yaml_COMPOSER_ERROR // Cannot compose a YAML document. + yaml_WRITER_ERROR // Cannot write to the output stream. + yaml_EMITTER_ERROR // Cannot emit a YAML stream. +) + +// The pointer position. +type yaml_mark_t struct { + index int // The position index. + line int // The position line. + column int // The position column. +} + +// Node Styles + +type yaml_style_t int8 + +type yaml_scalar_style_t yaml_style_t + +// Scalar styles. +const ( + // Let the emitter choose the style. + yaml_ANY_SCALAR_STYLE yaml_scalar_style_t = 0 + + yaml_PLAIN_SCALAR_STYLE yaml_scalar_style_t = 1 << iota // The plain scalar style. + yaml_SINGLE_QUOTED_SCALAR_STYLE // The single-quoted scalar style. + yaml_DOUBLE_QUOTED_SCALAR_STYLE // The double-quoted scalar style. + yaml_LITERAL_SCALAR_STYLE // The literal scalar style. + yaml_FOLDED_SCALAR_STYLE // The folded scalar style. +) + +type yaml_sequence_style_t yaml_style_t + +// Sequence styles. +const ( + // Let the emitter choose the style. + yaml_ANY_SEQUENCE_STYLE yaml_sequence_style_t = iota + + yaml_BLOCK_SEQUENCE_STYLE // The block sequence style. + yaml_FLOW_SEQUENCE_STYLE // The flow sequence style. +) + +type yaml_mapping_style_t yaml_style_t + +// Mapping styles. +const ( + // Let the emitter choose the style. + yaml_ANY_MAPPING_STYLE yaml_mapping_style_t = iota + + yaml_BLOCK_MAPPING_STYLE // The block mapping style. + yaml_FLOW_MAPPING_STYLE // The flow mapping style. +) + +// Tokens + +type yaml_token_type_t int + +// Token types. +const ( + // An empty token. + yaml_NO_TOKEN yaml_token_type_t = iota + + yaml_STREAM_START_TOKEN // A STREAM-START token. + yaml_STREAM_END_TOKEN // A STREAM-END token. + + yaml_VERSION_DIRECTIVE_TOKEN // A VERSION-DIRECTIVE token. + yaml_TAG_DIRECTIVE_TOKEN // A TAG-DIRECTIVE token. + yaml_DOCUMENT_START_TOKEN // A DOCUMENT-START token. + yaml_DOCUMENT_END_TOKEN // A DOCUMENT-END token. + + yaml_BLOCK_SEQUENCE_START_TOKEN // A BLOCK-SEQUENCE-START token. + yaml_BLOCK_MAPPING_START_TOKEN // A BLOCK-SEQUENCE-END token. + yaml_BLOCK_END_TOKEN // A BLOCK-END token. + + yaml_FLOW_SEQUENCE_START_TOKEN // A FLOW-SEQUENCE-START token. + yaml_FLOW_SEQUENCE_END_TOKEN // A FLOW-SEQUENCE-END token. + yaml_FLOW_MAPPING_START_TOKEN // A FLOW-MAPPING-START token. + yaml_FLOW_MAPPING_END_TOKEN // A FLOW-MAPPING-END token. + + yaml_BLOCK_ENTRY_TOKEN // A BLOCK-ENTRY token. + yaml_FLOW_ENTRY_TOKEN // A FLOW-ENTRY token. + yaml_KEY_TOKEN // A KEY token. + yaml_VALUE_TOKEN // A VALUE token. + + yaml_ALIAS_TOKEN // An ALIAS token. + yaml_ANCHOR_TOKEN // An ANCHOR token. + yaml_TAG_TOKEN // A TAG token. + yaml_SCALAR_TOKEN // A SCALAR token. +) + +func (tt yaml_token_type_t) String() string { + switch tt { + case yaml_NO_TOKEN: + return "yaml_NO_TOKEN" + case yaml_STREAM_START_TOKEN: + return "yaml_STREAM_START_TOKEN" + case yaml_STREAM_END_TOKEN: + return "yaml_STREAM_END_TOKEN" + case yaml_VERSION_DIRECTIVE_TOKEN: + return "yaml_VERSION_DIRECTIVE_TOKEN" + case yaml_TAG_DIRECTIVE_TOKEN: + return "yaml_TAG_DIRECTIVE_TOKEN" + case yaml_DOCUMENT_START_TOKEN: + return "yaml_DOCUMENT_START_TOKEN" + case yaml_DOCUMENT_END_TOKEN: + return "yaml_DOCUMENT_END_TOKEN" + case yaml_BLOCK_SEQUENCE_START_TOKEN: + return "yaml_BLOCK_SEQUENCE_START_TOKEN" + case yaml_BLOCK_MAPPING_START_TOKEN: + return "yaml_BLOCK_MAPPING_START_TOKEN" + case yaml_BLOCK_END_TOKEN: + return "yaml_BLOCK_END_TOKEN" + case yaml_FLOW_SEQUENCE_START_TOKEN: + return "yaml_FLOW_SEQUENCE_START_TOKEN" + case yaml_FLOW_SEQUENCE_END_TOKEN: + return "yaml_FLOW_SEQUENCE_END_TOKEN" + case yaml_FLOW_MAPPING_START_TOKEN: + return "yaml_FLOW_MAPPING_START_TOKEN" + case yaml_FLOW_MAPPING_END_TOKEN: + return "yaml_FLOW_MAPPING_END_TOKEN" + case yaml_BLOCK_ENTRY_TOKEN: + return "yaml_BLOCK_ENTRY_TOKEN" + case yaml_FLOW_ENTRY_TOKEN: + return "yaml_FLOW_ENTRY_TOKEN" + case yaml_KEY_TOKEN: + return "yaml_KEY_TOKEN" + case yaml_VALUE_TOKEN: + return "yaml_VALUE_TOKEN" + case yaml_ALIAS_TOKEN: + return "yaml_ALIAS_TOKEN" + case yaml_ANCHOR_TOKEN: + return "yaml_ANCHOR_TOKEN" + case yaml_TAG_TOKEN: + return "yaml_TAG_TOKEN" + case yaml_SCALAR_TOKEN: + return "yaml_SCALAR_TOKEN" + } + return "" +} + +// The token structure. +type yaml_token_t struct { + // The token type. + typ yaml_token_type_t + + // The start/end of the token. + start_mark, end_mark yaml_mark_t + + // The stream encoding (for yaml_STREAM_START_TOKEN). + encoding yaml_encoding_t + + // The alias/anchor/scalar value or tag/tag directive handle + // (for yaml_ALIAS_TOKEN, yaml_ANCHOR_TOKEN, yaml_SCALAR_TOKEN, yaml_TAG_TOKEN, yaml_TAG_DIRECTIVE_TOKEN). + value []byte + + // The tag suffix (for yaml_TAG_TOKEN). + suffix []byte + + // The tag directive prefix (for yaml_TAG_DIRECTIVE_TOKEN). + prefix []byte + + // The scalar style (for yaml_SCALAR_TOKEN). + style yaml_scalar_style_t + + // The version directive major/minor (for yaml_VERSION_DIRECTIVE_TOKEN). + major, minor int8 +} + +// Events + +type yaml_event_type_t int8 + +// Event types. +const ( + // An empty event. + yaml_NO_EVENT yaml_event_type_t = iota + + yaml_STREAM_START_EVENT // A STREAM-START event. + yaml_STREAM_END_EVENT // A STREAM-END event. + yaml_DOCUMENT_START_EVENT // A DOCUMENT-START event. + yaml_DOCUMENT_END_EVENT // A DOCUMENT-END event. + yaml_ALIAS_EVENT // An ALIAS event. + yaml_SCALAR_EVENT // A SCALAR event. + yaml_SEQUENCE_START_EVENT // A SEQUENCE-START event. + yaml_SEQUENCE_END_EVENT // A SEQUENCE-END event. + yaml_MAPPING_START_EVENT // A MAPPING-START event. + yaml_MAPPING_END_EVENT // A MAPPING-END event. + yaml_TAIL_COMMENT_EVENT +) + +var eventStrings = []string{ + yaml_NO_EVENT: "none", + yaml_STREAM_START_EVENT: "stream start", + yaml_STREAM_END_EVENT: "stream end", + yaml_DOCUMENT_START_EVENT: "document start", + yaml_DOCUMENT_END_EVENT: "document end", + yaml_ALIAS_EVENT: "alias", + yaml_SCALAR_EVENT: "scalar", + yaml_SEQUENCE_START_EVENT: "sequence start", + yaml_SEQUENCE_END_EVENT: "sequence end", + yaml_MAPPING_START_EVENT: "mapping start", + yaml_MAPPING_END_EVENT: "mapping end", + yaml_TAIL_COMMENT_EVENT: "tail comment", +} + +func (e yaml_event_type_t) String() string { + if e < 0 || int(e) >= len(eventStrings) { + return fmt.Sprintf("unknown event %d", e) + } + return eventStrings[e] +} + +// The event structure. +type yaml_event_t struct { + + // The event type. + typ yaml_event_type_t + + // The start and end of the event. + start_mark, end_mark yaml_mark_t + + // The document encoding (for yaml_STREAM_START_EVENT). + encoding yaml_encoding_t + + // The version directive (for yaml_DOCUMENT_START_EVENT). + version_directive *yaml_version_directive_t + + // The list of tag directives (for yaml_DOCUMENT_START_EVENT). + tag_directives []yaml_tag_directive_t + + // The comments + head_comment []byte + line_comment []byte + foot_comment []byte + tail_comment []byte + + // The anchor (for yaml_SCALAR_EVENT, yaml_SEQUENCE_START_EVENT, yaml_MAPPING_START_EVENT, yaml_ALIAS_EVENT). + anchor []byte + + // The tag (for yaml_SCALAR_EVENT, yaml_SEQUENCE_START_EVENT, yaml_MAPPING_START_EVENT). + tag []byte + + // The scalar value (for yaml_SCALAR_EVENT). + value []byte + + // Is the document start/end indicator implicit, or the tag optional? + // (for yaml_DOCUMENT_START_EVENT, yaml_DOCUMENT_END_EVENT, yaml_SEQUENCE_START_EVENT, yaml_MAPPING_START_EVENT, yaml_SCALAR_EVENT). + implicit bool + + // Is the tag optional for any non-plain style? (for yaml_SCALAR_EVENT). + quoted_implicit bool + + // The style (for yaml_SCALAR_EVENT, yaml_SEQUENCE_START_EVENT, yaml_MAPPING_START_EVENT). + style yaml_style_t +} + +func (e *yaml_event_t) scalar_style() yaml_scalar_style_t { return yaml_scalar_style_t(e.style) } +func (e *yaml_event_t) sequence_style() yaml_sequence_style_t { return yaml_sequence_style_t(e.style) } +func (e *yaml_event_t) mapping_style() yaml_mapping_style_t { return yaml_mapping_style_t(e.style) } + +// Nodes + +const ( + yaml_NULL_TAG = "tag:yaml.org,2002:null" // The tag !!null with the only possible value: null. + yaml_BOOL_TAG = "tag:yaml.org,2002:bool" // The tag !!bool with the values: true and false. + yaml_STR_TAG = "tag:yaml.org,2002:str" // The tag !!str for string values. + yaml_INT_TAG = "tag:yaml.org,2002:int" // The tag !!int for integer values. + yaml_FLOAT_TAG = "tag:yaml.org,2002:float" // The tag !!float for float values. + yaml_TIMESTAMP_TAG = "tag:yaml.org,2002:timestamp" // The tag !!timestamp for date and time values. + + yaml_SEQ_TAG = "tag:yaml.org,2002:seq" // The tag !!seq is used to denote sequences. + yaml_MAP_TAG = "tag:yaml.org,2002:map" // The tag !!map is used to denote mapping. + + // Not in original libyaml. + yaml_BINARY_TAG = "tag:yaml.org,2002:binary" + yaml_MERGE_TAG = "tag:yaml.org,2002:merge" + + yaml_DEFAULT_SCALAR_TAG = yaml_STR_TAG // The default scalar tag is !!str. + yaml_DEFAULT_SEQUENCE_TAG = yaml_SEQ_TAG // The default sequence tag is !!seq. + yaml_DEFAULT_MAPPING_TAG = yaml_MAP_TAG // The default mapping tag is !!map. +) + +type yaml_node_type_t int + +// Node types. +const ( + // An empty node. + yaml_NO_NODE yaml_node_type_t = iota + + yaml_SCALAR_NODE // A scalar node. + yaml_SEQUENCE_NODE // A sequence node. + yaml_MAPPING_NODE // A mapping node. +) + +// An element of a sequence node. +type yaml_node_item_t int + +// An element of a mapping node. +type yaml_node_pair_t struct { + key int // The key of the element. + value int // The value of the element. +} + +// The node structure. +type yaml_node_t struct { + typ yaml_node_type_t // The node type. + tag []byte // The node tag. + + // The node data. + + // The scalar parameters (for yaml_SCALAR_NODE). + scalar struct { + value []byte // The scalar value. + length int // The length of the scalar value. + style yaml_scalar_style_t // The scalar style. + } + + // The sequence parameters (for YAML_SEQUENCE_NODE). + sequence struct { + items_data []yaml_node_item_t // The stack of sequence items. + style yaml_sequence_style_t // The sequence style. + } + + // The mapping parameters (for yaml_MAPPING_NODE). + mapping struct { + pairs_data []yaml_node_pair_t // The stack of mapping pairs (key, value). + pairs_start *yaml_node_pair_t // The beginning of the stack. + pairs_end *yaml_node_pair_t // The end of the stack. + pairs_top *yaml_node_pair_t // The top of the stack. + style yaml_mapping_style_t // The mapping style. + } + + start_mark yaml_mark_t // The beginning of the node. + end_mark yaml_mark_t // The end of the node. + +} + +// The document structure. +type yaml_document_t struct { + + // The document nodes. + nodes []yaml_node_t + + // The version directive. + version_directive *yaml_version_directive_t + + // The list of tag directives. + tag_directives_data []yaml_tag_directive_t + tag_directives_start int // The beginning of the tag directives list. + tag_directives_end int // The end of the tag directives list. + + start_implicit int // Is the document start indicator implicit? + end_implicit int // Is the document end indicator implicit? + + // The start/end of the document. + start_mark, end_mark yaml_mark_t +} + +// The prototype of a read handler. +// +// The read handler is called when the parser needs to read more bytes from the +// source. The handler should write not more than size bytes to the buffer. +// The number of written bytes should be set to the size_read variable. +// +// [in,out] data A pointer to an application data specified by +// yaml_parser_set_input(). +// [out] buffer The buffer to write the data from the source. +// [in] size The size of the buffer. +// [out] size_read The actual number of bytes read from the source. +// +// On success, the handler should return 1. If the handler failed, +// the returned value should be 0. On EOF, the handler should set the +// size_read to 0 and return 1. +type yaml_read_handler_t func(parser *yaml_parser_t, buffer []byte) (n int, err error) + +// This structure holds information about a potential simple key. +type yaml_simple_key_t struct { + possible bool // Is a simple key possible? + required bool // Is a simple key required? + token_number int // The number of the token. + mark yaml_mark_t // The position mark. +} + +// The states of the parser. +type yaml_parser_state_t int + +const ( + yaml_PARSE_STREAM_START_STATE yaml_parser_state_t = iota + + yaml_PARSE_IMPLICIT_DOCUMENT_START_STATE // Expect the beginning of an implicit document. + yaml_PARSE_DOCUMENT_START_STATE // Expect DOCUMENT-START. + yaml_PARSE_DOCUMENT_CONTENT_STATE // Expect the content of a document. + yaml_PARSE_DOCUMENT_END_STATE // Expect DOCUMENT-END. + yaml_PARSE_BLOCK_NODE_STATE // Expect a block node. + yaml_PARSE_BLOCK_NODE_OR_INDENTLESS_SEQUENCE_STATE // Expect a block node or indentless sequence. + yaml_PARSE_FLOW_NODE_STATE // Expect a flow node. + yaml_PARSE_BLOCK_SEQUENCE_FIRST_ENTRY_STATE // Expect the first entry of a block sequence. + yaml_PARSE_BLOCK_SEQUENCE_ENTRY_STATE // Expect an entry of a block sequence. + yaml_PARSE_INDENTLESS_SEQUENCE_ENTRY_STATE // Expect an entry of an indentless sequence. + yaml_PARSE_BLOCK_MAPPING_FIRST_KEY_STATE // Expect the first key of a block mapping. + yaml_PARSE_BLOCK_MAPPING_KEY_STATE // Expect a block mapping key. + yaml_PARSE_BLOCK_MAPPING_VALUE_STATE // Expect a block mapping value. + yaml_PARSE_FLOW_SEQUENCE_FIRST_ENTRY_STATE // Expect the first entry of a flow sequence. + yaml_PARSE_FLOW_SEQUENCE_ENTRY_STATE // Expect an entry of a flow sequence. + yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_KEY_STATE // Expect a key of an ordered mapping. + yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_VALUE_STATE // Expect a value of an ordered mapping. + yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_END_STATE // Expect the and of an ordered mapping entry. + yaml_PARSE_FLOW_MAPPING_FIRST_KEY_STATE // Expect the first key of a flow mapping. + yaml_PARSE_FLOW_MAPPING_KEY_STATE // Expect a key of a flow mapping. + yaml_PARSE_FLOW_MAPPING_VALUE_STATE // Expect a value of a flow mapping. + yaml_PARSE_FLOW_MAPPING_EMPTY_VALUE_STATE // Expect an empty value of a flow mapping. + yaml_PARSE_END_STATE // Expect nothing. +) + +func (ps yaml_parser_state_t) String() string { + switch ps { + case yaml_PARSE_STREAM_START_STATE: + return "yaml_PARSE_STREAM_START_STATE" + case yaml_PARSE_IMPLICIT_DOCUMENT_START_STATE: + return "yaml_PARSE_IMPLICIT_DOCUMENT_START_STATE" + case yaml_PARSE_DOCUMENT_START_STATE: + return "yaml_PARSE_DOCUMENT_START_STATE" + case yaml_PARSE_DOCUMENT_CONTENT_STATE: + return "yaml_PARSE_DOCUMENT_CONTENT_STATE" + case yaml_PARSE_DOCUMENT_END_STATE: + return "yaml_PARSE_DOCUMENT_END_STATE" + case yaml_PARSE_BLOCK_NODE_STATE: + return "yaml_PARSE_BLOCK_NODE_STATE" + case yaml_PARSE_BLOCK_NODE_OR_INDENTLESS_SEQUENCE_STATE: + return "yaml_PARSE_BLOCK_NODE_OR_INDENTLESS_SEQUENCE_STATE" + case yaml_PARSE_FLOW_NODE_STATE: + return "yaml_PARSE_FLOW_NODE_STATE" + case yaml_PARSE_BLOCK_SEQUENCE_FIRST_ENTRY_STATE: + return "yaml_PARSE_BLOCK_SEQUENCE_FIRST_ENTRY_STATE" + case yaml_PARSE_BLOCK_SEQUENCE_ENTRY_STATE: + return "yaml_PARSE_BLOCK_SEQUENCE_ENTRY_STATE" + case yaml_PARSE_INDENTLESS_SEQUENCE_ENTRY_STATE: + return "yaml_PARSE_INDENTLESS_SEQUENCE_ENTRY_STATE" + case yaml_PARSE_BLOCK_MAPPING_FIRST_KEY_STATE: + return "yaml_PARSE_BLOCK_MAPPING_FIRST_KEY_STATE" + case yaml_PARSE_BLOCK_MAPPING_KEY_STATE: + return "yaml_PARSE_BLOCK_MAPPING_KEY_STATE" + case yaml_PARSE_BLOCK_MAPPING_VALUE_STATE: + return "yaml_PARSE_BLOCK_MAPPING_VALUE_STATE" + case yaml_PARSE_FLOW_SEQUENCE_FIRST_ENTRY_STATE: + return "yaml_PARSE_FLOW_SEQUENCE_FIRST_ENTRY_STATE" + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_STATE: + return "yaml_PARSE_FLOW_SEQUENCE_ENTRY_STATE" + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_KEY_STATE: + return "yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_KEY_STATE" + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_VALUE_STATE: + return "yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_VALUE_STATE" + case yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_END_STATE: + return "yaml_PARSE_FLOW_SEQUENCE_ENTRY_MAPPING_END_STATE" + case yaml_PARSE_FLOW_MAPPING_FIRST_KEY_STATE: + return "yaml_PARSE_FLOW_MAPPING_FIRST_KEY_STATE" + case yaml_PARSE_FLOW_MAPPING_KEY_STATE: + return "yaml_PARSE_FLOW_MAPPING_KEY_STATE" + case yaml_PARSE_FLOW_MAPPING_VALUE_STATE: + return "yaml_PARSE_FLOW_MAPPING_VALUE_STATE" + case yaml_PARSE_FLOW_MAPPING_EMPTY_VALUE_STATE: + return "yaml_PARSE_FLOW_MAPPING_EMPTY_VALUE_STATE" + case yaml_PARSE_END_STATE: + return "yaml_PARSE_END_STATE" + } + return "" +} + +// This structure holds aliases data. +type yaml_alias_data_t struct { + anchor []byte // The anchor. + index int // The node id. + mark yaml_mark_t // The anchor mark. +} + +// The parser structure. +// +// All members are internal. Manage the structure using the +// yaml_parser_ family of functions. +type yaml_parser_t struct { + + // Error handling + + error yaml_error_type_t // Error type. + + problem string // Error description. + + // The byte about which the problem occurred. + problem_offset int + problem_value int + problem_mark yaml_mark_t + + // The error context. + context string + context_mark yaml_mark_t + + // Reader stuff + + read_handler yaml_read_handler_t // Read handler. + + input_reader io.Reader // File input data. + input []byte // String input data. + input_pos int + + eof bool // EOF flag + + buffer []byte // The working buffer. + buffer_pos int // The current position of the buffer. + + unread int // The number of unread characters in the buffer. + + newlines int // The number of line breaks since last non-break/non-blank character + + raw_buffer []byte // The raw buffer. + raw_buffer_pos int // The current position of the buffer. + + encoding yaml_encoding_t // The input encoding. + + offset int // The offset of the current position (in bytes). + mark yaml_mark_t // The mark of the current position. + + // Comments + + head_comment []byte // The current head comments + line_comment []byte // The current line comments + foot_comment []byte // The current foot comments + tail_comment []byte // Foot comment that happens at the end of a block. + stem_comment []byte // Comment in item preceding a nested structure (list inside list item, etc) + + comments []yaml_comment_t // The folded comments for all parsed tokens + comments_head int + + // Scanner stuff + + stream_start_produced bool // Have we started to scan the input stream? + stream_end_produced bool // Have we reached the end of the input stream? + + flow_level int // The number of unclosed '[' and '{' indicators. + + tokens []yaml_token_t // The tokens queue. + tokens_head int // The head of the tokens queue. + tokens_parsed int // The number of tokens fetched from the queue. + token_available bool // Does the tokens queue contain a token ready for dequeueing. + + indent int // The current indentation level. + indents []int // The indentation levels stack. + + simple_key_allowed bool // May a simple key occur at the current position? + simple_keys []yaml_simple_key_t // The stack of simple keys. + simple_keys_by_tok map[int]int // possible simple_key indexes indexed by token_number + + // Parser stuff + + state yaml_parser_state_t // The current parser state. + states []yaml_parser_state_t // The parser states stack. + marks []yaml_mark_t // The stack of marks. + tag_directives []yaml_tag_directive_t // The list of TAG directives. + + // Dumper stuff + + aliases []yaml_alias_data_t // The alias data. + + document *yaml_document_t // The currently parsed document. +} + +type yaml_comment_t struct { + + scan_mark yaml_mark_t // Position where scanning for comments started + token_mark yaml_mark_t // Position after which tokens will be associated with this comment + start_mark yaml_mark_t // Position of '#' comment mark + end_mark yaml_mark_t // Position where comment terminated + + head []byte + line []byte + foot []byte +} + +// Emitter Definitions + +// The prototype of a write handler. +// +// The write handler is called when the emitter needs to flush the accumulated +// characters to the output. The handler should write @a size bytes of the +// @a buffer to the output. +// +// @param[in,out] data A pointer to an application data specified by +// yaml_emitter_set_output(). +// @param[in] buffer The buffer with bytes to be written. +// @param[in] size The size of the buffer. +// +// @returns On success, the handler should return @c 1. If the handler failed, +// the returned value should be @c 0. +// +type yaml_write_handler_t func(emitter *yaml_emitter_t, buffer []byte) error + +type yaml_emitter_state_t int + +// The emitter states. +const ( + // Expect STREAM-START. + yaml_EMIT_STREAM_START_STATE yaml_emitter_state_t = iota + + yaml_EMIT_FIRST_DOCUMENT_START_STATE // Expect the first DOCUMENT-START or STREAM-END. + yaml_EMIT_DOCUMENT_START_STATE // Expect DOCUMENT-START or STREAM-END. + yaml_EMIT_DOCUMENT_CONTENT_STATE // Expect the content of a document. + yaml_EMIT_DOCUMENT_END_STATE // Expect DOCUMENT-END. + yaml_EMIT_FLOW_SEQUENCE_FIRST_ITEM_STATE // Expect the first item of a flow sequence. + yaml_EMIT_FLOW_SEQUENCE_TRAIL_ITEM_STATE // Expect the next item of a flow sequence, with the comma already written out + yaml_EMIT_FLOW_SEQUENCE_ITEM_STATE // Expect an item of a flow sequence. + yaml_EMIT_FLOW_MAPPING_FIRST_KEY_STATE // Expect the first key of a flow mapping. + yaml_EMIT_FLOW_MAPPING_TRAIL_KEY_STATE // Expect the next key of a flow mapping, with the comma already written out + yaml_EMIT_FLOW_MAPPING_KEY_STATE // Expect a key of a flow mapping. + yaml_EMIT_FLOW_MAPPING_SIMPLE_VALUE_STATE // Expect a value for a simple key of a flow mapping. + yaml_EMIT_FLOW_MAPPING_VALUE_STATE // Expect a value of a flow mapping. + yaml_EMIT_BLOCK_SEQUENCE_FIRST_ITEM_STATE // Expect the first item of a block sequence. + yaml_EMIT_BLOCK_SEQUENCE_ITEM_STATE // Expect an item of a block sequence. + yaml_EMIT_BLOCK_MAPPING_FIRST_KEY_STATE // Expect the first key of a block mapping. + yaml_EMIT_BLOCK_MAPPING_KEY_STATE // Expect the key of a block mapping. + yaml_EMIT_BLOCK_MAPPING_SIMPLE_VALUE_STATE // Expect a value for a simple key of a block mapping. + yaml_EMIT_BLOCK_MAPPING_VALUE_STATE // Expect a value of a block mapping. + yaml_EMIT_END_STATE // Expect nothing. +) + +// The emitter structure. +// +// All members are internal. Manage the structure using the @c yaml_emitter_ +// family of functions. +type yaml_emitter_t struct { + + // Error handling + + error yaml_error_type_t // Error type. + problem string // Error description. + + // Writer stuff + + write_handler yaml_write_handler_t // Write handler. + + output_buffer *[]byte // String output data. + output_writer io.Writer // File output data. + + buffer []byte // The working buffer. + buffer_pos int // The current position of the buffer. + + raw_buffer []byte // The raw buffer. + raw_buffer_pos int // The current position of the buffer. + + encoding yaml_encoding_t // The stream encoding. + + // Emitter stuff + + canonical bool // If the output is in the canonical style? + best_indent int // The number of indentation spaces. + best_width int // The preferred width of the output lines. + unicode bool // Allow unescaped non-ASCII characters? + line_break yaml_break_t // The preferred line break. + + state yaml_emitter_state_t // The current emitter state. + states []yaml_emitter_state_t // The stack of states. + + events []yaml_event_t // The event queue. + events_head int // The head of the event queue. + + indents []int // The stack of indentation levels. + + tag_directives []yaml_tag_directive_t // The list of tag directives. + + indent int // The current indentation level. + + flow_level int // The current flow level. + + root_context bool // Is it the document root context? + sequence_context bool // Is it a sequence context? + mapping_context bool // Is it a mapping context? + simple_key_context bool // Is it a simple mapping key context? + + line int // The current line. + column int // The current column. + whitespace bool // If the last character was a whitespace? + indention bool // If the last character was an indentation character (' ', '-', '?', ':')? + open_ended bool // If an explicit document end is required? + + space_above bool // Is there's an empty line above? + foot_indent int // The indent used to write the foot comment above, or -1 if none. + + // Anchor analysis. + anchor_data struct { + anchor []byte // The anchor value. + alias bool // Is it an alias? + } + + // Tag analysis. + tag_data struct { + handle []byte // The tag handle. + suffix []byte // The tag suffix. + } + + // Scalar analysis. + scalar_data struct { + value []byte // The scalar value. + multiline bool // Does the scalar contain line breaks? + flow_plain_allowed bool // Can the scalar be expessed in the flow plain style? + block_plain_allowed bool // Can the scalar be expressed in the block plain style? + single_quoted_allowed bool // Can the scalar be expressed in the single quoted style? + block_allowed bool // Can the scalar be expressed in the literal or folded styles? + style yaml_scalar_style_t // The output style. + } + + // Comments + head_comment []byte + line_comment []byte + foot_comment []byte + tail_comment []byte + + key_line_comment []byte + + // Dumper stuff + + opened bool // If the stream was already opened? + closed bool // If the stream was already closed? + + // The information associated with the document nodes. + anchors *struct { + references int // The number of references. + anchor int // The anchor id. + serialized bool // If the node has been emitted? + } + + last_anchor_id int // The last assigned anchor id. + + document *yaml_document_t // The currently emitted document. +} diff --git a/vendor/gopkg.in/yaml.v3/yamlprivateh.go b/vendor/gopkg.in/yaml.v3/yamlprivateh.go new file mode 100644 index 00000000..e88f9c54 --- /dev/null +++ b/vendor/gopkg.in/yaml.v3/yamlprivateh.go @@ -0,0 +1,198 @@ +// +// Copyright (c) 2011-2019 Canonical Ltd +// Copyright (c) 2006-2010 Kirill Simonov +// +// Permission is hereby granted, free of charge, to any person obtaining a copy of +// this software and associated documentation files (the "Software"), to deal in +// the Software without restriction, including without limitation the rights to +// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +// of the Software, and to permit persons to whom the Software is furnished to do +// so, subject to the following conditions: +// +// The above copyright notice and this permission notice shall be included in all +// copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +// SOFTWARE. + +package yaml + +const ( + // The size of the input raw buffer. + input_raw_buffer_size = 512 + + // The size of the input buffer. + // It should be possible to decode the whole raw buffer. + input_buffer_size = input_raw_buffer_size * 3 + + // The size of the output buffer. + output_buffer_size = 128 + + // The size of the output raw buffer. + // It should be possible to encode the whole output buffer. + output_raw_buffer_size = (output_buffer_size*2 + 2) + + // The size of other stacks and queues. + initial_stack_size = 16 + initial_queue_size = 16 + initial_string_size = 16 +) + +// Check if the character at the specified position is an alphabetical +// character, a digit, '_', or '-'. +func is_alpha(b []byte, i int) bool { + return b[i] >= '0' && b[i] <= '9' || b[i] >= 'A' && b[i] <= 'Z' || b[i] >= 'a' && b[i] <= 'z' || b[i] == '_' || b[i] == '-' +} + +// Check if the character at the specified position is a digit. +func is_digit(b []byte, i int) bool { + return b[i] >= '0' && b[i] <= '9' +} + +// Get the value of a digit. +func as_digit(b []byte, i int) int { + return int(b[i]) - '0' +} + +// Check if the character at the specified position is a hex-digit. +func is_hex(b []byte, i int) bool { + return b[i] >= '0' && b[i] <= '9' || b[i] >= 'A' && b[i] <= 'F' || b[i] >= 'a' && b[i] <= 'f' +} + +// Get the value of a hex-digit. +func as_hex(b []byte, i int) int { + bi := b[i] + if bi >= 'A' && bi <= 'F' { + return int(bi) - 'A' + 10 + } + if bi >= 'a' && bi <= 'f' { + return int(bi) - 'a' + 10 + } + return int(bi) - '0' +} + +// Check if the character is ASCII. +func is_ascii(b []byte, i int) bool { + return b[i] <= 0x7F +} + +// Check if the character at the start of the buffer can be printed unescaped. +func is_printable(b []byte, i int) bool { + return ((b[i] == 0x0A) || // . == #x0A + (b[i] >= 0x20 && b[i] <= 0x7E) || // #x20 <= . <= #x7E + (b[i] == 0xC2 && b[i+1] >= 0xA0) || // #0xA0 <= . <= #xD7FF + (b[i] > 0xC2 && b[i] < 0xED) || + (b[i] == 0xED && b[i+1] < 0xA0) || + (b[i] == 0xEE) || + (b[i] == 0xEF && // #xE000 <= . <= #xFFFD + !(b[i+1] == 0xBB && b[i+2] == 0xBF) && // && . != #xFEFF + !(b[i+1] == 0xBF && (b[i+2] == 0xBE || b[i+2] == 0xBF)))) +} + +// Check if the character at the specified position is NUL. +func is_z(b []byte, i int) bool { + return b[i] == 0x00 +} + +// Check if the beginning of the buffer is a BOM. +func is_bom(b []byte, i int) bool { + return b[0] == 0xEF && b[1] == 0xBB && b[2] == 0xBF +} + +// Check if the character at the specified position is space. +func is_space(b []byte, i int) bool { + return b[i] == ' ' +} + +// Check if the character at the specified position is tab. +func is_tab(b []byte, i int) bool { + return b[i] == '\t' +} + +// Check if the character at the specified position is blank (space or tab). +func is_blank(b []byte, i int) bool { + //return is_space(b, i) || is_tab(b, i) + return b[i] == ' ' || b[i] == '\t' +} + +// Check if the character at the specified position is a line break. +func is_break(b []byte, i int) bool { + return (b[i] == '\r' || // CR (#xD) + b[i] == '\n' || // LF (#xA) + b[i] == 0xC2 && b[i+1] == 0x85 || // NEL (#x85) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA8 || // LS (#x2028) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA9) // PS (#x2029) +} + +func is_crlf(b []byte, i int) bool { + return b[i] == '\r' && b[i+1] == '\n' +} + +// Check if the character is a line break or NUL. +func is_breakz(b []byte, i int) bool { + //return is_break(b, i) || is_z(b, i) + return ( + // is_break: + b[i] == '\r' || // CR (#xD) + b[i] == '\n' || // LF (#xA) + b[i] == 0xC2 && b[i+1] == 0x85 || // NEL (#x85) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA8 || // LS (#x2028) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA9 || // PS (#x2029) + // is_z: + b[i] == 0) +} + +// Check if the character is a line break, space, or NUL. +func is_spacez(b []byte, i int) bool { + //return is_space(b, i) || is_breakz(b, i) + return ( + // is_space: + b[i] == ' ' || + // is_breakz: + b[i] == '\r' || // CR (#xD) + b[i] == '\n' || // LF (#xA) + b[i] == 0xC2 && b[i+1] == 0x85 || // NEL (#x85) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA8 || // LS (#x2028) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA9 || // PS (#x2029) + b[i] == 0) +} + +// Check if the character is a line break, space, tab, or NUL. +func is_blankz(b []byte, i int) bool { + //return is_blank(b, i) || is_breakz(b, i) + return ( + // is_blank: + b[i] == ' ' || b[i] == '\t' || + // is_breakz: + b[i] == '\r' || // CR (#xD) + b[i] == '\n' || // LF (#xA) + b[i] == 0xC2 && b[i+1] == 0x85 || // NEL (#x85) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA8 || // LS (#x2028) + b[i] == 0xE2 && b[i+1] == 0x80 && b[i+2] == 0xA9 || // PS (#x2029) + b[i] == 0) +} + +// Determine the width of the character. +func width(b byte) int { + // Don't replace these by a switch without first + // confirming that it is being inlined. + if b&0x80 == 0x00 { + return 1 + } + if b&0xE0 == 0xC0 { + return 2 + } + if b&0xF0 == 0xE0 { + return 3 + } + if b&0xF8 == 0xF0 { + return 4 + } + return 0 + +} diff --git a/vendor/k8s.io/api/admissionregistration/v1/generated.pb.go b/vendor/k8s.io/api/admissionregistration/v1/generated.pb.go index 1acb6345..adc47be7 100644 --- a/vendor/k8s.io/api/admissionregistration/v1/generated.pb.go +++ b/vendor/k8s.io/api/admissionregistration/v1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *MutatingWebhook) Reset() { *m = MutatingWebhook{} } func (*MutatingWebhook) ProtoMessage() {} @@ -3360,6 +3360,7 @@ func (m *WebhookClientConfig) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -3391,10 +3392,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -3415,55 +3414,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/admissionregistration/v1/generated.proto b/vendor/k8s.io/api/admissionregistration/v1/generated.proto index f102f3a7..7f9772e7 100644 --- a/vendor/k8s.io/api/admissionregistration/v1/generated.proto +++ b/vendor/k8s.io/api/admissionregistration/v1/generated.proto @@ -245,8 +245,8 @@ message Rule { // RuleWithOperations is a tuple of Operations and Resources. It is recommended to make // sure that all the tuple expansions are valid. message RuleWithOperations { - // Operations is the operations the admission hook cares about - CREATE, UPDATE, or * - // for all operations. + // Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * + // for all of those operations and any future admission operations that are added. // If '*' is present, the length of the slice must be one. // Required. repeated string operations = 1; diff --git a/vendor/k8s.io/api/admissionregistration/v1/types.go b/vendor/k8s.io/api/admissionregistration/v1/types.go index 114a4c68..74b87828 100644 --- a/vendor/k8s.io/api/admissionregistration/v1/types.go +++ b/vendor/k8s.io/api/admissionregistration/v1/types.go @@ -462,8 +462,8 @@ const ( // RuleWithOperations is a tuple of Operations and Resources. It is recommended to make // sure that all the tuple expansions are valid. type RuleWithOperations struct { - // Operations is the operations the admission hook cares about - CREATE, UPDATE, or * - // for all operations. + // Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * + // for all of those operations and any future admission operations that are added. // If '*' is present, the length of the slice must be one. // Required. Operations []OperationType `json:"operations,omitempty" protobuf:"bytes,1,rep,name=operations,casttype=OperationType"` diff --git a/vendor/k8s.io/api/admissionregistration/v1/types_swagger_doc_generated.go b/vendor/k8s.io/api/admissionregistration/v1/types_swagger_doc_generated.go index 2fde0ce3..5ec59304 100644 --- a/vendor/k8s.io/api/admissionregistration/v1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/admissionregistration/v1/types_swagger_doc_generated.go @@ -80,7 +80,7 @@ func (Rule) SwaggerDoc() map[string]string { var map_RuleWithOperations = map[string]string{ "": "RuleWithOperations is a tuple of Operations and Resources. It is recommended to make sure that all the tuple expansions are valid.", - "operations": "Operations is the operations the admission hook cares about - CREATE, UPDATE, or * for all operations. If '*' is present, the length of the slice must be one. Required.", + "operations": "Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * for all of those operations and any future admission operations that are added. If '*' is present, the length of the slice must be one. Required.", } func (RuleWithOperations) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/admissionregistration/v1beta1/doc.go b/vendor/k8s.io/api/admissionregistration/v1beta1/doc.go index 0a40726f..0095cb25 100644 --- a/vendor/k8s.io/api/admissionregistration/v1beta1/doc.go +++ b/vendor/k8s.io/api/admissionregistration/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=admissionregistration.k8s.io // Package v1beta1 is the v1beta1 version of the API. diff --git a/vendor/k8s.io/api/admissionregistration/v1beta1/generated.pb.go b/vendor/k8s.io/api/admissionregistration/v1beta1/generated.pb.go index d84d8b63..c98aa747 100644 --- a/vendor/k8s.io/api/admissionregistration/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/admissionregistration/v1beta1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *MutatingWebhook) Reset() { *m = MutatingWebhook{} } func (*MutatingWebhook) ProtoMessage() {} @@ -3361,6 +3361,7 @@ func (m *WebhookClientConfig) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -3392,10 +3393,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -3416,55 +3415,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/admissionregistration/v1beta1/generated.proto b/vendor/k8s.io/api/admissionregistration/v1beta1/generated.proto index 086cbcc7..70ffa921 100644 --- a/vendor/k8s.io/api/admissionregistration/v1beta1/generated.proto +++ b/vendor/k8s.io/api/admissionregistration/v1beta1/generated.proto @@ -249,8 +249,8 @@ message Rule { // RuleWithOperations is a tuple of Operations and Resources. It is recommended to make // sure that all the tuple expansions are valid. message RuleWithOperations { - // Operations is the operations the admission hook cares about - CREATE, UPDATE, or * - // for all operations. + // Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * + // for all of those operations and any future admission operations that are added. // If '*' is present, the length of the slice must be one. // Required. repeated string operations = 1; diff --git a/vendor/k8s.io/api/admissionregistration/v1beta1/types.go b/vendor/k8s.io/api/admissionregistration/v1beta1/types.go index 37a993e3..2297b7e1 100644 --- a/vendor/k8s.io/api/admissionregistration/v1beta1/types.go +++ b/vendor/k8s.io/api/admissionregistration/v1beta1/types.go @@ -113,6 +113,10 @@ const ( // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.9 +// +k8s:prerelease-lifecycle-gen:deprecated=1.16 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=admissionregistration.k8s.io,v1,ValidatingWebhookConfiguration // ValidatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and object without changing it. // Deprecated in v1.16, planned for removal in v1.19. Use admissionregistration.k8s.io/v1 ValidatingWebhookConfiguration instead. @@ -129,6 +133,10 @@ type ValidatingWebhookConfiguration struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.9 +// +k8s:prerelease-lifecycle-gen:deprecated=1.16 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=admissionregistration.k8s.io,v1,ValidatingWebhookConfigurationList // ValidatingWebhookConfigurationList is a list of ValidatingWebhookConfiguration. type ValidatingWebhookConfigurationList struct { @@ -144,6 +152,10 @@ type ValidatingWebhookConfigurationList struct { // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.9 +// +k8s:prerelease-lifecycle-gen:deprecated=1.16 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=admissionregistration.k8s.io,v1,MutatingWebhookConfiguration // MutatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and may change the object. // Deprecated in v1.16, planned for removal in v1.19. Use admissionregistration.k8s.io/v1 MutatingWebhookConfiguration instead. @@ -160,6 +172,10 @@ type MutatingWebhookConfiguration struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.9 +// +k8s:prerelease-lifecycle-gen:deprecated=1.16 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=admissionregistration.k8s.io,v1,MutatingWebhookConfigurationList // MutatingWebhookConfigurationList is a list of MutatingWebhookConfiguration. type MutatingWebhookConfigurationList struct { @@ -470,8 +486,8 @@ const ( // RuleWithOperations is a tuple of Operations and Resources. It is recommended to make // sure that all the tuple expansions are valid. type RuleWithOperations struct { - // Operations is the operations the admission hook cares about - CREATE, UPDATE, or * - // for all operations. + // Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * + // for all of those operations and any future admission operations that are added. // If '*' is present, the length of the slice must be one. // Required. Operations []OperationType `json:"operations,omitempty" protobuf:"bytes,1,rep,name=operations,casttype=OperationType"` diff --git a/vendor/k8s.io/api/admissionregistration/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/admissionregistration/v1beta1/types_swagger_doc_generated.go index d9fb5af8..f682172b 100644 --- a/vendor/k8s.io/api/admissionregistration/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/admissionregistration/v1beta1/types_swagger_doc_generated.go @@ -80,7 +80,7 @@ func (Rule) SwaggerDoc() map[string]string { var map_RuleWithOperations = map[string]string{ "": "RuleWithOperations is a tuple of Operations and Resources. It is recommended to make sure that all the tuple expansions are valid.", - "operations": "Operations is the operations the admission hook cares about - CREATE, UPDATE, or * for all operations. If '*' is present, the length of the slice must be one. Required.", + "operations": "Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * for all of those operations and any future admission operations that are added. If '*' is present, the length of the slice must be one. Required.", } func (RuleWithOperations) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/admissionregistration/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/admissionregistration/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..bfd93a05 --- /dev/null +++ b/vendor/k8s.io/api/admissionregistration/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,121 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *MutatingWebhookConfiguration) APILifecycleIntroduced() (major, minor int) { + return 1, 9 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *MutatingWebhookConfiguration) APILifecycleDeprecated() (major, minor int) { + return 1, 16 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *MutatingWebhookConfiguration) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "admissionregistration.k8s.io", Version: "v1", Kind: "MutatingWebhookConfiguration"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *MutatingWebhookConfiguration) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *MutatingWebhookConfigurationList) APILifecycleIntroduced() (major, minor int) { + return 1, 9 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *MutatingWebhookConfigurationList) APILifecycleDeprecated() (major, minor int) { + return 1, 16 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *MutatingWebhookConfigurationList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "admissionregistration.k8s.io", Version: "v1", Kind: "MutatingWebhookConfigurationList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *MutatingWebhookConfigurationList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ValidatingWebhookConfiguration) APILifecycleIntroduced() (major, minor int) { + return 1, 9 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ValidatingWebhookConfiguration) APILifecycleDeprecated() (major, minor int) { + return 1, 16 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ValidatingWebhookConfiguration) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "admissionregistration.k8s.io", Version: "v1", Kind: "ValidatingWebhookConfiguration"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ValidatingWebhookConfiguration) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ValidatingWebhookConfigurationList) APILifecycleIntroduced() (major, minor int) { + return 1, 9 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ValidatingWebhookConfigurationList) APILifecycleDeprecated() (major, minor int) { + return 1, 16 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ValidatingWebhookConfigurationList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "admissionregistration.k8s.io", Version: "v1", Kind: "ValidatingWebhookConfigurationList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ValidatingWebhookConfigurationList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/apps/v1/generated.pb.go b/vendor/k8s.io/api/apps/v1/generated.pb.go index 425144d8..6ef25f50 100644 --- a/vendor/k8s.io/api/apps/v1/generated.pb.go +++ b/vendor/k8s.io/api/apps/v1/generated.pb.go @@ -46,7 +46,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *ControllerRevision) Reset() { *m = ControllerRevision{} } func (*ControllerRevision) ProtoMessage() {} @@ -8155,6 +8155,7 @@ func (m *StatefulSetUpdateStrategy) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -8186,10 +8187,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -8210,55 +8209,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/apps/v1beta1/doc.go b/vendor/k8s.io/api/apps/v1beta1/doc.go index 9072bab6..38a35855 100644 --- a/vendor/k8s.io/api/apps/v1beta1/doc.go +++ b/vendor/k8s.io/api/apps/v1beta1/doc.go @@ -17,5 +17,6 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v1beta1 // import "k8s.io/api/apps/v1beta1" diff --git a/vendor/k8s.io/api/apps/v1beta1/generated.pb.go b/vendor/k8s.io/api/apps/v1beta1/generated.pb.go index 921e055c..f81b5590 100644 --- a/vendor/k8s.io/api/apps/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/apps/v1beta1/generated.pb.go @@ -47,7 +47,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *ControllerRevision) Reset() { *m = ControllerRevision{} } func (*ControllerRevision) ProtoMessage() {} @@ -6163,6 +6163,7 @@ func (m *StatefulSetUpdateStrategy) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -6194,10 +6195,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -6218,55 +6217,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/apps/v1beta1/types.go b/vendor/k8s.io/api/apps/v1beta1/types.go index b77fcf7a..9f822fae 100644 --- a/vendor/k8s.io/api/apps/v1beta1/types.go +++ b/vendor/k8s.io/api/apps/v1beta1/types.go @@ -56,6 +56,10 @@ type ScaleStatus struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=autoscaling,v1,Scale // Scale represents a scaling request for a resource. type Scale struct { @@ -75,6 +79,10 @@ type Scale struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.5 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,StatefulSet // DEPRECATED - This group version of StatefulSet is deprecated by apps/v1beta2/StatefulSet. See the release notes for // more information. @@ -274,6 +282,10 @@ type StatefulSetCondition struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.5 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,StatefulSetList // StatefulSetList is a collection of StatefulSets. type StatefulSetList struct { @@ -285,6 +297,10 @@ type StatefulSetList struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,Deployment // DEPRECATED - This group version of Deployment is deprecated by apps/v1beta2/Deployment. See the release notes for // more information. @@ -355,6 +371,10 @@ type DeploymentSpec struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DeploymentRollback // DEPRECATED. // DeploymentRollback stores the information required to rollback a deployment. @@ -512,6 +532,10 @@ type DeploymentCondition struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DeploymentList // DeploymentList is a list of Deployments. type DeploymentList struct { @@ -526,6 +550,10 @@ type DeploymentList struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.7 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ControllerRevision // DEPRECATED - This group version of ControllerRevision is deprecated by apps/v1beta2/ControllerRevision. See the // release notes for more information. @@ -553,6 +581,10 @@ type ControllerRevision struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.7 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ControllerRevisionList // ControllerRevisionList is a resource containing a list of ControllerRevision objects. type ControllerRevisionList struct { diff --git a/vendor/k8s.io/api/apps/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/apps/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..f3850fc9 --- /dev/null +++ b/vendor/k8s.io/api/apps/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,217 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ControllerRevision) APILifecycleIntroduced() (major, minor int) { + return 1, 7 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ControllerRevision) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ControllerRevision) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ControllerRevision"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ControllerRevision) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ControllerRevisionList) APILifecycleIntroduced() (major, minor int) { + return 1, 7 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ControllerRevisionList) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ControllerRevisionList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ControllerRevisionList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ControllerRevisionList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Deployment) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Deployment) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Deployment) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "Deployment"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Deployment) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DeploymentList) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DeploymentList) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DeploymentList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DeploymentList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DeploymentList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DeploymentRollback) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DeploymentRollback) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DeploymentRollback) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DeploymentRollback"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DeploymentRollback) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Scale) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Scale) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Scale) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "autoscaling", Version: "v1", Kind: "Scale"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Scale) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *StatefulSet) APILifecycleIntroduced() (major, minor int) { + return 1, 5 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *StatefulSet) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *StatefulSet) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "StatefulSet"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *StatefulSet) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *StatefulSetList) APILifecycleIntroduced() (major, minor int) { + return 1, 5 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *StatefulSetList) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *StatefulSetList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "StatefulSetList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *StatefulSetList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} diff --git a/vendor/k8s.io/api/apps/v1beta2/doc.go b/vendor/k8s.io/api/apps/v1beta2/doc.go index 9f499869..ac91fddf 100644 --- a/vendor/k8s.io/api/apps/v1beta2/doc.go +++ b/vendor/k8s.io/api/apps/v1beta2/doc.go @@ -17,5 +17,6 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v1beta2 // import "k8s.io/api/apps/v1beta2" diff --git a/vendor/k8s.io/api/apps/v1beta2/generated.pb.go b/vendor/k8s.io/api/apps/v1beta2/generated.pb.go index 624bb942..8a9f2005 100644 --- a/vendor/k8s.io/api/apps/v1beta2/generated.pb.go +++ b/vendor/k8s.io/api/apps/v1beta2/generated.pb.go @@ -47,7 +47,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *ControllerRevision) Reset() { *m = ControllerRevision{} } func (*ControllerRevision) ProtoMessage() {} @@ -8931,6 +8931,7 @@ func (m *StatefulSetUpdateStrategy) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -8962,10 +8963,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -8986,55 +8985,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/apps/v1beta2/types.go b/vendor/k8s.io/api/apps/v1beta2/types.go index d358455f..fc542ac1 100644 --- a/vendor/k8s.io/api/apps/v1beta2/types.go +++ b/vendor/k8s.io/api/apps/v1beta2/types.go @@ -58,6 +58,10 @@ type ScaleStatus struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=autoscaling,v1,Scale // Scale represents a scaling request for a resource. type Scale struct { @@ -79,6 +83,10 @@ type Scale struct { // +genclient:method=GetScale,verb=get,subresource=scale,result=Scale // +genclient:method=UpdateScale,verb=update,subresource=scale,input=Scale,result=Scale // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,StatefulSet // DEPRECATED - This group version of StatefulSet is deprecated by apps/v1/StatefulSet. See the release notes for // more information. @@ -282,6 +290,10 @@ type StatefulSetCondition struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,StatefulSetList // StatefulSetList is a collection of StatefulSets. type StatefulSetList struct { @@ -293,6 +305,10 @@ type StatefulSetList struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,Deployment // DEPRECATED - This group version of Deployment is deprecated by apps/v1/Deployment. See the release notes for // more information. @@ -492,6 +508,10 @@ type DeploymentCondition struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DeploymentList // DeploymentList is a list of Deployments. type DeploymentList struct { @@ -659,6 +679,10 @@ type DaemonSetCondition struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DaemonSet // DEPRECATED - This group version of DaemonSet is deprecated by apps/v1/DaemonSet. See the release notes for // more information. @@ -692,6 +716,10 @@ const ( ) // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DaemonSetList // DaemonSetList is a collection of daemon sets. type DaemonSetList struct { @@ -707,6 +735,10 @@ type DaemonSetList struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ReplicaSet // DEPRECATED - This group version of ReplicaSet is deprecated by apps/v1/ReplicaSet. See the release notes for // more information. @@ -735,6 +767,10 @@ type ReplicaSet struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ReplicaSetList // ReplicaSetList is a collection of ReplicaSets. type ReplicaSetList struct { @@ -835,6 +871,10 @@ type ReplicaSetCondition struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ControllerRevision // DEPRECATED - This group version of ControllerRevision is deprecated by apps/v1/ControllerRevision. See the // release notes for more information. @@ -862,6 +902,10 @@ type ControllerRevision struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ControllerRevisionList // ControllerRevisionList is a resource containing a list of ControllerRevision objects. type ControllerRevisionList struct { diff --git a/vendor/k8s.io/api/apps/v1beta2/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/apps/v1beta2/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..3368a189 --- /dev/null +++ b/vendor/k8s.io/api/apps/v1beta2/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,289 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta2 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ControllerRevision) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ControllerRevision) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ControllerRevision) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ControllerRevision"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ControllerRevision) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ControllerRevisionList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ControllerRevisionList) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ControllerRevisionList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ControllerRevisionList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ControllerRevisionList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DaemonSet) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DaemonSet) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DaemonSet) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DaemonSet"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DaemonSet) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DaemonSetList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DaemonSetList) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DaemonSetList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DaemonSetList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DaemonSetList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Deployment) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Deployment) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Deployment) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "Deployment"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Deployment) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DeploymentList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DeploymentList) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DeploymentList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DeploymentList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DeploymentList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ReplicaSet) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ReplicaSet) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ReplicaSet) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ReplicaSet"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ReplicaSet) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ReplicaSetList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ReplicaSetList) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ReplicaSetList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ReplicaSetList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ReplicaSetList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Scale) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Scale) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Scale) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "autoscaling", Version: "v1", Kind: "Scale"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Scale) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *StatefulSet) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *StatefulSet) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *StatefulSet) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "StatefulSet"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *StatefulSet) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *StatefulSetList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *StatefulSetList) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *StatefulSetList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "StatefulSetList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *StatefulSetList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} diff --git a/vendor/k8s.io/api/auditregistration/v1alpha1/generated.pb.go b/vendor/k8s.io/api/auditregistration/v1alpha1/generated.pb.go deleted file mode 100644 index 003cc30b..00000000 --- a/vendor/k8s.io/api/auditregistration/v1alpha1/generated.pb.go +++ /dev/null @@ -1,2056 +0,0 @@ -/* -Copyright The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -// Code generated by protoc-gen-gogo. DO NOT EDIT. -// source: k8s.io/kubernetes/vendor/k8s.io/api/auditregistration/v1alpha1/generated.proto - -package v1alpha1 - -import ( - fmt "fmt" - - io "io" - - proto "github.com/gogo/protobuf/proto" - - math "math" - math_bits "math/bits" - reflect "reflect" - strings "strings" -) - -// Reference imports to suppress errors if they are not otherwise used. -var _ = proto.Marshal -var _ = fmt.Errorf -var _ = math.Inf - -// This is a compile-time assertion to ensure that this generated file -// is compatible with the proto package it is being compiled against. -// A compilation error at this line likely means your copy of the -// proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package - -func (m *AuditSink) Reset() { *m = AuditSink{} } -func (*AuditSink) ProtoMessage() {} -func (*AuditSink) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{0} -} -func (m *AuditSink) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *AuditSink) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *AuditSink) XXX_Merge(src proto.Message) { - xxx_messageInfo_AuditSink.Merge(m, src) -} -func (m *AuditSink) XXX_Size() int { - return m.Size() -} -func (m *AuditSink) XXX_DiscardUnknown() { - xxx_messageInfo_AuditSink.DiscardUnknown(m) -} - -var xxx_messageInfo_AuditSink proto.InternalMessageInfo - -func (m *AuditSinkList) Reset() { *m = AuditSinkList{} } -func (*AuditSinkList) ProtoMessage() {} -func (*AuditSinkList) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{1} -} -func (m *AuditSinkList) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *AuditSinkList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *AuditSinkList) XXX_Merge(src proto.Message) { - xxx_messageInfo_AuditSinkList.Merge(m, src) -} -func (m *AuditSinkList) XXX_Size() int { - return m.Size() -} -func (m *AuditSinkList) XXX_DiscardUnknown() { - xxx_messageInfo_AuditSinkList.DiscardUnknown(m) -} - -var xxx_messageInfo_AuditSinkList proto.InternalMessageInfo - -func (m *AuditSinkSpec) Reset() { *m = AuditSinkSpec{} } -func (*AuditSinkSpec) ProtoMessage() {} -func (*AuditSinkSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{2} -} -func (m *AuditSinkSpec) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *AuditSinkSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *AuditSinkSpec) XXX_Merge(src proto.Message) { - xxx_messageInfo_AuditSinkSpec.Merge(m, src) -} -func (m *AuditSinkSpec) XXX_Size() int { - return m.Size() -} -func (m *AuditSinkSpec) XXX_DiscardUnknown() { - xxx_messageInfo_AuditSinkSpec.DiscardUnknown(m) -} - -var xxx_messageInfo_AuditSinkSpec proto.InternalMessageInfo - -func (m *Policy) Reset() { *m = Policy{} } -func (*Policy) ProtoMessage() {} -func (*Policy) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{3} -} -func (m *Policy) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *Policy) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *Policy) XXX_Merge(src proto.Message) { - xxx_messageInfo_Policy.Merge(m, src) -} -func (m *Policy) XXX_Size() int { - return m.Size() -} -func (m *Policy) XXX_DiscardUnknown() { - xxx_messageInfo_Policy.DiscardUnknown(m) -} - -var xxx_messageInfo_Policy proto.InternalMessageInfo - -func (m *ServiceReference) Reset() { *m = ServiceReference{} } -func (*ServiceReference) ProtoMessage() {} -func (*ServiceReference) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{4} -} -func (m *ServiceReference) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *ServiceReference) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *ServiceReference) XXX_Merge(src proto.Message) { - xxx_messageInfo_ServiceReference.Merge(m, src) -} -func (m *ServiceReference) XXX_Size() int { - return m.Size() -} -func (m *ServiceReference) XXX_DiscardUnknown() { - xxx_messageInfo_ServiceReference.DiscardUnknown(m) -} - -var xxx_messageInfo_ServiceReference proto.InternalMessageInfo - -func (m *Webhook) Reset() { *m = Webhook{} } -func (*Webhook) ProtoMessage() {} -func (*Webhook) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{5} -} -func (m *Webhook) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *Webhook) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *Webhook) XXX_Merge(src proto.Message) { - xxx_messageInfo_Webhook.Merge(m, src) -} -func (m *Webhook) XXX_Size() int { - return m.Size() -} -func (m *Webhook) XXX_DiscardUnknown() { - xxx_messageInfo_Webhook.DiscardUnknown(m) -} - -var xxx_messageInfo_Webhook proto.InternalMessageInfo - -func (m *WebhookClientConfig) Reset() { *m = WebhookClientConfig{} } -func (*WebhookClientConfig) ProtoMessage() {} -func (*WebhookClientConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{6} -} -func (m *WebhookClientConfig) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *WebhookClientConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *WebhookClientConfig) XXX_Merge(src proto.Message) { - xxx_messageInfo_WebhookClientConfig.Merge(m, src) -} -func (m *WebhookClientConfig) XXX_Size() int { - return m.Size() -} -func (m *WebhookClientConfig) XXX_DiscardUnknown() { - xxx_messageInfo_WebhookClientConfig.DiscardUnknown(m) -} - -var xxx_messageInfo_WebhookClientConfig proto.InternalMessageInfo - -func (m *WebhookThrottleConfig) Reset() { *m = WebhookThrottleConfig{} } -func (*WebhookThrottleConfig) ProtoMessage() {} -func (*WebhookThrottleConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_642d3597c6afa8ba, []int{7} -} -func (m *WebhookThrottleConfig) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *WebhookThrottleConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *WebhookThrottleConfig) XXX_Merge(src proto.Message) { - xxx_messageInfo_WebhookThrottleConfig.Merge(m, src) -} -func (m *WebhookThrottleConfig) XXX_Size() int { - return m.Size() -} -func (m *WebhookThrottleConfig) XXX_DiscardUnknown() { - xxx_messageInfo_WebhookThrottleConfig.DiscardUnknown(m) -} - -var xxx_messageInfo_WebhookThrottleConfig proto.InternalMessageInfo - -func init() { - proto.RegisterType((*AuditSink)(nil), "k8s.io.api.auditregistration.v1alpha1.AuditSink") - proto.RegisterType((*AuditSinkList)(nil), "k8s.io.api.auditregistration.v1alpha1.AuditSinkList") - proto.RegisterType((*AuditSinkSpec)(nil), "k8s.io.api.auditregistration.v1alpha1.AuditSinkSpec") - proto.RegisterType((*Policy)(nil), "k8s.io.api.auditregistration.v1alpha1.Policy") - proto.RegisterType((*ServiceReference)(nil), "k8s.io.api.auditregistration.v1alpha1.ServiceReference") - proto.RegisterType((*Webhook)(nil), "k8s.io.api.auditregistration.v1alpha1.Webhook") - proto.RegisterType((*WebhookClientConfig)(nil), "k8s.io.api.auditregistration.v1alpha1.WebhookClientConfig") - proto.RegisterType((*WebhookThrottleConfig)(nil), "k8s.io.api.auditregistration.v1alpha1.WebhookThrottleConfig") -} - -func init() { - proto.RegisterFile("k8s.io/kubernetes/vendor/k8s.io/api/auditregistration/v1alpha1/generated.proto", fileDescriptor_642d3597c6afa8ba) -} - -var fileDescriptor_642d3597c6afa8ba = []byte{ - // 765 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x52, 0x41, 0x6f, 0x13, 0x47, - 0x14, 0xf6, 0xc6, 0x76, 0x6c, 0x4f, 0x9c, 0x36, 0x9d, 0xb4, 0x95, 0x1b, 0x55, 0x6b, 0x6b, 0xa5, - 0x4a, 0x91, 0xda, 0xcc, 0x36, 0x55, 0xd4, 0x56, 0x88, 0x4b, 0x36, 0x27, 0xa4, 0x10, 0xc2, 0x98, - 0x80, 0x40, 0x08, 0x31, 0x5e, 0x3f, 0xef, 0x0e, 0xb6, 0x77, 0x97, 0xdd, 0x59, 0xa3, 0xdc, 0xf8, - 0x09, 0xfc, 0x05, 0xfe, 0x06, 0x37, 0x24, 0x90, 0x72, 0xcc, 0x31, 0xa7, 0x88, 0x98, 0x03, 0xff, - 0x81, 0x13, 0x9a, 0xd9, 0x59, 0xdb, 0xc4, 0x41, 0x38, 0xb7, 0x79, 0xdf, 0x7b, 0xdf, 0xf7, 0xbe, - 0xf7, 0xde, 0xa0, 0x83, 0xfe, 0xff, 0x09, 0xe1, 0xa1, 0xdd, 0x4f, 0x3b, 0x10, 0x07, 0x20, 0x20, - 0xb1, 0x47, 0x10, 0x74, 0xc3, 0xd8, 0xd6, 0x09, 0x16, 0x71, 0x9b, 0xa5, 0x5d, 0x2e, 0x62, 0xf0, - 0x78, 0x22, 0x62, 0x26, 0x78, 0x18, 0xd8, 0xa3, 0x6d, 0x36, 0x88, 0x7c, 0xb6, 0x6d, 0x7b, 0x10, - 0x40, 0xcc, 0x04, 0x74, 0x49, 0x14, 0x87, 0x22, 0xc4, 0x7f, 0x64, 0x34, 0xc2, 0x22, 0x4e, 0xe6, - 0x68, 0x24, 0xa7, 0x6d, 0x6c, 0x79, 0x5c, 0xf8, 0x69, 0x87, 0xb8, 0xe1, 0xd0, 0xf6, 0x42, 0x2f, - 0xb4, 0x15, 0xbb, 0x93, 0xf6, 0x54, 0xa4, 0x02, 0xf5, 0xca, 0x54, 0x37, 0x76, 0xa6, 0x66, 0x86, - 0xcc, 0xf5, 0x79, 0x00, 0xf1, 0xb1, 0x1d, 0xf5, 0x3d, 0x09, 0x24, 0xf6, 0x10, 0x04, 0xb3, 0x47, - 0x73, 0x5e, 0x36, 0xec, 0x6f, 0xb1, 0xe2, 0x34, 0x10, 0x7c, 0x08, 0x73, 0x84, 0x7f, 0xbf, 0x47, - 0x48, 0x5c, 0x1f, 0x86, 0xec, 0x32, 0xcf, 0x7a, 0x6f, 0xa0, 0xda, 0xae, 0x1c, 0xb6, 0xcd, 0x83, - 0x3e, 0x7e, 0x8a, 0xaa, 0xd2, 0x51, 0x97, 0x09, 0xd6, 0x30, 0x5a, 0xc6, 0xe6, 0xca, 0x3f, 0x7f, - 0x93, 0xe9, 0x56, 0x26, 0xc2, 0x24, 0xea, 0x7b, 0x12, 0x48, 0x88, 0xac, 0x26, 0xa3, 0x6d, 0x72, - 0xa7, 0xf3, 0x0c, 0x5c, 0x71, 0x1b, 0x04, 0x73, 0xf0, 0xc9, 0x79, 0xb3, 0x30, 0x3e, 0x6f, 0xa2, - 0x29, 0x46, 0x27, 0xaa, 0xf8, 0x3e, 0x2a, 0x25, 0x11, 0xb8, 0x8d, 0x25, 0xa5, 0xbe, 0x43, 0x16, - 0xda, 0x39, 0x99, 0x38, 0x6c, 0x47, 0xe0, 0x3a, 0x75, 0xdd, 0xa1, 0x24, 0x23, 0xaa, 0xf4, 0xac, - 0x77, 0x06, 0x5a, 0x9d, 0x54, 0xed, 0xf3, 0x44, 0xe0, 0xc7, 0x73, 0xb3, 0x90, 0xc5, 0x66, 0x91, - 0x6c, 0x35, 0xc9, 0x9a, 0xee, 0x53, 0xcd, 0x91, 0x99, 0x39, 0x8e, 0x50, 0x99, 0x0b, 0x18, 0x26, - 0x8d, 0xa5, 0x56, 0xf1, 0xd2, 0x9a, 0x16, 0x1a, 0xc4, 0x59, 0xd5, 0xe2, 0xe5, 0x5b, 0x52, 0x86, - 0x66, 0x6a, 0xd6, 0xdb, 0xd9, 0x31, 0xe4, 0x78, 0xf8, 0x08, 0x2d, 0x47, 0xe1, 0x80, 0xbb, 0xc7, - 0x7a, 0x88, 0xad, 0x05, 0x3b, 0x1d, 0x2a, 0x92, 0xf3, 0x83, 0x6e, 0xb3, 0x9c, 0xc5, 0x54, 0x8b, - 0xe1, 0x87, 0xa8, 0xf2, 0x02, 0x3a, 0x7e, 0x18, 0xf6, 0xf5, 0x29, 0xc8, 0x82, 0xba, 0x0f, 0x32, - 0x96, 0xf3, 0xa3, 0x16, 0xae, 0x68, 0x80, 0xe6, 0x7a, 0x96, 0x8b, 0x74, 0x33, 0xfc, 0x17, 0x2a, - 0x0f, 0x60, 0x04, 0x03, 0x65, 0xbd, 0xe6, 0xfc, 0x9a, 0x8f, 0xbc, 0x2f, 0xc1, 0xcf, 0xf9, 0x83, - 0x66, 0x45, 0xf8, 0x4f, 0xb4, 0x9c, 0x08, 0xe6, 0x41, 0xb6, 0xd3, 0x9a, 0xb3, 0x2e, 0x6d, 0xb7, - 0x15, 0x22, 0x6b, 0xd5, 0x8b, 0xea, 0x12, 0xeb, 0xb5, 0x81, 0xd6, 0xda, 0x10, 0x8f, 0xb8, 0x0b, - 0x14, 0x7a, 0x10, 0x43, 0xe0, 0x02, 0xb6, 0x51, 0x2d, 0x60, 0x43, 0x48, 0x22, 0xe6, 0x82, 0xee, - 0xf9, 0x93, 0xee, 0x59, 0x3b, 0xc8, 0x13, 0x74, 0x5a, 0x83, 0x5b, 0xa8, 0x24, 0x03, 0xb5, 0x82, - 0xda, 0xf4, 0x5f, 0xc9, 0x5a, 0xaa, 0x32, 0xf8, 0x77, 0x54, 0x8a, 0x98, 0xf0, 0x1b, 0x45, 0x55, - 0x51, 0x95, 0xd9, 0x43, 0x26, 0x7c, 0xaa, 0x50, 0x95, 0x0d, 0x63, 0xd1, 0x28, 0xb5, 0x8c, 0xcd, - 0xb2, 0xce, 0x86, 0xb1, 0xa0, 0x0a, 0xb5, 0x3e, 0x19, 0x28, 0xdf, 0x0e, 0xee, 0xa1, 0xaa, 0xf0, - 0xe3, 0x50, 0x88, 0x01, 0xe8, 0x43, 0xde, 0xbc, 0xde, 0xc2, 0xef, 0x69, 0xf6, 0x5e, 0x18, 0xf4, - 0xb8, 0xe7, 0xd4, 0xe5, 0xbf, 0xcc, 0x31, 0x3a, 0xd1, 0xc6, 0x02, 0xd5, 0xdd, 0x01, 0x87, 0x40, - 0x64, 0x75, 0xfa, 0xb8, 0x37, 0xae, 0xd7, 0x6b, 0x6f, 0x46, 0xc1, 0xf9, 0x59, 0x6f, 0xa5, 0x3e, - 0x8b, 0xd2, 0xaf, 0xba, 0x58, 0x6f, 0x0c, 0xb4, 0x7e, 0x05, 0x17, 0xff, 0x86, 0x8a, 0x69, 0x9c, - 0x9f, 0xbf, 0x32, 0x3e, 0x6f, 0x16, 0x8f, 0xe8, 0x3e, 0x95, 0x18, 0x7e, 0x82, 0x2a, 0x49, 0x76, - 0x3f, 0xed, 0xf1, 0xbf, 0x05, 0x3d, 0x5e, 0xbe, 0xba, 0xb3, 0x22, 0x7f, 0x61, 0x8e, 0xe6, 0xa2, - 0x78, 0x13, 0x55, 0x5d, 0xe6, 0xa4, 0x41, 0x77, 0x00, 0xea, 0x78, 0xf5, 0x6c, 0x65, 0x7b, 0xbb, - 0x19, 0x46, 0x27, 0x59, 0xab, 0x8d, 0x7e, 0xb9, 0x72, 0xc7, 0xd2, 0xfd, 0xf3, 0x28, 0x51, 0xee, - 0x8b, 0x99, 0xfb, 0xbb, 0x87, 0x6d, 0x2a, 0x31, 0xdc, 0x44, 0xe5, 0x4e, 0x1a, 0x27, 0x42, 0x79, - 0x2f, 0x3a, 0x35, 0xf9, 0xab, 0x1d, 0x09, 0xd0, 0x0c, 0x77, 0xc8, 0xc9, 0x85, 0x59, 0x38, 0xbd, - 0x30, 0x0b, 0x67, 0x17, 0x66, 0xe1, 0xe5, 0xd8, 0x34, 0x4e, 0xc6, 0xa6, 0x71, 0x3a, 0x36, 0x8d, - 0xb3, 0xb1, 0x69, 0x7c, 0x18, 0x9b, 0xc6, 0xab, 0x8f, 0x66, 0xe1, 0x51, 0x35, 0x9f, 0xea, 0x4b, - 0x00, 0x00, 0x00, 0xff, 0xff, 0x0a, 0x6c, 0xff, 0x86, 0xcd, 0x06, 0x00, 0x00, -} - -func (m *AuditSink) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *AuditSink) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *AuditSink) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - { - size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x12 - { - size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0xa - return len(dAtA) - i, nil -} - -func (m *AuditSinkList) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *AuditSinkList) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *AuditSinkList) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - if len(m.Items) > 0 { - for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { - { - size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x12 - } - } - { - size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0xa - return len(dAtA) - i, nil -} - -func (m *AuditSinkSpec) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *AuditSinkSpec) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *AuditSinkSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - { - size, err := m.Webhook.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x12 - { - size, err := m.Policy.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0xa - return len(dAtA) - i, nil -} - -func (m *Policy) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *Policy) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *Policy) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - if len(m.Stages) > 0 { - for iNdEx := len(m.Stages) - 1; iNdEx >= 0; iNdEx-- { - i -= len(m.Stages[iNdEx]) - copy(dAtA[i:], m.Stages[iNdEx]) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.Stages[iNdEx]))) - i-- - dAtA[i] = 0x12 - } - } - i -= len(m.Level) - copy(dAtA[i:], m.Level) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.Level))) - i-- - dAtA[i] = 0xa - return len(dAtA) - i, nil -} - -func (m *ServiceReference) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *ServiceReference) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *ServiceReference) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - if m.Port != nil { - i = encodeVarintGenerated(dAtA, i, uint64(*m.Port)) - i-- - dAtA[i] = 0x20 - } - if m.Path != nil { - i -= len(*m.Path) - copy(dAtA[i:], *m.Path) - i = encodeVarintGenerated(dAtA, i, uint64(len(*m.Path))) - i-- - dAtA[i] = 0x1a - } - i -= len(m.Name) - copy(dAtA[i:], m.Name) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.Name))) - i-- - dAtA[i] = 0x12 - i -= len(m.Namespace) - copy(dAtA[i:], m.Namespace) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.Namespace))) - i-- - dAtA[i] = 0xa - return len(dAtA) - i, nil -} - -func (m *Webhook) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *Webhook) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *Webhook) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - { - size, err := m.ClientConfig.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x12 - if m.Throttle != nil { - { - size, err := m.Throttle.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0xa - } - return len(dAtA) - i, nil -} - -func (m *WebhookClientConfig) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *WebhookClientConfig) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *WebhookClientConfig) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - if m.CABundle != nil { - i -= len(m.CABundle) - copy(dAtA[i:], m.CABundle) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.CABundle))) - i-- - dAtA[i] = 0x1a - } - if m.Service != nil { - { - size, err := m.Service.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x12 - } - if m.URL != nil { - i -= len(*m.URL) - copy(dAtA[i:], *m.URL) - i = encodeVarintGenerated(dAtA, i, uint64(len(*m.URL))) - i-- - dAtA[i] = 0xa - } - return len(dAtA) - i, nil -} - -func (m *WebhookThrottleConfig) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *WebhookThrottleConfig) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *WebhookThrottleConfig) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - if m.Burst != nil { - i = encodeVarintGenerated(dAtA, i, uint64(*m.Burst)) - i-- - dAtA[i] = 0x10 - } - if m.QPS != nil { - i = encodeVarintGenerated(dAtA, i, uint64(*m.QPS)) - i-- - dAtA[i] = 0x8 - } - return len(dAtA) - i, nil -} - -func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { - offset -= sovGenerated(v) - base := offset - for v >= 1<<7 { - dAtA[offset] = uint8(v&0x7f | 0x80) - v >>= 7 - offset++ - } - dAtA[offset] = uint8(v) - return base -} -func (m *AuditSink) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - l = m.ObjectMeta.Size() - n += 1 + l + sovGenerated(uint64(l)) - l = m.Spec.Size() - n += 1 + l + sovGenerated(uint64(l)) - return n -} - -func (m *AuditSinkList) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - l = m.ListMeta.Size() - n += 1 + l + sovGenerated(uint64(l)) - if len(m.Items) > 0 { - for _, e := range m.Items { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - } - return n -} - -func (m *AuditSinkSpec) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - l = m.Policy.Size() - n += 1 + l + sovGenerated(uint64(l)) - l = m.Webhook.Size() - n += 1 + l + sovGenerated(uint64(l)) - return n -} - -func (m *Policy) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - l = len(m.Level) - n += 1 + l + sovGenerated(uint64(l)) - if len(m.Stages) > 0 { - for _, s := range m.Stages { - l = len(s) - n += 1 + l + sovGenerated(uint64(l)) - } - } - return n -} - -func (m *ServiceReference) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - l = len(m.Namespace) - n += 1 + l + sovGenerated(uint64(l)) - l = len(m.Name) - n += 1 + l + sovGenerated(uint64(l)) - if m.Path != nil { - l = len(*m.Path) - n += 1 + l + sovGenerated(uint64(l)) - } - if m.Port != nil { - n += 1 + sovGenerated(uint64(*m.Port)) - } - return n -} - -func (m *Webhook) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - if m.Throttle != nil { - l = m.Throttle.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - l = m.ClientConfig.Size() - n += 1 + l + sovGenerated(uint64(l)) - return n -} - -func (m *WebhookClientConfig) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - if m.URL != nil { - l = len(*m.URL) - n += 1 + l + sovGenerated(uint64(l)) - } - if m.Service != nil { - l = m.Service.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - if m.CABundle != nil { - l = len(m.CABundle) - n += 1 + l + sovGenerated(uint64(l)) - } - return n -} - -func (m *WebhookThrottleConfig) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - if m.QPS != nil { - n += 1 + sovGenerated(uint64(*m.QPS)) - } - if m.Burst != nil { - n += 1 + sovGenerated(uint64(*m.Burst)) - } - return n -} - -func sovGenerated(x uint64) (n int) { - return (math_bits.Len64(x|1) + 6) / 7 -} -func sozGenerated(x uint64) (n int) { - return sovGenerated(uint64((x << 1) ^ uint64((int64(x) >> 63)))) -} -func (this *AuditSink) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&AuditSink{`, - `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, - `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "AuditSinkSpec", "AuditSinkSpec", 1), `&`, ``, 1) + `,`, - `}`, - }, "") - return s -} -func (this *AuditSinkList) String() string { - if this == nil { - return "nil" - } - repeatedStringForItems := "[]AuditSink{" - for _, f := range this.Items { - repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "AuditSink", "AuditSink", 1), `&`, ``, 1) + "," - } - repeatedStringForItems += "}" - s := strings.Join([]string{`&AuditSinkList{`, - `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, - `Items:` + repeatedStringForItems + `,`, - `}`, - }, "") - return s -} -func (this *AuditSinkSpec) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&AuditSinkSpec{`, - `Policy:` + strings.Replace(strings.Replace(this.Policy.String(), "Policy", "Policy", 1), `&`, ``, 1) + `,`, - `Webhook:` + strings.Replace(strings.Replace(this.Webhook.String(), "Webhook", "Webhook", 1), `&`, ``, 1) + `,`, - `}`, - }, "") - return s -} -func (this *Policy) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&Policy{`, - `Level:` + fmt.Sprintf("%v", this.Level) + `,`, - `Stages:` + fmt.Sprintf("%v", this.Stages) + `,`, - `}`, - }, "") - return s -} -func (this *ServiceReference) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&ServiceReference{`, - `Namespace:` + fmt.Sprintf("%v", this.Namespace) + `,`, - `Name:` + fmt.Sprintf("%v", this.Name) + `,`, - `Path:` + valueToStringGenerated(this.Path) + `,`, - `Port:` + valueToStringGenerated(this.Port) + `,`, - `}`, - }, "") - return s -} -func (this *Webhook) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&Webhook{`, - `Throttle:` + strings.Replace(this.Throttle.String(), "WebhookThrottleConfig", "WebhookThrottleConfig", 1) + `,`, - `ClientConfig:` + strings.Replace(strings.Replace(this.ClientConfig.String(), "WebhookClientConfig", "WebhookClientConfig", 1), `&`, ``, 1) + `,`, - `}`, - }, "") - return s -} -func (this *WebhookClientConfig) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&WebhookClientConfig{`, - `URL:` + valueToStringGenerated(this.URL) + `,`, - `Service:` + strings.Replace(this.Service.String(), "ServiceReference", "ServiceReference", 1) + `,`, - `CABundle:` + valueToStringGenerated(this.CABundle) + `,`, - `}`, - }, "") - return s -} -func (this *WebhookThrottleConfig) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&WebhookThrottleConfig{`, - `QPS:` + valueToStringGenerated(this.QPS) + `,`, - `Burst:` + valueToStringGenerated(this.Burst) + `,`, - `}`, - }, "") - return s -} -func valueToStringGenerated(v interface{}) string { - rv := reflect.ValueOf(v) - if rv.IsNil() { - return "nil" - } - pv := reflect.Indirect(rv).Interface() - return fmt.Sprintf("*%v", pv) -} -func (m *AuditSink) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: AuditSink: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: AuditSink: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func (m *AuditSinkList) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: AuditSinkList: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: AuditSinkList: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.Items = append(m.Items, AuditSink{}) - if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func (m *AuditSinkSpec) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: AuditSinkSpec: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: AuditSinkSpec: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Policy", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if err := m.Policy.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Webhook", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if err := m.Webhook.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func (m *Policy) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: Policy: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: Policy: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Level", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.Level = Level(dAtA[iNdEx:postIndex]) - iNdEx = postIndex - case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Stages", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.Stages = append(m.Stages, Stage(dAtA[iNdEx:postIndex])) - iNdEx = postIndex - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func (m *ServiceReference) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: ServiceReference: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: ServiceReference: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Namespace", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.Namespace = string(dAtA[iNdEx:postIndex]) - iNdEx = postIndex - case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Name", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.Name = string(dAtA[iNdEx:postIndex]) - iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Path", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - s := string(dAtA[iNdEx:postIndex]) - m.Path = &s - iNdEx = postIndex - case 4: - if wireType != 0 { - return fmt.Errorf("proto: wrong wireType = %d for field Port", wireType) - } - var v int32 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - v |= int32(b&0x7F) << shift - if b < 0x80 { - break - } - } - m.Port = &v - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func (m *Webhook) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: Webhook: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: Webhook: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Throttle", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if m.Throttle == nil { - m.Throttle = &WebhookThrottleConfig{} - } - if err := m.Throttle.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ClientConfig", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if err := m.ClientConfig.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func (m *WebhookClientConfig) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: WebhookClientConfig: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: WebhookClientConfig: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field URL", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - s := string(dAtA[iNdEx:postIndex]) - m.URL = &s - iNdEx = postIndex - case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Service", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if m.Service == nil { - m.Service = &ServiceReference{} - } - if err := m.Service.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field CABundle", wireType) - } - var byteLen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - byteLen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if byteLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + byteLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.CABundle = append(m.CABundle[:0], dAtA[iNdEx:postIndex]...) - if m.CABundle == nil { - m.CABundle = []byte{} - } - iNdEx = postIndex - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func (m *WebhookThrottleConfig) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: WebhookThrottleConfig: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: WebhookThrottleConfig: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - case 1: - if wireType != 0 { - return fmt.Errorf("proto: wrong wireType = %d for field QPS", wireType) - } - var v int64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - v |= int64(b&0x7F) << shift - if b < 0x80 { - break - } - } - m.QPS = &v - case 2: - if wireType != 0 { - return fmt.Errorf("proto: wrong wireType = %d for field Burst", wireType) - } - var v int64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - v |= int64(b&0x7F) << shift - if b < 0x80 { - break - } - } - m.Burst = &v - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} -func skipGenerated(dAtA []byte) (n int, err error) { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - wireType := int(wire & 0x7) - switch wireType { - case 0: - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - iNdEx++ - if dAtA[iNdEx-1] < 0x80 { - break - } - } - return iNdEx, nil - case 1: - iNdEx += 8 - return iNdEx, nil - case 2: - var length int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - length |= (int(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - if length < 0 { - return 0, ErrInvalidLengthGenerated - } - iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil - case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil - case 4: - return iNdEx, nil - case 5: - iNdEx += 4 - return iNdEx, nil - default: - return 0, fmt.Errorf("proto: illegal wireType %d", wireType) - } - } - panic("unreachable") -} - -var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") -) diff --git a/vendor/k8s.io/api/auditregistration/v1alpha1/generated.proto b/vendor/k8s.io/api/auditregistration/v1alpha1/generated.proto deleted file mode 100644 index 674debee..00000000 --- a/vendor/k8s.io/api/auditregistration/v1alpha1/generated.proto +++ /dev/null @@ -1,162 +0,0 @@ -/* -Copyright The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - - -// This file was autogenerated by go-to-protobuf. Do not edit it manually! - -syntax = 'proto2'; - -package k8s.io.api.auditregistration.v1alpha1; - -import "k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto"; -import "k8s.io/apimachinery/pkg/runtime/generated.proto"; -import "k8s.io/apimachinery/pkg/runtime/schema/generated.proto"; - -// Package-wide variables from generator "generated". -option go_package = "v1alpha1"; - -// AuditSink represents a cluster level audit sink -message AuditSink { - // +optional - optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; - - // Spec defines the audit configuration spec - optional AuditSinkSpec spec = 2; -} - -// AuditSinkList is a list of AuditSink items. -message AuditSinkList { - // +optional - optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; - - // List of audit configurations. - repeated AuditSink items = 2; -} - -// AuditSinkSpec holds the spec for the audit sink -message AuditSinkSpec { - // Policy defines the policy for selecting which events should be sent to the webhook - // required - optional Policy policy = 1; - - // Webhook to send events - // required - optional Webhook webhook = 2; -} - -// Policy defines the configuration of how audit events are logged -message Policy { - // The Level that all requests are recorded at. - // available options: None, Metadata, Request, RequestResponse - // required - optional string level = 1; - - // Stages is a list of stages for which events are created. - // +optional - repeated string stages = 2; -} - -// ServiceReference holds a reference to Service.legacy.k8s.io -message ServiceReference { - // `namespace` is the namespace of the service. - // Required - optional string namespace = 1; - - // `name` is the name of the service. - // Required - optional string name = 2; - - // `path` is an optional URL path which will be sent in any request to - // this service. - // +optional - optional string path = 3; - - // If specified, the port on the service that hosting webhook. - // Default to 443 for backward compatibility. - // `port` should be a valid port number (1-65535, inclusive). - // +optional - optional int32 port = 4; -} - -// Webhook holds the configuration of the webhook -message Webhook { - // Throttle holds the options for throttling the webhook - // +optional - optional WebhookThrottleConfig throttle = 1; - - // ClientConfig holds the connection parameters for the webhook - // required - optional WebhookClientConfig clientConfig = 2; -} - -// WebhookClientConfig contains the information to make a connection with the webhook -message WebhookClientConfig { - // `url` gives the location of the webhook, in standard URL form - // (`scheme://host:port/path`). Exactly one of `url` or `service` - // must be specified. - // - // The `host` should not refer to a service running in the cluster; use - // the `service` field instead. The host might be resolved via external - // DNS in some apiservers (e.g., `kube-apiserver` cannot resolve - // in-cluster DNS as that would be a layering violation). `host` may - // also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is - // risky unless you take great care to run this webhook on all hosts - // which run an apiserver which might need to make calls to this - // webhook. Such installs are likely to be non-portable, i.e., not easy - // to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in - // a URL. You may use the path to pass an arbitrary string to the - // webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not - // allowed. Fragments ("#...") and query parameters ("?...") are not - // allowed, either. - // - // +optional - optional string url = 1; - - // `service` is a reference to the service for this webhook. Either - // `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // +optional - optional ServiceReference service = 2; - - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. - // If unspecified, system trust roots on the apiserver are used. - // +optional - optional bytes caBundle = 3; -} - -// WebhookThrottleConfig holds the configuration for throttling events -message WebhookThrottleConfig { - // ThrottleQPS maximum number of batches per second - // default 10 QPS - // +optional - optional int64 qps = 1; - - // ThrottleBurst is the maximum number of events sent at the same moment - // default 15 QPS - // +optional - optional int64 burst = 2; -} - diff --git a/vendor/k8s.io/api/auditregistration/v1alpha1/types.go b/vendor/k8s.io/api/auditregistration/v1alpha1/types.go deleted file mode 100644 index a0fb48c3..00000000 --- a/vendor/k8s.io/api/auditregistration/v1alpha1/types.go +++ /dev/null @@ -1,198 +0,0 @@ -/* -Copyright 2018 The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -// +k8s:openapi-gen=true - -package v1alpha1 - -import ( - metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" -) - -// Level defines the amount of information logged during auditing -type Level string - -// Valid audit levels -const ( - // LevelNone disables auditing - LevelNone Level = "None" - // LevelMetadata provides the basic level of auditing. - LevelMetadata Level = "Metadata" - // LevelRequest provides Metadata level of auditing, and additionally - // logs the request object (does not apply for non-resource requests). - LevelRequest Level = "Request" - // LevelRequestResponse provides Request level of auditing, and additionally - // logs the response object (does not apply for non-resource requests and watches). - LevelRequestResponse Level = "RequestResponse" -) - -// Stage defines the stages in request handling during which audit events may be generated. -type Stage string - -// Valid audit stages. -const ( - // The stage for events generated after the audit handler receives the request, but before it - // is delegated down the handler chain. - StageRequestReceived = "RequestReceived" - // The stage for events generated after the response headers are sent, but before the response body - // is sent. This stage is only generated for long-running requests (e.g. watch). - StageResponseStarted = "ResponseStarted" - // The stage for events generated after the response body has been completed, and no more bytes - // will be sent. - StageResponseComplete = "ResponseComplete" - // The stage for events generated when a panic occurred. - StagePanic = "Panic" -) - -// +genclient -// +genclient:nonNamespaced -// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object - -// AuditSink represents a cluster level audit sink -type AuditSink struct { - metav1.TypeMeta `json:",inline"` - // +optional - metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` - - // Spec defines the audit configuration spec - Spec AuditSinkSpec `json:"spec,omitempty" protobuf:"bytes,2,opt,name=spec"` -} - -// AuditSinkSpec holds the spec for the audit sink -type AuditSinkSpec struct { - // Policy defines the policy for selecting which events should be sent to the webhook - // required - Policy Policy `json:"policy" protobuf:"bytes,1,opt,name=policy"` - - // Webhook to send events - // required - Webhook Webhook `json:"webhook" protobuf:"bytes,2,opt,name=webhook"` -} - -// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object - -// AuditSinkList is a list of AuditSink items. -type AuditSinkList struct { - metav1.TypeMeta `json:",inline"` - // +optional - metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` - - // List of audit configurations. - Items []AuditSink `json:"items" protobuf:"bytes,2,rep,name=items"` -} - -// Policy defines the configuration of how audit events are logged -type Policy struct { - // The Level that all requests are recorded at. - // available options: None, Metadata, Request, RequestResponse - // required - Level Level `json:"level" protobuf:"bytes,1,opt,name=level"` - - // Stages is a list of stages for which events are created. - // +optional - Stages []Stage `json:"stages" protobuf:"bytes,2,opt,name=stages"` -} - -// Webhook holds the configuration of the webhook -type Webhook struct { - // Throttle holds the options for throttling the webhook - // +optional - Throttle *WebhookThrottleConfig `json:"throttle,omitempty" protobuf:"bytes,1,opt,name=throttle"` - - // ClientConfig holds the connection parameters for the webhook - // required - ClientConfig WebhookClientConfig `json:"clientConfig" protobuf:"bytes,2,opt,name=clientConfig"` -} - -// WebhookThrottleConfig holds the configuration for throttling events -type WebhookThrottleConfig struct { - // ThrottleQPS maximum number of batches per second - // default 10 QPS - // +optional - QPS *int64 `json:"qps,omitempty" protobuf:"bytes,1,opt,name=qps"` - - // ThrottleBurst is the maximum number of events sent at the same moment - // default 15 QPS - // +optional - Burst *int64 `json:"burst,omitempty" protobuf:"bytes,2,opt,name=burst"` -} - -// WebhookClientConfig contains the information to make a connection with the webhook -type WebhookClientConfig struct { - // `url` gives the location of the webhook, in standard URL form - // (`scheme://host:port/path`). Exactly one of `url` or `service` - // must be specified. - // - // The `host` should not refer to a service running in the cluster; use - // the `service` field instead. The host might be resolved via external - // DNS in some apiservers (e.g., `kube-apiserver` cannot resolve - // in-cluster DNS as that would be a layering violation). `host` may - // also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is - // risky unless you take great care to run this webhook on all hosts - // which run an apiserver which might need to make calls to this - // webhook. Such installs are likely to be non-portable, i.e., not easy - // to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in - // a URL. You may use the path to pass an arbitrary string to the - // webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not - // allowed. Fragments ("#...") and query parameters ("?...") are not - // allowed, either. - // - // +optional - URL *string `json:"url,omitempty" protobuf:"bytes,1,opt,name=url"` - - // `service` is a reference to the service for this webhook. Either - // `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // +optional - Service *ServiceReference `json:"service,omitempty" protobuf:"bytes,2,opt,name=service"` - - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. - // If unspecified, system trust roots on the apiserver are used. - // +optional - CABundle []byte `json:"caBundle,omitempty" protobuf:"bytes,3,opt,name=caBundle"` -} - -// ServiceReference holds a reference to Service.legacy.k8s.io -type ServiceReference struct { - // `namespace` is the namespace of the service. - // Required - Namespace string `json:"namespace" protobuf:"bytes,1,opt,name=namespace"` - - // `name` is the name of the service. - // Required - Name string `json:"name" protobuf:"bytes,2,opt,name=name"` - - // `path` is an optional URL path which will be sent in any request to - // this service. - // +optional - Path *string `json:"path,omitempty" protobuf:"bytes,3,opt,name=path"` - - // If specified, the port on the service that hosting webhook. - // Default to 443 for backward compatibility. - // `port` should be a valid port number (1-65535, inclusive). - // +optional - Port *int32 `json:"port,omitempty" protobuf:"varint,4,opt,name=port"` -} diff --git a/vendor/k8s.io/api/auditregistration/v1alpha1/types_swagger_doc_generated.go b/vendor/k8s.io/api/auditregistration/v1alpha1/types_swagger_doc_generated.go deleted file mode 100644 index 1a86f4da..00000000 --- a/vendor/k8s.io/api/auditregistration/v1alpha1/types_swagger_doc_generated.go +++ /dev/null @@ -1,111 +0,0 @@ -/* -Copyright The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -package v1alpha1 - -// This file contains a collection of methods that can be used from go-restful to -// generate Swagger API documentation for its models. Please read this PR for more -// information on the implementation: https://github.com/emicklei/go-restful/pull/215 -// -// TODOs are ignored from the parser (e.g. TODO(andronat):... || TODO:...) if and only if -// they are on one line! For multiple line or blocks that you want to ignore use ---. -// Any context after a --- is ignored. -// -// Those methods can be generated by using hack/update-generated-swagger-docs.sh - -// AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. -var map_AuditSink = map[string]string{ - "": "AuditSink represents a cluster level audit sink", - "spec": "Spec defines the audit configuration spec", -} - -func (AuditSink) SwaggerDoc() map[string]string { - return map_AuditSink -} - -var map_AuditSinkList = map[string]string{ - "": "AuditSinkList is a list of AuditSink items.", - "items": "List of audit configurations.", -} - -func (AuditSinkList) SwaggerDoc() map[string]string { - return map_AuditSinkList -} - -var map_AuditSinkSpec = map[string]string{ - "": "AuditSinkSpec holds the spec for the audit sink", - "policy": "Policy defines the policy for selecting which events should be sent to the webhook required", - "webhook": "Webhook to send events required", -} - -func (AuditSinkSpec) SwaggerDoc() map[string]string { - return map_AuditSinkSpec -} - -var map_Policy = map[string]string{ - "": "Policy defines the configuration of how audit events are logged", - "level": "The Level that all requests are recorded at. available options: None, Metadata, Request, RequestResponse required", - "stages": "Stages is a list of stages for which events are created.", -} - -func (Policy) SwaggerDoc() map[string]string { - return map_Policy -} - -var map_ServiceReference = map[string]string{ - "": "ServiceReference holds a reference to Service.legacy.k8s.io", - "namespace": "`namespace` is the namespace of the service. Required", - "name": "`name` is the name of the service. Required", - "path": "`path` is an optional URL path which will be sent in any request to this service.", - "port": "If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).", -} - -func (ServiceReference) SwaggerDoc() map[string]string { - return map_ServiceReference -} - -var map_Webhook = map[string]string{ - "": "Webhook holds the configuration of the webhook", - "throttle": "Throttle holds the options for throttling the webhook", - "clientConfig": "ClientConfig holds the connection parameters for the webhook required", -} - -func (Webhook) SwaggerDoc() map[string]string { - return map_Webhook -} - -var map_WebhookClientConfig = map[string]string{ - "": "WebhookClientConfig contains the information to make a connection with the webhook", - "url": "`url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified.\n\nThe `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address.\n\nPlease note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster.\n\nThe scheme must be \"https\"; the URL must begin with \"https://\".\n\nA path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier.\n\nAttempting to use a user or basic auth e.g. \"user:password@\" is not allowed. Fragments (\"#...\") and query parameters (\"?...\") are not allowed, either.", - "service": "`service` is a reference to the service for this webhook. Either `service` or `url` must be specified.\n\nIf the webhook is running within the cluster, then you should use `service`.", - "caBundle": "`caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used.", -} - -func (WebhookClientConfig) SwaggerDoc() map[string]string { - return map_WebhookClientConfig -} - -var map_WebhookThrottleConfig = map[string]string{ - "": "WebhookThrottleConfig holds the configuration for throttling events", - "qps": "ThrottleQPS maximum number of batches per second default 10 QPS", - "burst": "ThrottleBurst is the maximum number of events sent at the same moment default 15 QPS", -} - -func (WebhookThrottleConfig) SwaggerDoc() map[string]string { - return map_WebhookThrottleConfig -} - -// AUTO-GENERATED FUNCTIONS END HERE diff --git a/vendor/k8s.io/api/auditregistration/v1alpha1/zz_generated.deepcopy.go b/vendor/k8s.io/api/auditregistration/v1alpha1/zz_generated.deepcopy.go deleted file mode 100644 index 621a19e8..00000000 --- a/vendor/k8s.io/api/auditregistration/v1alpha1/zz_generated.deepcopy.go +++ /dev/null @@ -1,229 +0,0 @@ -// +build !ignore_autogenerated - -/* -Copyright The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -// Code generated by deepcopy-gen. DO NOT EDIT. - -package v1alpha1 - -import ( - runtime "k8s.io/apimachinery/pkg/runtime" -) - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *AuditSink) DeepCopyInto(out *AuditSink) { - *out = *in - out.TypeMeta = in.TypeMeta - in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) - in.Spec.DeepCopyInto(&out.Spec) - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AuditSink. -func (in *AuditSink) DeepCopy() *AuditSink { - if in == nil { - return nil - } - out := new(AuditSink) - in.DeepCopyInto(out) - return out -} - -// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. -func (in *AuditSink) DeepCopyObject() runtime.Object { - if c := in.DeepCopy(); c != nil { - return c - } - return nil -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *AuditSinkList) DeepCopyInto(out *AuditSinkList) { - *out = *in - out.TypeMeta = in.TypeMeta - in.ListMeta.DeepCopyInto(&out.ListMeta) - if in.Items != nil { - in, out := &in.Items, &out.Items - *out = make([]AuditSink, len(*in)) - for i := range *in { - (*in)[i].DeepCopyInto(&(*out)[i]) - } - } - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AuditSinkList. -func (in *AuditSinkList) DeepCopy() *AuditSinkList { - if in == nil { - return nil - } - out := new(AuditSinkList) - in.DeepCopyInto(out) - return out -} - -// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. -func (in *AuditSinkList) DeepCopyObject() runtime.Object { - if c := in.DeepCopy(); c != nil { - return c - } - return nil -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *AuditSinkSpec) DeepCopyInto(out *AuditSinkSpec) { - *out = *in - in.Policy.DeepCopyInto(&out.Policy) - in.Webhook.DeepCopyInto(&out.Webhook) - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new AuditSinkSpec. -func (in *AuditSinkSpec) DeepCopy() *AuditSinkSpec { - if in == nil { - return nil - } - out := new(AuditSinkSpec) - in.DeepCopyInto(out) - return out -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *Policy) DeepCopyInto(out *Policy) { - *out = *in - if in.Stages != nil { - in, out := &in.Stages, &out.Stages - *out = make([]Stage, len(*in)) - copy(*out, *in) - } - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Policy. -func (in *Policy) DeepCopy() *Policy { - if in == nil { - return nil - } - out := new(Policy) - in.DeepCopyInto(out) - return out -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *ServiceReference) DeepCopyInto(out *ServiceReference) { - *out = *in - if in.Path != nil { - in, out := &in.Path, &out.Path - *out = new(string) - **out = **in - } - if in.Port != nil { - in, out := &in.Port, &out.Port - *out = new(int32) - **out = **in - } - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceReference. -func (in *ServiceReference) DeepCopy() *ServiceReference { - if in == nil { - return nil - } - out := new(ServiceReference) - in.DeepCopyInto(out) - return out -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *Webhook) DeepCopyInto(out *Webhook) { - *out = *in - if in.Throttle != nil { - in, out := &in.Throttle, &out.Throttle - *out = new(WebhookThrottleConfig) - (*in).DeepCopyInto(*out) - } - in.ClientConfig.DeepCopyInto(&out.ClientConfig) - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Webhook. -func (in *Webhook) DeepCopy() *Webhook { - if in == nil { - return nil - } - out := new(Webhook) - in.DeepCopyInto(out) - return out -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *WebhookClientConfig) DeepCopyInto(out *WebhookClientConfig) { - *out = *in - if in.URL != nil { - in, out := &in.URL, &out.URL - *out = new(string) - **out = **in - } - if in.Service != nil { - in, out := &in.Service, &out.Service - *out = new(ServiceReference) - (*in).DeepCopyInto(*out) - } - if in.CABundle != nil { - in, out := &in.CABundle, &out.CABundle - *out = make([]byte, len(*in)) - copy(*out, *in) - } - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new WebhookClientConfig. -func (in *WebhookClientConfig) DeepCopy() *WebhookClientConfig { - if in == nil { - return nil - } - out := new(WebhookClientConfig) - in.DeepCopyInto(out) - return out -} - -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *WebhookThrottleConfig) DeepCopyInto(out *WebhookThrottleConfig) { - *out = *in - if in.QPS != nil { - in, out := &in.QPS, &out.QPS - *out = new(int64) - **out = **in - } - if in.Burst != nil { - in, out := &in.Burst, &out.Burst - *out = new(int64) - **out = **in - } - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new WebhookThrottleConfig. -func (in *WebhookThrottleConfig) DeepCopy() *WebhookThrottleConfig { - if in == nil { - return nil - } - out := new(WebhookThrottleConfig) - in.DeepCopyInto(out) - return out -} diff --git a/vendor/k8s.io/api/authentication/v1/generated.pb.go b/vendor/k8s.io/api/authentication/v1/generated.pb.go index 02be20de..6524f8ca 100644 --- a/vendor/k8s.io/api/authentication/v1/generated.pb.go +++ b/vendor/k8s.io/api/authentication/v1/generated.pb.go @@ -44,7 +44,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *BoundObjectReference) Reset() { *m = BoundObjectReference{} } func (*BoundObjectReference) ProtoMessage() {} @@ -2498,6 +2498,7 @@ func (m *UserInfo) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -2529,10 +2530,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -2553,55 +2552,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/authentication/v1/types.go b/vendor/k8s.io/api/authentication/v1/types.go index c48b0369..668b7203 100644 --- a/vendor/k8s.io/api/authentication/v1/types.go +++ b/vendor/k8s.io/api/authentication/v1/types.go @@ -40,7 +40,7 @@ const ( // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // TokenReview attempts to authenticate a token to a known user. diff --git a/vendor/k8s.io/api/authentication/v1beta1/doc.go b/vendor/k8s.io/api/authentication/v1beta1/doc.go index 185a2240..2a2b176e 100644 --- a/vendor/k8s.io/api/authentication/v1beta1/doc.go +++ b/vendor/k8s.io/api/authentication/v1beta1/doc.go @@ -18,5 +18,6 @@ limitations under the License. // +k8s:protobuf-gen=package // +groupName=authentication.k8s.io // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v1beta1 // import "k8s.io/api/authentication/v1beta1" diff --git a/vendor/k8s.io/api/authentication/v1beta1/generated.pb.go b/vendor/k8s.io/api/authentication/v1beta1/generated.pb.go index 0721bda8..6c391dbf 100644 --- a/vendor/k8s.io/api/authentication/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/authentication/v1beta1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *ExtraValue) Reset() { *m = ExtraValue{} } func (*ExtraValue) ProtoMessage() {} @@ -1475,6 +1475,7 @@ func (m *UserInfo) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1506,10 +1507,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1530,55 +1529,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/authentication/v1beta1/types.go b/vendor/k8s.io/api/authentication/v1beta1/types.go index 0b6cba82..121b3461 100644 --- a/vendor/k8s.io/api/authentication/v1beta1/types.go +++ b/vendor/k8s.io/api/authentication/v1beta1/types.go @@ -24,8 +24,11 @@ import ( // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.4 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=authentication.k8s.io,v1,TokenReview // TokenReview attempts to authenticate a token to a known user. // Note: TokenReview requests may be cached by the webhook token authenticator diff --git a/vendor/k8s.io/api/authentication/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/authentication/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..9b5744db --- /dev/null +++ b/vendor/k8s.io/api/authentication/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,49 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *TokenReview) APILifecycleIntroduced() (major, minor int) { + return 1, 4 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *TokenReview) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *TokenReview) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "authentication.k8s.io", Version: "v1", Kind: "TokenReview"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *TokenReview) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/authorization/v1/generated.pb.go b/vendor/k8s.io/api/authorization/v1/generated.pb.go index 0dc01bc9..dbc0bdc7 100644 --- a/vendor/k8s.io/api/authorization/v1/generated.pb.go +++ b/vendor/k8s.io/api/authorization/v1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *ExtraValue) Reset() { *m = ExtraValue{} } func (*ExtraValue) ProtoMessage() {} @@ -4004,6 +4004,7 @@ func (m *SubjectRulesReviewStatus) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -4035,10 +4036,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -4059,55 +4058,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/authorization/v1/types.go b/vendor/k8s.io/api/authorization/v1/types.go index 86b05c54..be8913eb 100644 --- a/vendor/k8s.io/api/authorization/v1/types.go +++ b/vendor/k8s.io/api/authorization/v1/types.go @@ -24,7 +24,7 @@ import ( // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // SubjectAccessReview checks whether or not a user or group can perform an action. @@ -43,7 +43,7 @@ type SubjectAccessReview struct { // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // SelfSubjectAccessReview checks whether or the current user can perform an action. Not filling in a @@ -63,7 +63,7 @@ type SelfSubjectAccessReview struct { } // +genclient -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // LocalSubjectAccessReview checks whether or not a user or group can perform an action in a given namespace. @@ -189,7 +189,7 @@ type SubjectAccessReviewStatus struct { // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // SelfSubjectRulesReview enumerates the set of actions the current user can perform within a namespace. diff --git a/vendor/k8s.io/api/authorization/v1beta1/doc.go b/vendor/k8s.io/api/authorization/v1beta1/doc.go index 7046f111..c996e35c 100644 --- a/vendor/k8s.io/api/authorization/v1beta1/doc.go +++ b/vendor/k8s.io/api/authorization/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=authorization.k8s.io diff --git a/vendor/k8s.io/api/authorization/v1beta1/generated.pb.go b/vendor/k8s.io/api/authorization/v1beta1/generated.pb.go index f0def20b..647c0c58 100644 --- a/vendor/k8s.io/api/authorization/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/authorization/v1beta1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *ExtraValue) Reset() { *m = ExtraValue{} } func (*ExtraValue) ProtoMessage() {} @@ -4004,6 +4004,7 @@ func (m *SubjectRulesReviewStatus) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -4035,10 +4036,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -4059,55 +4058,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/authorization/v1beta1/types.go b/vendor/k8s.io/api/authorization/v1beta1/types.go index 618ff8c0..c62b5ea2 100644 --- a/vendor/k8s.io/api/authorization/v1beta1/types.go +++ b/vendor/k8s.io/api/authorization/v1beta1/types.go @@ -24,8 +24,11 @@ import ( // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=authorization.k8s.io,v1,SubjectAccessReview // SubjectAccessReview checks whether or not a user or group can perform an action. type SubjectAccessReview struct { @@ -43,8 +46,11 @@ type SubjectAccessReview struct { // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=authorization.k8s.io,v1,SelfSubjectAccessReview // SelfSubjectAccessReview checks whether or the current user can perform an action. Not filling in a // spec.namespace means "in all namespaces". Self is a special case, because users should always be able @@ -63,8 +69,11 @@ type SelfSubjectAccessReview struct { } // +genclient -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=authorization.k8s.io,v1,LocalSubjectAccessReview // LocalSubjectAccessReview checks whether or not a user or group can perform an action in a given namespace. // Having a namespace scoped resource makes it much easier to grant namespace scoped policy that includes permissions @@ -189,8 +198,11 @@ type SubjectAccessReviewStatus struct { // +genclient // +genclient:nonNamespaced -// +genclient:noVerbs +// +genclient:onlyVerbs=create // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=authorization.k8s.io,v1,SelfSubjectRulesReview // SelfSubjectRulesReview enumerates the set of actions the current user can perform within a namespace. // The returned list of actions may be incomplete depending on the server's authorization mode, diff --git a/vendor/k8s.io/api/authorization/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/authorization/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..fcb75dd9 --- /dev/null +++ b/vendor/k8s.io/api/authorization/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,121 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *LocalSubjectAccessReview) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *LocalSubjectAccessReview) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *LocalSubjectAccessReview) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "authorization.k8s.io", Version: "v1", Kind: "LocalSubjectAccessReview"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *LocalSubjectAccessReview) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *SelfSubjectAccessReview) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *SelfSubjectAccessReview) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *SelfSubjectAccessReview) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "authorization.k8s.io", Version: "v1", Kind: "SelfSubjectAccessReview"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *SelfSubjectAccessReview) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *SelfSubjectRulesReview) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *SelfSubjectRulesReview) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *SelfSubjectRulesReview) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "authorization.k8s.io", Version: "v1", Kind: "SelfSubjectRulesReview"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *SelfSubjectRulesReview) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *SubjectAccessReview) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *SubjectAccessReview) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *SubjectAccessReview) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "authorization.k8s.io", Version: "v1", Kind: "SubjectAccessReview"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *SubjectAccessReview) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/autoscaling/v1/generated.pb.go b/vendor/k8s.io/api/autoscaling/v1/generated.pb.go index 174e6f5f..1e3d8907 100644 --- a/vendor/k8s.io/api/autoscaling/v1/generated.pb.go +++ b/vendor/k8s.io/api/autoscaling/v1/generated.pb.go @@ -45,7 +45,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *CrossVersionObjectReference) Reset() { *m = CrossVersionObjectReference{} } func (*CrossVersionObjectReference) ProtoMessage() {} @@ -5487,6 +5487,7 @@ func (m *ScaleStatus) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -5518,10 +5519,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -5542,55 +5541,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/autoscaling/v2beta1/doc.go b/vendor/k8s.io/api/autoscaling/v2beta1/doc.go index 2cc9f11e..25ca507b 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta1/doc.go +++ b/vendor/k8s.io/api/autoscaling/v2beta1/doc.go @@ -17,5 +17,6 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v2beta1 // import "k8s.io/api/autoscaling/v2beta1" diff --git a/vendor/k8s.io/api/autoscaling/v2beta1/generated.pb.go b/vendor/k8s.io/api/autoscaling/v2beta1/generated.pb.go index 0b6ed381..e129e41b 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta1/generated.pb.go +++ b/vendor/k8s.io/api/autoscaling/v2beta1/generated.pb.go @@ -45,7 +45,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *CrossVersionObjectReference) Reset() { *m = CrossVersionObjectReference{} } func (*CrossVersionObjectReference) ProtoMessage() {} @@ -5012,6 +5012,7 @@ func (m *ResourceMetricStatus) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -5043,10 +5044,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -5067,55 +5066,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/autoscaling/v2beta1/types.go b/vendor/k8s.io/api/autoscaling/v2beta1/types.go index 53a53a3a..d76e8796 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta1/types.go +++ b/vendor/k8s.io/api/autoscaling/v2beta1/types.go @@ -373,6 +373,9 @@ type ExternalMetricStatus struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=autoscaling,v2beta2,HorizontalPodAutoscaler // HorizontalPodAutoscaler is the configuration for a horizontal pod // autoscaler, which automatically manages the replica count of any resource @@ -395,6 +398,9 @@ type HorizontalPodAutoscaler struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=autoscaling,v2beta2,HorizontalPodAutoscalerList // HorizontalPodAutoscaler is a list of horizontal pod autoscaler objects. type HorizontalPodAutoscalerList struct { diff --git a/vendor/k8s.io/api/autoscaling/v2beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/autoscaling/v2beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..f6baef69 --- /dev/null +++ b/vendor/k8s.io/api/autoscaling/v2beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,73 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v2beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *HorizontalPodAutoscaler) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *HorizontalPodAutoscaler) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *HorizontalPodAutoscaler) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "autoscaling", Version: "v2beta2", Kind: "HorizontalPodAutoscaler"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *HorizontalPodAutoscaler) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *HorizontalPodAutoscalerList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *HorizontalPodAutoscalerList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *HorizontalPodAutoscalerList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "autoscaling", Version: "v2beta2", Kind: "HorizontalPodAutoscalerList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *HorizontalPodAutoscalerList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} diff --git a/vendor/k8s.io/api/autoscaling/v2beta2/doc.go b/vendor/k8s.io/api/autoscaling/v2beta2/doc.go index 6d275f6d..76fb0aff 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta2/doc.go +++ b/vendor/k8s.io/api/autoscaling/v2beta2/doc.go @@ -17,5 +17,6 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v2beta2 // import "k8s.io/api/autoscaling/v2beta2" diff --git a/vendor/k8s.io/api/autoscaling/v2beta2/generated.pb.go b/vendor/k8s.io/api/autoscaling/v2beta2/generated.pb.go index 23bc5b98..c69d6cb9 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta2/generated.pb.go +++ b/vendor/k8s.io/api/autoscaling/v2beta2/generated.pb.go @@ -45,7 +45,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *CrossVersionObjectReference) Reset() { *m = CrossVersionObjectReference{} } func (*CrossVersionObjectReference) ProtoMessage() {} @@ -131,10 +131,66 @@ func (m *ExternalMetricStatus) XXX_DiscardUnknown() { var xxx_messageInfo_ExternalMetricStatus proto.InternalMessageInfo +func (m *HPAScalingPolicy) Reset() { *m = HPAScalingPolicy{} } +func (*HPAScalingPolicy) ProtoMessage() {} +func (*HPAScalingPolicy) Descriptor() ([]byte, []int) { + return fileDescriptor_592ad94d7d6be24f, []int{3} +} +func (m *HPAScalingPolicy) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *HPAScalingPolicy) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *HPAScalingPolicy) XXX_Merge(src proto.Message) { + xxx_messageInfo_HPAScalingPolicy.Merge(m, src) +} +func (m *HPAScalingPolicy) XXX_Size() int { + return m.Size() +} +func (m *HPAScalingPolicy) XXX_DiscardUnknown() { + xxx_messageInfo_HPAScalingPolicy.DiscardUnknown(m) +} + +var xxx_messageInfo_HPAScalingPolicy proto.InternalMessageInfo + +func (m *HPAScalingRules) Reset() { *m = HPAScalingRules{} } +func (*HPAScalingRules) ProtoMessage() {} +func (*HPAScalingRules) Descriptor() ([]byte, []int) { + return fileDescriptor_592ad94d7d6be24f, []int{4} +} +func (m *HPAScalingRules) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *HPAScalingRules) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *HPAScalingRules) XXX_Merge(src proto.Message) { + xxx_messageInfo_HPAScalingRules.Merge(m, src) +} +func (m *HPAScalingRules) XXX_Size() int { + return m.Size() +} +func (m *HPAScalingRules) XXX_DiscardUnknown() { + xxx_messageInfo_HPAScalingRules.DiscardUnknown(m) +} + +var xxx_messageInfo_HPAScalingRules proto.InternalMessageInfo + func (m *HorizontalPodAutoscaler) Reset() { *m = HorizontalPodAutoscaler{} } func (*HorizontalPodAutoscaler) ProtoMessage() {} func (*HorizontalPodAutoscaler) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{3} + return fileDescriptor_592ad94d7d6be24f, []int{5} } func (m *HorizontalPodAutoscaler) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -159,10 +215,38 @@ func (m *HorizontalPodAutoscaler) XXX_DiscardUnknown() { var xxx_messageInfo_HorizontalPodAutoscaler proto.InternalMessageInfo +func (m *HorizontalPodAutoscalerBehavior) Reset() { *m = HorizontalPodAutoscalerBehavior{} } +func (*HorizontalPodAutoscalerBehavior) ProtoMessage() {} +func (*HorizontalPodAutoscalerBehavior) Descriptor() ([]byte, []int) { + return fileDescriptor_592ad94d7d6be24f, []int{6} +} +func (m *HorizontalPodAutoscalerBehavior) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *HorizontalPodAutoscalerBehavior) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *HorizontalPodAutoscalerBehavior) XXX_Merge(src proto.Message) { + xxx_messageInfo_HorizontalPodAutoscalerBehavior.Merge(m, src) +} +func (m *HorizontalPodAutoscalerBehavior) XXX_Size() int { + return m.Size() +} +func (m *HorizontalPodAutoscalerBehavior) XXX_DiscardUnknown() { + xxx_messageInfo_HorizontalPodAutoscalerBehavior.DiscardUnknown(m) +} + +var xxx_messageInfo_HorizontalPodAutoscalerBehavior proto.InternalMessageInfo + func (m *HorizontalPodAutoscalerCondition) Reset() { *m = HorizontalPodAutoscalerCondition{} } func (*HorizontalPodAutoscalerCondition) ProtoMessage() {} func (*HorizontalPodAutoscalerCondition) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{4} + return fileDescriptor_592ad94d7d6be24f, []int{7} } func (m *HorizontalPodAutoscalerCondition) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -190,7 +274,7 @@ var xxx_messageInfo_HorizontalPodAutoscalerCondition proto.InternalMessageInfo func (m *HorizontalPodAutoscalerList) Reset() { *m = HorizontalPodAutoscalerList{} } func (*HorizontalPodAutoscalerList) ProtoMessage() {} func (*HorizontalPodAutoscalerList) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{5} + return fileDescriptor_592ad94d7d6be24f, []int{8} } func (m *HorizontalPodAutoscalerList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -218,7 +302,7 @@ var xxx_messageInfo_HorizontalPodAutoscalerList proto.InternalMessageInfo func (m *HorizontalPodAutoscalerSpec) Reset() { *m = HorizontalPodAutoscalerSpec{} } func (*HorizontalPodAutoscalerSpec) ProtoMessage() {} func (*HorizontalPodAutoscalerSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{6} + return fileDescriptor_592ad94d7d6be24f, []int{9} } func (m *HorizontalPodAutoscalerSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -246,7 +330,7 @@ var xxx_messageInfo_HorizontalPodAutoscalerSpec proto.InternalMessageInfo func (m *HorizontalPodAutoscalerStatus) Reset() { *m = HorizontalPodAutoscalerStatus{} } func (*HorizontalPodAutoscalerStatus) ProtoMessage() {} func (*HorizontalPodAutoscalerStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{7} + return fileDescriptor_592ad94d7d6be24f, []int{10} } func (m *HorizontalPodAutoscalerStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -274,7 +358,7 @@ var xxx_messageInfo_HorizontalPodAutoscalerStatus proto.InternalMessageInfo func (m *MetricIdentifier) Reset() { *m = MetricIdentifier{} } func (*MetricIdentifier) ProtoMessage() {} func (*MetricIdentifier) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{8} + return fileDescriptor_592ad94d7d6be24f, []int{11} } func (m *MetricIdentifier) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -302,7 +386,7 @@ var xxx_messageInfo_MetricIdentifier proto.InternalMessageInfo func (m *MetricSpec) Reset() { *m = MetricSpec{} } func (*MetricSpec) ProtoMessage() {} func (*MetricSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{9} + return fileDescriptor_592ad94d7d6be24f, []int{12} } func (m *MetricSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -330,7 +414,7 @@ var xxx_messageInfo_MetricSpec proto.InternalMessageInfo func (m *MetricStatus) Reset() { *m = MetricStatus{} } func (*MetricStatus) ProtoMessage() {} func (*MetricStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{10} + return fileDescriptor_592ad94d7d6be24f, []int{13} } func (m *MetricStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -358,7 +442,7 @@ var xxx_messageInfo_MetricStatus proto.InternalMessageInfo func (m *MetricTarget) Reset() { *m = MetricTarget{} } func (*MetricTarget) ProtoMessage() {} func (*MetricTarget) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{11} + return fileDescriptor_592ad94d7d6be24f, []int{14} } func (m *MetricTarget) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -386,7 +470,7 @@ var xxx_messageInfo_MetricTarget proto.InternalMessageInfo func (m *MetricValueStatus) Reset() { *m = MetricValueStatus{} } func (*MetricValueStatus) ProtoMessage() {} func (*MetricValueStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{12} + return fileDescriptor_592ad94d7d6be24f, []int{15} } func (m *MetricValueStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -414,7 +498,7 @@ var xxx_messageInfo_MetricValueStatus proto.InternalMessageInfo func (m *ObjectMetricSource) Reset() { *m = ObjectMetricSource{} } func (*ObjectMetricSource) ProtoMessage() {} func (*ObjectMetricSource) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{13} + return fileDescriptor_592ad94d7d6be24f, []int{16} } func (m *ObjectMetricSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -442,7 +526,7 @@ var xxx_messageInfo_ObjectMetricSource proto.InternalMessageInfo func (m *ObjectMetricStatus) Reset() { *m = ObjectMetricStatus{} } func (*ObjectMetricStatus) ProtoMessage() {} func (*ObjectMetricStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{14} + return fileDescriptor_592ad94d7d6be24f, []int{17} } func (m *ObjectMetricStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -470,7 +554,7 @@ var xxx_messageInfo_ObjectMetricStatus proto.InternalMessageInfo func (m *PodsMetricSource) Reset() { *m = PodsMetricSource{} } func (*PodsMetricSource) ProtoMessage() {} func (*PodsMetricSource) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{15} + return fileDescriptor_592ad94d7d6be24f, []int{18} } func (m *PodsMetricSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -498,7 +582,7 @@ var xxx_messageInfo_PodsMetricSource proto.InternalMessageInfo func (m *PodsMetricStatus) Reset() { *m = PodsMetricStatus{} } func (*PodsMetricStatus) ProtoMessage() {} func (*PodsMetricStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{16} + return fileDescriptor_592ad94d7d6be24f, []int{19} } func (m *PodsMetricStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -526,7 +610,7 @@ var xxx_messageInfo_PodsMetricStatus proto.InternalMessageInfo func (m *ResourceMetricSource) Reset() { *m = ResourceMetricSource{} } func (*ResourceMetricSource) ProtoMessage() {} func (*ResourceMetricSource) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{17} + return fileDescriptor_592ad94d7d6be24f, []int{20} } func (m *ResourceMetricSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -554,7 +638,7 @@ var xxx_messageInfo_ResourceMetricSource proto.InternalMessageInfo func (m *ResourceMetricStatus) Reset() { *m = ResourceMetricStatus{} } func (*ResourceMetricStatus) ProtoMessage() {} func (*ResourceMetricStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_592ad94d7d6be24f, []int{18} + return fileDescriptor_592ad94d7d6be24f, []int{21} } func (m *ResourceMetricStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -583,7 +667,10 @@ func init() { proto.RegisterType((*CrossVersionObjectReference)(nil), "k8s.io.api.autoscaling.v2beta2.CrossVersionObjectReference") proto.RegisterType((*ExternalMetricSource)(nil), "k8s.io.api.autoscaling.v2beta2.ExternalMetricSource") proto.RegisterType((*ExternalMetricStatus)(nil), "k8s.io.api.autoscaling.v2beta2.ExternalMetricStatus") + proto.RegisterType((*HPAScalingPolicy)(nil), "k8s.io.api.autoscaling.v2beta2.HPAScalingPolicy") + proto.RegisterType((*HPAScalingRules)(nil), "k8s.io.api.autoscaling.v2beta2.HPAScalingRules") proto.RegisterType((*HorizontalPodAutoscaler)(nil), "k8s.io.api.autoscaling.v2beta2.HorizontalPodAutoscaler") + proto.RegisterType((*HorizontalPodAutoscalerBehavior)(nil), "k8s.io.api.autoscaling.v2beta2.HorizontalPodAutoscalerBehavior") proto.RegisterType((*HorizontalPodAutoscalerCondition)(nil), "k8s.io.api.autoscaling.v2beta2.HorizontalPodAutoscalerCondition") proto.RegisterType((*HorizontalPodAutoscalerList)(nil), "k8s.io.api.autoscaling.v2beta2.HorizontalPodAutoscalerList") proto.RegisterType((*HorizontalPodAutoscalerSpec)(nil), "k8s.io.api.autoscaling.v2beta2.HorizontalPodAutoscalerSpec") @@ -606,97 +693,111 @@ func init() { } var fileDescriptor_592ad94d7d6be24f = []byte{ - // 1425 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x58, 0xdd, 0x6f, 0x1b, 0xc5, - 0x16, 0xcf, 0xda, 0x8e, 0x93, 0x8e, 0xd3, 0x24, 0x9d, 0x5b, 0xb5, 0x56, 0xaa, 0x6b, 0x47, 0xab, - 0xab, 0xab, 0x52, 0xd1, 0x35, 0x31, 0xe1, 0x43, 0x42, 0x48, 0xc4, 0x01, 0xda, 0x8a, 0xa4, 0x2d, - 0x93, 0xb4, 0x42, 0xa8, 0x45, 0x8c, 0x77, 0x4f, 0xdc, 0x21, 0xde, 0x5d, 0x6b, 0x76, 0x6c, 0x35, - 0x45, 0x42, 0xbc, 0xf0, 0x8e, 0x40, 0xfc, 0x13, 0x88, 0x17, 0x5e, 0x90, 0x78, 0xe4, 0x43, 0xa8, - 0x42, 0x08, 0xf5, 0xb1, 0x08, 0xc9, 0xa2, 0xe6, 0xbf, 0xe8, 0x13, 0xda, 0x99, 0xd9, 0xf5, 0xae, - 0xed, 0xc4, 0x4e, 0x95, 0x14, 0xf5, 0xcd, 0x33, 0xe7, 0x9c, 0xdf, 0xf9, 0x9c, 0x73, 0xce, 0x1a, - 0x5d, 0xda, 0x7d, 0x35, 0xb0, 0x98, 0x5f, 0xd9, 0x6d, 0xd7, 0x81, 0x7b, 0x20, 0x20, 0xa8, 0x74, - 0xc0, 0x73, 0x7c, 0x5e, 0xd1, 0x04, 0xda, 0x62, 0x15, 0xda, 0x16, 0x7e, 0x60, 0xd3, 0x26, 0xf3, - 0x1a, 0x95, 0x4e, 0xb5, 0x0e, 0x82, 0x56, 0x2b, 0x0d, 0xf0, 0x80, 0x53, 0x01, 0x8e, 0xd5, 0xe2, - 0xbe, 0xf0, 0x71, 0x49, 0xf1, 0x5b, 0xb4, 0xc5, 0xac, 0x04, 0xbf, 0xa5, 0xf9, 0x97, 0x2e, 0x36, - 0x98, 0xb8, 0xd3, 0xae, 0x5b, 0xb6, 0xef, 0x56, 0x1a, 0x7e, 0xc3, 0xaf, 0x48, 0xb1, 0x7a, 0x7b, - 0x47, 0x9e, 0xe4, 0x41, 0xfe, 0x52, 0x70, 0x4b, 0x66, 0x42, 0xbd, 0xed, 0x73, 0xa8, 0x74, 0x56, - 0x06, 0x55, 0x2e, 0xad, 0xf6, 0x79, 0x5c, 0x6a, 0xdf, 0x61, 0x1e, 0xf0, 0xbd, 0x4a, 0x6b, 0xb7, - 0x21, 0x85, 0x38, 0x04, 0x7e, 0x9b, 0xdb, 0x70, 0x28, 0xa9, 0xa0, 0xe2, 0x82, 0xa0, 0xa3, 0x74, - 0x55, 0xf6, 0x93, 0xe2, 0x6d, 0x4f, 0x30, 0x77, 0x58, 0xcd, 0xcb, 0xe3, 0x04, 0x02, 0xfb, 0x0e, - 0xb8, 0x74, 0x50, 0xce, 0xfc, 0xca, 0x40, 0xe7, 0xd6, 0xb9, 0x1f, 0x04, 0x37, 0x81, 0x07, 0xcc, - 0xf7, 0xae, 0xd5, 0x3f, 0x02, 0x5b, 0x10, 0xd8, 0x01, 0x0e, 0x9e, 0x0d, 0x78, 0x19, 0xe5, 0x76, - 0x99, 0xe7, 0x14, 0x8d, 0x65, 0xe3, 0xfc, 0x89, 0xda, 0xdc, 0xfd, 0x6e, 0x79, 0xaa, 0xd7, 0x2d, - 0xe7, 0xde, 0x61, 0x9e, 0x43, 0x24, 0x25, 0xe4, 0xf0, 0xa8, 0x0b, 0xc5, 0x4c, 0x9a, 0xe3, 0x2a, - 0x75, 0x81, 0x48, 0x0a, 0xae, 0x22, 0x44, 0x5b, 0x4c, 0x2b, 0x28, 0x66, 0x25, 0x1f, 0xd6, 0x7c, - 0x68, 0xed, 0xfa, 0x15, 0x4d, 0x21, 0x09, 0x2e, 0xf3, 0x17, 0x03, 0x9d, 0x7e, 0xeb, 0xae, 0x00, - 0xee, 0xd1, 0xe6, 0x26, 0x08, 0xce, 0xec, 0x2d, 0x19, 0x5f, 0xfc, 0x1e, 0xca, 0xbb, 0xf2, 0x2c, - 0x4d, 0x2a, 0x54, 0x5f, 0xb0, 0x0e, 0xae, 0x04, 0x4b, 0x49, 0x5f, 0x71, 0xc0, 0x13, 0x6c, 0x87, - 0x01, 0xaf, 0xcd, 0x6b, 0xd5, 0x79, 0x45, 0x21, 0x1a, 0x0f, 0x6f, 0xa3, 0xbc, 0xa0, 0xbc, 0x01, - 0x42, 0xba, 0x52, 0xa8, 0x3e, 0x3f, 0x19, 0xf2, 0xb6, 0x94, 0xe9, 0xa3, 0xaa, 0x33, 0xd1, 0x58, - 0xe6, 0xef, 0xc3, 0x8e, 0x08, 0x2a, 0xda, 0xc1, 0x31, 0x3a, 0x72, 0x0b, 0xcd, 0xd8, 0x6d, 0xce, - 0xc1, 0x8b, 0x3c, 0x59, 0x99, 0x0c, 0xfa, 0x26, 0x6d, 0xb6, 0x41, 0x59, 0x57, 0x5b, 0xd0, 0xd8, - 0x33, 0xeb, 0x0a, 0x89, 0x44, 0x90, 0xe6, 0x0f, 0x19, 0x74, 0xf6, 0xb2, 0xcf, 0xd9, 0x3d, 0xdf, - 0x13, 0xb4, 0x79, 0xdd, 0x77, 0xd6, 0x34, 0x20, 0x70, 0xfc, 0x21, 0x9a, 0x0d, 0x2b, 0xda, 0xa1, - 0x82, 0x8e, 0xf0, 0x2a, 0x2e, 0x4c, 0xab, 0xb5, 0xdb, 0x08, 0x2f, 0x02, 0x2b, 0xe4, 0xb6, 0x3a, - 0x2b, 0x96, 0x2a, 0xbb, 0x4d, 0x10, 0xb4, 0x5f, 0x19, 0xfd, 0x3b, 0x12, 0xa3, 0xe2, 0xdb, 0x28, - 0x17, 0xb4, 0xc0, 0xd6, 0x8e, 0xbd, 0x36, 0xce, 0xb1, 0x7d, 0x0c, 0xdd, 0x6a, 0x81, 0xdd, 0x2f, - 0xd5, 0xf0, 0x44, 0x24, 0x2c, 0x06, 0x94, 0x0f, 0x64, 0x00, 0x64, 0x99, 0x16, 0xaa, 0xaf, 0x3f, - 0xa9, 0x02, 0x15, 0xc5, 0x38, 0x43, 0xea, 0x4c, 0x34, 0xb8, 0xf9, 0x59, 0x16, 0x2d, 0xef, 0x23, - 0xb9, 0xee, 0x7b, 0x0e, 0x13, 0xcc, 0xf7, 0xf0, 0x65, 0x94, 0x13, 0x7b, 0x2d, 0xd0, 0x4f, 0x6f, - 0x35, 0xb2, 0x76, 0x7b, 0xaf, 0x05, 0x8f, 0xbb, 0xe5, 0xff, 0x8d, 0x93, 0x0f, 0xf9, 0x88, 0x44, - 0xc0, 0x1b, 0xb1, 0x57, 0x99, 0x14, 0x96, 0x36, 0xeb, 0x71, 0xb7, 0x3c, 0xa2, 0xff, 0x59, 0x31, - 0x52, 0xda, 0x78, 0xdc, 0x41, 0xb8, 0x49, 0x03, 0xb1, 0xcd, 0xa9, 0x17, 0x28, 0x4d, 0xcc, 0x05, - 0x1d, 0xaf, 0x0b, 0x93, 0xa5, 0x3b, 0x94, 0xa8, 0x2d, 0x69, 0x2b, 0xf0, 0xc6, 0x10, 0x1a, 0x19, - 0xa1, 0x01, 0xff, 0x1f, 0xe5, 0x39, 0xd0, 0xc0, 0xf7, 0x8a, 0x39, 0xe9, 0x45, 0x1c, 0x5c, 0x22, - 0x6f, 0x89, 0xa6, 0xe2, 0xe7, 0xd0, 0x8c, 0x0b, 0x41, 0x40, 0x1b, 0x50, 0x9c, 0x96, 0x8c, 0x71, - 0x2d, 0x6f, 0xaa, 0x6b, 0x12, 0xd1, 0xcd, 0x3f, 0x0c, 0x74, 0x6e, 0x9f, 0x38, 0x6e, 0xb0, 0x40, - 0xe0, 0x5b, 0x43, 0xf5, 0x6c, 0x4d, 0xe6, 0x60, 0x28, 0x2d, 0xab, 0x79, 0x51, 0xeb, 0x9e, 0x8d, - 0x6e, 0x12, 0xb5, 0x7c, 0x0b, 0x4d, 0x33, 0x01, 0x6e, 0x98, 0x95, 0xec, 0xf9, 0x42, 0xf5, 0x95, - 0x27, 0xac, 0xb5, 0xda, 0x49, 0xad, 0x63, 0xfa, 0x4a, 0x88, 0x46, 0x14, 0xa8, 0xf9, 0x67, 0x66, - 0x5f, 0xdf, 0xc2, 0x82, 0xc7, 0x1f, 0xa3, 0x79, 0x79, 0xd2, 0xfd, 0x0a, 0x76, 0xb4, 0x87, 0x63, - 0xdf, 0xd4, 0x01, 0xe3, 0xa2, 0x76, 0x46, 0x9b, 0x32, 0xbf, 0x95, 0x82, 0x26, 0x03, 0xaa, 0xf0, - 0x0a, 0x2a, 0xb8, 0xcc, 0x23, 0xd0, 0x6a, 0x32, 0x9b, 0xaa, 0xb2, 0x9c, 0xae, 0x2d, 0xf4, 0xba, - 0xe5, 0xc2, 0x66, 0xff, 0x9a, 0x24, 0x79, 0xf0, 0x4b, 0xa8, 0xe0, 0xd2, 0xbb, 0xb1, 0x48, 0x56, - 0x8a, 0xfc, 0x47, 0xeb, 0x2b, 0x6c, 0xf6, 0x49, 0x24, 0xc9, 0x87, 0x6f, 0x84, 0xd5, 0x10, 0x76, - 0xb7, 0xa0, 0x98, 0x93, 0x61, 0xbe, 0x30, 0x59, 0x33, 0x94, 0x2d, 0x22, 0x51, 0x39, 0x12, 0x82, - 0x44, 0x58, 0xe6, 0x77, 0x39, 0xf4, 0xdf, 0x03, 0xdf, 0x3e, 0x7e, 0x1b, 0x61, 0xbf, 0x1e, 0x00, - 0xef, 0x80, 0x73, 0x49, 0x0d, 0xdd, 0x70, 0xfa, 0x85, 0x31, 0xce, 0xd6, 0xce, 0x84, 0x65, 0x7f, - 0x6d, 0x88, 0x4a, 0x46, 0x48, 0x60, 0x1b, 0x9d, 0x0c, 0x1f, 0x83, 0x0a, 0x28, 0xd3, 0x83, 0xf6, - 0x70, 0x2f, 0xed, 0x54, 0xaf, 0x5b, 0x3e, 0xb9, 0x91, 0x04, 0x21, 0x69, 0x4c, 0xbc, 0x86, 0x16, - 0x74, 0x7f, 0x1f, 0x08, 0xf0, 0x59, 0x1d, 0x81, 0x85, 0xf5, 0x34, 0x99, 0x0c, 0xf2, 0x87, 0x10, - 0x0e, 0x04, 0x8c, 0x83, 0x13, 0x43, 0xe4, 0xd2, 0x10, 0x6f, 0xa6, 0xc9, 0x64, 0x90, 0x1f, 0x37, - 0xd1, 0xbc, 0x46, 0xd5, 0xf1, 0x2e, 0x4e, 0xcb, 0x94, 0x4d, 0x38, 0x89, 0x75, 0xd3, 0x8d, 0x6b, - 0x70, 0x3d, 0x85, 0x45, 0x06, 0xb0, 0xb1, 0x40, 0xc8, 0x8e, 0x5a, 0x5c, 0x50, 0xcc, 0x4b, 0x4d, - 0x6f, 0x3c, 0xe1, 0x1b, 0x8c, 0x7b, 0x65, 0x7f, 0x7c, 0xc5, 0x57, 0x01, 0x49, 0xe8, 0x31, 0xbf, - 0x34, 0xd0, 0xe2, 0xe0, 0x24, 0x8f, 0x77, 0x28, 0x63, 0xdf, 0x1d, 0xea, 0x36, 0x9a, 0x0d, 0xa0, - 0x09, 0xb6, 0xf0, 0xb9, 0x2e, 0x80, 0x17, 0x27, 0xec, 0x44, 0xb4, 0x0e, 0xcd, 0x2d, 0x2d, 0x5a, - 0x9b, 0x0b, 0x5b, 0x51, 0x74, 0x22, 0x31, 0xa4, 0xf9, 0x75, 0x16, 0xa1, 0x7e, 0xdd, 0xe3, 0xd5, - 0xd4, 0xe8, 0x59, 0x1e, 0x18, 0x3d, 0x8b, 0xc9, 0x85, 0x2c, 0x31, 0x66, 0x6e, 0xa2, 0xbc, 0x2f, - 0xfb, 0x81, 0xb6, 0xb0, 0x3a, 0x2e, 0x98, 0xf1, 0x84, 0x8f, 0xd1, 0x6a, 0x28, 0x6c, 0xe8, 0xba, - 0xab, 0x68, 0x34, 0x7c, 0x15, 0xe5, 0x5a, 0xbe, 0x13, 0x8d, 0xe4, 0xb1, 0x7b, 0xd2, 0x75, 0xdf, - 0x09, 0x52, 0x98, 0xb3, 0xa1, 0xed, 0xe1, 0x2d, 0x91, 0x38, 0xf8, 0x03, 0x34, 0x1b, 0xad, 0xeb, - 0xb2, 0x44, 0x0b, 0xd5, 0xd5, 0x71, 0x98, 0x44, 0xf3, 0xa7, 0x70, 0x65, 0x30, 0x23, 0x0a, 0x89, - 0x31, 0x43, 0x7c, 0xd0, 0x1b, 0x9f, 0x9c, 0x40, 0x13, 0xe0, 0x8f, 0x5a, 0x75, 0x15, 0x7e, 0x44, - 0x21, 0x31, 0xa6, 0xf9, 0x4d, 0x16, 0xcd, 0xa5, 0x56, 0xc9, 0x7f, 0x23, 0x5d, 0xea, 0xad, 0x1d, - 0x6d, 0xba, 0x14, 0xe6, 0xd1, 0xa7, 0x4b, 0xe1, 0x1e, 0x5f, 0xba, 0x12, 0xf8, 0x23, 0xd2, 0xf5, - 0x53, 0x26, 0x4a, 0x97, 0x9a, 0x7f, 0x93, 0xa5, 0x4b, 0xf1, 0x26, 0xd2, 0x75, 0x0d, 0x4d, 0x77, - 0xc2, 0x05, 0x5d, 0x67, 0xeb, 0xc0, 0x45, 0xc4, 0x8a, 0x9c, 0xb3, 0xde, 0x6d, 0x53, 0x4f, 0x30, - 0xb1, 0x57, 0x3b, 0x11, 0x2e, 0x08, 0x72, 0xc3, 0x27, 0x0a, 0x07, 0x3b, 0x68, 0x8e, 0x76, 0x80, - 0xd3, 0x06, 0xc8, 0x6b, 0x9d, 0xaf, 0xc3, 0xe2, 0x2e, 0xf6, 0xba, 0xe5, 0xb9, 0xb5, 0x04, 0x0e, - 0x49, 0xa1, 0x86, 0x63, 0x50, 0x9f, 0x6f, 0x08, 0xd6, 0x64, 0xf7, 0xd4, 0x18, 0x54, 0x93, 0x41, - 0x8e, 0xc1, 0xb5, 0x21, 0x2a, 0x19, 0x21, 0x61, 0x7e, 0x91, 0x41, 0xa7, 0x86, 0x3e, 0x53, 0xfa, - 0x41, 0x31, 0x8e, 0x29, 0x28, 0x99, 0xa7, 0x18, 0x94, 0xec, 0xa1, 0x83, 0xf2, 0x73, 0x06, 0xe1, - 0xe1, 0x26, 0x8a, 0x3f, 0x91, 0xa3, 0xd8, 0xe6, 0xac, 0x0e, 0x8e, 0x22, 0x1f, 0xc5, 0x6e, 0x97, - 0x9c, 0xe3, 0x49, 0x6c, 0x32, 0xa8, 0xec, 0x78, 0xbe, 0xa4, 0x13, 0x1f, 0xcc, 0xd9, 0xa3, 0xfd, - 0x60, 0x36, 0x7f, 0x1b, 0x0c, 0xe3, 0x33, 0xfd, 0x85, 0x3e, 0x2a, 0xfd, 0xd9, 0xa7, 0x98, 0x7e, - 0xf3, 0x47, 0x03, 0x2d, 0x0e, 0x0e, 0xe1, 0x67, 0xee, 0x7f, 0x9b, 0x5f, 0xd3, 0x4e, 0x3c, 0xdb, - 0xff, 0xd9, 0x7c, 0x6b, 0xa0, 0xd3, 0xa3, 0x56, 0x18, 0xbc, 0x9e, 0x5a, 0x3c, 0x2b, 0xc9, 0xc5, - 0xf3, 0x71, 0xb7, 0x5c, 0x1e, 0xf1, 0xaf, 0x40, 0x04, 0x93, 0xd8, 0x4d, 0x8f, 0x27, 0x01, 0xdf, - 0x0f, 0xdb, 0xac, 0x92, 0x70, 0x24, 0x36, 0x1f, 0x6b, 0xbc, 0x6b, 0x17, 0xef, 0x3f, 0x2a, 0x4d, - 0x3d, 0x78, 0x54, 0x9a, 0x7a, 0xf8, 0xa8, 0x34, 0xf5, 0x69, 0xaf, 0x64, 0xdc, 0xef, 0x95, 0x8c, - 0x07, 0xbd, 0x92, 0xf1, 0xb0, 0x57, 0x32, 0xfe, 0xea, 0x95, 0x8c, 0xcf, 0xff, 0x2e, 0x4d, 0xbd, - 0x3f, 0xa3, 0xa1, 0xff, 0x09, 0x00, 0x00, 0xff, 0xff, 0x7e, 0xa0, 0xce, 0xf5, 0x16, 0x17, 0x00, - 0x00, + // 1657 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x58, 0xdb, 0x6f, 0x1b, 0x45, + 0x17, 0xcf, 0xda, 0xce, 0x6d, 0x9c, 0x5b, 0xa7, 0xfd, 0x5a, 0x2b, 0xd5, 0x67, 0x47, 0xfb, 0x55, + 0x1f, 0x50, 0xd1, 0x35, 0x31, 0x01, 0x2a, 0x55, 0x08, 0xe2, 0x14, 0xda, 0xaa, 0x49, 0x1b, 0xc6, + 0x69, 0x40, 0x28, 0xad, 0x18, 0xef, 0x4e, 0x9c, 0x21, 0xf6, 0xae, 0xb5, 0xb3, 0x76, 0x9b, 0x22, + 0x21, 0x5e, 0x78, 0x47, 0x20, 0x5e, 0xf9, 0x03, 0x10, 0x42, 0xe2, 0x05, 0x89, 0x47, 0x2e, 0xaa, + 0x2a, 0x84, 0x50, 0xdf, 0x28, 0x2f, 0x16, 0x35, 0xff, 0x45, 0x9e, 0xd0, 0x5c, 0x76, 0xbd, 0xbb, + 0x76, 0x62, 0x27, 0x4a, 0x8a, 0xfa, 0xb6, 0x33, 0xe7, 0x9c, 0xdf, 0x99, 0x39, 0xf7, 0x59, 0x70, + 0x65, 0xfb, 0x22, 0x33, 0xa8, 0x93, 0xdf, 0x6e, 0x94, 0x89, 0x6b, 0x13, 0x8f, 0xb0, 0x7c, 0x93, + 0xd8, 0x96, 0xe3, 0xe6, 0x15, 0x01, 0xd7, 0x69, 0x1e, 0x37, 0x3c, 0x87, 0x99, 0xb8, 0x4a, 0xed, + 0x4a, 0xbe, 0x59, 0x28, 0x13, 0x0f, 0x17, 0xf2, 0x15, 0x62, 0x13, 0x17, 0x7b, 0xc4, 0x32, 0xea, + 0xae, 0xe3, 0x39, 0x30, 0x2b, 0xf9, 0x0d, 0x5c, 0xa7, 0x46, 0x88, 0xdf, 0x50, 0xfc, 0xb3, 0x17, + 0x2a, 0xd4, 0xdb, 0x6a, 0x94, 0x0d, 0xd3, 0xa9, 0xe5, 0x2b, 0x4e, 0xc5, 0xc9, 0x0b, 0xb1, 0x72, + 0x63, 0x53, 0xac, 0xc4, 0x42, 0x7c, 0x49, 0xb8, 0x59, 0x3d, 0xa4, 0xde, 0x74, 0x5c, 0x92, 0x6f, + 0xce, 0xc7, 0x55, 0xce, 0x2e, 0x74, 0x78, 0x6a, 0xd8, 0xdc, 0xa2, 0x36, 0x71, 0x77, 0xf2, 0xf5, + 0xed, 0x8a, 0x10, 0x72, 0x09, 0x73, 0x1a, 0xae, 0x49, 0x0e, 0x24, 0xc5, 0xf2, 0x35, 0xe2, 0xe1, + 0x5e, 0xba, 0xf2, 0x7b, 0x49, 0xb9, 0x0d, 0xdb, 0xa3, 0xb5, 0x6e, 0x35, 0xaf, 0xf6, 0x13, 0x60, + 0xe6, 0x16, 0xa9, 0xe1, 0xb8, 0x9c, 0xfe, 0xa5, 0x06, 0xce, 0x2e, 0xb9, 0x0e, 0x63, 0xeb, 0xc4, + 0x65, 0xd4, 0xb1, 0x6f, 0x96, 0x3f, 0x24, 0xa6, 0x87, 0xc8, 0x26, 0x71, 0x89, 0x6d, 0x12, 0x38, + 0x07, 0x52, 0xdb, 0xd4, 0xb6, 0x32, 0xda, 0x9c, 0xf6, 0xfc, 0x78, 0x71, 0xe2, 0x61, 0x2b, 0x37, + 0xd4, 0x6e, 0xe5, 0x52, 0xd7, 0xa9, 0x6d, 0x21, 0x41, 0xe1, 0x1c, 0x36, 0xae, 0x91, 0x4c, 0x22, + 0xca, 0x71, 0x03, 0xd7, 0x08, 0x12, 0x14, 0x58, 0x00, 0x00, 0xd7, 0xa9, 0x52, 0x90, 0x49, 0x0a, + 0x3e, 0xa8, 0xf8, 0xc0, 0xe2, 0xea, 0x35, 0x45, 0x41, 0x21, 0x2e, 0xfd, 0x81, 0x06, 0x4e, 0xbd, + 0x75, 0xcf, 0x23, 0xae, 0x8d, 0xab, 0x2b, 0xc4, 0x73, 0xa9, 0x59, 0x12, 0xf6, 0x85, 0xef, 0x81, + 0x91, 0x9a, 0x58, 0x8b, 0x23, 0xa5, 0x0b, 0x2f, 0x19, 0xfb, 0x47, 0x82, 0x21, 0xa5, 0xaf, 0x59, + 0xc4, 0xf6, 0xe8, 0x26, 0x25, 0x6e, 0x71, 0x4a, 0xa9, 0x1e, 0x91, 0x14, 0xa4, 0xf0, 0xe0, 0x1a, + 0x18, 0xf1, 0xb0, 0x5b, 0x21, 0x9e, 0xb8, 0x4a, 0xba, 0xf0, 0xe2, 0x60, 0xc8, 0x6b, 0x42, 0xa6, + 0x83, 0x2a, 0xd7, 0x48, 0x61, 0xe9, 0xbf, 0x77, 0x5f, 0xc4, 0xc3, 0x5e, 0x83, 0x1d, 0xe3, 0x45, + 0x36, 0xc0, 0xa8, 0xd9, 0x70, 0x5d, 0x62, 0xfb, 0x37, 0x99, 0x1f, 0x0c, 0x7a, 0x1d, 0x57, 0x1b, + 0x44, 0x9e, 0xae, 0x38, 0xad, 0xb0, 0x47, 0x97, 0x24, 0x12, 0xf2, 0x21, 0xf5, 0x6f, 0x35, 0x30, + 0x73, 0x75, 0x75, 0xb1, 0x24, 0x21, 0x56, 0x9d, 0x2a, 0x35, 0x77, 0xe0, 0x45, 0x90, 0xf2, 0x76, + 0xea, 0x44, 0x85, 0xc9, 0x39, 0x3f, 0x08, 0xd6, 0x76, 0xea, 0x64, 0xb7, 0x95, 0x3b, 0x15, 0xe7, + 0xe7, 0xfb, 0x48, 0x48, 0xc0, 0xff, 0x81, 0xe1, 0x26, 0xd7, 0x2b, 0x8e, 0x3a, 0x5c, 0x9c, 0x54, + 0xa2, 0xc3, 0xe2, 0x30, 0x48, 0xd2, 0xe0, 0x25, 0x30, 0x59, 0x27, 0x2e, 0x75, 0xac, 0x12, 0x31, + 0x1d, 0xdb, 0x62, 0x22, 0x88, 0x86, 0x8b, 0xff, 0x51, 0xcc, 0x93, 0xab, 0x61, 0x22, 0x8a, 0xf2, + 0xea, 0x5f, 0x25, 0xc0, 0x74, 0xe7, 0x00, 0xa8, 0x51, 0x25, 0x0c, 0xde, 0x01, 0xb3, 0xcc, 0xc3, + 0x65, 0x5a, 0xa5, 0xf7, 0xb1, 0x47, 0x1d, 0xfb, 0x5d, 0x6a, 0x5b, 0xce, 0xdd, 0x28, 0x7a, 0xb6, + 0xdd, 0xca, 0xcd, 0x96, 0xf6, 0xe4, 0x42, 0xfb, 0x20, 0xc0, 0xeb, 0x60, 0x82, 0x91, 0x2a, 0x31, + 0x3d, 0x79, 0x5f, 0x65, 0x97, 0xe7, 0xda, 0xad, 0xdc, 0x44, 0x29, 0xb4, 0xbf, 0xdb, 0xca, 0x9d, + 0x8c, 0x18, 0x46, 0x12, 0x51, 0x44, 0x18, 0xde, 0x01, 0x63, 0x75, 0xfe, 0x45, 0x09, 0xcb, 0x24, + 0xe6, 0x92, 0x83, 0xc4, 0x4a, 0xdc, 0xe0, 0xc5, 0x19, 0x65, 0xaa, 0xb1, 0x55, 0x85, 0x84, 0x02, + 0x4c, 0xfd, 0xc7, 0x04, 0x38, 0x73, 0xd5, 0x71, 0xe9, 0x7d, 0xc7, 0xf6, 0x70, 0x75, 0xd5, 0xb1, + 0x16, 0x15, 0x22, 0x71, 0xe1, 0x07, 0x60, 0x8c, 0xd7, 0x28, 0x0b, 0x7b, 0xb8, 0x47, 0x9c, 0x06, + 0xa5, 0xc6, 0xa8, 0x6f, 0x57, 0xf8, 0x06, 0x33, 0x38, 0xb7, 0xd1, 0x9c, 0x37, 0x64, 0x21, 0x59, + 0x21, 0x1e, 0xee, 0xe4, 0x7a, 0x67, 0x0f, 0x05, 0xa8, 0xf0, 0x36, 0x48, 0xb1, 0x3a, 0x31, 0x55, + 0xa8, 0x5e, 0xea, 0x7b, 0xb3, 0xde, 0x07, 0x2d, 0xd5, 0x89, 0xd9, 0x29, 0x3e, 0x7c, 0x85, 0x04, + 0x2c, 0x24, 0x60, 0x84, 0x89, 0x90, 0x16, 0x5e, 0x4d, 0x17, 0x5e, 0x3f, 0xac, 0x02, 0x99, 0x17, + 0x41, 0xce, 0xc9, 0x35, 0x52, 0xe0, 0xfa, 0x1f, 0x1a, 0xc8, 0xed, 0x21, 0x59, 0x24, 0x5b, 0xb8, + 0x49, 0x1d, 0x17, 0xae, 0x83, 0x51, 0xb1, 0x73, 0xab, 0xae, 0x4c, 0x99, 0x1f, 0xdc, 0x8d, 0x22, + 0x6c, 0x8b, 0x69, 0x9e, 0x91, 0x25, 0x89, 0x81, 0x7c, 0x30, 0xb8, 0x01, 0xc6, 0xc5, 0xe7, 0x65, + 0xe7, 0xae, 0xad, 0xcc, 0x78, 0x60, 0xe4, 0xc9, 0x76, 0x2b, 0x37, 0x5e, 0xf2, 0x51, 0x50, 0x07, + 0x50, 0xff, 0x34, 0x09, 0xe6, 0xf6, 0xb8, 0xd9, 0x92, 0x63, 0x5b, 0x94, 0x07, 0x3f, 0xbc, 0x1a, + 0xc9, 0xff, 0x85, 0x58, 0xfe, 0x9f, 0xeb, 0x27, 0x1f, 0xaa, 0x07, 0xcb, 0x81, 0xbf, 0x12, 0x11, + 0x2c, 0x65, 0xf0, 0xdd, 0x56, 0xae, 0x47, 0xaf, 0x36, 0x02, 0xa4, 0xa8, 0x5b, 0x60, 0x13, 0xc0, + 0x2a, 0x66, 0xde, 0x9a, 0x8b, 0x6d, 0x26, 0x35, 0xd1, 0x1a, 0x51, 0x91, 0x70, 0x7e, 0xb0, 0x40, + 0xe6, 0x12, 0xc5, 0x59, 0x75, 0x0a, 0xb8, 0xdc, 0x85, 0x86, 0x7a, 0x68, 0x80, 0xff, 0x07, 0x23, + 0x2e, 0xc1, 0xcc, 0xb1, 0x33, 0x29, 0x71, 0x8b, 0x20, 0x6c, 0x90, 0xd8, 0x45, 0x8a, 0x0a, 0x5f, + 0x00, 0xa3, 0x35, 0xc2, 0x18, 0xae, 0x90, 0xcc, 0xb0, 0x60, 0x0c, 0xea, 0xee, 0x8a, 0xdc, 0x46, + 0x3e, 0x5d, 0xff, 0x53, 0x03, 0x67, 0xf7, 0xb0, 0xe3, 0x32, 0x65, 0x1e, 0xdc, 0xe8, 0xca, 0x54, + 0x63, 0xb0, 0x0b, 0x72, 0x69, 0x91, 0xa7, 0x41, 0x8d, 0xf0, 0x77, 0x42, 0x59, 0xba, 0x01, 0x86, + 0xa9, 0x47, 0x6a, 0x7e, 0x01, 0x7a, 0xed, 0x90, 0x59, 0xd4, 0xa9, 0xef, 0xd7, 0x38, 0x1a, 0x92, + 0xa0, 0xfa, 0x83, 0xe4, 0x9e, 0x77, 0xe3, 0xa9, 0x0c, 0x3f, 0x02, 0x53, 0x62, 0xa5, 0x7a, 0x2b, + 0xd9, 0x54, 0x37, 0xec, 0x5b, 0x2d, 0xf6, 0x19, 0x6d, 0x8a, 0xa7, 0xd5, 0x51, 0xa6, 0x4a, 0x11, + 0x68, 0x14, 0x53, 0x05, 0xe7, 0x41, 0xba, 0x46, 0x6d, 0x44, 0xea, 0x55, 0x6a, 0x62, 0xa6, 0xfa, + 0xd4, 0x74, 0xbb, 0x95, 0x4b, 0xaf, 0x74, 0xb6, 0x51, 0x98, 0x07, 0xbe, 0x02, 0xd2, 0x35, 0x7c, + 0x2f, 0x10, 0x91, 0xfd, 0xe4, 0xa4, 0xd2, 0x97, 0x5e, 0xe9, 0x90, 0x50, 0x98, 0x0f, 0xde, 0xe2, + 0xd1, 0xc0, 0x3b, 0x31, 0xcb, 0xa4, 0x84, 0x99, 0xcf, 0x0f, 0xd6, 0xb8, 0x45, 0xf1, 0x0b, 0x45, + 0x8e, 0x80, 0x40, 0x3e, 0x16, 0xa4, 0x60, 0xac, 0xac, 0x6a, 0x90, 0x88, 0xb2, 0x74, 0xe1, 0x8d, + 0xc3, 0xba, 0x4f, 0xc1, 0x14, 0x27, 0x78, 0x98, 0xf8, 0x2b, 0x14, 0xc0, 0xeb, 0xdf, 0xa7, 0xc0, + 0x7f, 0xf7, 0x2d, 0xa0, 0xf0, 0x6d, 0x00, 0x9d, 0x32, 0x23, 0x6e, 0x93, 0x58, 0x57, 0xe4, 0x2c, + 0xca, 0x87, 0x42, 0xee, 0xce, 0x64, 0xf1, 0x34, 0xcf, 0xb0, 0x9b, 0x5d, 0x54, 0xd4, 0x43, 0x02, + 0x9a, 0x60, 0x92, 0xe7, 0x9d, 0xf4, 0x1d, 0x55, 0xf3, 0xe7, 0xc1, 0x92, 0xfa, 0x04, 0x1f, 0x1d, + 0x96, 0xc3, 0x20, 0x28, 0x8a, 0x09, 0x17, 0xc1, 0xb4, 0x1a, 0x7b, 0x62, 0xbe, 0x3c, 0xa3, 0x8c, + 0x3d, 0xbd, 0x14, 0x25, 0xa3, 0x38, 0x3f, 0x87, 0xb0, 0x08, 0xa3, 0x2e, 0xb1, 0x02, 0x88, 0x54, + 0x14, 0xe2, 0x72, 0x94, 0x8c, 0xe2, 0xfc, 0xb0, 0x0a, 0xa6, 0x14, 0xaa, 0x72, 0x6d, 0x66, 0x58, + 0x44, 0xc7, 0x80, 0x03, 0xaa, 0xea, 0x5c, 0x41, 0xb8, 0x2f, 0x45, 0xb0, 0x50, 0x0c, 0x1b, 0x7a, + 0x00, 0x98, 0x7e, 0x35, 0x65, 0x99, 0x11, 0xa1, 0xe9, 0xcd, 0x43, 0xc6, 0x4b, 0x50, 0x96, 0x3b, + 0x33, 0x40, 0xb0, 0xc5, 0x50, 0x48, 0x8f, 0xfe, 0x85, 0x06, 0x66, 0xe2, 0x03, 0x6e, 0xf0, 0xb4, + 0xd0, 0xf6, 0x7c, 0x5a, 0xdc, 0x06, 0x63, 0x72, 0x54, 0x72, 0x5c, 0x15, 0x00, 0x2f, 0x0f, 0x58, + 0xf4, 0x70, 0x99, 0x54, 0x4b, 0x4a, 0x54, 0x86, 0xb3, 0xbf, 0x42, 0x01, 0xa4, 0xfe, 0x75, 0x12, + 0x80, 0x4e, 0x8a, 0xc1, 0x85, 0x48, 0x97, 0x9b, 0x8b, 0x75, 0xb9, 0x99, 0xf0, 0x3b, 0x25, 0xd4, + 0xd1, 0xd6, 0xc1, 0x88, 0x23, 0x4a, 0x8f, 0x3a, 0x61, 0xa1, 0x9f, 0x31, 0x83, 0x31, 0x29, 0x40, + 0x2b, 0x02, 0xde, 0x3b, 0x54, 0x01, 0x53, 0x68, 0xf0, 0x06, 0x48, 0xd5, 0x1d, 0xcb, 0x9f, 0x6b, + 0xfa, 0x8e, 0x84, 0xab, 0x8e, 0xc5, 0x22, 0x98, 0x63, 0xfc, 0xec, 0x7c, 0x17, 0x09, 0x1c, 0x3e, + 0x66, 0xfa, 0xaf, 0x58, 0x11, 0xa2, 0xe9, 0xc2, 0x42, 0x3f, 0x4c, 0xa4, 0xf8, 0x23, 0xb8, 0xc2, + 0x98, 0x3e, 0x05, 0x05, 0x98, 0x1c, 0x9f, 0xa8, 0x87, 0x90, 0x2a, 0x43, 0x7d, 0xf1, 0x7b, 0xbd, + 0x00, 0x25, 0xbe, 0x4f, 0x41, 0x01, 0xa6, 0xfe, 0x4d, 0x12, 0x4c, 0x44, 0x5e, 0x58, 0xff, 0x86, + 0xbb, 0x64, 0xae, 0x1d, 0xad, 0xbb, 0x24, 0xe6, 0xd1, 0xbb, 0x4b, 0xe2, 0x1e, 0x9f, 0xbb, 0x42, + 0xf8, 0x3d, 0xdc, 0xf5, 0x73, 0xc2, 0x77, 0x97, 0x6c, 0xb5, 0x83, 0xb9, 0x4b, 0xf2, 0x86, 0xdc, + 0x75, 0x33, 0xfc, 0x7e, 0xec, 0x33, 0xf3, 0x18, 0xfe, 0xe5, 0x8c, 0x77, 0x1a, 0xd8, 0xf6, 0xa8, + 0xb7, 0x53, 0x1c, 0xef, 0x7a, 0x6b, 0x5a, 0x60, 0x02, 0x37, 0x89, 0x8b, 0x2b, 0x44, 0x6c, 0x2b, + 0x7f, 0x1d, 0x14, 0x77, 0x86, 0x3f, 0xf5, 0x16, 0x43, 0x38, 0x28, 0x82, 0xca, 0xdb, 0xa0, 0x5a, + 0xdf, 0xf2, 0x82, 0x37, 0xa4, 0xea, 0x0c, 0xa2, 0x0d, 0x2e, 0x76, 0x51, 0x51, 0x0f, 0x09, 0xfd, + 0xf3, 0x04, 0x38, 0xd1, 0xf5, 0x7a, 0xef, 0x18, 0x45, 0x3b, 0x26, 0xa3, 0x24, 0x9e, 0xa2, 0x51, + 0x92, 0x07, 0x36, 0xca, 0x2f, 0x09, 0x00, 0xbb, 0x8b, 0x28, 0xfc, 0x58, 0xb4, 0x62, 0xd3, 0xa5, + 0x65, 0x62, 0x49, 0xf2, 0x51, 0x8c, 0x91, 0xe1, 0x3e, 0x1e, 0xc6, 0x46, 0x71, 0x65, 0xc7, 0xf3, + 0x83, 0x29, 0xf4, 0x1f, 0x29, 0x79, 0xb4, 0xff, 0x91, 0xf4, 0xdf, 0xe2, 0x66, 0x7c, 0xa6, 0x7f, + 0x5c, 0xf5, 0x72, 0x7f, 0xf2, 0x29, 0xba, 0x5f, 0xff, 0x49, 0x03, 0x33, 0xf1, 0x26, 0xfc, 0xcc, + 0xfd, 0xce, 0xfc, 0x35, 0x7a, 0x89, 0x67, 0xfb, 0x57, 0xe6, 0x77, 0x1a, 0x38, 0xd5, 0x6b, 0x84, + 0x81, 0x4b, 0x91, 0xc1, 0x33, 0x1f, 0x1e, 0x3c, 0x77, 0x5b, 0xb9, 0x5c, 0x8f, 0x1f, 0x10, 0x3e, + 0x4c, 0x68, 0x36, 0x3d, 0x1e, 0x07, 0xfc, 0xd0, 0x7d, 0x66, 0xe9, 0x84, 0x23, 0x39, 0xf3, 0xb1, + 0xda, 0xbb, 0x78, 0xe1, 0xe1, 0x93, 0xec, 0xd0, 0xa3, 0x27, 0xd9, 0xa1, 0xc7, 0x4f, 0xb2, 0x43, + 0x9f, 0xb4, 0xb3, 0xda, 0xc3, 0x76, 0x56, 0x7b, 0xd4, 0xce, 0x6a, 0x8f, 0xdb, 0x59, 0xed, 0xaf, + 0x76, 0x56, 0xfb, 0xec, 0xef, 0xec, 0xd0, 0xfb, 0xa3, 0x0a, 0xfa, 0x9f, 0x00, 0x00, 0x00, 0xff, + 0xff, 0x79, 0xae, 0x08, 0x04, 0x2d, 0x1a, 0x00, 0x00, } func (m *CrossVersionObjectReference) Marshal() (dAtA []byte, err error) { @@ -823,6 +924,89 @@ func (m *ExternalMetricStatus) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } +func (m *HPAScalingPolicy) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *HPAScalingPolicy) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *HPAScalingPolicy) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i = encodeVarintGenerated(dAtA, i, uint64(m.PeriodSeconds)) + i-- + dAtA[i] = 0x18 + i = encodeVarintGenerated(dAtA, i, uint64(m.Value)) + i-- + dAtA[i] = 0x10 + i -= len(m.Type) + copy(dAtA[i:], m.Type) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Type))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *HPAScalingRules) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *HPAScalingRules) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *HPAScalingRules) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.StabilizationWindowSeconds != nil { + i = encodeVarintGenerated(dAtA, i, uint64(*m.StabilizationWindowSeconds)) + i-- + dAtA[i] = 0x18 + } + if len(m.Policies) > 0 { + for iNdEx := len(m.Policies) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Policies[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + if m.SelectPolicy != nil { + i -= len(*m.SelectPolicy) + copy(dAtA[i:], *m.SelectPolicy) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.SelectPolicy))) + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil +} + func (m *HorizontalPodAutoscaler) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -876,6 +1060,53 @@ func (m *HorizontalPodAutoscaler) MarshalToSizedBuffer(dAtA []byte) (int, error) return len(dAtA) - i, nil } +func (m *HorizontalPodAutoscalerBehavior) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *HorizontalPodAutoscalerBehavior) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *HorizontalPodAutoscalerBehavior) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.ScaleDown != nil { + { + size, err := m.ScaleDown.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + if m.ScaleUp != nil { + { + size, err := m.ScaleUp.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil +} + func (m *HorizontalPodAutoscalerCondition) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -996,6 +1227,18 @@ func (m *HorizontalPodAutoscalerSpec) MarshalToSizedBuffer(dAtA []byte) (int, er _ = i var l int _ = l + if m.Behavior != nil { + { + size, err := m.Behavior.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x2a + } if len(m.Metrics) > 0 { for iNdEx := len(m.Metrics) - 1; iNdEx >= 0; iNdEx-- { { @@ -1726,6 +1969,41 @@ func (m *ExternalMetricStatus) Size() (n int) { return n } +func (m *HPAScalingPolicy) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Type) + n += 1 + l + sovGenerated(uint64(l)) + n += 1 + sovGenerated(uint64(m.Value)) + n += 1 + sovGenerated(uint64(m.PeriodSeconds)) + return n +} + +func (m *HPAScalingRules) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.SelectPolicy != nil { + l = len(*m.SelectPolicy) + n += 1 + l + sovGenerated(uint64(l)) + } + if len(m.Policies) > 0 { + for _, e := range m.Policies { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if m.StabilizationWindowSeconds != nil { + n += 1 + sovGenerated(uint64(*m.StabilizationWindowSeconds)) + } + return n +} + func (m *HorizontalPodAutoscaler) Size() (n int) { if m == nil { return 0 @@ -1741,6 +2019,23 @@ func (m *HorizontalPodAutoscaler) Size() (n int) { return n } +func (m *HorizontalPodAutoscalerBehavior) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.ScaleUp != nil { + l = m.ScaleUp.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + if m.ScaleDown != nil { + l = m.ScaleDown.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + func (m *HorizontalPodAutoscalerCondition) Size() (n int) { if m == nil { return 0 @@ -1795,6 +2090,10 @@ func (m *HorizontalPodAutoscalerSpec) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) } } + if m.Behavior != nil { + l = m.Behavior.Size() + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -2061,42 +2360,82 @@ func (this *ExternalMetricStatus) String() string { }, "") return s } -func (this *HorizontalPodAutoscaler) String() string { +func (this *HPAScalingPolicy) String() string { if this == nil { return "nil" } - s := strings.Join([]string{`&HorizontalPodAutoscaler{`, - `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, - `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "HorizontalPodAutoscalerSpec", "HorizontalPodAutoscalerSpec", 1), `&`, ``, 1) + `,`, - `Status:` + strings.Replace(strings.Replace(this.Status.String(), "HorizontalPodAutoscalerStatus", "HorizontalPodAutoscalerStatus", 1), `&`, ``, 1) + `,`, + s := strings.Join([]string{`&HPAScalingPolicy{`, + `Type:` + fmt.Sprintf("%v", this.Type) + `,`, + `Value:` + fmt.Sprintf("%v", this.Value) + `,`, + `PeriodSeconds:` + fmt.Sprintf("%v", this.PeriodSeconds) + `,`, `}`, }, "") return s } -func (this *HorizontalPodAutoscalerCondition) String() string { +func (this *HPAScalingRules) String() string { if this == nil { return "nil" } - s := strings.Join([]string{`&HorizontalPodAutoscalerCondition{`, - `Type:` + fmt.Sprintf("%v", this.Type) + `,`, - `Status:` + fmt.Sprintf("%v", this.Status) + `,`, - `LastTransitionTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastTransitionTime), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, - `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`, - `Message:` + fmt.Sprintf("%v", this.Message) + `,`, + repeatedStringForPolicies := "[]HPAScalingPolicy{" + for _, f := range this.Policies { + repeatedStringForPolicies += strings.Replace(strings.Replace(f.String(), "HPAScalingPolicy", "HPAScalingPolicy", 1), `&`, ``, 1) + "," + } + repeatedStringForPolicies += "}" + s := strings.Join([]string{`&HPAScalingRules{`, + `SelectPolicy:` + valueToStringGenerated(this.SelectPolicy) + `,`, + `Policies:` + repeatedStringForPolicies + `,`, + `StabilizationWindowSeconds:` + valueToStringGenerated(this.StabilizationWindowSeconds) + `,`, `}`, }, "") return s } -func (this *HorizontalPodAutoscalerList) String() string { +func (this *HorizontalPodAutoscaler) String() string { if this == nil { return "nil" } - repeatedStringForItems := "[]HorizontalPodAutoscaler{" - for _, f := range this.Items { - repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "HorizontalPodAutoscaler", "HorizontalPodAutoscaler", 1), `&`, ``, 1) + "," - } - repeatedStringForItems += "}" - s := strings.Join([]string{`&HorizontalPodAutoscalerList{`, + s := strings.Join([]string{`&HorizontalPodAutoscaler{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "HorizontalPodAutoscalerSpec", "HorizontalPodAutoscalerSpec", 1), `&`, ``, 1) + `,`, + `Status:` + strings.Replace(strings.Replace(this.Status.String(), "HorizontalPodAutoscalerStatus", "HorizontalPodAutoscalerStatus", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *HorizontalPodAutoscalerBehavior) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&HorizontalPodAutoscalerBehavior{`, + `ScaleUp:` + strings.Replace(this.ScaleUp.String(), "HPAScalingRules", "HPAScalingRules", 1) + `,`, + `ScaleDown:` + strings.Replace(this.ScaleDown.String(), "HPAScalingRules", "HPAScalingRules", 1) + `,`, + `}`, + }, "") + return s +} +func (this *HorizontalPodAutoscalerCondition) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&HorizontalPodAutoscalerCondition{`, + `Type:` + fmt.Sprintf("%v", this.Type) + `,`, + `Status:` + fmt.Sprintf("%v", this.Status) + `,`, + `LastTransitionTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastTransitionTime), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`, + `Message:` + fmt.Sprintf("%v", this.Message) + `,`, + `}`, + }, "") + return s +} +func (this *HorizontalPodAutoscalerList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]HorizontalPodAutoscaler{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "HorizontalPodAutoscaler", "HorizontalPodAutoscaler", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&HorizontalPodAutoscalerList{`, `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, `Items:` + repeatedStringForItems + `,`, `}`, @@ -2117,6 +2456,7 @@ func (this *HorizontalPodAutoscalerSpec) String() string { `MinReplicas:` + valueToStringGenerated(this.MinReplicas) + `,`, `MaxReplicas:` + fmt.Sprintf("%v", this.MaxReplicas) + `,`, `Metrics:` + repeatedStringForMetrics + `,`, + `Behavior:` + strings.Replace(this.Behavior.String(), "HorizontalPodAutoscalerBehavior", "HorizontalPodAutoscalerBehavior", 1) + `,`, `}`, }, "") return s @@ -2271,22 +2611,409 @@ func (this *ResourceMetricStatus) String() string { if this == nil { return "nil" } - s := strings.Join([]string{`&ResourceMetricStatus{`, - `Name:` + fmt.Sprintf("%v", this.Name) + `,`, - `Current:` + strings.Replace(strings.Replace(this.Current.String(), "MetricValueStatus", "MetricValueStatus", 1), `&`, ``, 1) + `,`, - `}`, - }, "") - return s -} -func valueToStringGenerated(v interface{}) string { - rv := reflect.ValueOf(v) - if rv.IsNil() { - return "nil" + s := strings.Join([]string{`&ResourceMetricStatus{`, + `Name:` + fmt.Sprintf("%v", this.Name) + `,`, + `Current:` + strings.Replace(strings.Replace(this.Current.String(), "MetricValueStatus", "MetricValueStatus", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func valueToStringGenerated(v interface{}) string { + rv := reflect.ValueOf(v) + if rv.IsNil() { + return "nil" + } + pv := reflect.Indirect(rv).Interface() + return fmt.Sprintf("*%v", pv) +} +func (m *CrossVersionObjectReference) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CrossVersionObjectReference: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CrossVersionObjectReference: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Kind", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Kind = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Name", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Name = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field APIVersion", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.APIVersion = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *ExternalMetricSource) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: ExternalMetricSource: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: ExternalMetricSource: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Metric", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Metric.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Target", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Target.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *ExternalMetricStatus) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: ExternalMetricStatus: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: ExternalMetricStatus: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Metric", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Metric.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Current", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Current.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF } - pv := reflect.Indirect(rv).Interface() - return fmt.Sprintf("*%v", pv) + return nil } -func (m *CrossVersionObjectReference) Unmarshal(dAtA []byte) error { +func (m *HPAScalingPolicy) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -2309,15 +3036,15 @@ func (m *CrossVersionObjectReference) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: CrossVersionObjectReference: wiretype end group for non-group") + return fmt.Errorf("proto: HPAScalingPolicy: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: CrossVersionObjectReference: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: HPAScalingPolicy: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Kind", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Type", wireType) } var stringLen uint64 for shift := uint(0); ; shift += 7 { @@ -2345,13 +3072,13 @@ func (m *CrossVersionObjectReference) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - m.Kind = string(dAtA[iNdEx:postIndex]) + m.Type = HPAScalingPolicyType(dAtA[iNdEx:postIndex]) iNdEx = postIndex case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Name", wireType) + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field Value", wireType) } - var stringLen uint64 + m.Value = 0 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -2361,29 +3088,16 @@ func (m *CrossVersionObjectReference) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - stringLen |= uint64(b&0x7F) << shift + m.Value |= int32(b&0x7F) << shift if b < 0x80 { break } } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.Name = string(dAtA[iNdEx:postIndex]) - iNdEx = postIndex case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field APIVersion", wireType) + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field PeriodSeconds", wireType) } - var stringLen uint64 + m.PeriodSeconds = 0 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -2393,24 +3107,11 @@ func (m *CrossVersionObjectReference) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - stringLen |= uint64(b&0x7F) << shift + m.PeriodSeconds |= int32(b&0x7F) << shift if b < 0x80 { break } } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.APIVersion = string(dAtA[iNdEx:postIndex]) - iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -2435,7 +3136,7 @@ func (m *CrossVersionObjectReference) Unmarshal(dAtA []byte) error { } return nil } -func (m *ExternalMetricSource) Unmarshal(dAtA []byte) error { +func (m *HPAScalingRules) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -2458,17 +3159,17 @@ func (m *ExternalMetricSource) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: ExternalMetricSource: wiretype end group for non-group") + return fmt.Errorf("proto: HPAScalingRules: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: ExternalMetricSource: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: HPAScalingRules: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Metric", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field SelectPolicy", wireType) } - var msglen int + var stringLen uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -2478,28 +3179,28 @@ func (m *ExternalMetricSource) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - msglen |= int(b&0x7F) << shift + stringLen |= uint64(b&0x7F) << shift if b < 0x80 { break } } - if msglen < 0 { + intStringLen := int(stringLen) + if intStringLen < 0 { return ErrInvalidLengthGenerated } - postIndex := iNdEx + msglen + postIndex := iNdEx + intStringLen if postIndex < 0 { return ErrInvalidLengthGenerated } if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Metric.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } + s := ScalingPolicySelect(dAtA[iNdEx:postIndex]) + m.SelectPolicy = &s iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Target", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Policies", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -2526,10 +3227,31 @@ func (m *ExternalMetricSource) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Target.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + m.Policies = append(m.Policies, HPAScalingPolicy{}) + if err := m.Policies[len(m.Policies)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex + case 3: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field StabilizationWindowSeconds", wireType) + } + var v int32 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int32(b&0x7F) << shift + if b < 0x80 { + break + } + } + m.StabilizationWindowSeconds = &v default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -2554,7 +3276,7 @@ func (m *ExternalMetricSource) Unmarshal(dAtA []byte) error { } return nil } -func (m *ExternalMetricStatus) Unmarshal(dAtA []byte) error { +func (m *HorizontalPodAutoscaler) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -2577,15 +3299,15 @@ func (m *ExternalMetricStatus) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: ExternalMetricStatus: wiretype end group for non-group") + return fmt.Errorf("proto: HorizontalPodAutoscaler: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: ExternalMetricStatus: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: HorizontalPodAutoscaler: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Metric", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -2612,13 +3334,13 @@ func (m *ExternalMetricStatus) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Metric.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Current", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -2645,7 +3367,40 @@ func (m *ExternalMetricStatus) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Current.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex @@ -2673,7 +3428,7 @@ func (m *ExternalMetricStatus) Unmarshal(dAtA []byte) error { } return nil } -func (m *HorizontalPodAutoscaler) Unmarshal(dAtA []byte) error { +func (m *HorizontalPodAutoscalerBehavior) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -2696,15 +3451,15 @@ func (m *HorizontalPodAutoscaler) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: HorizontalPodAutoscaler: wiretype end group for non-group") + return fmt.Errorf("proto: HorizontalPodAutoscalerBehavior: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: HorizontalPodAutoscaler: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: HorizontalPodAutoscalerBehavior: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ScaleUp", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -2731,13 +3486,16 @@ func (m *HorizontalPodAutoscaler) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if m.ScaleUp == nil { + m.ScaleUp = &HPAScalingRules{} + } + if err := m.ScaleUp.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ScaleDown", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -2764,40 +3522,10 @@ func (m *HorizontalPodAutoscaler) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF + if m.ScaleDown == nil { + m.ScaleDown = &HPAScalingRules{} } - if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if err := m.ScaleDown.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex @@ -3294,6 +4022,42 @@ func (m *HorizontalPodAutoscalerSpec) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Behavior", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Behavior == nil { + m.Behavior = &HorizontalPodAutoscalerBehavior{} + } + if err := m.Behavior.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -5215,6 +5979,7 @@ func (m *ResourceMetricStatus) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -5246,10 +6011,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -5270,55 +6033,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/autoscaling/v2beta2/generated.proto b/vendor/k8s.io/api/autoscaling/v2beta2/generated.proto index 80f1d345..24dc5882 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta2/generated.proto +++ b/vendor/k8s.io/api/autoscaling/v2beta2/generated.proto @@ -64,6 +64,47 @@ message ExternalMetricStatus { optional MetricValueStatus current = 2; } +// HPAScalingPolicy is a single policy which must hold true for a specified past interval. +message HPAScalingPolicy { + // Type is used to specify the scaling policy. + optional string type = 1; + + // Value contains the amount of change which is permitted by the policy. + // It must be greater than zero + optional int32 value = 2; + + // PeriodSeconds specifies the window of time for which the policy should hold true. + // PeriodSeconds must be greater than zero and less than or equal to 1800 (30 min). + optional int32 periodSeconds = 3; +} + +// HPAScalingRules configures the scaling behavior for one direction. +// These Rules are applied after calculating DesiredReplicas from metrics for the HPA. +// They can limit the scaling velocity by specifying scaling policies. +// They can prevent flapping by specifying the stabilization window, so that the +// number of replicas is not set instantly, instead, the safest value from the stabilization +// window is chosen. +message HPAScalingRules { + // StabilizationWindowSeconds is the number of seconds for which past recommendations should be + // considered while scaling up or scaling down. + // StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). + // If not set, use the default values: + // - For scale up: 0 (i.e. no stabilization is done). + // - For scale down: 300 (i.e. the stabilization window is 300 seconds long). + // +optional + optional int32 stabilizationWindowSeconds = 3; + + // selectPolicy is used to specify which policy should be used. + // If not set, the default value MaxPolicySelect is used. + // +optional + optional string selectPolicy = 1; + + // policies is a list of potential scaling polices which can be used during scaling. + // At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid + // +optional + repeated HPAScalingPolicy policies = 2; +} + // HorizontalPodAutoscaler is the configuration for a horizontal pod // autoscaler, which automatically manages the replica count of any resource // implementing the scale subresource based on the metrics specified. @@ -83,6 +124,25 @@ message HorizontalPodAutoscaler { optional HorizontalPodAutoscalerStatus status = 3; } +// HorizontalPodAutoscalerBehavior configures the scaling behavior of the target +// in both Up and Down directions (scaleUp and scaleDown fields respectively). +message HorizontalPodAutoscalerBehavior { + // scaleUp is scaling policy for scaling Up. + // If not set, the default value is the higher of: + // * increase no more than 4 pods per 60 seconds + // * double the number of pods per 60 seconds + // No stabilization is used. + // +optional + optional HPAScalingRules scaleUp = 1; + + // scaleDown is scaling policy for scaling Down. + // If not set, the default value is to allow to scale down to minReplicas pods, with a + // 300 second stabilization window (i.e., the highest recommendation for + // the last 300sec is used). + // +optional + optional HPAScalingRules scaleDown = 2; +} + // HorizontalPodAutoscalerCondition describes the state of // a HorizontalPodAutoscaler at a certain point. message HorizontalPodAutoscalerCondition { @@ -145,6 +205,12 @@ message HorizontalPodAutoscalerSpec { // If not set, the default metric will be set to 80% average CPU utilization. // +optional repeated MetricSpec metrics = 4; + + // behavior configures the scaling behavior of the target + // in both Up and Down directions (scaleUp and scaleDown fields respectively). + // If not set, the default HPAScalingRules for scale up and scale down are used. + // +optional + optional HorizontalPodAutoscalerBehavior behavior = 5; } // HorizontalPodAutoscalerStatus describes the current status of a horizontal pod autoscaler. diff --git a/vendor/k8s.io/api/autoscaling/v2beta2/types.go b/vendor/k8s.io/api/autoscaling/v2beta2/types.go index 4480c7da..6e5b8f68 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta2/types.go +++ b/vendor/k8s.io/api/autoscaling/v2beta2/types.go @@ -26,6 +26,8 @@ import ( // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.12 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // HorizontalPodAutoscaler is the configuration for a horizontal pod // autoscaler, which automatically manages the replica count of any resource @@ -72,6 +74,12 @@ type HorizontalPodAutoscalerSpec struct { // If not set, the default metric will be set to 80% average CPU utilization. // +optional Metrics []MetricSpec `json:"metrics,omitempty" protobuf:"bytes,4,rep,name=metrics"` + + // behavior configures the scaling behavior of the target + // in both Up and Down directions (scaleUp and scaleDown fields respectively). + // If not set, the default HPAScalingRules for scale up and scale down are used. + // +optional + Behavior *HorizontalPodAutoscalerBehavior `json:"behavior,omitempty" protobuf:"bytes,5,opt,name=behavior"` } // CrossVersionObjectReference contains enough information to let you identify the referred resource. @@ -117,6 +125,84 @@ type MetricSpec struct { External *ExternalMetricSource `json:"external,omitempty" protobuf:"bytes,5,opt,name=external"` } +// HorizontalPodAutoscalerBehavior configures the scaling behavior of the target +// in both Up and Down directions (scaleUp and scaleDown fields respectively). +type HorizontalPodAutoscalerBehavior struct { + // scaleUp is scaling policy for scaling Up. + // If not set, the default value is the higher of: + // * increase no more than 4 pods per 60 seconds + // * double the number of pods per 60 seconds + // No stabilization is used. + // +optional + ScaleUp *HPAScalingRules `json:"scaleUp,omitempty" protobuf:"bytes,1,opt,name=scaleUp"` + // scaleDown is scaling policy for scaling Down. + // If not set, the default value is to allow to scale down to minReplicas pods, with a + // 300 second stabilization window (i.e., the highest recommendation for + // the last 300sec is used). + // +optional + ScaleDown *HPAScalingRules `json:"scaleDown,omitempty" protobuf:"bytes,2,opt,name=scaleDown"` +} + +// ScalingPolicySelect is used to specify which policy should be used while scaling in a certain direction +type ScalingPolicySelect string + +const ( + // MaxPolicySelect selects the policy with the highest possible change. + MaxPolicySelect ScalingPolicySelect = "Max" + // MinPolicySelect selects the policy with the lowest possible change. + MinPolicySelect ScalingPolicySelect = "Min" + // DisabledPolicySelect disables the scaling in this direction. + DisabledPolicySelect ScalingPolicySelect = "Disabled" +) + +// HPAScalingRules configures the scaling behavior for one direction. +// These Rules are applied after calculating DesiredReplicas from metrics for the HPA. +// They can limit the scaling velocity by specifying scaling policies. +// They can prevent flapping by specifying the stabilization window, so that the +// number of replicas is not set instantly, instead, the safest value from the stabilization +// window is chosen. +type HPAScalingRules struct { + // StabilizationWindowSeconds is the number of seconds for which past recommendations should be + // considered while scaling up or scaling down. + // StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). + // If not set, use the default values: + // - For scale up: 0 (i.e. no stabilization is done). + // - For scale down: 300 (i.e. the stabilization window is 300 seconds long). + // +optional + StabilizationWindowSeconds *int32 `json:"stabilizationWindowSeconds" protobuf:"varint,3,opt,name=stabilizationWindowSeconds"` + // selectPolicy is used to specify which policy should be used. + // If not set, the default value MaxPolicySelect is used. + // +optional + SelectPolicy *ScalingPolicySelect `json:"selectPolicy,omitempty" protobuf:"bytes,1,opt,name=selectPolicy"` + // policies is a list of potential scaling polices which can be used during scaling. + // At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid + // +optional + Policies []HPAScalingPolicy `json:"policies,omitempty" protobuf:"bytes,2,rep,name=policies"` +} + +// HPAScalingPolicyType is the type of the policy which could be used while making scaling decisions. +type HPAScalingPolicyType string + +const ( + // PodsScalingPolicy is a policy used to specify a change in absolute number of pods. + PodsScalingPolicy HPAScalingPolicyType = "Pods" + // PercentScalingPolicy is a policy used to specify a relative amount of change with respect to + // the current number of pods. + PercentScalingPolicy HPAScalingPolicyType = "Percent" +) + +// HPAScalingPolicy is a single policy which must hold true for a specified past interval. +type HPAScalingPolicy struct { + // Type is used to specify the scaling policy. + Type HPAScalingPolicyType `json:"type" protobuf:"bytes,1,opt,name=type,casttype=HPAScalingPolicyType"` + // Value contains the amount of change which is permitted by the policy. + // It must be greater than zero + Value int32 `json:"value" protobuf:"varint,2,opt,name=value"` + // PeriodSeconds specifies the window of time for which the policy should hold true. + // PeriodSeconds must be greater than zero and less than or equal to 1800 (30 min). + PeriodSeconds int32 `json:"periodSeconds" protobuf:"varint,3,opt,name=periodSeconds"` +} + // MetricSourceType indicates the type of metric. type MetricSourceType string @@ -383,6 +469,8 @@ type MetricValueStatus struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.12 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // HorizontalPodAutoscalerList is a list of horizontal pod autoscaler objects. type HorizontalPodAutoscalerList struct { diff --git a/vendor/k8s.io/api/autoscaling/v2beta2/types_swagger_doc_generated.go b/vendor/k8s.io/api/autoscaling/v2beta2/types_swagger_doc_generated.go index bb85b9f0..3f38880f 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta2/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/autoscaling/v2beta2/types_swagger_doc_generated.go @@ -58,6 +58,28 @@ func (ExternalMetricStatus) SwaggerDoc() map[string]string { return map_ExternalMetricStatus } +var map_HPAScalingPolicy = map[string]string{ + "": "HPAScalingPolicy is a single policy which must hold true for a specified past interval.", + "type": "Type is used to specify the scaling policy.", + "value": "Value contains the amount of change which is permitted by the policy. It must be greater than zero", + "periodSeconds": "PeriodSeconds specifies the window of time for which the policy should hold true. PeriodSeconds must be greater than zero and less than or equal to 1800 (30 min).", +} + +func (HPAScalingPolicy) SwaggerDoc() map[string]string { + return map_HPAScalingPolicy +} + +var map_HPAScalingRules = map[string]string{ + "": "HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.", + "stabilizationWindowSeconds": "StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).", + "selectPolicy": "selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.", + "policies": "policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid", +} + +func (HPAScalingRules) SwaggerDoc() map[string]string { + return map_HPAScalingRules +} + var map_HorizontalPodAutoscaler = map[string]string{ "": "HorizontalPodAutoscaler is the configuration for a horizontal pod autoscaler, which automatically manages the replica count of any resource implementing the scale subresource based on the metrics specified.", "metadata": "metadata is the standard object metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", @@ -69,6 +91,16 @@ func (HorizontalPodAutoscaler) SwaggerDoc() map[string]string { return map_HorizontalPodAutoscaler } +var map_HorizontalPodAutoscalerBehavior = map[string]string{ + "": "HorizontalPodAutoscalerBehavior configures the scaling behavior of the target in both Up and Down directions (scaleUp and scaleDown fields respectively).", + "scaleUp": "scaleUp is scaling policy for scaling Up. If not set, the default value is the higher of:\n * increase no more than 4 pods per 60 seconds\n * double the number of pods per 60 seconds\nNo stabilization is used.", + "scaleDown": "scaleDown is scaling policy for scaling Down. If not set, the default value is to allow to scale down to minReplicas pods, with a 300 second stabilization window (i.e., the highest recommendation for the last 300sec is used).", +} + +func (HorizontalPodAutoscalerBehavior) SwaggerDoc() map[string]string { + return map_HorizontalPodAutoscalerBehavior +} + var map_HorizontalPodAutoscalerCondition = map[string]string{ "": "HorizontalPodAutoscalerCondition describes the state of a HorizontalPodAutoscaler at a certain point.", "type": "type describes the current condition", @@ -98,6 +130,7 @@ var map_HorizontalPodAutoscalerSpec = map[string]string{ "minReplicas": "minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. minReplicas is allowed to be 0 if the alpha feature gate HPAScaleToZero is enabled and at least one Object or External metric is configured. Scaling is active as long as at least one metric value is available.", "maxReplicas": "maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas.", "metrics": "metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization.", + "behavior": "behavior configures the scaling behavior of the target in both Up and Down directions (scaleUp and scaleDown fields respectively). If not set, the default HPAScalingRules for scale up and scale down are used.", } func (HorizontalPodAutoscalerSpec) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.deepcopy.go b/vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.deepcopy.go index 2dffa333..ca26fe92 100644 --- a/vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.deepcopy.go @@ -77,6 +77,53 @@ func (in *ExternalMetricStatus) DeepCopy() *ExternalMetricStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *HPAScalingPolicy) DeepCopyInto(out *HPAScalingPolicy) { + *out = *in + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HPAScalingPolicy. +func (in *HPAScalingPolicy) DeepCopy() *HPAScalingPolicy { + if in == nil { + return nil + } + out := new(HPAScalingPolicy) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *HPAScalingRules) DeepCopyInto(out *HPAScalingRules) { + *out = *in + if in.StabilizationWindowSeconds != nil { + in, out := &in.StabilizationWindowSeconds, &out.StabilizationWindowSeconds + *out = new(int32) + **out = **in + } + if in.SelectPolicy != nil { + in, out := &in.SelectPolicy, &out.SelectPolicy + *out = new(ScalingPolicySelect) + **out = **in + } + if in.Policies != nil { + in, out := &in.Policies, &out.Policies + *out = make([]HPAScalingPolicy, len(*in)) + copy(*out, *in) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HPAScalingRules. +func (in *HPAScalingRules) DeepCopy() *HPAScalingRules { + if in == nil { + return nil + } + out := new(HPAScalingRules) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *HorizontalPodAutoscaler) DeepCopyInto(out *HorizontalPodAutoscaler) { *out = *in @@ -105,6 +152,32 @@ func (in *HorizontalPodAutoscaler) DeepCopyObject() runtime.Object { return nil } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *HorizontalPodAutoscalerBehavior) DeepCopyInto(out *HorizontalPodAutoscalerBehavior) { + *out = *in + if in.ScaleUp != nil { + in, out := &in.ScaleUp, &out.ScaleUp + *out = new(HPAScalingRules) + (*in).DeepCopyInto(*out) + } + if in.ScaleDown != nil { + in, out := &in.ScaleDown, &out.ScaleDown + *out = new(HPAScalingRules) + (*in).DeepCopyInto(*out) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HorizontalPodAutoscalerBehavior. +func (in *HorizontalPodAutoscalerBehavior) DeepCopy() *HorizontalPodAutoscalerBehavior { + if in == nil { + return nil + } + out := new(HorizontalPodAutoscalerBehavior) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *HorizontalPodAutoscalerCondition) DeepCopyInto(out *HorizontalPodAutoscalerCondition) { *out = *in @@ -171,6 +244,11 @@ func (in *HorizontalPodAutoscalerSpec) DeepCopyInto(out *HorizontalPodAutoscaler (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.Behavior != nil { + in, out := &in.Behavior, &out.Behavior + *out = new(HorizontalPodAutoscalerBehavior) + (*in).DeepCopyInto(*out) + } return } diff --git a/vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..83926e3a --- /dev/null +++ b/vendor/k8s.io/api/autoscaling/v2beta2/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,57 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v2beta2 + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *HorizontalPodAutoscaler) APILifecycleIntroduced() (major, minor int) { + return 1, 12 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *HorizontalPodAutoscaler) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *HorizontalPodAutoscaler) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *HorizontalPodAutoscalerList) APILifecycleIntroduced() (major, minor int) { + return 1, 12 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *HorizontalPodAutoscalerList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *HorizontalPodAutoscalerList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} diff --git a/vendor/k8s.io/api/batch/v1/generated.pb.go b/vendor/k8s.io/api/batch/v1/generated.pb.go index fb9d21e1..35944e72 100644 --- a/vendor/k8s.io/api/batch/v1/generated.pb.go +++ b/vendor/k8s.io/api/batch/v1/generated.pb.go @@ -43,7 +43,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Job) Reset() { *m = Job{} } func (*Job) ProtoMessage() {} @@ -1771,6 +1771,7 @@ func (m *JobStatus) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1802,10 +1803,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1826,55 +1825,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/batch/v1beta1/doc.go b/vendor/k8s.io/api/batch/v1beta1/doc.go index 258ff028..cb2572f5 100644 --- a/vendor/k8s.io/api/batch/v1beta1/doc.go +++ b/vendor/k8s.io/api/batch/v1beta1/doc.go @@ -17,5 +17,6 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v1beta1 // import "k8s.io/api/batch/v1beta1" diff --git a/vendor/k8s.io/api/batch/v1beta1/generated.pb.go b/vendor/k8s.io/api/batch/v1beta1/generated.pb.go index 837a2f9c..69c4054b 100644 --- a/vendor/k8s.io/api/batch/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/batch/v1beta1/generated.pb.go @@ -43,7 +43,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *CronJob) Reset() { *m = CronJob{} } func (*CronJob) ProtoMessage() {} @@ -1660,6 +1660,7 @@ func (m *JobTemplateSpec) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1691,10 +1692,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1715,55 +1714,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/batch/v1beta1/types.go b/vendor/k8s.io/api/batch/v1beta1/types.go index 2978747a..6f49cc2a 100644 --- a/vendor/k8s.io/api/batch/v1beta1/types.go +++ b/vendor/k8s.io/api/batch/v1beta1/types.go @@ -18,11 +18,13 @@ package v1beta1 import ( batchv1 "k8s.io/api/batch/v1" - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" ) // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // JobTemplate describes a template for creating copies of a predefined pod. type JobTemplate struct { @@ -53,6 +55,8 @@ type JobTemplateSpec struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // CronJob represents the configuration of a single cron job. type CronJob struct { @@ -74,6 +78,8 @@ type CronJob struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // CronJobList is a collection of cron jobs. type CronJobList struct { diff --git a/vendor/k8s.io/api/batch/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/batch/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..63bae5f1 --- /dev/null +++ b/vendor/k8s.io/api/batch/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,75 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CronJob) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CronJob) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CronJob) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CronJobList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CronJobList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CronJobList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *JobTemplate) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *JobTemplate) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *JobTemplate) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} diff --git a/vendor/k8s.io/api/batch/v2alpha1/generated.pb.go b/vendor/k8s.io/api/batch/v2alpha1/generated.pb.go index 8271c841..3e58dbb9 100644 --- a/vendor/k8s.io/api/batch/v2alpha1/generated.pb.go +++ b/vendor/k8s.io/api/batch/v2alpha1/generated.pb.go @@ -43,7 +43,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *CronJob) Reset() { *m = CronJob{} } func (*CronJob) ProtoMessage() {} @@ -1660,6 +1660,7 @@ func (m *JobTemplateSpec) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1691,10 +1692,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1715,55 +1714,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/auditregistration/v1alpha1/doc.go b/vendor/k8s.io/api/certificates/v1/doc.go similarity index 80% rename from vendor/k8s.io/api/auditregistration/v1alpha1/doc.go rename to vendor/k8s.io/api/certificates/v1/doc.go index ae8f7671..fe3ea3af 100644 --- a/vendor/k8s.io/api/auditregistration/v1alpha1/doc.go +++ b/vendor/k8s.io/api/certificates/v1/doc.go @@ -1,5 +1,5 @@ /* -Copyright 2018 The Kubernetes Authors. +Copyright 2020 The Kubernetes Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. @@ -18,6 +18,6 @@ limitations under the License. // +k8s:protobuf-gen=package // +k8s:openapi-gen=true -// +groupName=auditregistration.k8s.io +// +groupName=certificates.k8s.io -package v1alpha1 // import "k8s.io/api/auditregistration/v1alpha1" +package v1 // import "k8s.io/api/certificates/v1" diff --git a/vendor/k8s.io/api/certificates/v1/generated.pb.go b/vendor/k8s.io/api/certificates/v1/generated.pb.go new file mode 100644 index 00000000..d2cf41a2 --- /dev/null +++ b/vendor/k8s.io/api/certificates/v1/generated.pb.go @@ -0,0 +1,2042 @@ +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by protoc-gen-gogo. DO NOT EDIT. +// source: k8s.io/kubernetes/vendor/k8s.io/api/certificates/v1/generated.proto + +package v1 + +import ( + fmt "fmt" + + io "io" + + proto "github.com/gogo/protobuf/proto" + github_com_gogo_protobuf_sortkeys "github.com/gogo/protobuf/sortkeys" + + k8s_io_api_core_v1 "k8s.io/api/core/v1" + + math "math" + math_bits "math/bits" + reflect "reflect" + strings "strings" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package + +func (m *CertificateSigningRequest) Reset() { *m = CertificateSigningRequest{} } +func (*CertificateSigningRequest) ProtoMessage() {} +func (*CertificateSigningRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_17e045d0de66f3c7, []int{0} +} +func (m *CertificateSigningRequest) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CertificateSigningRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CertificateSigningRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_CertificateSigningRequest.Merge(m, src) +} +func (m *CertificateSigningRequest) XXX_Size() int { + return m.Size() +} +func (m *CertificateSigningRequest) XXX_DiscardUnknown() { + xxx_messageInfo_CertificateSigningRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_CertificateSigningRequest proto.InternalMessageInfo + +func (m *CertificateSigningRequestCondition) Reset() { *m = CertificateSigningRequestCondition{} } +func (*CertificateSigningRequestCondition) ProtoMessage() {} +func (*CertificateSigningRequestCondition) Descriptor() ([]byte, []int) { + return fileDescriptor_17e045d0de66f3c7, []int{1} +} +func (m *CertificateSigningRequestCondition) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CertificateSigningRequestCondition) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CertificateSigningRequestCondition) XXX_Merge(src proto.Message) { + xxx_messageInfo_CertificateSigningRequestCondition.Merge(m, src) +} +func (m *CertificateSigningRequestCondition) XXX_Size() int { + return m.Size() +} +func (m *CertificateSigningRequestCondition) XXX_DiscardUnknown() { + xxx_messageInfo_CertificateSigningRequestCondition.DiscardUnknown(m) +} + +var xxx_messageInfo_CertificateSigningRequestCondition proto.InternalMessageInfo + +func (m *CertificateSigningRequestList) Reset() { *m = CertificateSigningRequestList{} } +func (*CertificateSigningRequestList) ProtoMessage() {} +func (*CertificateSigningRequestList) Descriptor() ([]byte, []int) { + return fileDescriptor_17e045d0de66f3c7, []int{2} +} +func (m *CertificateSigningRequestList) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CertificateSigningRequestList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CertificateSigningRequestList) XXX_Merge(src proto.Message) { + xxx_messageInfo_CertificateSigningRequestList.Merge(m, src) +} +func (m *CertificateSigningRequestList) XXX_Size() int { + return m.Size() +} +func (m *CertificateSigningRequestList) XXX_DiscardUnknown() { + xxx_messageInfo_CertificateSigningRequestList.DiscardUnknown(m) +} + +var xxx_messageInfo_CertificateSigningRequestList proto.InternalMessageInfo + +func (m *CertificateSigningRequestSpec) Reset() { *m = CertificateSigningRequestSpec{} } +func (*CertificateSigningRequestSpec) ProtoMessage() {} +func (*CertificateSigningRequestSpec) Descriptor() ([]byte, []int) { + return fileDescriptor_17e045d0de66f3c7, []int{3} +} +func (m *CertificateSigningRequestSpec) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CertificateSigningRequestSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CertificateSigningRequestSpec) XXX_Merge(src proto.Message) { + xxx_messageInfo_CertificateSigningRequestSpec.Merge(m, src) +} +func (m *CertificateSigningRequestSpec) XXX_Size() int { + return m.Size() +} +func (m *CertificateSigningRequestSpec) XXX_DiscardUnknown() { + xxx_messageInfo_CertificateSigningRequestSpec.DiscardUnknown(m) +} + +var xxx_messageInfo_CertificateSigningRequestSpec proto.InternalMessageInfo + +func (m *CertificateSigningRequestStatus) Reset() { *m = CertificateSigningRequestStatus{} } +func (*CertificateSigningRequestStatus) ProtoMessage() {} +func (*CertificateSigningRequestStatus) Descriptor() ([]byte, []int) { + return fileDescriptor_17e045d0de66f3c7, []int{4} +} +func (m *CertificateSigningRequestStatus) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CertificateSigningRequestStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CertificateSigningRequestStatus) XXX_Merge(src proto.Message) { + xxx_messageInfo_CertificateSigningRequestStatus.Merge(m, src) +} +func (m *CertificateSigningRequestStatus) XXX_Size() int { + return m.Size() +} +func (m *CertificateSigningRequestStatus) XXX_DiscardUnknown() { + xxx_messageInfo_CertificateSigningRequestStatus.DiscardUnknown(m) +} + +var xxx_messageInfo_CertificateSigningRequestStatus proto.InternalMessageInfo + +func (m *ExtraValue) Reset() { *m = ExtraValue{} } +func (*ExtraValue) ProtoMessage() {} +func (*ExtraValue) Descriptor() ([]byte, []int) { + return fileDescriptor_17e045d0de66f3c7, []int{5} +} +func (m *ExtraValue) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *ExtraValue) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *ExtraValue) XXX_Merge(src proto.Message) { + xxx_messageInfo_ExtraValue.Merge(m, src) +} +func (m *ExtraValue) XXX_Size() int { + return m.Size() +} +func (m *ExtraValue) XXX_DiscardUnknown() { + xxx_messageInfo_ExtraValue.DiscardUnknown(m) +} + +var xxx_messageInfo_ExtraValue proto.InternalMessageInfo + +func init() { + proto.RegisterType((*CertificateSigningRequest)(nil), "k8s.io.api.certificates.v1.CertificateSigningRequest") + proto.RegisterType((*CertificateSigningRequestCondition)(nil), "k8s.io.api.certificates.v1.CertificateSigningRequestCondition") + proto.RegisterType((*CertificateSigningRequestList)(nil), "k8s.io.api.certificates.v1.CertificateSigningRequestList") + proto.RegisterType((*CertificateSigningRequestSpec)(nil), "k8s.io.api.certificates.v1.CertificateSigningRequestSpec") + proto.RegisterMapType((map[string]ExtraValue)(nil), "k8s.io.api.certificates.v1.CertificateSigningRequestSpec.ExtraEntry") + proto.RegisterType((*CertificateSigningRequestStatus)(nil), "k8s.io.api.certificates.v1.CertificateSigningRequestStatus") + proto.RegisterType((*ExtraValue)(nil), "k8s.io.api.certificates.v1.ExtraValue") +} + +func init() { + proto.RegisterFile("k8s.io/kubernetes/vendor/k8s.io/api/certificates/v1/generated.proto", fileDescriptor_17e045d0de66f3c7) +} + +var fileDescriptor_17e045d0de66f3c7 = []byte{ + // 873 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x54, 0xcf, 0x6f, 0x1b, 0x45, + 0x14, 0xf6, 0xfa, 0x57, 0xec, 0x71, 0x49, 0xab, 0x11, 0xaa, 0x16, 0x4b, 0xdd, 0x8d, 0x56, 0x50, + 0x05, 0x04, 0xbb, 0x38, 0x2a, 0x10, 0x0a, 0xe2, 0xb0, 0x69, 0x85, 0x2a, 0x52, 0x90, 0x26, 0x09, + 0x87, 0xc2, 0xa1, 0x93, 0xf5, 0xeb, 0x66, 0xea, 0xee, 0x0f, 0x66, 0x66, 0x2d, 0x7c, 0xeb, 0x9f, + 0xc0, 0x91, 0x23, 0xff, 0x09, 0xd7, 0x1c, 0x7b, 0x2c, 0x12, 0xb2, 0x88, 0x7b, 0xe1, 0x6f, 0xc8, + 0x09, 0xcd, 0xec, 0x78, 0xed, 0xfc, 0x70, 0x5b, 0x72, 0xdb, 0xf9, 0xde, 0xf7, 0xbe, 0xef, 0xbd, + 0xb7, 0x6f, 0x06, 0xed, 0x8c, 0xb6, 0x85, 0xcf, 0xb2, 0x60, 0x54, 0x1c, 0x02, 0x4f, 0x41, 0x82, + 0x08, 0xc6, 0x90, 0x0e, 0x33, 0x1e, 0x98, 0x00, 0xcd, 0x59, 0x10, 0x01, 0x97, 0xec, 0x09, 0x8b, + 0xa8, 0x0e, 0x0f, 0x82, 0x18, 0x52, 0xe0, 0x54, 0xc2, 0xd0, 0xcf, 0x79, 0x26, 0x33, 0xdc, 0x2f, + 0xb9, 0x3e, 0xcd, 0x99, 0xbf, 0xcc, 0xf5, 0xc7, 0x83, 0xfe, 0x27, 0x31, 0x93, 0x47, 0xc5, 0xa1, + 0x1f, 0x65, 0x49, 0x10, 0x67, 0x71, 0x16, 0xe8, 0x94, 0xc3, 0xe2, 0x89, 0x3e, 0xe9, 0x83, 0xfe, + 0x2a, 0xa5, 0xfa, 0xde, 0xb2, 0x6d, 0xc6, 0xe1, 0x12, 0xbb, 0xfe, 0x9d, 0x05, 0x27, 0xa1, 0xd1, + 0x11, 0x4b, 0x81, 0x4f, 0x82, 0x7c, 0x14, 0x2b, 0x40, 0x04, 0x09, 0x48, 0x7a, 0x59, 0x56, 0xb0, + 0x2a, 0x8b, 0x17, 0xa9, 0x64, 0x09, 0x5c, 0x48, 0xf8, 0xfc, 0x4d, 0x09, 0x22, 0x3a, 0x82, 0x84, + 0x9e, 0xcf, 0xf3, 0xfe, 0xac, 0xa3, 0xf7, 0x76, 0x16, 0x53, 0xd8, 0x63, 0x71, 0xca, 0xd2, 0x98, + 0xc0, 0x2f, 0x05, 0x08, 0x89, 0x1f, 0xa3, 0x8e, 0xaa, 0x70, 0x48, 0x25, 0xb5, 0xad, 0x0d, 0x6b, + 0xb3, 0xb7, 0xf5, 0xa9, 0xbf, 0x18, 0x5f, 0x65, 0xe4, 0xe7, 0xa3, 0x58, 0x01, 0xc2, 0x57, 0x6c, + 0x7f, 0x3c, 0xf0, 0x7f, 0x38, 0x7c, 0x0a, 0x91, 0x7c, 0x08, 0x92, 0x86, 0xf8, 0x78, 0xea, 0xd6, + 0x66, 0x53, 0x17, 0x2d, 0x30, 0x52, 0xa9, 0xe2, 0x9f, 0x50, 0x53, 0xe4, 0x10, 0xd9, 0x75, 0xad, + 0xfe, 0xa5, 0xbf, 0xfa, 0xe7, 0xf8, 0x2b, 0xcb, 0xdc, 0xcb, 0x21, 0x0a, 0xaf, 0x19, 0x9b, 0xa6, + 0x3a, 0x11, 0x2d, 0x8a, 0x23, 0xd4, 0x16, 0x92, 0xca, 0x42, 0xd8, 0x0d, 0x2d, 0xff, 0xd5, 0xd5, + 0xe4, 0xb5, 0x44, 0xb8, 0x6e, 0x0c, 0xda, 0xe5, 0x99, 0x18, 0x69, 0xef, 0x55, 0x03, 0x79, 0x2b, + 0x73, 0x77, 0xb2, 0x74, 0xc8, 0x24, 0xcb, 0x52, 0xbc, 0x8d, 0x9a, 0x72, 0x92, 0x83, 0x1e, 0x63, + 0x37, 0x7c, 0x7f, 0x5e, 0xed, 0xfe, 0x24, 0x87, 0xd3, 0xa9, 0xfb, 0xee, 0x79, 0xbe, 0xc2, 0x89, + 0xce, 0xc0, 0xbb, 0x55, 0x17, 0x6d, 0x9d, 0x7b, 0xe7, 0x6c, 0x21, 0xa7, 0x53, 0xf7, 0x92, 0x3d, + 0xf4, 0x2b, 0xa5, 0xb3, 0xe5, 0xe2, 0xdb, 0xa8, 0xcd, 0x81, 0x8a, 0x2c, 0xd5, 0x23, 0xef, 0x2e, + 0xda, 0x22, 0x1a, 0x25, 0x26, 0x8a, 0x3f, 0x44, 0x6b, 0x09, 0x08, 0x41, 0x63, 0xd0, 0xc3, 0xeb, + 0x86, 0xd7, 0x0d, 0x71, 0xed, 0x61, 0x09, 0x93, 0x79, 0x1c, 0x3f, 0x45, 0xeb, 0xcf, 0xa8, 0x90, + 0x07, 0xf9, 0x90, 0x4a, 0xd8, 0x67, 0x09, 0xd8, 0x4d, 0x3d, 0xee, 0x8f, 0xde, 0x6e, 0x57, 0x54, + 0x46, 0x78, 0xd3, 0xa8, 0xaf, 0xef, 0x9e, 0x51, 0x22, 0xe7, 0x94, 0xf1, 0x18, 0x61, 0x85, 0xec, + 0x73, 0x9a, 0x8a, 0x72, 0x50, 0xca, 0xaf, 0xf5, 0xbf, 0xfd, 0xfa, 0xc6, 0x0f, 0xef, 0x5e, 0x50, + 0x23, 0x97, 0x38, 0x78, 0x7f, 0x59, 0xe8, 0xd6, 0xca, 0xbf, 0xbc, 0xcb, 0x84, 0xc4, 0x3f, 0x5f, + 0xb8, 0x2b, 0xfe, 0xdb, 0xd5, 0xa3, 0xb2, 0xf5, 0x4d, 0xb9, 0x61, 0x6a, 0xea, 0xcc, 0x91, 0xa5, + 0x7b, 0xf2, 0x08, 0xb5, 0x98, 0x84, 0x44, 0xd8, 0xf5, 0x8d, 0xc6, 0x66, 0x6f, 0xeb, 0xb3, 0x2b, + 0x6d, 0x72, 0xf8, 0x8e, 0x71, 0x68, 0x3d, 0x50, 0x5a, 0xa4, 0x94, 0xf4, 0xfe, 0x6d, 0xbc, 0xa6, + 0x37, 0x75, 0x9d, 0xf0, 0x07, 0x68, 0x8d, 0x97, 0x47, 0xdd, 0xda, 0xb5, 0xb0, 0xa7, 0x16, 0xc1, + 0x30, 0xc8, 0x3c, 0x86, 0xb7, 0x10, 0x12, 0x2c, 0x4e, 0x81, 0x7f, 0x4f, 0x13, 0xb0, 0xd7, 0xf4, + 0xda, 0x54, 0xd7, 0x7f, 0xaf, 0x8a, 0x90, 0x25, 0x16, 0xf6, 0x51, 0xbb, 0x50, 0x5b, 0x24, 0xec, + 0xd6, 0x46, 0x63, 0xb3, 0x1b, 0xde, 0x54, 0xbb, 0x78, 0xa0, 0x91, 0xd3, 0xa9, 0xdb, 0xf9, 0x0e, + 0x26, 0xfa, 0x40, 0x0c, 0x0b, 0x7f, 0x8c, 0x3a, 0x85, 0x00, 0x9e, 0x2a, 0x87, 0x72, 0x83, 0xab, + 0xb1, 0x1d, 0x18, 0x9c, 0x54, 0x0c, 0x7c, 0x0b, 0x35, 0x0a, 0x36, 0x34, 0x1b, 0xdc, 0x33, 0xc4, + 0xc6, 0xc1, 0x83, 0x7b, 0x44, 0xe1, 0xd8, 0x43, 0xed, 0x98, 0x67, 0x45, 0x2e, 0xec, 0xa6, 0x36, + 0x47, 0xca, 0xfc, 0x5b, 0x8d, 0x10, 0x13, 0xc1, 0x0c, 0xb5, 0xe0, 0x57, 0xc9, 0xa9, 0xdd, 0xd6, + 0x93, 0xbf, 0x77, 0xe5, 0x27, 0xca, 0xbf, 0xaf, 0x64, 0xee, 0xa7, 0x92, 0x4f, 0x16, 0x3f, 0x42, + 0x63, 0xa4, 0x74, 0xe8, 0x3f, 0x46, 0x68, 0xc1, 0xc1, 0x37, 0x50, 0x63, 0x04, 0x93, 0xf2, 0xc1, + 0x20, 0xea, 0x13, 0x7f, 0x8d, 0x5a, 0x63, 0xfa, 0xac, 0x00, 0xf3, 0x5a, 0xde, 0x7e, 0x5d, 0x29, + 0x5a, 0xe8, 0x47, 0xc5, 0x26, 0x65, 0xd2, 0xdd, 0xfa, 0xb6, 0xe5, 0x1d, 0x5b, 0xc8, 0x7d, 0xc3, + 0x43, 0x87, 0x39, 0x42, 0xd1, 0xfc, 0xf1, 0x10, 0xb6, 0xa5, 0xbb, 0xfe, 0xe6, 0x4a, 0x5d, 0x57, + 0x6f, 0xd0, 0x62, 0x0b, 0x2a, 0x48, 0x90, 0x25, 0x17, 0x3c, 0x40, 0xbd, 0x25, 0x55, 0xdd, 0xdf, + 0xb5, 0xf0, 0xfa, 0x6c, 0xea, 0xf6, 0x96, 0xc4, 0xc9, 0x32, 0xc7, 0xfb, 0xc2, 0x0c, 0x4b, 0xf7, + 0x88, 0xdd, 0xf9, 0xfd, 0xb0, 0xf4, 0x8f, 0xec, 0x9e, 0x5f, 0xf2, 0xbb, 0x9d, 0xdf, 0xff, 0x70, + 0x6b, 0xcf, 0xff, 0xde, 0xa8, 0x85, 0x9b, 0xc7, 0x27, 0x4e, 0xed, 0xc5, 0x89, 0x53, 0x7b, 0x79, + 0xe2, 0xd4, 0x9e, 0xcf, 0x1c, 0xeb, 0x78, 0xe6, 0x58, 0x2f, 0x66, 0x8e, 0xf5, 0x72, 0xe6, 0x58, + 0xff, 0xcc, 0x1c, 0xeb, 0xb7, 0x57, 0x4e, 0xed, 0x51, 0x7d, 0x3c, 0xf8, 0x2f, 0x00, 0x00, 0xff, + 0xff, 0x9d, 0x8f, 0x4c, 0xfa, 0x70, 0x08, 0x00, 0x00, +} + +func (m *CertificateSigningRequest) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CertificateSigningRequest) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CertificateSigningRequest) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + { + size, err := m.Status.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1a + { + size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *CertificateSigningRequestCondition) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CertificateSigningRequestCondition) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CertificateSigningRequestCondition) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i -= len(m.Status) + copy(dAtA[i:], m.Status) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Status))) + i-- + dAtA[i] = 0x32 + { + size, err := m.LastTransitionTime.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x2a + { + size, err := m.LastUpdateTime.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x22 + i -= len(m.Message) + copy(dAtA[i:], m.Message) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Message))) + i-- + dAtA[i] = 0x1a + i -= len(m.Reason) + copy(dAtA[i:], m.Reason) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Reason))) + i-- + dAtA[i] = 0x12 + i -= len(m.Type) + copy(dAtA[i:], m.Type) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Type))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *CertificateSigningRequestList) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CertificateSigningRequestList) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CertificateSigningRequestList) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + { + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *CertificateSigningRequestSpec) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CertificateSigningRequestSpec) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CertificateSigningRequestSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i -= len(m.SignerName) + copy(dAtA[i:], m.SignerName) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.SignerName))) + i-- + dAtA[i] = 0x3a + if len(m.Extra) > 0 { + keysForExtra := make([]string, 0, len(m.Extra)) + for k := range m.Extra { + keysForExtra = append(keysForExtra, string(k)) + } + github_com_gogo_protobuf_sortkeys.Strings(keysForExtra) + for iNdEx := len(keysForExtra) - 1; iNdEx >= 0; iNdEx-- { + v := m.Extra[string(keysForExtra[iNdEx])] + baseI := i + { + size, err := (&v).MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + i -= len(keysForExtra[iNdEx]) + copy(dAtA[i:], keysForExtra[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(keysForExtra[iNdEx]))) + i-- + dAtA[i] = 0xa + i = encodeVarintGenerated(dAtA, i, uint64(baseI-i)) + i-- + dAtA[i] = 0x32 + } + } + if len(m.Usages) > 0 { + for iNdEx := len(m.Usages) - 1; iNdEx >= 0; iNdEx-- { + i -= len(m.Usages[iNdEx]) + copy(dAtA[i:], m.Usages[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Usages[iNdEx]))) + i-- + dAtA[i] = 0x2a + } + } + if len(m.Groups) > 0 { + for iNdEx := len(m.Groups) - 1; iNdEx >= 0; iNdEx-- { + i -= len(m.Groups[iNdEx]) + copy(dAtA[i:], m.Groups[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Groups[iNdEx]))) + i-- + dAtA[i] = 0x22 + } + } + i -= len(m.UID) + copy(dAtA[i:], m.UID) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.UID))) + i-- + dAtA[i] = 0x1a + i -= len(m.Username) + copy(dAtA[i:], m.Username) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Username))) + i-- + dAtA[i] = 0x12 + if m.Request != nil { + i -= len(m.Request) + copy(dAtA[i:], m.Request) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Request))) + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil +} + +func (m *CertificateSigningRequestStatus) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CertificateSigningRequestStatus) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CertificateSigningRequestStatus) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.Certificate != nil { + i -= len(m.Certificate) + copy(dAtA[i:], m.Certificate) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Certificate))) + i-- + dAtA[i] = 0x12 + } + if len(m.Conditions) > 0 { + for iNdEx := len(m.Conditions) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Conditions[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + } + } + return len(dAtA) - i, nil +} + +func (m ExtraValue) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m ExtraValue) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m ExtraValue) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m) > 0 { + for iNdEx := len(m) - 1; iNdEx >= 0; iNdEx-- { + i -= len(m[iNdEx]) + copy(dAtA[i:], m[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(m[iNdEx]))) + i-- + dAtA[i] = 0xa + } + } + return len(dAtA) - i, nil +} + +func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { + offset -= sovGenerated(v) + base := offset + for v >= 1<<7 { + dAtA[offset] = uint8(v&0x7f | 0x80) + v >>= 7 + offset++ + } + dAtA[offset] = uint8(v) + return base +} +func (m *CertificateSigningRequest) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Spec.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Status.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *CertificateSigningRequestCondition) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Type) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Reason) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Message) + n += 1 + l + sovGenerated(uint64(l)) + l = m.LastUpdateTime.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.LastTransitionTime.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Status) + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *CertificateSigningRequestList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *CertificateSigningRequestSpec) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.Request != nil { + l = len(m.Request) + n += 1 + l + sovGenerated(uint64(l)) + } + l = len(m.Username) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.UID) + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Groups) > 0 { + for _, s := range m.Groups { + l = len(s) + n += 1 + l + sovGenerated(uint64(l)) + } + } + if len(m.Usages) > 0 { + for _, s := range m.Usages { + l = len(s) + n += 1 + l + sovGenerated(uint64(l)) + } + } + if len(m.Extra) > 0 { + for k, v := range m.Extra { + _ = k + _ = v + l = v.Size() + mapEntrySize := 1 + len(k) + sovGenerated(uint64(len(k))) + 1 + l + sovGenerated(uint64(l)) + n += mapEntrySize + 1 + sovGenerated(uint64(mapEntrySize)) + } + } + l = len(m.SignerName) + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *CertificateSigningRequestStatus) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if len(m.Conditions) > 0 { + for _, e := range m.Conditions { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if m.Certificate != nil { + l = len(m.Certificate) + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m ExtraValue) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if len(m) > 0 { + for _, s := range m { + l = len(s) + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func sovGenerated(x uint64) (n int) { + return (math_bits.Len64(x|1) + 6) / 7 +} +func sozGenerated(x uint64) (n int) { + return sovGenerated(uint64((x << 1) ^ uint64((int64(x) >> 63)))) +} +func (this *CertificateSigningRequest) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&CertificateSigningRequest{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "CertificateSigningRequestSpec", "CertificateSigningRequestSpec", 1), `&`, ``, 1) + `,`, + `Status:` + strings.Replace(strings.Replace(this.Status.String(), "CertificateSigningRequestStatus", "CertificateSigningRequestStatus", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *CertificateSigningRequestCondition) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&CertificateSigningRequestCondition{`, + `Type:` + fmt.Sprintf("%v", this.Type) + `,`, + `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`, + `Message:` + fmt.Sprintf("%v", this.Message) + `,`, + `LastUpdateTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastUpdateTime), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `LastTransitionTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastTransitionTime), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `Status:` + fmt.Sprintf("%v", this.Status) + `,`, + `}`, + }, "") + return s +} +func (this *CertificateSigningRequestList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]CertificateSigningRequest{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "CertificateSigningRequest", "CertificateSigningRequest", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&CertificateSigningRequestList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} +func (this *CertificateSigningRequestSpec) String() string { + if this == nil { + return "nil" + } + keysForExtra := make([]string, 0, len(this.Extra)) + for k := range this.Extra { + keysForExtra = append(keysForExtra, k) + } + github_com_gogo_protobuf_sortkeys.Strings(keysForExtra) + mapStringForExtra := "map[string]ExtraValue{" + for _, k := range keysForExtra { + mapStringForExtra += fmt.Sprintf("%v: %v,", k, this.Extra[k]) + } + mapStringForExtra += "}" + s := strings.Join([]string{`&CertificateSigningRequestSpec{`, + `Request:` + valueToStringGenerated(this.Request) + `,`, + `Username:` + fmt.Sprintf("%v", this.Username) + `,`, + `UID:` + fmt.Sprintf("%v", this.UID) + `,`, + `Groups:` + fmt.Sprintf("%v", this.Groups) + `,`, + `Usages:` + fmt.Sprintf("%v", this.Usages) + `,`, + `Extra:` + mapStringForExtra + `,`, + `SignerName:` + fmt.Sprintf("%v", this.SignerName) + `,`, + `}`, + }, "") + return s +} +func (this *CertificateSigningRequestStatus) String() string { + if this == nil { + return "nil" + } + repeatedStringForConditions := "[]CertificateSigningRequestCondition{" + for _, f := range this.Conditions { + repeatedStringForConditions += strings.Replace(strings.Replace(f.String(), "CertificateSigningRequestCondition", "CertificateSigningRequestCondition", 1), `&`, ``, 1) + "," + } + repeatedStringForConditions += "}" + s := strings.Join([]string{`&CertificateSigningRequestStatus{`, + `Conditions:` + repeatedStringForConditions + `,`, + `Certificate:` + valueToStringGenerated(this.Certificate) + `,`, + `}`, + }, "") + return s +} +func valueToStringGenerated(v interface{}) string { + rv := reflect.ValueOf(v) + if rv.IsNil() { + return "nil" + } + pv := reflect.Indirect(rv).Interface() + return fmt.Sprintf("*%v", pv) +} +func (m *CertificateSigningRequest) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CertificateSigningRequest: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CertificateSigningRequest: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *CertificateSigningRequestCondition) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CertificateSigningRequestCondition: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CertificateSigningRequestCondition: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Type", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Type = RequestConditionType(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Reason", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Reason = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Message", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Message = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field LastUpdateTime", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.LastUpdateTime.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field LastTransitionTime", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.LastTransitionTime.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 6: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Status = k8s_io_api_core_v1.ConditionStatus(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *CertificateSigningRequestList) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CertificateSigningRequestList: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CertificateSigningRequestList: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Items = append(m.Items, CertificateSigningRequest{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *CertificateSigningRequestSpec) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CertificateSigningRequestSpec: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CertificateSigningRequestSpec: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Request", wireType) + } + var byteLen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + byteLen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if byteLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + byteLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Request = append(m.Request[:0], dAtA[iNdEx:postIndex]...) + if m.Request == nil { + m.Request = []byte{} + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Username", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Username = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field UID", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.UID = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Groups", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Groups = append(m.Groups, string(dAtA[iNdEx:postIndex])) + iNdEx = postIndex + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Usages", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Usages = append(m.Usages, KeyUsage(dAtA[iNdEx:postIndex])) + iNdEx = postIndex + case 6: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Extra", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Extra == nil { + m.Extra = make(map[string]ExtraValue) + } + var mapkey string + mapvalue := &ExtraValue{} + for iNdEx < postIndex { + entryPreIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + if fieldNum == 1 { + var stringLenmapkey uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLenmapkey |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLenmapkey := int(stringLenmapkey) + if intStringLenmapkey < 0 { + return ErrInvalidLengthGenerated + } + postStringIndexmapkey := iNdEx + intStringLenmapkey + if postStringIndexmapkey < 0 { + return ErrInvalidLengthGenerated + } + if postStringIndexmapkey > l { + return io.ErrUnexpectedEOF + } + mapkey = string(dAtA[iNdEx:postStringIndexmapkey]) + iNdEx = postStringIndexmapkey + } else if fieldNum == 2 { + var mapmsglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + mapmsglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if mapmsglen < 0 { + return ErrInvalidLengthGenerated + } + postmsgIndex := iNdEx + mapmsglen + if postmsgIndex < 0 { + return ErrInvalidLengthGenerated + } + if postmsgIndex > l { + return io.ErrUnexpectedEOF + } + mapvalue = &ExtraValue{} + if err := mapvalue.Unmarshal(dAtA[iNdEx:postmsgIndex]); err != nil { + return err + } + iNdEx = postmsgIndex + } else { + iNdEx = entryPreIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > postIndex { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + m.Extra[mapkey] = *mapvalue + iNdEx = postIndex + case 7: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field SignerName", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.SignerName = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *CertificateSigningRequestStatus) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CertificateSigningRequestStatus: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CertificateSigningRequestStatus: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Conditions", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Conditions = append(m.Conditions, CertificateSigningRequestCondition{}) + if err := m.Conditions[len(m.Conditions)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Certificate", wireType) + } + var byteLen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + byteLen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if byteLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + byteLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Certificate = append(m.Certificate[:0], dAtA[iNdEx:postIndex]...) + if m.Certificate == nil { + m.Certificate = []byte{} + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *ExtraValue) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: ExtraValue: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: ExtraValue: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + *m = append(*m, string(dAtA[iNdEx:postIndex])) + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func skipGenerated(dAtA []byte) (n int, err error) { + l := len(dAtA) + iNdEx := 0 + depth := 0 + for iNdEx < l { + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return 0, ErrIntOverflowGenerated + } + if iNdEx >= l { + return 0, io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } + } + wireType := int(wire & 0x7) + switch wireType { + case 0: + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return 0, ErrIntOverflowGenerated + } + if iNdEx >= l { + return 0, io.ErrUnexpectedEOF + } + iNdEx++ + if dAtA[iNdEx-1] < 0x80 { + break + } + } + case 1: + iNdEx += 8 + case 2: + var length int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return 0, ErrIntOverflowGenerated + } + if iNdEx >= l { + return 0, io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + length |= (int(b) & 0x7F) << shift + if b < 0x80 { + break + } + } + if length < 0 { + return 0, ErrInvalidLengthGenerated + } + iNdEx += length + case 3: + depth++ + case 4: + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- + case 5: + iNdEx += 4 + default: + return 0, fmt.Errorf("proto: illegal wireType %d", wireType) + } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } + } + return 0, io.ErrUnexpectedEOF +} + +var ( + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") +) diff --git a/vendor/k8s.io/api/certificates/v1/generated.proto b/vendor/k8s.io/api/certificates/v1/generated.proto new file mode 100644 index 00000000..8427424a --- /dev/null +++ b/vendor/k8s.io/api/certificates/v1/generated.proto @@ -0,0 +1,226 @@ +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + + +// This file was autogenerated by go-to-protobuf. Do not edit it manually! + +syntax = 'proto2'; + +package k8s.io.api.certificates.v1; + +import "k8s.io/api/core/v1/generated.proto"; +import "k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto"; +import "k8s.io/apimachinery/pkg/runtime/generated.proto"; +import "k8s.io/apimachinery/pkg/runtime/schema/generated.proto"; + +// Package-wide variables from generator "generated". +option go_package = "v1"; + +// CertificateSigningRequest objects provide a mechanism to obtain x509 certificates +// by submitting a certificate signing request, and having it asynchronously approved and issued. +// +// Kubelets use this API to obtain: +// 1. client certificates to authenticate to kube-apiserver (with the "kubernetes.io/kube-apiserver-client-kubelet" signerName). +// 2. serving certificates for TLS endpoints kube-apiserver can connect to securely (with the "kubernetes.io/kubelet-serving" signerName). +// +// This API can be used to request client certificates to authenticate to kube-apiserver +// (with the "kubernetes.io/kube-apiserver-client" signerName), +// or to obtain certificates from custom non-Kubernetes signers. +message CertificateSigningRequest { + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // spec contains the certificate request, and is immutable after creation. + // Only the request, signerName, and usages fields can be set on creation. + // Other fields are derived by Kubernetes and cannot be modified by users. + optional CertificateSigningRequestSpec spec = 2; + + // status contains information about whether the request is approved or denied, + // and the certificate issued by the signer, or the failure condition indicating signer failure. + // +optional + optional CertificateSigningRequestStatus status = 3; +} + +// CertificateSigningRequestCondition describes a condition of a CertificateSigningRequest object +message CertificateSigningRequestCondition { + // type of the condition. Known conditions are "Approved", "Denied", and "Failed". + // + // An "Approved" condition is added via the /approval subresource, + // indicating the request was approved and should be issued by the signer. + // + // A "Denied" condition is added via the /approval subresource, + // indicating the request was denied and should not be issued by the signer. + // + // A "Failed" condition is added via the /status subresource, + // indicating the signer failed to issue the certificate. + // + // Approved and Denied conditions are mutually exclusive. + // Approved, Denied, and Failed conditions cannot be removed once added. + // + // Only one condition of a given type is allowed. + optional string type = 1; + + // status of the condition, one of True, False, Unknown. + // Approved, Denied, and Failed conditions may not be "False" or "Unknown". + optional string status = 6; + + // reason indicates a brief reason for the request state + // +optional + optional string reason = 2; + + // message contains a human readable message with details about the request state + // +optional + optional string message = 3; + + // lastUpdateTime is the time of the last update to this condition + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.Time lastUpdateTime = 4; + + // lastTransitionTime is the time the condition last transitioned from one status to another. + // If unset, when a new condition type is added or an existing condition's status is changed, + // the server defaults this to the current time. + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.Time lastTransitionTime = 5; +} + +// CertificateSigningRequestList is a collection of CertificateSigningRequest objects +message CertificateSigningRequestList { + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; + + // items is a collection of CertificateSigningRequest objects + repeated CertificateSigningRequest items = 2; +} + +// CertificateSigningRequestSpec contains the certificate request. +message CertificateSigningRequestSpec { + // request contains an x509 certificate signing request encoded in a "CERTIFICATE REQUEST" PEM block. + // When serialized as JSON or YAML, the data is additionally base64-encoded. + // +listType=atomic + optional bytes request = 1; + + // signerName indicates the requested signer, and is a qualified name. + // + // List/watch requests for CertificateSigningRequests can filter on this field using a "spec.signerName=NAME" fieldSelector. + // + // Well-known Kubernetes signers are: + // 1. "kubernetes.io/kube-apiserver-client": issues client certificates that can be used to authenticate to kube-apiserver. + // Requests for this signer are never auto-approved by kube-controller-manager, can be issued by the "csrsigning" controller in kube-controller-manager. + // 2. "kubernetes.io/kube-apiserver-client-kubelet": issues client certificates that kubelets use to authenticate to kube-apiserver. + // Requests for this signer can be auto-approved by the "csrapproving" controller in kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager. + // 3. "kubernetes.io/kubelet-serving" issues serving certificates that kubelets use to serve TLS endpoints, which kube-apiserver can connect to securely. + // Requests for this signer are never auto-approved by kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager. + // + // More details are available at https://k8s.io/docs/reference/access-authn-authz/certificate-signing-requests/#kubernetes-signers + // + // Custom signerNames can also be specified. The signer defines: + // 1. Trust distribution: how trust (CA bundles) are distributed. + // 2. Permitted subjects: and behavior when a disallowed subject is requested. + // 3. Required, permitted, or forbidden x509 extensions in the request (including whether subjectAltNames are allowed, which types, restrictions on allowed values) and behavior when a disallowed extension is requested. + // 4. Required, permitted, or forbidden key usages / extended key usages. + // 5. Expiration/certificate lifetime: whether it is fixed by the signer, configurable by the admin. + // 6. Whether or not requests for CA certificates are allowed. + optional string signerName = 7; + + // usages specifies a set of key usages requested in the issued certificate. + // + // Requests for TLS client certificates typically request: "digital signature", "key encipherment", "client auth". + // + // Requests for TLS serving certificates typically request: "key encipherment", "digital signature", "server auth". + // + // Valid values are: + // "signing", "digital signature", "content commitment", + // "key encipherment", "key agreement", "data encipherment", + // "cert sign", "crl sign", "encipher only", "decipher only", "any", + // "server auth", "client auth", + // "code signing", "email protection", "s/mime", + // "ipsec end system", "ipsec tunnel", "ipsec user", + // "timestamping", "ocsp signing", "microsoft sgc", "netscape sgc" + // +listType=atomic + repeated string usages = 5; + + // username contains the name of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +optional + optional string username = 2; + + // uid contains the uid of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +optional + optional string uid = 3; + + // groups contains group membership of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +listType=atomic + // +optional + repeated string groups = 4; + + // extra contains extra attributes of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +optional + map extra = 6; +} + +// CertificateSigningRequestStatus contains conditions used to indicate +// approved/denied/failed status of the request, and the issued certificate. +message CertificateSigningRequestStatus { + // conditions applied to the request. Known conditions are "Approved", "Denied", and "Failed". + // +listType=map + // +listMapKey=type + // +optional + repeated CertificateSigningRequestCondition conditions = 1; + + // certificate is populated with an issued certificate by the signer after an Approved condition is present. + // This field is set via the /status subresource. Once populated, this field is immutable. + // + // If the certificate signing request is denied, a condition of type "Denied" is added and this field remains empty. + // If the signer cannot issue the certificate, a condition of type "Failed" is added and this field remains empty. + // + // Validation requirements: + // 1. certificate must contain one or more PEM blocks. + // 2. All PEM blocks must have the "CERTIFICATE" label, contain no headers, and the encoded data + // must be a BER-encoded ASN.1 Certificate structure as described in section 4 of RFC5280. + // 3. Non-PEM content may appear before or after the "CERTIFICATE" PEM blocks and is unvalidated, + // to allow for explanatory text as described in section 5.2 of RFC7468. + // + // If more than one PEM block is present, and the definition of the requested spec.signerName + // does not indicate otherwise, the first block is the issued certificate, + // and subsequent blocks should be treated as intermediate certificates and presented in TLS handshakes. + // + // The certificate is encoded in PEM format. + // + // When serialized as JSON or YAML, the data is additionally base64-encoded, so it consists of: + // + // base64( + // -----BEGIN CERTIFICATE----- + // ... + // -----END CERTIFICATE----- + // ) + // + // +listType=atomic + // +optional + optional bytes certificate = 2; +} + +// ExtraValue masks the value so protobuf can generate +// +protobuf.nullable=true +// +protobuf.options.(gogoproto.goproto_stringer)=false +message ExtraValue { + // items, if empty, will result in an empty slice + + repeated string items = 1; +} + diff --git a/vendor/k8s.io/api/certificates/v1/register.go b/vendor/k8s.io/api/certificates/v1/register.go new file mode 100644 index 00000000..2dac94ad --- /dev/null +++ b/vendor/k8s.io/api/certificates/v1/register.go @@ -0,0 +1,61 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package v1 + +import ( + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" + "k8s.io/apimachinery/pkg/runtime" + "k8s.io/apimachinery/pkg/runtime/schema" +) + +// GroupName is the group name use in this package +const GroupName = "certificates.k8s.io" + +// SchemeGroupVersion is group version used to register these objects +var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: "v1"} + +// Kind takes an unqualified kind and returns a Group qualified GroupKind +func Kind(kind string) schema.GroupKind { + return SchemeGroupVersion.WithKind(kind).GroupKind() +} + +// Resource takes an unqualified resource and returns a Group qualified GroupResource +func Resource(resource string) schema.GroupResource { + return SchemeGroupVersion.WithResource(resource).GroupResource() +} + +var ( + // SchemeBuilder is the scheme builder with scheme init functions to run for this API package + SchemeBuilder = runtime.NewSchemeBuilder(addKnownTypes) + + localSchemeBuilder = &SchemeBuilder + + // AddToScheme is a global function that registers this API group & version to a scheme + AddToScheme = localSchemeBuilder.AddToScheme +) + +// Adds the list of known types to the given scheme. +func addKnownTypes(scheme *runtime.Scheme) error { + scheme.AddKnownTypes(SchemeGroupVersion, + &CertificateSigningRequest{}, + &CertificateSigningRequestList{}, + ) + + // Add the watch version that applies + metav1.AddToGroupVersion(scheme, SchemeGroupVersion) + return nil +} diff --git a/vendor/k8s.io/api/certificates/v1/types.go b/vendor/k8s.io/api/certificates/v1/types.go new file mode 100644 index 00000000..8d3b2305 --- /dev/null +++ b/vendor/k8s.io/api/certificates/v1/types.go @@ -0,0 +1,284 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package v1 + +import ( + "fmt" + + v1 "k8s.io/api/core/v1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" +) + +// +genclient +// +genclient:nonNamespaced +// +genclient:method=UpdateApproval,verb=update,subresource=approval,input=k8s.io/api/certificates/v1.CertificateSigningRequest,result=k8s.io/api/certificates/v1.CertificateSigningRequest +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// CertificateSigningRequest objects provide a mechanism to obtain x509 certificates +// by submitting a certificate signing request, and having it asynchronously approved and issued. +// +// Kubelets use this API to obtain: +// 1. client certificates to authenticate to kube-apiserver (with the "kubernetes.io/kube-apiserver-client-kubelet" signerName). +// 2. serving certificates for TLS endpoints kube-apiserver can connect to securely (with the "kubernetes.io/kubelet-serving" signerName). +// +// This API can be used to request client certificates to authenticate to kube-apiserver +// (with the "kubernetes.io/kube-apiserver-client" signerName), +// or to obtain certificates from custom non-Kubernetes signers. +type CertificateSigningRequest struct { + metav1.TypeMeta `json:",inline"` + // +optional + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // spec contains the certificate request, and is immutable after creation. + // Only the request, signerName, and usages fields can be set on creation. + // Other fields are derived by Kubernetes and cannot be modified by users. + Spec CertificateSigningRequestSpec `json:"spec" protobuf:"bytes,2,opt,name=spec"` + + // status contains information about whether the request is approved or denied, + // and the certificate issued by the signer, or the failure condition indicating signer failure. + // +optional + Status CertificateSigningRequestStatus `json:"status,omitempty" protobuf:"bytes,3,opt,name=status"` +} + +// CertificateSigningRequestSpec contains the certificate request. +type CertificateSigningRequestSpec struct { + // request contains an x509 certificate signing request encoded in a "CERTIFICATE REQUEST" PEM block. + // When serialized as JSON or YAML, the data is additionally base64-encoded. + // +listType=atomic + Request []byte `json:"request" protobuf:"bytes,1,opt,name=request"` + + // signerName indicates the requested signer, and is a qualified name. + // + // List/watch requests for CertificateSigningRequests can filter on this field using a "spec.signerName=NAME" fieldSelector. + // + // Well-known Kubernetes signers are: + // 1. "kubernetes.io/kube-apiserver-client": issues client certificates that can be used to authenticate to kube-apiserver. + // Requests for this signer are never auto-approved by kube-controller-manager, can be issued by the "csrsigning" controller in kube-controller-manager. + // 2. "kubernetes.io/kube-apiserver-client-kubelet": issues client certificates that kubelets use to authenticate to kube-apiserver. + // Requests for this signer can be auto-approved by the "csrapproving" controller in kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager. + // 3. "kubernetes.io/kubelet-serving" issues serving certificates that kubelets use to serve TLS endpoints, which kube-apiserver can connect to securely. + // Requests for this signer are never auto-approved by kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager. + // + // More details are available at https://k8s.io/docs/reference/access-authn-authz/certificate-signing-requests/#kubernetes-signers + // + // Custom signerNames can also be specified. The signer defines: + // 1. Trust distribution: how trust (CA bundles) are distributed. + // 2. Permitted subjects: and behavior when a disallowed subject is requested. + // 3. Required, permitted, or forbidden x509 extensions in the request (including whether subjectAltNames are allowed, which types, restrictions on allowed values) and behavior when a disallowed extension is requested. + // 4. Required, permitted, or forbidden key usages / extended key usages. + // 5. Expiration/certificate lifetime: whether it is fixed by the signer, configurable by the admin. + // 6. Whether or not requests for CA certificates are allowed. + SignerName string `json:"signerName" protobuf:"bytes,7,opt,name=signerName"` + + // usages specifies a set of key usages requested in the issued certificate. + // + // Requests for TLS client certificates typically request: "digital signature", "key encipherment", "client auth". + // + // Requests for TLS serving certificates typically request: "key encipherment", "digital signature", "server auth". + // + // Valid values are: + // "signing", "digital signature", "content commitment", + // "key encipherment", "key agreement", "data encipherment", + // "cert sign", "crl sign", "encipher only", "decipher only", "any", + // "server auth", "client auth", + // "code signing", "email protection", "s/mime", + // "ipsec end system", "ipsec tunnel", "ipsec user", + // "timestamping", "ocsp signing", "microsoft sgc", "netscape sgc" + // +listType=atomic + Usages []KeyUsage `json:"usages,omitempty" protobuf:"bytes,5,opt,name=usages"` + + // username contains the name of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +optional + Username string `json:"username,omitempty" protobuf:"bytes,2,opt,name=username"` + // uid contains the uid of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +optional + UID string `json:"uid,omitempty" protobuf:"bytes,3,opt,name=uid"` + // groups contains group membership of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +listType=atomic + // +optional + Groups []string `json:"groups,omitempty" protobuf:"bytes,4,rep,name=groups"` + // extra contains extra attributes of the user that created the CertificateSigningRequest. + // Populated by the API server on creation and immutable. + // +optional + Extra map[string]ExtraValue `json:"extra,omitempty" protobuf:"bytes,6,rep,name=extra"` +} + +// Built in signerName values that are honored by kube-controller-manager. +const ( + // "kubernetes.io/kube-apiserver-client" signer issues client certificates that can be used to authenticate to kube-apiserver. + // Never auto-approved by kube-controller-manager. + // Can be issued by the "csrsigning" controller in kube-controller-manager. + KubeAPIServerClientSignerName = "kubernetes.io/kube-apiserver-client" + + // "kubernetes.io/kube-apiserver-client-kubelet" issues client certificates that kubelets use to authenticate to kube-apiserver. + // Can be auto-approved by the "csrapproving" controller in kube-controller-manager. + // Can be issued by the "csrsigning" controller in kube-controller-manager. + KubeAPIServerClientKubeletSignerName = "kubernetes.io/kube-apiserver-client-kubelet" + + // "kubernetes.io/kubelet-serving" issues serving certificates that kubelets use to serve TLS endpoints, + // which kube-apiserver can connect to securely. + // Never auto-approved by kube-controller-manager. + // Can be issued by the "csrsigning" controller in kube-controller-manager. + KubeletServingSignerName = "kubernetes.io/kubelet-serving" +) + +// ExtraValue masks the value so protobuf can generate +// +protobuf.nullable=true +// +protobuf.options.(gogoproto.goproto_stringer)=false +type ExtraValue []string + +func (t ExtraValue) String() string { + return fmt.Sprintf("%v", []string(t)) +} + +// CertificateSigningRequestStatus contains conditions used to indicate +// approved/denied/failed status of the request, and the issued certificate. +type CertificateSigningRequestStatus struct { + // conditions applied to the request. Known conditions are "Approved", "Denied", and "Failed". + // +listType=map + // +listMapKey=type + // +optional + Conditions []CertificateSigningRequestCondition `json:"conditions,omitempty" protobuf:"bytes,1,rep,name=conditions"` + + // certificate is populated with an issued certificate by the signer after an Approved condition is present. + // This field is set via the /status subresource. Once populated, this field is immutable. + // + // If the certificate signing request is denied, a condition of type "Denied" is added and this field remains empty. + // If the signer cannot issue the certificate, a condition of type "Failed" is added and this field remains empty. + // + // Validation requirements: + // 1. certificate must contain one or more PEM blocks. + // 2. All PEM blocks must have the "CERTIFICATE" label, contain no headers, and the encoded data + // must be a BER-encoded ASN.1 Certificate structure as described in section 4 of RFC5280. + // 3. Non-PEM content may appear before or after the "CERTIFICATE" PEM blocks and is unvalidated, + // to allow for explanatory text as described in section 5.2 of RFC7468. + // + // If more than one PEM block is present, and the definition of the requested spec.signerName + // does not indicate otherwise, the first block is the issued certificate, + // and subsequent blocks should be treated as intermediate certificates and presented in TLS handshakes. + // + // The certificate is encoded in PEM format. + // + // When serialized as JSON or YAML, the data is additionally base64-encoded, so it consists of: + // + // base64( + // -----BEGIN CERTIFICATE----- + // ... + // -----END CERTIFICATE----- + // ) + // + // +listType=atomic + // +optional + Certificate []byte `json:"certificate,omitempty" protobuf:"bytes,2,opt,name=certificate"` +} + +// RequestConditionType is the type of a CertificateSigningRequestCondition +type RequestConditionType string + +// Well-known condition types for certificate requests. +const ( + // Approved indicates the request was approved and should be issued by the signer. + CertificateApproved RequestConditionType = "Approved" + // Denied indicates the request was denied and should not be issued by the signer. + CertificateDenied RequestConditionType = "Denied" + // Failed indicates the signer failed to issue the certificate. + CertificateFailed RequestConditionType = "Failed" +) + +// CertificateSigningRequestCondition describes a condition of a CertificateSigningRequest object +type CertificateSigningRequestCondition struct { + // type of the condition. Known conditions are "Approved", "Denied", and "Failed". + // + // An "Approved" condition is added via the /approval subresource, + // indicating the request was approved and should be issued by the signer. + // + // A "Denied" condition is added via the /approval subresource, + // indicating the request was denied and should not be issued by the signer. + // + // A "Failed" condition is added via the /status subresource, + // indicating the signer failed to issue the certificate. + // + // Approved and Denied conditions are mutually exclusive. + // Approved, Denied, and Failed conditions cannot be removed once added. + // + // Only one condition of a given type is allowed. + Type RequestConditionType `json:"type" protobuf:"bytes,1,opt,name=type,casttype=RequestConditionType"` + // status of the condition, one of True, False, Unknown. + // Approved, Denied, and Failed conditions may not be "False" or "Unknown". + Status v1.ConditionStatus `json:"status" protobuf:"bytes,6,opt,name=status,casttype=k8s.io/api/core/v1.ConditionStatus"` + // reason indicates a brief reason for the request state + // +optional + Reason string `json:"reason,omitempty" protobuf:"bytes,2,opt,name=reason"` + // message contains a human readable message with details about the request state + // +optional + Message string `json:"message,omitempty" protobuf:"bytes,3,opt,name=message"` + // lastUpdateTime is the time of the last update to this condition + // +optional + LastUpdateTime metav1.Time `json:"lastUpdateTime,omitempty" protobuf:"bytes,4,opt,name=lastUpdateTime"` + // lastTransitionTime is the time the condition last transitioned from one status to another. + // If unset, when a new condition type is added or an existing condition's status is changed, + // the server defaults this to the current time. + // +optional + LastTransitionTime metav1.Time `json:"lastTransitionTime,omitempty" protobuf:"bytes,5,opt,name=lastTransitionTime"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// CertificateSigningRequestList is a collection of CertificateSigningRequest objects +type CertificateSigningRequestList struct { + metav1.TypeMeta `json:",inline"` + // +optional + metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // items is a collection of CertificateSigningRequest objects + Items []CertificateSigningRequest `json:"items" protobuf:"bytes,2,rep,name=items"` +} + +// KeyUsage specifies valid usage contexts for keys. +// See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3 +// https://tools.ietf.org/html/rfc5280#section-4.2.1.12 +type KeyUsage string + +// Valid key usages +const ( + UsageSigning KeyUsage = "signing" + UsageDigitalSignature KeyUsage = "digital signature" + UsageContentCommitment KeyUsage = "content commitment" + UsageKeyEncipherment KeyUsage = "key encipherment" + UsageKeyAgreement KeyUsage = "key agreement" + UsageDataEncipherment KeyUsage = "data encipherment" + UsageCertSign KeyUsage = "cert sign" + UsageCRLSign KeyUsage = "crl sign" + UsageEncipherOnly KeyUsage = "encipher only" + UsageDecipherOnly KeyUsage = "decipher only" + UsageAny KeyUsage = "any" + UsageServerAuth KeyUsage = "server auth" + UsageClientAuth KeyUsage = "client auth" + UsageCodeSigning KeyUsage = "code signing" + UsageEmailProtection KeyUsage = "email protection" + UsageSMIME KeyUsage = "s/mime" + UsageIPsecEndSystem KeyUsage = "ipsec end system" + UsageIPsecTunnel KeyUsage = "ipsec tunnel" + UsageIPsecUser KeyUsage = "ipsec user" + UsageTimestamping KeyUsage = "timestamping" + UsageOCSPSigning KeyUsage = "ocsp signing" + UsageMicrosoftSGC KeyUsage = "microsoft sgc" + UsageNetscapeSGC KeyUsage = "netscape sgc" +) diff --git a/vendor/k8s.io/api/certificates/v1/types_swagger_doc_generated.go b/vendor/k8s.io/api/certificates/v1/types_swagger_doc_generated.go new file mode 100644 index 00000000..9a078fa0 --- /dev/null +++ b/vendor/k8s.io/api/certificates/v1/types_swagger_doc_generated.go @@ -0,0 +1,88 @@ +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package v1 + +// This file contains a collection of methods that can be used from go-restful to +// generate Swagger API documentation for its models. Please read this PR for more +// information on the implementation: https://github.com/emicklei/go-restful/pull/215 +// +// TODOs are ignored from the parser (e.g. TODO(andronat):... || TODO:...) if and only if +// they are on one line! For multiple line or blocks that you want to ignore use ---. +// Any context after a --- is ignored. +// +// Those methods can be generated by using hack/update-generated-swagger-docs.sh + +// AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. +var map_CertificateSigningRequest = map[string]string{ + "": "CertificateSigningRequest objects provide a mechanism to obtain x509 certificates by submitting a certificate signing request, and having it asynchronously approved and issued.\n\nKubelets use this API to obtain:\n 1. client certificates to authenticate to kube-apiserver (with the \"kubernetes.io/kube-apiserver-client-kubelet\" signerName).\n 2. serving certificates for TLS endpoints kube-apiserver can connect to securely (with the \"kubernetes.io/kubelet-serving\" signerName).\n\nThis API can be used to request client certificates to authenticate to kube-apiserver (with the \"kubernetes.io/kube-apiserver-client\" signerName), or to obtain certificates from custom non-Kubernetes signers.", + "spec": "spec contains the certificate request, and is immutable after creation. Only the request, signerName, and usages fields can be set on creation. Other fields are derived by Kubernetes and cannot be modified by users.", + "status": "status contains information about whether the request is approved or denied, and the certificate issued by the signer, or the failure condition indicating signer failure.", +} + +func (CertificateSigningRequest) SwaggerDoc() map[string]string { + return map_CertificateSigningRequest +} + +var map_CertificateSigningRequestCondition = map[string]string{ + "": "CertificateSigningRequestCondition describes a condition of a CertificateSigningRequest object", + "type": "type of the condition. Known conditions are \"Approved\", \"Denied\", and \"Failed\".\n\nAn \"Approved\" condition is added via the /approval subresource, indicating the request was approved and should be issued by the signer.\n\nA \"Denied\" condition is added via the /approval subresource, indicating the request was denied and should not be issued by the signer.\n\nA \"Failed\" condition is added via the /status subresource, indicating the signer failed to issue the certificate.\n\nApproved and Denied conditions are mutually exclusive. Approved, Denied, and Failed conditions cannot be removed once added.\n\nOnly one condition of a given type is allowed.", + "status": "status of the condition, one of True, False, Unknown. Approved, Denied, and Failed conditions may not be \"False\" or \"Unknown\".", + "reason": "reason indicates a brief reason for the request state", + "message": "message contains a human readable message with details about the request state", + "lastUpdateTime": "lastUpdateTime is the time of the last update to this condition", + "lastTransitionTime": "lastTransitionTime is the time the condition last transitioned from one status to another. If unset, when a new condition type is added or an existing condition's status is changed, the server defaults this to the current time.", +} + +func (CertificateSigningRequestCondition) SwaggerDoc() map[string]string { + return map_CertificateSigningRequestCondition +} + +var map_CertificateSigningRequestList = map[string]string{ + "": "CertificateSigningRequestList is a collection of CertificateSigningRequest objects", + "items": "items is a collection of CertificateSigningRequest objects", +} + +func (CertificateSigningRequestList) SwaggerDoc() map[string]string { + return map_CertificateSigningRequestList +} + +var map_CertificateSigningRequestSpec = map[string]string{ + "": "CertificateSigningRequestSpec contains the certificate request.", + "request": "request contains an x509 certificate signing request encoded in a \"CERTIFICATE REQUEST\" PEM block. When serialized as JSON or YAML, the data is additionally base64-encoded.", + "signerName": "signerName indicates the requested signer, and is a qualified name.\n\nList/watch requests for CertificateSigningRequests can filter on this field using a \"spec.signerName=NAME\" fieldSelector.\n\nWell-known Kubernetes signers are:\n 1. \"kubernetes.io/kube-apiserver-client\": issues client certificates that can be used to authenticate to kube-apiserver.\n Requests for this signer are never auto-approved by kube-controller-manager, can be issued by the \"csrsigning\" controller in kube-controller-manager.\n 2. \"kubernetes.io/kube-apiserver-client-kubelet\": issues client certificates that kubelets use to authenticate to kube-apiserver.\n Requests for this signer can be auto-approved by the \"csrapproving\" controller in kube-controller-manager, and can be issued by the \"csrsigning\" controller in kube-controller-manager.\n 3. \"kubernetes.io/kubelet-serving\" issues serving certificates that kubelets use to serve TLS endpoints, which kube-apiserver can connect to securely.\n Requests for this signer are never auto-approved by kube-controller-manager, and can be issued by the \"csrsigning\" controller in kube-controller-manager.\n\nMore details are available at https://k8s.io/docs/reference/access-authn-authz/certificate-signing-requests/#kubernetes-signers\n\nCustom signerNames can also be specified. The signer defines:\n 1. Trust distribution: how trust (CA bundles) are distributed.\n 2. Permitted subjects: and behavior when a disallowed subject is requested.\n 3. Required, permitted, or forbidden x509 extensions in the request (including whether subjectAltNames are allowed, which types, restrictions on allowed values) and behavior when a disallowed extension is requested.\n 4. Required, permitted, or forbidden key usages / extended key usages.\n 5. Expiration/certificate lifetime: whether it is fixed by the signer, configurable by the admin.\n 6. Whether or not requests for CA certificates are allowed.", + "usages": "usages specifies a set of key usages requested in the issued certificate.\n\nRequests for TLS client certificates typically request: \"digital signature\", \"key encipherment\", \"client auth\".\n\nRequests for TLS serving certificates typically request: \"key encipherment\", \"digital signature\", \"server auth\".\n\nValid values are:\n \"signing\", \"digital signature\", \"content commitment\",\n \"key encipherment\", \"key agreement\", \"data encipherment\",\n \"cert sign\", \"crl sign\", \"encipher only\", \"decipher only\", \"any\",\n \"server auth\", \"client auth\",\n \"code signing\", \"email protection\", \"s/mime\",\n \"ipsec end system\", \"ipsec tunnel\", \"ipsec user\",\n \"timestamping\", \"ocsp signing\", \"microsoft sgc\", \"netscape sgc\"", + "username": "username contains the name of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.", + "uid": "uid contains the uid of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.", + "groups": "groups contains group membership of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.", + "extra": "extra contains extra attributes of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.", +} + +func (CertificateSigningRequestSpec) SwaggerDoc() map[string]string { + return map_CertificateSigningRequestSpec +} + +var map_CertificateSigningRequestStatus = map[string]string{ + "": "CertificateSigningRequestStatus contains conditions used to indicate approved/denied/failed status of the request, and the issued certificate.", + "conditions": "conditions applied to the request. Known conditions are \"Approved\", \"Denied\", and \"Failed\".", + "certificate": "certificate is populated with an issued certificate by the signer after an Approved condition is present. This field is set via the /status subresource. Once populated, this field is immutable.\n\nIf the certificate signing request is denied, a condition of type \"Denied\" is added and this field remains empty. If the signer cannot issue the certificate, a condition of type \"Failed\" is added and this field remains empty.\n\nValidation requirements:\n 1. certificate must contain one or more PEM blocks.\n 2. All PEM blocks must have the \"CERTIFICATE\" label, contain no headers, and the encoded data\n must be a BER-encoded ASN.1 Certificate structure as described in section 4 of RFC5280.\n 3. Non-PEM content may appear before or after the \"CERTIFICATE\" PEM blocks and is unvalidated,\n to allow for explanatory text as described in section 5.2 of RFC7468.\n\nIf more than one PEM block is present, and the definition of the requested spec.signerName does not indicate otherwise, the first block is the issued certificate, and subsequent blocks should be treated as intermediate certificates and presented in TLS handshakes.\n\nThe certificate is encoded in PEM format.\n\nWhen serialized as JSON or YAML, the data is additionally base64-encoded, so it consists of:\n\n base64(", +} + +func (CertificateSigningRequestStatus) SwaggerDoc() map[string]string { + return map_CertificateSigningRequestStatus +} + +// AUTO-GENERATED FUNCTIONS END HERE diff --git a/vendor/k8s.io/api/certificates/v1/zz_generated.deepcopy.go b/vendor/k8s.io/api/certificates/v1/zz_generated.deepcopy.go new file mode 100644 index 00000000..cc6a60be --- /dev/null +++ b/vendor/k8s.io/api/certificates/v1/zz_generated.deepcopy.go @@ -0,0 +1,198 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by deepcopy-gen. DO NOT EDIT. + +package v1 + +import ( + runtime "k8s.io/apimachinery/pkg/runtime" +) + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CertificateSigningRequest) DeepCopyInto(out *CertificateSigningRequest) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + in.Spec.DeepCopyInto(&out.Spec) + in.Status.DeepCopyInto(&out.Status) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CertificateSigningRequest. +func (in *CertificateSigningRequest) DeepCopy() *CertificateSigningRequest { + if in == nil { + return nil + } + out := new(CertificateSigningRequest) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *CertificateSigningRequest) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CertificateSigningRequestCondition) DeepCopyInto(out *CertificateSigningRequestCondition) { + *out = *in + in.LastUpdateTime.DeepCopyInto(&out.LastUpdateTime) + in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CertificateSigningRequestCondition. +func (in *CertificateSigningRequestCondition) DeepCopy() *CertificateSigningRequestCondition { + if in == nil { + return nil + } + out := new(CertificateSigningRequestCondition) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CertificateSigningRequestList) DeepCopyInto(out *CertificateSigningRequestList) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ListMeta.DeepCopyInto(&out.ListMeta) + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]CertificateSigningRequest, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CertificateSigningRequestList. +func (in *CertificateSigningRequestList) DeepCopy() *CertificateSigningRequestList { + if in == nil { + return nil + } + out := new(CertificateSigningRequestList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *CertificateSigningRequestList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CertificateSigningRequestSpec) DeepCopyInto(out *CertificateSigningRequestSpec) { + *out = *in + if in.Request != nil { + in, out := &in.Request, &out.Request + *out = make([]byte, len(*in)) + copy(*out, *in) + } + if in.Usages != nil { + in, out := &in.Usages, &out.Usages + *out = make([]KeyUsage, len(*in)) + copy(*out, *in) + } + if in.Groups != nil { + in, out := &in.Groups, &out.Groups + *out = make([]string, len(*in)) + copy(*out, *in) + } + if in.Extra != nil { + in, out := &in.Extra, &out.Extra + *out = make(map[string]ExtraValue, len(*in)) + for key, val := range *in { + var outVal []string + if val == nil { + (*out)[key] = nil + } else { + in, out := &val, &outVal + *out = make(ExtraValue, len(*in)) + copy(*out, *in) + } + (*out)[key] = outVal + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CertificateSigningRequestSpec. +func (in *CertificateSigningRequestSpec) DeepCopy() *CertificateSigningRequestSpec { + if in == nil { + return nil + } + out := new(CertificateSigningRequestSpec) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CertificateSigningRequestStatus) DeepCopyInto(out *CertificateSigningRequestStatus) { + *out = *in + if in.Conditions != nil { + in, out := &in.Conditions, &out.Conditions + *out = make([]CertificateSigningRequestCondition, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.Certificate != nil { + in, out := &in.Certificate, &out.Certificate + *out = make([]byte, len(*in)) + copy(*out, *in) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CertificateSigningRequestStatus. +func (in *CertificateSigningRequestStatus) DeepCopy() *CertificateSigningRequestStatus { + if in == nil { + return nil + } + out := new(CertificateSigningRequestStatus) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in ExtraValue) DeepCopyInto(out *ExtraValue) { + { + in := &in + *out = make(ExtraValue, len(*in)) + copy(*out, *in) + return + } +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ExtraValue. +func (in ExtraValue) DeepCopy() ExtraValue { + if in == nil { + return nil + } + out := new(ExtraValue) + in.DeepCopyInto(out) + return *out +} diff --git a/vendor/k8s.io/api/certificates/v1beta1/doc.go b/vendor/k8s.io/api/certificates/v1beta1/doc.go index 9055248b..1165518c 100644 --- a/vendor/k8s.io/api/certificates/v1beta1/doc.go +++ b/vendor/k8s.io/api/certificates/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=certificates.k8s.io diff --git a/vendor/k8s.io/api/certificates/v1beta1/generated.pb.go b/vendor/k8s.io/api/certificates/v1beta1/generated.pb.go index 2e61b568..1729931b 100644 --- a/vendor/k8s.io/api/certificates/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/certificates/v1beta1/generated.pb.go @@ -27,6 +27,8 @@ import ( proto "github.com/gogo/protobuf/proto" github_com_gogo_protobuf_sortkeys "github.com/gogo/protobuf/sortkeys" + k8s_io_api_core_v1 "k8s.io/api/core/v1" + math "math" math_bits "math/bits" reflect "reflect" @@ -42,7 +44,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *CertificateSigningRequest) Reset() { *m = CertificateSigningRequest{} } func (*CertificateSigningRequest) ProtoMessage() {} @@ -227,58 +229,62 @@ func init() { } var fileDescriptor_09d156762b8218ef = []byte{ - // 805 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x54, 0x4b, 0x8f, 0x1b, 0x45, - 0x10, 0xf6, 0xf8, 0xb5, 0x76, 0x7b, 0xd9, 0x44, 0x2d, 0x14, 0x0d, 0x2b, 0x65, 0x66, 0x35, 0x02, - 0xb4, 0x3c, 0xd2, 0xc3, 0x46, 0x08, 0x56, 0x7b, 0x40, 0x30, 0x4b, 0x04, 0x2b, 0x12, 0x21, 0x75, - 0x62, 0x0e, 0x08, 0x89, 0xb4, 0xc7, 0x95, 0x71, 0xc7, 0x99, 0x07, 0xd3, 0x3d, 0x06, 0xdf, 0xf2, - 0x13, 0x38, 0x72, 0x41, 0xe2, 0x97, 0x70, 0x5e, 0x0e, 0x48, 0x39, 0xe6, 0x80, 0x2c, 0xd6, 0xfc, - 0x8b, 0x9c, 0x50, 0xf7, 0xb4, 0x3d, 0xc6, 0x2b, 0xe3, 0x28, 0x7b, 0x9b, 0xfa, 0xaa, 0xbe, 0xaf, - 0x1e, 0x5d, 0x35, 0xe8, 0xcb, 0xf1, 0xb1, 0x20, 0x3c, 0xf5, 0xc7, 0xc5, 0x00, 0xf2, 0x04, 0x24, - 0x08, 0x7f, 0x02, 0xc9, 0x30, 0xcd, 0x7d, 0xe3, 0x60, 0x19, 0xf7, 0x43, 0xc8, 0x25, 0x7f, 0xc4, - 0x43, 0xa6, 0xdd, 0x47, 0x03, 0x90, 0xec, 0xc8, 0x8f, 0x20, 0x81, 0x9c, 0x49, 0x18, 0x92, 0x2c, - 0x4f, 0x65, 0x8a, 0xdd, 0x92, 0x40, 0x58, 0xc6, 0xc9, 0x2a, 0x81, 0x18, 0xc2, 0xfe, 0xad, 0x88, - 0xcb, 0x51, 0x31, 0x20, 0x61, 0x1a, 0xfb, 0x51, 0x1a, 0xa5, 0xbe, 0xe6, 0x0d, 0x8a, 0x47, 0xda, - 0xd2, 0x86, 0xfe, 0x2a, 0xf5, 0xf6, 0x3f, 0xac, 0x0a, 0x88, 0x59, 0x38, 0xe2, 0x09, 0xe4, 0x53, - 0x3f, 0x1b, 0x47, 0x0a, 0x10, 0x7e, 0x0c, 0x92, 0xf9, 0x93, 0x4b, 0x55, 0xec, 0xfb, 0x9b, 0x58, - 0x79, 0x91, 0x48, 0x1e, 0xc3, 0x25, 0xc2, 0x47, 0xdb, 0x08, 0x22, 0x1c, 0x41, 0xcc, 0xd6, 0x79, - 0xde, 0x1f, 0x75, 0xf4, 0xc6, 0x69, 0xd5, 0xe6, 0x7d, 0x1e, 0x25, 0x3c, 0x89, 0x28, 0xfc, 0x50, - 0x80, 0x90, 0xf8, 0x21, 0xea, 0xa8, 0x0a, 0x87, 0x4c, 0x32, 0xdb, 0x3a, 0xb0, 0x0e, 0x7b, 0xb7, - 0x3f, 0x20, 0xd5, 0x7c, 0x96, 0x89, 0x48, 0x36, 0x8e, 0x14, 0x20, 0x88, 0x8a, 0x26, 0x93, 0x23, - 0xf2, 0xf5, 0xe0, 0x31, 0x84, 0xf2, 0x1e, 0x48, 0x16, 0xe0, 0xf3, 0x99, 0x5b, 0x9b, 0xcf, 0x5c, - 0x54, 0x61, 0x74, 0xa9, 0x8a, 0x1f, 0xa2, 0xa6, 0xc8, 0x20, 0xb4, 0xeb, 0x5a, 0xfd, 0x13, 0xb2, - 0x65, 0xfa, 0x64, 0x63, 0xad, 0xf7, 0x33, 0x08, 0x83, 0x5d, 0x93, 0xab, 0xa9, 0x2c, 0xaa, 0x95, - 0xf1, 0x08, 0xb5, 0x85, 0x64, 0xb2, 0x10, 0x76, 0x43, 0xe7, 0xf8, 0xf4, 0x0a, 0x39, 0xb4, 0x4e, - 0xb0, 0x67, 0xb2, 0xb4, 0x4b, 0x9b, 0x1a, 0x7d, 0xef, 0xd7, 0x3a, 0xf2, 0x36, 0x72, 0x4f, 0xd3, - 0x64, 0xc8, 0x25, 0x4f, 0x13, 0x7c, 0x8c, 0x9a, 0x72, 0x9a, 0x81, 0x1e, 0x68, 0x37, 0x78, 0x73, - 0x51, 0xf2, 0x83, 0x69, 0x06, 0x2f, 0x66, 0xee, 0xeb, 0xeb, 0xf1, 0x0a, 0xa7, 0x9a, 0x81, 0xdf, - 0x46, 0xed, 0x1c, 0x98, 0x48, 0x13, 0x3d, 0xae, 0x6e, 0x55, 0x08, 0xd5, 0x28, 0x35, 0x5e, 0xfc, - 0x0e, 0xda, 0x89, 0x41, 0x08, 0x16, 0x81, 0xee, 0xb9, 0x1b, 0x5c, 0x33, 0x81, 0x3b, 0xf7, 0x4a, - 0x98, 0x2e, 0xfc, 0xf8, 0x31, 0xda, 0x7b, 0xc2, 0x84, 0xec, 0x67, 0x43, 0x26, 0xe1, 0x01, 0x8f, - 0xc1, 0x6e, 0xea, 0x29, 0xbd, 0xfb, 0x72, 0xef, 0xac, 0x18, 0xc1, 0x0d, 0xa3, 0xbe, 0x77, 0xf7, - 0x3f, 0x4a, 0x74, 0x4d, 0xd9, 0x9b, 0x59, 0xe8, 0xe6, 0xc6, 0xf9, 0xdc, 0xe5, 0x42, 0xe2, 0xef, - 0x2e, 0xed, 0x1b, 0x79, 0xb9, 0x3a, 0x14, 0x5b, 0x6f, 0xdb, 0x75, 0x53, 0x4b, 0x67, 0x81, 0xac, - 0xec, 0xda, 0xf7, 0xa8, 0xc5, 0x25, 0xc4, 0xc2, 0xae, 0x1f, 0x34, 0x0e, 0x7b, 0xb7, 0x4f, 0x5e, - 0x7d, 0x11, 0x82, 0xd7, 0x4c, 0x9a, 0xd6, 0x99, 0x12, 0xa4, 0xa5, 0xae, 0xf7, 0x7b, 0xe3, 0x7f, - 0x1a, 0x54, 0x2b, 0x89, 0xdf, 0x42, 0x3b, 0x79, 0x69, 0xea, 0xfe, 0x76, 0x83, 0x9e, 0x7a, 0x15, - 0x13, 0x41, 0x17, 0x3e, 0x4c, 0x50, 0xbb, 0x50, 0xcf, 0x23, 0xec, 0xd6, 0x41, 0xe3, 0xb0, 0x1b, - 0xdc, 0x50, 0x8f, 0xdc, 0xd7, 0xc8, 0x8b, 0x99, 0xdb, 0xf9, 0x0a, 0xa6, 0xda, 0xa0, 0x26, 0x0a, - 0xbf, 0x8f, 0x3a, 0x85, 0x80, 0x3c, 0x61, 0x31, 0x98, 0xd5, 0x58, 0xce, 0xa1, 0x6f, 0x70, 0xba, - 0x8c, 0xc0, 0x37, 0x51, 0xa3, 0xe0, 0x43, 0xb3, 0x1a, 0x3d, 0x13, 0xd8, 0xe8, 0x9f, 0x7d, 0x4e, - 0x15, 0x8e, 0x3d, 0xd4, 0x8e, 0xf2, 0xb4, 0xc8, 0x84, 0xdd, 0xd4, 0xc9, 0x91, 0x4a, 0xfe, 0x85, - 0x46, 0xa8, 0xf1, 0xe0, 0x04, 0xb5, 0xe0, 0x27, 0x99, 0x33, 0xbb, 0xad, 0x47, 0x79, 0x76, 0xb5, - 0xbb, 0x25, 0x77, 0x94, 0xd6, 0x9d, 0x44, 0xe6, 0xd3, 0x6a, 0xb2, 0x1a, 0xa3, 0x65, 0x9a, 0x7d, - 0x40, 0xa8, 0x8a, 0xc1, 0xd7, 0x51, 0x63, 0x0c, 0xd3, 0xf2, 0x80, 0xa8, 0xfa, 0xc4, 0x9f, 0xa1, - 0xd6, 0x84, 0x3d, 0x29, 0xc0, 0xfc, 0x47, 0xde, 0xdb, 0x5a, 0x8f, 0x56, 0xfb, 0x46, 0x51, 0x68, - 0xc9, 0x3c, 0xa9, 0x1f, 0x5b, 0xde, 0x9f, 0x16, 0x72, 0xb7, 0x5c, 0x3f, 0xfe, 0x11, 0xa1, 0x70, - 0x71, 0x9b, 0xc2, 0xb6, 0x74, 0xff, 0xa7, 0xaf, 0xde, 0xff, 0xf2, 0xce, 0xab, 0x1f, 0xe5, 0x12, - 0x12, 0x74, 0x25, 0x15, 0x3e, 0x42, 0xbd, 0x15, 0x69, 0xdd, 0xe9, 0x6e, 0x70, 0x6d, 0x3e, 0x73, - 0x7b, 0x2b, 0xe2, 0x74, 0x35, 0xc6, 0xfb, 0xd8, 0x8c, 0x4d, 0x37, 0x8a, 0xdd, 0xc5, 0xfe, 0x5b, - 0xfa, 0x5d, 0xbb, 0xeb, 0xfb, 0x7b, 0xd2, 0xf9, 0xe5, 0x37, 0xb7, 0xf6, 0xf4, 0xaf, 0x83, 0x5a, - 0x70, 0xeb, 0xfc, 0xc2, 0xa9, 0x3d, 0xbb, 0x70, 0x6a, 0xcf, 0x2f, 0x9c, 0xda, 0xd3, 0xb9, 0x63, - 0x9d, 0xcf, 0x1d, 0xeb, 0xd9, 0xdc, 0xb1, 0x9e, 0xcf, 0x1d, 0xeb, 0xef, 0xb9, 0x63, 0xfd, 0xfc, - 0x8f, 0x53, 0xfb, 0x76, 0xc7, 0x74, 0xf7, 0x6f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x39, 0x0e, 0xb6, - 0xcd, 0x7f, 0x07, 0x00, 0x00, + // 878 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x55, 0x4b, 0x6f, 0x1c, 0x45, + 0x10, 0xde, 0xf1, 0xbe, 0x7b, 0x8d, 0x13, 0xb5, 0x50, 0x34, 0xac, 0x94, 0x19, 0x6b, 0x04, 0xc8, + 0x3c, 0xd2, 0x83, 0xa3, 0x08, 0x2c, 0x1f, 0x10, 0x8c, 0x89, 0xc0, 0xc2, 0x01, 0xa9, 0x6d, 0x73, + 0x40, 0x48, 0xa4, 0x77, 0xb6, 0x32, 0xee, 0x6c, 0xe6, 0xc1, 0x74, 0xcf, 0xc2, 0xde, 0xf2, 0x13, + 0x38, 0x72, 0xe4, 0xe7, 0x98, 0x03, 0x52, 0x8e, 0x39, 0xa0, 0x15, 0xde, 0xdc, 0xf9, 0x01, 0x3e, + 0xa1, 0xee, 0xe9, 0x9d, 0x5d, 0xbf, 0x70, 0x48, 0x6e, 0xdb, 0x5f, 0xd7, 0xf7, 0x7d, 0x55, 0x35, + 0xd5, 0xb5, 0xe8, 0xab, 0xd1, 0x96, 0x20, 0x3c, 0xf5, 0x47, 0xc5, 0x00, 0xf2, 0x04, 0x24, 0x08, + 0x7f, 0x0c, 0xc9, 0x30, 0xcd, 0x7d, 0x73, 0xc1, 0x32, 0xee, 0x87, 0x90, 0x4b, 0xfe, 0x88, 0x87, + 0x4c, 0x5f, 0x6f, 0x0e, 0x40, 0xb2, 0x4d, 0x3f, 0x82, 0x04, 0x72, 0x26, 0x61, 0x48, 0xb2, 0x3c, + 0x95, 0x29, 0x76, 0x4b, 0x02, 0x61, 0x19, 0x27, 0xcb, 0x04, 0x62, 0x08, 0xfd, 0x3b, 0x11, 0x97, + 0x47, 0xc5, 0x80, 0x84, 0x69, 0xec, 0x47, 0x69, 0x94, 0xfa, 0x9a, 0x37, 0x28, 0x1e, 0xe9, 0x93, + 0x3e, 0xe8, 0x5f, 0xa5, 0x5e, 0xdf, 0x5b, 0x4e, 0x20, 0xcd, 0xc1, 0x1f, 0x5f, 0xf0, 0xec, 0xdf, + 0x5b, 0xc4, 0xc4, 0x2c, 0x3c, 0xe2, 0x09, 0xe4, 0x13, 0x3f, 0x1b, 0x45, 0x0a, 0x10, 0x7e, 0x0c, + 0x92, 0x5d, 0xc6, 0xf2, 0xaf, 0x62, 0xe5, 0x45, 0x22, 0x79, 0x0c, 0x17, 0x08, 0x1f, 0x5f, 0x47, + 0x10, 0xe1, 0x11, 0xc4, 0xec, 0x3c, 0xcf, 0xfb, 0x63, 0x05, 0xbd, 0xb5, 0xb3, 0x68, 0xc5, 0x3e, + 0x8f, 0x12, 0x9e, 0x44, 0x14, 0x7e, 0x2a, 0x40, 0x48, 0xfc, 0x10, 0x75, 0x54, 0x86, 0x43, 0x26, + 0x99, 0x6d, 0xad, 0x5b, 0x1b, 0xbd, 0xbb, 0x1f, 0x91, 0x45, 0x0f, 0x2b, 0x23, 0x92, 0x8d, 0x22, + 0x05, 0x08, 0xa2, 0xa2, 0xc9, 0x78, 0x93, 0x7c, 0x3b, 0x78, 0x0c, 0xa1, 0x7c, 0x00, 0x92, 0x05, + 0xf8, 0x78, 0xea, 0xd6, 0x66, 0x53, 0x17, 0x2d, 0x30, 0x5a, 0xa9, 0xe2, 0x87, 0xa8, 0x21, 0x32, + 0x08, 0xed, 0x15, 0xad, 0xfe, 0x29, 0xb9, 0xe6, 0x0b, 0x91, 0x2b, 0x73, 0xdd, 0xcf, 0x20, 0x0c, + 0x56, 0x8d, 0x57, 0x43, 0x9d, 0xa8, 0x56, 0xc6, 0x47, 0xa8, 0x25, 0x24, 0x93, 0x85, 0xb0, 0xeb, + 0xda, 0xe3, 0xb3, 0xd7, 0xf0, 0xd0, 0x3a, 0xc1, 0x9a, 0x71, 0x69, 0x95, 0x67, 0x6a, 0xf4, 0xbd, + 0x17, 0x75, 0xe4, 0x5d, 0xc9, 0xdd, 0x49, 0x93, 0x21, 0x97, 0x3c, 0x4d, 0xf0, 0x16, 0x6a, 0xc8, + 0x49, 0x06, 0xba, 0xa1, 0xdd, 0xe0, 0xed, 0x79, 0xca, 0x07, 0x93, 0x0c, 0x4e, 0xa7, 0xee, 0x9b, + 0xe7, 0xe3, 0x15, 0x4e, 0x35, 0x03, 0xef, 0x55, 0xa5, 0xb4, 0x34, 0xf7, 0xde, 0xd9, 0x44, 0x4e, + 0xa7, 0xee, 0x25, 0x13, 0x49, 0x2a, 0xa5, 0xb3, 0xe9, 0xe2, 0x77, 0x51, 0x2b, 0x07, 0x26, 0xd2, + 0x44, 0x37, 0xbf, 0xbb, 0x28, 0x8b, 0x6a, 0x94, 0x9a, 0x5b, 0xfc, 0x1e, 0x6a, 0xc7, 0x20, 0x04, + 0x8b, 0x40, 0x77, 0xb0, 0x1b, 0xdc, 0x30, 0x81, 0xed, 0x07, 0x25, 0x4c, 0xe7, 0xf7, 0xf8, 0x31, + 0x5a, 0x7b, 0xc2, 0x84, 0x3c, 0xcc, 0x86, 0x4c, 0xc2, 0x01, 0x8f, 0xc1, 0x6e, 0xe8, 0x9e, 0xbf, + 0xff, 0x72, 0x53, 0xa3, 0x18, 0xc1, 0x2d, 0xa3, 0xbe, 0xb6, 0x77, 0x46, 0x89, 0x9e, 0x53, 0xc6, + 0x63, 0x84, 0x15, 0x72, 0x90, 0xb3, 0x44, 0x94, 0x8d, 0x52, 0x7e, 0xcd, 0xff, 0xed, 0xd7, 0x37, + 0x7e, 0x78, 0xef, 0x82, 0x1a, 0xbd, 0xc4, 0xc1, 0x9b, 0x5a, 0xe8, 0xf6, 0x95, 0x5f, 0x79, 0x8f, + 0x0b, 0x89, 0x7f, 0xb8, 0xf0, 0x6a, 0xc8, 0xcb, 0xe5, 0xa3, 0xd8, 0xfa, 0xcd, 0xdc, 0x34, 0x39, + 0x75, 0xe6, 0xc8, 0xd2, 0x8b, 0xf9, 0x11, 0x35, 0xb9, 0x84, 0x58, 0xd8, 0x2b, 0xeb, 0xf5, 0x8d, + 0xde, 0xdd, 0xed, 0x57, 0x1f, 0xe7, 0xe0, 0x0d, 0x63, 0xd3, 0xdc, 0x55, 0x82, 0xb4, 0xd4, 0xf5, + 0xfe, 0xa9, 0xff, 0x47, 0x81, 0xea, 0x61, 0xe1, 0x77, 0x50, 0x3b, 0x2f, 0x8f, 0xba, 0xbe, 0xd5, + 0xa0, 0xa7, 0xa6, 0xc1, 0x44, 0xd0, 0xf9, 0x1d, 0x26, 0x08, 0x09, 0x1e, 0x25, 0x90, 0x7f, 0xc3, + 0x62, 0xb0, 0xdb, 0xe5, 0x90, 0xa9, 0x4d, 0xb0, 0x5f, 0xa1, 0x74, 0x29, 0x02, 0x13, 0xd4, 0x2a, + 0xd4, 0x18, 0x09, 0xbb, 0xb9, 0x5e, 0xdf, 0xe8, 0x06, 0xb7, 0xd4, 0x30, 0x1e, 0x6a, 0xe4, 0x74, + 0xea, 0x76, 0xbe, 0x86, 0x89, 0x3e, 0x50, 0x13, 0x85, 0x3f, 0x44, 0x9d, 0x42, 0x40, 0x9e, 0x28, + 0xf5, 0x72, 0x84, 0xab, 0xbe, 0x1d, 0x1a, 0x9c, 0x56, 0x11, 0xf8, 0x36, 0xaa, 0x17, 0x7c, 0x68, + 0x46, 0xb8, 0x67, 0x02, 0xeb, 0x87, 0xbb, 0x5f, 0x50, 0x85, 0x63, 0x0f, 0xb5, 0xa2, 0x3c, 0x2d, + 0x32, 0x61, 0x37, 0xb4, 0x39, 0x52, 0xe6, 0x5f, 0x6a, 0x84, 0x9a, 0x1b, 0x9c, 0xa0, 0x26, 0xfc, + 0x22, 0x73, 0x66, 0xb7, 0x74, 0xeb, 0x77, 0x5f, 0x6f, 0x5b, 0x91, 0xfb, 0x4a, 0xeb, 0x7e, 0x22, + 0xf3, 0xc9, 0xe2, 0x4b, 0x68, 0x8c, 0x96, 0x36, 0x7d, 0x40, 0x68, 0x11, 0x83, 0x6f, 0xa2, 0xfa, + 0x08, 0x26, 0xe5, 0xda, 0xa0, 0xea, 0x27, 0xfe, 0x1c, 0x35, 0xc7, 0xec, 0x49, 0x01, 0x66, 0x7b, + 0x7e, 0x70, 0x6d, 0x3e, 0x5a, 0xed, 0x3b, 0x45, 0xa1, 0x25, 0x73, 0x7b, 0x65, 0xcb, 0xf2, 0xfe, + 0xb4, 0x90, 0x7b, 0xcd, 0xce, 0xc3, 0x3f, 0x23, 0x14, 0xce, 0xf7, 0x88, 0xb0, 0x2d, 0x5d, 0xff, + 0xce, 0xab, 0xd7, 0x5f, 0xed, 0xa4, 0xc5, 0xdf, 0x43, 0x05, 0x09, 0xba, 0x64, 0x85, 0x37, 0x51, + 0x6f, 0x49, 0x5a, 0x57, 0xba, 0x1a, 0xdc, 0x98, 0x4d, 0xdd, 0xde, 0x92, 0x38, 0x5d, 0x8e, 0xf1, + 0x3e, 0x31, 0x6d, 0xd3, 0x85, 0x62, 0x77, 0xfe, 0x5e, 0x2c, 0xfd, 0x5d, 0xbb, 0xe7, 0xe7, 0x7d, + 0xbb, 0xf3, 0xdb, 0xef, 0x6e, 0xed, 0xe9, 0x5f, 0xeb, 0xb5, 0xe0, 0xce, 0xf1, 0x89, 0x53, 0x7b, + 0x76, 0xe2, 0xd4, 0x9e, 0x9f, 0x38, 0xb5, 0xa7, 0x33, 0xc7, 0x3a, 0x9e, 0x39, 0xd6, 0xb3, 0x99, + 0x63, 0x3d, 0x9f, 0x39, 0xd6, 0xdf, 0x33, 0xc7, 0xfa, 0xf5, 0x85, 0x53, 0xfb, 0xbe, 0x6d, 0xaa, + 0xfb, 0x37, 0x00, 0x00, 0xff, 0xff, 0x21, 0x97, 0x54, 0xe9, 0x99, 0x08, 0x00, 0x00, } func (m *CertificateSigningRequest) Marshal() (dAtA []byte, err error) { @@ -354,6 +360,21 @@ func (m *CertificateSigningRequestCondition) MarshalToSizedBuffer(dAtA []byte) ( _ = i var l int _ = l + i -= len(m.Status) + copy(dAtA[i:], m.Status) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Status))) + i-- + dAtA[i] = 0x32 + { + size, err := m.LastTransitionTime.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x2a { size, err := m.LastUpdateTime.MarshalToSizedBuffer(dAtA[:i]) if err != nil { @@ -449,6 +470,13 @@ func (m *CertificateSigningRequestSpec) MarshalToSizedBuffer(dAtA []byte) (int, _ = i var l int _ = l + if m.SignerName != nil { + i -= len(*m.SignerName) + copy(dAtA[i:], *m.SignerName) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.SignerName))) + i-- + dAtA[i] = 0x3a + } if len(m.Extra) > 0 { keysForExtra := make([]string, 0, len(m.Extra)) for k := range m.Extra { @@ -632,6 +660,10 @@ func (m *CertificateSigningRequestCondition) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) l = m.LastUpdateTime.Size() n += 1 + l + sovGenerated(uint64(l)) + l = m.LastTransitionTime.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Status) + n += 1 + l + sovGenerated(uint64(l)) return n } @@ -687,6 +719,10 @@ func (m *CertificateSigningRequestSpec) Size() (n int) { n += mapEntrySize + 1 + sovGenerated(uint64(mapEntrySize)) } } + if m.SignerName != nil { + l = len(*m.SignerName) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -751,6 +787,8 @@ func (this *CertificateSigningRequestCondition) String() string { `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`, `Message:` + fmt.Sprintf("%v", this.Message) + `,`, `LastUpdateTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastUpdateTime), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `LastTransitionTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastTransitionTime), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `Status:` + fmt.Sprintf("%v", this.Status) + `,`, `}`, }, "") return s @@ -792,6 +830,7 @@ func (this *CertificateSigningRequestSpec) String() string { `Groups:` + fmt.Sprintf("%v", this.Groups) + `,`, `Usages:` + fmt.Sprintf("%v", this.Usages) + `,`, `Extra:` + mapStringForExtra + `,`, + `SignerName:` + valueToStringGenerated(this.SignerName) + `,`, `}`, }, "") return s @@ -1130,6 +1169,71 @@ func (m *CertificateSigningRequestCondition) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field LastTransitionTime", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.LastTransitionTime.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 6: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Status = k8s_io_api_core_v1.ConditionStatus(dAtA[iNdEx:postIndex]) + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -1594,6 +1698,39 @@ func (m *CertificateSigningRequestSpec) Unmarshal(dAtA []byte) error { } m.Extra[mapkey] = *mapvalue iNdEx = postIndex + case 7: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field SignerName", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.SignerName = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -1827,6 +1964,7 @@ func (m *ExtraValue) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1858,10 +1996,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1882,55 +2018,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/certificates/v1beta1/generated.proto b/vendor/k8s.io/api/certificates/v1beta1/generated.proto index 5200224a..2fb4dc4e 100644 --- a/vendor/k8s.io/api/certificates/v1beta1/generated.proto +++ b/vendor/k8s.io/api/certificates/v1beta1/generated.proto @@ -21,6 +21,7 @@ syntax = 'proto2'; package k8s.io.api.certificates.v1beta1; +import "k8s.io/api/core/v1/generated.proto"; import "k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto"; import "k8s.io/apimachinery/pkg/runtime/generated.proto"; import "k8s.io/apimachinery/pkg/runtime/schema/generated.proto"; @@ -43,9 +44,16 @@ message CertificateSigningRequest { } message CertificateSigningRequestCondition { - // request approval state, currently Approved or Denied. + // type of the condition. Known conditions include "Approved", "Denied", and "Failed". optional string type = 1; + // Status of the condition, one of True, False, Unknown. + // Approved, Denied, and Failed conditions may not be "False" or "Unknown". + // Defaults to "True". + // If unset, should be treated as "True". + // +optional + optional string status = 6; + // brief reason for the request state // +optional optional string reason = 2; @@ -57,6 +65,12 @@ message CertificateSigningRequestCondition { // timestamp for the last update to this condition // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.Time lastUpdateTime = 4; + + // lastTransitionTime is the time the condition last transitioned from one status to another. + // If unset, when a new condition type is added or an existing condition's status is changed, + // the server defaults this to the current time. + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.Time lastTransitionTime = 5; } message CertificateSigningRequestList { @@ -71,12 +85,51 @@ message CertificateSigningRequestList { // Kubernetes and cannot be modified by users. message CertificateSigningRequestSpec { // Base64-encoded PKCS#10 CSR data + // +listType=atomic optional bytes request = 1; + // Requested signer for the request. It is a qualified name in the form: + // `scope-hostname.io/name`. + // If empty, it will be defaulted: + // 1. If it's a kubelet client certificate, it is assigned + // "kubernetes.io/kube-apiserver-client-kubelet". + // 2. If it's a kubelet serving certificate, it is assigned + // "kubernetes.io/kubelet-serving". + // 3. Otherwise, it is assigned "kubernetes.io/legacy-unknown". + // Distribution of trust for signers happens out of band. + // You can select on this field using `spec.signerName`. + // +optional + optional string signerName = 7; + // allowedUsages specifies a set of usage contexts the key will be // valid for. // See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3 // https://tools.ietf.org/html/rfc5280#section-4.2.1.12 + // Valid values are: + // "signing", + // "digital signature", + // "content commitment", + // "key encipherment", + // "key agreement", + // "data encipherment", + // "cert sign", + // "crl sign", + // "encipher only", + // "decipher only", + // "any", + // "server auth", + // "client auth", + // "code signing", + // "email protection", + // "s/mime", + // "ipsec end system", + // "ipsec tunnel", + // "ipsec user", + // "timestamping", + // "ocsp signing", + // "microsoft sgc", + // "netscape sgc" + // +listType=atomic repeated string usages = 5; // Information about the requesting user. @@ -91,6 +144,7 @@ message CertificateSigningRequestSpec { // Group information about the requesting user. // See user.Info interface for details. + // +listType=atomic // +optional repeated string groups = 4; @@ -102,10 +156,13 @@ message CertificateSigningRequestSpec { message CertificateSigningRequestStatus { // Conditions applied to the request, such as approval or denial. + // +listType=map + // +listMapKey=type // +optional repeated CertificateSigningRequestCondition conditions = 1; // If request was approved, the controller will place the issued certificate here. + // +listType=atomic // +optional optional bytes certificate = 2; } diff --git a/vendor/k8s.io/api/certificates/v1beta1/types.go b/vendor/k8s.io/api/certificates/v1beta1/types.go index 93f81cd5..9e61c67f 100644 --- a/vendor/k8s.io/api/certificates/v1beta1/types.go +++ b/vendor/k8s.io/api/certificates/v1beta1/types.go @@ -19,12 +19,16 @@ package v1beta1 import ( "fmt" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" ) // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.12 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=certificates.k8s.io,v1,CertificateSigningRequest // Describes a certificate signing request type CertificateSigningRequest struct { @@ -46,12 +50,51 @@ type CertificateSigningRequest struct { // Kubernetes and cannot be modified by users. type CertificateSigningRequestSpec struct { // Base64-encoded PKCS#10 CSR data + // +listType=atomic Request []byte `json:"request" protobuf:"bytes,1,opt,name=request"` + // Requested signer for the request. It is a qualified name in the form: + // `scope-hostname.io/name`. + // If empty, it will be defaulted: + // 1. If it's a kubelet client certificate, it is assigned + // "kubernetes.io/kube-apiserver-client-kubelet". + // 2. If it's a kubelet serving certificate, it is assigned + // "kubernetes.io/kubelet-serving". + // 3. Otherwise, it is assigned "kubernetes.io/legacy-unknown". + // Distribution of trust for signers happens out of band. + // You can select on this field using `spec.signerName`. + // +optional + SignerName *string `json:"signerName,omitempty" protobuf:"bytes,7,opt,name=signerName"` + // allowedUsages specifies a set of usage contexts the key will be // valid for. // See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3 // https://tools.ietf.org/html/rfc5280#section-4.2.1.12 + // Valid values are: + // "signing", + // "digital signature", + // "content commitment", + // "key encipherment", + // "key agreement", + // "data encipherment", + // "cert sign", + // "crl sign", + // "encipher only", + // "decipher only", + // "any", + // "server auth", + // "client auth", + // "code signing", + // "email protection", + // "s/mime", + // "ipsec end system", + // "ipsec tunnel", + // "ipsec user", + // "timestamping", + // "ocsp signing", + // "microsoft sgc", + // "netscape sgc" + // +listType=atomic Usages []KeyUsage `json:"usages,omitempty" protobuf:"bytes,5,opt,name=usages"` // Information about the requesting user. @@ -64,6 +107,7 @@ type CertificateSigningRequestSpec struct { UID string `json:"uid,omitempty" protobuf:"bytes,3,opt,name=uid"` // Group information about the requesting user. // See user.Info interface for details. + // +listType=atomic // +optional Groups []string `json:"groups,omitempty" protobuf:"bytes,4,rep,name=groups"` // Extra information about the requesting user. @@ -72,6 +116,28 @@ type CertificateSigningRequestSpec struct { Extra map[string]ExtraValue `json:"extra,omitempty" protobuf:"bytes,6,rep,name=extra"` } +// Built in signerName values that are honoured by kube-controller-manager. +// None of these usages are related to ServiceAccount token secrets +// `.data[ca.crt]` in any way. +const ( + // Signs certificates that will be honored as client-certs by the + // kube-apiserver. Never auto-approved by kube-controller-manager. + KubeAPIServerClientSignerName = "kubernetes.io/kube-apiserver-client" + + // Signs client certificates that will be honored as client-certs by the + // kube-apiserver for a kubelet. + // May be auto-approved by kube-controller-manager. + KubeAPIServerClientKubeletSignerName = "kubernetes.io/kube-apiserver-client-kubelet" + + // Signs serving certificates that are honored as a valid kubelet serving + // certificate by the kube-apiserver, but has no other guarantees. + KubeletServingSignerName = "kubernetes.io/kubelet-serving" + + // Has no guarantees for trust at all. Some distributions may honor these + // as client certs, but that behavior is not standard kubernetes behavior. + LegacyUnknownSignerName = "kubernetes.io/legacy-unknown" +) + // ExtraValue masks the value so protobuf can generate // +protobuf.nullable=true // +protobuf.options.(gogoproto.goproto_stringer)=false @@ -83,10 +149,13 @@ func (t ExtraValue) String() string { type CertificateSigningRequestStatus struct { // Conditions applied to the request, such as approval or denial. + // +listType=map + // +listMapKey=type // +optional Conditions []CertificateSigningRequestCondition `json:"conditions,omitempty" protobuf:"bytes,1,rep,name=conditions"` // If request was approved, the controller will place the issued certificate here. + // +listType=atomic // +optional Certificate []byte `json:"certificate,omitempty" protobuf:"bytes,2,opt,name=certificate"` } @@ -97,11 +166,18 @@ type RequestConditionType string const ( CertificateApproved RequestConditionType = "Approved" CertificateDenied RequestConditionType = "Denied" + CertificateFailed RequestConditionType = "Failed" ) type CertificateSigningRequestCondition struct { - // request approval state, currently Approved or Denied. + // type of the condition. Known conditions include "Approved", "Denied", and "Failed". Type RequestConditionType `json:"type" protobuf:"bytes,1,opt,name=type,casttype=RequestConditionType"` + // Status of the condition, one of True, False, Unknown. + // Approved, Denied, and Failed conditions may not be "False" or "Unknown". + // Defaults to "True". + // If unset, should be treated as "True". + // +optional + Status v1.ConditionStatus `json:"status" protobuf:"bytes,6,opt,name=status,casttype=k8s.io/api/core/v1.ConditionStatus"` // brief reason for the request state // +optional Reason string `json:"reason,omitempty" protobuf:"bytes,2,opt,name=reason"` @@ -111,9 +187,17 @@ type CertificateSigningRequestCondition struct { // timestamp for the last update to this condition // +optional LastUpdateTime metav1.Time `json:"lastUpdateTime,omitempty" protobuf:"bytes,4,opt,name=lastUpdateTime"` + // lastTransitionTime is the time the condition last transitioned from one status to another. + // If unset, when a new condition type is added or an existing condition's status is changed, + // the server defaults this to the current time. + // +optional + LastTransitionTime metav1.Time `json:"lastTransitionTime,omitempty" protobuf:"bytes,5,opt,name=lastTransitionTime"` } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.12 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=certificates.k8s.io,v1,CertificateSigningRequestList type CertificateSigningRequestList struct { metav1.TypeMeta `json:",inline"` diff --git a/vendor/k8s.io/api/certificates/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/certificates/v1beta1/types_swagger_doc_generated.go index f6a7e16a..396fee0b 100644 --- a/vendor/k8s.io/api/certificates/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/certificates/v1beta1/types_swagger_doc_generated.go @@ -38,10 +38,12 @@ func (CertificateSigningRequest) SwaggerDoc() map[string]string { } var map_CertificateSigningRequestCondition = map[string]string{ - "type": "request approval state, currently Approved or Denied.", - "reason": "brief reason for the request state", - "message": "human readable message with details about the request state", - "lastUpdateTime": "timestamp for the last update to this condition", + "type": "type of the condition. Known conditions include \"Approved\", \"Denied\", and \"Failed\".", + "status": "Status of the condition, one of True, False, Unknown. Approved, Denied, and Failed conditions may not be \"False\" or \"Unknown\". Defaults to \"True\". If unset, should be treated as \"True\".", + "reason": "brief reason for the request state", + "message": "human readable message with details about the request state", + "lastUpdateTime": "timestamp for the last update to this condition", + "lastTransitionTime": "lastTransitionTime is the time the condition last transitioned from one status to another. If unset, when a new condition type is added or an existing condition's status is changed, the server defaults this to the current time.", } func (CertificateSigningRequestCondition) SwaggerDoc() map[string]string { @@ -49,13 +51,14 @@ func (CertificateSigningRequestCondition) SwaggerDoc() map[string]string { } var map_CertificateSigningRequestSpec = map[string]string{ - "": "This information is immutable after the request is created. Only the Request and Usages fields can be set on creation, other fields are derived by Kubernetes and cannot be modified by users.", - "request": "Base64-encoded PKCS#10 CSR data", - "usages": "allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3\n https://tools.ietf.org/html/rfc5280#section-4.2.1.12", - "username": "Information about the requesting user. See user.Info interface for details.", - "uid": "UID information about the requesting user. See user.Info interface for details.", - "groups": "Group information about the requesting user. See user.Info interface for details.", - "extra": "Extra information about the requesting user. See user.Info interface for details.", + "": "This information is immutable after the request is created. Only the Request and Usages fields can be set on creation, other fields are derived by Kubernetes and cannot be modified by users.", + "request": "Base64-encoded PKCS#10 CSR data", + "signerName": "Requested signer for the request. It is a qualified name in the form: `scope-hostname.io/name`. If empty, it will be defaulted:\n 1. If it's a kubelet client certificate, it is assigned\n \"kubernetes.io/kube-apiserver-client-kubelet\".\n 2. If it's a kubelet serving certificate, it is assigned\n \"kubernetes.io/kubelet-serving\".\n 3. Otherwise, it is assigned \"kubernetes.io/legacy-unknown\".\nDistribution of trust for signers happens out of band. You can select on this field using `spec.signerName`.", + "usages": "allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3\n https://tools.ietf.org/html/rfc5280#section-4.2.1.12\nValid values are:\n \"signing\",\n \"digital signature\",\n \"content commitment\",\n \"key encipherment\",\n \"key agreement\",\n \"data encipherment\",\n \"cert sign\",\n \"crl sign\",\n \"encipher only\",\n \"decipher only\",\n \"any\",\n \"server auth\",\n \"client auth\",\n \"code signing\",\n \"email protection\",\n \"s/mime\",\n \"ipsec end system\",\n \"ipsec tunnel\",\n \"ipsec user\",\n \"timestamping\",\n \"ocsp signing\",\n \"microsoft sgc\",\n \"netscape sgc\"", + "username": "Information about the requesting user. See user.Info interface for details.", + "uid": "UID information about the requesting user. See user.Info interface for details.", + "groups": "Group information about the requesting user. See user.Info interface for details.", + "extra": "Extra information about the requesting user. See user.Info interface for details.", } func (CertificateSigningRequestSpec) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/certificates/v1beta1/zz_generated.deepcopy.go b/vendor/k8s.io/api/certificates/v1beta1/zz_generated.deepcopy.go index b3e0aeb5..0463f5bb 100644 --- a/vendor/k8s.io/api/certificates/v1beta1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/certificates/v1beta1/zz_generated.deepcopy.go @@ -56,6 +56,7 @@ func (in *CertificateSigningRequest) DeepCopyObject() runtime.Object { func (in *CertificateSigningRequestCondition) DeepCopyInto(out *CertificateSigningRequestCondition) { *out = *in in.LastUpdateTime.DeepCopyInto(&out.LastUpdateTime) + in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime) return } @@ -110,6 +111,11 @@ func (in *CertificateSigningRequestSpec) DeepCopyInto(out *CertificateSigningReq *out = make([]byte, len(*in)) copy(*out, *in) } + if in.SignerName != nil { + in, out := &in.SignerName, &out.SignerName + *out = new(string) + **out = **in + } if in.Usages != nil { in, out := &in.Usages, &out.Usages *out = make([]KeyUsage, len(*in)) diff --git a/vendor/k8s.io/api/certificates/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/certificates/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..6ccebfe6 --- /dev/null +++ b/vendor/k8s.io/api/certificates/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,73 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CertificateSigningRequest) APILifecycleIntroduced() (major, minor int) { + return 1, 12 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CertificateSigningRequest) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CertificateSigningRequest) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "certificates.k8s.io", Version: "v1", Kind: "CertificateSigningRequest"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CertificateSigningRequest) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CertificateSigningRequestList) APILifecycleIntroduced() (major, minor int) { + return 1, 12 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CertificateSigningRequestList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CertificateSigningRequestList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "certificates.k8s.io", Version: "v1", Kind: "CertificateSigningRequestList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CertificateSigningRequestList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/coordination/v1/generated.pb.go b/vendor/k8s.io/api/coordination/v1/generated.pb.go index 7e78be19..22c3d624 100644 --- a/vendor/k8s.io/api/coordination/v1/generated.pb.go +++ b/vendor/k8s.io/api/coordination/v1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Lease) Reset() { *m = Lease{} } func (*Lease) ProtoMessage() {} @@ -893,6 +893,7 @@ func (m *LeaseSpec) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -924,10 +925,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -948,55 +947,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/coordination/v1beta1/doc.go b/vendor/k8s.io/api/coordination/v1beta1/doc.go index 304732d5..e733411a 100644 --- a/vendor/k8s.io/api/coordination/v1beta1/doc.go +++ b/vendor/k8s.io/api/coordination/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=coordination.k8s.io diff --git a/vendor/k8s.io/api/coordination/v1beta1/generated.pb.go b/vendor/k8s.io/api/coordination/v1beta1/generated.pb.go index 2463d625..57a314cf 100644 --- a/vendor/k8s.io/api/coordination/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/coordination/v1beta1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Lease) Reset() { *m = Lease{} } func (*Lease) ProtoMessage() {} @@ -893,6 +893,7 @@ func (m *LeaseSpec) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -924,10 +925,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -948,55 +947,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/coordination/v1beta1/types.go b/vendor/k8s.io/api/coordination/v1beta1/types.go index da88f675..8f300fca 100644 --- a/vendor/k8s.io/api/coordination/v1beta1/types.go +++ b/vendor/k8s.io/api/coordination/v1beta1/types.go @@ -22,6 +22,9 @@ import ( // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.12 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=coordination.k8s.io,v1,Lease // Lease defines a lease concept. type Lease struct { @@ -60,6 +63,9 @@ type LeaseSpec struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.12 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=coordination.k8s.io,v1,LeaseList // LeaseList is a list of Lease objects. type LeaseList struct { diff --git a/vendor/k8s.io/api/coordination/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/coordination/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..9d18b922 --- /dev/null +++ b/vendor/k8s.io/api/coordination/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,73 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Lease) APILifecycleIntroduced() (major, minor int) { + return 1, 12 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Lease) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Lease) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "coordination.k8s.io", Version: "v1", Kind: "Lease"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Lease) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *LeaseList) APILifecycleIntroduced() (major, minor int) { + return 1, 12 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *LeaseList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *LeaseList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "coordination.k8s.io", Version: "v1", Kind: "LeaseList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *LeaseList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/core/v1/annotation_key_constants.go b/vendor/k8s.io/api/core/v1/annotation_key_constants.go index edc9b4d6..d3ebf862 100644 --- a/vendor/k8s.io/api/core/v1/annotation_key_constants.go +++ b/vendor/k8s.io/api/core/v1/annotation_key_constants.go @@ -39,17 +39,42 @@ const ( // SeccompPodAnnotationKey represents the key of a seccomp profile applied // to all containers of a pod. + // Deprecated: set a pod security context `seccompProfile` field. SeccompPodAnnotationKey string = "seccomp.security.alpha.kubernetes.io/pod" // SeccompContainerAnnotationKeyPrefix represents the key of a seccomp profile applied // to one container of a pod. + // Deprecated: set a container security context `seccompProfile` field. SeccompContainerAnnotationKeyPrefix string = "container.seccomp.security.alpha.kubernetes.io/" // SeccompProfileRuntimeDefault represents the default seccomp profile used by container runtime. + // Deprecated: set a pod or container security context `seccompProfile` of type "RuntimeDefault" instead. SeccompProfileRuntimeDefault string = "runtime/default" + // SeccompProfileNameUnconfined is the unconfined seccomp profile. + SeccompProfileNameUnconfined string = "unconfined" + + // SeccompLocalhostProfileNamePrefix is the prefix for specifying profiles loaded from the node's disk. + SeccompLocalhostProfileNamePrefix = "localhost/" + + // AppArmorBetaContainerAnnotationKeyPrefix is the prefix to an annotation key specifying a container's apparmor profile. + AppArmorBetaContainerAnnotationKeyPrefix = "container.apparmor.security.beta.kubernetes.io/" + // AppArmorBetaDefaultProfileAnnotatoinKey is the annotation key specifying the default AppArmor profile. + AppArmorBetaDefaultProfileAnnotationKey = "apparmor.security.beta.kubernetes.io/defaultProfileName" + // AppArmorBetaAllowedProfileAnnotationKey is the annotation key specifying the allowed AppArmor profiles. + AppArmorBetaAllowedProfilesAnnotationKey = "apparmor.security.beta.kubernetes.io/allowedProfileNames" + + // AppArmorBetaProfileRuntimeDefault is the profile specifying the runtime default. + AppArmorBetaProfileRuntimeDefault = "runtime/default" + + // AppArmorBetaProfileNamePrefix is the prefix for specifying profiles loaded on the node. + AppArmorBetaProfileNamePrefix = "localhost/" + + // AppArmorBetaProfileNameUnconfined is the Unconfined AppArmor profile + AppArmorBetaProfileNameUnconfined = "unconfined" + // DeprecatedSeccompProfileDockerDefault represents the default seccomp profile used by docker. - // This is now deprecated and should be replaced by SeccompProfileRuntimeDefault. + // Deprecated: set a pod or container security context `seccompProfile` of type "RuntimeDefault" instead. DeprecatedSeccompProfileDockerDefault string = "docker/default" // PreferAvoidPodsAnnotationKey represents the key of preferAvoidPods data (json serialized) diff --git a/vendor/k8s.io/api/core/v1/generated.pb.go b/vendor/k8s.io/api/core/v1/generated.pb.go index 732385ce..9b29b21e 100644 --- a/vendor/k8s.io/api/core/v1/generated.pb.go +++ b/vendor/k8s.io/api/core/v1/generated.pb.go @@ -47,7 +47,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *AWSElasticBlockStoreVolumeSource) Reset() { *m = AWSElasticBlockStoreVolumeSource{} } func (*AWSElasticBlockStoreVolumeSource) ProtoMessage() {} @@ -1449,10 +1449,38 @@ func (m *EphemeralContainers) XXX_DiscardUnknown() { var xxx_messageInfo_EphemeralContainers proto.InternalMessageInfo +func (m *EphemeralVolumeSource) Reset() { *m = EphemeralVolumeSource{} } +func (*EphemeralVolumeSource) ProtoMessage() {} +func (*EphemeralVolumeSource) Descriptor() ([]byte, []int) { + return fileDescriptor_83c10c24ec417dc9, []int{50} +} +func (m *EphemeralVolumeSource) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *EphemeralVolumeSource) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *EphemeralVolumeSource) XXX_Merge(src proto.Message) { + xxx_messageInfo_EphemeralVolumeSource.Merge(m, src) +} +func (m *EphemeralVolumeSource) XXX_Size() int { + return m.Size() +} +func (m *EphemeralVolumeSource) XXX_DiscardUnknown() { + xxx_messageInfo_EphemeralVolumeSource.DiscardUnknown(m) +} + +var xxx_messageInfo_EphemeralVolumeSource proto.InternalMessageInfo + func (m *Event) Reset() { *m = Event{} } func (*Event) ProtoMessage() {} func (*Event) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{50} + return fileDescriptor_83c10c24ec417dc9, []int{51} } func (m *Event) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1480,7 +1508,7 @@ var xxx_messageInfo_Event proto.InternalMessageInfo func (m *EventList) Reset() { *m = EventList{} } func (*EventList) ProtoMessage() {} func (*EventList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{51} + return fileDescriptor_83c10c24ec417dc9, []int{52} } func (m *EventList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1508,7 +1536,7 @@ var xxx_messageInfo_EventList proto.InternalMessageInfo func (m *EventSeries) Reset() { *m = EventSeries{} } func (*EventSeries) ProtoMessage() {} func (*EventSeries) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{52} + return fileDescriptor_83c10c24ec417dc9, []int{53} } func (m *EventSeries) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1536,7 +1564,7 @@ var xxx_messageInfo_EventSeries proto.InternalMessageInfo func (m *EventSource) Reset() { *m = EventSource{} } func (*EventSource) ProtoMessage() {} func (*EventSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{53} + return fileDescriptor_83c10c24ec417dc9, []int{54} } func (m *EventSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1564,7 +1592,7 @@ var xxx_messageInfo_EventSource proto.InternalMessageInfo func (m *ExecAction) Reset() { *m = ExecAction{} } func (*ExecAction) ProtoMessage() {} func (*ExecAction) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{54} + return fileDescriptor_83c10c24ec417dc9, []int{55} } func (m *ExecAction) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1592,7 +1620,7 @@ var xxx_messageInfo_ExecAction proto.InternalMessageInfo func (m *FCVolumeSource) Reset() { *m = FCVolumeSource{} } func (*FCVolumeSource) ProtoMessage() {} func (*FCVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{55} + return fileDescriptor_83c10c24ec417dc9, []int{56} } func (m *FCVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1620,7 +1648,7 @@ var xxx_messageInfo_FCVolumeSource proto.InternalMessageInfo func (m *FlexPersistentVolumeSource) Reset() { *m = FlexPersistentVolumeSource{} } func (*FlexPersistentVolumeSource) ProtoMessage() {} func (*FlexPersistentVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{56} + return fileDescriptor_83c10c24ec417dc9, []int{57} } func (m *FlexPersistentVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1648,7 +1676,7 @@ var xxx_messageInfo_FlexPersistentVolumeSource proto.InternalMessageInfo func (m *FlexVolumeSource) Reset() { *m = FlexVolumeSource{} } func (*FlexVolumeSource) ProtoMessage() {} func (*FlexVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{57} + return fileDescriptor_83c10c24ec417dc9, []int{58} } func (m *FlexVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1676,7 +1704,7 @@ var xxx_messageInfo_FlexVolumeSource proto.InternalMessageInfo func (m *FlockerVolumeSource) Reset() { *m = FlockerVolumeSource{} } func (*FlockerVolumeSource) ProtoMessage() {} func (*FlockerVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{58} + return fileDescriptor_83c10c24ec417dc9, []int{59} } func (m *FlockerVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1704,7 +1732,7 @@ var xxx_messageInfo_FlockerVolumeSource proto.InternalMessageInfo func (m *GCEPersistentDiskVolumeSource) Reset() { *m = GCEPersistentDiskVolumeSource{} } func (*GCEPersistentDiskVolumeSource) ProtoMessage() {} func (*GCEPersistentDiskVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{59} + return fileDescriptor_83c10c24ec417dc9, []int{60} } func (m *GCEPersistentDiskVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1732,7 +1760,7 @@ var xxx_messageInfo_GCEPersistentDiskVolumeSource proto.InternalMessageInfo func (m *GitRepoVolumeSource) Reset() { *m = GitRepoVolumeSource{} } func (*GitRepoVolumeSource) ProtoMessage() {} func (*GitRepoVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{60} + return fileDescriptor_83c10c24ec417dc9, []int{61} } func (m *GitRepoVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1760,7 +1788,7 @@ var xxx_messageInfo_GitRepoVolumeSource proto.InternalMessageInfo func (m *GlusterfsPersistentVolumeSource) Reset() { *m = GlusterfsPersistentVolumeSource{} } func (*GlusterfsPersistentVolumeSource) ProtoMessage() {} func (*GlusterfsPersistentVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{61} + return fileDescriptor_83c10c24ec417dc9, []int{62} } func (m *GlusterfsPersistentVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1788,7 +1816,7 @@ var xxx_messageInfo_GlusterfsPersistentVolumeSource proto.InternalMessageInfo func (m *GlusterfsVolumeSource) Reset() { *m = GlusterfsVolumeSource{} } func (*GlusterfsVolumeSource) ProtoMessage() {} func (*GlusterfsVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{62} + return fileDescriptor_83c10c24ec417dc9, []int{63} } func (m *GlusterfsVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1816,7 +1844,7 @@ var xxx_messageInfo_GlusterfsVolumeSource proto.InternalMessageInfo func (m *HTTPGetAction) Reset() { *m = HTTPGetAction{} } func (*HTTPGetAction) ProtoMessage() {} func (*HTTPGetAction) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{63} + return fileDescriptor_83c10c24ec417dc9, []int{64} } func (m *HTTPGetAction) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1844,7 +1872,7 @@ var xxx_messageInfo_HTTPGetAction proto.InternalMessageInfo func (m *HTTPHeader) Reset() { *m = HTTPHeader{} } func (*HTTPHeader) ProtoMessage() {} func (*HTTPHeader) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{64} + return fileDescriptor_83c10c24ec417dc9, []int{65} } func (m *HTTPHeader) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1872,7 +1900,7 @@ var xxx_messageInfo_HTTPHeader proto.InternalMessageInfo func (m *Handler) Reset() { *m = Handler{} } func (*Handler) ProtoMessage() {} func (*Handler) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{65} + return fileDescriptor_83c10c24ec417dc9, []int{66} } func (m *Handler) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1900,7 +1928,7 @@ var xxx_messageInfo_Handler proto.InternalMessageInfo func (m *HostAlias) Reset() { *m = HostAlias{} } func (*HostAlias) ProtoMessage() {} func (*HostAlias) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{66} + return fileDescriptor_83c10c24ec417dc9, []int{67} } func (m *HostAlias) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1928,7 +1956,7 @@ var xxx_messageInfo_HostAlias proto.InternalMessageInfo func (m *HostPathVolumeSource) Reset() { *m = HostPathVolumeSource{} } func (*HostPathVolumeSource) ProtoMessage() {} func (*HostPathVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{67} + return fileDescriptor_83c10c24ec417dc9, []int{68} } func (m *HostPathVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1956,7 +1984,7 @@ var xxx_messageInfo_HostPathVolumeSource proto.InternalMessageInfo func (m *ISCSIPersistentVolumeSource) Reset() { *m = ISCSIPersistentVolumeSource{} } func (*ISCSIPersistentVolumeSource) ProtoMessage() {} func (*ISCSIPersistentVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{68} + return fileDescriptor_83c10c24ec417dc9, []int{69} } func (m *ISCSIPersistentVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1984,7 +2012,7 @@ var xxx_messageInfo_ISCSIPersistentVolumeSource proto.InternalMessageInfo func (m *ISCSIVolumeSource) Reset() { *m = ISCSIVolumeSource{} } func (*ISCSIVolumeSource) ProtoMessage() {} func (*ISCSIVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{69} + return fileDescriptor_83c10c24ec417dc9, []int{70} } func (m *ISCSIVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2012,7 +2040,7 @@ var xxx_messageInfo_ISCSIVolumeSource proto.InternalMessageInfo func (m *KeyToPath) Reset() { *m = KeyToPath{} } func (*KeyToPath) ProtoMessage() {} func (*KeyToPath) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{70} + return fileDescriptor_83c10c24ec417dc9, []int{71} } func (m *KeyToPath) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2040,7 +2068,7 @@ var xxx_messageInfo_KeyToPath proto.InternalMessageInfo func (m *Lifecycle) Reset() { *m = Lifecycle{} } func (*Lifecycle) ProtoMessage() {} func (*Lifecycle) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{71} + return fileDescriptor_83c10c24ec417dc9, []int{72} } func (m *Lifecycle) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2068,7 +2096,7 @@ var xxx_messageInfo_Lifecycle proto.InternalMessageInfo func (m *LimitRange) Reset() { *m = LimitRange{} } func (*LimitRange) ProtoMessage() {} func (*LimitRange) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{72} + return fileDescriptor_83c10c24ec417dc9, []int{73} } func (m *LimitRange) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2096,7 +2124,7 @@ var xxx_messageInfo_LimitRange proto.InternalMessageInfo func (m *LimitRangeItem) Reset() { *m = LimitRangeItem{} } func (*LimitRangeItem) ProtoMessage() {} func (*LimitRangeItem) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{73} + return fileDescriptor_83c10c24ec417dc9, []int{74} } func (m *LimitRangeItem) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2124,7 +2152,7 @@ var xxx_messageInfo_LimitRangeItem proto.InternalMessageInfo func (m *LimitRangeList) Reset() { *m = LimitRangeList{} } func (*LimitRangeList) ProtoMessage() {} func (*LimitRangeList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{74} + return fileDescriptor_83c10c24ec417dc9, []int{75} } func (m *LimitRangeList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2152,7 +2180,7 @@ var xxx_messageInfo_LimitRangeList proto.InternalMessageInfo func (m *LimitRangeSpec) Reset() { *m = LimitRangeSpec{} } func (*LimitRangeSpec) ProtoMessage() {} func (*LimitRangeSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{75} + return fileDescriptor_83c10c24ec417dc9, []int{76} } func (m *LimitRangeSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2180,7 +2208,7 @@ var xxx_messageInfo_LimitRangeSpec proto.InternalMessageInfo func (m *List) Reset() { *m = List{} } func (*List) ProtoMessage() {} func (*List) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{76} + return fileDescriptor_83c10c24ec417dc9, []int{77} } func (m *List) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2208,7 +2236,7 @@ var xxx_messageInfo_List proto.InternalMessageInfo func (m *LoadBalancerIngress) Reset() { *m = LoadBalancerIngress{} } func (*LoadBalancerIngress) ProtoMessage() {} func (*LoadBalancerIngress) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{77} + return fileDescriptor_83c10c24ec417dc9, []int{78} } func (m *LoadBalancerIngress) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2236,7 +2264,7 @@ var xxx_messageInfo_LoadBalancerIngress proto.InternalMessageInfo func (m *LoadBalancerStatus) Reset() { *m = LoadBalancerStatus{} } func (*LoadBalancerStatus) ProtoMessage() {} func (*LoadBalancerStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{78} + return fileDescriptor_83c10c24ec417dc9, []int{79} } func (m *LoadBalancerStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2264,7 +2292,7 @@ var xxx_messageInfo_LoadBalancerStatus proto.InternalMessageInfo func (m *LocalObjectReference) Reset() { *m = LocalObjectReference{} } func (*LocalObjectReference) ProtoMessage() {} func (*LocalObjectReference) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{79} + return fileDescriptor_83c10c24ec417dc9, []int{80} } func (m *LocalObjectReference) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2292,7 +2320,7 @@ var xxx_messageInfo_LocalObjectReference proto.InternalMessageInfo func (m *LocalVolumeSource) Reset() { *m = LocalVolumeSource{} } func (*LocalVolumeSource) ProtoMessage() {} func (*LocalVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{80} + return fileDescriptor_83c10c24ec417dc9, []int{81} } func (m *LocalVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2320,7 +2348,7 @@ var xxx_messageInfo_LocalVolumeSource proto.InternalMessageInfo func (m *NFSVolumeSource) Reset() { *m = NFSVolumeSource{} } func (*NFSVolumeSource) ProtoMessage() {} func (*NFSVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{81} + return fileDescriptor_83c10c24ec417dc9, []int{82} } func (m *NFSVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2348,7 +2376,7 @@ var xxx_messageInfo_NFSVolumeSource proto.InternalMessageInfo func (m *Namespace) Reset() { *m = Namespace{} } func (*Namespace) ProtoMessage() {} func (*Namespace) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{82} + return fileDescriptor_83c10c24ec417dc9, []int{83} } func (m *Namespace) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2376,7 +2404,7 @@ var xxx_messageInfo_Namespace proto.InternalMessageInfo func (m *NamespaceCondition) Reset() { *m = NamespaceCondition{} } func (*NamespaceCondition) ProtoMessage() {} func (*NamespaceCondition) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{83} + return fileDescriptor_83c10c24ec417dc9, []int{84} } func (m *NamespaceCondition) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2404,7 +2432,7 @@ var xxx_messageInfo_NamespaceCondition proto.InternalMessageInfo func (m *NamespaceList) Reset() { *m = NamespaceList{} } func (*NamespaceList) ProtoMessage() {} func (*NamespaceList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{84} + return fileDescriptor_83c10c24ec417dc9, []int{85} } func (m *NamespaceList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2432,7 +2460,7 @@ var xxx_messageInfo_NamespaceList proto.InternalMessageInfo func (m *NamespaceSpec) Reset() { *m = NamespaceSpec{} } func (*NamespaceSpec) ProtoMessage() {} func (*NamespaceSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{85} + return fileDescriptor_83c10c24ec417dc9, []int{86} } func (m *NamespaceSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2460,7 +2488,7 @@ var xxx_messageInfo_NamespaceSpec proto.InternalMessageInfo func (m *NamespaceStatus) Reset() { *m = NamespaceStatus{} } func (*NamespaceStatus) ProtoMessage() {} func (*NamespaceStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{86} + return fileDescriptor_83c10c24ec417dc9, []int{87} } func (m *NamespaceStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2488,7 +2516,7 @@ var xxx_messageInfo_NamespaceStatus proto.InternalMessageInfo func (m *Node) Reset() { *m = Node{} } func (*Node) ProtoMessage() {} func (*Node) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{87} + return fileDescriptor_83c10c24ec417dc9, []int{88} } func (m *Node) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2516,7 +2544,7 @@ var xxx_messageInfo_Node proto.InternalMessageInfo func (m *NodeAddress) Reset() { *m = NodeAddress{} } func (*NodeAddress) ProtoMessage() {} func (*NodeAddress) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{88} + return fileDescriptor_83c10c24ec417dc9, []int{89} } func (m *NodeAddress) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2544,7 +2572,7 @@ var xxx_messageInfo_NodeAddress proto.InternalMessageInfo func (m *NodeAffinity) Reset() { *m = NodeAffinity{} } func (*NodeAffinity) ProtoMessage() {} func (*NodeAffinity) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{89} + return fileDescriptor_83c10c24ec417dc9, []int{90} } func (m *NodeAffinity) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2572,7 +2600,7 @@ var xxx_messageInfo_NodeAffinity proto.InternalMessageInfo func (m *NodeCondition) Reset() { *m = NodeCondition{} } func (*NodeCondition) ProtoMessage() {} func (*NodeCondition) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{90} + return fileDescriptor_83c10c24ec417dc9, []int{91} } func (m *NodeCondition) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2600,7 +2628,7 @@ var xxx_messageInfo_NodeCondition proto.InternalMessageInfo func (m *NodeConfigSource) Reset() { *m = NodeConfigSource{} } func (*NodeConfigSource) ProtoMessage() {} func (*NodeConfigSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{91} + return fileDescriptor_83c10c24ec417dc9, []int{92} } func (m *NodeConfigSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2628,7 +2656,7 @@ var xxx_messageInfo_NodeConfigSource proto.InternalMessageInfo func (m *NodeConfigStatus) Reset() { *m = NodeConfigStatus{} } func (*NodeConfigStatus) ProtoMessage() {} func (*NodeConfigStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{92} + return fileDescriptor_83c10c24ec417dc9, []int{93} } func (m *NodeConfigStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2656,7 +2684,7 @@ var xxx_messageInfo_NodeConfigStatus proto.InternalMessageInfo func (m *NodeDaemonEndpoints) Reset() { *m = NodeDaemonEndpoints{} } func (*NodeDaemonEndpoints) ProtoMessage() {} func (*NodeDaemonEndpoints) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{93} + return fileDescriptor_83c10c24ec417dc9, []int{94} } func (m *NodeDaemonEndpoints) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2684,7 +2712,7 @@ var xxx_messageInfo_NodeDaemonEndpoints proto.InternalMessageInfo func (m *NodeList) Reset() { *m = NodeList{} } func (*NodeList) ProtoMessage() {} func (*NodeList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{94} + return fileDescriptor_83c10c24ec417dc9, []int{95} } func (m *NodeList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2712,7 +2740,7 @@ var xxx_messageInfo_NodeList proto.InternalMessageInfo func (m *NodeProxyOptions) Reset() { *m = NodeProxyOptions{} } func (*NodeProxyOptions) ProtoMessage() {} func (*NodeProxyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{95} + return fileDescriptor_83c10c24ec417dc9, []int{96} } func (m *NodeProxyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2740,7 +2768,7 @@ var xxx_messageInfo_NodeProxyOptions proto.InternalMessageInfo func (m *NodeResources) Reset() { *m = NodeResources{} } func (*NodeResources) ProtoMessage() {} func (*NodeResources) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{96} + return fileDescriptor_83c10c24ec417dc9, []int{97} } func (m *NodeResources) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2768,7 +2796,7 @@ var xxx_messageInfo_NodeResources proto.InternalMessageInfo func (m *NodeSelector) Reset() { *m = NodeSelector{} } func (*NodeSelector) ProtoMessage() {} func (*NodeSelector) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{97} + return fileDescriptor_83c10c24ec417dc9, []int{98} } func (m *NodeSelector) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2796,7 +2824,7 @@ var xxx_messageInfo_NodeSelector proto.InternalMessageInfo func (m *NodeSelectorRequirement) Reset() { *m = NodeSelectorRequirement{} } func (*NodeSelectorRequirement) ProtoMessage() {} func (*NodeSelectorRequirement) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{98} + return fileDescriptor_83c10c24ec417dc9, []int{99} } func (m *NodeSelectorRequirement) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2824,7 +2852,7 @@ var xxx_messageInfo_NodeSelectorRequirement proto.InternalMessageInfo func (m *NodeSelectorTerm) Reset() { *m = NodeSelectorTerm{} } func (*NodeSelectorTerm) ProtoMessage() {} func (*NodeSelectorTerm) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{99} + return fileDescriptor_83c10c24ec417dc9, []int{100} } func (m *NodeSelectorTerm) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2852,7 +2880,7 @@ var xxx_messageInfo_NodeSelectorTerm proto.InternalMessageInfo func (m *NodeSpec) Reset() { *m = NodeSpec{} } func (*NodeSpec) ProtoMessage() {} func (*NodeSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{100} + return fileDescriptor_83c10c24ec417dc9, []int{101} } func (m *NodeSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2880,7 +2908,7 @@ var xxx_messageInfo_NodeSpec proto.InternalMessageInfo func (m *NodeStatus) Reset() { *m = NodeStatus{} } func (*NodeStatus) ProtoMessage() {} func (*NodeStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{101} + return fileDescriptor_83c10c24ec417dc9, []int{102} } func (m *NodeStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2908,7 +2936,7 @@ var xxx_messageInfo_NodeStatus proto.InternalMessageInfo func (m *NodeSystemInfo) Reset() { *m = NodeSystemInfo{} } func (*NodeSystemInfo) ProtoMessage() {} func (*NodeSystemInfo) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{102} + return fileDescriptor_83c10c24ec417dc9, []int{103} } func (m *NodeSystemInfo) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2936,7 +2964,7 @@ var xxx_messageInfo_NodeSystemInfo proto.InternalMessageInfo func (m *ObjectFieldSelector) Reset() { *m = ObjectFieldSelector{} } func (*ObjectFieldSelector) ProtoMessage() {} func (*ObjectFieldSelector) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{103} + return fileDescriptor_83c10c24ec417dc9, []int{104} } func (m *ObjectFieldSelector) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2964,7 +2992,7 @@ var xxx_messageInfo_ObjectFieldSelector proto.InternalMessageInfo func (m *ObjectReference) Reset() { *m = ObjectReference{} } func (*ObjectReference) ProtoMessage() {} func (*ObjectReference) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{104} + return fileDescriptor_83c10c24ec417dc9, []int{105} } func (m *ObjectReference) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -2992,7 +3020,7 @@ var xxx_messageInfo_ObjectReference proto.InternalMessageInfo func (m *PersistentVolume) Reset() { *m = PersistentVolume{} } func (*PersistentVolume) ProtoMessage() {} func (*PersistentVolume) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{105} + return fileDescriptor_83c10c24ec417dc9, []int{106} } func (m *PersistentVolume) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3020,7 +3048,7 @@ var xxx_messageInfo_PersistentVolume proto.InternalMessageInfo func (m *PersistentVolumeClaim) Reset() { *m = PersistentVolumeClaim{} } func (*PersistentVolumeClaim) ProtoMessage() {} func (*PersistentVolumeClaim) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{106} + return fileDescriptor_83c10c24ec417dc9, []int{107} } func (m *PersistentVolumeClaim) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3048,7 +3076,7 @@ var xxx_messageInfo_PersistentVolumeClaim proto.InternalMessageInfo func (m *PersistentVolumeClaimCondition) Reset() { *m = PersistentVolumeClaimCondition{} } func (*PersistentVolumeClaimCondition) ProtoMessage() {} func (*PersistentVolumeClaimCondition) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{107} + return fileDescriptor_83c10c24ec417dc9, []int{108} } func (m *PersistentVolumeClaimCondition) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3076,7 +3104,7 @@ var xxx_messageInfo_PersistentVolumeClaimCondition proto.InternalMessageInfo func (m *PersistentVolumeClaimList) Reset() { *m = PersistentVolumeClaimList{} } func (*PersistentVolumeClaimList) ProtoMessage() {} func (*PersistentVolumeClaimList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{108} + return fileDescriptor_83c10c24ec417dc9, []int{109} } func (m *PersistentVolumeClaimList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3104,7 +3132,7 @@ var xxx_messageInfo_PersistentVolumeClaimList proto.InternalMessageInfo func (m *PersistentVolumeClaimSpec) Reset() { *m = PersistentVolumeClaimSpec{} } func (*PersistentVolumeClaimSpec) ProtoMessage() {} func (*PersistentVolumeClaimSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{109} + return fileDescriptor_83c10c24ec417dc9, []int{110} } func (m *PersistentVolumeClaimSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3132,7 +3160,7 @@ var xxx_messageInfo_PersistentVolumeClaimSpec proto.InternalMessageInfo func (m *PersistentVolumeClaimStatus) Reset() { *m = PersistentVolumeClaimStatus{} } func (*PersistentVolumeClaimStatus) ProtoMessage() {} func (*PersistentVolumeClaimStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{110} + return fileDescriptor_83c10c24ec417dc9, []int{111} } func (m *PersistentVolumeClaimStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3157,10 +3185,38 @@ func (m *PersistentVolumeClaimStatus) XXX_DiscardUnknown() { var xxx_messageInfo_PersistentVolumeClaimStatus proto.InternalMessageInfo +func (m *PersistentVolumeClaimTemplate) Reset() { *m = PersistentVolumeClaimTemplate{} } +func (*PersistentVolumeClaimTemplate) ProtoMessage() {} +func (*PersistentVolumeClaimTemplate) Descriptor() ([]byte, []int) { + return fileDescriptor_83c10c24ec417dc9, []int{112} +} +func (m *PersistentVolumeClaimTemplate) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *PersistentVolumeClaimTemplate) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *PersistentVolumeClaimTemplate) XXX_Merge(src proto.Message) { + xxx_messageInfo_PersistentVolumeClaimTemplate.Merge(m, src) +} +func (m *PersistentVolumeClaimTemplate) XXX_Size() int { + return m.Size() +} +func (m *PersistentVolumeClaimTemplate) XXX_DiscardUnknown() { + xxx_messageInfo_PersistentVolumeClaimTemplate.DiscardUnknown(m) +} + +var xxx_messageInfo_PersistentVolumeClaimTemplate proto.InternalMessageInfo + func (m *PersistentVolumeClaimVolumeSource) Reset() { *m = PersistentVolumeClaimVolumeSource{} } func (*PersistentVolumeClaimVolumeSource) ProtoMessage() {} func (*PersistentVolumeClaimVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{111} + return fileDescriptor_83c10c24ec417dc9, []int{113} } func (m *PersistentVolumeClaimVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3188,7 +3244,7 @@ var xxx_messageInfo_PersistentVolumeClaimVolumeSource proto.InternalMessageInfo func (m *PersistentVolumeList) Reset() { *m = PersistentVolumeList{} } func (*PersistentVolumeList) ProtoMessage() {} func (*PersistentVolumeList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{112} + return fileDescriptor_83c10c24ec417dc9, []int{114} } func (m *PersistentVolumeList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3216,7 +3272,7 @@ var xxx_messageInfo_PersistentVolumeList proto.InternalMessageInfo func (m *PersistentVolumeSource) Reset() { *m = PersistentVolumeSource{} } func (*PersistentVolumeSource) ProtoMessage() {} func (*PersistentVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{113} + return fileDescriptor_83c10c24ec417dc9, []int{115} } func (m *PersistentVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3244,7 +3300,7 @@ var xxx_messageInfo_PersistentVolumeSource proto.InternalMessageInfo func (m *PersistentVolumeSpec) Reset() { *m = PersistentVolumeSpec{} } func (*PersistentVolumeSpec) ProtoMessage() {} func (*PersistentVolumeSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{114} + return fileDescriptor_83c10c24ec417dc9, []int{116} } func (m *PersistentVolumeSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3272,7 +3328,7 @@ var xxx_messageInfo_PersistentVolumeSpec proto.InternalMessageInfo func (m *PersistentVolumeStatus) Reset() { *m = PersistentVolumeStatus{} } func (*PersistentVolumeStatus) ProtoMessage() {} func (*PersistentVolumeStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{115} + return fileDescriptor_83c10c24ec417dc9, []int{117} } func (m *PersistentVolumeStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3300,7 +3356,7 @@ var xxx_messageInfo_PersistentVolumeStatus proto.InternalMessageInfo func (m *PhotonPersistentDiskVolumeSource) Reset() { *m = PhotonPersistentDiskVolumeSource{} } func (*PhotonPersistentDiskVolumeSource) ProtoMessage() {} func (*PhotonPersistentDiskVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{116} + return fileDescriptor_83c10c24ec417dc9, []int{118} } func (m *PhotonPersistentDiskVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3328,7 +3384,7 @@ var xxx_messageInfo_PhotonPersistentDiskVolumeSource proto.InternalMessageInfo func (m *Pod) Reset() { *m = Pod{} } func (*Pod) ProtoMessage() {} func (*Pod) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{117} + return fileDescriptor_83c10c24ec417dc9, []int{119} } func (m *Pod) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3356,7 +3412,7 @@ var xxx_messageInfo_Pod proto.InternalMessageInfo func (m *PodAffinity) Reset() { *m = PodAffinity{} } func (*PodAffinity) ProtoMessage() {} func (*PodAffinity) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{118} + return fileDescriptor_83c10c24ec417dc9, []int{120} } func (m *PodAffinity) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3384,7 +3440,7 @@ var xxx_messageInfo_PodAffinity proto.InternalMessageInfo func (m *PodAffinityTerm) Reset() { *m = PodAffinityTerm{} } func (*PodAffinityTerm) ProtoMessage() {} func (*PodAffinityTerm) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{119} + return fileDescriptor_83c10c24ec417dc9, []int{121} } func (m *PodAffinityTerm) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3412,7 +3468,7 @@ var xxx_messageInfo_PodAffinityTerm proto.InternalMessageInfo func (m *PodAntiAffinity) Reset() { *m = PodAntiAffinity{} } func (*PodAntiAffinity) ProtoMessage() {} func (*PodAntiAffinity) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{120} + return fileDescriptor_83c10c24ec417dc9, []int{122} } func (m *PodAntiAffinity) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3440,7 +3496,7 @@ var xxx_messageInfo_PodAntiAffinity proto.InternalMessageInfo func (m *PodAttachOptions) Reset() { *m = PodAttachOptions{} } func (*PodAttachOptions) ProtoMessage() {} func (*PodAttachOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{121} + return fileDescriptor_83c10c24ec417dc9, []int{123} } func (m *PodAttachOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3468,7 +3524,7 @@ var xxx_messageInfo_PodAttachOptions proto.InternalMessageInfo func (m *PodCondition) Reset() { *m = PodCondition{} } func (*PodCondition) ProtoMessage() {} func (*PodCondition) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{122} + return fileDescriptor_83c10c24ec417dc9, []int{124} } func (m *PodCondition) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3496,7 +3552,7 @@ var xxx_messageInfo_PodCondition proto.InternalMessageInfo func (m *PodDNSConfig) Reset() { *m = PodDNSConfig{} } func (*PodDNSConfig) ProtoMessage() {} func (*PodDNSConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{123} + return fileDescriptor_83c10c24ec417dc9, []int{125} } func (m *PodDNSConfig) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3524,7 +3580,7 @@ var xxx_messageInfo_PodDNSConfig proto.InternalMessageInfo func (m *PodDNSConfigOption) Reset() { *m = PodDNSConfigOption{} } func (*PodDNSConfigOption) ProtoMessage() {} func (*PodDNSConfigOption) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{124} + return fileDescriptor_83c10c24ec417dc9, []int{126} } func (m *PodDNSConfigOption) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3552,7 +3608,7 @@ var xxx_messageInfo_PodDNSConfigOption proto.InternalMessageInfo func (m *PodExecOptions) Reset() { *m = PodExecOptions{} } func (*PodExecOptions) ProtoMessage() {} func (*PodExecOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{125} + return fileDescriptor_83c10c24ec417dc9, []int{127} } func (m *PodExecOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3580,7 +3636,7 @@ var xxx_messageInfo_PodExecOptions proto.InternalMessageInfo func (m *PodIP) Reset() { *m = PodIP{} } func (*PodIP) ProtoMessage() {} func (*PodIP) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{126} + return fileDescriptor_83c10c24ec417dc9, []int{128} } func (m *PodIP) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3608,7 +3664,7 @@ var xxx_messageInfo_PodIP proto.InternalMessageInfo func (m *PodList) Reset() { *m = PodList{} } func (*PodList) ProtoMessage() {} func (*PodList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{127} + return fileDescriptor_83c10c24ec417dc9, []int{129} } func (m *PodList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3636,7 +3692,7 @@ var xxx_messageInfo_PodList proto.InternalMessageInfo func (m *PodLogOptions) Reset() { *m = PodLogOptions{} } func (*PodLogOptions) ProtoMessage() {} func (*PodLogOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{128} + return fileDescriptor_83c10c24ec417dc9, []int{130} } func (m *PodLogOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3664,7 +3720,7 @@ var xxx_messageInfo_PodLogOptions proto.InternalMessageInfo func (m *PodPortForwardOptions) Reset() { *m = PodPortForwardOptions{} } func (*PodPortForwardOptions) ProtoMessage() {} func (*PodPortForwardOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{129} + return fileDescriptor_83c10c24ec417dc9, []int{131} } func (m *PodPortForwardOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3692,7 +3748,7 @@ var xxx_messageInfo_PodPortForwardOptions proto.InternalMessageInfo func (m *PodProxyOptions) Reset() { *m = PodProxyOptions{} } func (*PodProxyOptions) ProtoMessage() {} func (*PodProxyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{130} + return fileDescriptor_83c10c24ec417dc9, []int{132} } func (m *PodProxyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3720,7 +3776,7 @@ var xxx_messageInfo_PodProxyOptions proto.InternalMessageInfo func (m *PodReadinessGate) Reset() { *m = PodReadinessGate{} } func (*PodReadinessGate) ProtoMessage() {} func (*PodReadinessGate) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{131} + return fileDescriptor_83c10c24ec417dc9, []int{133} } func (m *PodReadinessGate) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3748,7 +3804,7 @@ var xxx_messageInfo_PodReadinessGate proto.InternalMessageInfo func (m *PodSecurityContext) Reset() { *m = PodSecurityContext{} } func (*PodSecurityContext) ProtoMessage() {} func (*PodSecurityContext) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{132} + return fileDescriptor_83c10c24ec417dc9, []int{134} } func (m *PodSecurityContext) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3776,7 +3832,7 @@ var xxx_messageInfo_PodSecurityContext proto.InternalMessageInfo func (m *PodSignature) Reset() { *m = PodSignature{} } func (*PodSignature) ProtoMessage() {} func (*PodSignature) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{133} + return fileDescriptor_83c10c24ec417dc9, []int{135} } func (m *PodSignature) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3804,7 +3860,7 @@ var xxx_messageInfo_PodSignature proto.InternalMessageInfo func (m *PodSpec) Reset() { *m = PodSpec{} } func (*PodSpec) ProtoMessage() {} func (*PodSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{134} + return fileDescriptor_83c10c24ec417dc9, []int{136} } func (m *PodSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3832,7 +3888,7 @@ var xxx_messageInfo_PodSpec proto.InternalMessageInfo func (m *PodStatus) Reset() { *m = PodStatus{} } func (*PodStatus) ProtoMessage() {} func (*PodStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{135} + return fileDescriptor_83c10c24ec417dc9, []int{137} } func (m *PodStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3860,7 +3916,7 @@ var xxx_messageInfo_PodStatus proto.InternalMessageInfo func (m *PodStatusResult) Reset() { *m = PodStatusResult{} } func (*PodStatusResult) ProtoMessage() {} func (*PodStatusResult) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{136} + return fileDescriptor_83c10c24ec417dc9, []int{138} } func (m *PodStatusResult) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3888,7 +3944,7 @@ var xxx_messageInfo_PodStatusResult proto.InternalMessageInfo func (m *PodTemplate) Reset() { *m = PodTemplate{} } func (*PodTemplate) ProtoMessage() {} func (*PodTemplate) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{137} + return fileDescriptor_83c10c24ec417dc9, []int{139} } func (m *PodTemplate) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3916,7 +3972,7 @@ var xxx_messageInfo_PodTemplate proto.InternalMessageInfo func (m *PodTemplateList) Reset() { *m = PodTemplateList{} } func (*PodTemplateList) ProtoMessage() {} func (*PodTemplateList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{138} + return fileDescriptor_83c10c24ec417dc9, []int{140} } func (m *PodTemplateList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3944,7 +4000,7 @@ var xxx_messageInfo_PodTemplateList proto.InternalMessageInfo func (m *PodTemplateSpec) Reset() { *m = PodTemplateSpec{} } func (*PodTemplateSpec) ProtoMessage() {} func (*PodTemplateSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{139} + return fileDescriptor_83c10c24ec417dc9, []int{141} } func (m *PodTemplateSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -3972,7 +4028,7 @@ var xxx_messageInfo_PodTemplateSpec proto.InternalMessageInfo func (m *PortworxVolumeSource) Reset() { *m = PortworxVolumeSource{} } func (*PortworxVolumeSource) ProtoMessage() {} func (*PortworxVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{140} + return fileDescriptor_83c10c24ec417dc9, []int{142} } func (m *PortworxVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4000,7 +4056,7 @@ var xxx_messageInfo_PortworxVolumeSource proto.InternalMessageInfo func (m *Preconditions) Reset() { *m = Preconditions{} } func (*Preconditions) ProtoMessage() {} func (*Preconditions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{141} + return fileDescriptor_83c10c24ec417dc9, []int{143} } func (m *Preconditions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4028,7 +4084,7 @@ var xxx_messageInfo_Preconditions proto.InternalMessageInfo func (m *PreferAvoidPodsEntry) Reset() { *m = PreferAvoidPodsEntry{} } func (*PreferAvoidPodsEntry) ProtoMessage() {} func (*PreferAvoidPodsEntry) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{142} + return fileDescriptor_83c10c24ec417dc9, []int{144} } func (m *PreferAvoidPodsEntry) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4056,7 +4112,7 @@ var xxx_messageInfo_PreferAvoidPodsEntry proto.InternalMessageInfo func (m *PreferredSchedulingTerm) Reset() { *m = PreferredSchedulingTerm{} } func (*PreferredSchedulingTerm) ProtoMessage() {} func (*PreferredSchedulingTerm) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{143} + return fileDescriptor_83c10c24ec417dc9, []int{145} } func (m *PreferredSchedulingTerm) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4084,7 +4140,7 @@ var xxx_messageInfo_PreferredSchedulingTerm proto.InternalMessageInfo func (m *Probe) Reset() { *m = Probe{} } func (*Probe) ProtoMessage() {} func (*Probe) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{144} + return fileDescriptor_83c10c24ec417dc9, []int{146} } func (m *Probe) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4112,7 +4168,7 @@ var xxx_messageInfo_Probe proto.InternalMessageInfo func (m *ProjectedVolumeSource) Reset() { *m = ProjectedVolumeSource{} } func (*ProjectedVolumeSource) ProtoMessage() {} func (*ProjectedVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{145} + return fileDescriptor_83c10c24ec417dc9, []int{147} } func (m *ProjectedVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4140,7 +4196,7 @@ var xxx_messageInfo_ProjectedVolumeSource proto.InternalMessageInfo func (m *QuobyteVolumeSource) Reset() { *m = QuobyteVolumeSource{} } func (*QuobyteVolumeSource) ProtoMessage() {} func (*QuobyteVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{146} + return fileDescriptor_83c10c24ec417dc9, []int{148} } func (m *QuobyteVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4168,7 +4224,7 @@ var xxx_messageInfo_QuobyteVolumeSource proto.InternalMessageInfo func (m *RBDPersistentVolumeSource) Reset() { *m = RBDPersistentVolumeSource{} } func (*RBDPersistentVolumeSource) ProtoMessage() {} func (*RBDPersistentVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{147} + return fileDescriptor_83c10c24ec417dc9, []int{149} } func (m *RBDPersistentVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4196,7 +4252,7 @@ var xxx_messageInfo_RBDPersistentVolumeSource proto.InternalMessageInfo func (m *RBDVolumeSource) Reset() { *m = RBDVolumeSource{} } func (*RBDVolumeSource) ProtoMessage() {} func (*RBDVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{148} + return fileDescriptor_83c10c24ec417dc9, []int{150} } func (m *RBDVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4224,7 +4280,7 @@ var xxx_messageInfo_RBDVolumeSource proto.InternalMessageInfo func (m *RangeAllocation) Reset() { *m = RangeAllocation{} } func (*RangeAllocation) ProtoMessage() {} func (*RangeAllocation) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{149} + return fileDescriptor_83c10c24ec417dc9, []int{151} } func (m *RangeAllocation) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4252,7 +4308,7 @@ var xxx_messageInfo_RangeAllocation proto.InternalMessageInfo func (m *ReplicationController) Reset() { *m = ReplicationController{} } func (*ReplicationController) ProtoMessage() {} func (*ReplicationController) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{150} + return fileDescriptor_83c10c24ec417dc9, []int{152} } func (m *ReplicationController) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4280,7 +4336,7 @@ var xxx_messageInfo_ReplicationController proto.InternalMessageInfo func (m *ReplicationControllerCondition) Reset() { *m = ReplicationControllerCondition{} } func (*ReplicationControllerCondition) ProtoMessage() {} func (*ReplicationControllerCondition) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{151} + return fileDescriptor_83c10c24ec417dc9, []int{153} } func (m *ReplicationControllerCondition) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4308,7 +4364,7 @@ var xxx_messageInfo_ReplicationControllerCondition proto.InternalMessageInfo func (m *ReplicationControllerList) Reset() { *m = ReplicationControllerList{} } func (*ReplicationControllerList) ProtoMessage() {} func (*ReplicationControllerList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{152} + return fileDescriptor_83c10c24ec417dc9, []int{154} } func (m *ReplicationControllerList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4336,7 +4392,7 @@ var xxx_messageInfo_ReplicationControllerList proto.InternalMessageInfo func (m *ReplicationControllerSpec) Reset() { *m = ReplicationControllerSpec{} } func (*ReplicationControllerSpec) ProtoMessage() {} func (*ReplicationControllerSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{153} + return fileDescriptor_83c10c24ec417dc9, []int{155} } func (m *ReplicationControllerSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4364,7 +4420,7 @@ var xxx_messageInfo_ReplicationControllerSpec proto.InternalMessageInfo func (m *ReplicationControllerStatus) Reset() { *m = ReplicationControllerStatus{} } func (*ReplicationControllerStatus) ProtoMessage() {} func (*ReplicationControllerStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{154} + return fileDescriptor_83c10c24ec417dc9, []int{156} } func (m *ReplicationControllerStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4392,7 +4448,7 @@ var xxx_messageInfo_ReplicationControllerStatus proto.InternalMessageInfo func (m *ResourceFieldSelector) Reset() { *m = ResourceFieldSelector{} } func (*ResourceFieldSelector) ProtoMessage() {} func (*ResourceFieldSelector) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{155} + return fileDescriptor_83c10c24ec417dc9, []int{157} } func (m *ResourceFieldSelector) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4420,7 +4476,7 @@ var xxx_messageInfo_ResourceFieldSelector proto.InternalMessageInfo func (m *ResourceQuota) Reset() { *m = ResourceQuota{} } func (*ResourceQuota) ProtoMessage() {} func (*ResourceQuota) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{156} + return fileDescriptor_83c10c24ec417dc9, []int{158} } func (m *ResourceQuota) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4448,7 +4504,7 @@ var xxx_messageInfo_ResourceQuota proto.InternalMessageInfo func (m *ResourceQuotaList) Reset() { *m = ResourceQuotaList{} } func (*ResourceQuotaList) ProtoMessage() {} func (*ResourceQuotaList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{157} + return fileDescriptor_83c10c24ec417dc9, []int{159} } func (m *ResourceQuotaList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4476,7 +4532,7 @@ var xxx_messageInfo_ResourceQuotaList proto.InternalMessageInfo func (m *ResourceQuotaSpec) Reset() { *m = ResourceQuotaSpec{} } func (*ResourceQuotaSpec) ProtoMessage() {} func (*ResourceQuotaSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{158} + return fileDescriptor_83c10c24ec417dc9, []int{160} } func (m *ResourceQuotaSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4504,7 +4560,7 @@ var xxx_messageInfo_ResourceQuotaSpec proto.InternalMessageInfo func (m *ResourceQuotaStatus) Reset() { *m = ResourceQuotaStatus{} } func (*ResourceQuotaStatus) ProtoMessage() {} func (*ResourceQuotaStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{159} + return fileDescriptor_83c10c24ec417dc9, []int{161} } func (m *ResourceQuotaStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4532,7 +4588,7 @@ var xxx_messageInfo_ResourceQuotaStatus proto.InternalMessageInfo func (m *ResourceRequirements) Reset() { *m = ResourceRequirements{} } func (*ResourceRequirements) ProtoMessage() {} func (*ResourceRequirements) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{160} + return fileDescriptor_83c10c24ec417dc9, []int{162} } func (m *ResourceRequirements) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4560,7 +4616,7 @@ var xxx_messageInfo_ResourceRequirements proto.InternalMessageInfo func (m *SELinuxOptions) Reset() { *m = SELinuxOptions{} } func (*SELinuxOptions) ProtoMessage() {} func (*SELinuxOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{161} + return fileDescriptor_83c10c24ec417dc9, []int{163} } func (m *SELinuxOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4588,7 +4644,7 @@ var xxx_messageInfo_SELinuxOptions proto.InternalMessageInfo func (m *ScaleIOPersistentVolumeSource) Reset() { *m = ScaleIOPersistentVolumeSource{} } func (*ScaleIOPersistentVolumeSource) ProtoMessage() {} func (*ScaleIOPersistentVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{162} + return fileDescriptor_83c10c24ec417dc9, []int{164} } func (m *ScaleIOPersistentVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4616,7 +4672,7 @@ var xxx_messageInfo_ScaleIOPersistentVolumeSource proto.InternalMessageInfo func (m *ScaleIOVolumeSource) Reset() { *m = ScaleIOVolumeSource{} } func (*ScaleIOVolumeSource) ProtoMessage() {} func (*ScaleIOVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{163} + return fileDescriptor_83c10c24ec417dc9, []int{165} } func (m *ScaleIOVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4644,7 +4700,7 @@ var xxx_messageInfo_ScaleIOVolumeSource proto.InternalMessageInfo func (m *ScopeSelector) Reset() { *m = ScopeSelector{} } func (*ScopeSelector) ProtoMessage() {} func (*ScopeSelector) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{164} + return fileDescriptor_83c10c24ec417dc9, []int{166} } func (m *ScopeSelector) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4672,7 +4728,7 @@ var xxx_messageInfo_ScopeSelector proto.InternalMessageInfo func (m *ScopedResourceSelectorRequirement) Reset() { *m = ScopedResourceSelectorRequirement{} } func (*ScopedResourceSelectorRequirement) ProtoMessage() {} func (*ScopedResourceSelectorRequirement) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{165} + return fileDescriptor_83c10c24ec417dc9, []int{167} } func (m *ScopedResourceSelectorRequirement) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4697,10 +4753,38 @@ func (m *ScopedResourceSelectorRequirement) XXX_DiscardUnknown() { var xxx_messageInfo_ScopedResourceSelectorRequirement proto.InternalMessageInfo +func (m *SeccompProfile) Reset() { *m = SeccompProfile{} } +func (*SeccompProfile) ProtoMessage() {} +func (*SeccompProfile) Descriptor() ([]byte, []int) { + return fileDescriptor_83c10c24ec417dc9, []int{168} +} +func (m *SeccompProfile) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *SeccompProfile) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *SeccompProfile) XXX_Merge(src proto.Message) { + xxx_messageInfo_SeccompProfile.Merge(m, src) +} +func (m *SeccompProfile) XXX_Size() int { + return m.Size() +} +func (m *SeccompProfile) XXX_DiscardUnknown() { + xxx_messageInfo_SeccompProfile.DiscardUnknown(m) +} + +var xxx_messageInfo_SeccompProfile proto.InternalMessageInfo + func (m *Secret) Reset() { *m = Secret{} } func (*Secret) ProtoMessage() {} func (*Secret) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{166} + return fileDescriptor_83c10c24ec417dc9, []int{169} } func (m *Secret) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4728,7 +4812,7 @@ var xxx_messageInfo_Secret proto.InternalMessageInfo func (m *SecretEnvSource) Reset() { *m = SecretEnvSource{} } func (*SecretEnvSource) ProtoMessage() {} func (*SecretEnvSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{167} + return fileDescriptor_83c10c24ec417dc9, []int{170} } func (m *SecretEnvSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4756,7 +4840,7 @@ var xxx_messageInfo_SecretEnvSource proto.InternalMessageInfo func (m *SecretKeySelector) Reset() { *m = SecretKeySelector{} } func (*SecretKeySelector) ProtoMessage() {} func (*SecretKeySelector) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{168} + return fileDescriptor_83c10c24ec417dc9, []int{171} } func (m *SecretKeySelector) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4784,7 +4868,7 @@ var xxx_messageInfo_SecretKeySelector proto.InternalMessageInfo func (m *SecretList) Reset() { *m = SecretList{} } func (*SecretList) ProtoMessage() {} func (*SecretList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{169} + return fileDescriptor_83c10c24ec417dc9, []int{172} } func (m *SecretList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4812,7 +4896,7 @@ var xxx_messageInfo_SecretList proto.InternalMessageInfo func (m *SecretProjection) Reset() { *m = SecretProjection{} } func (*SecretProjection) ProtoMessage() {} func (*SecretProjection) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{170} + return fileDescriptor_83c10c24ec417dc9, []int{173} } func (m *SecretProjection) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4840,7 +4924,7 @@ var xxx_messageInfo_SecretProjection proto.InternalMessageInfo func (m *SecretReference) Reset() { *m = SecretReference{} } func (*SecretReference) ProtoMessage() {} func (*SecretReference) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{171} + return fileDescriptor_83c10c24ec417dc9, []int{174} } func (m *SecretReference) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4868,7 +4952,7 @@ var xxx_messageInfo_SecretReference proto.InternalMessageInfo func (m *SecretVolumeSource) Reset() { *m = SecretVolumeSource{} } func (*SecretVolumeSource) ProtoMessage() {} func (*SecretVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{172} + return fileDescriptor_83c10c24ec417dc9, []int{175} } func (m *SecretVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4896,7 +4980,7 @@ var xxx_messageInfo_SecretVolumeSource proto.InternalMessageInfo func (m *SecurityContext) Reset() { *m = SecurityContext{} } func (*SecurityContext) ProtoMessage() {} func (*SecurityContext) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{173} + return fileDescriptor_83c10c24ec417dc9, []int{176} } func (m *SecurityContext) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4924,7 +5008,7 @@ var xxx_messageInfo_SecurityContext proto.InternalMessageInfo func (m *SerializedReference) Reset() { *m = SerializedReference{} } func (*SerializedReference) ProtoMessage() {} func (*SerializedReference) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{174} + return fileDescriptor_83c10c24ec417dc9, []int{177} } func (m *SerializedReference) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4952,7 +5036,7 @@ var xxx_messageInfo_SerializedReference proto.InternalMessageInfo func (m *Service) Reset() { *m = Service{} } func (*Service) ProtoMessage() {} func (*Service) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{175} + return fileDescriptor_83c10c24ec417dc9, []int{178} } func (m *Service) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -4980,7 +5064,7 @@ var xxx_messageInfo_Service proto.InternalMessageInfo func (m *ServiceAccount) Reset() { *m = ServiceAccount{} } func (*ServiceAccount) ProtoMessage() {} func (*ServiceAccount) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{176} + return fileDescriptor_83c10c24ec417dc9, []int{179} } func (m *ServiceAccount) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5008,7 +5092,7 @@ var xxx_messageInfo_ServiceAccount proto.InternalMessageInfo func (m *ServiceAccountList) Reset() { *m = ServiceAccountList{} } func (*ServiceAccountList) ProtoMessage() {} func (*ServiceAccountList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{177} + return fileDescriptor_83c10c24ec417dc9, []int{180} } func (m *ServiceAccountList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5036,7 +5120,7 @@ var xxx_messageInfo_ServiceAccountList proto.InternalMessageInfo func (m *ServiceAccountTokenProjection) Reset() { *m = ServiceAccountTokenProjection{} } func (*ServiceAccountTokenProjection) ProtoMessage() {} func (*ServiceAccountTokenProjection) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{178} + return fileDescriptor_83c10c24ec417dc9, []int{181} } func (m *ServiceAccountTokenProjection) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5064,7 +5148,7 @@ var xxx_messageInfo_ServiceAccountTokenProjection proto.InternalMessageInfo func (m *ServiceList) Reset() { *m = ServiceList{} } func (*ServiceList) ProtoMessage() {} func (*ServiceList) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{179} + return fileDescriptor_83c10c24ec417dc9, []int{182} } func (m *ServiceList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5092,7 +5176,7 @@ var xxx_messageInfo_ServiceList proto.InternalMessageInfo func (m *ServicePort) Reset() { *m = ServicePort{} } func (*ServicePort) ProtoMessage() {} func (*ServicePort) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{180} + return fileDescriptor_83c10c24ec417dc9, []int{183} } func (m *ServicePort) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5120,7 +5204,7 @@ var xxx_messageInfo_ServicePort proto.InternalMessageInfo func (m *ServiceProxyOptions) Reset() { *m = ServiceProxyOptions{} } func (*ServiceProxyOptions) ProtoMessage() {} func (*ServiceProxyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{181} + return fileDescriptor_83c10c24ec417dc9, []int{184} } func (m *ServiceProxyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5148,7 +5232,7 @@ var xxx_messageInfo_ServiceProxyOptions proto.InternalMessageInfo func (m *ServiceSpec) Reset() { *m = ServiceSpec{} } func (*ServiceSpec) ProtoMessage() {} func (*ServiceSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{182} + return fileDescriptor_83c10c24ec417dc9, []int{185} } func (m *ServiceSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5176,7 +5260,7 @@ var xxx_messageInfo_ServiceSpec proto.InternalMessageInfo func (m *ServiceStatus) Reset() { *m = ServiceStatus{} } func (*ServiceStatus) ProtoMessage() {} func (*ServiceStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{183} + return fileDescriptor_83c10c24ec417dc9, []int{186} } func (m *ServiceStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5204,7 +5288,7 @@ var xxx_messageInfo_ServiceStatus proto.InternalMessageInfo func (m *SessionAffinityConfig) Reset() { *m = SessionAffinityConfig{} } func (*SessionAffinityConfig) ProtoMessage() {} func (*SessionAffinityConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{184} + return fileDescriptor_83c10c24ec417dc9, []int{187} } func (m *SessionAffinityConfig) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5232,7 +5316,7 @@ var xxx_messageInfo_SessionAffinityConfig proto.InternalMessageInfo func (m *StorageOSPersistentVolumeSource) Reset() { *m = StorageOSPersistentVolumeSource{} } func (*StorageOSPersistentVolumeSource) ProtoMessage() {} func (*StorageOSPersistentVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{185} + return fileDescriptor_83c10c24ec417dc9, []int{188} } func (m *StorageOSPersistentVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5260,7 +5344,7 @@ var xxx_messageInfo_StorageOSPersistentVolumeSource proto.InternalMessageInfo func (m *StorageOSVolumeSource) Reset() { *m = StorageOSVolumeSource{} } func (*StorageOSVolumeSource) ProtoMessage() {} func (*StorageOSVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{186} + return fileDescriptor_83c10c24ec417dc9, []int{189} } func (m *StorageOSVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5288,7 +5372,7 @@ var xxx_messageInfo_StorageOSVolumeSource proto.InternalMessageInfo func (m *Sysctl) Reset() { *m = Sysctl{} } func (*Sysctl) ProtoMessage() {} func (*Sysctl) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{187} + return fileDescriptor_83c10c24ec417dc9, []int{190} } func (m *Sysctl) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5316,7 +5400,7 @@ var xxx_messageInfo_Sysctl proto.InternalMessageInfo func (m *TCPSocketAction) Reset() { *m = TCPSocketAction{} } func (*TCPSocketAction) ProtoMessage() {} func (*TCPSocketAction) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{188} + return fileDescriptor_83c10c24ec417dc9, []int{191} } func (m *TCPSocketAction) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5344,7 +5428,7 @@ var xxx_messageInfo_TCPSocketAction proto.InternalMessageInfo func (m *Taint) Reset() { *m = Taint{} } func (*Taint) ProtoMessage() {} func (*Taint) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{189} + return fileDescriptor_83c10c24ec417dc9, []int{192} } func (m *Taint) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5372,7 +5456,7 @@ var xxx_messageInfo_Taint proto.InternalMessageInfo func (m *Toleration) Reset() { *m = Toleration{} } func (*Toleration) ProtoMessage() {} func (*Toleration) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{190} + return fileDescriptor_83c10c24ec417dc9, []int{193} } func (m *Toleration) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5400,7 +5484,7 @@ var xxx_messageInfo_Toleration proto.InternalMessageInfo func (m *TopologySelectorLabelRequirement) Reset() { *m = TopologySelectorLabelRequirement{} } func (*TopologySelectorLabelRequirement) ProtoMessage() {} func (*TopologySelectorLabelRequirement) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{191} + return fileDescriptor_83c10c24ec417dc9, []int{194} } func (m *TopologySelectorLabelRequirement) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5428,7 +5512,7 @@ var xxx_messageInfo_TopologySelectorLabelRequirement proto.InternalMessageInfo func (m *TopologySelectorTerm) Reset() { *m = TopologySelectorTerm{} } func (*TopologySelectorTerm) ProtoMessage() {} func (*TopologySelectorTerm) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{192} + return fileDescriptor_83c10c24ec417dc9, []int{195} } func (m *TopologySelectorTerm) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5456,7 +5540,7 @@ var xxx_messageInfo_TopologySelectorTerm proto.InternalMessageInfo func (m *TopologySpreadConstraint) Reset() { *m = TopologySpreadConstraint{} } func (*TopologySpreadConstraint) ProtoMessage() {} func (*TopologySpreadConstraint) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{193} + return fileDescriptor_83c10c24ec417dc9, []int{196} } func (m *TopologySpreadConstraint) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5484,7 +5568,7 @@ var xxx_messageInfo_TopologySpreadConstraint proto.InternalMessageInfo func (m *TypedLocalObjectReference) Reset() { *m = TypedLocalObjectReference{} } func (*TypedLocalObjectReference) ProtoMessage() {} func (*TypedLocalObjectReference) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{194} + return fileDescriptor_83c10c24ec417dc9, []int{197} } func (m *TypedLocalObjectReference) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5512,7 +5596,7 @@ var xxx_messageInfo_TypedLocalObjectReference proto.InternalMessageInfo func (m *Volume) Reset() { *m = Volume{} } func (*Volume) ProtoMessage() {} func (*Volume) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{195} + return fileDescriptor_83c10c24ec417dc9, []int{198} } func (m *Volume) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5540,7 +5624,7 @@ var xxx_messageInfo_Volume proto.InternalMessageInfo func (m *VolumeDevice) Reset() { *m = VolumeDevice{} } func (*VolumeDevice) ProtoMessage() {} func (*VolumeDevice) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{196} + return fileDescriptor_83c10c24ec417dc9, []int{199} } func (m *VolumeDevice) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5568,7 +5652,7 @@ var xxx_messageInfo_VolumeDevice proto.InternalMessageInfo func (m *VolumeMount) Reset() { *m = VolumeMount{} } func (*VolumeMount) ProtoMessage() {} func (*VolumeMount) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{197} + return fileDescriptor_83c10c24ec417dc9, []int{200} } func (m *VolumeMount) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5596,7 +5680,7 @@ var xxx_messageInfo_VolumeMount proto.InternalMessageInfo func (m *VolumeNodeAffinity) Reset() { *m = VolumeNodeAffinity{} } func (*VolumeNodeAffinity) ProtoMessage() {} func (*VolumeNodeAffinity) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{198} + return fileDescriptor_83c10c24ec417dc9, []int{201} } func (m *VolumeNodeAffinity) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5624,7 +5708,7 @@ var xxx_messageInfo_VolumeNodeAffinity proto.InternalMessageInfo func (m *VolumeProjection) Reset() { *m = VolumeProjection{} } func (*VolumeProjection) ProtoMessage() {} func (*VolumeProjection) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{199} + return fileDescriptor_83c10c24ec417dc9, []int{202} } func (m *VolumeProjection) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5652,7 +5736,7 @@ var xxx_messageInfo_VolumeProjection proto.InternalMessageInfo func (m *VolumeSource) Reset() { *m = VolumeSource{} } func (*VolumeSource) ProtoMessage() {} func (*VolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{200} + return fileDescriptor_83c10c24ec417dc9, []int{203} } func (m *VolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5680,7 +5764,7 @@ var xxx_messageInfo_VolumeSource proto.InternalMessageInfo func (m *VsphereVirtualDiskVolumeSource) Reset() { *m = VsphereVirtualDiskVolumeSource{} } func (*VsphereVirtualDiskVolumeSource) ProtoMessage() {} func (*VsphereVirtualDiskVolumeSource) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{201} + return fileDescriptor_83c10c24ec417dc9, []int{204} } func (m *VsphereVirtualDiskVolumeSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5708,7 +5792,7 @@ var xxx_messageInfo_VsphereVirtualDiskVolumeSource proto.InternalMessageInfo func (m *WeightedPodAffinityTerm) Reset() { *m = WeightedPodAffinityTerm{} } func (*WeightedPodAffinityTerm) ProtoMessage() {} func (*WeightedPodAffinityTerm) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{202} + return fileDescriptor_83c10c24ec417dc9, []int{205} } func (m *WeightedPodAffinityTerm) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5736,7 +5820,7 @@ var xxx_messageInfo_WeightedPodAffinityTerm proto.InternalMessageInfo func (m *WindowsSecurityContextOptions) Reset() { *m = WindowsSecurityContextOptions{} } func (*WindowsSecurityContextOptions) ProtoMessage() {} func (*WindowsSecurityContextOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_83c10c24ec417dc9, []int{203} + return fileDescriptor_83c10c24ec417dc9, []int{206} } func (m *WindowsSecurityContextOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -5816,6 +5900,7 @@ func init() { proto.RegisterType((*EphemeralContainer)(nil), "k8s.io.api.core.v1.EphemeralContainer") proto.RegisterType((*EphemeralContainerCommon)(nil), "k8s.io.api.core.v1.EphemeralContainerCommon") proto.RegisterType((*EphemeralContainers)(nil), "k8s.io.api.core.v1.EphemeralContainers") + proto.RegisterType((*EphemeralVolumeSource)(nil), "k8s.io.api.core.v1.EphemeralVolumeSource") proto.RegisterType((*Event)(nil), "k8s.io.api.core.v1.Event") proto.RegisterType((*EventList)(nil), "k8s.io.api.core.v1.EventList") proto.RegisterType((*EventSeries)(nil), "k8s.io.api.core.v1.EventSeries") @@ -5888,6 +5973,7 @@ func init() { proto.RegisterType((*PersistentVolumeClaimSpec)(nil), "k8s.io.api.core.v1.PersistentVolumeClaimSpec") proto.RegisterType((*PersistentVolumeClaimStatus)(nil), "k8s.io.api.core.v1.PersistentVolumeClaimStatus") proto.RegisterMapType((ResourceList)(nil), "k8s.io.api.core.v1.PersistentVolumeClaimStatus.CapacityEntry") + proto.RegisterType((*PersistentVolumeClaimTemplate)(nil), "k8s.io.api.core.v1.PersistentVolumeClaimTemplate") proto.RegisterType((*PersistentVolumeClaimVolumeSource)(nil), "k8s.io.api.core.v1.PersistentVolumeClaimVolumeSource") proto.RegisterType((*PersistentVolumeList)(nil), "k8s.io.api.core.v1.PersistentVolumeList") proto.RegisterType((*PersistentVolumeSource)(nil), "k8s.io.api.core.v1.PersistentVolumeSource") @@ -5952,6 +6038,7 @@ func init() { proto.RegisterType((*ScaleIOVolumeSource)(nil), "k8s.io.api.core.v1.ScaleIOVolumeSource") proto.RegisterType((*ScopeSelector)(nil), "k8s.io.api.core.v1.ScopeSelector") proto.RegisterType((*ScopedResourceSelectorRequirement)(nil), "k8s.io.api.core.v1.ScopedResourceSelectorRequirement") + proto.RegisterType((*SeccompProfile)(nil), "k8s.io.api.core.v1.SeccompProfile") proto.RegisterType((*Secret)(nil), "k8s.io.api.core.v1.Secret") proto.RegisterMapType((map[string][]byte)(nil), "k8s.io.api.core.v1.Secret.DataEntry") proto.RegisterMapType((map[string]string)(nil), "k8s.io.api.core.v1.Secret.StringDataEntry") @@ -6000,859 +6087,876 @@ func init() { } var fileDescriptor_83c10c24ec417dc9 = []byte{ - // 13620 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xec, 0xbd, 0x6b, 0x70, 0x24, 0x59, - 0x5a, 0x18, 0xba, 0x59, 0xa5, 0x47, 0xd5, 0xa7, 0xf7, 0xe9, 0xc7, 0xa8, 0x35, 0xdd, 0xad, 0x9e, - 0x9c, 0xdd, 0x9e, 0x9e, 0x9d, 0x19, 0xf5, 0xce, 0x6b, 0x67, 0x98, 0x99, 0x1d, 0x90, 0x54, 0x52, - 0x77, 0x4d, 0xb7, 0xd4, 0x35, 0xa7, 0xd4, 0xdd, 0xbb, 0xc3, 0xec, 0xde, 0x4d, 0x55, 0x1e, 0x49, - 0x39, 0x2a, 0x65, 0xd6, 0x64, 0x66, 0x49, 0xad, 0xb9, 0x10, 0x97, 0xbb, 0x3c, 0xf7, 0x02, 0x37, - 0x36, 0x6c, 0xc2, 0x0f, 0x20, 0xb0, 0x03, 0xe3, 0x00, 0x0c, 0x76, 0x18, 0x83, 0x01, 0xef, 0x62, - 0x1b, 0x83, 0xed, 0xc0, 0xfe, 0x81, 0xb1, 0xc3, 0xf6, 0x12, 0x41, 0x58, 0x86, 0xc6, 0x61, 0x62, - 0x7f, 0x18, 0x08, 0x83, 0x7f, 0x58, 0x26, 0x8c, 0xe3, 0x3c, 0xf3, 0x9c, 0xac, 0xcc, 0xaa, 0x52, - 0x8f, 0x5a, 0x3b, 0x6c, 0xcc, 0xbf, 0xaa, 0xf3, 0x7d, 0xe7, 0x3b, 0x27, 0xcf, 0xf3, 0x3b, 0xdf, - 0x13, 0x5e, 0xdd, 0x7e, 0x39, 0x9a, 0xf3, 0x82, 0xab, 0xdb, 0xed, 0x75, 0x12, 0xfa, 0x24, 0x26, - 0xd1, 0xd5, 0x5d, 0xe2, 0xbb, 0x41, 0x78, 0x55, 0x00, 0x9c, 0x96, 0x77, 0xb5, 0x11, 0x84, 0xe4, - 0xea, 0xee, 0xb3, 0x57, 0x37, 0x89, 0x4f, 0x42, 0x27, 0x26, 0xee, 0x5c, 0x2b, 0x0c, 0xe2, 0x00, - 0x21, 0x8e, 0x33, 0xe7, 0xb4, 0xbc, 0x39, 0x8a, 0x33, 0xb7, 0xfb, 0xec, 0xcc, 0x33, 0x9b, 0x5e, - 0xbc, 0xd5, 0x5e, 0x9f, 0x6b, 0x04, 0x3b, 0x57, 0x37, 0x83, 0xcd, 0xe0, 0x2a, 0x43, 0x5d, 0x6f, - 0x6f, 0xb0, 0x7f, 0xec, 0x0f, 0xfb, 0xc5, 0x49, 0xcc, 0xbc, 0x90, 0x34, 0xb3, 0xe3, 0x34, 0xb6, - 0x3c, 0x9f, 0x84, 0xfb, 0x57, 0x5b, 0xdb, 0x9b, 0xac, 0xdd, 0x90, 0x44, 0x41, 0x3b, 0x6c, 0x90, - 0x74, 0xc3, 0x5d, 0x6b, 0x45, 0x57, 0x77, 0x48, 0xec, 0x64, 0x74, 0x77, 0xe6, 0x6a, 0x5e, 0xad, - 0xb0, 0xed, 0xc7, 0xde, 0x4e, 0x67, 0x33, 0x9f, 0xec, 0x55, 0x21, 0x6a, 0x6c, 0x91, 0x1d, 0xa7, - 0xa3, 0xde, 0xf3, 0x79, 0xf5, 0xda, 0xb1, 0xd7, 0xbc, 0xea, 0xf9, 0x71, 0x14, 0x87, 0xe9, 0x4a, - 0xf6, 0x57, 0x2d, 0xb8, 0x34, 0x7f, 0xb7, 0xbe, 0xd4, 0x74, 0xa2, 0xd8, 0x6b, 0x2c, 0x34, 0x83, - 0xc6, 0x76, 0x3d, 0x0e, 0x42, 0x72, 0x27, 0x68, 0xb6, 0x77, 0x48, 0x9d, 0x0d, 0x04, 0x7a, 0x1a, - 0x4a, 0xbb, 0xec, 0x7f, 0xb5, 0x32, 0x6d, 0x5d, 0xb2, 0xae, 0x94, 0x17, 0x26, 0x7f, 0xe3, 0x60, - 0xf6, 0x23, 0xf7, 0x0f, 0x66, 0x4b, 0x77, 0x44, 0x39, 0x56, 0x18, 0xe8, 0x32, 0x0c, 0x6d, 0x44, - 0x6b, 0xfb, 0x2d, 0x32, 0x5d, 0x60, 0xb8, 0xe3, 0x02, 0x77, 0x68, 0xb9, 0x4e, 0x4b, 0xb1, 0x80, - 0xa2, 0xab, 0x50, 0x6e, 0x39, 0x61, 0xec, 0xc5, 0x5e, 0xe0, 0x4f, 0x17, 0x2f, 0x59, 0x57, 0x06, - 0x17, 0xa6, 0x04, 0x6a, 0xb9, 0x26, 0x01, 0x38, 0xc1, 0xa1, 0xdd, 0x08, 0x89, 0xe3, 0xde, 0xf2, - 0x9b, 0xfb, 0xd3, 0x03, 0x97, 0xac, 0x2b, 0xa5, 0xa4, 0x1b, 0x58, 0x94, 0x63, 0x85, 0x61, 0xff, - 0x70, 0x01, 0x4a, 0xf3, 0x1b, 0x1b, 0x9e, 0xef, 0xc5, 0xfb, 0xe8, 0x0e, 0x8c, 0xfa, 0x81, 0x4b, - 0xe4, 0x7f, 0xf6, 0x15, 0x23, 0xcf, 0x5d, 0x9a, 0xeb, 0x5c, 0x4a, 0x73, 0xab, 0x1a, 0xde, 0xc2, - 0xe4, 0xfd, 0x83, 0xd9, 0x51, 0xbd, 0x04, 0x1b, 0x74, 0x10, 0x86, 0x91, 0x56, 0xe0, 0x2a, 0xb2, - 0x05, 0x46, 0x76, 0x36, 0x8b, 0x6c, 0x2d, 0x41, 0x5b, 0x98, 0xb8, 0x7f, 0x30, 0x3b, 0xa2, 0x15, - 0x60, 0x9d, 0x08, 0x5a, 0x87, 0x09, 0xfa, 0xd7, 0x8f, 0x3d, 0x45, 0xb7, 0xc8, 0xe8, 0x3e, 0x9e, - 0x47, 0x57, 0x43, 0x5d, 0x38, 0x75, 0xff, 0x60, 0x76, 0x22, 0x55, 0x88, 0xd3, 0x04, 0xed, 0xf7, - 0x60, 0x7c, 0x3e, 0x8e, 0x9d, 0xc6, 0x16, 0x71, 0xf9, 0x0c, 0xa2, 0x17, 0x60, 0xc0, 0x77, 0x76, - 0x88, 0x98, 0xdf, 0x4b, 0x62, 0x60, 0x07, 0x56, 0x9d, 0x1d, 0x72, 0x78, 0x30, 0x3b, 0x79, 0xdb, - 0xf7, 0xde, 0x6d, 0x8b, 0x55, 0x41, 0xcb, 0x30, 0xc3, 0x46, 0xcf, 0x01, 0xb8, 0x64, 0xd7, 0x6b, - 0x90, 0x9a, 0x13, 0x6f, 0x89, 0xf9, 0x46, 0xa2, 0x2e, 0x54, 0x14, 0x04, 0x6b, 0x58, 0xf6, 0x3d, - 0x28, 0xcf, 0xef, 0x06, 0x9e, 0x5b, 0x0b, 0xdc, 0x08, 0x6d, 0xc3, 0x44, 0x2b, 0x24, 0x1b, 0x24, - 0x54, 0x45, 0xd3, 0xd6, 0xa5, 0xe2, 0x95, 0x91, 0xe7, 0xae, 0x64, 0x7e, 0xac, 0x89, 0xba, 0xe4, - 0xc7, 0xe1, 0xfe, 0xc2, 0x23, 0xa2, 0xbd, 0x89, 0x14, 0x14, 0xa7, 0x29, 0xdb, 0xff, 0xbc, 0x00, - 0x67, 0xe6, 0xdf, 0x6b, 0x87, 0xa4, 0xe2, 0x45, 0xdb, 0xe9, 0x15, 0xee, 0x7a, 0xd1, 0xf6, 0x6a, - 0x32, 0x02, 0x6a, 0x69, 0x55, 0x44, 0x39, 0x56, 0x18, 0xe8, 0x19, 0x18, 0xa6, 0xbf, 0x6f, 0xe3, - 0xaa, 0xf8, 0xe4, 0x53, 0x02, 0x79, 0xa4, 0xe2, 0xc4, 0x4e, 0x85, 0x83, 0xb0, 0xc4, 0x41, 0x2b, - 0x30, 0xd2, 0x60, 0x1b, 0x72, 0x73, 0x25, 0x70, 0x09, 0x9b, 0xcc, 0xf2, 0xc2, 0x53, 0x14, 0x7d, - 0x31, 0x29, 0x3e, 0x3c, 0x98, 0x9d, 0xe6, 0x7d, 0x13, 0x24, 0x34, 0x18, 0xd6, 0xeb, 0x23, 0x5b, - 0xed, 0xaf, 0x01, 0x46, 0x09, 0x32, 0xf6, 0xd6, 0x15, 0x6d, 0xab, 0x0c, 0xb2, 0xad, 0x32, 0x9a, - 0xbd, 0x4d, 0xd0, 0xb3, 0x30, 0xb0, 0xed, 0xf9, 0xee, 0xf4, 0x10, 0xa3, 0x75, 0x81, 0xce, 0xf9, - 0x0d, 0xcf, 0x77, 0x0f, 0x0f, 0x66, 0xa7, 0x8c, 0xee, 0xd0, 0x42, 0xcc, 0x50, 0xed, 0x3f, 0xb1, - 0x60, 0x96, 0xc1, 0x96, 0xbd, 0x26, 0xa9, 0x91, 0x30, 0xf2, 0xa2, 0x98, 0xf8, 0xb1, 0x31, 0xa0, - 0xcf, 0x01, 0x44, 0xa4, 0x11, 0x92, 0x58, 0x1b, 0x52, 0xb5, 0x30, 0xea, 0x0a, 0x82, 0x35, 0x2c, - 0x7a, 0x20, 0x44, 0x5b, 0x4e, 0xc8, 0xd6, 0x97, 0x18, 0x58, 0x75, 0x20, 0xd4, 0x25, 0x00, 0x27, - 0x38, 0xc6, 0x81, 0x50, 0xec, 0x75, 0x20, 0xa0, 0x4f, 0xc1, 0x44, 0xd2, 0x58, 0xd4, 0x72, 0x1a, - 0x72, 0x00, 0xd9, 0x96, 0xa9, 0x9b, 0x20, 0x9c, 0xc6, 0xb5, 0xff, 0x8e, 0x25, 0x16, 0x0f, 0xfd, - 0xea, 0x0f, 0xf8, 0xb7, 0xda, 0xbf, 0x6c, 0xc1, 0xf0, 0x82, 0xe7, 0xbb, 0x9e, 0xbf, 0x89, 0x3e, - 0x0f, 0x25, 0x7a, 0x37, 0xb9, 0x4e, 0xec, 0x88, 0x73, 0xef, 0x13, 0xda, 0xde, 0x52, 0x57, 0xc5, - 0x5c, 0x6b, 0x7b, 0x93, 0x16, 0x44, 0x73, 0x14, 0x9b, 0xee, 0xb6, 0x5b, 0xeb, 0xef, 0x90, 0x46, - 0xbc, 0x42, 0x62, 0x27, 0xf9, 0x9c, 0xa4, 0x0c, 0x2b, 0xaa, 0xe8, 0x06, 0x0c, 0xc5, 0x4e, 0xb8, - 0x49, 0x62, 0x71, 0x00, 0x66, 0x1e, 0x54, 0xbc, 0x26, 0xa6, 0x3b, 0x92, 0xf8, 0x0d, 0x92, 0x5c, - 0x0b, 0x6b, 0xac, 0x2a, 0x16, 0x24, 0xec, 0x1f, 0x1c, 0x86, 0x73, 0x8b, 0xf5, 0x6a, 0xce, 0xba, - 0xba, 0x0c, 0x43, 0x6e, 0xe8, 0xed, 0x92, 0x50, 0x8c, 0xb3, 0xa2, 0x52, 0x61, 0xa5, 0x58, 0x40, - 0xd1, 0xcb, 0x30, 0xca, 0x2f, 0xa4, 0xeb, 0x8e, 0xef, 0x36, 0xe5, 0x10, 0x9f, 0x16, 0xd8, 0xa3, - 0x77, 0x34, 0x18, 0x36, 0x30, 0x8f, 0xb8, 0xa8, 0x2e, 0xa7, 0x36, 0x63, 0xde, 0x65, 0xf7, 0x45, - 0x0b, 0x26, 0x79, 0x33, 0xf3, 0x71, 0x1c, 0x7a, 0xeb, 0xed, 0x98, 0x44, 0xd3, 0x83, 0xec, 0xa4, - 0x5b, 0xcc, 0x1a, 0xad, 0xdc, 0x11, 0x98, 0xbb, 0x93, 0xa2, 0xc2, 0x0f, 0xc1, 0x69, 0xd1, 0xee, - 0x64, 0x1a, 0x8c, 0x3b, 0x9a, 0x45, 0xdf, 0x69, 0xc1, 0x4c, 0x23, 0xf0, 0xe3, 0x30, 0x68, 0x36, - 0x49, 0x58, 0x6b, 0xaf, 0x37, 0xbd, 0x68, 0x8b, 0xaf, 0x53, 0x4c, 0x36, 0xd8, 0x49, 0x90, 0x33, - 0x87, 0x0a, 0x49, 0xcc, 0xe1, 0xc5, 0xfb, 0x07, 0xb3, 0x33, 0x8b, 0xb9, 0xa4, 0x70, 0x97, 0x66, - 0xd0, 0x36, 0x20, 0x7a, 0x95, 0xd6, 0x63, 0x67, 0x93, 0x24, 0x8d, 0x0f, 0xf7, 0xdf, 0xf8, 0xd9, - 0xfb, 0x07, 0xb3, 0x68, 0xb5, 0x83, 0x04, 0xce, 0x20, 0x8b, 0xde, 0x85, 0xd3, 0xb4, 0xb4, 0xe3, - 0x5b, 0x4b, 0xfd, 0x37, 0x37, 0x7d, 0xff, 0x60, 0xf6, 0xf4, 0x6a, 0x06, 0x11, 0x9c, 0x49, 0x1a, - 0x7d, 0x87, 0x05, 0xe7, 0x92, 0xcf, 0x5f, 0xba, 0xd7, 0x72, 0x7c, 0x37, 0x69, 0xb8, 0xdc, 0x7f, - 0xc3, 0xf4, 0x4c, 0x3e, 0xb7, 0x98, 0x47, 0x09, 0xe7, 0x37, 0x32, 0xb3, 0x08, 0x67, 0x32, 0x57, - 0x0b, 0x9a, 0x84, 0xe2, 0x36, 0xe1, 0x5c, 0x50, 0x19, 0xd3, 0x9f, 0xe8, 0x34, 0x0c, 0xee, 0x3a, - 0xcd, 0xb6, 0xd8, 0x28, 0x98, 0xff, 0x79, 0xa5, 0xf0, 0xb2, 0x65, 0xff, 0x8b, 0x22, 0x4c, 0x2c, - 0xd6, 0xab, 0x0f, 0xb4, 0x0b, 0xf5, 0x6b, 0xa8, 0xd0, 0xf5, 0x1a, 0x4a, 0x2e, 0xb5, 0x62, 0xee, - 0xa5, 0xf6, 0xff, 0x64, 0x6c, 0xa1, 0x01, 0xb6, 0x85, 0xbe, 0x29, 0x67, 0x0b, 0x1d, 0xf3, 0xc6, - 0xd9, 0xcd, 0x59, 0x45, 0x83, 0x6c, 0x32, 0x33, 0x39, 0x96, 0x9b, 0x41, 0xc3, 0x69, 0xa6, 0x8f, - 0xbe, 0x23, 0x2e, 0xa5, 0xe3, 0x99, 0xc7, 0x06, 0x8c, 0x2e, 0x3a, 0x2d, 0x67, 0xdd, 0x6b, 0x7a, - 0xb1, 0x47, 0x22, 0xf4, 0x04, 0x14, 0x1d, 0xd7, 0x65, 0xdc, 0x56, 0x79, 0xe1, 0xcc, 0xfd, 0x83, - 0xd9, 0xe2, 0xbc, 0x4b, 0xaf, 0x7d, 0x50, 0x58, 0xfb, 0x98, 0x62, 0xa0, 0x8f, 0xc3, 0x80, 0x1b, - 0x06, 0xad, 0xe9, 0x02, 0xc3, 0xa4, 0xbb, 0x6e, 0xa0, 0x12, 0x06, 0xad, 0x14, 0x2a, 0xc3, 0xb1, - 0x7f, 0xb5, 0x00, 0xe7, 0x17, 0x49, 0x6b, 0x6b, 0xb9, 0x9e, 0x73, 0x7e, 0x5f, 0x81, 0xd2, 0x4e, - 0xe0, 0x7b, 0x71, 0x10, 0x46, 0xa2, 0x69, 0xb6, 0x22, 0x56, 0x44, 0x19, 0x56, 0x50, 0x74, 0x09, - 0x06, 0x5a, 0x09, 0x53, 0x39, 0x2a, 0x19, 0x52, 0xc6, 0x4e, 0x32, 0x08, 0xc5, 0x68, 0x47, 0x24, - 0x14, 0x2b, 0x46, 0x61, 0xdc, 0x8e, 0x48, 0x88, 0x19, 0x24, 0xb9, 0x99, 0xe9, 0x9d, 0x2d, 0x4e, - 0xe8, 0xd4, 0xcd, 0x4c, 0x21, 0x58, 0xc3, 0x42, 0x35, 0x28, 0x47, 0xa9, 0x99, 0xed, 0x6b, 0x9b, - 0x8e, 0xb1, 0xab, 0x5b, 0xcd, 0x64, 0x42, 0xc4, 0xb8, 0x51, 0x86, 0x7a, 0x5e, 0xdd, 0x5f, 0x29, - 0x00, 0xe2, 0x43, 0xf8, 0x17, 0x6c, 0xe0, 0x6e, 0x77, 0x0e, 0x5c, 0xff, 0x5b, 0xe2, 0xb8, 0x46, - 0xef, 0x4f, 0x2d, 0x38, 0xbf, 0xe8, 0xf9, 0x2e, 0x09, 0x73, 0x16, 0xe0, 0xc3, 0x79, 0xcb, 0x1e, - 0x8d, 0x69, 0x30, 0x96, 0xd8, 0xc0, 0x31, 0x2c, 0x31, 0xfb, 0x8f, 0x2c, 0x40, 0xfc, 0xb3, 0x3f, - 0x70, 0x1f, 0x7b, 0xbb, 0xf3, 0x63, 0x8f, 0x61, 0x59, 0xd8, 0x37, 0x61, 0x7c, 0xb1, 0xe9, 0x11, - 0x3f, 0xae, 0xd6, 0x16, 0x03, 0x7f, 0xc3, 0xdb, 0x44, 0xaf, 0xc0, 0x78, 0xec, 0xed, 0x90, 0xa0, - 0x1d, 0xd7, 0x49, 0x23, 0xf0, 0xd9, 0x4b, 0xd2, 0xba, 0x32, 0xb8, 0x80, 0xee, 0x1f, 0xcc, 0x8e, - 0xaf, 0x19, 0x10, 0x9c, 0xc2, 0xb4, 0x7f, 0x87, 0x8e, 0x5f, 0xb0, 0xd3, 0x0a, 0x7c, 0xe2, 0xc7, - 0x8b, 0x81, 0xef, 0x72, 0x89, 0xc3, 0x2b, 0x30, 0x10, 0xd3, 0xf1, 0xe0, 0x63, 0x77, 0x59, 0x6e, - 0x14, 0x3a, 0x0a, 0x87, 0x07, 0xb3, 0x67, 0x3b, 0x6b, 0xb0, 0x71, 0x62, 0x75, 0xd0, 0x37, 0xc1, - 0x50, 0x14, 0x3b, 0x71, 0x3b, 0x12, 0xa3, 0xf9, 0x98, 0x1c, 0xcd, 0x3a, 0x2b, 0x3d, 0x3c, 0x98, - 0x9d, 0x50, 0xd5, 0x78, 0x11, 0x16, 0x15, 0xd0, 0x93, 0x30, 0xbc, 0x43, 0xa2, 0xc8, 0xd9, 0x94, - 0xb7, 0xe1, 0x84, 0xa8, 0x3b, 0xbc, 0xc2, 0x8b, 0xb1, 0x84, 0xa3, 0xc7, 0x61, 0x90, 0x84, 0x61, - 0x10, 0x8a, 0x3d, 0x3a, 0x26, 0x10, 0x07, 0x97, 0x68, 0x21, 0xe6, 0x30, 0xfb, 0xdf, 0x58, 0x30, - 0xa1, 0xfa, 0xca, 0xdb, 0x3a, 0x81, 0x57, 0xc1, 0x5b, 0x00, 0x0d, 0xf9, 0x81, 0x11, 0xbb, 0x3d, - 0x46, 0x9e, 0xbb, 0x9c, 0x79, 0x51, 0x77, 0x0c, 0x63, 0x42, 0x59, 0x15, 0x45, 0x58, 0xa3, 0x66, - 0xff, 0x63, 0x0b, 0x4e, 0xa5, 0xbe, 0xe8, 0xa6, 0x17, 0xc5, 0xe8, 0xed, 0x8e, 0xaf, 0x9a, 0xeb, - 0xef, 0xab, 0x68, 0x6d, 0xf6, 0x4d, 0x6a, 0x29, 0xcb, 0x12, 0xed, 0x8b, 0xae, 0xc3, 0xa0, 0x17, - 0x93, 0x1d, 0xf9, 0x31, 0x8f, 0x77, 0xfd, 0x18, 0xde, 0xab, 0x64, 0x46, 0xaa, 0xb4, 0x26, 0xe6, - 0x04, 0xec, 0xbf, 0x5c, 0x84, 0x32, 0x5f, 0xb6, 0x2b, 0x4e, 0xeb, 0x04, 0xe6, 0xa2, 0x0a, 0x03, - 0x8c, 0x3a, 0xef, 0xf8, 0x13, 0xd9, 0x1d, 0x17, 0xdd, 0x99, 0xa3, 0x4f, 0x7e, 0xce, 0x1c, 0xa9, - 0xab, 0x81, 0x16, 0x61, 0x46, 0x02, 0x39, 0x00, 0xeb, 0x9e, 0xef, 0x84, 0xfb, 0xb4, 0x6c, 0xba, - 0xc8, 0x08, 0x3e, 0xd3, 0x9d, 0xe0, 0x82, 0xc2, 0xe7, 0x64, 0x55, 0x5f, 0x13, 0x00, 0xd6, 0x88, - 0xce, 0xbc, 0x04, 0x65, 0x85, 0x7c, 0x14, 0x1e, 0x67, 0xe6, 0x53, 0x30, 0x91, 0x6a, 0xab, 0x57, - 0xf5, 0x51, 0x9d, 0x45, 0xfa, 0x32, 0x3b, 0x05, 0x44, 0xaf, 0x97, 0xfc, 0x5d, 0x71, 0x8a, 0xbe, - 0x07, 0xa7, 0x9b, 0x19, 0x87, 0x93, 0x98, 0xaa, 0xfe, 0x0f, 0xb3, 0xf3, 0xe2, 0xb3, 0x4f, 0x67, - 0x41, 0x71, 0x66, 0x1b, 0xf4, 0xda, 0x0f, 0x5a, 0x74, 0xcd, 0x3b, 0x4d, 0x9d, 0x83, 0xbe, 0x25, - 0xca, 0xb0, 0x82, 0xd2, 0x23, 0xec, 0xb4, 0xea, 0xfc, 0x0d, 0xb2, 0x5f, 0x27, 0x4d, 0xd2, 0x88, - 0x83, 0xf0, 0xeb, 0xda, 0xfd, 0x0b, 0x7c, 0xf4, 0xf9, 0x09, 0x38, 0x22, 0x08, 0x14, 0x6f, 0x90, - 0x7d, 0x3e, 0x15, 0xfa, 0xd7, 0x15, 0xbb, 0x7e, 0xdd, 0xcf, 0x59, 0x30, 0xa6, 0xbe, 0xee, 0x04, - 0xb6, 0xfa, 0x82, 0xb9, 0xd5, 0x2f, 0x74, 0x5d, 0xe0, 0x39, 0x9b, 0xfc, 0x2b, 0x05, 0x38, 0xa7, - 0x70, 0x28, 0xbb, 0xcf, 0xff, 0x88, 0x55, 0x75, 0x15, 0xca, 0xbe, 0x12, 0x44, 0x59, 0xa6, 0x04, - 0x28, 0x11, 0x43, 0x25, 0x38, 0x94, 0x6b, 0xf3, 0x13, 0x69, 0xd1, 0xa8, 0x2e, 0xa1, 0x15, 0xd2, - 0xd8, 0x05, 0x28, 0xb6, 0x3d, 0x57, 0xdc, 0x19, 0x9f, 0x90, 0xa3, 0x7d, 0xbb, 0x5a, 0x39, 0x3c, - 0x98, 0x7d, 0x2c, 0x4f, 0x3b, 0x40, 0x2f, 0xab, 0x68, 0xee, 0x76, 0xb5, 0x82, 0x69, 0x65, 0x34, - 0x0f, 0x13, 0x52, 0x01, 0x72, 0x87, 0x72, 0x50, 0x81, 0x2f, 0xae, 0x16, 0x25, 0x66, 0xc5, 0x26, - 0x18, 0xa7, 0xf1, 0x51, 0x05, 0x26, 0xb7, 0xdb, 0xeb, 0xa4, 0x49, 0x62, 0xfe, 0xc1, 0x37, 0x08, - 0x17, 0x42, 0x96, 0x93, 0xc7, 0xd6, 0x8d, 0x14, 0x1c, 0x77, 0xd4, 0xb0, 0xff, 0x9c, 0x1d, 0xf1, - 0x62, 0xf4, 0x6a, 0x61, 0x40, 0x17, 0x16, 0xa5, 0xfe, 0xf5, 0x5c, 0xce, 0xfd, 0xac, 0x8a, 0x1b, - 0x64, 0x7f, 0x2d, 0xa0, 0xcc, 0x76, 0xf6, 0xaa, 0x30, 0xd6, 0xfc, 0x40, 0xd7, 0x35, 0xff, 0x0b, - 0x05, 0x38, 0xa3, 0x46, 0xc0, 0xe0, 0xeb, 0xfe, 0xa2, 0x8f, 0xc1, 0xb3, 0x30, 0xe2, 0x92, 0x0d, - 0xa7, 0xdd, 0x8c, 0x95, 0x44, 0x7c, 0x90, 0x6b, 0x45, 0x2a, 0x49, 0x31, 0xd6, 0x71, 0x8e, 0x30, - 0x6c, 0xff, 0x63, 0x84, 0xdd, 0xad, 0xb1, 0x43, 0xd7, 0xb8, 0xda, 0x35, 0x56, 0xee, 0xae, 0x79, - 0x1c, 0x06, 0xbd, 0x1d, 0xca, 0x6b, 0x15, 0x4c, 0x16, 0xaa, 0x4a, 0x0b, 0x31, 0x87, 0xa1, 0x8f, - 0xc1, 0x70, 0x23, 0xd8, 0xd9, 0x71, 0x7c, 0x97, 0x5d, 0x79, 0xe5, 0x85, 0x11, 0xca, 0x8e, 0x2d, - 0xf2, 0x22, 0x2c, 0x61, 0xe8, 0x3c, 0x0c, 0x38, 0xe1, 0x26, 0x17, 0x4b, 0x94, 0x17, 0x4a, 0xb4, - 0xa5, 0xf9, 0x70, 0x33, 0xc2, 0xac, 0x94, 0xbe, 0xaa, 0xf6, 0x82, 0x70, 0xdb, 0xf3, 0x37, 0x2b, - 0x5e, 0x28, 0xb6, 0x84, 0xba, 0x0b, 0xef, 0x2a, 0x08, 0xd6, 0xb0, 0xd0, 0x32, 0x0c, 0xb6, 0x82, - 0x30, 0x8e, 0xa6, 0x87, 0xd8, 0x70, 0x3f, 0x96, 0x73, 0x10, 0xf1, 0xaf, 0xad, 0x05, 0x61, 0x9c, - 0x7c, 0x00, 0xfd, 0x17, 0x61, 0x5e, 0x1d, 0xdd, 0x84, 0x61, 0xe2, 0xef, 0x2e, 0x87, 0xc1, 0xce, - 0xf4, 0xa9, 0x7c, 0x4a, 0x4b, 0x1c, 0x85, 0x2f, 0xb3, 0x84, 0xed, 0x14, 0xc5, 0x58, 0x92, 0x40, - 0xdf, 0x04, 0x45, 0xe2, 0xef, 0x4e, 0x0f, 0x33, 0x4a, 0x33, 0x39, 0x94, 0xee, 0x38, 0x61, 0x72, - 0xe6, 0x2f, 0xf9, 0xbb, 0x98, 0xd6, 0x41, 0x9f, 0x81, 0xb2, 0x3c, 0x30, 0x22, 0x21, 0x7f, 0xcb, - 0x5c, 0xb0, 0xf2, 0x98, 0xc1, 0xe4, 0xdd, 0xb6, 0x17, 0x92, 0x1d, 0xe2, 0xc7, 0x51, 0x72, 0x42, - 0x4a, 0x68, 0x84, 0x13, 0x6a, 0xe8, 0x33, 0x52, 0xe8, 0xbb, 0x12, 0xb4, 0xfd, 0x38, 0x9a, 0x2e, - 0xb3, 0xee, 0x65, 0xaa, 0xe3, 0xee, 0x24, 0x78, 0x69, 0xa9, 0x30, 0xaf, 0x8c, 0x0d, 0x52, 0xe8, - 0xb3, 0x30, 0xc6, 0xff, 0x73, 0xa5, 0x56, 0x34, 0x7d, 0x86, 0xd1, 0xbe, 0x94, 0x4f, 0x9b, 0x23, - 0x2e, 0x9c, 0x11, 0xc4, 0xc7, 0xf4, 0xd2, 0x08, 0x9b, 0xd4, 0x10, 0x86, 0xb1, 0xa6, 0xb7, 0x4b, - 0x7c, 0x12, 0x45, 0xb5, 0x30, 0x58, 0x27, 0xd3, 0xc0, 0x06, 0xe6, 0x5c, 0xb6, 0x12, 0x2c, 0x58, - 0x27, 0x0b, 0x53, 0x94, 0xe6, 0x4d, 0xbd, 0x0e, 0x36, 0x49, 0xa0, 0xdb, 0x30, 0x4e, 0x1f, 0x61, - 0x5e, 0x42, 0x74, 0xa4, 0x17, 0x51, 0xf6, 0x54, 0xc2, 0x46, 0x25, 0x9c, 0x22, 0x82, 0x6e, 0xc1, - 0x68, 0x14, 0x3b, 0x61, 0xdc, 0x6e, 0x71, 0xa2, 0x67, 0x7b, 0x11, 0x65, 0x3a, 0xd4, 0xba, 0x56, - 0x05, 0x1b, 0x04, 0xd0, 0x1b, 0x50, 0x6e, 0x7a, 0x1b, 0xa4, 0xb1, 0xdf, 0x68, 0x92, 0xe9, 0x51, - 0x46, 0x2d, 0xf3, 0x50, 0xb9, 0x29, 0x91, 0xf8, 0xab, 0x50, 0xfd, 0xc5, 0x49, 0x75, 0x74, 0x07, - 0xce, 0xc6, 0x24, 0xdc, 0xf1, 0x7c, 0x87, 0x1e, 0x06, 0xe2, 0xb5, 0xc4, 0x74, 0x93, 0x63, 0x6c, - 0xb7, 0x5d, 0x14, 0xb3, 0x71, 0x76, 0x2d, 0x13, 0x0b, 0xe7, 0xd4, 0x46, 0xf7, 0x60, 0x3a, 0x03, - 0x12, 0x34, 0xbd, 0xc6, 0xfe, 0xf4, 0x69, 0x46, 0xf9, 0x35, 0x41, 0x79, 0x7a, 0x2d, 0x07, 0xef, - 0xb0, 0x0b, 0x0c, 0xe7, 0x52, 0x47, 0xb7, 0x60, 0x82, 0x9d, 0x40, 0xb5, 0x76, 0xb3, 0x29, 0x1a, - 0x1c, 0x67, 0x0d, 0x7e, 0x4c, 0xde, 0xc7, 0x55, 0x13, 0x7c, 0x78, 0x30, 0x0b, 0xc9, 0x3f, 0x9c, - 0xae, 0x8d, 0xd6, 0x99, 0x1a, 0xac, 0x1d, 0x7a, 0xf1, 0x3e, 0x3d, 0x37, 0xc8, 0xbd, 0x78, 0x7a, - 0xa2, 0xab, 0x08, 0x42, 0x47, 0x55, 0xba, 0x32, 0xbd, 0x10, 0xa7, 0x09, 0xd2, 0x23, 0x35, 0x8a, - 0x5d, 0xcf, 0x9f, 0x9e, 0x64, 0x27, 0xb5, 0x3a, 0x91, 0xea, 0xb4, 0x10, 0x73, 0x18, 0x53, 0x81, - 0xd1, 0x1f, 0xb7, 0xe8, 0xcd, 0x35, 0xc5, 0x10, 0x13, 0x15, 0x98, 0x04, 0xe0, 0x04, 0x87, 0x32, - 0x93, 0x71, 0xbc, 0x3f, 0x8d, 0x18, 0xaa, 0x3a, 0x58, 0xd6, 0xd6, 0x3e, 0x83, 0x69, 0xb9, 0xbd, - 0x0e, 0xe3, 0xea, 0x20, 0x64, 0x63, 0x82, 0x66, 0x61, 0x90, 0xb1, 0x4f, 0x42, 0x60, 0x56, 0xa6, - 0x5d, 0x60, 0xac, 0x15, 0xe6, 0xe5, 0xac, 0x0b, 0xde, 0x7b, 0x64, 0x61, 0x3f, 0x26, 0xfc, 0x99, - 0x5e, 0xd4, 0xba, 0x20, 0x01, 0x38, 0xc1, 0xb1, 0xff, 0x37, 0x67, 0x43, 0x93, 0xd3, 0xb6, 0x8f, - 0xfb, 0xe5, 0x69, 0x28, 0x6d, 0x05, 0x51, 0x4c, 0xb1, 0x59, 0x1b, 0x83, 0x09, 0xe3, 0x79, 0x5d, - 0x94, 0x63, 0x85, 0x81, 0x5e, 0x85, 0xb1, 0x86, 0xde, 0x80, 0xb8, 0x1c, 0xd5, 0x31, 0x62, 0xb4, - 0x8e, 0x4d, 0x5c, 0xf4, 0x32, 0x94, 0x98, 0x59, 0x47, 0x23, 0x68, 0x0a, 0xae, 0x4d, 0xde, 0xf0, - 0xa5, 0x9a, 0x28, 0x3f, 0xd4, 0x7e, 0x63, 0x85, 0x8d, 0x2e, 0xc3, 0x10, 0xed, 0x42, 0xb5, 0x26, - 0xae, 0x25, 0x25, 0xfb, 0xb9, 0xce, 0x4a, 0xb1, 0x80, 0xda, 0x7f, 0xa9, 0xa0, 0x8d, 0x32, 0x7d, - 0xe2, 0x12, 0x54, 0x83, 0xe1, 0x3d, 0xc7, 0x8b, 0x3d, 0x7f, 0x53, 0xf0, 0x1f, 0x4f, 0x76, 0xbd, - 0xa3, 0x58, 0xa5, 0xbb, 0xbc, 0x02, 0xbf, 0x45, 0xc5, 0x1f, 0x2c, 0xc9, 0x50, 0x8a, 0x61, 0xdb, - 0xf7, 0x29, 0xc5, 0x42, 0xbf, 0x14, 0x31, 0xaf, 0xc0, 0x29, 0x8a, 0x3f, 0x58, 0x92, 0x41, 0x6f, - 0x03, 0xc8, 0x1d, 0x46, 0x5c, 0x61, 0x4e, 0xf1, 0x74, 0x6f, 0xa2, 0x6b, 0xaa, 0xce, 0xc2, 0x38, - 0xbd, 0xa3, 0x93, 0xff, 0x58, 0xa3, 0x67, 0xc7, 0x8c, 0x4f, 0xeb, 0xec, 0x0c, 0xfa, 0x56, 0xba, - 0xc4, 0x9d, 0x30, 0x26, 0xee, 0x7c, 0x2c, 0x06, 0xe7, 0xe3, 0xfd, 0x3d, 0x52, 0xd6, 0xbc, 0x1d, - 0xa2, 0x6f, 0x07, 0x41, 0x04, 0x27, 0xf4, 0xec, 0x5f, 0x2a, 0xc2, 0x74, 0x5e, 0x77, 0xe9, 0xa2, - 0x23, 0xf7, 0xbc, 0x78, 0x91, 0xb2, 0x57, 0x96, 0xb9, 0xe8, 0x96, 0x44, 0x39, 0x56, 0x18, 0x74, - 0xf6, 0x23, 0x6f, 0x53, 0xbe, 0x31, 0x07, 0x93, 0xd9, 0xaf, 0xb3, 0x52, 0x2c, 0xa0, 0x14, 0x2f, - 0x24, 0x4e, 0x24, 0xec, 0x75, 0xb4, 0x55, 0x82, 0x59, 0x29, 0x16, 0x50, 0x5d, 0x80, 0x35, 0xd0, - 0x43, 0x80, 0x65, 0x0c, 0xd1, 0xe0, 0xf1, 0x0e, 0x11, 0xfa, 0x1c, 0xc0, 0x86, 0xe7, 0x7b, 0xd1, - 0x16, 0xa3, 0x3e, 0x74, 0x64, 0xea, 0x8a, 0x39, 0x5b, 0x56, 0x54, 0xb0, 0x46, 0x11, 0xbd, 0x08, - 0x23, 0x6a, 0x03, 0x56, 0x2b, 0x4c, 0x79, 0xa9, 0x19, 0x83, 0x24, 0xa7, 0x51, 0x05, 0xeb, 0x78, - 0xf6, 0x3b, 0xe9, 0xf5, 0x22, 0x76, 0x80, 0x36, 0xbe, 0x56, 0xbf, 0xe3, 0x5b, 0xe8, 0x3e, 0xbe, - 0xf6, 0xd7, 0x8a, 0x30, 0x61, 0x34, 0xd6, 0x8e, 0xfa, 0x38, 0xb3, 0xae, 0xd1, 0x03, 0xdc, 0x89, - 0x89, 0xd8, 0x7f, 0x76, 0xef, 0xad, 0xa2, 0x1f, 0xf2, 0x74, 0x07, 0xf0, 0xfa, 0xe8, 0x73, 0x50, - 0x6e, 0x3a, 0x11, 0x13, 0x86, 0x11, 0xb1, 0xef, 0xfa, 0x21, 0x96, 0x3c, 0x4c, 0x9c, 0x28, 0xd6, - 0x6e, 0x4d, 0x4e, 0x3b, 0x21, 0x49, 0x6f, 0x1a, 0xca, 0x9f, 0x48, 0x83, 0x30, 0xd5, 0x09, 0xca, - 0xc4, 0xec, 0x63, 0x0e, 0x43, 0x2f, 0xc3, 0x68, 0x48, 0xd8, 0xaa, 0x58, 0xa4, 0xdc, 0x1c, 0x5b, - 0x66, 0x83, 0x09, 0xdb, 0x87, 0x35, 0x18, 0x36, 0x30, 0x93, 0xb7, 0xc1, 0x50, 0x97, 0xb7, 0xc1, - 0x93, 0x30, 0xcc, 0x7e, 0xa8, 0x15, 0xa0, 0x66, 0xa3, 0xca, 0x8b, 0xb1, 0x84, 0xa7, 0x17, 0x4c, - 0xa9, 0xbf, 0x05, 0x43, 0x5f, 0x1f, 0x62, 0x51, 0x33, 0xc5, 0x71, 0x89, 0x9f, 0x72, 0x62, 0xc9, - 0x63, 0x09, 0xb3, 0x3f, 0x0e, 0xe3, 0x15, 0x87, 0xec, 0x04, 0xfe, 0x92, 0xef, 0xb6, 0x02, 0xcf, - 0x8f, 0xd1, 0x34, 0x0c, 0xb0, 0x4b, 0x84, 0x1f, 0x01, 0x03, 0xb4, 0x21, 0x3c, 0x40, 0x1f, 0x04, - 0xf6, 0x26, 0x9c, 0xa9, 0x04, 0x7b, 0xfe, 0x9e, 0x13, 0xba, 0xf3, 0xb5, 0xaa, 0xf6, 0xbe, 0x5e, - 0x95, 0xef, 0x3b, 0x6e, 0x87, 0x95, 0x79, 0xf4, 0x6a, 0x35, 0x39, 0x5b, 0xbb, 0xec, 0x35, 0x49, - 0x8e, 0x14, 0xe4, 0xaf, 0x16, 0x8c, 0x96, 0x12, 0x7c, 0xa5, 0xa8, 0xb2, 0x72, 0x15, 0x55, 0x6f, - 0x42, 0x69, 0xc3, 0x23, 0x4d, 0x17, 0x93, 0x0d, 0xb1, 0x12, 0x9f, 0xc8, 0x37, 0x2d, 0x59, 0xa6, - 0x98, 0x52, 0xea, 0xc5, 0x5f, 0x87, 0xcb, 0xa2, 0x32, 0x56, 0x64, 0xd0, 0x36, 0x4c, 0xca, 0x07, - 0x83, 0x84, 0x8a, 0x75, 0xf9, 0x64, 0xb7, 0x57, 0x88, 0x49, 0xfc, 0xf4, 0xfd, 0x83, 0xd9, 0x49, - 0x9c, 0x22, 0x83, 0x3b, 0x08, 0xd3, 0xe7, 0xe0, 0x0e, 0x3d, 0x81, 0x07, 0xd8, 0xf0, 0xb3, 0xe7, - 0x20, 0x7b, 0xd9, 0xb2, 0x52, 0xfb, 0x47, 0x2d, 0x78, 0xa4, 0x63, 0x64, 0xc4, 0x0b, 0xff, 0x98, - 0x67, 0x21, 0xfd, 0xe2, 0x2e, 0xf4, 0x7e, 0x71, 0xdb, 0x3f, 0x6b, 0xc1, 0xe9, 0xa5, 0x9d, 0x56, - 0xbc, 0x5f, 0xf1, 0x4c, 0xad, 0xd2, 0x4b, 0x30, 0xb4, 0x43, 0x5c, 0xaf, 0xbd, 0x23, 0x66, 0x6e, - 0x56, 0x9e, 0x52, 0x2b, 0xac, 0xf4, 0xf0, 0x60, 0x76, 0xac, 0x1e, 0x07, 0xa1, 0xb3, 0x49, 0x78, - 0x01, 0x16, 0xe8, 0xec, 0xac, 0xf7, 0xde, 0x23, 0x37, 0xbd, 0x1d, 0x4f, 0x9a, 0x0a, 0x75, 0x95, - 0xd9, 0xcd, 0xc9, 0x01, 0x9d, 0x7b, 0xb3, 0xed, 0xf8, 0xb1, 0x17, 0xef, 0x0b, 0x85, 0x90, 0x24, - 0x82, 0x13, 0x7a, 0xf6, 0x57, 0x2d, 0x98, 0x90, 0xeb, 0x7e, 0xde, 0x75, 0x43, 0x12, 0x45, 0x68, - 0x06, 0x0a, 0x5e, 0x4b, 0xf4, 0x12, 0x44, 0x2f, 0x0b, 0xd5, 0x1a, 0x2e, 0x78, 0x2d, 0xc9, 0x96, - 0xb1, 0x83, 0xb0, 0x68, 0xea, 0xc6, 0xae, 0x8b, 0x72, 0xac, 0x30, 0xd0, 0x15, 0x28, 0xf9, 0x81, - 0xcb, 0xcd, 0xb5, 0xf8, 0x95, 0xc6, 0x16, 0xd8, 0xaa, 0x28, 0xc3, 0x0a, 0x8a, 0x6a, 0x50, 0xe6, - 0x96, 0x4c, 0xc9, 0xa2, 0xed, 0xcb, 0x1e, 0x8a, 0x7d, 0xd9, 0x9a, 0xac, 0x89, 0x13, 0x22, 0xf6, - 0x0f, 0x58, 0x30, 0x2a, 0xbf, 0xac, 0x4f, 0x9e, 0x93, 0x6e, 0xad, 0x84, 0xdf, 0x4c, 0xb6, 0x16, - 0xe5, 0x19, 0x19, 0xc4, 0x60, 0x15, 0x8b, 0x47, 0x61, 0x15, 0xed, 0x1f, 0x29, 0xc0, 0xb8, 0xec, - 0x4e, 0xbd, 0xbd, 0x1e, 0x91, 0x18, 0xad, 0x41, 0xd9, 0xe1, 0x43, 0x4e, 0xe4, 0x8a, 0x7d, 0x3c, - 0x5b, 0x28, 0x60, 0xcc, 0x4f, 0x72, 0x7b, 0xcf, 0xcb, 0xda, 0x38, 0x21, 0x84, 0x9a, 0x30, 0xe5, - 0x07, 0x31, 0x3b, 0xc9, 0x15, 0xbc, 0x9b, 0xea, 0x25, 0x4d, 0xfd, 0x9c, 0xa0, 0x3e, 0xb5, 0x9a, - 0xa6, 0x82, 0x3b, 0x09, 0xa3, 0x25, 0x29, 0x68, 0x29, 0xe6, 0xbf, 0xec, 0xf5, 0x59, 0xc8, 0x96, - 0xb3, 0xd8, 0xbf, 0x62, 0x41, 0x59, 0xa2, 0x9d, 0x84, 0x96, 0x6d, 0x05, 0x86, 0x23, 0x36, 0x09, - 0x72, 0x68, 0xec, 0x6e, 0x1d, 0xe7, 0xf3, 0x95, 0x5c, 0x50, 0xfc, 0x7f, 0x84, 0x25, 0x0d, 0x26, - 0x67, 0x57, 0xdd, 0xff, 0x80, 0xc8, 0xd9, 0x55, 0x7f, 0x72, 0x6e, 0x98, 0x3f, 0x60, 0x7d, 0xd6, - 0x04, 0x57, 0x94, 0x8f, 0x6a, 0x85, 0x64, 0xc3, 0xbb, 0x97, 0xe6, 0xa3, 0x6a, 0xac, 0x14, 0x0b, - 0x28, 0x7a, 0x1b, 0x46, 0x1b, 0x52, 0xc0, 0x9a, 0x6c, 0xd7, 0xcb, 0x5d, 0x85, 0xfd, 0x4a, 0x2f, - 0xc4, 0x05, 0x1b, 0x8b, 0x5a, 0x7d, 0x6c, 0x50, 0x33, 0xd5, 0xfc, 0xc5, 0x5e, 0x6a, 0xfe, 0x84, - 0x6e, 0xbe, 0xd2, 0xfb, 0xc7, 0x2c, 0x18, 0xe2, 0x82, 0xb5, 0xfe, 0xe4, 0x9a, 0x9a, 0x9a, 0x2c, - 0x19, 0xbb, 0x3b, 0xb4, 0x50, 0xa8, 0xbd, 0xd0, 0x0a, 0x94, 0xd9, 0x0f, 0x26, 0x18, 0x2c, 0xe6, - 0x5b, 0xc5, 0xf3, 0x56, 0xf5, 0x0e, 0xde, 0x91, 0xd5, 0x70, 0x42, 0xc1, 0xfe, 0xa1, 0x22, 0x3d, - 0xaa, 0x12, 0x54, 0xe3, 0x06, 0xb7, 0x1e, 0xde, 0x0d, 0x5e, 0x78, 0x58, 0x37, 0xf8, 0x26, 0x4c, - 0x34, 0x34, 0xa5, 0x5a, 0x32, 0x93, 0x57, 0xba, 0x2e, 0x12, 0x4d, 0xff, 0xc6, 0x45, 0x26, 0x8b, - 0x26, 0x11, 0x9c, 0xa6, 0x8a, 0xbe, 0x15, 0x46, 0xf9, 0x3c, 0x8b, 0x56, 0xb8, 0xa5, 0xc4, 0xc7, - 0xf2, 0xd7, 0x8b, 0xde, 0x04, 0x17, 0xb1, 0x69, 0xd5, 0xb1, 0x41, 0xcc, 0xfe, 0x63, 0x0b, 0xd0, - 0x52, 0x6b, 0x8b, 0xec, 0x90, 0xd0, 0x69, 0x26, 0xb2, 0xf1, 0xff, 0xcf, 0x82, 0x69, 0xd2, 0x51, - 0xbc, 0x18, 0xec, 0xec, 0x88, 0x17, 0x48, 0xce, 0x23, 0x79, 0x29, 0xa7, 0x8e, 0x72, 0x1b, 0x98, - 0xce, 0xc3, 0xc0, 0xb9, 0xed, 0xa1, 0x15, 0x38, 0xc5, 0xaf, 0x3c, 0x05, 0xd0, 0x6c, 0xa3, 0x1f, - 0x15, 0x84, 0x4f, 0xad, 0x75, 0xa2, 0xe0, 0xac, 0x7a, 0xf6, 0x77, 0x8d, 0x42, 0x6e, 0x2f, 0x3e, - 0x54, 0x0a, 0x7c, 0xa8, 0x14, 0xf8, 0x50, 0x29, 0xf0, 0xa1, 0x52, 0xe0, 0x43, 0xa5, 0xc0, 0x37, - 0xbc, 0x52, 0xe0, 0x0f, 0x2d, 0x38, 0xd5, 0x79, 0x0d, 0x9c, 0x04, 0x63, 0xde, 0x86, 0x53, 0x9d, - 0x77, 0x5d, 0x57, 0x3b, 0xb8, 0xce, 0x7e, 0x26, 0xf7, 0x5e, 0xc6, 0x37, 0xe0, 0x2c, 0xfa, 0xf6, - 0x2f, 0x95, 0x60, 0x70, 0x69, 0x97, 0xf8, 0xf1, 0x09, 0x7c, 0x62, 0x03, 0xc6, 0x3d, 0x7f, 0x37, - 0x68, 0xee, 0x12, 0x97, 0xc3, 0x8f, 0xf2, 0xde, 0x3d, 0x2b, 0x48, 0x8f, 0x57, 0x0d, 0x12, 0x38, - 0x45, 0xf2, 0x61, 0xc8, 0x9c, 0xaf, 0xc1, 0x10, 0xbf, 0x1d, 0x84, 0xc0, 0x39, 0xf3, 0x32, 0x60, - 0x83, 0x28, 0xee, 0xbc, 0x44, 0x1e, 0xce, 0x6f, 0x1f, 0x51, 0x1d, 0xbd, 0x03, 0xe3, 0x1b, 0x5e, - 0x18, 0xc5, 0x6b, 0xde, 0x0e, 0x89, 0x62, 0x67, 0xa7, 0xf5, 0x00, 0x32, 0x66, 0x35, 0x0e, 0xcb, - 0x06, 0x25, 0x9c, 0xa2, 0x8c, 0x36, 0x61, 0xac, 0xe9, 0xe8, 0x4d, 0x0d, 0x1f, 0xb9, 0x29, 0x75, - 0xed, 0xdc, 0xd4, 0x09, 0x61, 0x93, 0x2e, 0xdd, 0xa7, 0x0d, 0x26, 0x26, 0x2d, 0x31, 0xe1, 0x81, - 0xda, 0xa7, 0x5c, 0x3e, 0xca, 0x61, 0x94, 0x83, 0x62, 0x96, 0xb1, 0x65, 0x93, 0x83, 0xd2, 0xec, - 0x5f, 0x3f, 0x0f, 0x65, 0x42, 0x87, 0x90, 0x12, 0x16, 0x37, 0xd7, 0xd5, 0xfe, 0xfa, 0xba, 0xe2, - 0x35, 0xc2, 0xc0, 0x94, 0xee, 0x2f, 0x49, 0x4a, 0x38, 0x21, 0x8a, 0x16, 0x61, 0x28, 0x22, 0xa1, - 0x47, 0x22, 0x71, 0x87, 0x75, 0x99, 0x46, 0x86, 0xc6, 0x9d, 0x4a, 0xf8, 0x6f, 0x2c, 0xaa, 0xd2, - 0xe5, 0xe5, 0x30, 0xc1, 0x27, 0xbb, 0x65, 0xb4, 0xe5, 0x35, 0xcf, 0x4a, 0xb1, 0x80, 0xa2, 0x37, - 0x60, 0x38, 0x24, 0x4d, 0xa6, 0x3e, 0x1a, 0xeb, 0x7f, 0x91, 0x73, 0x6d, 0x14, 0xaf, 0x87, 0x25, - 0x01, 0x74, 0x03, 0x50, 0x48, 0x28, 0x07, 0xe6, 0xf9, 0x9b, 0xca, 0x5e, 0x54, 0x9c, 0xe0, 0x6a, - 0xc7, 0xe3, 0x04, 0x43, 0xfa, 0xf7, 0xe0, 0x8c, 0x6a, 0xe8, 0x1a, 0x4c, 0xa9, 0xd2, 0xaa, 0x1f, - 0xc5, 0x0e, 0x3d, 0x39, 0x27, 0x18, 0x2d, 0x25, 0x00, 0xc1, 0x69, 0x04, 0xdc, 0x59, 0xc7, 0xfe, - 0x69, 0x0b, 0xf8, 0x38, 0x9f, 0xc0, 0xb3, 0xff, 0x75, 0xf3, 0xd9, 0x7f, 0x2e, 0x77, 0xe6, 0x72, - 0x9e, 0xfc, 0xf7, 0x2d, 0x18, 0xd1, 0x66, 0x36, 0x59, 0xb3, 0x56, 0x97, 0x35, 0xdb, 0x86, 0x49, - 0xba, 0xd2, 0x6f, 0xad, 0x47, 0x24, 0xdc, 0x25, 0x2e, 0x5b, 0x98, 0x85, 0x07, 0x5b, 0x98, 0xca, - 0x90, 0xed, 0x66, 0x8a, 0x20, 0xee, 0x68, 0x02, 0xbd, 0x24, 0x75, 0x29, 0x45, 0xc3, 0x0e, 0x9c, - 0xeb, 0x49, 0x0e, 0x0f, 0x66, 0x27, 0xb5, 0x0f, 0xd1, 0x75, 0x27, 0xf6, 0xe7, 0xe5, 0x37, 0x2a, - 0x83, 0xc1, 0x86, 0x5a, 0x2c, 0x29, 0x83, 0x41, 0xb5, 0x1c, 0x70, 0x82, 0x43, 0xf7, 0xe8, 0x56, - 0x10, 0xc5, 0x69, 0x83, 0xc1, 0xeb, 0x41, 0x14, 0x63, 0x06, 0xb1, 0x9f, 0x07, 0x58, 0xba, 0x47, - 0x1a, 0x7c, 0xa9, 0xeb, 0xcf, 0x19, 0x2b, 0xff, 0x39, 0x63, 0xff, 0x3b, 0x0b, 0xc6, 0x97, 0x17, - 0x0d, 0x89, 0xf0, 0x1c, 0x00, 0x7f, 0x83, 0xdd, 0xbd, 0xbb, 0x2a, 0xb5, 0xed, 0x5c, 0x61, 0xaa, - 0x4a, 0xb1, 0x86, 0x81, 0xce, 0x41, 0xb1, 0xd9, 0xf6, 0x85, 0x74, 0x72, 0x98, 0x5e, 0xd8, 0x37, - 0xdb, 0x3e, 0xa6, 0x65, 0x9a, 0x13, 0x42, 0xb1, 0x6f, 0x27, 0x84, 0x9e, 0xc1, 0x00, 0xd0, 0x2c, - 0x0c, 0xee, 0xed, 0x79, 0x2e, 0x77, 0xb9, 0x14, 0x96, 0x00, 0x77, 0xef, 0x56, 0x2b, 0x11, 0xe6, - 0xe5, 0xf6, 0x97, 0x8a, 0x30, 0xb3, 0xdc, 0x24, 0xf7, 0xde, 0xa7, 0xdb, 0x69, 0xbf, 0x2e, 0x14, - 0x47, 0x13, 0x0d, 0x1d, 0xd5, 0x4d, 0xa6, 0xf7, 0x78, 0x6c, 0xc0, 0x30, 0xb7, 0x97, 0x93, 0x4e, - 0xa8, 0xaf, 0x66, 0xb5, 0x9e, 0x3f, 0x20, 0x73, 0xdc, 0xee, 0x4e, 0xf8, 0xd0, 0xa9, 0x9b, 0x56, - 0x94, 0x62, 0x49, 0x7c, 0xe6, 0x15, 0x18, 0xd5, 0x31, 0x8f, 0xe4, 0xb0, 0xf6, 0xff, 0x16, 0x61, - 0x92, 0xf6, 0xe0, 0xa1, 0x4e, 0xc4, 0xed, 0xce, 0x89, 0x38, 0x6e, 0xa7, 0xa5, 0xde, 0xb3, 0xf1, - 0x76, 0x7a, 0x36, 0x9e, 0xcd, 0x9b, 0x8d, 0x93, 0x9e, 0x83, 0xef, 0xb4, 0xe0, 0xd4, 0x72, 0x33, - 0x68, 0x6c, 0xa7, 0x1c, 0x8b, 0x5e, 0x84, 0x11, 0x7a, 0x8e, 0x47, 0x86, 0xcf, 0xbb, 0x11, 0x05, - 0x41, 0x80, 0xb0, 0x8e, 0xa7, 0x55, 0xbb, 0x7d, 0xbb, 0x5a, 0xc9, 0x0a, 0x9e, 0x20, 0x40, 0x58, - 0xc7, 0xb3, 0x7f, 0xd3, 0x82, 0x0b, 0xd7, 0x16, 0x97, 0x92, 0xa5, 0xd8, 0x11, 0xbf, 0xe1, 0x32, - 0x0c, 0xb5, 0x5c, 0xad, 0x2b, 0x89, 0xc0, 0xb7, 0xc2, 0x7a, 0x21, 0xa0, 0x1f, 0x94, 0xd8, 0x24, - 0x3f, 0x65, 0xc1, 0xa9, 0x6b, 0x5e, 0x4c, 0xaf, 0xe5, 0x74, 0x24, 0x01, 0x7a, 0x2f, 0x47, 0x5e, - 0x1c, 0x84, 0xfb, 0xe9, 0x48, 0x02, 0x58, 0x41, 0xb0, 0x86, 0xc5, 0x5b, 0xde, 0xf5, 0x98, 0xa5, - 0x76, 0xc1, 0xd4, 0x63, 0x61, 0x51, 0x8e, 0x15, 0x06, 0xfd, 0x30, 0xd7, 0x0b, 0x99, 0xd4, 0x70, - 0x5f, 0x9c, 0xb0, 0xea, 0xc3, 0x2a, 0x12, 0x80, 0x13, 0x1c, 0xfa, 0x80, 0x9a, 0xbd, 0xd6, 0x6c, - 0x47, 0x31, 0x09, 0x37, 0xa2, 0x9c, 0xd3, 0xf1, 0x79, 0x28, 0x13, 0x29, 0xa3, 0x17, 0xbd, 0x56, - 0xac, 0xa6, 0x12, 0xde, 0xf3, 0x80, 0x06, 0x0a, 0xaf, 0x0f, 0x37, 0xc5, 0xa3, 0xf9, 0x99, 0x2d, - 0x03, 0x22, 0x7a, 0x5b, 0x7a, 0x84, 0x07, 0xe6, 0x2a, 0xbe, 0xd4, 0x01, 0xc5, 0x19, 0x35, 0xec, - 0x1f, 0xb5, 0xe0, 0x8c, 0xfa, 0xe0, 0x0f, 0xdc, 0x67, 0xda, 0x3f, 0x5f, 0x80, 0xb1, 0xeb, 0x6b, - 0x6b, 0xb5, 0x6b, 0x24, 0x16, 0xd7, 0x76, 0x6f, 0x35, 0x3a, 0xd6, 0xb4, 0x81, 0xdd, 0x5e, 0x81, - 0xed, 0xd8, 0x6b, 0xce, 0xf1, 0x40, 0x41, 0x73, 0x55, 0x3f, 0xbe, 0x15, 0xd6, 0xe3, 0xd0, 0xf3, - 0x37, 0x33, 0xf5, 0x87, 0x92, 0xb9, 0x28, 0xe6, 0x31, 0x17, 0xe8, 0x79, 0x18, 0x62, 0x91, 0x8a, - 0xe4, 0x24, 0x3c, 0xaa, 0x1e, 0x51, 0xac, 0xf4, 0xf0, 0x60, 0xb6, 0x7c, 0x1b, 0x57, 0xf9, 0x1f, - 0x2c, 0x50, 0xd1, 0x6d, 0x18, 0xd9, 0x8a, 0xe3, 0xd6, 0x75, 0xe2, 0xb8, 0xf4, 0xb5, 0xcc, 0x8f, - 0xc3, 0x8b, 0x59, 0xc7, 0x21, 0x1d, 0x04, 0x8e, 0x96, 0x9c, 0x20, 0x49, 0x59, 0x84, 0x75, 0x3a, - 0x76, 0x1d, 0x20, 0x81, 0x1d, 0x93, 0xee, 0xc4, 0xfe, 0x7d, 0x0b, 0x86, 0x79, 0xd0, 0x88, 0x10, - 0xbd, 0x06, 0x03, 0xe4, 0x1e, 0x69, 0x08, 0x56, 0x39, 0xb3, 0xc3, 0x09, 0xa7, 0xc5, 0x65, 0xc0, - 0xf4, 0x3f, 0x66, 0xb5, 0xd0, 0x75, 0x18, 0xa6, 0xbd, 0xbd, 0xa6, 0x22, 0x68, 0x3c, 0x96, 0xf7, - 0xc5, 0x6a, 0xda, 0x39, 0x73, 0x26, 0x8a, 0xb0, 0xac, 0xce, 0xb4, 0xcf, 0x8d, 0x56, 0x9d, 0x9e, - 0xd8, 0x71, 0x37, 0xc6, 0x62, 0x6d, 0xb1, 0xc6, 0x91, 0x04, 0x35, 0xae, 0x7d, 0x96, 0x85, 0x38, - 0x21, 0x62, 0xaf, 0x41, 0x99, 0x4e, 0xea, 0x7c, 0xd3, 0x73, 0xba, 0x2b, 0xd4, 0x9f, 0x82, 0xb2, - 0x54, 0x97, 0x47, 0xc2, 0x59, 0x9c, 0x51, 0x95, 0xda, 0xf4, 0x08, 0x27, 0x70, 0x7b, 0x03, 0x4e, - 0x33, 0xe3, 0x47, 0x27, 0xde, 0x32, 0xf6, 0x58, 0xef, 0xc5, 0xfc, 0xb4, 0x78, 0x79, 0xf2, 0x99, - 0x99, 0xd6, 0xfc, 0x31, 0x47, 0x25, 0xc5, 0xe4, 0x15, 0x6a, 0x7f, 0x6d, 0x00, 0x1e, 0xad, 0xd6, - 0xf3, 0xe3, 0x89, 0xbc, 0x0c, 0xa3, 0x9c, 0x2f, 0xa5, 0x4b, 0xdb, 0x69, 0x8a, 0x76, 0x95, 0xf0, - 0x77, 0x4d, 0x83, 0x61, 0x03, 0x13, 0x5d, 0x80, 0xa2, 0xf7, 0xae, 0x9f, 0x76, 0x6d, 0xaa, 0xbe, - 0xb9, 0x8a, 0x69, 0x39, 0x05, 0x53, 0x16, 0x97, 0xdf, 0x1d, 0x0a, 0xac, 0xd8, 0xdc, 0xd7, 0x61, - 0xdc, 0x8b, 0x1a, 0x91, 0x57, 0xf5, 0xe9, 0x39, 0xa3, 0x9d, 0x54, 0x4a, 0x2a, 0x42, 0x3b, 0xad, - 0xa0, 0x38, 0x85, 0xad, 0x5d, 0x64, 0x83, 0x7d, 0xb3, 0xc9, 0x3d, 0xbd, 0xa7, 0xe9, 0x0b, 0xa0, - 0xc5, 0xbe, 0x2e, 0x62, 0x52, 0x7c, 0xf1, 0x02, 0xe0, 0x1f, 0x1c, 0x61, 0x09, 0xa3, 0x4f, 0xce, - 0xc6, 0x96, 0xd3, 0x9a, 0x6f, 0xc7, 0x5b, 0x15, 0x2f, 0x6a, 0x04, 0xbb, 0x24, 0xdc, 0x67, 0xd2, - 0x82, 0x52, 0xf2, 0xe4, 0x54, 0x80, 0xc5, 0xeb, 0xf3, 0x35, 0x8a, 0x89, 0x3b, 0xeb, 0xa0, 0x79, - 0x98, 0x90, 0x85, 0x75, 0x12, 0xb1, 0x2b, 0x6c, 0x84, 0x91, 0x51, 0xce, 0x46, 0xa2, 0x58, 0x11, - 0x49, 0xe3, 0x9b, 0x9c, 0x34, 0x1c, 0x07, 0x27, 0xfd, 0x12, 0x8c, 0x79, 0xbe, 0x17, 0x7b, 0x4e, - 0x1c, 0x70, 0x15, 0x14, 0x17, 0x0c, 0x30, 0xd9, 0x7a, 0x55, 0x07, 0x60, 0x13, 0xcf, 0xfe, 0x2f, - 0x03, 0x30, 0xc5, 0xa6, 0xed, 0xc3, 0x15, 0xf6, 0x8d, 0xb4, 0xc2, 0x6e, 0x77, 0xae, 0xb0, 0xe3, - 0x78, 0x22, 0x3c, 0xf0, 0x32, 0x7b, 0x07, 0xca, 0xca, 0xbf, 0x4a, 0x3a, 0x58, 0x5a, 0x39, 0x0e, - 0x96, 0xbd, 0xb9, 0x0f, 0x69, 0xa2, 0x56, 0xcc, 0x34, 0x51, 0xfb, 0xeb, 0x16, 0x24, 0x3a, 0x15, - 0x74, 0x1d, 0xca, 0xad, 0x80, 0x59, 0x5e, 0x86, 0xd2, 0x9c, 0xf9, 0xd1, 0xcc, 0x8b, 0x8a, 0x5f, - 0x8a, 0xfc, 0xe3, 0x6b, 0xb2, 0x06, 0x4e, 0x2a, 0xa3, 0x05, 0x18, 0x6e, 0x85, 0xa4, 0x1e, 0xb3, - 0xb0, 0x22, 0x3d, 0xe9, 0xf0, 0x35, 0xc2, 0xf1, 0xb1, 0xac, 0x68, 0xff, 0x82, 0x05, 0xc0, 0xad, - 0xc0, 0x1c, 0x7f, 0x93, 0x9c, 0x80, 0xb8, 0xbb, 0x02, 0x03, 0x51, 0x8b, 0x34, 0xba, 0xd9, 0xc4, - 0x26, 0xfd, 0xa9, 0xb7, 0x48, 0x23, 0x19, 0x70, 0xfa, 0x0f, 0xb3, 0xda, 0xf6, 0x77, 0x03, 0x8c, - 0x27, 0x68, 0xd5, 0x98, 0xec, 0xa0, 0x67, 0x8c, 0x30, 0x03, 0xe7, 0x52, 0x61, 0x06, 0xca, 0x0c, - 0x5b, 0x93, 0xac, 0xbe, 0x03, 0xc5, 0x1d, 0xe7, 0x9e, 0x10, 0x9d, 0x3d, 0xd5, 0xbd, 0x1b, 0x94, - 0xfe, 0xdc, 0x8a, 0x73, 0x8f, 0x3f, 0x12, 0x9f, 0x92, 0x0b, 0x64, 0xc5, 0xb9, 0x77, 0xc8, 0x2d, - 0x5f, 0xd9, 0x21, 0x75, 0xd3, 0x8b, 0xe2, 0x2f, 0xfc, 0xe7, 0xe4, 0x3f, 0x5b, 0x76, 0xb4, 0x11, - 0xd6, 0x96, 0xe7, 0x0b, 0x9b, 0xa8, 0xbe, 0xda, 0xf2, 0xfc, 0x74, 0x5b, 0x9e, 0xdf, 0x47, 0x5b, - 0x9e, 0x8f, 0xde, 0x83, 0x61, 0x61, 0x7f, 0x28, 0xc2, 0xfa, 0x5c, 0xed, 0xa3, 0x3d, 0x61, 0xbe, - 0xc8, 0xdb, 0xbc, 0x2a, 0x1f, 0xc1, 0xa2, 0xb4, 0x67, 0xbb, 0xb2, 0x41, 0xf4, 0x57, 0x2c, 0x18, - 0x17, 0xbf, 0x31, 0x79, 0xb7, 0x4d, 0xa2, 0x58, 0xf0, 0x9e, 0x9f, 0xec, 0xbf, 0x0f, 0xa2, 0x22, - 0xef, 0xca, 0x27, 0xe5, 0x31, 0x6b, 0x02, 0x7b, 0xf6, 0x28, 0xd5, 0x0b, 0xf4, 0xf7, 0x2c, 0x38, - 0xbd, 0xe3, 0xdc, 0xe3, 0x2d, 0xf2, 0x32, 0xec, 0xc4, 0x5e, 0x20, 0x54, 0xff, 0xaf, 0xf5, 0x37, - 0xfd, 0x1d, 0xd5, 0x79, 0x27, 0xa5, 0x7e, 0xf2, 0x74, 0x16, 0x4a, 0xcf, 0xae, 0x66, 0xf6, 0x6b, - 0x66, 0x03, 0x4a, 0x72, 0xbd, 0x65, 0x88, 0x1a, 0x2a, 0x3a, 0x63, 0x7d, 0x64, 0xf3, 0x4f, 0xdd, - 0xd7, 0x9f, 0xb6, 0x23, 0xd6, 0xda, 0x43, 0x6d, 0xe7, 0x1d, 0x18, 0xd5, 0xd7, 0xd8, 0x43, 0x6d, - 0xeb, 0x5d, 0x38, 0x95, 0xb1, 0x96, 0x1e, 0x6a, 0x93, 0x7b, 0x70, 0x2e, 0x77, 0x7d, 0x3c, 0xcc, - 0x86, 0xed, 0x9f, 0xb7, 0xf4, 0x73, 0xf0, 0x04, 0x74, 0x0e, 0x8b, 0xa6, 0xce, 0xe1, 0x62, 0xf7, - 0x9d, 0x93, 0xa3, 0x78, 0x78, 0x5b, 0xef, 0x34, 0x3d, 0xd5, 0xd1, 0x1b, 0x30, 0xd4, 0xa4, 0x25, - 0xd2, 0xf0, 0xd5, 0xee, 0xbd, 0x23, 0x13, 0x5e, 0x8a, 0x95, 0x47, 0x58, 0x50, 0xb0, 0x7f, 0xd9, - 0x82, 0x81, 0x13, 0x18, 0x09, 0x6c, 0x8e, 0xc4, 0x33, 0xb9, 0xa4, 0x45, 0xc4, 0xe1, 0x39, 0xec, - 0xec, 0x2d, 0xdd, 0x8b, 0x89, 0x1f, 0xb1, 0xa7, 0x62, 0xe6, 0xc0, 0xfc, 0x5f, 0x70, 0xea, 0x66, - 0xe0, 0xb8, 0x0b, 0x4e, 0xd3, 0xf1, 0x1b, 0x24, 0xac, 0xfa, 0x9b, 0x47, 0xb2, 0xc0, 0x2e, 0xf4, - 0xb2, 0xc0, 0xb6, 0xb7, 0x00, 0xe9, 0x0d, 0x08, 0x57, 0x16, 0x0c, 0xc3, 0x1e, 0x6f, 0x4a, 0x0c, - 0xff, 0x13, 0xd9, 0xac, 0x59, 0x47, 0xcf, 0x34, 0x27, 0x0d, 0x5e, 0x80, 0x25, 0x21, 0xfb, 0x65, - 0xc8, 0xf4, 0x87, 0xef, 0x2d, 0x36, 0xb0, 0x3f, 0x03, 0x53, 0xac, 0xe6, 0x11, 0x9f, 0xb4, 0x76, - 0x4a, 0x2a, 0x99, 0x11, 0xfc, 0xce, 0xfe, 0xa2, 0x05, 0x13, 0xab, 0xa9, 0x98, 0x60, 0x97, 0x99, - 0x02, 0x34, 0x43, 0x18, 0x5e, 0x67, 0xa5, 0x58, 0x40, 0x8f, 0x5d, 0x06, 0xf5, 0xe7, 0x16, 0x24, - 0x21, 0x2a, 0x4e, 0x80, 0xf1, 0x5a, 0x34, 0x18, 0xaf, 0x4c, 0xd9, 0x88, 0xea, 0x4e, 0x1e, 0xdf, - 0x85, 0x6e, 0xa8, 0x78, 0x4c, 0x5d, 0xc4, 0x22, 0x09, 0x19, 0x1e, 0xbd, 0x67, 0xdc, 0x0c, 0xda, - 0x24, 0x23, 0x34, 0xd9, 0xff, 0xb1, 0x00, 0x48, 0xe1, 0xf6, 0x1d, 0x2f, 0xaa, 0xb3, 0xc6, 0xf1, - 0xc4, 0x8b, 0xda, 0x05, 0xc4, 0x54, 0xf8, 0xa1, 0xe3, 0x47, 0x9c, 0xac, 0x27, 0xa4, 0x6e, 0x47, - 0xb3, 0x0f, 0x98, 0x11, 0x4d, 0xa2, 0x9b, 0x1d, 0xd4, 0x70, 0x46, 0x0b, 0x9a, 0x69, 0xc6, 0x60, - 0xbf, 0xa6, 0x19, 0x43, 0x3d, 0xdc, 0xd5, 0x7e, 0xce, 0x82, 0x31, 0x35, 0x4c, 0x1f, 0x10, 0xfb, - 0x73, 0xd5, 0x9f, 0x9c, 0xa3, 0xaf, 0xa6, 0x75, 0x99, 0x5d, 0x09, 0xdf, 0xcc, 0xdc, 0x0e, 0x9d, - 0xa6, 0xf7, 0x1e, 0x51, 0xd1, 0xfa, 0x66, 0x85, 0x1b, 0xa1, 0x28, 0x3d, 0x3c, 0x98, 0x1d, 0x53, - 0xff, 0x78, 0x74, 0xe0, 0xa4, 0x8a, 0xfd, 0x13, 0x74, 0xb3, 0x9b, 0x4b, 0x11, 0xbd, 0x08, 0x83, - 0xad, 0x2d, 0x27, 0x22, 0x29, 0xa7, 0x9b, 0xc1, 0x1a, 0x2d, 0x3c, 0x3c, 0x98, 0x1d, 0x57, 0x15, - 0x58, 0x09, 0xe6, 0xd8, 0xfd, 0x47, 0xe1, 0xea, 0x5c, 0x9c, 0x3d, 0xa3, 0x70, 0xfd, 0xb1, 0x05, - 0x03, 0xab, 0x81, 0x7b, 0x12, 0x47, 0xc0, 0xeb, 0xc6, 0x11, 0x70, 0x3e, 0x2f, 0x70, 0x7b, 0xee, - 0xee, 0x5f, 0x4e, 0xed, 0xfe, 0x8b, 0xb9, 0x14, 0xba, 0x6f, 0xfc, 0x1d, 0x18, 0x61, 0xe1, 0xe0, - 0x85, 0x83, 0xd1, 0xf3, 0xc6, 0x86, 0x9f, 0x4d, 0x6d, 0xf8, 0x09, 0x0d, 0x55, 0xdb, 0xe9, 0x4f, - 0xc2, 0xb0, 0x70, 0x72, 0x49, 0x7b, 0x6f, 0x0a, 0x5c, 0x2c, 0xe1, 0xf6, 0x8f, 0x15, 0xc1, 0x08, - 0x3f, 0x8f, 0x7e, 0xc5, 0x82, 0xb9, 0x90, 0x1b, 0xbf, 0xba, 0x95, 0x76, 0xe8, 0xf9, 0x9b, 0xf5, - 0xc6, 0x16, 0x71, 0xdb, 0x4d, 0xcf, 0xdf, 0xac, 0x6e, 0xfa, 0x81, 0x2a, 0x5e, 0xba, 0x47, 0x1a, - 0x6d, 0xa6, 0xbe, 0xea, 0x11, 0xeb, 0x5e, 0x19, 0x91, 0x3f, 0x77, 0xff, 0x60, 0x76, 0x0e, 0x1f, - 0x89, 0x36, 0x3e, 0x62, 0x5f, 0xd0, 0x6f, 0x5a, 0x70, 0x95, 0x47, 0x65, 0xef, 0xbf, 0xff, 0x5d, - 0xde, 0xb9, 0x35, 0x49, 0x2a, 0x21, 0xb2, 0x46, 0xc2, 0x9d, 0x85, 0x97, 0xc4, 0x80, 0x5e, 0xad, - 0x1d, 0xad, 0x2d, 0x7c, 0xd4, 0xce, 0xd9, 0xff, 0xac, 0x08, 0x63, 0x22, 0xb4, 0x93, 0xb8, 0x03, - 0x5e, 0x34, 0x96, 0xc4, 0x63, 0xa9, 0x25, 0x31, 0x65, 0x20, 0x1f, 0xcf, 0xf1, 0x1f, 0xc1, 0x14, - 0x3d, 0x9c, 0xaf, 0x13, 0x27, 0x8c, 0xd7, 0x89, 0xc3, 0x2d, 0xae, 0x8a, 0x47, 0x3e, 0xfd, 0x95, - 0x60, 0xed, 0x66, 0x9a, 0x18, 0xee, 0xa4, 0xff, 0x8d, 0x74, 0xe7, 0xf8, 0x30, 0xd9, 0x11, 0x9d, - 0xeb, 0x2d, 0x28, 0x2b, 0x0f, 0x0d, 0x71, 0xe8, 0x74, 0x0f, 0x72, 0x97, 0xa6, 0xc0, 0x85, 0x5f, - 0x89, 0x77, 0x50, 0x42, 0xce, 0xfe, 0xfb, 0x05, 0xa3, 0x41, 0x3e, 0x89, 0xab, 0x50, 0x72, 0xa2, - 0xc8, 0xdb, 0xf4, 0x89, 0x2b, 0x76, 0xec, 0x47, 0xf3, 0x76, 0xac, 0xd1, 0x0c, 0xf3, 0x92, 0x99, - 0x17, 0x35, 0xb1, 0xa2, 0x81, 0xae, 0x73, 0xbb, 0xb6, 0x5d, 0xf9, 0x52, 0xeb, 0x8f, 0x1a, 0x48, - 0xcb, 0xb7, 0x5d, 0x82, 0x45, 0x7d, 0xf4, 0x59, 0x6e, 0x78, 0x78, 0xc3, 0x0f, 0xf6, 0xfc, 0x6b, - 0x41, 0x20, 0xc3, 0x27, 0xf4, 0x47, 0x70, 0x4a, 0x9a, 0x1b, 0xaa, 0xea, 0xd8, 0xa4, 0xd6, 0x5f, - 0x04, 0xcb, 0x6f, 0x83, 0x53, 0x94, 0xb4, 0xe9, 0xdd, 0x1c, 0x21, 0x02, 0x13, 0x22, 0x6e, 0x98, - 0x2c, 0x13, 0x63, 0x97, 0xf9, 0x08, 0x33, 0x6b, 0x27, 0x12, 0xe0, 0x1b, 0x26, 0x09, 0x9c, 0xa6, - 0x69, 0xff, 0xa4, 0x05, 0xcc, 0xd3, 0xf3, 0x04, 0xf8, 0x91, 0x4f, 0x99, 0xfc, 0xc8, 0x74, 0xde, - 0x20, 0xe7, 0xb0, 0x22, 0x2f, 0xf0, 0x95, 0x55, 0x0b, 0x83, 0x7b, 0xfb, 0xc2, 0xe8, 0xa3, 0xf7, - 0xfb, 0xc3, 0xfe, 0x5f, 0x16, 0x3f, 0xc4, 0x94, 0xff, 0x04, 0xfa, 0x76, 0x28, 0x35, 0x9c, 0x96, - 0xd3, 0xe0, 0xb9, 0x52, 0x72, 0x65, 0x71, 0x46, 0xa5, 0xb9, 0x45, 0x51, 0x83, 0xcb, 0x96, 0x64, - 0xfc, 0xb9, 0x92, 0x2c, 0xee, 0x29, 0x4f, 0x52, 0x4d, 0xce, 0x6c, 0xc3, 0x98, 0x41, 0xec, 0xa1, - 0x0a, 0x22, 0xbe, 0x9d, 0x5f, 0xb1, 0x2a, 0x5e, 0xe2, 0x0e, 0x4c, 0xf9, 0xda, 0x7f, 0x7a, 0xa1, - 0xc8, 0xc7, 0xe5, 0x47, 0x7b, 0x5d, 0xa2, 0xec, 0xf6, 0xd1, 0xfc, 0x4e, 0x53, 0x64, 0x70, 0x27, - 0x65, 0xfb, 0xc7, 0x2d, 0x78, 0x44, 0x47, 0xd4, 0x5c, 0x5b, 0x7a, 0x49, 0xf7, 0x2b, 0x50, 0x0a, - 0x5a, 0x24, 0x74, 0xe2, 0x20, 0x14, 0xb7, 0xc6, 0x15, 0x39, 0xe8, 0xb7, 0x44, 0xf9, 0xa1, 0x88, - 0x34, 0x2e, 0xa9, 0xcb, 0x72, 0xac, 0x6a, 0xd2, 0xd7, 0x27, 0x1b, 0x8c, 0x48, 0x38, 0x31, 0xb1, - 0x33, 0x80, 0x29, 0xba, 0x23, 0x2c, 0x20, 0xf6, 0xd7, 0x2c, 0xbe, 0xb0, 0xf4, 0xae, 0xa3, 0x77, - 0x61, 0x72, 0xc7, 0x89, 0x1b, 0x5b, 0x4b, 0xf7, 0x5a, 0x21, 0xd7, 0x95, 0xc8, 0x71, 0x7a, 0xaa, - 0xd7, 0x38, 0x69, 0x1f, 0x99, 0xd8, 0x52, 0xae, 0xa4, 0x88, 0xe1, 0x0e, 0xf2, 0x68, 0x1d, 0x46, - 0x58, 0x19, 0xf3, 0xcf, 0x8b, 0xba, 0xb1, 0x06, 0x79, 0xad, 0x29, 0x5b, 0x81, 0x95, 0x84, 0x0e, - 0xd6, 0x89, 0xda, 0x3f, 0x53, 0xe4, 0xbb, 0x9d, 0xb1, 0xf2, 0x4f, 0xc2, 0x70, 0x2b, 0x70, 0x17, - 0xab, 0x15, 0x2c, 0x66, 0x41, 0x5d, 0x23, 0x35, 0x5e, 0x8c, 0x25, 0x1c, 0x5d, 0x81, 0x92, 0xf8, - 0x29, 0x75, 0x5b, 0xec, 0x6c, 0x16, 0x78, 0x11, 0x56, 0x50, 0xf4, 0x1c, 0x40, 0x2b, 0x0c, 0x76, - 0x3d, 0x97, 0x05, 0x81, 0x28, 0x9a, 0x66, 0x3e, 0x35, 0x05, 0xc1, 0x1a, 0x16, 0x7a, 0x15, 0xc6, - 0xda, 0x7e, 0xc4, 0xd9, 0x11, 0x67, 0x5d, 0x04, 0xe5, 0x2e, 0x25, 0x06, 0x28, 0xb7, 0x75, 0x20, - 0x36, 0x71, 0xd1, 0x3c, 0x0c, 0xc5, 0x0e, 0x33, 0x5b, 0x19, 0xcc, 0xb7, 0xb7, 0x5d, 0xa3, 0x18, - 0x7a, 0x5a, 0x0e, 0x5a, 0x01, 0x8b, 0x8a, 0xe8, 0x2d, 0xe9, 0x2a, 0xcb, 0x0f, 0x76, 0x61, 0xe8, - 0xde, 0xdf, 0x25, 0xa0, 0x39, 0xca, 0x0a, 0x03, 0x7a, 0x83, 0x16, 0x7a, 0x05, 0x80, 0xdc, 0x8b, - 0x49, 0xe8, 0x3b, 0x4d, 0x65, 0x15, 0xa6, 0xf8, 0x82, 0x4a, 0xb0, 0x1a, 0xc4, 0xb7, 0x23, 0xb2, - 0xa4, 0x30, 0xb0, 0x86, 0x6d, 0xff, 0x66, 0x19, 0x20, 0xe1, 0xdb, 0xd1, 0x7b, 0x1d, 0x07, 0xd7, - 0xd3, 0xdd, 0x39, 0xfd, 0xe3, 0x3b, 0xb5, 0xd0, 0xf7, 0x58, 0x30, 0xe2, 0x34, 0x9b, 0x41, 0xc3, - 0x89, 0xd9, 0x0c, 0x15, 0xba, 0x1f, 0x9c, 0xa2, 0xfd, 0xf9, 0xa4, 0x06, 0xef, 0xc2, 0xf3, 0x72, - 0x85, 0x6a, 0x90, 0x9e, 0xbd, 0xd0, 0x1b, 0x46, 0x9f, 0x90, 0x4f, 0xc5, 0xa2, 0x31, 0x94, 0xea, - 0xa9, 0x58, 0x66, 0x77, 0x84, 0xfe, 0x4a, 0xbc, 0x6d, 0xbc, 0x12, 0x07, 0xf2, 0x7d, 0x01, 0x0d, - 0xf6, 0xb5, 0xd7, 0x03, 0x11, 0xd5, 0xf4, 0xb8, 0x00, 0x83, 0xf9, 0x8e, 0x77, 0xda, 0x3b, 0xa9, - 0x47, 0x4c, 0x80, 0x77, 0x60, 0xc2, 0x35, 0x99, 0x00, 0xb1, 0x12, 0x9f, 0xc8, 0xa3, 0x9b, 0xe2, - 0x19, 0x92, 0x6b, 0x3f, 0x05, 0xc0, 0x69, 0xc2, 0xa8, 0xc6, 0x63, 0x3e, 0x54, 0xfd, 0x8d, 0x40, - 0x38, 0x5b, 0xd8, 0xb9, 0x73, 0xb9, 0x1f, 0xc5, 0x64, 0x87, 0x62, 0x26, 0xb7, 0xfb, 0xaa, 0xa8, - 0x8b, 0x15, 0x15, 0xf4, 0x06, 0x0c, 0x31, 0xcf, 0xab, 0x68, 0xba, 0x94, 0x2f, 0x2b, 0x36, 0x83, - 0x98, 0x25, 0x1b, 0x92, 0xfd, 0x8d, 0xb0, 0xa0, 0x80, 0xae, 0x4b, 0xbf, 0xc6, 0xa8, 0xea, 0xdf, - 0x8e, 0x08, 0xf3, 0x6b, 0x2c, 0x2f, 0x7c, 0x34, 0x71, 0x59, 0xe4, 0xe5, 0x99, 0xc9, 0xbb, 0x8c, - 0x9a, 0x94, 0x8b, 0x12, 0xff, 0x65, 0x4e, 0xb0, 0x69, 0xc8, 0xef, 0x9e, 0x99, 0x37, 0x2c, 0x19, - 0xce, 0x3b, 0x26, 0x09, 0x9c, 0xa6, 0x49, 0x39, 0x52, 0xbe, 0xeb, 0x85, 0xbb, 0x46, 0xaf, 0xb3, - 0x83, 0x3f, 0xc4, 0xd9, 0x6d, 0xc4, 0x4b, 0xb0, 0xa8, 0x7f, 0xa2, 0xec, 0xc1, 0x8c, 0x0f, 0x93, - 0xe9, 0x2d, 0xfa, 0x50, 0xd9, 0x91, 0xdf, 0x1f, 0x80, 0x71, 0x73, 0x49, 0xa1, 0xab, 0x50, 0x16, - 0x44, 0x54, 0x1c, 0x7f, 0xb5, 0x4b, 0x56, 0x24, 0x00, 0x27, 0x38, 0x2c, 0x7d, 0x03, 0xab, 0xae, - 0x99, 0xd9, 0x26, 0xe9, 0x1b, 0x14, 0x04, 0x6b, 0x58, 0xf4, 0x61, 0xb5, 0x1e, 0x04, 0xb1, 0xba, - 0x90, 0xd4, 0xba, 0x5b, 0x60, 0xa5, 0x58, 0x40, 0xe9, 0x45, 0xb4, 0x4d, 0x42, 0x9f, 0x34, 0xcd, - 0xf0, 0xc0, 0xea, 0x22, 0xba, 0xa1, 0x03, 0xb1, 0x89, 0x4b, 0xaf, 0xd3, 0x20, 0x62, 0x0b, 0x59, - 0x3c, 0xdf, 0x12, 0xb3, 0xe5, 0x3a, 0x77, 0xad, 0x96, 0x70, 0xf4, 0x19, 0x78, 0x44, 0x85, 0x40, - 0xc2, 0x5c, 0x0f, 0x21, 0x5b, 0x1c, 0x32, 0xa4, 0x2d, 0x8f, 0x2c, 0x66, 0xa3, 0xe1, 0xbc, 0xfa, - 0xe8, 0x75, 0x18, 0x17, 0x2c, 0xbe, 0xa4, 0x38, 0x6c, 0x9a, 0xc6, 0xdc, 0x30, 0xa0, 0x38, 0x85, - 0x2d, 0x03, 0x1c, 0x33, 0x2e, 0x5b, 0x52, 0x28, 0x75, 0x06, 0x38, 0xd6, 0xe1, 0xb8, 0xa3, 0x06, - 0x9a, 0x87, 0x09, 0xce, 0x83, 0x79, 0xfe, 0x26, 0x9f, 0x13, 0xe1, 0x4d, 0xa5, 0xb6, 0xd4, 0x2d, - 0x13, 0x8c, 0xd3, 0xf8, 0xe8, 0x65, 0x18, 0x75, 0xc2, 0xc6, 0x96, 0x17, 0x93, 0x46, 0xdc, 0x0e, - 0xb9, 0x9b, 0x95, 0x66, 0x5b, 0x34, 0xaf, 0xc1, 0xb0, 0x81, 0x69, 0xbf, 0x07, 0xa7, 0x32, 0x62, - 0x2e, 0xd0, 0x85, 0xe3, 0xb4, 0x3c, 0xf9, 0x4d, 0x29, 0x03, 0xe4, 0xf9, 0x5a, 0x55, 0x7e, 0x8d, - 0x86, 0x45, 0x57, 0x27, 0x8b, 0xcd, 0xa0, 0xa5, 0x00, 0x54, 0xab, 0x73, 0x59, 0x02, 0x70, 0x82, - 0x63, 0xff, 0xf7, 0x02, 0x4c, 0x64, 0xe8, 0x56, 0x58, 0x1a, 0xba, 0xd4, 0x23, 0x25, 0xc9, 0x3a, - 0x67, 0xc6, 0xcb, 0x2e, 0x1c, 0x21, 0x5e, 0x76, 0xb1, 0x57, 0xbc, 0xec, 0x81, 0xf7, 0x13, 0x2f, - 0xdb, 0x1c, 0xb1, 0xc1, 0xbe, 0x46, 0x2c, 0x23, 0xc6, 0xf6, 0xd0, 0x11, 0x63, 0x6c, 0x1b, 0x83, - 0x3e, 0xdc, 0xc7, 0xa0, 0xff, 0x50, 0x01, 0x26, 0xd3, 0x36, 0x90, 0x27, 0x20, 0xb7, 0x7d, 0xc3, - 0x90, 0xdb, 0x66, 0x27, 0x75, 0x4c, 0x5b, 0x66, 0xe6, 0xc9, 0x70, 0x71, 0x4a, 0x86, 0xfb, 0xf1, - 0xbe, 0xa8, 0x75, 0x97, 0xe7, 0xfe, 0xad, 0x02, 0x9c, 0x49, 0x57, 0x59, 0x6c, 0x3a, 0xde, 0xce, - 0x09, 0x8c, 0xcd, 0x2d, 0x63, 0x6c, 0x9e, 0xe9, 0xe7, 0x6b, 0x58, 0xd7, 0x72, 0x07, 0xe8, 0x6e, - 0x6a, 0x80, 0xae, 0xf6, 0x4f, 0xb2, 0xfb, 0x28, 0x7d, 0xb5, 0x08, 0x17, 0x33, 0xeb, 0x25, 0x62, - 0xcf, 0x65, 0x43, 0xec, 0xf9, 0x5c, 0x4a, 0xec, 0x69, 0x77, 0xaf, 0x7d, 0x3c, 0x72, 0x50, 0xe1, - 0x21, 0xcb, 0x02, 0x08, 0x3c, 0xa0, 0x0c, 0xd4, 0xf0, 0x90, 0x55, 0x84, 0xb0, 0x49, 0xf7, 0x1b, - 0x49, 0xf6, 0xf9, 0xaf, 0x2c, 0x38, 0x97, 0x39, 0x37, 0x27, 0x20, 0xeb, 0x5a, 0x35, 0x65, 0x5d, - 0x4f, 0xf6, 0xbd, 0x5a, 0x73, 0x84, 0x5f, 0xbf, 0x3e, 0x90, 0xf3, 0x2d, 0xec, 0x25, 0x7f, 0x0b, - 0x46, 0x9c, 0x46, 0x83, 0x44, 0xd1, 0x4a, 0xe0, 0xaa, 0x90, 0xc0, 0xcf, 0xb0, 0x77, 0x56, 0x52, - 0x7c, 0x78, 0x30, 0x3b, 0x93, 0x26, 0x91, 0x80, 0xb1, 0x4e, 0x01, 0x7d, 0x16, 0x4a, 0x91, 0xb8, - 0x37, 0xc5, 0xdc, 0x3f, 0xdf, 0xe7, 0xe0, 0x38, 0xeb, 0xa4, 0x69, 0x86, 0x39, 0x52, 0x92, 0x0a, - 0x45, 0xd2, 0x0c, 0x89, 0x52, 0x38, 0xd6, 0x90, 0x28, 0xcf, 0x01, 0xec, 0xaa, 0xc7, 0x40, 0x5a, - 0xfe, 0xa0, 0x3d, 0x13, 0x34, 0x2c, 0xf4, 0x2d, 0x30, 0x19, 0xf1, 0xa0, 0x7e, 0x8b, 0x4d, 0x27, - 0x62, 0x6e, 0x2e, 0x62, 0x15, 0xb2, 0x50, 0x4a, 0xf5, 0x14, 0x0c, 0x77, 0x60, 0xa3, 0x65, 0xd9, - 0x2a, 0x8b, 0x40, 0xc8, 0x17, 0xe6, 0xe5, 0xa4, 0x45, 0x91, 0x04, 0xf7, 0x74, 0x7a, 0xf8, 0xd9, - 0xc0, 0x6b, 0x35, 0xd1, 0x67, 0x01, 0xe8, 0xf2, 0x11, 0x72, 0x88, 0xe1, 0xfc, 0xc3, 0x93, 0x9e, - 0x2a, 0x6e, 0xa6, 0x55, 0x2e, 0xf3, 0x4d, 0xad, 0x28, 0x22, 0x58, 0x23, 0x68, 0xff, 0xd0, 0x00, - 0x3c, 0xda, 0xe5, 0x8c, 0x44, 0xf3, 0xa6, 0x1e, 0xf6, 0xa9, 0xf4, 0xe3, 0x7a, 0x26, 0xb3, 0xb2, - 0xf1, 0xda, 0x4e, 0x2d, 0xc5, 0xc2, 0xfb, 0x5e, 0x8a, 0xdf, 0x6f, 0x69, 0x62, 0x0f, 0x6e, 0xab, - 0xf9, 0xa9, 0x23, 0x9e, 0xfd, 0xc7, 0x28, 0x07, 0xd9, 0xc8, 0x10, 0x26, 0x3c, 0xd7, 0x77, 0x77, - 0xfa, 0x96, 0x2e, 0x9c, 0xac, 0x94, 0xf8, 0x0b, 0x16, 0x3c, 0x96, 0xd9, 0x5f, 0xc3, 0x22, 0xe7, - 0x2a, 0x94, 0x1b, 0xb4, 0x50, 0x73, 0x45, 0x4c, 0x7c, 0xb4, 0x25, 0x00, 0x27, 0x38, 0x86, 0xe1, - 0x4d, 0xa1, 0xa7, 0xe1, 0xcd, 0x3f, 0xb5, 0xa0, 0x63, 0x7f, 0x9c, 0xc0, 0x41, 0x5d, 0x35, 0x0f, - 0xea, 0x8f, 0xf6, 0x33, 0x97, 0x39, 0x67, 0xf4, 0x1f, 0x4d, 0xc0, 0xd9, 0x1c, 0x57, 0x9c, 0x5d, - 0x98, 0xda, 0x6c, 0x10, 0xd3, 0xc9, 0x53, 0x7c, 0x4c, 0xa6, 0x3f, 0x6c, 0x57, 0x8f, 0x50, 0x96, - 0xd1, 0x72, 0xaa, 0x03, 0x05, 0x77, 0x36, 0x81, 0xbe, 0x60, 0xc1, 0x69, 0x67, 0x2f, 0xea, 0x48, - 0x81, 0x2f, 0xd6, 0xcc, 0x0b, 0x99, 0x42, 0x90, 0x1e, 0x29, 0xf3, 0x79, 0x8a, 0xcf, 0x2c, 0x2c, - 0x9c, 0xd9, 0x16, 0xc2, 0x22, 0x48, 0x3c, 0x65, 0xe7, 0xbb, 0xb8, 0x21, 0x67, 0xf9, 0x4c, 0xf1, - 0x1b, 0x44, 0x42, 0xb0, 0xa2, 0x83, 0x3e, 0x0f, 0xe5, 0x4d, 0xe9, 0xc8, 0x98, 0x71, 0x43, 0x25, - 0x03, 0xd9, 0xdd, 0xbd, 0x93, 0x6b, 0x32, 0x15, 0x12, 0x4e, 0x88, 0xa2, 0xd7, 0xa1, 0xe8, 0x6f, - 0x44, 0xdd, 0xb2, 0x64, 0xa6, 0x4c, 0xd6, 0xb8, 0xb3, 0xff, 0xea, 0x72, 0x1d, 0xd3, 0x8a, 0xe8, - 0x3a, 0x14, 0xc3, 0x75, 0x57, 0x48, 0xf0, 0x32, 0xcf, 0x70, 0xbc, 0x50, 0xc9, 0xe9, 0x15, 0xa3, - 0x84, 0x17, 0x2a, 0x98, 0x92, 0x40, 0x35, 0x18, 0x64, 0xfe, 0x2b, 0xe2, 0x3e, 0xc8, 0xe4, 0x7c, - 0xbb, 0xf8, 0x81, 0xf1, 0x88, 0x00, 0x0c, 0x01, 0x73, 0x42, 0x68, 0x0d, 0x86, 0x1a, 0x2c, 0xa3, - 0xa2, 0x88, 0x47, 0xf6, 0x89, 0x4c, 0x59, 0x5d, 0x97, 0x54, 0x93, 0x42, 0x74, 0xc5, 0x30, 0xb0, - 0xa0, 0xc5, 0xa8, 0x92, 0xd6, 0xd6, 0x46, 0x24, 0x32, 0x00, 0x67, 0x53, 0xed, 0x92, 0x41, 0x55, - 0x50, 0x65, 0x18, 0x58, 0xd0, 0x42, 0xaf, 0x40, 0x61, 0xa3, 0x21, 0x7c, 0x53, 0x32, 0x85, 0x76, - 0x66, 0xbc, 0x86, 0x85, 0xa1, 0xfb, 0x07, 0xb3, 0x85, 0xe5, 0x45, 0x5c, 0xd8, 0x68, 0xa0, 0x55, - 0x18, 0xde, 0xe0, 0x1e, 0xde, 0x42, 0x2e, 0xf7, 0x44, 0xb6, 0xf3, 0x79, 0x87, 0x13, 0x38, 0x77, - 0xcb, 0x10, 0x00, 0x2c, 0x89, 0xb0, 0x98, 0xeb, 0xca, 0x53, 0x5d, 0x84, 0xee, 0x9a, 0x3b, 0x5a, - 0x74, 0x01, 0x7e, 0x3f, 0x27, 0xfe, 0xee, 0x58, 0xa3, 0x48, 0x57, 0xb5, 0x23, 0xd3, 0xb0, 0x8b, - 0x50, 0x2c, 0x99, 0xab, 0xba, 0x47, 0x86, 0x7a, 0xbe, 0xaa, 0x15, 0x12, 0x4e, 0x88, 0xa2, 0x6d, - 0x18, 0xdb, 0x8d, 0x5a, 0x5b, 0x44, 0x6e, 0x69, 0x16, 0x99, 0x25, 0xe7, 0x0a, 0xbb, 0x23, 0x10, - 0xbd, 0x30, 0x6e, 0x3b, 0xcd, 0x8e, 0x53, 0x88, 0xa9, 0xbf, 0xef, 0xe8, 0xc4, 0xb0, 0x49, 0x9b, - 0x0e, 0xff, 0xbb, 0xed, 0x60, 0x7d, 0x3f, 0x26, 0x22, 0xe2, 0x56, 0xe6, 0xf0, 0xbf, 0xc9, 0x51, - 0x3a, 0x87, 0x5f, 0x00, 0xb0, 0x24, 0x82, 0xee, 0x88, 0xe1, 0x61, 0xa7, 0xe7, 0x64, 0x7e, 0x58, - 0xcc, 0x79, 0x89, 0x94, 0x33, 0x28, 0xec, 0xb4, 0x4c, 0x48, 0xb1, 0x53, 0xb2, 0xb5, 0x15, 0xc4, - 0x81, 0x9f, 0x3a, 0xa1, 0xa7, 0xf2, 0x4f, 0xc9, 0x5a, 0x06, 0x7e, 0xe7, 0x29, 0x99, 0x85, 0x85, - 0x33, 0xdb, 0x42, 0x2e, 0x8c, 0xb7, 0x82, 0x30, 0xde, 0x0b, 0x42, 0xb9, 0xbe, 0x50, 0x17, 0xb9, - 0x82, 0x81, 0x29, 0x5a, 0x64, 0xc1, 0xec, 0x4c, 0x08, 0x4e, 0xd1, 0x44, 0x9f, 0x86, 0xe1, 0xa8, - 0xe1, 0x34, 0x49, 0xf5, 0xd6, 0xf4, 0xa9, 0xfc, 0xeb, 0xa7, 0xce, 0x51, 0x72, 0x56, 0x17, 0x0f, - 0xd0, 0xce, 0x51, 0xb0, 0x24, 0x87, 0x96, 0x61, 0x90, 0xe5, 0xd4, 0x62, 0xe1, 0xe1, 0x72, 0xa2, - 0x7b, 0x76, 0x18, 0x10, 0xf3, 0xb3, 0x89, 0x15, 0x63, 0x5e, 0x9d, 0xee, 0x01, 0xc1, 0x5e, 0x07, - 0xd1, 0xf4, 0x99, 0xfc, 0x3d, 0x20, 0xb8, 0xf2, 0x5b, 0xf5, 0x6e, 0x7b, 0x40, 0x21, 0xe1, 0x84, - 0x28, 0x3d, 0x99, 0xe9, 0x69, 0x7a, 0xb6, 0x8b, 0xe5, 0x4b, 0xee, 0x59, 0xca, 0x4e, 0x66, 0x7a, - 0x92, 0x52, 0x12, 0xf6, 0xef, 0x0e, 0x77, 0xf2, 0x2c, 0xec, 0x41, 0xf6, 0x5d, 0x56, 0x87, 0xae, - 0xee, 0x93, 0xfd, 0xca, 0x87, 0x8e, 0x91, 0x5b, 0xfd, 0x82, 0x05, 0x67, 0x5b, 0x99, 0x1f, 0x22, - 0x18, 0x80, 0xfe, 0xc4, 0x4c, 0xfc, 0xd3, 0x55, 0x28, 0xc1, 0x6c, 0x38, 0xce, 0x69, 0x29, 0xfd, - 0x22, 0x28, 0xbe, 0xef, 0x17, 0xc1, 0x0a, 0x94, 0x18, 0x93, 0xd9, 0x23, 0xc3, 0x70, 0xfa, 0x61, - 0xc4, 0x58, 0x89, 0x45, 0x51, 0x11, 0x2b, 0x12, 0xe8, 0x07, 0x2c, 0xb8, 0x90, 0xee, 0x3a, 0x26, - 0x0c, 0x2c, 0xe2, 0x0f, 0xf2, 0xb7, 0xe0, 0xb2, 0xf8, 0xfe, 0x0b, 0xb5, 0x6e, 0xc8, 0x87, 0xbd, - 0x10, 0x70, 0xf7, 0xc6, 0x50, 0x25, 0xe3, 0x31, 0x3a, 0x64, 0x0a, 0xe0, 0xfb, 0x78, 0x90, 0xbe, - 0x00, 0xa3, 0x3b, 0x41, 0xdb, 0x8f, 0x85, 0xa1, 0x8c, 0x50, 0xda, 0x33, 0x65, 0xf5, 0x8a, 0x56, - 0x8e, 0x0d, 0xac, 0xd4, 0x33, 0xb6, 0xf4, 0xc0, 0xcf, 0xd8, 0xb7, 0x61, 0xd4, 0xd7, 0x2c, 0x3b, - 0x05, 0x3f, 0x70, 0x39, 0x3f, 0x76, 0xa8, 0x6e, 0x07, 0xca, 0x7b, 0xa9, 0x97, 0x60, 0x83, 0xda, - 0xc9, 0xbe, 0x8d, 0x7e, 0xda, 0xca, 0x60, 0xea, 0xf9, 0x6b, 0xf9, 0x35, 0xf3, 0xb5, 0x7c, 0x39, - 0xfd, 0x5a, 0xee, 0x10, 0xbe, 0x1a, 0x0f, 0xe5, 0xfe, 0xf3, 0x9c, 0xf4, 0x1b, 0x26, 0xd0, 0x6e, - 0xc2, 0xa5, 0x5e, 0xd7, 0x12, 0xb3, 0x98, 0x72, 0x95, 0xaa, 0x2d, 0xb1, 0x98, 0x72, 0xab, 0x15, - 0xcc, 0x20, 0xfd, 0xc6, 0x91, 0xb1, 0xff, 0x9b, 0x05, 0xc5, 0x5a, 0xe0, 0x9e, 0x80, 0x30, 0xf9, - 0x53, 0x86, 0x30, 0xf9, 0xd1, 0xec, 0x0b, 0xd1, 0xcd, 0x15, 0x1d, 0x2f, 0xa5, 0x44, 0xc7, 0x17, - 0xf2, 0x08, 0x74, 0x17, 0x14, 0xff, 0x44, 0x11, 0x46, 0x6a, 0x81, 0xab, 0xcc, 0x95, 0x7f, 0xfd, - 0x41, 0xcc, 0x95, 0x73, 0x03, 0xfc, 0x6b, 0x94, 0x99, 0xa1, 0x95, 0xf4, 0xb1, 0xfc, 0x0b, 0x66, - 0xb5, 0x7c, 0x97, 0x78, 0x9b, 0x5b, 0x31, 0x71, 0xd3, 0x9f, 0x73, 0x72, 0x56, 0xcb, 0xff, 0xd5, - 0x82, 0x89, 0x54, 0xeb, 0xa8, 0x09, 0x63, 0x4d, 0x5d, 0x30, 0x29, 0xd6, 0xe9, 0x03, 0xc9, 0x34, - 0x85, 0xd5, 0xa7, 0x56, 0x84, 0x4d, 0xe2, 0x68, 0x0e, 0x40, 0x69, 0xea, 0xa4, 0x04, 0x8c, 0x71, - 0xfd, 0x4a, 0x95, 0x17, 0x61, 0x0d, 0x03, 0xbd, 0x08, 0x23, 0x71, 0xd0, 0x0a, 0x9a, 0xc1, 0xe6, - 0xfe, 0x0d, 0x22, 0x23, 0x17, 0x29, 0x5b, 0xae, 0xb5, 0x04, 0x84, 0x75, 0x3c, 0xfb, 0xa7, 0x8a, - 0xfc, 0x43, 0xfd, 0xd8, 0xfb, 0x70, 0x4d, 0x7e, 0xb0, 0xd7, 0xe4, 0x57, 0x2d, 0x98, 0xa4, 0xad, - 0x33, 0x73, 0x11, 0x79, 0xd9, 0xaa, 0x98, 0xc1, 0x56, 0x97, 0x98, 0xc1, 0x97, 0xe9, 0xd9, 0xe5, - 0x06, 0xed, 0x58, 0x48, 0xd0, 0xb4, 0xc3, 0x89, 0x96, 0x62, 0x01, 0x15, 0x78, 0x24, 0x0c, 0x85, - 0x8b, 0x9b, 0x8e, 0x47, 0xc2, 0x10, 0x0b, 0xa8, 0x0c, 0x29, 0x3c, 0x90, 0x1d, 0x52, 0x98, 0xc7, - 0x61, 0x14, 0x86, 0x05, 0x82, 0xed, 0xd1, 0xe2, 0x30, 0x4a, 0x8b, 0x83, 0x04, 0xc7, 0xfe, 0xf9, - 0x22, 0x8c, 0xd6, 0x02, 0x37, 0xd1, 0x95, 0xbd, 0x60, 0xe8, 0xca, 0x2e, 0xa5, 0x74, 0x65, 0x93, - 0x3a, 0xee, 0x87, 0x9a, 0xb1, 0xaf, 0x97, 0x66, 0xec, 0x9f, 0x58, 0x6c, 0xd6, 0x2a, 0xab, 0x75, - 0x6e, 0x7d, 0x84, 0x9e, 0x85, 0x11, 0x76, 0x20, 0x31, 0x9f, 0x4a, 0xa9, 0x40, 0x62, 0x29, 0x94, - 0x56, 0x93, 0x62, 0xac, 0xe3, 0xa0, 0x2b, 0x50, 0x8a, 0x88, 0x13, 0x36, 0xb6, 0xd4, 0x19, 0x27, - 0xb4, 0x3d, 0xbc, 0x0c, 0x2b, 0x28, 0x7a, 0x33, 0x09, 0x01, 0x58, 0xcc, 0xf7, 0xd1, 0xd2, 0xfb, - 0xc3, 0xb7, 0x48, 0x7e, 0xdc, 0x3f, 0xfb, 0x2e, 0xa0, 0x4e, 0xfc, 0x3e, 0x62, 0x5f, 0xcd, 0x9a, - 0xb1, 0xaf, 0xca, 0x1d, 0x71, 0xaf, 0xfe, 0xcc, 0x82, 0xf1, 0x5a, 0xe0, 0xd2, 0xad, 0xfb, 0x8d, - 0xb4, 0x4f, 0xf5, 0xf8, 0xa7, 0x43, 0x5d, 0xe2, 0x9f, 0x3e, 0x0e, 0x83, 0xb5, 0xc0, 0xad, 0xd6, - 0xba, 0xf9, 0x36, 0xdb, 0x7f, 0xdb, 0x82, 0xe1, 0x5a, 0xe0, 0x9e, 0x80, 0x70, 0xfe, 0x35, 0x53, - 0x38, 0xff, 0x48, 0xce, 0xba, 0xc9, 0x91, 0xc7, 0xff, 0xcd, 0x01, 0x18, 0xa3, 0xfd, 0x0c, 0x36, - 0xe5, 0x54, 0x1a, 0xc3, 0x66, 0xf5, 0x31, 0x6c, 0x94, 0x17, 0x0e, 0x9a, 0xcd, 0x60, 0x2f, 0x3d, - 0xad, 0xcb, 0xac, 0x14, 0x0b, 0x28, 0x7a, 0x1a, 0x4a, 0xad, 0x90, 0xec, 0x7a, 0x81, 0x60, 0x32, - 0x35, 0x55, 0x47, 0x4d, 0x94, 0x63, 0x85, 0x41, 0x1f, 0x67, 0x91, 0xe7, 0x37, 0x48, 0x9d, 0x34, - 0x02, 0xdf, 0xe5, 0xf2, 0xeb, 0xa2, 0x48, 0x1b, 0xa0, 0x95, 0x63, 0x03, 0x0b, 0xdd, 0x85, 0x32, - 0xfb, 0xcf, 0x8e, 0x9d, 0xa3, 0x67, 0x93, 0x14, 0xd9, 0xc5, 0x04, 0x01, 0x9c, 0xd0, 0x42, 0xcf, - 0x01, 0xc4, 0x32, 0x42, 0x76, 0x24, 0xe2, 0x1c, 0x29, 0x86, 0x5c, 0xc5, 0xce, 0x8e, 0xb0, 0x86, - 0x85, 0x9e, 0x82, 0x72, 0xec, 0x78, 0xcd, 0x9b, 0x9e, 0x4f, 0x22, 0x26, 0x97, 0x2e, 0xca, 0x24, - 0x5f, 0xa2, 0x10, 0x27, 0x70, 0xca, 0x10, 0xb1, 0x20, 0x00, 0x3c, 0x17, 0x6d, 0x89, 0x61, 0x33, - 0x86, 0xe8, 0xa6, 0x2a, 0xc5, 0x1a, 0x06, 0xda, 0x82, 0xf3, 0x9e, 0xcf, 0x42, 0xec, 0x93, 0xfa, - 0xb6, 0xd7, 0x5a, 0xbb, 0x59, 0xbf, 0x43, 0x42, 0x6f, 0x63, 0x7f, 0xc1, 0x69, 0x6c, 0x13, 0x5f, - 0xe6, 0x09, 0xfc, 0xa8, 0xe8, 0xe2, 0xf9, 0x6a, 0x17, 0x5c, 0xdc, 0x95, 0x92, 0xfd, 0x32, 0x9c, - 0xa9, 0x05, 0x6e, 0x2d, 0x08, 0xe3, 0xe5, 0x20, 0xdc, 0x73, 0x42, 0x57, 0xae, 0x94, 0x59, 0x99, - 0x85, 0x84, 0x1e, 0x85, 0x83, 0xfc, 0xa0, 0x30, 0x72, 0x61, 0x3d, 0xcf, 0x98, 0xaf, 0x23, 0x3a, - 0xa3, 0x34, 0x18, 0x1b, 0xa0, 0xf2, 0x4d, 0x5c, 0x73, 0x62, 0x82, 0x6e, 0xb1, 0xa4, 0xb8, 0xc9, - 0x8d, 0x28, 0xaa, 0x3f, 0xa9, 0x25, 0xc5, 0x4d, 0x80, 0x99, 0x57, 0xa8, 0x59, 0xdf, 0xfe, 0xd9, - 0x01, 0x76, 0x38, 0xa6, 0x72, 0x16, 0xa0, 0xcf, 0xc1, 0x78, 0x44, 0x6e, 0x7a, 0x7e, 0xfb, 0x9e, - 0x94, 0x09, 0x74, 0x71, 0x27, 0xaa, 0x2f, 0xe9, 0x98, 0x5c, 0xb2, 0x68, 0x96, 0xe1, 0x14, 0x35, - 0xb4, 0x03, 0xe3, 0x7b, 0x9e, 0xef, 0x06, 0x7b, 0x91, 0xa4, 0x5f, 0xca, 0x17, 0x30, 0xde, 0xe5, - 0x98, 0xa9, 0x3e, 0x1a, 0xcd, 0xdd, 0x35, 0x88, 0xe1, 0x14, 0x71, 0xba, 0x00, 0xc3, 0xb6, 0x3f, - 0x1f, 0xdd, 0x8e, 0x48, 0x28, 0xd2, 0x1b, 0xb3, 0x05, 0x88, 0x65, 0x21, 0x4e, 0xe0, 0x74, 0x01, - 0xb2, 0x3f, 0xd7, 0xc2, 0xa0, 0xcd, 0xe3, 0xd8, 0x8b, 0x05, 0x88, 0x55, 0x29, 0xd6, 0x30, 0xe8, - 0x06, 0x65, 0xff, 0x56, 0x03, 0x1f, 0x07, 0x41, 0x2c, 0xb7, 0x34, 0x4b, 0xa8, 0xa9, 0x95, 0x63, - 0x03, 0x0b, 0x2d, 0x03, 0x8a, 0xda, 0xad, 0x56, 0x93, 0xd9, 0x29, 0x38, 0x4d, 0x46, 0x8a, 0xeb, - 0x88, 0x8b, 0x3c, 0x4a, 0x67, 0xbd, 0x03, 0x8a, 0x33, 0x6a, 0xd0, 0xb3, 0x7a, 0x43, 0x74, 0x75, - 0x90, 0x75, 0x95, 0x2b, 0x23, 0xea, 0xbc, 0x9f, 0x12, 0x86, 0x96, 0x60, 0x38, 0xda, 0x8f, 0x1a, - 0xb1, 0x08, 0x37, 0x96, 0x93, 0x96, 0xa6, 0xce, 0x50, 0xb4, 0xac, 0x68, 0xbc, 0x0a, 0x96, 0x75, - 0xed, 0x6f, 0x67, 0xac, 0x00, 0x4b, 0x86, 0x1b, 0xb7, 0x43, 0x82, 0x76, 0x60, 0xac, 0xc5, 0x56, - 0x98, 0x08, 0xcc, 0x2e, 0x96, 0xc9, 0x0b, 0x7d, 0xbe, 0xe9, 0xf7, 0xe8, 0x09, 0xaa, 0x64, 0x6e, - 0xec, 0xb1, 0x54, 0xd3, 0xc9, 0x61, 0x93, 0xba, 0xfd, 0xd5, 0xb3, 0xec, 0x32, 0xa9, 0xf3, 0x87, - 0xfa, 0xb0, 0x30, 0xac, 0x16, 0xaf, 0x92, 0x99, 0x7c, 0x89, 0x51, 0xf2, 0x45, 0xc2, 0x38, 0x1b, - 0xcb, 0xba, 0xe8, 0xb3, 0x30, 0x4e, 0x99, 0x7c, 0x2d, 0x31, 0xc5, 0xe9, 0x7c, 0x07, 0xf8, 0x24, - 0x1f, 0x85, 0x96, 0xb4, 0x41, 0xaf, 0x8c, 0x53, 0xc4, 0xd0, 0x9b, 0xcc, 0x04, 0xc0, 0xcc, 0x79, - 0xd1, 0x83, 0xb4, 0xae, 0xed, 0x97, 0x64, 0x35, 0x22, 0x79, 0xf9, 0x34, 0xec, 0x87, 0x9b, 0x4f, - 0x03, 0xdd, 0x84, 0x31, 0x91, 0x11, 0x56, 0x08, 0x3a, 0x8b, 0x86, 0x20, 0x6b, 0x0c, 0xeb, 0xc0, - 0xc3, 0x74, 0x01, 0x36, 0x2b, 0xa3, 0x4d, 0xb8, 0xa0, 0x25, 0x75, 0xb9, 0x16, 0x3a, 0x4c, 0x1b, - 0xed, 0xb1, 0x93, 0x48, 0xbb, 0xe6, 0x1e, 0xbb, 0x7f, 0x30, 0x7b, 0x61, 0xad, 0x1b, 0x22, 0xee, - 0x4e, 0x07, 0xdd, 0x82, 0x33, 0xdc, 0x7d, 0xb3, 0x42, 0x1c, 0xb7, 0xe9, 0xf9, 0xea, 0x1e, 0xe5, - 0xbb, 0xe5, 0xdc, 0xfd, 0x83, 0xd9, 0x33, 0xf3, 0x59, 0x08, 0x38, 0xbb, 0x1e, 0x7a, 0x0d, 0xca, - 0xae, 0x1f, 0x89, 0x31, 0x18, 0x32, 0xf2, 0xe6, 0x94, 0x2b, 0xab, 0x75, 0xf5, 0xfd, 0xc9, 0x1f, - 0x9c, 0x54, 0x40, 0x9b, 0x5c, 0xd8, 0xa9, 0x64, 0x0b, 0xc3, 0x1d, 0x81, 0x67, 0xd2, 0x52, 0x2a, - 0xc3, 0x81, 0x8b, 0x4b, 0xf9, 0x95, 0x5d, 0xb3, 0xe1, 0xdb, 0x65, 0x10, 0x46, 0x6f, 0x00, 0xa2, - 0xcc, 0xb7, 0xd7, 0x20, 0xf3, 0x0d, 0x16, 0xf5, 0x9f, 0xc9, 0x86, 0x4b, 0xa6, 0x4b, 0x51, 0xbd, - 0x03, 0x03, 0x67, 0xd4, 0x42, 0xd7, 0xe9, 0x6d, 0xa0, 0x97, 0x0a, 0xfb, 0x6c, 0x95, 0xe5, 0xac, - 0x42, 0x5a, 0x21, 0x69, 0x38, 0x31, 0x71, 0x4d, 0x8a, 0x38, 0x55, 0x0f, 0xb9, 0x70, 0xde, 0x69, - 0xc7, 0x01, 0x93, 0x23, 0x9b, 0xa8, 0x6b, 0xc1, 0x36, 0xf1, 0x99, 0x0a, 0xa7, 0xb4, 0x70, 0x89, - 0x5e, 0xd4, 0xf3, 0x5d, 0xf0, 0x70, 0x57, 0x2a, 0x94, 0xc1, 0x52, 0x39, 0x4a, 0xc1, 0x8c, 0xa7, - 0x93, 0x91, 0xa7, 0xf4, 0x45, 0x18, 0xd9, 0x0a, 0xa2, 0x78, 0x95, 0xc4, 0x7b, 0x41, 0xb8, 0x2d, - 0xa2, 0x22, 0x26, 0x91, 0x74, 0x13, 0x10, 0xd6, 0xf1, 0xe8, 0x0b, 0x8a, 0x19, 0x18, 0x54, 0x2b, - 0x4c, 0xb7, 0x5b, 0x4a, 0xce, 0x98, 0xeb, 0xbc, 0x18, 0x4b, 0xb8, 0x44, 0xad, 0xd6, 0x16, 0x99, - 0x9e, 0x36, 0x85, 0x5a, 0xad, 0x2d, 0x62, 0x09, 0xa7, 0xcb, 0x35, 0xda, 0x72, 0x42, 0x52, 0x0b, - 0x83, 0x06, 0x89, 0xb4, 0xf8, 0xcd, 0x8f, 0xf2, 0x98, 0x8f, 0x74, 0xb9, 0xd6, 0xb3, 0x10, 0x70, - 0x76, 0x3d, 0x44, 0x3a, 0x13, 0x1a, 0x8d, 0xe7, 0x0b, 0xd8, 0x3b, 0x59, 0x81, 0x3e, 0x73, 0x1a, - 0xf9, 0x30, 0xa9, 0x52, 0x29, 0xf1, 0x28, 0x8f, 0xd1, 0xf4, 0x04, 0x5b, 0xdb, 0xfd, 0x87, 0x88, - 0x54, 0x2a, 0x8b, 0x6a, 0x8a, 0x12, 0xee, 0xa0, 0x6d, 0x84, 0x4c, 0x9a, 0xec, 0x99, 0xb4, 0xf6, - 0x2a, 0x94, 0xa3, 0xf6, 0xba, 0x1b, 0xec, 0x38, 0x9e, 0xcf, 0xf4, 0xb4, 0x1a, 0x2b, 0x5f, 0x97, - 0x00, 0x9c, 0xe0, 0xa0, 0x65, 0x28, 0x39, 0x52, 0x1f, 0x81, 0xf2, 0x23, 0x6d, 0x28, 0x2d, 0x04, - 0x77, 0x3e, 0x97, 0x1a, 0x08, 0x55, 0x17, 0xbd, 0x0a, 0x63, 0xc2, 0xfd, 0x50, 0x64, 0xf1, 0x3b, - 0x65, 0xfa, 0x88, 0xd4, 0x75, 0x20, 0x36, 0x71, 0xd1, 0x6d, 0x18, 0x89, 0x83, 0x26, 0x73, 0x74, - 0xa0, 0x1c, 0xd2, 0xd9, 0xfc, 0x68, 0x5d, 0x6b, 0x0a, 0x4d, 0x17, 0x05, 0xaa, 0xaa, 0x58, 0xa7, - 0x83, 0xd6, 0xf8, 0x7a, 0x67, 0x71, 0x8c, 0x49, 0x34, 0xfd, 0x48, 0xfe, 0x9d, 0xa4, 0xc2, 0x1d, - 0x9b, 0xdb, 0x41, 0xd4, 0xc4, 0x3a, 0x19, 0x74, 0x0d, 0xa6, 0x5a, 0xa1, 0x17, 0xb0, 0x35, 0xa1, - 0x54, 0x51, 0xd3, 0x66, 0xf6, 0x95, 0x5a, 0x1a, 0x01, 0x77, 0xd6, 0x61, 0xde, 0xa3, 0xa2, 0x70, - 0xfa, 0x1c, 0xcf, 0xda, 0xcb, 0x5f, 0x46, 0xbc, 0x0c, 0x2b, 0x28, 0x5a, 0x61, 0x27, 0x31, 0x7f, - 0xd4, 0x4f, 0xcf, 0xe4, 0x07, 0xf7, 0xd0, 0x1f, 0xff, 0x9c, 0xef, 0x53, 0x7f, 0x71, 0x42, 0x01, - 0xb9, 0x5a, 0x46, 0x38, 0xca, 0x6c, 0x47, 0xd3, 0xe7, 0xbb, 0x58, 0x79, 0xa5, 0x38, 0xf3, 0x84, - 0x21, 0x30, 0x8a, 0x23, 0x9c, 0xa2, 0x89, 0xbe, 0x05, 0x26, 0x45, 0x30, 0xb1, 0x64, 0x98, 0x2e, - 0x24, 0xe6, 0xa3, 0x38, 0x05, 0xc3, 0x1d, 0xd8, 0x3c, 0xbe, 0xbb, 0xb3, 0xde, 0x24, 0xe2, 0xe8, - 0xbb, 0xe9, 0xf9, 0xdb, 0xd1, 0xf4, 0x45, 0x76, 0x3e, 0x88, 0xf8, 0xee, 0x69, 0x28, 0xce, 0xa8, - 0x81, 0xd6, 0x60, 0xb2, 0x15, 0x12, 0xb2, 0xc3, 0x78, 0x64, 0x71, 0x9f, 0xcd, 0x72, 0xe7, 0x69, - 0xda, 0x93, 0x5a, 0x0a, 0x76, 0x98, 0x51, 0x86, 0x3b, 0x28, 0xa0, 0x3d, 0x28, 0x05, 0xbb, 0x24, - 0xdc, 0x22, 0x8e, 0x3b, 0x7d, 0xa9, 0x8b, 0x39, 0xb3, 0xb8, 0xdc, 0x6e, 0x09, 0xdc, 0x94, 0xfa, - 0x5a, 0x16, 0xf7, 0x56, 0x5f, 0xcb, 0xc6, 0xd0, 0x0f, 0x5a, 0x70, 0x4e, 0x4a, 0xbc, 0xeb, 0x2d, - 0x3a, 0xea, 0x8b, 0x81, 0x1f, 0xc5, 0x21, 0x77, 0xf7, 0x7d, 0x2c, 0xdf, 0x05, 0x76, 0x2d, 0xa7, - 0x92, 0x92, 0x2b, 0x9e, 0xcb, 0xc3, 0x88, 0x70, 0x7e, 0x8b, 0x33, 0xdf, 0x0c, 0x53, 0x1d, 0x37, - 0xf7, 0x51, 0x52, 0x4e, 0xcc, 0x6c, 0xc3, 0x98, 0x31, 0x3a, 0x0f, 0x55, 0x73, 0xf9, 0x2f, 0x87, - 0xa1, 0xac, 0xb4, 0x5a, 0xe8, 0xaa, 0xa9, 0xac, 0x3c, 0x97, 0x56, 0x56, 0x96, 0xe8, 0x6b, 0x56, - 0xd7, 0x4f, 0xae, 0x65, 0x04, 0x57, 0xca, 0xdb, 0x8b, 0xfd, 0x7b, 0xcd, 0x6a, 0x42, 0xca, 0x62, - 0xdf, 0x5a, 0xcf, 0x81, 0xae, 0x72, 0xcf, 0x6b, 0x30, 0xe5, 0x07, 0x8c, 0x5d, 0x24, 0xae, 0xe4, - 0x05, 0xd8, 0x95, 0x5f, 0xd6, 0xa3, 0x15, 0xa4, 0x10, 0x70, 0x67, 0x1d, 0xda, 0x20, 0xbf, 0xb3, - 0xd3, 0x82, 0x56, 0x7e, 0xa5, 0x63, 0x01, 0x45, 0x8f, 0xc3, 0x60, 0x2b, 0x70, 0xab, 0x35, 0xc1, - 0x2a, 0x6a, 0xe9, 0x47, 0xdd, 0x6a, 0x0d, 0x73, 0x18, 0x9a, 0x87, 0x21, 0xf6, 0x23, 0x9a, 0x1e, - 0xcd, 0x77, 0x4b, 0x67, 0x35, 0xb4, 0x84, 0x1e, 0xac, 0x02, 0x16, 0x15, 0x99, 0xc0, 0x87, 0xf2, - 0xd7, 0x4c, 0xe0, 0x33, 0xfc, 0x80, 0x02, 0x1f, 0x49, 0x00, 0x27, 0xb4, 0xd0, 0x3d, 0x38, 0x63, - 0xbc, 0x69, 0xf8, 0x12, 0x21, 0x91, 0x70, 0x8d, 0x7d, 0xbc, 0xeb, 0x63, 0x46, 0x68, 0x49, 0x2f, - 0x88, 0x4e, 0x9f, 0xa9, 0x66, 0x51, 0xc2, 0xd9, 0x0d, 0xa0, 0x26, 0x4c, 0x35, 0x3a, 0x5a, 0x2d, - 0xf5, 0xdf, 0xaa, 0x9a, 0xd0, 0xce, 0x16, 0x3b, 0x09, 0xa3, 0x57, 0xa1, 0xf4, 0x6e, 0x10, 0xb1, - 0x63, 0x56, 0xb0, 0xb7, 0xd2, 0xaf, 0xb2, 0xf4, 0xe6, 0xad, 0x3a, 0x2b, 0x3f, 0x3c, 0x98, 0x1d, - 0xa9, 0x05, 0xae, 0xfc, 0x8b, 0x55, 0x05, 0xf4, 0xbd, 0x16, 0xcc, 0x74, 0x3e, 0x9a, 0x54, 0xa7, - 0xc7, 0xfa, 0xef, 0xb4, 0x2d, 0x1a, 0x9d, 0x59, 0xca, 0x25, 0x87, 0xbb, 0x34, 0x65, 0x7f, 0x99, - 0x6b, 0x34, 0x85, 0xde, 0x83, 0x44, 0xed, 0xe6, 0x49, 0x24, 0x40, 0x5c, 0x32, 0x54, 0x32, 0x0f, - 0xac, 0x35, 0xff, 0x35, 0x8b, 0x69, 0xcd, 0xd7, 0xc8, 0x4e, 0xab, 0xe9, 0xc4, 0x27, 0xe1, 0x96, - 0xf7, 0x26, 0x94, 0x62, 0xd1, 0x5a, 0xb7, 0x9c, 0x8d, 0x5a, 0xa7, 0x98, 0xe5, 0x80, 0x62, 0x36, - 0x65, 0x29, 0x56, 0x64, 0xec, 0x7f, 0xc8, 0x67, 0x40, 0x42, 0x4e, 0x40, 0xf2, 0x5d, 0x31, 0x25, - 0xdf, 0xb3, 0x3d, 0xbe, 0x20, 0x47, 0x02, 0xfe, 0x0f, 0xcc, 0x7e, 0x33, 0x21, 0xcb, 0x07, 0xdd, - 0x5c, 0xc3, 0xfe, 0x61, 0x0b, 0x4e, 0x67, 0xd9, 0x37, 0xd2, 0x07, 0x02, 0x17, 0xf1, 0x28, 0xf3, - 0x15, 0x35, 0x82, 0x77, 0x44, 0x39, 0x56, 0x18, 0x7d, 0xa7, 0x43, 0x3a, 0x5a, 0x78, 0xd0, 0x5b, - 0x30, 0x56, 0x0b, 0x89, 0x76, 0xa1, 0xbd, 0xce, 0xfd, 0x6c, 0x79, 0x7f, 0x9e, 0x3e, 0xb2, 0x8f, - 0xad, 0xfd, 0x33, 0x05, 0x38, 0xcd, 0xf5, 0xcf, 0xf3, 0xbb, 0x81, 0xe7, 0xd6, 0x02, 0x57, 0xa4, - 0xb2, 0x7a, 0x0b, 0x46, 0x5b, 0x9a, 0x5c, 0xae, 0x5b, 0xa8, 0x3b, 0x5d, 0x7e, 0x97, 0x48, 0x12, - 0xf4, 0x52, 0x6c, 0xd0, 0x42, 0x2e, 0x8c, 0x92, 0x5d, 0xaf, 0xa1, 0x94, 0x98, 0x85, 0x23, 0x5f, - 0x2e, 0xaa, 0x95, 0x25, 0x8d, 0x0e, 0x36, 0xa8, 0x3e, 0x84, 0xec, 0xa6, 0xf6, 0x8f, 0x58, 0xf0, - 0x48, 0x4e, 0x60, 0x3c, 0xda, 0xdc, 0x1e, 0xd3, 0xf4, 0x8b, 0x44, 0x89, 0xaa, 0x39, 0xae, 0xff, - 0xc7, 0x02, 0x8a, 0x3e, 0x0d, 0xc0, 0xf5, 0xf7, 0xf4, 0x85, 0xda, 0x2b, 0x82, 0x98, 0x11, 0xfc, - 0x48, 0x8b, 0x63, 0x23, 0xeb, 0x63, 0x8d, 0x96, 0xfd, 0x93, 0x45, 0x18, 0xe4, 0x29, 0x9e, 0x97, - 0x61, 0x78, 0x8b, 0x07, 0xf8, 0xef, 0x27, 0x97, 0x40, 0x22, 0x3b, 0xe0, 0x05, 0x58, 0x56, 0x46, - 0x2b, 0x70, 0x8a, 0x27, 0x48, 0x68, 0x56, 0x48, 0xd3, 0xd9, 0x97, 0x82, 0x2e, 0x9e, 0x5c, 0x50, - 0x09, 0xfc, 0xaa, 0x9d, 0x28, 0x38, 0xab, 0x1e, 0x7a, 0x1d, 0xc6, 0xe9, 0xc3, 0x23, 0x68, 0xc7, - 0x92, 0x12, 0x4f, 0x8d, 0xa0, 0x5e, 0x3a, 0x6b, 0x06, 0x14, 0xa7, 0xb0, 0xe9, 0xdb, 0xb7, 0xd5, - 0x21, 0xd2, 0x1b, 0x4c, 0xde, 0xbe, 0xa6, 0x18, 0xcf, 0xc4, 0x65, 0x86, 0x8d, 0x6d, 0x66, 0xc6, - 0xb9, 0xb6, 0x15, 0x92, 0x68, 0x2b, 0x68, 0xba, 0x8c, 0xd1, 0x1a, 0xd4, 0x0c, 0x1b, 0x53, 0x70, - 0xdc, 0x51, 0x83, 0x52, 0xd9, 0x70, 0xbc, 0x66, 0x3b, 0x24, 0x09, 0x95, 0x21, 0x93, 0xca, 0x72, - 0x0a, 0x8e, 0x3b, 0x6a, 0xd0, 0x75, 0x74, 0xa6, 0x16, 0x06, 0xf4, 0xf0, 0x92, 0xd1, 0x3e, 0x94, - 0xb5, 0xea, 0xb0, 0x74, 0x4c, 0xec, 0x12, 0x17, 0x4b, 0xd8, 0xf3, 0x71, 0x0a, 0x86, 0xaa, 0xba, - 0x2e, 0x5c, 0x12, 0x25, 0x15, 0xf4, 0x2c, 0x8c, 0x88, 0xb0, 0xf7, 0xcc, 0xa8, 0x92, 0x4f, 0x1d, - 0x53, 0xad, 0x57, 0x92, 0x62, 0xac, 0xe3, 0xd8, 0xdf, 0x57, 0x80, 0x53, 0x19, 0x56, 0xf1, 0xfc, - 0xa8, 0xda, 0xf4, 0xa2, 0x58, 0x25, 0x50, 0xd3, 0x8e, 0x2a, 0x5e, 0x8e, 0x15, 0x06, 0xdd, 0x0f, - 0xfc, 0x30, 0x4c, 0x1f, 0x80, 0xc2, 0xea, 0x54, 0x40, 0x8f, 0x98, 0x8a, 0xec, 0x12, 0x0c, 0xb4, - 0x23, 0x22, 0x23, 0xda, 0xa9, 0xf3, 0x9b, 0x69, 0x5c, 0x18, 0x84, 0xb2, 0xc7, 0x9b, 0x4a, 0x79, - 0xa1, 0xb1, 0xc7, 0x5c, 0x7d, 0xc1, 0x61, 0xb4, 0x73, 0x31, 0xf1, 0x1d, 0x3f, 0x16, 0x4c, 0x74, - 0x12, 0x9a, 0x89, 0x95, 0x62, 0x01, 0xb5, 0xbf, 0x54, 0x84, 0x73, 0xb9, 0x7e, 0x32, 0xb4, 0xeb, - 0x3b, 0x81, 0xef, 0xc5, 0x81, 0xb2, 0x59, 0xe0, 0xe1, 0x98, 0x48, 0x6b, 0x6b, 0x45, 0x94, 0x63, - 0x85, 0x81, 0x2e, 0xc3, 0x20, 0x13, 0x3a, 0x75, 0xa4, 0x92, 0x5b, 0xa8, 0xf0, 0xf8, 0x1c, 0x1c, - 0xdc, 0x77, 0x9a, 0xce, 0xc7, 0x61, 0xa0, 0x15, 0x04, 0xcd, 0xf4, 0xa1, 0x45, 0xbb, 0x1b, 0x04, - 0x4d, 0xcc, 0x80, 0xe8, 0x63, 0x62, 0xbc, 0x52, 0x4a, 0x7a, 0xec, 0xb8, 0x41, 0xa4, 0x0d, 0xda, - 0x93, 0x30, 0xbc, 0x4d, 0xf6, 0x43, 0xcf, 0xdf, 0x4c, 0x1b, 0x6f, 0xdc, 0xe0, 0xc5, 0x58, 0xc2, - 0xcd, 0xac, 0x40, 0xc3, 0xc7, 0x9d, 0x5f, 0xb3, 0xd4, 0xf3, 0x0a, 0xfc, 0xfe, 0x22, 0x4c, 0xe0, - 0x85, 0xca, 0x87, 0x13, 0x71, 0xbb, 0x73, 0x22, 0x8e, 0x3b, 0xbf, 0x66, 0xef, 0xd9, 0xf8, 0x45, - 0x0b, 0x26, 0x58, 0xf0, 0x7d, 0x11, 0xc8, 0xc7, 0x0b, 0xfc, 0x13, 0x60, 0xf1, 0x1e, 0x87, 0xc1, - 0x90, 0x36, 0x9a, 0xce, 0x21, 0xc7, 0x7a, 0x82, 0x39, 0x0c, 0x9d, 0x87, 0x01, 0xd6, 0x05, 0x3a, - 0x79, 0xa3, 0x3c, 0xfd, 0x4e, 0xc5, 0x89, 0x1d, 0xcc, 0x4a, 0x59, 0x74, 0x0a, 0x4c, 0x5a, 0x4d, - 0x8f, 0x77, 0x3a, 0x51, 0x09, 0x7e, 0x30, 0xa2, 0x53, 0x64, 0x76, 0xed, 0xfd, 0x45, 0xa7, 0xc8, - 0x26, 0xd9, 0xfd, 0xf9, 0xf4, 0x87, 0x05, 0xb8, 0x98, 0x59, 0xaf, 0xef, 0xe8, 0x14, 0xdd, 0x6b, - 0x3f, 0xcc, 0x20, 0xed, 0xc5, 0x13, 0x34, 0x8d, 0x1b, 0xe8, 0x97, 0xc3, 0x1c, 0xec, 0x23, 0x68, - 0x44, 0xe6, 0x90, 0x7d, 0x40, 0x82, 0x46, 0x64, 0xf6, 0x2d, 0xe7, 0xf9, 0xf7, 0xe7, 0x85, 0x9c, - 0x6f, 0x61, 0x0f, 0xc1, 0x2b, 0xf4, 0x9c, 0x61, 0xc0, 0x48, 0x70, 0xcc, 0xa3, 0xfc, 0x8c, 0xe1, - 0x65, 0x58, 0x41, 0xd1, 0x3c, 0x4c, 0xec, 0x78, 0x3e, 0x3d, 0x7c, 0xf6, 0x4d, 0xc6, 0x4f, 0xc5, - 0xf4, 0x59, 0x31, 0xc1, 0x38, 0x8d, 0x8f, 0x3c, 0x2d, 0xa0, 0x44, 0x21, 0x3f, 0x2b, 0x73, 0x6e, - 0x6f, 0xe7, 0x4c, 0x75, 0xa9, 0x1a, 0xc5, 0x8c, 0xe0, 0x12, 0x2b, 0xda, 0xfb, 0xbf, 0xd8, 0xff, - 0xfb, 0x7f, 0x34, 0xfb, 0xed, 0x3f, 0xf3, 0x2a, 0x8c, 0x3d, 0xb0, 0xc0, 0xd7, 0xfe, 0x6a, 0x11, - 0x1e, 0xed, 0xb2, 0xed, 0xf9, 0x59, 0x6f, 0xcc, 0x81, 0x76, 0xd6, 0x77, 0xcc, 0x43, 0x0d, 0x4e, - 0x6f, 0xb4, 0x9b, 0xcd, 0x7d, 0x66, 0x7d, 0x4e, 0x5c, 0x89, 0x21, 0x78, 0xca, 0xf3, 0x32, 0xe1, - 0xd1, 0x72, 0x06, 0x0e, 0xce, 0xac, 0x49, 0x19, 0x7a, 0x7a, 0x93, 0xec, 0x2b, 0x52, 0x29, 0x86, - 0x1e, 0xeb, 0x40, 0x6c, 0xe2, 0xa2, 0x6b, 0x30, 0xe5, 0xec, 0x3a, 0x1e, 0x8f, 0xca, 0x29, 0x09, - 0x70, 0x8e, 0x5e, 0xc9, 0xe9, 0xe6, 0xd3, 0x08, 0xb8, 0xb3, 0x0e, 0x7a, 0x03, 0x50, 0x20, 0xb2, - 0xca, 0x5f, 0x23, 0xbe, 0xd0, 0x6a, 0xb1, 0xb9, 0x2b, 0x26, 0x47, 0xc2, 0xad, 0x0e, 0x0c, 0x9c, - 0x51, 0x2b, 0x15, 0xa0, 0x61, 0x28, 0x3f, 0x40, 0x43, 0xf7, 0x73, 0xb1, 0x67, 0x7e, 0x80, 0xff, - 0x64, 0xd1, 0xeb, 0x8b, 0x33, 0xf9, 0x66, 0x9c, 0xb1, 0x57, 0x99, 0x41, 0x17, 0x97, 0xe1, 0x69, - 0xb1, 0x12, 0xce, 0x68, 0x06, 0x5d, 0x09, 0x10, 0x9b, 0xb8, 0x7c, 0x41, 0x44, 0x89, 0x8b, 0x9e, - 0xc1, 0xe2, 0x8b, 0x60, 0x28, 0x0a, 0x03, 0x7d, 0x06, 0x86, 0x5d, 0x6f, 0xd7, 0x8b, 0x82, 0x50, - 0xac, 0xf4, 0x23, 0xaa, 0x0b, 0x92, 0x73, 0xb0, 0xc2, 0xc9, 0x60, 0x49, 0xcf, 0xfe, 0xfe, 0x02, - 0x8c, 0xc9, 0x16, 0xdf, 0x6c, 0x07, 0xb1, 0x73, 0x02, 0xd7, 0xf2, 0x35, 0xe3, 0x5a, 0xfe, 0x58, - 0xb7, 0x88, 0x30, 0xac, 0x4b, 0xb9, 0xd7, 0xf1, 0xad, 0xd4, 0x75, 0xfc, 0x44, 0x6f, 0x52, 0xdd, - 0xaf, 0xe1, 0x7f, 0x64, 0xc1, 0x94, 0x81, 0x7f, 0x02, 0xb7, 0xc1, 0xb2, 0x79, 0x1b, 0x3c, 0xd6, - 0xf3, 0x1b, 0x72, 0x6e, 0x81, 0xef, 0x2e, 0xa6, 0xfa, 0xce, 0x4e, 0xff, 0x77, 0x61, 0x60, 0xcb, - 0x09, 0xdd, 0x6e, 0x11, 0xb0, 0x3b, 0x2a, 0xcd, 0x5d, 0x77, 0x42, 0xa1, 0xd6, 0x7b, 0x5a, 0x25, - 0x45, 0x76, 0xc2, 0xde, 0x2a, 0x3d, 0xd6, 0x14, 0x7a, 0x19, 0x86, 0xa2, 0x46, 0xd0, 0x52, 0xf6, - 0xe2, 0x97, 0x78, 0xc2, 0x64, 0x5a, 0x72, 0x78, 0x30, 0x8b, 0xcc, 0xe6, 0x68, 0x31, 0x16, 0xf8, - 0xe8, 0x2d, 0x18, 0x63, 0xbf, 0x94, 0x8d, 0x4d, 0x31, 0x3f, 0x5b, 0x4e, 0x5d, 0x47, 0xe4, 0x06, - 0x68, 0x46, 0x11, 0x36, 0x49, 0xcd, 0x6c, 0x42, 0x59, 0x7d, 0xd6, 0x43, 0xd5, 0xc7, 0xfd, 0xdb, - 0x22, 0x9c, 0xca, 0x58, 0x73, 0x28, 0x32, 0x66, 0xe2, 0xd9, 0x3e, 0x97, 0xea, 0xfb, 0x9c, 0x8b, - 0x88, 0xbd, 0x86, 0x5c, 0xb1, 0xb6, 0xfa, 0x6e, 0xf4, 0x76, 0x44, 0xd2, 0x8d, 0xd2, 0xa2, 0xde, - 0x8d, 0xd2, 0xc6, 0x4e, 0x6c, 0xa8, 0x69, 0x43, 0xaa, 0xa7, 0x0f, 0x75, 0x4e, 0xff, 0xa4, 0x08, - 0xa7, 0xb3, 0x82, 0x54, 0xa1, 0x6f, 0x4b, 0x65, 0x4e, 0x7b, 0xa1, 0xdf, 0xf0, 0x56, 0x3c, 0x9d, - 0x1a, 0x97, 0x01, 0x2f, 0xcc, 0x99, 0xb9, 0xd4, 0x7a, 0x0e, 0xb3, 0x68, 0x93, 0xb9, 0x9f, 0x87, - 0x3c, 0xe3, 0x9d, 0x3c, 0x3e, 0x3e, 0xd9, 0x77, 0x07, 0x44, 0xaa, 0xbc, 0x28, 0xa5, 0xbf, 0x97, - 0xc5, 0xbd, 0xf5, 0xf7, 0xb2, 0xe5, 0x19, 0x0f, 0x46, 0xb4, 0xaf, 0x79, 0xa8, 0x33, 0xbe, 0x4d, - 0x6f, 0x2b, 0xad, 0xdf, 0x0f, 0x75, 0xd6, 0x7f, 0xc4, 0x82, 0x94, 0x35, 0xb4, 0x12, 0x8b, 0x59, - 0xb9, 0x62, 0xb1, 0x4b, 0x30, 0x10, 0x06, 0x4d, 0x92, 0x4e, 0x54, 0x86, 0x83, 0x26, 0xc1, 0x0c, - 0x42, 0x31, 0xe2, 0x44, 0xd8, 0x31, 0xaa, 0x3f, 0xe4, 0xc4, 0x13, 0xed, 0x71, 0x18, 0x6c, 0x92, - 0x5d, 0xd2, 0x4c, 0xe7, 0x93, 0xb8, 0x49, 0x0b, 0x31, 0x87, 0xd9, 0xbf, 0x38, 0x00, 0x17, 0xba, - 0x06, 0x70, 0xa0, 0xcf, 0xa1, 0x4d, 0x27, 0x26, 0x7b, 0xce, 0x7e, 0x3a, 0xf0, 0xfb, 0x35, 0x5e, - 0x8c, 0x25, 0x9c, 0xf9, 0xab, 0xf0, 0xf8, 0xad, 0x29, 0x21, 0xa2, 0x08, 0xdb, 0x2a, 0xa0, 0xa6, - 0x50, 0xaa, 0x78, 0x1c, 0x42, 0xa9, 0xe7, 0x00, 0xa2, 0xa8, 0xc9, 0x0d, 0x5f, 0x5c, 0xe1, 0x08, - 0x93, 0xc4, 0xf9, 0xad, 0xdf, 0x14, 0x10, 0xac, 0x61, 0xa1, 0x0a, 0x4c, 0xb6, 0xc2, 0x20, 0xe6, - 0x32, 0xd9, 0x0a, 0xb7, 0x0d, 0x1b, 0x34, 0x7d, 0xe7, 0x6b, 0x29, 0x38, 0xee, 0xa8, 0x81, 0x5e, - 0x84, 0x11, 0xe1, 0x4f, 0x5f, 0x0b, 0x82, 0xa6, 0x10, 0x03, 0x29, 0x73, 0xa9, 0x7a, 0x02, 0xc2, - 0x3a, 0x9e, 0x56, 0x8d, 0x09, 0x7a, 0x87, 0x33, 0xab, 0x71, 0x61, 0xaf, 0x86, 0x97, 0x0a, 0x58, - 0x57, 0xea, 0x2b, 0x60, 0x5d, 0x22, 0x18, 0x2b, 0xf7, 0xad, 0xdb, 0x82, 0x9e, 0xa2, 0xa4, 0x9f, - 0x1b, 0x80, 0x53, 0x62, 0xe1, 0x3c, 0xec, 0xe5, 0x72, 0xbb, 0x73, 0xb9, 0x1c, 0x87, 0xe8, 0xec, - 0xc3, 0x35, 0x73, 0xd2, 0x6b, 0xe6, 0x07, 0x2c, 0x30, 0xd9, 0x2b, 0xf4, 0x7f, 0xe7, 0x66, 0xce, - 0x78, 0x31, 0x97, 0x5d, 0x73, 0xe5, 0x05, 0xf2, 0x3e, 0x73, 0x68, 0xd8, 0xff, 0xc1, 0x82, 0xc7, - 0x7a, 0x52, 0x44, 0x4b, 0x50, 0x66, 0x3c, 0xa0, 0xf6, 0x3a, 0x7b, 0x42, 0xd9, 0x8e, 0x4a, 0x40, - 0x0e, 0x4b, 0x9a, 0xd4, 0x44, 0x4b, 0x1d, 0x29, 0x4a, 0x9e, 0xcc, 0x48, 0x51, 0x72, 0xc6, 0x18, - 0x9e, 0x07, 0xcc, 0x51, 0xf2, 0xe5, 0x22, 0x0c, 0xf1, 0x15, 0x7f, 0x02, 0xcf, 0xb0, 0x65, 0x21, - 0xb7, 0xed, 0x12, 0x11, 0x8f, 0xf7, 0x65, 0xae, 0xe2, 0xc4, 0x0e, 0x67, 0x13, 0xd4, 0x6d, 0x95, - 0x48, 0x78, 0xd1, 0xe7, 0x00, 0xa2, 0x38, 0xf4, 0xfc, 0x4d, 0x5a, 0x26, 0x62, 0x25, 0x7e, 0xbc, - 0x0b, 0xb5, 0xba, 0x42, 0xe6, 0x34, 0x93, 0x9d, 0xab, 0x00, 0x58, 0xa3, 0x88, 0xe6, 0x8c, 0xfb, - 0x72, 0x26, 0x25, 0xf8, 0x04, 0x4e, 0x35, 0xb9, 0x3d, 0x67, 0x5e, 0x82, 0xb2, 0x22, 0xde, 0x4b, - 0x8a, 0x33, 0xaa, 0x33, 0x17, 0x9f, 0x82, 0x89, 0x54, 0xdf, 0x8e, 0x24, 0x04, 0xfa, 0x25, 0x0b, - 0x26, 0x78, 0x67, 0x96, 0xfc, 0x5d, 0x71, 0xa6, 0xbe, 0x07, 0xa7, 0x9b, 0x19, 0x67, 0x9b, 0x98, - 0xd1, 0xfe, 0xcf, 0x42, 0x25, 0xf4, 0xc9, 0x82, 0xe2, 0xcc, 0x36, 0xd0, 0x15, 0xba, 0x6e, 0xe9, - 0xd9, 0xe5, 0x34, 0x85, 0x5b, 0xe3, 0x28, 0x5f, 0xb3, 0xbc, 0x0c, 0x2b, 0xa8, 0xfd, 0xdb, 0x16, - 0x4c, 0xf1, 0x9e, 0xdf, 0x20, 0xfb, 0x6a, 0x87, 0x7f, 0x3d, 0xfb, 0x2e, 0xb2, 0x06, 0x15, 0x72, - 0xb2, 0x06, 0xe9, 0x9f, 0x56, 0xec, 0xfa, 0x69, 0x3f, 0x63, 0x81, 0x58, 0x21, 0x27, 0xf0, 0x94, - 0xff, 0x66, 0xf3, 0x29, 0x3f, 0x93, 0xbf, 0x09, 0x72, 0xde, 0xf0, 0x7f, 0x66, 0xc1, 0x24, 0x47, - 0x48, 0x74, 0xce, 0x5f, 0xd7, 0x79, 0xe8, 0x27, 0xb7, 0xe8, 0x0d, 0xb2, 0xbf, 0x16, 0xd4, 0x9c, - 0x78, 0x2b, 0xfb, 0xa3, 0x8c, 0xc9, 0x1a, 0xe8, 0x3a, 0x59, 0xae, 0xdc, 0x40, 0x47, 0x48, 0x58, - 0x7c, 0xe4, 0xa0, 0xfa, 0xf6, 0xd7, 0x2c, 0x40, 0xbc, 0x19, 0x83, 0xfd, 0xa1, 0x4c, 0x05, 0x2b, - 0xd5, 0xae, 0x8b, 0xe4, 0x68, 0x52, 0x10, 0xac, 0x61, 0x1d, 0xcb, 0xf0, 0xa4, 0x0c, 0x07, 0x8a, - 0xbd, 0x0d, 0x07, 0x8e, 0x30, 0xa2, 0x7f, 0x30, 0x08, 0x69, 0x0f, 0x10, 0x74, 0x07, 0x46, 0x1b, - 0x4e, 0xcb, 0x59, 0xf7, 0x9a, 0x5e, 0xec, 0x91, 0xa8, 0x9b, 0xc5, 0xd1, 0xa2, 0x86, 0x27, 0x54, - 0xbd, 0x5a, 0x09, 0x36, 0xe8, 0xa0, 0x39, 0x80, 0x56, 0xe8, 0xed, 0x7a, 0x4d, 0xb2, 0xc9, 0x24, - 0x0e, 0xcc, 0x91, 0x9a, 0x9b, 0xd1, 0xc8, 0x52, 0xac, 0x61, 0x64, 0x78, 0xaa, 0x16, 0x1f, 0xb2, - 0xa7, 0x2a, 0x9c, 0x98, 0xa7, 0xea, 0xc0, 0x91, 0x3c, 0x55, 0x4b, 0x47, 0xf6, 0x54, 0x1d, 0xec, - 0xcb, 0x53, 0x15, 0xc3, 0x59, 0xc9, 0xc1, 0xd1, 0xff, 0xcb, 0x5e, 0x93, 0x08, 0xb6, 0x9d, 0x7b, - 0x7f, 0xcf, 0xdc, 0x3f, 0x98, 0x3d, 0x8b, 0x33, 0x31, 0x70, 0x4e, 0x4d, 0xf4, 0x69, 0x98, 0x76, - 0x9a, 0xcd, 0x60, 0x4f, 0x4d, 0xea, 0x52, 0xd4, 0x70, 0x9a, 0x5c, 0x94, 0x3f, 0xcc, 0xa8, 0x9e, - 0xbf, 0x7f, 0x30, 0x3b, 0x3d, 0x9f, 0x83, 0x83, 0x73, 0x6b, 0xa3, 0xd7, 0xa0, 0xdc, 0x0a, 0x83, - 0xc6, 0x8a, 0xe6, 0xa6, 0x76, 0x91, 0x0e, 0x60, 0x4d, 0x16, 0x1e, 0x1e, 0xcc, 0x8e, 0xa9, 0x3f, - 0xec, 0xc2, 0x4f, 0x2a, 0xd8, 0xdb, 0x70, 0xaa, 0x4e, 0x42, 0x8f, 0xa5, 0x1f, 0x76, 0x93, 0xf3, - 0x63, 0x0d, 0xca, 0x61, 0xea, 0xc4, 0xec, 0x2b, 0x8a, 0x9c, 0x16, 0x7d, 0x5c, 0x9e, 0x90, 0x09, - 0x21, 0xfb, 0x7f, 0x5a, 0x30, 0x2c, 0x3c, 0x32, 0x4e, 0x80, 0x51, 0x9b, 0x37, 0xe4, 0xe5, 0xb3, - 0xd9, 0xb7, 0x0a, 0xeb, 0x4c, 0xae, 0xa4, 0xbc, 0x9a, 0x92, 0x94, 0x3f, 0xd6, 0x8d, 0x48, 0x77, - 0x19, 0xf9, 0x5f, 0x2b, 0xc2, 0xb8, 0xe9, 0xba, 0x77, 0x02, 0x43, 0xb0, 0x0a, 0xc3, 0x91, 0xf0, - 0x4d, 0x2b, 0xe4, 0x5b, 0x64, 0xa7, 0x27, 0x31, 0xb1, 0xd6, 0x12, 0xde, 0x68, 0x92, 0x48, 0xa6, - 0xd3, 0x5b, 0xf1, 0x21, 0x3a, 0xbd, 0xf5, 0xf2, 0x9e, 0x1c, 0x38, 0x0e, 0xef, 0x49, 0xfb, 0x2b, - 0xec, 0x66, 0xd3, 0xcb, 0x4f, 0x80, 0xe9, 0xb9, 0x66, 0xde, 0x81, 0x76, 0x97, 0x95, 0x25, 0x3a, - 0x95, 0xc3, 0xfc, 0xfc, 0x82, 0x05, 0x17, 0x32, 0xbe, 0x4a, 0xe3, 0x84, 0x9e, 0x86, 0x92, 0xd3, - 0x76, 0x3d, 0xb5, 0x97, 0x35, 0xad, 0xd9, 0xbc, 0x28, 0xc7, 0x0a, 0x03, 0x2d, 0xc2, 0x14, 0xb9, - 0xd7, 0xf2, 0xb8, 0xc2, 0x50, 0x37, 0xa9, 0x2c, 0xf2, 0xc8, 0xda, 0x4b, 0x69, 0x20, 0xee, 0xc4, - 0x57, 0xc1, 0x1e, 0x8a, 0xb9, 0xc1, 0x1e, 0xfe, 0xae, 0x05, 0x23, 0xca, 0x3b, 0xeb, 0xa1, 0x8f, - 0xf6, 0xb7, 0x98, 0xa3, 0xfd, 0x68, 0x97, 0xd1, 0xce, 0x19, 0xe6, 0xbf, 0x51, 0x50, 0xfd, 0xad, - 0x05, 0x61, 0xdc, 0x07, 0x87, 0xf5, 0x32, 0x94, 0x5a, 0x61, 0x10, 0x07, 0x8d, 0xa0, 0x29, 0x18, - 0xac, 0xf3, 0x49, 0xd4, 0x13, 0x5e, 0x7e, 0xa8, 0xfd, 0xc6, 0x0a, 0x9b, 0x8d, 0x5e, 0x10, 0xc6, - 0x82, 0xa9, 0x49, 0x46, 0x2f, 0x08, 0x63, 0xcc, 0x20, 0xc8, 0x05, 0x88, 0x9d, 0x70, 0x93, 0xc4, - 0xb4, 0x4c, 0x44, 0x59, 0xca, 0x3f, 0x3c, 0xda, 0xb1, 0xd7, 0x9c, 0xf3, 0xfc, 0x38, 0x8a, 0xc3, - 0xb9, 0xaa, 0x1f, 0xdf, 0x0a, 0xf9, 0x7b, 0x4d, 0x0b, 0x63, 0xa2, 0x68, 0x61, 0x8d, 0xae, 0x74, - 0x2b, 0x66, 0x6d, 0x0c, 0x9a, 0xfa, 0xf7, 0x55, 0x51, 0x8e, 0x15, 0x86, 0xfd, 0x12, 0xbb, 0x4a, - 0xd8, 0x00, 0x1d, 0x2d, 0xee, 0xc7, 0x97, 0xcb, 0x6a, 0x68, 0x99, 0xf2, 0xad, 0xa2, 0x47, 0x17, - 0xe9, 0x7e, 0x72, 0xd3, 0x86, 0x75, 0x17, 0xa3, 0x24, 0x04, 0x09, 0xfa, 0xd6, 0x0e, 0x9b, 0x8a, - 0x67, 0x7a, 0x5c, 0x01, 0x47, 0xb0, 0xa2, 0x60, 0xd1, 0xfe, 0x59, 0x2c, 0xf4, 0x6a, 0x4d, 0x2c, - 0x72, 0x2d, 0xda, 0xbf, 0x00, 0xe0, 0x04, 0x07, 0x5d, 0x15, 0xaf, 0x71, 0x2e, 0x9a, 0x7e, 0x34, - 0xf5, 0x1a, 0x97, 0x9f, 0xaf, 0x09, 0xb3, 0x9f, 0x85, 0x11, 0x95, 0xeb, 0xb2, 0xc6, 0x53, 0x28, - 0x8a, 0x98, 0x53, 0x4b, 0x49, 0x31, 0xd6, 0x71, 0xd0, 0x1a, 0x4c, 0x44, 0x5c, 0xd4, 0xa3, 0x42, - 0x8b, 0x72, 0x91, 0xd9, 0xc7, 0xa5, 0x21, 0x4a, 0xdd, 0x04, 0x1f, 0xb2, 0x22, 0x7e, 0x74, 0x48, - 0x57, 0xde, 0x34, 0x09, 0xf4, 0x3a, 0x8c, 0x37, 0x03, 0xc7, 0x5d, 0x70, 0x9a, 0x8e, 0xdf, 0x60, - 0xdf, 0x5b, 0x32, 0x53, 0xa6, 0xdd, 0x34, 0xa0, 0x38, 0x85, 0x4d, 0x39, 0x1f, 0xbd, 0x44, 0x84, - 0xc3, 0x75, 0xfc, 0x4d, 0x12, 0x89, 0xcc, 0x85, 0x8c, 0xf3, 0xb9, 0x99, 0x83, 0x83, 0x73, 0x6b, - 0xa3, 0x97, 0x61, 0x54, 0x7e, 0xbe, 0xe6, 0xf9, 0x9e, 0xd8, 0xde, 0x6b, 0x30, 0x6c, 0x60, 0xa2, - 0x3d, 0x38, 0x23, 0xff, 0xaf, 0x85, 0xce, 0xc6, 0x86, 0xd7, 0x10, 0xee, 0xa0, 0xdc, 0x31, 0x6e, - 0x5e, 0x7a, 0x6f, 0x2d, 0x65, 0x21, 0x1d, 0x1e, 0xcc, 0x5e, 0x12, 0xa3, 0x96, 0x09, 0x67, 0x93, - 0x98, 0x4d, 0x1f, 0xad, 0xc0, 0xa9, 0x2d, 0xe2, 0x34, 0xe3, 0xad, 0xc5, 0x2d, 0xd2, 0xd8, 0x96, - 0x9b, 0x88, 0xf9, 0xd3, 0x6b, 0x16, 0xeb, 0xd7, 0x3b, 0x51, 0x70, 0x56, 0x3d, 0xf4, 0x36, 0x4c, - 0xb7, 0xda, 0xeb, 0x4d, 0x2f, 0xda, 0x5a, 0x0d, 0x62, 0x66, 0x8d, 0xa2, 0x52, 0x67, 0x0a, 0xc7, - 0x7b, 0x15, 0xb1, 0xa0, 0x96, 0x83, 0x87, 0x73, 0x29, 0xa0, 0xf7, 0xe0, 0x4c, 0x6a, 0x31, 0x08, - 0xd7, 0xe3, 0xf1, 0xfc, 0xe0, 0xe2, 0xf5, 0xac, 0x0a, 0xc2, 0x8b, 0x3f, 0x0b, 0x84, 0xb3, 0x9b, - 0x40, 0x2f, 0x40, 0xc9, 0x6b, 0x2d, 0x3b, 0x3b, 0x5e, 0x73, 0x9f, 0x45, 0x47, 0x2f, 0xb3, 0x88, - 0xe1, 0xa5, 0x6a, 0x8d, 0x97, 0x1d, 0x6a, 0xbf, 0xb1, 0xc2, 0xa4, 0xfc, 0xbe, 0x16, 0x03, 0x32, - 0x9a, 0x9e, 0x4c, 0x8c, 0x6d, 0xb5, 0x40, 0x91, 0x11, 0x36, 0xb0, 0xde, 0x9f, 0x0d, 0xd3, 0xbb, - 0xb4, 0xb2, 0xc6, 0x00, 0xa2, 0xcf, 0xc3, 0xa8, 0xbe, 0x62, 0xc5, 0x65, 0x76, 0x39, 0x9b, 0x3f, - 0xd2, 0x56, 0x36, 0x67, 0x1f, 0xd5, 0xea, 0xd5, 0x61, 0xd8, 0xa0, 0x68, 0x13, 0xc8, 0x1e, 0x4b, - 0x74, 0x13, 0x4a, 0x8d, 0xa6, 0x47, 0xfc, 0xb8, 0x5a, 0xeb, 0x16, 0xbe, 0x68, 0x51, 0xe0, 0x88, - 0xc9, 0x11, 0x91, 0x9f, 0x79, 0x19, 0x56, 0x14, 0xec, 0x5f, 0x2d, 0xc0, 0x6c, 0x8f, 0x30, 0xe2, - 0x29, 0x51, 0xbb, 0xd5, 0x97, 0xa8, 0x7d, 0x5e, 0x26, 0x1d, 0x5d, 0x4d, 0xc9, 0x1f, 0x52, 0x09, - 0x45, 0x13, 0x29, 0x44, 0x1a, 0xbf, 0x6f, 0xd3, 0x67, 0x5d, 0x5a, 0x3f, 0xd0, 0xd3, 0x78, 0xdf, - 0xd0, 0xd2, 0x0d, 0xf6, 0xff, 0xe8, 0xc9, 0xd5, 0xb8, 0xd8, 0x5f, 0x29, 0xc0, 0x19, 0x35, 0x84, - 0xdf, 0xb8, 0x03, 0x77, 0xbb, 0x73, 0xe0, 0x8e, 0x41, 0x5f, 0x65, 0xdf, 0x82, 0x21, 0x1e, 0x8f, - 0xa9, 0x0f, 0x66, 0xeb, 0x71, 0x33, 0x74, 0xa1, 0x62, 0x09, 0x8c, 0xf0, 0x85, 0xdf, 0x6b, 0xc1, - 0xc4, 0xda, 0x62, 0xad, 0x1e, 0x34, 0xb6, 0x49, 0x3c, 0xcf, 0x99, 0x63, 0x2c, 0x78, 0x2d, 0xeb, - 0x01, 0x79, 0xa8, 0x2c, 0xee, 0xec, 0x12, 0x0c, 0x6c, 0x05, 0x51, 0x9c, 0x56, 0x66, 0x5f, 0x0f, - 0xa2, 0x18, 0x33, 0x88, 0xfd, 0x3b, 0x16, 0x0c, 0xb2, 0x34, 0xdb, 0xbd, 0x12, 0xbd, 0xf7, 0xf3, - 0x5d, 0xe8, 0x45, 0x18, 0x22, 0x1b, 0x1b, 0xa4, 0x11, 0x8b, 0x59, 0x95, 0xde, 0xc7, 0x43, 0x4b, - 0xac, 0x94, 0x32, 0x18, 0xac, 0x31, 0xfe, 0x17, 0x0b, 0x64, 0x74, 0x17, 0xca, 0xb1, 0xb7, 0x43, - 0xe6, 0x5d, 0x57, 0xa8, 0x03, 0x1f, 0xc0, 0x83, 0x7a, 0x4d, 0x12, 0xc0, 0x09, 0x2d, 0xfb, 0x4b, - 0x05, 0x80, 0x24, 0x1a, 0x47, 0xaf, 0x4f, 0x5c, 0xe8, 0x50, 0x14, 0x5d, 0xce, 0x50, 0x14, 0xa1, - 0x84, 0x60, 0x86, 0x96, 0x48, 0x0d, 0x53, 0xb1, 0xaf, 0x61, 0x1a, 0x38, 0xca, 0x30, 0x2d, 0xc2, - 0x54, 0x12, 0x4d, 0xc4, 0x0c, 0xa6, 0xc4, 0x1e, 0x44, 0x6b, 0x69, 0x20, 0xee, 0xc4, 0xb7, 0x09, - 0x5c, 0x52, 0x41, 0x15, 0xc4, 0x5d, 0xc3, 0xac, 0x4d, 0x8f, 0x90, 0xf3, 0x3f, 0xd1, 0x84, 0x15, - 0x72, 0x35, 0x61, 0x3f, 0x6e, 0xc1, 0xe9, 0x74, 0x3b, 0xcc, 0xfd, 0xef, 0x8b, 0x16, 0x9c, 0x61, - 0xfa, 0x40, 0xd6, 0x6a, 0xa7, 0xf6, 0xf1, 0x85, 0xae, 0x81, 0x22, 0x72, 0x7a, 0x9c, 0xb8, 0xb9, - 0xaf, 0x64, 0x91, 0xc6, 0xd9, 0x2d, 0xda, 0xff, 0xbe, 0x00, 0xd3, 0x79, 0x11, 0x26, 0x98, 0x31, - 0xba, 0x73, 0xaf, 0xbe, 0x4d, 0xf6, 0x84, 0xc9, 0x6f, 0x62, 0x8c, 0xce, 0x8b, 0xb1, 0x84, 0xa7, - 0x23, 0x43, 0x17, 0xfa, 0x8b, 0x0c, 0x8d, 0xb6, 0x60, 0x6a, 0x6f, 0x8b, 0xf8, 0xb7, 0xfd, 0xc8, - 0x89, 0xbd, 0x68, 0xc3, 0x63, 0x19, 0xdb, 0xf9, 0xba, 0x79, 0x45, 0x1a, 0xe6, 0xde, 0x4d, 0x23, - 0x1c, 0x1e, 0xcc, 0x5e, 0x30, 0x0a, 0x92, 0x2e, 0xf3, 0x83, 0x04, 0x77, 0x12, 0xed, 0x0c, 0xac, - 0x3d, 0xf0, 0x10, 0x03, 0x6b, 0xdb, 0x5f, 0xb4, 0xe0, 0x5c, 0x6e, 0xe2, 0x3b, 0x74, 0x05, 0x4a, - 0x4e, 0xcb, 0xe3, 0x82, 0x53, 0x71, 0x8c, 0x32, 0x01, 0x40, 0xad, 0xca, 0xc5, 0xa6, 0x0a, 0xaa, - 0x12, 0xf2, 0x16, 0x72, 0x13, 0xf2, 0xf6, 0xcc, 0xaf, 0x6b, 0x7f, 0x8f, 0x05, 0xc2, 0x91, 0xae, - 0x8f, 0xb3, 0xfb, 0x2d, 0x99, 0xcf, 0xdc, 0x48, 0xbe, 0x71, 0x29, 0xdf, 0xb3, 0x50, 0xa4, 0xdc, - 0x50, 0xbc, 0x92, 0x91, 0x68, 0xc3, 0xa0, 0x65, 0xbb, 0x20, 0xa0, 0x15, 0xc2, 0xc4, 0x8e, 0xbd, - 0x7b, 0xf3, 0x1c, 0x80, 0xcb, 0x70, 0xb5, 0xac, 0xc6, 0xea, 0x66, 0xae, 0x28, 0x08, 0xd6, 0xb0, - 0xec, 0x7f, 0x5d, 0x80, 0x11, 0x99, 0xec, 0xa1, 0xed, 0xf7, 0x23, 0x1c, 0x38, 0x52, 0xf6, 0x37, - 0x96, 0x06, 0x9c, 0x12, 0xae, 0x25, 0x32, 0x95, 0x24, 0x0d, 0xb8, 0x04, 0xe0, 0x04, 0x87, 0xee, - 0xa2, 0xa8, 0xbd, 0xce, 0xd0, 0x53, 0x6e, 0x5f, 0x75, 0x5e, 0x8c, 0x25, 0x1c, 0x7d, 0x1a, 0x26, - 0x79, 0xbd, 0x30, 0x68, 0x39, 0x9b, 0x5c, 0x22, 0x3d, 0xa8, 0xfc, 0xb5, 0x27, 0x57, 0x52, 0xb0, - 0xc3, 0x83, 0xd9, 0xd3, 0xe9, 0x32, 0xa6, 0x6a, 0xe9, 0xa0, 0xc2, 0xcc, 0x37, 0x78, 0x23, 0x74, - 0xf7, 0x77, 0x58, 0x7d, 0x24, 0x20, 0xac, 0xe3, 0xd9, 0x9f, 0x07, 0xd4, 0x99, 0xf6, 0x02, 0xbd, - 0xc1, 0x6d, 0xf6, 0xbc, 0x90, 0xb8, 0xdd, 0x54, 0x2f, 0xba, 0x57, 0xb2, 0xf4, 0xd8, 0xe0, 0xb5, - 0xb0, 0xaa, 0x6f, 0xff, 0xff, 0x45, 0x98, 0x4c, 0xfb, 0xa8, 0xa2, 0xeb, 0x30, 0xc4, 0x59, 0x0f, - 0x41, 0xbe, 0x8b, 0x66, 0x5f, 0xf3, 0x6c, 0x65, 0x87, 0xb0, 0xe0, 0x5e, 0x44, 0x7d, 0xf4, 0x36, - 0x8c, 0xb8, 0xc1, 0x9e, 0xbf, 0xe7, 0x84, 0xee, 0x7c, 0xad, 0x2a, 0x96, 0x73, 0xe6, 0x6b, 0xa9, - 0x92, 0xa0, 0xe9, 0xde, 0xb2, 0x4c, 0x8b, 0x95, 0x80, 0xb0, 0x4e, 0x0e, 0xad, 0xb1, 0x28, 0xbd, - 0x1b, 0xde, 0xe6, 0x8a, 0xd3, 0xea, 0x66, 0xc0, 0xbd, 0x28, 0x91, 0x34, 0xca, 0x63, 0x22, 0x94, - 0x2f, 0x07, 0xe0, 0x84, 0x10, 0xfa, 0x36, 0x38, 0x15, 0xe5, 0x08, 0x58, 0xf3, 0xb2, 0x20, 0x75, - 0x93, 0x39, 0x2e, 0x3c, 0x42, 0xdf, 0xb1, 0x59, 0xa2, 0xd8, 0xac, 0x66, 0xec, 0x5f, 0x3b, 0x05, - 0xc6, 0x26, 0x36, 0x92, 0xe2, 0x59, 0xc7, 0x94, 0x14, 0x0f, 0x43, 0x89, 0xec, 0xb4, 0xe2, 0xfd, - 0x8a, 0x17, 0x76, 0xcb, 0xaa, 0xba, 0x24, 0x70, 0x3a, 0x69, 0x4a, 0x08, 0x56, 0x74, 0xb2, 0x33, - 0x17, 0x16, 0xbf, 0x8e, 0x99, 0x0b, 0x07, 0x4e, 0x30, 0x73, 0xe1, 0x2a, 0x0c, 0x6f, 0x7a, 0x31, - 0x26, 0xad, 0x40, 0x30, 0xfd, 0x99, 0xeb, 0xf0, 0x1a, 0x47, 0xe9, 0xcc, 0x91, 0x25, 0x00, 0x58, - 0x12, 0x41, 0x6f, 0xa8, 0x1d, 0x38, 0x94, 0xff, 0x66, 0xee, 0x54, 0x41, 0x67, 0xee, 0x41, 0x91, - 0x9f, 0x70, 0xf8, 0x41, 0xf3, 0x13, 0x2e, 0xcb, 0xac, 0x82, 0xa5, 0x7c, 0x6f, 0x0b, 0x96, 0x34, - 0xb0, 0x47, 0x2e, 0xc1, 0x3b, 0x7a, 0x26, 0xc6, 0x72, 0xfe, 0x49, 0xa0, 0x92, 0x2c, 0xf6, 0x99, - 0x7f, 0xf1, 0x7b, 0x2c, 0x38, 0xd3, 0xca, 0x4a, 0x4a, 0x2a, 0xb4, 0xb5, 0x2f, 0xf6, 0x9d, 0x75, - 0xd5, 0x68, 0x90, 0x09, 0x6a, 0x32, 0xd1, 0x70, 0x76, 0x73, 0x74, 0xa0, 0xc3, 0x75, 0x57, 0x24, - 0x10, 0x7c, 0x3c, 0x27, 0x91, 0x63, 0x97, 0xf4, 0x8d, 0x6b, 0x19, 0x49, 0x03, 0x3f, 0x9a, 0x97, - 0x34, 0xb0, 0xef, 0x54, 0x81, 0x6f, 0xa8, 0x14, 0x8e, 0x63, 0xf9, 0x4b, 0x89, 0x27, 0x68, 0xec, - 0x99, 0xb8, 0xf1, 0x0d, 0x95, 0xb8, 0xb1, 0x4b, 0x1c, 0x49, 0x9e, 0x96, 0xb1, 0x67, 0xba, 0x46, - 0x2d, 0xe5, 0xe2, 0xc4, 0xf1, 0xa4, 0x5c, 0x34, 0xae, 0x1a, 0x9e, 0xf5, 0xef, 0xa9, 0x1e, 0x57, - 0x8d, 0x41, 0xb7, 0xfb, 0x65, 0xc3, 0xd3, 0x4b, 0x4e, 0x3d, 0x50, 0x7a, 0xc9, 0x3b, 0x7a, 0xba, - 0x46, 0xd4, 0x23, 0x1f, 0x21, 0x45, 0xea, 0x33, 0x49, 0xe3, 0x1d, 0xfd, 0x02, 0x3c, 0x95, 0x4f, - 0x57, 0xdd, 0x73, 0x9d, 0x74, 0x33, 0xaf, 0xc0, 0x8e, 0xe4, 0x8f, 0xa7, 0x4f, 0x26, 0xf9, 0xe3, - 0x99, 0x63, 0x4f, 0xfe, 0x78, 0xf6, 0x04, 0x92, 0x3f, 0x3e, 0x72, 0x82, 0xc9, 0x1f, 0xef, 0x30, - 0x13, 0x07, 0x1e, 0x8e, 0x44, 0xc4, 0xbd, 0xcc, 0x8e, 0xb1, 0x98, 0x15, 0xb3, 0x84, 0x7f, 0x9c, - 0x02, 0xe1, 0x84, 0x54, 0x46, 0x52, 0xc9, 0xe9, 0x87, 0x90, 0x54, 0x72, 0x35, 0x49, 0x2a, 0x79, - 0x2e, 0x7f, 0xaa, 0x33, 0x4c, 0xcb, 0x73, 0x52, 0x49, 0xde, 0xd1, 0x53, 0x40, 0x3e, 0xda, 0x45, - 0x14, 0x9f, 0x25, 0x78, 0xec, 0x92, 0xf8, 0xf1, 0x75, 0x9e, 0xf8, 0xf1, 0x7c, 0xfe, 0x49, 0x9e, - 0xbe, 0xee, 0xcc, 0x74, 0x8f, 0xdf, 0x57, 0x80, 0x8b, 0xdd, 0xf7, 0x45, 0x22, 0xf5, 0xac, 0x25, - 0x1a, 0xc1, 0x94, 0xd4, 0x93, 0xbf, 0xad, 0x12, 0xac, 0xbe, 0x23, 0x55, 0x5d, 0x83, 0x29, 0x65, - 0x3b, 0xde, 0xf4, 0x1a, 0xfb, 0x5a, 0x86, 0x7b, 0xe5, 0x6f, 0x5b, 0x4f, 0x23, 0xe0, 0xce, 0x3a, - 0x68, 0x1e, 0x26, 0x8c, 0xc2, 0x6a, 0x45, 0xbc, 0xa1, 0x94, 0x98, 0xb5, 0x6e, 0x82, 0x71, 0x1a, - 0xdf, 0xfe, 0x69, 0x0b, 0x1e, 0xc9, 0xc9, 0xab, 0xd4, 0x77, 0x20, 0xa6, 0x0d, 0x98, 0x68, 0x99, - 0x55, 0x7b, 0xc4, 0x6b, 0x33, 0xb2, 0x37, 0xa9, 0xbe, 0xa6, 0x00, 0x38, 0x4d, 0xd4, 0xfe, 0x53, - 0x0b, 0x2e, 0x74, 0x35, 0xe3, 0x42, 0x18, 0xce, 0x6e, 0xee, 0x44, 0xce, 0x62, 0x48, 0x5c, 0xe2, - 0xc7, 0x9e, 0xd3, 0xac, 0xb7, 0x48, 0x43, 0x93, 0x5b, 0x33, 0x7b, 0xa8, 0x6b, 0x2b, 0xf5, 0xf9, - 0x4e, 0x0c, 0x9c, 0x53, 0x13, 0x2d, 0x03, 0xea, 0x84, 0x88, 0x19, 0x66, 0x31, 0x5d, 0x3b, 0xe9, - 0xe1, 0x8c, 0x1a, 0xe8, 0x25, 0x18, 0x53, 0xe6, 0x61, 0xda, 0x8c, 0xb3, 0x03, 0x18, 0xeb, 0x00, - 0x6c, 0xe2, 0x2d, 0x5c, 0xf9, 0x8d, 0xdf, 0xbb, 0xf8, 0x91, 0xdf, 0xfa, 0xbd, 0x8b, 0x1f, 0xf9, - 0xed, 0xdf, 0xbb, 0xf8, 0x91, 0xef, 0xb8, 0x7f, 0xd1, 0xfa, 0x8d, 0xfb, 0x17, 0xad, 0xdf, 0xba, - 0x7f, 0xd1, 0xfa, 0xed, 0xfb, 0x17, 0xad, 0xdf, 0xbd, 0x7f, 0xd1, 0xfa, 0xd2, 0xef, 0x5f, 0xfc, - 0xc8, 0x5b, 0x85, 0xdd, 0x67, 0xff, 0x4f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x5e, 0x40, 0x10, 0x5c, - 0xb3, 0xfc, 0x00, 0x00, + // 13889 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xec, 0xbd, 0x6b, 0x70, 0x24, 0xd7, + 0x75, 0x18, 0xac, 0x9e, 0xc1, 0x63, 0xe6, 0xe0, 0x7d, 0xb1, 0xbb, 0xc4, 0x82, 0xbb, 0x8b, 0x65, + 0xaf, 0xb4, 0x5c, 0x8a, 0x24, 0x56, 0x7c, 0x89, 0x34, 0x49, 0xd1, 0x02, 0x30, 0xc0, 0xee, 0x70, + 0x17, 0xd8, 0xe1, 0x1d, 0xec, 0xae, 0x44, 0x53, 0xfa, 0xd4, 0x98, 0xb9, 0x00, 0x9a, 0x98, 0xe9, + 0x1e, 0x76, 0xf7, 0x60, 0x17, 0xfc, 0xe4, 0xfa, 0xfc, 0xc9, 0x4f, 0xf9, 0x91, 0x52, 0xa5, 0x5c, + 0x79, 0xd8, 0x2e, 0x57, 0xca, 0x71, 0xca, 0x56, 0x9c, 0xa4, 0xe2, 0xd8, 0xb1, 0x1d, 0xcb, 0x89, + 0x9d, 0x38, 0x0f, 0x27, 0x3f, 0x1c, 0xc7, 0x95, 0x44, 0xae, 0x72, 0x05, 0xb1, 0xd7, 0xa9, 0xb8, + 0xf4, 0x23, 0xb6, 0x13, 0x3b, 0x3f, 0x82, 0xb8, 0xe2, 0xd4, 0x7d, 0xf6, 0xbd, 0x3d, 0xdd, 0x33, + 0x83, 0x25, 0x00, 0x51, 0x2a, 0xfe, 0x9b, 0xb9, 0xe7, 0xdc, 0x73, 0x6f, 0xdf, 0xe7, 0xb9, 0xe7, + 0x09, 0xaf, 0xec, 0xbc, 0x14, 0xce, 0xbb, 0xfe, 0xd5, 0x9d, 0xf6, 0x06, 0x09, 0x3c, 0x12, 0x91, + 0xf0, 0xea, 0x2e, 0xf1, 0xea, 0x7e, 0x70, 0x55, 0x00, 0x9c, 0x96, 0x7b, 0xb5, 0xe6, 0x07, 0xe4, + 0xea, 0xee, 0x33, 0x57, 0xb7, 0x88, 0x47, 0x02, 0x27, 0x22, 0xf5, 0xf9, 0x56, 0xe0, 0x47, 0x3e, + 0x42, 0x1c, 0x67, 0xde, 0x69, 0xb9, 0xf3, 0x14, 0x67, 0x7e, 0xf7, 0x99, 0xd9, 0xa7, 0xb7, 0xdc, + 0x68, 0xbb, 0xbd, 0x31, 0x5f, 0xf3, 0x9b, 0x57, 0xb7, 0xfc, 0x2d, 0xff, 0x2a, 0x43, 0xdd, 0x68, + 0x6f, 0xb2, 0x7f, 0xec, 0x0f, 0xfb, 0xc5, 0x49, 0xcc, 0x3e, 0x1f, 0x37, 0xd3, 0x74, 0x6a, 0xdb, + 0xae, 0x47, 0x82, 0xbd, 0xab, 0xad, 0x9d, 0x2d, 0xd6, 0x6e, 0x40, 0x42, 0xbf, 0x1d, 0xd4, 0x48, + 0xb2, 0xe1, 0xae, 0xb5, 0xc2, 0xab, 0x4d, 0x12, 0x39, 0x29, 0xdd, 0x9d, 0xbd, 0x9a, 0x55, 0x2b, + 0x68, 0x7b, 0x91, 0xdb, 0xec, 0x6c, 0xe6, 0xe3, 0xbd, 0x2a, 0x84, 0xb5, 0x6d, 0xd2, 0x74, 0x3a, + 0xea, 0x3d, 0x97, 0x55, 0xaf, 0x1d, 0xb9, 0x8d, 0xab, 0xae, 0x17, 0x85, 0x51, 0x90, 0xac, 0x64, + 0x7f, 0xd5, 0x82, 0x8b, 0x0b, 0x77, 0xab, 0xcb, 0x0d, 0x27, 0x8c, 0xdc, 0xda, 0x62, 0xc3, 0xaf, + 0xed, 0x54, 0x23, 0x3f, 0x20, 0x77, 0xfc, 0x46, 0xbb, 0x49, 0xaa, 0x6c, 0x20, 0xd0, 0x53, 0x50, + 0xd8, 0x65, 0xff, 0xcb, 0xa5, 0x19, 0xeb, 0xa2, 0x75, 0xa5, 0xb8, 0x38, 0xf9, 0x1b, 0xfb, 0x73, + 0x1f, 0x7a, 0xb0, 0x3f, 0x57, 0xb8, 0x23, 0xca, 0xb1, 0xc2, 0x40, 0x97, 0x61, 0x68, 0x33, 0x5c, + 0xdf, 0x6b, 0x91, 0x99, 0x1c, 0xc3, 0x1d, 0x17, 0xb8, 0x43, 0x2b, 0x55, 0x5a, 0x8a, 0x05, 0x14, + 0x5d, 0x85, 0x62, 0xcb, 0x09, 0x22, 0x37, 0x72, 0x7d, 0x6f, 0x26, 0x7f, 0xd1, 0xba, 0x32, 0xb8, + 0x38, 0x25, 0x50, 0x8b, 0x15, 0x09, 0xc0, 0x31, 0x0e, 0xed, 0x46, 0x40, 0x9c, 0xfa, 0x2d, 0xaf, + 0xb1, 0x37, 0x33, 0x70, 0xd1, 0xba, 0x52, 0x88, 0xbb, 0x81, 0x45, 0x39, 0x56, 0x18, 0xf6, 0x8f, + 0xe4, 0xa0, 0xb0, 0xb0, 0xb9, 0xe9, 0x7a, 0x6e, 0xb4, 0x87, 0xee, 0xc0, 0xa8, 0xe7, 0xd7, 0x89, + 0xfc, 0xcf, 0xbe, 0x62, 0xe4, 0xd9, 0x8b, 0xf3, 0x9d, 0x4b, 0x69, 0x7e, 0x4d, 0xc3, 0x5b, 0x9c, + 0x7c, 0xb0, 0x3f, 0x37, 0xaa, 0x97, 0x60, 0x83, 0x0e, 0xc2, 0x30, 0xd2, 0xf2, 0xeb, 0x8a, 0x6c, + 0x8e, 0x91, 0x9d, 0x4b, 0x23, 0x5b, 0x89, 0xd1, 0x16, 0x27, 0x1e, 0xec, 0xcf, 0x8d, 0x68, 0x05, + 0x58, 0x27, 0x82, 0x36, 0x60, 0x82, 0xfe, 0xf5, 0x22, 0x57, 0xd1, 0xcd, 0x33, 0xba, 0x97, 0xb2, + 0xe8, 0x6a, 0xa8, 0x8b, 0xd3, 0x0f, 0xf6, 0xe7, 0x26, 0x12, 0x85, 0x38, 0x49, 0xd0, 0x7e, 0x17, + 0xc6, 0x17, 0xa2, 0xc8, 0xa9, 0x6d, 0x93, 0x3a, 0x9f, 0x41, 0xf4, 0x3c, 0x0c, 0x78, 0x4e, 0x93, + 0x88, 0xf9, 0xbd, 0x28, 0x06, 0x76, 0x60, 0xcd, 0x69, 0x92, 0x83, 0xfd, 0xb9, 0xc9, 0xdb, 0x9e, + 0xfb, 0x4e, 0x5b, 0xac, 0x0a, 0x5a, 0x86, 0x19, 0x36, 0x7a, 0x16, 0xa0, 0x4e, 0x76, 0xdd, 0x1a, + 0xa9, 0x38, 0xd1, 0xb6, 0x98, 0x6f, 0x24, 0xea, 0x42, 0x49, 0x41, 0xb0, 0x86, 0x65, 0xdf, 0x87, + 0xe2, 0xc2, 0xae, 0xef, 0xd6, 0x2b, 0x7e, 0x3d, 0x44, 0x3b, 0x30, 0xd1, 0x0a, 0xc8, 0x26, 0x09, + 0x54, 0xd1, 0x8c, 0x75, 0x31, 0x7f, 0x65, 0xe4, 0xd9, 0x2b, 0xa9, 0x1f, 0x6b, 0xa2, 0x2e, 0x7b, + 0x51, 0xb0, 0xb7, 0xf8, 0x88, 0x68, 0x6f, 0x22, 0x01, 0xc5, 0x49, 0xca, 0xf6, 0x3f, 0xcf, 0xc1, + 0xe9, 0x85, 0x77, 0xdb, 0x01, 0x29, 0xb9, 0xe1, 0x4e, 0x72, 0x85, 0xd7, 0xdd, 0x70, 0x67, 0x2d, + 0x1e, 0x01, 0xb5, 0xb4, 0x4a, 0xa2, 0x1c, 0x2b, 0x0c, 0xf4, 0x34, 0x0c, 0xd3, 0xdf, 0xb7, 0x71, + 0x59, 0x7c, 0xf2, 0xb4, 0x40, 0x1e, 0x29, 0x39, 0x91, 0x53, 0xe2, 0x20, 0x2c, 0x71, 0xd0, 0x2a, + 0x8c, 0xd4, 0xd8, 0x86, 0xdc, 0x5a, 0xf5, 0xeb, 0x84, 0x4d, 0x66, 0x71, 0xf1, 0x49, 0x8a, 0xbe, + 0x14, 0x17, 0x1f, 0xec, 0xcf, 0xcd, 0xf0, 0xbe, 0x09, 0x12, 0x1a, 0x0c, 0xeb, 0xf5, 0x91, 0xad, + 0xf6, 0xd7, 0x00, 0xa3, 0x04, 0x29, 0x7b, 0xeb, 0x8a, 0xb6, 0x55, 0x06, 0xd9, 0x56, 0x19, 0x4d, + 0xdf, 0x26, 0xe8, 0x19, 0x18, 0xd8, 0x71, 0xbd, 0xfa, 0xcc, 0x10, 0xa3, 0x75, 0x9e, 0xce, 0xf9, + 0x0d, 0xd7, 0xab, 0x1f, 0xec, 0xcf, 0x4d, 0x19, 0xdd, 0xa1, 0x85, 0x98, 0xa1, 0xda, 0x7f, 0x6a, + 0xc1, 0x1c, 0x83, 0xad, 0xb8, 0x0d, 0x52, 0x21, 0x41, 0xe8, 0x86, 0x11, 0xf1, 0x22, 0x63, 0x40, + 0x9f, 0x05, 0x08, 0x49, 0x2d, 0x20, 0x91, 0x36, 0xa4, 0x6a, 0x61, 0x54, 0x15, 0x04, 0x6b, 0x58, + 0xf4, 0x40, 0x08, 0xb7, 0x9d, 0x80, 0xad, 0x2f, 0x31, 0xb0, 0xea, 0x40, 0xa8, 0x4a, 0x00, 0x8e, + 0x71, 0x8c, 0x03, 0x21, 0xdf, 0xeb, 0x40, 0x40, 0x9f, 0x80, 0x89, 0xb8, 0xb1, 0xb0, 0xe5, 0xd4, + 0xe4, 0x00, 0xb2, 0x2d, 0x53, 0x35, 0x41, 0x38, 0x89, 0x6b, 0xff, 0x6d, 0x4b, 0x2c, 0x1e, 0xfa, + 0xd5, 0xef, 0xf3, 0x6f, 0xb5, 0x7f, 0xc9, 0x82, 0xe1, 0x45, 0xd7, 0xab, 0xbb, 0xde, 0x16, 0xfa, + 0x1c, 0x14, 0xe8, 0xdd, 0x54, 0x77, 0x22, 0x47, 0x9c, 0x7b, 0x1f, 0xd3, 0xf6, 0x96, 0xba, 0x2a, + 0xe6, 0x5b, 0x3b, 0x5b, 0xb4, 0x20, 0x9c, 0xa7, 0xd8, 0x74, 0xb7, 0xdd, 0xda, 0x78, 0x9b, 0xd4, + 0xa2, 0x55, 0x12, 0x39, 0xf1, 0xe7, 0xc4, 0x65, 0x58, 0x51, 0x45, 0x37, 0x60, 0x28, 0x72, 0x82, + 0x2d, 0x12, 0x89, 0x03, 0x30, 0xf5, 0xa0, 0xe2, 0x35, 0x31, 0xdd, 0x91, 0xc4, 0xab, 0x91, 0xf8, + 0x5a, 0x58, 0x67, 0x55, 0xb1, 0x20, 0x61, 0xff, 0xd0, 0x30, 0x9c, 0x5d, 0xaa, 0x96, 0x33, 0xd6, + 0xd5, 0x65, 0x18, 0xaa, 0x07, 0xee, 0x2e, 0x09, 0xc4, 0x38, 0x2b, 0x2a, 0x25, 0x56, 0x8a, 0x05, + 0x14, 0xbd, 0x04, 0xa3, 0xfc, 0x42, 0xba, 0xee, 0x78, 0xf5, 0x86, 0x1c, 0xe2, 0x53, 0x02, 0x7b, + 0xf4, 0x8e, 0x06, 0xc3, 0x06, 0xe6, 0x21, 0x17, 0xd5, 0xe5, 0xc4, 0x66, 0xcc, 0xba, 0xec, 0xbe, + 0x68, 0xc1, 0x24, 0x6f, 0x66, 0x21, 0x8a, 0x02, 0x77, 0xa3, 0x1d, 0x91, 0x70, 0x66, 0x90, 0x9d, + 0x74, 0x4b, 0x69, 0xa3, 0x95, 0x39, 0x02, 0xf3, 0x77, 0x12, 0x54, 0xf8, 0x21, 0x38, 0x23, 0xda, + 0x9d, 0x4c, 0x82, 0x71, 0x47, 0xb3, 0xe8, 0x3b, 0x2d, 0x98, 0xad, 0xf9, 0x5e, 0x14, 0xf8, 0x8d, + 0x06, 0x09, 0x2a, 0xed, 0x8d, 0x86, 0x1b, 0x6e, 0xf3, 0x75, 0x8a, 0xc9, 0x26, 0x3b, 0x09, 0x32, + 0xe6, 0x50, 0x21, 0x89, 0x39, 0xbc, 0xf0, 0x60, 0x7f, 0x6e, 0x76, 0x29, 0x93, 0x14, 0xee, 0xd2, + 0x0c, 0xda, 0x01, 0x44, 0xaf, 0xd2, 0x6a, 0xe4, 0x6c, 0x91, 0xb8, 0xf1, 0xe1, 0xfe, 0x1b, 0x3f, + 0xf3, 0x60, 0x7f, 0x0e, 0xad, 0x75, 0x90, 0xc0, 0x29, 0x64, 0xd1, 0x3b, 0x70, 0x8a, 0x96, 0x76, + 0x7c, 0x6b, 0xa1, 0xff, 0xe6, 0x66, 0x1e, 0xec, 0xcf, 0x9d, 0x5a, 0x4b, 0x21, 0x82, 0x53, 0x49, + 0xa3, 0xef, 0xb0, 0xe0, 0x6c, 0xfc, 0xf9, 0xcb, 0xf7, 0x5b, 0x8e, 0x57, 0x8f, 0x1b, 0x2e, 0xf6, + 0xdf, 0x30, 0x3d, 0x93, 0xcf, 0x2e, 0x65, 0x51, 0xc2, 0xd9, 0x8d, 0xcc, 0x2e, 0xc1, 0xe9, 0xd4, + 0xd5, 0x82, 0x26, 0x21, 0xbf, 0x43, 0x38, 0x17, 0x54, 0xc4, 0xf4, 0x27, 0x3a, 0x05, 0x83, 0xbb, + 0x4e, 0xa3, 0x2d, 0x36, 0x0a, 0xe6, 0x7f, 0x5e, 0xce, 0xbd, 0x64, 0xd9, 0xff, 0x22, 0x0f, 0x13, + 0x4b, 0xd5, 0xf2, 0x43, 0xed, 0x42, 0xfd, 0x1a, 0xca, 0x75, 0xbd, 0x86, 0xe2, 0x4b, 0x2d, 0x9f, + 0x79, 0xa9, 0xfd, 0x7f, 0x29, 0x5b, 0x68, 0x80, 0x6d, 0xa1, 0x6f, 0xc9, 0xd8, 0x42, 0x47, 0xbc, + 0x71, 0x76, 0x33, 0x56, 0xd1, 0x20, 0x9b, 0xcc, 0x54, 0x8e, 0xe5, 0xa6, 0x5f, 0x73, 0x1a, 0xc9, + 0xa3, 0xef, 0x90, 0x4b, 0xe9, 0x68, 0xe6, 0xb1, 0x06, 0xa3, 0x4b, 0x4e, 0xcb, 0xd9, 0x70, 0x1b, + 0x6e, 0xe4, 0x92, 0x10, 0x3d, 0x0e, 0x79, 0xa7, 0x5e, 0x67, 0xdc, 0x56, 0x71, 0xf1, 0xf4, 0x83, + 0xfd, 0xb9, 0xfc, 0x42, 0x9d, 0x5e, 0xfb, 0xa0, 0xb0, 0xf6, 0x30, 0xc5, 0x40, 0x1f, 0x85, 0x81, + 0x7a, 0xe0, 0xb7, 0x66, 0x72, 0x0c, 0x93, 0xee, 0xba, 0x81, 0x52, 0xe0, 0xb7, 0x12, 0xa8, 0x0c, + 0xc7, 0xfe, 0xb5, 0x1c, 0x9c, 0x5b, 0x22, 0xad, 0xed, 0x95, 0x6a, 0xc6, 0xf9, 0x7d, 0x05, 0x0a, + 0x4d, 0xdf, 0x73, 0x23, 0x3f, 0x08, 0x45, 0xd3, 0x6c, 0x45, 0xac, 0x8a, 0x32, 0xac, 0xa0, 0xe8, + 0x22, 0x0c, 0xb4, 0x62, 0xa6, 0x72, 0x54, 0x32, 0xa4, 0x8c, 0x9d, 0x64, 0x10, 0x8a, 0xd1, 0x0e, + 0x49, 0x20, 0x56, 0x8c, 0xc2, 0xb8, 0x1d, 0x92, 0x00, 0x33, 0x48, 0x7c, 0x33, 0xd3, 0x3b, 0x5b, + 0x9c, 0xd0, 0x89, 0x9b, 0x99, 0x42, 0xb0, 0x86, 0x85, 0x2a, 0x50, 0x0c, 0x13, 0x33, 0xdb, 0xd7, + 0x36, 0x1d, 0x63, 0x57, 0xb7, 0x9a, 0xc9, 0x98, 0x88, 0x71, 0xa3, 0x0c, 0xf5, 0xbc, 0xba, 0xbf, + 0x92, 0x03, 0xc4, 0x87, 0xf0, 0x1b, 0x6c, 0xe0, 0x6e, 0x77, 0x0e, 0x5c, 0xff, 0x5b, 0xe2, 0xa8, + 0x46, 0xef, 0xcf, 0x2c, 0x38, 0xb7, 0xe4, 0x7a, 0x75, 0x12, 0x64, 0x2c, 0xc0, 0xe3, 0x79, 0xcb, + 0x1e, 0x8e, 0x69, 0x30, 0x96, 0xd8, 0xc0, 0x11, 0x2c, 0x31, 0xfb, 0x8f, 0x2d, 0x40, 0xfc, 0xb3, + 0xdf, 0x77, 0x1f, 0x7b, 0xbb, 0xf3, 0x63, 0x8f, 0x60, 0x59, 0xd8, 0x37, 0x61, 0x7c, 0xa9, 0xe1, + 0x12, 0x2f, 0x2a, 0x57, 0x96, 0x7c, 0x6f, 0xd3, 0xdd, 0x42, 0x2f, 0xc3, 0x78, 0xe4, 0x36, 0x89, + 0xdf, 0x8e, 0xaa, 0xa4, 0xe6, 0x7b, 0xec, 0x25, 0x69, 0x5d, 0x19, 0x5c, 0x44, 0x0f, 0xf6, 0xe7, + 0xc6, 0xd7, 0x0d, 0x08, 0x4e, 0x60, 0xda, 0xbf, 0x4b, 0xc7, 0xcf, 0x6f, 0xb6, 0x7c, 0x8f, 0x78, + 0xd1, 0x92, 0xef, 0xd5, 0xb9, 0xc4, 0xe1, 0x65, 0x18, 0x88, 0xe8, 0x78, 0xf0, 0xb1, 0xbb, 0x2c, + 0x37, 0x0a, 0x1d, 0x85, 0x83, 0xfd, 0xb9, 0x33, 0x9d, 0x35, 0xd8, 0x38, 0xb1, 0x3a, 0xe8, 0x5b, + 0x60, 0x28, 0x8c, 0x9c, 0xa8, 0x1d, 0x8a, 0xd1, 0x7c, 0x4c, 0x8e, 0x66, 0x95, 0x95, 0x1e, 0xec, + 0xcf, 0x4d, 0xa8, 0x6a, 0xbc, 0x08, 0x8b, 0x0a, 0xe8, 0x09, 0x18, 0x6e, 0x92, 0x30, 0x74, 0xb6, + 0xe4, 0x6d, 0x38, 0x21, 0xea, 0x0e, 0xaf, 0xf2, 0x62, 0x2c, 0xe1, 0xe8, 0x12, 0x0c, 0x92, 0x20, + 0xf0, 0x03, 0xb1, 0x47, 0xc7, 0x04, 0xe2, 0xe0, 0x32, 0x2d, 0xc4, 0x1c, 0x66, 0xff, 0x5b, 0x0b, + 0x26, 0x54, 0x5f, 0x79, 0x5b, 0x27, 0xf0, 0x2a, 0x78, 0x13, 0xa0, 0x26, 0x3f, 0x30, 0x64, 0xb7, + 0xc7, 0xc8, 0xb3, 0x97, 0x53, 0x2f, 0xea, 0x8e, 0x61, 0x8c, 0x29, 0xab, 0xa2, 0x10, 0x6b, 0xd4, + 0xec, 0x7f, 0x6c, 0xc1, 0x74, 0xe2, 0x8b, 0x6e, 0xba, 0x61, 0x84, 0xde, 0xea, 0xf8, 0xaa, 0xf9, + 0xfe, 0xbe, 0x8a, 0xd6, 0x66, 0xdf, 0xa4, 0x96, 0xb2, 0x2c, 0xd1, 0xbe, 0xe8, 0x3a, 0x0c, 0xba, + 0x11, 0x69, 0xca, 0x8f, 0xb9, 0xd4, 0xf5, 0x63, 0x78, 0xaf, 0xe2, 0x19, 0x29, 0xd3, 0x9a, 0x98, + 0x13, 0xb0, 0x7f, 0x2d, 0x0f, 0x45, 0xbe, 0x6c, 0x57, 0x9d, 0xd6, 0x09, 0xcc, 0xc5, 0x93, 0x50, + 0x74, 0x9b, 0xcd, 0x76, 0xe4, 0x6c, 0x88, 0xe3, 0xbc, 0xc0, 0xb7, 0x56, 0x59, 0x16, 0xe2, 0x18, + 0x8e, 0xca, 0x30, 0xc0, 0xba, 0xc2, 0xbf, 0xf2, 0xf1, 0xf4, 0xaf, 0x14, 0x7d, 0x9f, 0x2f, 0x39, + 0x91, 0xc3, 0x39, 0x29, 0x75, 0x8f, 0xd0, 0x22, 0xcc, 0x48, 0x20, 0x07, 0x60, 0xc3, 0xf5, 0x9c, + 0x60, 0x8f, 0x96, 0xcd, 0xe4, 0x19, 0xc1, 0xa7, 0xbb, 0x13, 0x5c, 0x54, 0xf8, 0x9c, 0xac, 0xfa, + 0xb0, 0x18, 0x80, 0x35, 0xa2, 0xb3, 0x2f, 0x42, 0x51, 0x21, 0x1f, 0x86, 0x21, 0x9a, 0xfd, 0x04, + 0x4c, 0x24, 0xda, 0xea, 0x55, 0x7d, 0x54, 0xe7, 0xa7, 0x7e, 0x99, 0x1d, 0x19, 0xa2, 0xd7, 0xcb, + 0xde, 0xae, 0x38, 0x72, 0xdf, 0x85, 0x53, 0x8d, 0x94, 0x93, 0x4c, 0xcc, 0x6b, 0xff, 0x27, 0xdf, + 0x39, 0xf1, 0xd9, 0xa7, 0xd2, 0xa0, 0x38, 0xb5, 0x0d, 0xca, 0x23, 0xf8, 0x2d, 0xba, 0x41, 0x9c, + 0x86, 0xce, 0x6e, 0xdf, 0x12, 0x65, 0x58, 0x41, 0xe9, 0x79, 0x77, 0x4a, 0x75, 0xfe, 0x06, 0xd9, + 0xab, 0x92, 0x06, 0xa9, 0x45, 0x7e, 0xf0, 0x75, 0xed, 0xfe, 0x79, 0x3e, 0xfa, 0xfc, 0xb8, 0x1c, + 0x11, 0x04, 0xf2, 0x37, 0xc8, 0x1e, 0x9f, 0x0a, 0xfd, 0xeb, 0xf2, 0x5d, 0xbf, 0xee, 0x67, 0x2d, + 0x18, 0x53, 0x5f, 0x77, 0x02, 0xe7, 0xc2, 0xa2, 0x79, 0x2e, 0x9c, 0xef, 0xba, 0xc0, 0x33, 0x4e, + 0x84, 0xaf, 0xe4, 0xe0, 0xac, 0xc2, 0xa1, 0x6f, 0x03, 0xfe, 0x47, 0xac, 0xaa, 0xab, 0x50, 0xf4, + 0x94, 0xd4, 0xca, 0x32, 0xc5, 0x45, 0xb1, 0xcc, 0x2a, 0xc6, 0xa1, 0x2c, 0x9e, 0x17, 0x8b, 0x96, + 0x46, 0x75, 0x71, 0xae, 0x10, 0xdd, 0x2e, 0x42, 0xbe, 0xed, 0xd6, 0xc5, 0x05, 0xf3, 0x31, 0x39, + 0xda, 0xb7, 0xcb, 0xa5, 0x83, 0xfd, 0xb9, 0xc7, 0xb2, 0x54, 0x09, 0xf4, 0x66, 0x0b, 0xe7, 0x6f, + 0x97, 0x4b, 0x98, 0x56, 0x46, 0x0b, 0x30, 0x21, 0xb5, 0x25, 0x77, 0x28, 0xbb, 0xe5, 0x7b, 0xe2, + 0x1e, 0x52, 0x32, 0x59, 0x6c, 0x82, 0x71, 0x12, 0x1f, 0x95, 0x60, 0x72, 0xa7, 0xbd, 0x41, 0x1a, + 0x24, 0xe2, 0x1f, 0x7c, 0x83, 0x70, 0x89, 0x65, 0x31, 0x7e, 0x99, 0xdd, 0x48, 0xc0, 0x71, 0x47, + 0x0d, 0xfb, 0x2f, 0xd8, 0x7d, 0x20, 0x46, 0xaf, 0x12, 0xf8, 0x74, 0x61, 0x51, 0xea, 0x5f, 0xcf, + 0xe5, 0xdc, 0xcf, 0xaa, 0xb8, 0x41, 0xf6, 0xd6, 0x7d, 0xca, 0x99, 0xa7, 0xaf, 0x0a, 0x63, 0xcd, + 0x0f, 0x74, 0x5d, 0xf3, 0x3f, 0x9f, 0x83, 0xd3, 0x6a, 0x04, 0x0c, 0x26, 0xf0, 0x1b, 0x7d, 0x0c, + 0x9e, 0x81, 0x91, 0x3a, 0xd9, 0x74, 0xda, 0x8d, 0x48, 0x89, 0xcf, 0x07, 0xb9, 0x0a, 0xa5, 0x14, + 0x17, 0x63, 0x1d, 0xe7, 0x10, 0xc3, 0xf6, 0x3f, 0x47, 0xd8, 0x45, 0x1c, 0x39, 0x74, 0x8d, 0xab, + 0x5d, 0x63, 0x65, 0xee, 0x9a, 0x4b, 0x30, 0xe8, 0x36, 0x29, 0x63, 0x96, 0x33, 0xf9, 0xad, 0x32, + 0x2d, 0xc4, 0x1c, 0x86, 0x3e, 0x02, 0xc3, 0x35, 0xbf, 0xd9, 0x74, 0xbc, 0x3a, 0xbb, 0xf2, 0x8a, + 0x8b, 0x23, 0x94, 0x77, 0x5b, 0xe2, 0x45, 0x58, 0xc2, 0xd0, 0x39, 0x18, 0x70, 0x82, 0x2d, 0x2e, + 0xc3, 0x28, 0x2e, 0x16, 0x68, 0x4b, 0x0b, 0xc1, 0x56, 0x88, 0x59, 0x29, 0x7d, 0x82, 0xdd, 0xf3, + 0x83, 0x1d, 0xd7, 0xdb, 0x2a, 0xb9, 0x81, 0xd8, 0x12, 0xea, 0x2e, 0xbc, 0xab, 0x20, 0x58, 0xc3, + 0x42, 0x2b, 0x30, 0xd8, 0xf2, 0x83, 0x28, 0x9c, 0x19, 0x62, 0xc3, 0xfd, 0x58, 0xc6, 0x41, 0xc4, + 0xbf, 0xb6, 0xe2, 0x07, 0x51, 0xfc, 0x01, 0xf4, 0x5f, 0x88, 0x79, 0x75, 0x74, 0x13, 0x86, 0x89, + 0xb7, 0xbb, 0x12, 0xf8, 0xcd, 0x99, 0xe9, 0x6c, 0x4a, 0xcb, 0x1c, 0x85, 0x2f, 0xb3, 0x98, 0x47, + 0x15, 0xc5, 0x58, 0x92, 0x40, 0xdf, 0x02, 0x79, 0xe2, 0xed, 0xce, 0x0c, 0x33, 0x4a, 0xb3, 0x19, + 0x94, 0xee, 0x38, 0x41, 0x7c, 0xe6, 0x2f, 0x7b, 0xbb, 0x98, 0xd6, 0x41, 0x9f, 0x86, 0xa2, 0x3c, + 0x30, 0x42, 0x21, 0xac, 0x4b, 0x5d, 0xb0, 0xf2, 0x98, 0xc1, 0xe4, 0x9d, 0xb6, 0x1b, 0x90, 0x26, + 0xf1, 0xa2, 0x30, 0x3e, 0x21, 0x25, 0x34, 0xc4, 0x31, 0x35, 0xf4, 0x69, 0x29, 0x21, 0x5e, 0xf5, + 0xdb, 0x5e, 0x14, 0xce, 0x14, 0x59, 0xf7, 0x52, 0x75, 0x77, 0x77, 0x62, 0xbc, 0xa4, 0x08, 0x99, + 0x57, 0xc6, 0x06, 0x29, 0xf4, 0x19, 0x18, 0xe3, 0xff, 0xb9, 0x06, 0x2c, 0x9c, 0x39, 0xcd, 0x68, + 0x5f, 0xcc, 0xa6, 0xcd, 0x11, 0x17, 0x4f, 0x0b, 0xe2, 0x63, 0x7a, 0x69, 0x88, 0x4d, 0x6a, 0x08, + 0xc3, 0x58, 0xc3, 0xdd, 0x25, 0x1e, 0x09, 0xc3, 0x4a, 0xe0, 0x6f, 0x90, 0x19, 0x60, 0x03, 0x73, + 0x36, 0x5d, 0x63, 0xe6, 0x6f, 0x90, 0xc5, 0x29, 0x4a, 0xf3, 0xa6, 0x5e, 0x07, 0x9b, 0x24, 0xd0, + 0x6d, 0x18, 0xa7, 0x2f, 0x36, 0x37, 0x26, 0x3a, 0xd2, 0x8b, 0x28, 0x7b, 0x57, 0x61, 0xa3, 0x12, + 0x4e, 0x10, 0x41, 0xb7, 0x60, 0x34, 0x8c, 0x9c, 0x20, 0x6a, 0xb7, 0x38, 0xd1, 0x33, 0xbd, 0x88, + 0x32, 0x85, 0x6b, 0x55, 0xab, 0x82, 0x0d, 0x02, 0xe8, 0x75, 0x28, 0x36, 0xdc, 0x4d, 0x52, 0xdb, + 0xab, 0x35, 0xc8, 0xcc, 0x28, 0xa3, 0x96, 0x7a, 0xa8, 0xdc, 0x94, 0x48, 0x9c, 0xcf, 0x55, 0x7f, + 0x71, 0x5c, 0x1d, 0xdd, 0x81, 0x33, 0x11, 0x09, 0x9a, 0xae, 0xe7, 0xd0, 0xc3, 0x40, 0x3c, 0xad, + 0x98, 0x22, 0x73, 0x8c, 0xed, 0xb6, 0x0b, 0x62, 0x36, 0xce, 0xac, 0xa7, 0x62, 0xe1, 0x8c, 0xda, + 0xe8, 0x3e, 0xcc, 0xa4, 0x40, 0xfc, 0x86, 0x5b, 0xdb, 0x9b, 0x39, 0xc5, 0x28, 0xbf, 0x2a, 0x28, + 0xcf, 0xac, 0x67, 0xe0, 0x1d, 0x74, 0x81, 0xe1, 0x4c, 0xea, 0xe8, 0x16, 0x4c, 0xb0, 0x13, 0xa8, + 0xd2, 0x6e, 0x34, 0x44, 0x83, 0xe3, 0xac, 0xc1, 0x8f, 0xc8, 0xfb, 0xb8, 0x6c, 0x82, 0x0f, 0xf6, + 0xe7, 0x20, 0xfe, 0x87, 0x93, 0xb5, 0xd1, 0x06, 0xd3, 0x99, 0xb5, 0x03, 0x37, 0xda, 0xa3, 0xe7, + 0x06, 0xb9, 0x1f, 0xcd, 0x4c, 0x74, 0x95, 0x57, 0xe8, 0xa8, 0x4a, 0xb1, 0xa6, 0x17, 0xe2, 0x24, + 0x41, 0x7a, 0xa4, 0x86, 0x51, 0xdd, 0xf5, 0x66, 0x26, 0xf9, 0xbb, 0x44, 0x9e, 0x48, 0x55, 0x5a, + 0x88, 0x39, 0x8c, 0xe9, 0xcb, 0xe8, 0x8f, 0x5b, 0xf4, 0xe6, 0x9a, 0x62, 0x88, 0xb1, 0xbe, 0x4c, + 0x02, 0x70, 0x8c, 0x43, 0x99, 0xc9, 0x28, 0xda, 0x9b, 0x41, 0x0c, 0x55, 0x1d, 0x2c, 0xeb, 0xeb, + 0x9f, 0xc6, 0xb4, 0xdc, 0xde, 0x80, 0x71, 0x75, 0x10, 0xb2, 0x31, 0x41, 0x73, 0x30, 0xc8, 0xd8, + 0x27, 0x21, 0x5d, 0x2b, 0xd2, 0x2e, 0x30, 0xd6, 0x0a, 0xf3, 0x72, 0xd6, 0x05, 0xf7, 0x5d, 0xb2, + 0xb8, 0x17, 0x11, 0xfe, 0xa6, 0xcf, 0x6b, 0x5d, 0x90, 0x00, 0x1c, 0xe3, 0xd8, 0xff, 0x87, 0xb3, + 0xa1, 0xf1, 0x69, 0xdb, 0xc7, 0xfd, 0xf2, 0x14, 0x14, 0xb6, 0xfd, 0x30, 0xa2, 0xd8, 0xac, 0x8d, + 0xc1, 0x98, 0xf1, 0xbc, 0x2e, 0xca, 0xb1, 0xc2, 0x40, 0xaf, 0xc0, 0x58, 0x4d, 0x6f, 0x40, 0x5c, + 0x8e, 0xea, 0x18, 0x31, 0x5a, 0xc7, 0x26, 0x2e, 0x7a, 0x09, 0x0a, 0xcc, 0x06, 0xa4, 0xe6, 0x37, + 0x04, 0xd7, 0x26, 0x6f, 0xf8, 0x42, 0x45, 0x94, 0x1f, 0x68, 0xbf, 0xb1, 0xc2, 0x46, 0x97, 0x61, + 0x88, 0x76, 0xa1, 0x5c, 0x11, 0xd7, 0x92, 0x12, 0x14, 0x5d, 0x67, 0xa5, 0x58, 0x40, 0xed, 0xbf, + 0x9c, 0xd3, 0x46, 0x99, 0xbe, 0x87, 0x09, 0xaa, 0xc0, 0xf0, 0x3d, 0xc7, 0x8d, 0x5c, 0x6f, 0x4b, + 0xf0, 0x1f, 0x4f, 0x74, 0xbd, 0xa3, 0x58, 0xa5, 0xbb, 0xbc, 0x02, 0xbf, 0x45, 0xc5, 0x1f, 0x2c, + 0xc9, 0x50, 0x8a, 0x41, 0xdb, 0xf3, 0x28, 0xc5, 0x5c, 0xbf, 0x14, 0x31, 0xaf, 0xc0, 0x29, 0x8a, + 0x3f, 0x58, 0x92, 0x41, 0x6f, 0x01, 0xc8, 0x1d, 0x46, 0xea, 0xc2, 0xf6, 0xe2, 0xa9, 0xde, 0x44, + 0xd7, 0x55, 0x9d, 0xc5, 0x71, 0x7a, 0x47, 0xc7, 0xff, 0xb1, 0x46, 0xcf, 0x8e, 0x18, 0x9f, 0xd6, + 0xd9, 0x19, 0xf4, 0x6d, 0x74, 0x89, 0x3b, 0x41, 0x44, 0xea, 0x0b, 0x91, 0x18, 0x9c, 0x8f, 0xf6, + 0xf7, 0x48, 0x59, 0x77, 0x9b, 0x44, 0xdf, 0x0e, 0x82, 0x08, 0x8e, 0xe9, 0xd9, 0xbf, 0x98, 0x87, + 0x99, 0xac, 0xee, 0xd2, 0x45, 0x47, 0xee, 0xbb, 0xd1, 0x12, 0x65, 0xaf, 0x2c, 0x73, 0xd1, 0x2d, + 0x8b, 0x72, 0xac, 0x30, 0xe8, 0xec, 0x87, 0xee, 0x96, 0x7c, 0x63, 0x0e, 0xc6, 0xb3, 0x5f, 0x65, + 0xa5, 0x58, 0x40, 0x29, 0x5e, 0x40, 0x9c, 0x50, 0x18, 0xf7, 0x68, 0xab, 0x04, 0xb3, 0x52, 0x2c, + 0xa0, 0xba, 0xb4, 0x6b, 0xa0, 0x87, 0xb4, 0xcb, 0x18, 0xa2, 0xc1, 0xa3, 0x1d, 0x22, 0xf4, 0x59, + 0x80, 0x4d, 0xd7, 0x73, 0xc3, 0x6d, 0x46, 0x7d, 0xe8, 0xd0, 0xd4, 0x15, 0x73, 0xb6, 0xa2, 0xa8, + 0x60, 0x8d, 0x22, 0x7a, 0x01, 0x46, 0xd4, 0x06, 0x2c, 0x97, 0x98, 0xa6, 0x53, 0xb3, 0x1c, 0x89, + 0x4f, 0xa3, 0x12, 0xd6, 0xf1, 0xec, 0xb7, 0x93, 0xeb, 0x45, 0xec, 0x00, 0x6d, 0x7c, 0xad, 0x7e, + 0xc7, 0x37, 0xd7, 0x7d, 0x7c, 0xed, 0xaf, 0xe5, 0x61, 0xc2, 0x68, 0xac, 0x1d, 0xf6, 0x71, 0x66, + 0x5d, 0xa3, 0x07, 0xb8, 0x13, 0x11, 0xb1, 0xff, 0xec, 0xde, 0x5b, 0x45, 0x3f, 0xe4, 0xe9, 0x0e, + 0xe0, 0xf5, 0xd1, 0x67, 0xa1, 0xd8, 0x70, 0x42, 0x26, 0x39, 0x23, 0x62, 0xdf, 0xf5, 0x43, 0x2c, + 0x7e, 0x98, 0x38, 0x61, 0xa4, 0xdd, 0x9a, 0x9c, 0x76, 0x4c, 0x92, 0xde, 0x34, 0x94, 0x3f, 0x91, + 0xd6, 0x63, 0xaa, 0x13, 0x94, 0x89, 0xd9, 0xc3, 0x1c, 0x86, 0x5e, 0x82, 0xd1, 0x80, 0xb0, 0x55, + 0xb1, 0x44, 0xb9, 0x39, 0xb6, 0xcc, 0x06, 0x63, 0xb6, 0x0f, 0x6b, 0x30, 0x6c, 0x60, 0xc6, 0x6f, + 0x83, 0xa1, 0x2e, 0x6f, 0x83, 0x27, 0x60, 0x98, 0xfd, 0x50, 0x2b, 0x40, 0xcd, 0x46, 0x99, 0x17, + 0x63, 0x09, 0x4f, 0x2e, 0x98, 0x42, 0x7f, 0x0b, 0x86, 0xbe, 0x3e, 0xc4, 0xa2, 0x66, 0x5a, 0xe6, + 0x02, 0x3f, 0xe5, 0xc4, 0x92, 0xc7, 0x12, 0x66, 0x7f, 0x14, 0xc6, 0x4b, 0x0e, 0x69, 0xfa, 0xde, + 0xb2, 0x57, 0x6f, 0xf9, 0xae, 0x17, 0xa1, 0x19, 0x18, 0x60, 0x97, 0x08, 0x3f, 0x02, 0x06, 0x68, + 0x43, 0x78, 0x80, 0x3e, 0x08, 0xec, 0x2d, 0x38, 0x5d, 0xf2, 0xef, 0x79, 0xf7, 0x9c, 0xa0, 0xbe, + 0x50, 0x29, 0x6b, 0xef, 0xeb, 0x35, 0xf9, 0xbe, 0xe3, 0x46, 0x5b, 0xa9, 0x47, 0xaf, 0x56, 0x93, + 0xb3, 0xb5, 0x2b, 0x6e, 0x83, 0x64, 0x48, 0x41, 0xfe, 0x6a, 0xce, 0x68, 0x29, 0xc6, 0x57, 0x5a, + 0x2d, 0x2b, 0x53, 0xab, 0xf5, 0x06, 0x14, 0x36, 0x5d, 0xd2, 0xa8, 0x63, 0xb2, 0x29, 0x56, 0xe2, + 0xe3, 0xd9, 0x76, 0x28, 0x2b, 0x14, 0x53, 0x4a, 0xbd, 0xf8, 0xeb, 0x70, 0x45, 0x54, 0xc6, 0x8a, + 0x0c, 0xda, 0x81, 0x49, 0xf9, 0x60, 0x90, 0x50, 0xb1, 0x2e, 0x9f, 0xe8, 0xf6, 0x0a, 0x31, 0x89, + 0x9f, 0x7a, 0xb0, 0x3f, 0x37, 0x89, 0x13, 0x64, 0x70, 0x07, 0x61, 0xfa, 0x1c, 0x6c, 0xd2, 0x13, + 0x78, 0x80, 0x0d, 0x3f, 0x7b, 0x0e, 0xb2, 0x97, 0x2d, 0x2b, 0xb5, 0x7f, 0xcc, 0x82, 0x47, 0x3a, + 0x46, 0x46, 0xbc, 0xf0, 0x8f, 0x78, 0x16, 0x92, 0x2f, 0xee, 0x5c, 0xef, 0x17, 0xb7, 0xfd, 0x77, + 0x2c, 0x38, 0xb5, 0xdc, 0x6c, 0x45, 0x7b, 0x25, 0xd7, 0x54, 0x41, 0xbd, 0x08, 0x43, 0x4d, 0x52, + 0x77, 0xdb, 0x4d, 0x31, 0x73, 0x73, 0xf2, 0x94, 0x5a, 0x65, 0xa5, 0x07, 0xfb, 0x73, 0x63, 0xd5, + 0xc8, 0x0f, 0x9c, 0x2d, 0xc2, 0x0b, 0xb0, 0x40, 0x67, 0x67, 0xbd, 0xfb, 0x2e, 0xb9, 0xe9, 0x36, + 0x5d, 0x69, 0x57, 0xd4, 0x55, 0x66, 0x37, 0x2f, 0x07, 0x74, 0xfe, 0x8d, 0xb6, 0xe3, 0x45, 0x6e, + 0xb4, 0x27, 0xb4, 0x47, 0x92, 0x08, 0x8e, 0xe9, 0xd9, 0x5f, 0xb5, 0x60, 0x42, 0xae, 0xfb, 0x85, + 0x7a, 0x3d, 0x20, 0x61, 0x88, 0x66, 0x21, 0xe7, 0xb6, 0x44, 0x2f, 0x41, 0xf4, 0x32, 0x57, 0xae, + 0xe0, 0x9c, 0xdb, 0x92, 0x6c, 0x19, 0x3b, 0x08, 0xf3, 0xa6, 0x22, 0xed, 0xba, 0x28, 0xc7, 0x0a, + 0x03, 0x5d, 0x81, 0x82, 0xe7, 0xd7, 0xb9, 0x6d, 0x17, 0xbf, 0xd2, 0xd8, 0x02, 0x5b, 0x13, 0x65, + 0x58, 0x41, 0x51, 0x05, 0x8a, 0xdc, 0xec, 0x29, 0x5e, 0xb4, 0x7d, 0x19, 0x4f, 0xb1, 0x2f, 0x5b, + 0x97, 0x35, 0x71, 0x4c, 0xc4, 0xfe, 0x55, 0x0b, 0x46, 0xe5, 0x97, 0xf5, 0xc9, 0x73, 0xd2, 0xad, + 0x15, 0xf3, 0x9b, 0xf1, 0xd6, 0xa2, 0x3c, 0x23, 0x83, 0x18, 0xac, 0x62, 0xfe, 0x50, 0xac, 0xe2, + 0x33, 0x30, 0xe2, 0xb4, 0x5a, 0x15, 0x93, 0xcf, 0x64, 0x4b, 0x69, 0x21, 0x2e, 0xc6, 0x3a, 0x8e, + 0xfd, 0xa3, 0x39, 0x18, 0x97, 0x5f, 0x50, 0x6d, 0x6f, 0x84, 0x24, 0x42, 0xeb, 0x50, 0x74, 0xf8, + 0x2c, 0x11, 0xb9, 0xc8, 0x2f, 0xa5, 0xcb, 0x11, 0x8c, 0x29, 0x8d, 0x2f, 0xfc, 0x05, 0x59, 0x1b, + 0xc7, 0x84, 0x50, 0x03, 0xa6, 0x3c, 0x3f, 0x62, 0x87, 0xbf, 0x82, 0x77, 0x53, 0xed, 0x24, 0xa9, + 0x9f, 0x15, 0xd4, 0xa7, 0xd6, 0x92, 0x54, 0x70, 0x27, 0x61, 0xb4, 0x2c, 0x65, 0x33, 0xf9, 0x6c, + 0x61, 0x80, 0x3e, 0x71, 0xe9, 0xa2, 0x19, 0xfb, 0x57, 0x2c, 0x28, 0x4a, 0xb4, 0x93, 0xd0, 0xe2, + 0xad, 0xc2, 0x70, 0xc8, 0x26, 0x41, 0x0e, 0x8d, 0xdd, 0xad, 0xe3, 0x7c, 0xbe, 0xe2, 0x3b, 0x8d, + 0xff, 0x0f, 0xb1, 0xa4, 0xc1, 0x44, 0xf3, 0xaa, 0xfb, 0xef, 0x13, 0xd1, 0xbc, 0xea, 0x4f, 0xc6, + 0xa5, 0xf4, 0x87, 0xac, 0xcf, 0x9a, 0xac, 0x8b, 0xb2, 0x5e, 0xad, 0x80, 0x6c, 0xba, 0xf7, 0x93, + 0xac, 0x57, 0x85, 0x95, 0x62, 0x01, 0x45, 0x6f, 0xc1, 0x68, 0x4d, 0xca, 0x64, 0xe3, 0x1d, 0x7e, + 0xb9, 0xab, 0x7e, 0x40, 0xa9, 0x92, 0xb8, 0x2c, 0x64, 0x49, 0xab, 0x8f, 0x0d, 0x6a, 0xa6, 0x19, + 0x41, 0xbe, 0x97, 0x19, 0x41, 0x4c, 0x37, 0x5b, 0xa9, 0xfe, 0xe3, 0x16, 0x0c, 0x71, 0x59, 0x5c, + 0x7f, 0xa2, 0x50, 0x4d, 0xb3, 0x16, 0x8f, 0xdd, 0x1d, 0x5a, 0x28, 0x34, 0x65, 0x68, 0x15, 0x8a, + 0xec, 0x07, 0x93, 0x25, 0xe6, 0xb3, 0xad, 0xee, 0x79, 0xab, 0x7a, 0x07, 0xef, 0xc8, 0x6a, 0x38, + 0xa6, 0x60, 0xff, 0x70, 0x9e, 0x9e, 0x6e, 0x31, 0xaa, 0x71, 0xe9, 0x5b, 0xc7, 0x77, 0xe9, 0xe7, + 0x8e, 0xeb, 0xd2, 0xdf, 0x82, 0x89, 0x9a, 0xa6, 0x87, 0x8b, 0x67, 0xf2, 0x4a, 0xd7, 0x45, 0xa2, + 0xa9, 0xec, 0xb8, 0x94, 0x65, 0xc9, 0x24, 0x82, 0x93, 0x54, 0xd1, 0xb7, 0xc1, 0x28, 0x9f, 0x67, + 0xd1, 0x0a, 0xb7, 0xc4, 0xf8, 0x48, 0xf6, 0x7a, 0xd1, 0x9b, 0xe0, 0x52, 0x39, 0xad, 0x3a, 0x36, + 0x88, 0xd9, 0x7f, 0x62, 0x01, 0x5a, 0x6e, 0x6d, 0x93, 0x26, 0x09, 0x9c, 0x46, 0x2c, 0x4e, 0xff, + 0x7e, 0x0b, 0x66, 0x48, 0x47, 0xf1, 0x92, 0xdf, 0x6c, 0x8a, 0x47, 0x4b, 0xc6, 0xbb, 0x7a, 0x39, + 0xa3, 0x8e, 0x72, 0x4b, 0x98, 0xc9, 0xc2, 0xc0, 0x99, 0xed, 0xa1, 0x55, 0x98, 0xe6, 0xb7, 0xa4, + 0x02, 0x68, 0xb6, 0xd7, 0x8f, 0x0a, 0xc2, 0xd3, 0xeb, 0x9d, 0x28, 0x38, 0xad, 0x9e, 0xfd, 0x5d, + 0xa3, 0x90, 0xd9, 0x8b, 0x0f, 0xf4, 0x08, 0x1f, 0xe8, 0x11, 0x3e, 0xd0, 0x23, 0x7c, 0xa0, 0x47, + 0xf8, 0x40, 0x8f, 0xf0, 0x4d, 0xaf, 0x47, 0xf8, 0x23, 0x0b, 0xa6, 0x3b, 0xaf, 0x81, 0x93, 0x60, + 0xcc, 0xdb, 0x30, 0xdd, 0x79, 0xd7, 0x75, 0xb5, 0xb3, 0xeb, 0xec, 0x67, 0x7c, 0xef, 0xa5, 0x7c, + 0x03, 0x4e, 0xa3, 0x6f, 0xff, 0xba, 0x05, 0xa7, 0x15, 0xb2, 0xf1, 0xd2, 0xff, 0x3c, 0x4c, 0xf3, + 0xf3, 0x65, 0xa9, 0xe1, 0xb8, 0xcd, 0x75, 0xd2, 0x6c, 0x35, 0x9c, 0x48, 0x9a, 0x19, 0x3c, 0x93, + 0xba, 0x55, 0x13, 0x26, 0xba, 0x46, 0xc5, 0xc5, 0x47, 0x68, 0xbf, 0x52, 0x00, 0x38, 0xad, 0x19, + 0xc3, 0x28, 0x35, 0xd7, 0xd3, 0x4c, 0xf8, 0x17, 0x0b, 0x30, 0xb8, 0xbc, 0x4b, 0xbc, 0xe8, 0x04, + 0x26, 0xaa, 0x06, 0xe3, 0xae, 0xb7, 0xeb, 0x37, 0x76, 0x49, 0x9d, 0xc3, 0x0f, 0xf3, 0xd0, 0x3f, + 0x23, 0x48, 0x8f, 0x97, 0x0d, 0x12, 0x38, 0x41, 0xf2, 0x38, 0x84, 0xed, 0xd7, 0x60, 0x88, 0xdf, + 0x71, 0x42, 0xd2, 0x9e, 0x7a, 0xa5, 0xb1, 0x41, 0x14, 0x37, 0x77, 0xac, 0x08, 0xe0, 0x77, 0xa8, + 0xa8, 0x8e, 0xde, 0x86, 0xf1, 0x4d, 0x37, 0x08, 0xa3, 0x75, 0xb7, 0x49, 0xc2, 0xc8, 0x69, 0xb6, + 0x1e, 0x42, 0xb8, 0xae, 0xc6, 0x61, 0xc5, 0xa0, 0x84, 0x13, 0x94, 0xd1, 0x16, 0x8c, 0x35, 0x1c, + 0xbd, 0xa9, 0xe1, 0x43, 0x37, 0xa5, 0x2e, 0xcf, 0x9b, 0x3a, 0x21, 0x6c, 0xd2, 0xa5, 0xa7, 0x4d, + 0x8d, 0xc9, 0x87, 0x0b, 0x4c, 0x6a, 0xa2, 0x4e, 0x1b, 0x2e, 0x18, 0xe6, 0x30, 0xca, 0x07, 0x32, + 0xfb, 0xe1, 0xa2, 0xc9, 0x07, 0x6a, 0x56, 0xc2, 0x9f, 0x83, 0x22, 0xa1, 0x43, 0x48, 0x09, 0x8b, + 0xfb, 0xf7, 0x6a, 0x7f, 0x7d, 0x5d, 0x75, 0x6b, 0x81, 0x6f, 0xaa, 0x35, 0x96, 0x25, 0x25, 0x1c, + 0x13, 0x45, 0x4b, 0x30, 0x14, 0x92, 0xc0, 0x25, 0xa1, 0xb8, 0x89, 0xbb, 0x4c, 0x23, 0x43, 0xe3, + 0xae, 0x37, 0xfc, 0x37, 0x16, 0x55, 0xe9, 0xf2, 0x72, 0x98, 0xc4, 0x97, 0xdd, 0x95, 0xda, 0xf2, + 0x5a, 0x60, 0xa5, 0x58, 0x40, 0xd1, 0xeb, 0x30, 0x1c, 0x90, 0x06, 0xd3, 0x9b, 0x8d, 0xf5, 0xbf, + 0xc8, 0xb9, 0x1a, 0x8e, 0xd7, 0xc3, 0x92, 0x00, 0xba, 0x01, 0x28, 0x20, 0x94, 0x8f, 0x74, 0xbd, + 0x2d, 0x65, 0x55, 0x2b, 0xee, 0x21, 0x75, 0x6e, 0xe1, 0x18, 0x43, 0x7a, 0x41, 0xe1, 0x94, 0x6a, + 0xe8, 0x1a, 0x4c, 0xa9, 0xd2, 0xb2, 0x17, 0x46, 0x0e, 0x3d, 0xff, 0x27, 0x18, 0x2d, 0x25, 0xc6, + 0xc1, 0x49, 0x04, 0xdc, 0x59, 0xc7, 0xfe, 0xb2, 0x05, 0x7c, 0x9c, 0x4f, 0x40, 0x78, 0xf1, 0x9a, + 0x29, 0xbc, 0x38, 0x9b, 0x39, 0x73, 0x19, 0x82, 0x8b, 0x2f, 0x5b, 0x30, 0xa2, 0xcd, 0x6c, 0xbc, + 0x66, 0xad, 0x2e, 0x6b, 0xb6, 0x0d, 0x93, 0x74, 0xa5, 0xdf, 0xda, 0x08, 0x49, 0xb0, 0x4b, 0xea, + 0x6c, 0x61, 0xe6, 0x1e, 0x6e, 0x61, 0x2a, 0x0b, 0xbe, 0x9b, 0x09, 0x82, 0xb8, 0xa3, 0x09, 0xfb, + 0x73, 0xb2, 0xab, 0xca, 0xe0, 0xb1, 0xa6, 0xe6, 0x3c, 0x61, 0xf0, 0xa8, 0x66, 0x15, 0xc7, 0x38, + 0x74, 0xab, 0x6d, 0xfb, 0x61, 0x94, 0x34, 0x78, 0xbc, 0xee, 0x87, 0x11, 0x66, 0x10, 0xfb, 0x39, + 0x80, 0xe5, 0xfb, 0xa4, 0xc6, 0x57, 0xac, 0xfe, 0xb6, 0xb2, 0xb2, 0xdf, 0x56, 0xf6, 0x6f, 0x5b, + 0x30, 0xbe, 0xb2, 0x64, 0xdc, 0x73, 0xf3, 0x00, 0xfc, 0x41, 0x78, 0xf7, 0xee, 0x9a, 0xb4, 0x16, + 0xe0, 0x0a, 0x5f, 0x55, 0x8a, 0x35, 0x0c, 0x74, 0x16, 0xf2, 0x8d, 0xb6, 0x27, 0xa4, 0xab, 0xc3, + 0x94, 0x7b, 0xb8, 0xd9, 0xf6, 0x30, 0x2d, 0xd3, 0x3c, 0x2e, 0xf2, 0x7d, 0x7b, 0x5c, 0xf4, 0x8c, + 0x7c, 0x80, 0xe6, 0x60, 0xf0, 0xde, 0x3d, 0xb7, 0xce, 0xfd, 0x4b, 0x85, 0x25, 0xc3, 0xdd, 0xbb, + 0xe5, 0x52, 0x88, 0x79, 0xb9, 0xfd, 0xa5, 0x3c, 0xcc, 0xae, 0x34, 0xc8, 0xfd, 0xf7, 0xe8, 0x63, + 0xdb, 0xaf, 0xbf, 0xc8, 0xe1, 0xe4, 0x54, 0x87, 0xf5, 0x09, 0xea, 0x3d, 0x1e, 0x9b, 0x30, 0xcc, + 0xed, 0xfd, 0xa4, 0xc7, 0xed, 0x2b, 0x69, 0xad, 0x67, 0x0f, 0xc8, 0x3c, 0xb7, 0x1b, 0x14, 0x0e, + 0x83, 0xea, 0xc2, 0x14, 0xa5, 0x58, 0x12, 0x9f, 0x7d, 0x19, 0x46, 0x75, 0xcc, 0x43, 0x79, 0xe7, + 0xfd, 0xff, 0x79, 0x98, 0xa4, 0x3d, 0x38, 0xd6, 0x89, 0xb8, 0xdd, 0x39, 0x11, 0x47, 0xed, 0xa1, + 0xd5, 0x7b, 0x36, 0xde, 0x4a, 0xce, 0xc6, 0x33, 0x59, 0xb3, 0x71, 0xd2, 0x73, 0xf0, 0x9d, 0x16, + 0x4c, 0xaf, 0x34, 0xfc, 0xda, 0x4e, 0xc2, 0x8b, 0xea, 0x05, 0x18, 0xa1, 0xc7, 0x71, 0x68, 0x38, + 0xf8, 0x1b, 0x21, 0x1f, 0x04, 0x08, 0xeb, 0x78, 0x5a, 0xb5, 0xdb, 0xb7, 0xcb, 0xa5, 0xb4, 0x48, + 0x11, 0x02, 0x84, 0x75, 0x3c, 0xfb, 0x37, 0x2d, 0x38, 0x7f, 0x6d, 0x69, 0x39, 0x5e, 0x8a, 0x1d, + 0xc1, 0x2a, 0x2e, 0xc3, 0x50, 0xab, 0xae, 0x75, 0x25, 0x96, 0x3e, 0x97, 0x58, 0x2f, 0x04, 0xf4, + 0xfd, 0x12, 0x88, 0xe5, 0xa7, 0x2d, 0x98, 0xbe, 0xe6, 0x46, 0xf4, 0x76, 0x4d, 0x86, 0x4d, 0xa0, + 0xd7, 0x6b, 0xe8, 0x46, 0x7e, 0xb0, 0x97, 0x0c, 0x9b, 0x80, 0x15, 0x04, 0x6b, 0x58, 0xbc, 0xe5, + 0x5d, 0x97, 0x59, 0x9a, 0xe7, 0x4c, 0x3d, 0x1c, 0x16, 0xe5, 0x58, 0x61, 0xd0, 0x0f, 0xab, 0xbb, + 0x01, 0x13, 0x61, 0xee, 0x89, 0x13, 0x56, 0x7d, 0x58, 0x49, 0x02, 0x70, 0x8c, 0x43, 0x5f, 0x73, + 0x73, 0xd7, 0x1a, 0xed, 0x30, 0x22, 0xc1, 0x66, 0x98, 0x71, 0x3a, 0x3e, 0x07, 0x45, 0x22, 0x15, + 0x06, 0xa2, 0xd7, 0x8a, 0x63, 0x54, 0x9a, 0x04, 0x1e, 0xbd, 0x41, 0xe1, 0xf5, 0xe1, 0x93, 0x79, + 0x38, 0xa7, 0xba, 0x15, 0x40, 0x44, 0x6f, 0x4b, 0x0f, 0x67, 0xc1, 0xfc, 0xe2, 0x97, 0x3b, 0xa0, + 0x38, 0xa5, 0x86, 0xfd, 0x63, 0x16, 0x9c, 0x56, 0x1f, 0xfc, 0xbe, 0xfb, 0x4c, 0xfb, 0xe7, 0x72, + 0x30, 0x76, 0x7d, 0x7d, 0xbd, 0x72, 0x8d, 0x44, 0xe2, 0xda, 0xee, 0x6d, 0x06, 0x80, 0x35, 0x6d, + 0x66, 0xb7, 0xc7, 0x5c, 0x3b, 0x72, 0x1b, 0xf3, 0x3c, 0x2a, 0xd2, 0x7c, 0xd9, 0x8b, 0x6e, 0x05, + 0xd5, 0x28, 0x70, 0xbd, 0xad, 0x54, 0xfd, 0xa7, 0x64, 0x2e, 0xf2, 0x59, 0xcc, 0x05, 0x7a, 0x0e, + 0x86, 0x58, 0x58, 0x26, 0x39, 0x09, 0x8f, 0xaa, 0xb7, 0x10, 0x2b, 0x3d, 0xd8, 0x9f, 0x2b, 0xde, + 0xc6, 0x65, 0xfe, 0x07, 0x0b, 0x54, 0x74, 0x1b, 0x46, 0xb6, 0xa3, 0xa8, 0x75, 0x9d, 0x38, 0x75, + 0xfa, 0x74, 0xe7, 0xc7, 0xe1, 0x85, 0xb4, 0xe3, 0x90, 0x0e, 0x02, 0x47, 0x8b, 0x4f, 0x90, 0xb8, + 0x2c, 0xc4, 0x3a, 0x1d, 0xbb, 0x0a, 0x10, 0xc3, 0x8e, 0x48, 0x91, 0x63, 0xff, 0x81, 0x05, 0xc3, + 0x3c, 0x42, 0x46, 0x80, 0x5e, 0x85, 0x01, 0x72, 0x9f, 0xd4, 0x04, 0xc7, 0x9b, 0xda, 0xe1, 0x98, + 0xd3, 0xe2, 0x02, 0x69, 0xfa, 0x1f, 0xb3, 0x5a, 0xe8, 0x3a, 0x0c, 0xd3, 0xde, 0x5e, 0x53, 0xe1, + 0x42, 0x1e, 0xcb, 0xfa, 0x62, 0x35, 0xed, 0x9c, 0x39, 0x13, 0x45, 0x58, 0x56, 0x67, 0xda, 0xf3, + 0x5a, 0xab, 0x4a, 0x4f, 0xec, 0xa8, 0x1b, 0x63, 0xb1, 0xbe, 0x54, 0xe1, 0x48, 0x82, 0x1a, 0xd7, + 0x9e, 0xcb, 0x42, 0x1c, 0x13, 0xb1, 0xd7, 0xa1, 0x48, 0x27, 0x75, 0xa1, 0xe1, 0x3a, 0xdd, 0x0d, + 0x02, 0x9e, 0x84, 0xa2, 0x54, 0xf7, 0x87, 0xc2, 0x33, 0x9e, 0x51, 0x95, 0xd6, 0x00, 0x21, 0x8e, + 0xe1, 0xf6, 0x26, 0x9c, 0x62, 0xc6, 0x9b, 0x4e, 0xb4, 0x6d, 0xec, 0xb1, 0xde, 0x8b, 0xf9, 0x29, + 0xf1, 0x80, 0xe4, 0x33, 0x33, 0xa3, 0x39, 0x9f, 0x8e, 0x4a, 0x8a, 0xf1, 0x63, 0xd2, 0xfe, 0xda, + 0x00, 0x3c, 0x5a, 0xae, 0x66, 0x07, 0x4f, 0x79, 0x09, 0x46, 0x39, 0x5f, 0x4a, 0x97, 0xb6, 0xd3, + 0x10, 0xed, 0x2a, 0x49, 0xf4, 0xba, 0x06, 0xc3, 0x06, 0x26, 0x3a, 0x0f, 0x79, 0xf7, 0x1d, 0x2f, + 0xe9, 0x9a, 0x55, 0x7e, 0x63, 0x0d, 0xd3, 0x72, 0x0a, 0xa6, 0x2c, 0x2e, 0xbf, 0x3b, 0x14, 0x58, + 0xb1, 0xb9, 0xaf, 0xc1, 0xb8, 0x1b, 0xd6, 0x42, 0xb7, 0xec, 0xd1, 0x73, 0x46, 0x3b, 0xa9, 0x94, + 0x70, 0x83, 0x76, 0x5a, 0x41, 0x71, 0x02, 0x5b, 0xbb, 0xc8, 0x06, 0xfb, 0x66, 0x93, 0x7b, 0xba, + 0x8a, 0xd3, 0x17, 0x40, 0x8b, 0x7d, 0x5d, 0xc8, 0x54, 0x0a, 0xe2, 0x05, 0xc0, 0x3f, 0x38, 0xc4, + 0x12, 0x46, 0x5f, 0x8e, 0xb5, 0x6d, 0xa7, 0xb5, 0xd0, 0x8e, 0xb6, 0x4b, 0x6e, 0x58, 0xf3, 0x77, + 0x49, 0xb0, 0xc7, 0x1e, 0xfd, 0x85, 0xf8, 0xe5, 0xa8, 0x00, 0x4b, 0xd7, 0x17, 0x2a, 0x14, 0x13, + 0x77, 0xd6, 0x41, 0x0b, 0x30, 0x21, 0x0b, 0xab, 0x24, 0x64, 0x57, 0xd8, 0x08, 0x23, 0xa3, 0x9c, + 0xa5, 0x44, 0xb1, 0x22, 0x92, 0xc4, 0x37, 0x39, 0x69, 0x38, 0x0a, 0x4e, 0xfa, 0x45, 0x18, 0x73, + 0x3d, 0x37, 0x72, 0x9d, 0xc8, 0xe7, 0xfa, 0x30, 0xfe, 0xbe, 0x67, 0x82, 0xfe, 0xb2, 0x0e, 0xc0, + 0x26, 0x9e, 0xfd, 0x5f, 0x06, 0x60, 0x8a, 0x4d, 0xdb, 0x07, 0x2b, 0xec, 0x9b, 0x69, 0x85, 0xdd, + 0xee, 0x5c, 0x61, 0x47, 0xf1, 0x44, 0x78, 0xe8, 0x65, 0xf6, 0x36, 0x14, 0x95, 0x7f, 0x98, 0x74, + 0x10, 0xb5, 0x32, 0x1c, 0x44, 0x7b, 0x73, 0x1f, 0xd2, 0xc4, 0x2e, 0x9f, 0x6a, 0x62, 0xf7, 0xd7, + 0x2d, 0x88, 0x15, 0x3c, 0xe8, 0x3a, 0x14, 0x5b, 0x3e, 0xb3, 0x1c, 0x0d, 0xa4, 0x39, 0xf6, 0xa3, + 0xa9, 0x17, 0x15, 0xbf, 0x14, 0xf9, 0xc7, 0x57, 0x64, 0x0d, 0x1c, 0x57, 0x46, 0x8b, 0x30, 0xdc, + 0x0a, 0x48, 0x35, 0x62, 0x31, 0x54, 0x7a, 0xd2, 0xe1, 0x6b, 0x84, 0xe3, 0x63, 0x59, 0xd1, 0xfe, + 0x79, 0x0b, 0x80, 0x5b, 0xb1, 0x39, 0xde, 0x16, 0x39, 0x01, 0xa9, 0x75, 0x09, 0x06, 0xc2, 0x16, + 0xa9, 0x75, 0xb3, 0xe9, 0x8d, 0xfb, 0x53, 0x6d, 0x91, 0x5a, 0x3c, 0xe0, 0xf4, 0x1f, 0x66, 0xb5, + 0xed, 0xef, 0x06, 0x18, 0x8f, 0xd1, 0xca, 0x11, 0x69, 0xa2, 0xa7, 0x8d, 0x98, 0x0a, 0x67, 0x13, + 0x31, 0x15, 0x8a, 0x0c, 0x5b, 0x13, 0x90, 0xbe, 0x0d, 0xf9, 0xa6, 0x73, 0x5f, 0x48, 0xc0, 0x9e, + 0xec, 0xde, 0x0d, 0x4a, 0x7f, 0x7e, 0xd5, 0xb9, 0xcf, 0x1f, 0x89, 0x4f, 0xca, 0x05, 0xb2, 0xea, + 0xdc, 0x3f, 0xe0, 0x96, 0xbb, 0xec, 0x90, 0xba, 0xe9, 0x86, 0xd1, 0x17, 0xfe, 0x73, 0xfc, 0x9f, + 0x2d, 0x3b, 0xda, 0x08, 0x6b, 0xcb, 0xf5, 0x84, 0x81, 0x56, 0x5f, 0x6d, 0xb9, 0x5e, 0xb2, 0x2d, + 0xd7, 0xeb, 0xa3, 0x2d, 0xd7, 0x43, 0xef, 0xc2, 0xb0, 0xb0, 0x9f, 0x14, 0x31, 0x8c, 0xae, 0xf6, + 0xd1, 0x9e, 0x30, 0xbf, 0xe4, 0x6d, 0x5e, 0x95, 0x8f, 0x60, 0x51, 0xda, 0xb3, 0x5d, 0xd9, 0x20, + 0xfa, 0x2b, 0x16, 0x8c, 0x8b, 0xdf, 0x98, 0xbc, 0xd3, 0x26, 0x61, 0x24, 0x78, 0xcf, 0x8f, 0xf7, + 0xdf, 0x07, 0x51, 0x91, 0x77, 0xe5, 0xe3, 0xf2, 0x98, 0x35, 0x81, 0x3d, 0x7b, 0x94, 0xe8, 0x05, + 0xfa, 0x7b, 0x16, 0x9c, 0x6a, 0x3a, 0xf7, 0x79, 0x8b, 0xbc, 0x0c, 0x3b, 0x91, 0xeb, 0x0b, 0x3b, + 0x84, 0x57, 0xfb, 0x9b, 0xfe, 0x8e, 0xea, 0xbc, 0x93, 0x52, 0x59, 0x7a, 0x2a, 0x0d, 0xa5, 0x67, + 0x57, 0x53, 0xfb, 0x35, 0xbb, 0x09, 0x05, 0xb9, 0xde, 0x52, 0x44, 0x0d, 0x25, 0x9d, 0xb1, 0x3e, + 0xb4, 0xf9, 0xaa, 0x1e, 0xab, 0x80, 0xb6, 0x23, 0xd6, 0xda, 0xb1, 0xb6, 0xf3, 0x36, 0x8c, 0xea, + 0x6b, 0xec, 0x58, 0xdb, 0x7a, 0x07, 0xa6, 0x53, 0xd6, 0xd2, 0xb1, 0x36, 0x79, 0x0f, 0xce, 0x66, + 0xae, 0x8f, 0xe3, 0x6c, 0xd8, 0xfe, 0x39, 0x4b, 0x3f, 0x07, 0x4f, 0x40, 0x75, 0xb0, 0x64, 0xaa, + 0x0e, 0x2e, 0x74, 0xdf, 0x39, 0x19, 0xfa, 0x83, 0xb7, 0xf4, 0x4e, 0xd3, 0x53, 0x1d, 0xbd, 0x0e, + 0x43, 0x0d, 0x5a, 0x22, 0xad, 0x70, 0xed, 0xde, 0x3b, 0x32, 0xe6, 0xa5, 0x58, 0x79, 0x88, 0x05, + 0x05, 0xfb, 0x97, 0x2c, 0x18, 0x38, 0x81, 0x91, 0xc0, 0xe6, 0x48, 0x3c, 0x9d, 0x49, 0x5a, 0x84, + 0x57, 0x9e, 0xc7, 0xce, 0xbd, 0xe5, 0xfb, 0x11, 0xf1, 0x42, 0xf6, 0x54, 0x4c, 0x1d, 0x98, 0xff, + 0x07, 0xa6, 0x6f, 0xfa, 0x4e, 0x7d, 0xd1, 0x69, 0x38, 0x5e, 0x8d, 0x04, 0x65, 0x6f, 0xeb, 0x50, + 0x16, 0xe4, 0xb9, 0x5e, 0x16, 0xe4, 0xf6, 0x36, 0x20, 0xbd, 0x01, 0xe1, 0x8a, 0x83, 0x61, 0xd8, + 0xe5, 0x4d, 0x89, 0xe1, 0x7f, 0x3c, 0x9d, 0x35, 0xeb, 0xe8, 0x99, 0xe6, 0x64, 0xc2, 0x0b, 0xb0, + 0x24, 0x64, 0xbf, 0x04, 0xa9, 0xfe, 0xfc, 0xbd, 0xc5, 0x06, 0xf6, 0xa7, 0x61, 0x8a, 0xd5, 0x3c, + 0xe4, 0x93, 0xd6, 0x4e, 0x48, 0x25, 0x53, 0x22, 0xfd, 0xd9, 0x5f, 0xb4, 0x60, 0x62, 0x2d, 0x11, + 0x00, 0xed, 0x32, 0xd3, 0x63, 0xa6, 0x08, 0xc3, 0xab, 0xac, 0x14, 0x0b, 0xe8, 0x91, 0xcb, 0xa0, + 0xfe, 0xc2, 0x82, 0x38, 0xc4, 0xc6, 0x09, 0x30, 0x5e, 0x4b, 0x06, 0xe3, 0x95, 0x2a, 0x1b, 0x51, + 0xdd, 0xc9, 0xe2, 0xbb, 0xd0, 0x0d, 0x15, 0x7c, 0xaa, 0x8b, 0x58, 0x24, 0x26, 0xc3, 0x43, 0x15, + 0x8d, 0x9b, 0x11, 0xaa, 0x64, 0x38, 0x2a, 0xfb, 0x3f, 0xe6, 0x00, 0x29, 0xdc, 0xbe, 0x83, 0x63, + 0x75, 0xd6, 0x38, 0x9a, 0xe0, 0x58, 0xbb, 0x80, 0x98, 0x26, 0x3e, 0x70, 0xbc, 0x90, 0x93, 0x75, + 0x85, 0xd4, 0xed, 0x70, 0x6a, 0xfe, 0x59, 0xd1, 0x24, 0xba, 0xd9, 0x41, 0x0d, 0xa7, 0xb4, 0xa0, + 0x59, 0x58, 0x0c, 0xf6, 0x6b, 0x61, 0x31, 0xd4, 0xc3, 0xdd, 0xee, 0x67, 0x2d, 0x18, 0x53, 0xc3, + 0xf4, 0x3e, 0x31, 0x86, 0x57, 0xfd, 0xc9, 0x38, 0xfa, 0x2a, 0x5a, 0x97, 0xd9, 0x95, 0xf0, 0xad, + 0xcc, 0x6d, 0xd2, 0x69, 0xb8, 0xef, 0x12, 0x15, 0x9a, 0x70, 0x4e, 0xb8, 0x41, 0x8a, 0xd2, 0x83, + 0xfd, 0xb9, 0x31, 0xf5, 0x8f, 0x87, 0x42, 0x8e, 0xab, 0xd8, 0x3f, 0x49, 0x37, 0xbb, 0xb9, 0x14, + 0xd1, 0x0b, 0x30, 0xd8, 0xda, 0x76, 0x42, 0x92, 0x70, 0x1a, 0x1a, 0xac, 0xd0, 0xc2, 0x83, 0xfd, + 0xb9, 0x71, 0x55, 0x81, 0x95, 0x60, 0x8e, 0xdd, 0x7f, 0xc8, 0xb1, 0xce, 0xc5, 0xd9, 0x33, 0xe4, + 0xd8, 0x9f, 0x58, 0x30, 0xb0, 0xe6, 0xd7, 0x4f, 0xe2, 0x08, 0x78, 0xcd, 0x38, 0x02, 0xce, 0x65, + 0x45, 0xa9, 0xcf, 0xdc, 0xfd, 0x2b, 0x89, 0xdd, 0x7f, 0x21, 0x93, 0x42, 0xf7, 0x8d, 0xdf, 0x84, + 0x11, 0x16, 0xfb, 0x5e, 0x38, 0x48, 0x3d, 0x67, 0x6c, 0xf8, 0xb9, 0xc4, 0x86, 0x9f, 0xd0, 0x50, + 0xb5, 0x9d, 0xfe, 0x04, 0x0c, 0x0b, 0x8f, 0x9b, 0xa4, 0xf7, 0xa9, 0xc0, 0xc5, 0x12, 0x6e, 0xff, + 0x78, 0x1e, 0x8c, 0x58, 0xfb, 0xe8, 0x57, 0x2c, 0x98, 0x0f, 0xb8, 0x25, 0x6e, 0xbd, 0xd4, 0x0e, + 0x5c, 0x6f, 0xab, 0x5a, 0xdb, 0x26, 0xf5, 0x76, 0xc3, 0xf5, 0xb6, 0xca, 0x5b, 0x9e, 0xaf, 0x8a, + 0x97, 0xef, 0x93, 0x5a, 0x9b, 0xa9, 0xaf, 0x7a, 0x04, 0xf6, 0x57, 0x16, 0xed, 0xcf, 0x3e, 0xd8, + 0x9f, 0x9b, 0xc7, 0x87, 0xa2, 0x8d, 0x0f, 0xd9, 0x17, 0xf4, 0x9b, 0x16, 0x5c, 0xe5, 0x21, 0xe8, + 0xfb, 0xef, 0x7f, 0x97, 0x77, 0x6e, 0x45, 0x92, 0x8a, 0x89, 0xac, 0x93, 0xa0, 0xb9, 0xf8, 0xa2, + 0x18, 0xd0, 0xab, 0x95, 0xc3, 0xb5, 0x85, 0x0f, 0xdb, 0x39, 0xfb, 0x9f, 0xe6, 0x61, 0x4c, 0x84, + 0xa6, 0x12, 0x77, 0xc0, 0x0b, 0xc6, 0x92, 0x78, 0x2c, 0xb1, 0x24, 0xa6, 0x0c, 0xe4, 0xa3, 0x39, + 0xfe, 0x43, 0x98, 0xa2, 0x87, 0xf3, 0x75, 0xe2, 0x04, 0xd1, 0x06, 0x71, 0xb8, 0xe1, 0x54, 0xfe, + 0xd0, 0xa7, 0xbf, 0x12, 0xac, 0xdd, 0x4c, 0x12, 0xc3, 0x9d, 0xf4, 0xbf, 0x99, 0xee, 0x1c, 0x0f, + 0x26, 0x3b, 0xa2, 0x8b, 0xbd, 0x09, 0x45, 0xe5, 0x2e, 0x22, 0x0e, 0x9d, 0xee, 0x41, 0xfa, 0x92, + 0x14, 0xb8, 0xf0, 0x2b, 0x76, 0x55, 0x8a, 0xc9, 0xd9, 0x7f, 0x3f, 0x67, 0x34, 0xc8, 0x27, 0x71, + 0x0d, 0x0a, 0x4e, 0x18, 0xba, 0x5b, 0x1e, 0xa9, 0x8b, 0x1d, 0xfb, 0xe1, 0xac, 0x1d, 0x6b, 0x34, + 0xc3, 0x5c, 0x76, 0x16, 0x44, 0x4d, 0xac, 0x68, 0xa0, 0xeb, 0xdc, 0x3c, 0x6d, 0x57, 0xbe, 0xd4, + 0xfa, 0xa3, 0x06, 0xd2, 0x80, 0x6d, 0x97, 0x60, 0x51, 0x1f, 0x7d, 0x86, 0xdb, 0x0f, 0xde, 0xf0, + 0xfc, 0x7b, 0xde, 0x35, 0xdf, 0x97, 0xe1, 0x1f, 0xfa, 0x23, 0x38, 0x25, 0xad, 0x06, 0x55, 0x75, + 0x6c, 0x52, 0xeb, 0x2f, 0x5c, 0xe7, 0xe7, 0x61, 0x9a, 0x92, 0x36, 0xbd, 0xb3, 0x43, 0x44, 0x60, + 0x42, 0xc4, 0x3d, 0x93, 0x65, 0x62, 0xec, 0x52, 0x1f, 0x61, 0x66, 0xed, 0x58, 0x02, 0x7c, 0xc3, + 0x24, 0x81, 0x93, 0x34, 0xed, 0x9f, 0xb2, 0x80, 0x79, 0xaa, 0x9e, 0x00, 0x3f, 0xf2, 0x09, 0x93, + 0x1f, 0x99, 0xc9, 0x1a, 0xe4, 0x0c, 0x56, 0xe4, 0x79, 0xbe, 0xb2, 0x2a, 0x81, 0x7f, 0x7f, 0x4f, + 0x18, 0x7d, 0xf4, 0x7e, 0x7f, 0xd8, 0xff, 0xdb, 0xe2, 0x87, 0x98, 0x72, 0xe6, 0x40, 0xdf, 0x0e, + 0x85, 0x9a, 0xd3, 0x72, 0x6a, 0x3c, 0x31, 0x4c, 0xa6, 0x2c, 0xce, 0xa8, 0x34, 0xbf, 0x24, 0x6a, + 0x70, 0xd9, 0x92, 0x8c, 0x9f, 0x57, 0x90, 0xc5, 0x3d, 0xe5, 0x49, 0xaa, 0xc9, 0xd9, 0x1d, 0x18, + 0x33, 0x88, 0x1d, 0xab, 0x20, 0xe2, 0xdb, 0xf9, 0x15, 0xab, 0xe2, 0x3d, 0x36, 0x61, 0xca, 0xd3, + 0xfe, 0xd3, 0x0b, 0x45, 0x3e, 0x2e, 0x3f, 0xdc, 0xeb, 0x12, 0x65, 0xb7, 0x8f, 0xe6, 0x04, 0x9b, + 0x20, 0x83, 0x3b, 0x29, 0xdb, 0x3f, 0x61, 0xc1, 0x23, 0x3a, 0xa2, 0xe6, 0x67, 0xd3, 0x4b, 0xba, + 0x5f, 0x82, 0x82, 0xdf, 0x22, 0x81, 0x13, 0xf9, 0x81, 0xb8, 0x35, 0xae, 0xc8, 0x41, 0xbf, 0x25, + 0xca, 0x0f, 0x44, 0x58, 0x75, 0x49, 0x5d, 0x96, 0x63, 0x55, 0x93, 0xbe, 0x3e, 0xd9, 0x60, 0x84, + 0xc2, 0xa3, 0x8a, 0x9d, 0x01, 0x4c, 0xd1, 0x1d, 0x62, 0x01, 0xb1, 0xbf, 0x66, 0xf1, 0x85, 0xa5, + 0x77, 0x1d, 0xbd, 0x03, 0x93, 0x4d, 0x27, 0xaa, 0x6d, 0x2f, 0xdf, 0x6f, 0x05, 0x5c, 0x57, 0x22, + 0xc7, 0xe9, 0xc9, 0x5e, 0xe3, 0xa4, 0x7d, 0x64, 0x6c, 0x12, 0xb9, 0x9a, 0x20, 0x86, 0x3b, 0xc8, + 0xa3, 0x0d, 0x18, 0x61, 0x65, 0xcc, 0x59, 0x30, 0xec, 0xc6, 0x1a, 0x64, 0xb5, 0xa6, 0x6c, 0x05, + 0x56, 0x63, 0x3a, 0x58, 0x27, 0x6a, 0xff, 0x4c, 0x9e, 0xef, 0x76, 0xc6, 0xca, 0x3f, 0x01, 0xc3, + 0x2d, 0xbf, 0xbe, 0x54, 0x2e, 0x61, 0x31, 0x0b, 0xea, 0x1a, 0xa9, 0xf0, 0x62, 0x2c, 0xe1, 0xe8, + 0x0a, 0x14, 0xc4, 0x4f, 0xa9, 0xdb, 0x62, 0x67, 0xb3, 0xc0, 0x0b, 0xb1, 0x82, 0xa2, 0x67, 0x01, + 0x5a, 0x81, 0xbf, 0xeb, 0xd6, 0x59, 0x10, 0x8b, 0xbc, 0x69, 0xe6, 0x53, 0x51, 0x10, 0xac, 0x61, + 0xa1, 0x57, 0x60, 0xac, 0xed, 0x85, 0x9c, 0x1d, 0xd1, 0x42, 0xd6, 0x2a, 0x03, 0x94, 0xdb, 0x3a, + 0x10, 0x9b, 0xb8, 0x68, 0x01, 0x86, 0x22, 0x87, 0x99, 0xad, 0x0c, 0x66, 0x9b, 0xcd, 0xae, 0x53, + 0x0c, 0x3d, 0x07, 0x09, 0xad, 0x80, 0x45, 0x45, 0xf4, 0xa6, 0xf4, 0xdb, 0xe5, 0x07, 0xbb, 0xb0, + 0x57, 0xef, 0xef, 0x12, 0xd0, 0xbc, 0x76, 0x85, 0x1d, 0xbc, 0x41, 0x0b, 0xbd, 0x0c, 0x40, 0xee, + 0x47, 0x24, 0xf0, 0x9c, 0x86, 0xb2, 0x0a, 0x53, 0x7c, 0x41, 0xc9, 0x5f, 0xf3, 0xa3, 0xdb, 0x21, + 0x59, 0x56, 0x18, 0x58, 0xc3, 0xb6, 0x7f, 0xb3, 0x08, 0x10, 0xf3, 0xed, 0xe8, 0xdd, 0x8e, 0x83, + 0xeb, 0xa9, 0xee, 0x9c, 0xfe, 0xd1, 0x9d, 0x5a, 0xe8, 0x7b, 0x2c, 0x18, 0x71, 0x1a, 0x0d, 0xbf, + 0xe6, 0xf0, 0xa0, 0xc2, 0xb9, 0xee, 0x07, 0xa7, 0x68, 0x7f, 0x21, 0xae, 0xc1, 0xbb, 0xf0, 0x9c, + 0x5c, 0xa1, 0x1a, 0xa4, 0x67, 0x2f, 0xf4, 0x86, 0xd1, 0xc7, 0xe4, 0x53, 0x31, 0x6f, 0x0c, 0xa5, + 0x7a, 0x2a, 0x16, 0xd9, 0x1d, 0xa1, 0xbf, 0x12, 0x6f, 0x1b, 0xaf, 0xc4, 0x81, 0x6c, 0xc7, 0x44, + 0x83, 0x7d, 0xed, 0xf5, 0x40, 0x44, 0x15, 0x3d, 0x48, 0xc1, 0x60, 0xb6, 0x17, 0xa0, 0xf6, 0x4e, + 0xea, 0x11, 0xa0, 0xe0, 0x6d, 0x98, 0xa8, 0x9b, 0x4c, 0x80, 0x58, 0x89, 0x8f, 0x67, 0xd1, 0x4d, + 0xf0, 0x0c, 0xf1, 0xb5, 0x9f, 0x00, 0xe0, 0x24, 0x61, 0x54, 0xe1, 0x31, 0x2b, 0xca, 0xde, 0xa6, + 0x2f, 0x7c, 0x26, 0xec, 0xcc, 0xb9, 0xdc, 0x0b, 0x23, 0xd2, 0xa4, 0x98, 0xf1, 0xed, 0xbe, 0x26, + 0xea, 0x62, 0x45, 0x05, 0xbd, 0x0e, 0x43, 0xcc, 0x0d, 0x2c, 0x9c, 0x29, 0x64, 0xcb, 0x8a, 0xcd, + 0x20, 0x6c, 0xf1, 0x86, 0x64, 0x7f, 0x43, 0x2c, 0x28, 0xa0, 0xeb, 0xd2, 0xc9, 0x32, 0x2c, 0x7b, + 0xb7, 0x43, 0xc2, 0x9c, 0x2c, 0x8b, 0x8b, 0x1f, 0x8e, 0xfd, 0x27, 0x79, 0x79, 0x6a, 0xa6, 0x32, + 0xa3, 0x26, 0xe5, 0xa2, 0xc4, 0x7f, 0x99, 0x00, 0x6d, 0x06, 0xb2, 0xbb, 0x67, 0x26, 0x49, 0x8b, + 0x87, 0xf3, 0x8e, 0x49, 0x02, 0x27, 0x69, 0x52, 0x8e, 0x94, 0xef, 0x7a, 0xe1, 0x75, 0xd1, 0xeb, + 0xec, 0xe0, 0x0f, 0x71, 0x76, 0x1b, 0xf1, 0x12, 0x2c, 0xea, 0x9f, 0x28, 0x7b, 0x30, 0xeb, 0xc1, + 0x64, 0x72, 0x8b, 0x1e, 0x2b, 0x3b, 0xf2, 0x07, 0x03, 0x30, 0x6e, 0x2e, 0x29, 0x74, 0x15, 0x8a, + 0x82, 0x88, 0x4a, 0x5a, 0xa0, 0x76, 0xc9, 0xaa, 0x04, 0xe0, 0x18, 0x87, 0xe5, 0xaa, 0x60, 0xd5, + 0x35, 0x33, 0xdb, 0x38, 0x57, 0x85, 0x82, 0x60, 0x0d, 0x8b, 0x3e, 0xac, 0x36, 0x7c, 0x3f, 0x52, + 0x17, 0x92, 0x5a, 0x77, 0x8b, 0xac, 0x14, 0x0b, 0x28, 0xbd, 0x88, 0x76, 0x48, 0xe0, 0x91, 0x86, + 0x19, 0xde, 0x58, 0x5d, 0x44, 0x37, 0x74, 0x20, 0x36, 0x71, 0xe9, 0x75, 0xea, 0x87, 0x6c, 0x21, + 0x8b, 0xe7, 0x5b, 0x6c, 0xb6, 0x5c, 0xe5, 0x7e, 0xde, 0x12, 0x8e, 0x3e, 0x0d, 0x8f, 0xa8, 0x10, + 0x4e, 0x98, 0xeb, 0x21, 0x64, 0x8b, 0x43, 0x86, 0xb4, 0xe5, 0x91, 0xa5, 0x74, 0x34, 0x9c, 0x55, + 0x1f, 0xbd, 0x06, 0xe3, 0x82, 0xc5, 0x97, 0x14, 0x87, 0x4d, 0xd3, 0x98, 0x1b, 0x06, 0x14, 0x27, + 0xb0, 0x65, 0x80, 0x66, 0xc6, 0x65, 0x4b, 0x0a, 0x85, 0xce, 0x00, 0xcd, 0x3a, 0x1c, 0x77, 0xd4, + 0x40, 0x0b, 0x30, 0xc1, 0x79, 0x30, 0xd7, 0xdb, 0xe2, 0x73, 0x22, 0x9c, 0xa2, 0xd4, 0x96, 0xba, + 0x65, 0x82, 0x71, 0x12, 0x1f, 0xbd, 0x04, 0xa3, 0x4e, 0x50, 0xdb, 0x76, 0x23, 0x52, 0x8b, 0xda, + 0x01, 0xf7, 0x96, 0xd2, 0x6c, 0x8b, 0x16, 0x34, 0x18, 0x36, 0x30, 0xed, 0x77, 0x61, 0x3a, 0x25, + 0x00, 0x04, 0x5d, 0x38, 0x4e, 0xcb, 0x95, 0xdf, 0x94, 0x30, 0x40, 0x5e, 0xa8, 0x94, 0xe5, 0xd7, + 0x68, 0x58, 0x74, 0x75, 0xb2, 0x40, 0x11, 0x5a, 0xbe, 0x43, 0xb5, 0x3a, 0x57, 0x24, 0x00, 0xc7, + 0x38, 0xf6, 0xff, 0xc8, 0xc1, 0x44, 0x8a, 0x6e, 0x85, 0xe5, 0xdc, 0x4b, 0x3c, 0x52, 0xe2, 0x14, + 0x7b, 0x66, 0xbc, 0xef, 0xdc, 0x21, 0xe2, 0x7d, 0xe7, 0x7b, 0xc5, 0xfb, 0x1e, 0x78, 0x2f, 0xf1, + 0xbe, 0xcd, 0x11, 0x1b, 0xec, 0x6b, 0xc4, 0x52, 0x62, 0x84, 0x0f, 0x1d, 0x32, 0x46, 0xb8, 0x31, + 0xe8, 0xc3, 0x7d, 0x0c, 0xfa, 0x0f, 0xe7, 0x60, 0x32, 0x69, 0x03, 0x79, 0x02, 0x72, 0xdb, 0xd7, + 0x0d, 0xb9, 0xed, 0x95, 0x7e, 0x5c, 0x5e, 0x33, 0x65, 0xb8, 0x38, 0x21, 0xc3, 0xfd, 0x68, 0x5f, + 0xd4, 0xba, 0xcb, 0x73, 0xff, 0x66, 0x0e, 0x4e, 0xa7, 0xfa, 0xdc, 0x9e, 0xc0, 0xd8, 0xdc, 0x32, + 0xc6, 0xe6, 0xe9, 0xbe, 0xdd, 0x81, 0x33, 0x07, 0xe8, 0x6e, 0x62, 0x80, 0xae, 0xf6, 0x4f, 0xb2, + 0xfb, 0x28, 0x7d, 0x35, 0x0f, 0x17, 0x52, 0xeb, 0xc5, 0x62, 0xcf, 0x15, 0x43, 0xec, 0xf9, 0x6c, + 0x42, 0xec, 0x69, 0x77, 0xaf, 0x7d, 0x34, 0x72, 0x50, 0xe1, 0xe8, 0xca, 0xa2, 0x19, 0x3c, 0xa4, + 0x0c, 0xd4, 0x70, 0x74, 0x55, 0x84, 0xb0, 0x49, 0xf7, 0x9b, 0x49, 0xf6, 0xf9, 0xaf, 0x2d, 0x38, + 0x9b, 0x3a, 0x37, 0x27, 0x20, 0xeb, 0x5a, 0x33, 0x65, 0x5d, 0x4f, 0xf4, 0xbd, 0x5a, 0x33, 0x84, + 0x5f, 0xbf, 0x3e, 0x90, 0xf1, 0x2d, 0xec, 0x25, 0x7f, 0x0b, 0x46, 0x9c, 0x5a, 0x8d, 0x84, 0xe1, + 0xaa, 0x5f, 0x57, 0x21, 0x8d, 0x9f, 0x66, 0xef, 0xac, 0xb8, 0xf8, 0x60, 0x7f, 0x6e, 0x36, 0x49, + 0x22, 0x06, 0x63, 0x9d, 0x02, 0xfa, 0x0c, 0x14, 0x42, 0x71, 0x6f, 0x8a, 0xb9, 0x7f, 0xae, 0xcf, + 0xc1, 0x71, 0x36, 0x48, 0xc3, 0x8c, 0xb9, 0xa4, 0x24, 0x15, 0x8a, 0xa4, 0x19, 0x9f, 0x25, 0x77, + 0xa4, 0xf1, 0x59, 0x9e, 0x05, 0xd8, 0x55, 0x8f, 0x81, 0xa4, 0xfc, 0x41, 0x7b, 0x26, 0x68, 0x58, + 0xe8, 0x93, 0x30, 0x19, 0xf2, 0xa0, 0x84, 0x4b, 0x0d, 0x27, 0x64, 0x6e, 0x2e, 0x62, 0x15, 0xb2, + 0xb8, 0x4e, 0xd5, 0x04, 0x0c, 0x77, 0x60, 0xa3, 0x15, 0xd9, 0x2a, 0x8b, 0xa0, 0xc8, 0x17, 0xe6, + 0xe5, 0xb8, 0x45, 0x91, 0xf1, 0xf7, 0x54, 0x72, 0xf8, 0xd9, 0xc0, 0x6b, 0x35, 0xd1, 0x67, 0x00, + 0xe8, 0xf2, 0x11, 0x72, 0x88, 0xe1, 0xec, 0xc3, 0x93, 0x9e, 0x2a, 0xf5, 0x54, 0xab, 0x5c, 0xe6, + 0x9b, 0x5a, 0x52, 0x44, 0xb0, 0x46, 0xd0, 0xfe, 0xe1, 0x01, 0x78, 0xb4, 0xcb, 0x19, 0x89, 0x16, + 0x4c, 0x3d, 0xec, 0x93, 0xc9, 0xc7, 0xf5, 0x6c, 0x6a, 0x65, 0xe3, 0xb5, 0x9d, 0x58, 0x8a, 0xb9, + 0xf7, 0xbc, 0x14, 0x7f, 0xc0, 0xd2, 0xc4, 0x1e, 0xdc, 0x56, 0xf3, 0x13, 0x87, 0x3c, 0xfb, 0x8f, + 0x50, 0x0e, 0xb2, 0x99, 0x22, 0x4c, 0x78, 0xb6, 0xef, 0xee, 0xf4, 0x2d, 0x5d, 0x38, 0x59, 0x29, + 0xf1, 0x6f, 0x5b, 0x70, 0xbe, 0x6b, 0x70, 0x8e, 0x6f, 0x40, 0x86, 0xc1, 0xfe, 0x82, 0x05, 0x8f, + 0xa5, 0xd6, 0x30, 0xcc, 0x8c, 0xae, 0x42, 0xb1, 0x46, 0x0b, 0x35, 0xff, 0xca, 0xd8, 0xf1, 0x5c, + 0x02, 0x70, 0x8c, 0x73, 0xc8, 0xc0, 0x23, 0xbf, 0x6a, 0x41, 0xc7, 0xa6, 0x3f, 0x81, 0xdb, 0xa7, + 0x6c, 0xde, 0x3e, 0x1f, 0xee, 0x67, 0x34, 0x33, 0x2e, 0x9e, 0x3f, 0x9e, 0x80, 0x33, 0x19, 0xfe, + 0x45, 0xbb, 0x30, 0xb5, 0x55, 0x23, 0xa6, 0xe7, 0x6a, 0xb7, 0xf8, 0x2f, 0x5d, 0xdd, 0x5c, 0x59, + 0x4e, 0xd2, 0xa9, 0x0e, 0x14, 0xdc, 0xd9, 0x04, 0xfa, 0x82, 0x05, 0xa7, 0x9c, 0x7b, 0xe1, 0x32, + 0xe5, 0x22, 0xdc, 0xda, 0x62, 0xc3, 0xaf, 0xed, 0xd0, 0x23, 0x5a, 0x6e, 0x84, 0xe7, 0x53, 0x25, + 0x3b, 0x77, 0xab, 0x1d, 0xf8, 0x46, 0xf3, 0x2c, 0x49, 0x6b, 0x1a, 0x16, 0x4e, 0x6d, 0x0b, 0x61, + 0x11, 0xb9, 0x9f, 0xbe, 0x51, 0xba, 0xf8, 0x56, 0xa7, 0x39, 0x82, 0xf1, 0x6b, 0x51, 0x42, 0xb0, + 0xa2, 0x83, 0x3e, 0x07, 0xc5, 0x2d, 0xe9, 0x9d, 0x99, 0x72, 0xed, 0xc6, 0x03, 0xd9, 0xdd, 0x67, + 0x95, 0xab, 0x67, 0x15, 0x12, 0x8e, 0x89, 0xa2, 0xd7, 0x20, 0xef, 0x6d, 0x86, 0xdd, 0xf2, 0x9c, + 0x26, 0xec, 0xf0, 0x78, 0x04, 0x83, 0xb5, 0x95, 0x2a, 0xa6, 0x15, 0xd1, 0x75, 0xc8, 0x07, 0x1b, + 0x75, 0x21, 0x96, 0x4c, 0xdd, 0xa4, 0x78, 0xb1, 0x94, 0xd1, 0x2b, 0x46, 0x09, 0x2f, 0x96, 0x30, + 0x25, 0x81, 0x2a, 0x30, 0xc8, 0x9c, 0x72, 0xc4, 0x25, 0x97, 0xca, 0xce, 0x77, 0x71, 0x6e, 0xe3, + 0x61, 0x0e, 0x18, 0x02, 0xe6, 0x84, 0xd0, 0x3a, 0x0c, 0xd5, 0x58, 0x4e, 0x4c, 0x11, 0xf1, 0xed, + 0x63, 0xa9, 0x02, 0xc8, 0x2e, 0xc9, 0x42, 0x85, 0x3c, 0x8e, 0x61, 0x60, 0x41, 0x8b, 0x51, 0x25, + 0xad, 0xed, 0xcd, 0x50, 0xe4, 0x70, 0x4e, 0xa7, 0xda, 0x25, 0x07, 0xae, 0xa0, 0xca, 0x30, 0xb0, + 0xa0, 0x85, 0x5e, 0x86, 0xdc, 0x66, 0x4d, 0x38, 0xdc, 0xa4, 0x4a, 0x22, 0xcd, 0x20, 0x14, 0x8b, + 0x43, 0x0f, 0xf6, 0xe7, 0x72, 0x2b, 0x4b, 0x38, 0xb7, 0x59, 0x43, 0x6b, 0x30, 0xbc, 0xc9, 0xdd, + 0xd6, 0x85, 0xb0, 0xf1, 0xf1, 0x74, 0x8f, 0xfa, 0x0e, 0xcf, 0x76, 0xee, 0x6b, 0x22, 0x00, 0x58, + 0x12, 0x61, 0x81, 0xf0, 0x95, 0xfb, 0xbd, 0x08, 0x8e, 0x36, 0x7f, 0xb8, 0x90, 0x09, 0x9c, 0xe9, + 0x88, 0x9d, 0xf8, 0xb1, 0x46, 0x91, 0xae, 0x6a, 0x47, 0x26, 0xd2, 0x17, 0x61, 0x62, 0x52, 0x57, + 0xb5, 0xca, 0xb6, 0xdf, 0x6d, 0x55, 0x2b, 0x24, 0x1c, 0x13, 0x45, 0x3b, 0x30, 0xb6, 0x1b, 0xb6, + 0xb6, 0x89, 0xdc, 0xd2, 0x2c, 0x6a, 0x4c, 0xc6, 0xbd, 0x7c, 0x47, 0x20, 0xba, 0x41, 0xd4, 0x76, + 0x1a, 0x1d, 0xa7, 0x10, 0xd3, 0xe9, 0xdf, 0xd1, 0x89, 0x61, 0x93, 0x36, 0x1d, 0xfe, 0x77, 0xda, + 0xfe, 0xc6, 0x5e, 0x44, 0x44, 0x4c, 0xb3, 0xd4, 0xe1, 0x7f, 0x83, 0xa3, 0x74, 0x0e, 0xbf, 0x00, + 0x60, 0x49, 0x04, 0xdd, 0x11, 0xc3, 0xc3, 0x4e, 0xcf, 0xc9, 0xec, 0xc0, 0xa3, 0x0b, 0x12, 0x29, + 0x63, 0x50, 0xd8, 0x69, 0x19, 0x93, 0x62, 0xa7, 0x64, 0x6b, 0xdb, 0x8f, 0x7c, 0x2f, 0x71, 0x42, + 0x4f, 0x65, 0x9f, 0x92, 0x95, 0x14, 0xfc, 0xce, 0x53, 0x32, 0x0d, 0x0b, 0xa7, 0xb6, 0x85, 0xea, + 0x30, 0xde, 0xf2, 0x83, 0xe8, 0x9e, 0x1f, 0xc8, 0xf5, 0x85, 0xba, 0x08, 0x4b, 0x0c, 0x4c, 0xd1, + 0x22, 0x0b, 0x17, 0x68, 0x42, 0x70, 0x82, 0x26, 0xfa, 0x14, 0x0c, 0x87, 0x35, 0xa7, 0x41, 0xca, + 0xb7, 0x66, 0xa6, 0xb3, 0xaf, 0x9f, 0x2a, 0x47, 0xc9, 0x58, 0x5d, 0x3c, 0x6a, 0x3e, 0x47, 0xc1, + 0x92, 0x1c, 0x5a, 0x81, 0x41, 0x96, 0xe8, 0x8c, 0x05, 0xe0, 0xcb, 0x88, 0x9f, 0xda, 0x61, 0x15, + 0xcd, 0xcf, 0x26, 0x56, 0x8c, 0x79, 0x75, 0xba, 0x07, 0xc4, 0x9b, 0xc1, 0x0f, 0x67, 0x4e, 0x67, + 0xef, 0x01, 0xf1, 0xd4, 0xb8, 0x55, 0xed, 0xb6, 0x07, 0x14, 0x12, 0x8e, 0x89, 0xd2, 0x93, 0x99, + 0x9e, 0xa6, 0x67, 0xba, 0x98, 0xf3, 0x64, 0x9e, 0xa5, 0xec, 0x64, 0xa6, 0x27, 0x29, 0x25, 0x61, + 0xff, 0xde, 0x70, 0x27, 0xcf, 0xc2, 0x5e, 0x99, 0xdf, 0x65, 0x75, 0x28, 0x20, 0x3f, 0xde, 0xaf, + 0xd0, 0xeb, 0x08, 0x59, 0xf0, 0x2f, 0x58, 0x70, 0xa6, 0x95, 0xfa, 0x21, 0x82, 0x01, 0xe8, 0x4f, + 0x76, 0xc6, 0x3f, 0x5d, 0x05, 0x6b, 0x4c, 0x87, 0xe3, 0x8c, 0x96, 0x92, 0xcf, 0x9c, 0xfc, 0x7b, + 0x7e, 0xe6, 0xac, 0x42, 0x81, 0x31, 0x99, 0x3d, 0x72, 0x44, 0x27, 0x5f, 0x7b, 0x8c, 0x95, 0x58, + 0x12, 0x15, 0xb1, 0x22, 0x81, 0x7e, 0xd0, 0x82, 0xf3, 0xc9, 0xae, 0x63, 0xc2, 0xc0, 0x22, 0xc2, + 0x23, 0x7f, 0xe0, 0xae, 0x88, 0xef, 0xef, 0xe0, 0xff, 0x0d, 0xe4, 0x83, 0x5e, 0x08, 0xb8, 0x7b, + 0x63, 0xa8, 0x94, 0xf2, 0xc2, 0x1e, 0x32, 0xb5, 0x0a, 0x7d, 0xbc, 0xb2, 0x9f, 0x87, 0xd1, 0xa6, + 0xdf, 0xf6, 0x22, 0x61, 0xfd, 0x23, 0x2c, 0x11, 0x98, 0x06, 0x7e, 0x55, 0x2b, 0xc7, 0x06, 0x56, + 0xe2, 0x6d, 0x5e, 0x78, 0xe8, 0xb7, 0xf9, 0x5b, 0x30, 0xea, 0x69, 0xe6, 0xaa, 0x82, 0x1f, 0xb8, + 0x9c, 0x1d, 0x9d, 0x55, 0x37, 0x6e, 0xe5, 0xbd, 0xd4, 0x4b, 0xb0, 0x41, 0xed, 0x64, 0x1f, 0x7c, + 0x5f, 0xb6, 0x52, 0x98, 0x7a, 0x2e, 0x02, 0x78, 0xd5, 0x14, 0x01, 0x5c, 0x4e, 0x8a, 0x00, 0x3a, + 0x24, 0xca, 0xc6, 0xeb, 0xbf, 0xff, 0xe4, 0x33, 0xfd, 0x86, 0x30, 0xb4, 0x1b, 0x70, 0xb1, 0xd7, + 0xb5, 0xc4, 0xcc, 0xc0, 0xea, 0x4a, 0x7f, 0x18, 0x9b, 0x81, 0xd5, 0xcb, 0x25, 0xcc, 0x20, 0xfd, + 0x06, 0xc7, 0xb1, 0xff, 0x9b, 0x05, 0xf9, 0x8a, 0x5f, 0x3f, 0x81, 0x07, 0xef, 0x27, 0x8c, 0x07, + 0xef, 0xa3, 0xe9, 0x17, 0x62, 0x3d, 0x53, 0x1e, 0xbe, 0x9c, 0x90, 0x87, 0x9f, 0xcf, 0x22, 0xd0, + 0x5d, 0xfa, 0xfd, 0x93, 0x79, 0x18, 0xa9, 0xf8, 0x75, 0x65, 0x83, 0xfd, 0xeb, 0x0f, 0x63, 0x83, + 0x9d, 0x99, 0x42, 0x41, 0xa3, 0xcc, 0xac, 0xc7, 0xa4, 0xe3, 0xe8, 0x37, 0x98, 0x29, 0xf6, 0x5d, + 0xe2, 0x6e, 0x6d, 0x47, 0xa4, 0x9e, 0xfc, 0x9c, 0x93, 0x33, 0xc5, 0xfe, 0xaf, 0x16, 0x4c, 0x24, + 0x5a, 0x47, 0x0d, 0x18, 0x6b, 0xe8, 0xd2, 0x56, 0xb1, 0x4e, 0x1f, 0x4a, 0x50, 0x2b, 0x4c, 0x59, + 0xb5, 0x22, 0x6c, 0x12, 0x47, 0xf3, 0x00, 0x4a, 0xfd, 0x28, 0xc5, 0x7a, 0x8c, 0xeb, 0x57, 0xfa, + 0xc9, 0x10, 0x6b, 0x18, 0xe8, 0x05, 0x18, 0x89, 0xfc, 0x96, 0xdf, 0xf0, 0xb7, 0xf6, 0x6e, 0x10, + 0x19, 0x8e, 0x49, 0x19, 0xa8, 0xad, 0xc7, 0x20, 0xac, 0xe3, 0xd9, 0x3f, 0x9d, 0xe7, 0x1f, 0xea, + 0x45, 0xee, 0x07, 0x6b, 0xf2, 0xfd, 0xbd, 0x26, 0xbf, 0x6a, 0xc1, 0x24, 0x6d, 0x9d, 0xd9, 0xc0, + 0xc8, 0xcb, 0x56, 0x45, 0x65, 0xb6, 0xba, 0x44, 0x65, 0xbe, 0x4c, 0xcf, 0xae, 0xba, 0xdf, 0x8e, + 0x84, 0x04, 0x4d, 0x3b, 0x9c, 0x68, 0x29, 0x16, 0x50, 0x81, 0x47, 0x82, 0x40, 0xf8, 0xed, 0xe9, + 0x78, 0x24, 0x08, 0xb0, 0x80, 0xca, 0xa0, 0xcd, 0x03, 0xe9, 0x41, 0x9b, 0x79, 0x70, 0x49, 0x61, + 0x2d, 0x21, 0xd8, 0x1e, 0x2d, 0xb8, 0xa4, 0x34, 0xa3, 0x88, 0x71, 0xec, 0x9f, 0xcb, 0xc3, 0x68, + 0xc5, 0xaf, 0xc7, 0x0a, 0xc0, 0xe7, 0x0d, 0x05, 0xe0, 0xc5, 0x84, 0x02, 0x70, 0x52, 0xc7, 0xfd, + 0x40, 0xdd, 0xf7, 0xf5, 0x52, 0xf7, 0xfd, 0x13, 0x8b, 0xcd, 0x5a, 0x69, 0xad, 0xca, 0x4d, 0xaa, + 0xd0, 0x33, 0x30, 0xc2, 0x0e, 0x24, 0xe6, 0x28, 0x2a, 0xb5, 0x62, 0x2c, 0x19, 0xd1, 0x5a, 0x5c, + 0x8c, 0x75, 0x1c, 0x74, 0x05, 0x0a, 0x21, 0x71, 0x82, 0xda, 0xb6, 0x3a, 0xe3, 0x84, 0x0a, 0x8b, + 0x97, 0x61, 0x05, 0x45, 0x6f, 0xc4, 0x71, 0x0d, 0xf3, 0xd9, 0x8e, 0x67, 0x7a, 0x7f, 0xf8, 0x16, + 0xc9, 0x0e, 0x66, 0x68, 0xdf, 0x05, 0xd4, 0x89, 0xdf, 0x47, 0x40, 0xaf, 0x39, 0x33, 0xa0, 0x57, + 0xb1, 0x23, 0x98, 0xd7, 0x9f, 0x5b, 0x30, 0x5e, 0xf1, 0xeb, 0x74, 0xeb, 0x7e, 0x33, 0xed, 0x53, + 0x3d, 0xa8, 0xeb, 0x50, 0x97, 0xa0, 0xae, 0x97, 0x60, 0xb0, 0xe2, 0xd7, 0xcb, 0x95, 0x6e, 0x0e, + 0xdb, 0xf6, 0xdf, 0xb2, 0x60, 0xb8, 0xe2, 0xd7, 0x4f, 0x40, 0x38, 0xff, 0xaa, 0x29, 0x9c, 0x7f, + 0x24, 0x63, 0xdd, 0x64, 0xc8, 0xe3, 0xff, 0xc6, 0x00, 0x8c, 0xd1, 0x7e, 0xfa, 0x5b, 0x72, 0x2a, + 0x8d, 0x61, 0xb3, 0xfa, 0x18, 0x36, 0xca, 0x0b, 0xfb, 0x8d, 0x86, 0x7f, 0x2f, 0x39, 0xad, 0x2b, + 0xac, 0x14, 0x0b, 0x28, 0x7a, 0x0a, 0x0a, 0xad, 0x80, 0xec, 0xba, 0xbe, 0x60, 0x32, 0x35, 0x55, + 0x47, 0x45, 0x94, 0x63, 0x85, 0x41, 0x1f, 0x67, 0xa1, 0xeb, 0xd5, 0x48, 0x95, 0xd4, 0x7c, 0xaf, + 0xce, 0xe5, 0xd7, 0x79, 0x91, 0x98, 0x41, 0x2b, 0xc7, 0x06, 0x16, 0xba, 0x0b, 0x45, 0xf6, 0x9f, + 0x1d, 0x3b, 0x87, 0x4f, 0xf1, 0x29, 0x52, 0xbe, 0x09, 0x02, 0x38, 0xa6, 0x85, 0x9e, 0x05, 0x88, + 0x64, 0xf4, 0xee, 0x50, 0x04, 0x6f, 0x52, 0x0c, 0xb9, 0x8a, 0xeb, 0x1d, 0x62, 0x0d, 0x0b, 0x3d, + 0x09, 0xc5, 0xc8, 0x71, 0x1b, 0x37, 0x5d, 0x8f, 0x84, 0x4c, 0x2e, 0x9d, 0x97, 0x99, 0xd7, 0x44, + 0x21, 0x8e, 0xe1, 0x94, 0x21, 0x62, 0x91, 0x0d, 0x78, 0x82, 0xe0, 0x02, 0xc3, 0x66, 0x0c, 0xd1, + 0x4d, 0x55, 0x8a, 0x35, 0x0c, 0xb4, 0x0d, 0xe7, 0x5c, 0x8f, 0x25, 0x31, 0x20, 0xd5, 0x1d, 0xb7, + 0xb5, 0x7e, 0xb3, 0x7a, 0x87, 0x04, 0xee, 0xe6, 0xde, 0xa2, 0x53, 0xdb, 0x21, 0x9e, 0x4c, 0xde, + 0xf8, 0x61, 0xd1, 0xc5, 0x73, 0xe5, 0x2e, 0xb8, 0xb8, 0x2b, 0x25, 0xfb, 0x25, 0x38, 0x5d, 0xf1, + 0xeb, 0x15, 0x3f, 0x88, 0x56, 0xfc, 0xe0, 0x9e, 0x13, 0xd4, 0xe5, 0x4a, 0x99, 0x93, 0x79, 0x5e, + 0xe8, 0x51, 0x38, 0xc8, 0x0f, 0x0a, 0x23, 0xdb, 0xd8, 0x73, 0x8c, 0xf9, 0x3a, 0xa4, 0x87, 0x4d, + 0x8d, 0xb1, 0x01, 0x2a, 0xa3, 0xc7, 0x35, 0x27, 0x22, 0xe8, 0x16, 0xcb, 0x54, 0x1c, 0xdf, 0x88, + 0xa2, 0xfa, 0x13, 0x5a, 0xa6, 0xe2, 0x18, 0x98, 0x7a, 0x85, 0x9a, 0xf5, 0xed, 0xff, 0x3e, 0xc8, + 0x0e, 0xc7, 0x44, 0x56, 0x08, 0xf4, 0x59, 0x18, 0x0f, 0xc9, 0x4d, 0xd7, 0x6b, 0xdf, 0x97, 0x32, + 0x81, 0x2e, 0x3e, 0x52, 0xd5, 0x65, 0x1d, 0x93, 0x4b, 0x16, 0xcd, 0x32, 0x9c, 0xa0, 0x86, 0x9a, + 0x30, 0x7e, 0xcf, 0xf5, 0xea, 0xfe, 0xbd, 0x50, 0xd2, 0x2f, 0x64, 0x0b, 0x18, 0xef, 0x72, 0xcc, + 0x44, 0x1f, 0x8d, 0xe6, 0xee, 0x1a, 0xc4, 0x70, 0x82, 0x38, 0x5d, 0x80, 0x41, 0xdb, 0x5b, 0x08, + 0x6f, 0x87, 0x24, 0x10, 0x39, 0xa7, 0xd9, 0x02, 0xc4, 0xb2, 0x10, 0xc7, 0x70, 0xba, 0x00, 0xd9, + 0x9f, 0x6b, 0x81, 0xdf, 0xe6, 0x31, 0xf6, 0xc5, 0x02, 0xc4, 0xaa, 0x14, 0x6b, 0x18, 0x74, 0x83, + 0xb2, 0x7f, 0x6b, 0xbe, 0x87, 0x7d, 0x3f, 0x92, 0x5b, 0x9a, 0x65, 0x39, 0xd5, 0xca, 0xb1, 0x81, + 0x85, 0x56, 0x00, 0x85, 0xed, 0x56, 0xab, 0xc1, 0x8c, 0x2f, 0x9c, 0x06, 0x23, 0xc5, 0x15, 0xdf, + 0x79, 0x1e, 0x7a, 0xb4, 0xda, 0x01, 0xc5, 0x29, 0x35, 0xe8, 0x59, 0xbd, 0x29, 0xba, 0x3a, 0xc8, + 0xba, 0xca, 0x95, 0x11, 0x55, 0xde, 0x4f, 0x09, 0x43, 0xcb, 0x30, 0x1c, 0xee, 0x85, 0xb5, 0x48, + 0xc4, 0x50, 0xcb, 0x48, 0xfc, 0x53, 0x65, 0x28, 0x5a, 0xde, 0x39, 0x5e, 0x05, 0xcb, 0xba, 0xa8, + 0x06, 0xd3, 0x82, 0xe2, 0xd2, 0xb6, 0xe3, 0xa9, 0x34, 0x2a, 0xdc, 0x06, 0xf5, 0x99, 0x07, 0xfb, + 0x73, 0xd3, 0xa2, 0x65, 0x1d, 0x7c, 0xb0, 0x3f, 0x77, 0xa6, 0xe2, 0xd7, 0x53, 0x20, 0x38, 0x8d, + 0x1a, 0x5f, 0x7c, 0xb5, 0x9a, 0xdf, 0x6c, 0x55, 0x02, 0x7f, 0xd3, 0x6d, 0x90, 0x6e, 0x0a, 0x9d, + 0xaa, 0x81, 0x29, 0x16, 0x9f, 0x51, 0x86, 0x13, 0xd4, 0xec, 0x6f, 0x67, 0xfc, 0x0c, 0x4b, 0xb3, + 0x1c, 0xb5, 0x03, 0x82, 0x9a, 0x30, 0xd6, 0x62, 0xdb, 0x44, 0x44, 0xbe, 0x17, 0x6b, 0xfd, 0xf9, + 0x3e, 0x05, 0x13, 0xf7, 0xe8, 0x35, 0xa0, 0x04, 0x87, 0xec, 0xc5, 0x57, 0xd1, 0xc9, 0x61, 0x93, + 0xba, 0xfd, 0x63, 0x8f, 0xb0, 0x1b, 0xb1, 0xca, 0xa5, 0x0d, 0xc3, 0xc2, 0xe4, 0x5d, 0x3c, 0xad, + 0x66, 0xb3, 0xc5, 0x5e, 0xf1, 0xb4, 0x08, 0xb3, 0x79, 0x2c, 0xeb, 0xa2, 0xcf, 0xc0, 0x38, 0x7d, + 0xa9, 0x68, 0xf9, 0x4b, 0x4e, 0x65, 0x87, 0x26, 0x88, 0xd3, 0x96, 0x68, 0x59, 0x31, 0xf4, 0xca, + 0x38, 0x41, 0x0c, 0xbd, 0xc1, 0x8c, 0x33, 0xcc, 0xd4, 0x28, 0x3d, 0x48, 0xeb, 0x76, 0x18, 0x92, + 0xac, 0x46, 0x24, 0x2b, 0xed, 0x8a, 0x7d, 0xbc, 0x69, 0x57, 0xd0, 0x4d, 0x18, 0x13, 0xb9, 0x86, + 0xc5, 0xca, 0xcd, 0x1b, 0xd2, 0xb8, 0x31, 0xac, 0x03, 0x0f, 0x92, 0x05, 0xd8, 0xac, 0x8c, 0xb6, + 0xe0, 0xbc, 0x96, 0xfb, 0xe7, 0x5a, 0xe0, 0x30, 0x95, 0xba, 0xcb, 0x8e, 0x53, 0xed, 0xae, 0x7e, + 0xec, 0xc1, 0xfe, 0xdc, 0xf9, 0xf5, 0x6e, 0x88, 0xb8, 0x3b, 0x1d, 0x74, 0x0b, 0x4e, 0x73, 0xc7, + 0xda, 0x12, 0x71, 0xea, 0x0d, 0xd7, 0x53, 0xcc, 0x00, 0xdf, 0xf2, 0x67, 0x1f, 0xec, 0xcf, 0x9d, + 0x5e, 0x48, 0x43, 0xc0, 0xe9, 0xf5, 0xd0, 0xab, 0x50, 0xac, 0x7b, 0xa1, 0x18, 0x83, 0x21, 0x23, + 0xbd, 0x52, 0xb1, 0xb4, 0x56, 0x55, 0xdf, 0x1f, 0xff, 0xc1, 0x71, 0x05, 0xb4, 0xc5, 0x25, 0xb6, + 0x4a, 0x40, 0x32, 0xdc, 0x11, 0x12, 0x28, 0x29, 0x6a, 0x33, 0x5c, 0xeb, 0xb8, 0xaa, 0x42, 0x59, + 0x9c, 0x1b, 0x5e, 0x77, 0x06, 0x61, 0xf4, 0x3a, 0x20, 0xfa, 0x82, 0x70, 0x6b, 0x64, 0xa1, 0xc6, + 0xd2, 0x2a, 0x30, 0x01, 0x77, 0xc1, 0x74, 0xf6, 0xaa, 0x76, 0x60, 0xe0, 0x94, 0x5a, 0xe8, 0x3a, + 0x3d, 0x55, 0xf4, 0x52, 0x71, 0x6a, 0xa9, 0x64, 0x78, 0x25, 0xd2, 0x0a, 0x48, 0xcd, 0x89, 0x48, + 0xdd, 0xa4, 0x88, 0x13, 0xf5, 0x50, 0x1d, 0xce, 0x39, 0xed, 0xc8, 0x67, 0xc2, 0x70, 0x13, 0x75, + 0xdd, 0xdf, 0x21, 0x1e, 0xd3, 0x43, 0x15, 0x16, 0x2f, 0x52, 0x6e, 0x63, 0xa1, 0x0b, 0x1e, 0xee, + 0x4a, 0x85, 0x72, 0x89, 0x2a, 0xfb, 0x2d, 0x98, 0x91, 0x8e, 0x52, 0x32, 0xe0, 0xbe, 0x00, 0x23, + 0xdb, 0x7e, 0x18, 0xad, 0x91, 0xe8, 0x9e, 0x1f, 0xec, 0x88, 0x78, 0x95, 0x71, 0x8c, 0xe3, 0x18, + 0x84, 0x75, 0x3c, 0xfa, 0x0c, 0x64, 0x56, 0x12, 0xe5, 0x12, 0x53, 0x50, 0x17, 0xe2, 0x33, 0xe6, + 0x3a, 0x2f, 0xc6, 0x12, 0x2e, 0x51, 0xcb, 0x95, 0x25, 0xa6, 0x6c, 0x4e, 0xa0, 0x96, 0x2b, 0x4b, + 0x58, 0xc2, 0xe9, 0x72, 0x0d, 0xb7, 0x9d, 0x80, 0x54, 0x02, 0xbf, 0x46, 0x42, 0x2d, 0xb2, 0xf6, + 0xa3, 0x3c, 0x1a, 0x27, 0x5d, 0xae, 0xd5, 0x34, 0x04, 0x9c, 0x5e, 0x0f, 0x91, 0xce, 0xbc, 0x57, + 0xe3, 0xd9, 0x5a, 0x82, 0x4e, 0x7e, 0xa6, 0xcf, 0xd4, 0x57, 0x1e, 0x4c, 0xaa, 0x8c, 0x5b, 0x3c, + 0xfe, 0x66, 0x38, 0x33, 0xc1, 0xd6, 0x76, 0xff, 0xc1, 0x3b, 0x95, 0xde, 0xa5, 0x9c, 0xa0, 0x84, + 0x3b, 0x68, 0x1b, 0xc1, 0xac, 0x26, 0x7b, 0xa6, 0x43, 0xbe, 0x0a, 0xc5, 0xb0, 0xbd, 0x51, 0xf7, + 0x9b, 0x8e, 0xeb, 0x31, 0x65, 0xb3, 0xf6, 0x1e, 0xa9, 0x4a, 0x00, 0x8e, 0x71, 0xd0, 0x0a, 0x14, + 0x1c, 0xa9, 0x54, 0x41, 0xd9, 0x31, 0x50, 0x94, 0x2a, 0x85, 0x87, 0x05, 0x90, 0x6a, 0x14, 0x55, + 0x17, 0xbd, 0x02, 0x63, 0xc2, 0x31, 0x54, 0x24, 0x7b, 0x9c, 0x36, 0xbd, 0x77, 0xaa, 0x3a, 0x10, + 0x9b, 0xb8, 0xe8, 0x36, 0x8c, 0x44, 0x7e, 0x83, 0xb9, 0xa0, 0x50, 0x36, 0xef, 0x4c, 0x76, 0x1c, + 0xb5, 0x75, 0x85, 0xa6, 0xcb, 0x33, 0x55, 0x55, 0xac, 0xd3, 0x41, 0xeb, 0x7c, 0xbd, 0xb3, 0x08, + 0xd3, 0x24, 0x9c, 0x79, 0x24, 0xfb, 0x4e, 0x52, 0x81, 0xa8, 0xcd, 0xed, 0x20, 0x6a, 0x62, 0x9d, + 0x0c, 0xba, 0x06, 0x53, 0xad, 0xc0, 0xf5, 0xd9, 0x9a, 0x50, 0xfa, 0xb4, 0x19, 0x33, 0xbd, 0x4d, + 0x25, 0x89, 0x80, 0x3b, 0xeb, 0x30, 0xbf, 0x5e, 0x51, 0x38, 0x73, 0x96, 0xe7, 0x83, 0xe6, 0xcf, + 0x3b, 0x5e, 0x86, 0x15, 0x14, 0xad, 0xb2, 0x93, 0x98, 0x4b, 0x26, 0x66, 0x66, 0xb3, 0xc3, 0xae, + 0xe8, 0x12, 0x0c, 0xce, 0xbc, 0xaa, 0xbf, 0x38, 0xa6, 0x80, 0xea, 0x5a, 0xe2, 0x40, 0xfa, 0x62, + 0x08, 0x67, 0xce, 0x75, 0x31, 0x55, 0x4b, 0x3c, 0x2f, 0x62, 0x86, 0xc0, 0x28, 0x0e, 0x71, 0x82, + 0x26, 0xfa, 0x24, 0x4c, 0x8a, 0x30, 0x6f, 0xf1, 0x30, 0x9d, 0x8f, 0x0d, 0x7b, 0x71, 0x02, 0x86, + 0x3b, 0xb0, 0x79, 0xe4, 0x7d, 0x67, 0xa3, 0x41, 0xc4, 0xd1, 0x77, 0xd3, 0xf5, 0x76, 0xc2, 0x99, + 0x0b, 0xec, 0x7c, 0x10, 0x91, 0xf7, 0x93, 0x50, 0x9c, 0x52, 0x03, 0xad, 0xc3, 0x64, 0x2b, 0x20, + 0xa4, 0xc9, 0x18, 0x7d, 0x71, 0x9f, 0xcd, 0x71, 0xb7, 0x76, 0xda, 0x93, 0x4a, 0x02, 0x76, 0x90, + 0x52, 0x86, 0x3b, 0x28, 0xa0, 0x7b, 0x50, 0xf0, 0x77, 0x49, 0xb0, 0x4d, 0x9c, 0xfa, 0xcc, 0xc5, + 0x2e, 0x86, 0xe6, 0xe2, 0x72, 0xbb, 0x25, 0x70, 0x13, 0x3a, 0x78, 0x59, 0xdc, 0x5b, 0x07, 0x2f, + 0x1b, 0x43, 0x3f, 0x64, 0xc1, 0x59, 0x29, 0xb6, 0xaf, 0xb6, 0xe8, 0xa8, 0x2f, 0xf9, 0x5e, 0x18, + 0x05, 0xdc, 0x11, 0xfb, 0xb1, 0x6c, 0xe7, 0xe4, 0xf5, 0x8c, 0x4a, 0x4a, 0x38, 0x7a, 0x36, 0x0b, + 0x23, 0xc4, 0xd9, 0x2d, 0xa2, 0x25, 0x98, 0x0a, 0x49, 0x24, 0x0f, 0xa3, 0x85, 0x70, 0xe5, 0x8d, + 0xd2, 0xda, 0xcc, 0x25, 0xee, 0x45, 0x4e, 0x37, 0x43, 0x35, 0x09, 0xc4, 0x9d, 0xf8, 0xb3, 0xdf, + 0x0a, 0x53, 0x1d, 0xd7, 0xff, 0x61, 0x32, 0x8a, 0xcc, 0xee, 0xc0, 0x98, 0x31, 0xc4, 0xc7, 0xaa, + 0xc3, 0xfd, 0x97, 0xc3, 0x50, 0x54, 0xfa, 0x3d, 0x74, 0xd5, 0x54, 0xdb, 0x9e, 0x4d, 0xaa, 0x6d, + 0x0b, 0xf4, 0x5d, 0xaf, 0x6b, 0x6a, 0xd7, 0x53, 0x62, 0x67, 0x65, 0x6d, 0xe8, 0xfe, 0x9d, 0xa2, + 0x35, 0x71, 0x6d, 0xbe, 0x6f, 0xfd, 0xef, 0x40, 0x57, 0x09, 0xf0, 0x35, 0x98, 0xf2, 0x7c, 0xc6, + 0x73, 0x92, 0xba, 0x64, 0x28, 0x18, 0xdf, 0x50, 0xd4, 0x83, 0x51, 0x24, 0x10, 0x70, 0x67, 0x1d, + 0xda, 0x20, 0xbf, 0xf8, 0x93, 0x22, 0x67, 0xce, 0x17, 0x60, 0x01, 0x45, 0x97, 0x60, 0xb0, 0xe5, + 0xd7, 0xcb, 0x15, 0xc1, 0x6f, 0x6a, 0xa9, 0x6e, 0xeb, 0xe5, 0x0a, 0xe6, 0x30, 0xb4, 0x00, 0x43, + 0xec, 0x47, 0x38, 0x33, 0x9a, 0x1d, 0x75, 0x80, 0xd5, 0xd0, 0xf2, 0xb5, 0xb0, 0x0a, 0x58, 0x54, + 0x64, 0xa2, 0x2f, 0xca, 0xa4, 0x33, 0xd1, 0xd7, 0xf0, 0x43, 0x8a, 0xbe, 0x24, 0x01, 0x1c, 0xd3, + 0x42, 0xf7, 0xe1, 0xb4, 0xf1, 0x30, 0xe2, 0x4b, 0x84, 0x84, 0xc2, 0xf3, 0xf9, 0x52, 0xd7, 0x17, + 0x91, 0xd0, 0x17, 0x9f, 0x17, 0x9d, 0x3e, 0x5d, 0x4e, 0xa3, 0x84, 0xd3, 0x1b, 0x40, 0x0d, 0x98, + 0xaa, 0x75, 0xb4, 0x5a, 0xe8, 0xbf, 0x55, 0x35, 0xa1, 0x9d, 0x2d, 0x76, 0x12, 0x46, 0xaf, 0x40, + 0xe1, 0x1d, 0x3f, 0x64, 0x67, 0xb5, 0xe0, 0x91, 0xa5, 0xdb, 0x6c, 0xe1, 0x8d, 0x5b, 0x55, 0x56, + 0x7e, 0xb0, 0x3f, 0x37, 0x52, 0xf1, 0xeb, 0xf2, 0x2f, 0x56, 0x15, 0xd0, 0xf7, 0x5a, 0x30, 0xdb, + 0xf9, 0xf2, 0x52, 0x9d, 0x1e, 0xeb, 0xbf, 0xd3, 0xb6, 0x68, 0x74, 0x76, 0x39, 0x93, 0x1c, 0xee, + 0xd2, 0x94, 0xfd, 0xcb, 0x5c, 0xb7, 0x2b, 0x34, 0x40, 0x24, 0x6c, 0x37, 0x4e, 0x22, 0x4d, 0xe5, + 0xb2, 0xa1, 0x9c, 0x7a, 0x68, 0xfb, 0x81, 0x7f, 0x66, 0x31, 0xfb, 0x81, 0x13, 0x74, 0x14, 0x78, + 0x03, 0x0a, 0x91, 0x4c, 0x36, 0xda, 0x25, 0xb3, 0xa6, 0xd6, 0x29, 0x66, 0x43, 0xa1, 0x38, 0x56, + 0x95, 0x57, 0x54, 0x91, 0xb1, 0xff, 0x21, 0x9f, 0x01, 0x09, 0x39, 0x01, 0x1d, 0x40, 0xc9, 0xd4, + 0x01, 0xcc, 0xf5, 0xf8, 0x82, 0x0c, 0x5d, 0xc0, 0x3f, 0x30, 0xfb, 0xcd, 0x24, 0x35, 0xef, 0x77, + 0xc3, 0x15, 0xfb, 0x47, 0x2c, 0x38, 0x95, 0x66, 0xe9, 0x49, 0x5f, 0x19, 0x5c, 0x4e, 0xa4, 0x0c, + 0x79, 0xd4, 0x08, 0xde, 0x11, 0xe5, 0x58, 0x61, 0xf4, 0x9d, 0xed, 0xea, 0x70, 0xd1, 0x5f, 0x6f, + 0xc1, 0x58, 0x25, 0x20, 0xda, 0x85, 0xf6, 0x1a, 0x77, 0xa3, 0xe6, 0xfd, 0x79, 0xea, 0xd0, 0x2e, + 0xd4, 0xf6, 0xcf, 0xe4, 0xe0, 0x14, 0xd7, 0xc4, 0x2f, 0xec, 0xfa, 0x6e, 0xbd, 0xe2, 0xd7, 0x45, + 0xa6, 0xb2, 0x37, 0x61, 0xb4, 0xa5, 0x09, 0xf7, 0xba, 0x45, 0x32, 0xd4, 0x85, 0x80, 0xb1, 0x38, + 0x42, 0x2f, 0xc5, 0x06, 0x2d, 0x54, 0x87, 0x51, 0xb2, 0xeb, 0xd6, 0x94, 0x3a, 0x37, 0x77, 0xe8, + 0xcb, 0x45, 0xb5, 0xb2, 0xac, 0xd1, 0xc1, 0x06, 0xd5, 0x63, 0xc8, 0x41, 0x6b, 0xff, 0xa8, 0x05, + 0x8f, 0x64, 0xc4, 0x3d, 0xa4, 0xcd, 0xdd, 0x63, 0x36, 0x0f, 0x22, 0x9d, 0xa5, 0x6a, 0x8e, 0x5b, + 0x42, 0x60, 0x01, 0x45, 0x9f, 0x02, 0xe0, 0x96, 0x0c, 0xf4, 0x99, 0xdb, 0x2b, 0x40, 0x9c, 0x11, + 0xdb, 0x4a, 0x0b, 0x53, 0x24, 0xeb, 0x63, 0x8d, 0x96, 0xfd, 0x53, 0x79, 0x18, 0xe4, 0xe9, 0xc4, + 0x57, 0x60, 0x78, 0x9b, 0xe7, 0x6f, 0xe8, 0x27, 0x55, 0x44, 0x2c, 0x80, 0xe0, 0x05, 0x58, 0x56, + 0x46, 0xab, 0x30, 0xcd, 0xf3, 0x5f, 0x34, 0x4a, 0xa4, 0xe1, 0xec, 0x49, 0x69, 0x19, 0xcf, 0x1d, + 0xa9, 0xa4, 0x86, 0xe5, 0x4e, 0x14, 0x9c, 0x56, 0x0f, 0xbd, 0x06, 0xe3, 0xf4, 0xf5, 0xe2, 0xb7, + 0x23, 0x49, 0x89, 0x67, 0xbe, 0x50, 0xcf, 0xa5, 0x75, 0x03, 0x8a, 0x13, 0xd8, 0xf4, 0x01, 0xdd, + 0xea, 0x90, 0x0b, 0x0e, 0xc6, 0x0f, 0x68, 0x53, 0x16, 0x68, 0xe2, 0x32, 0x13, 0xcf, 0x36, 0x33, + 0x68, 0x5d, 0xdf, 0x0e, 0x48, 0xb8, 0xed, 0x37, 0xea, 0x8c, 0xd1, 0x1a, 0xd4, 0x4c, 0x3c, 0x13, + 0x70, 0xdc, 0x51, 0x83, 0x52, 0xd9, 0x74, 0xdc, 0x46, 0x3b, 0x20, 0x31, 0x95, 0x21, 0x93, 0xca, + 0x4a, 0x02, 0x8e, 0x3b, 0x6a, 0xd0, 0x75, 0x74, 0xba, 0x12, 0xf8, 0xf4, 0xf0, 0x92, 0xc1, 0x5c, + 0x94, 0xdd, 0xee, 0xb0, 0xf4, 0x3b, 0xed, 0x12, 0xf6, 0x4c, 0x58, 0x36, 0x72, 0x0a, 0x86, 0xd2, + 0xbe, 0x2a, 0x3c, 0x4e, 0x25, 0x15, 0xf4, 0x0c, 0x8c, 0x88, 0xac, 0x06, 0xcc, 0xbc, 0x94, 0x4f, + 0x1d, 0x33, 0x32, 0x28, 0xc5, 0xc5, 0x58, 0xc7, 0xb1, 0xbf, 0x2f, 0x07, 0xd3, 0x29, 0xfe, 0x01, + 0xfc, 0xa8, 0xda, 0x72, 0xc3, 0x48, 0xe5, 0xc7, 0xd3, 0x8e, 0x2a, 0x5e, 0x8e, 0x15, 0x06, 0xdd, + 0x0f, 0xfc, 0x30, 0x4c, 0x1e, 0x80, 0xc2, 0xfe, 0x56, 0x40, 0x0f, 0x99, 0x69, 0xee, 0x22, 0x0c, + 0xb4, 0x43, 0x22, 0x03, 0x16, 0xaa, 0xf3, 0x9b, 0xe9, 0x9e, 0x18, 0x84, 0xb2, 0xc7, 0x5b, 0x4a, + 0x8d, 0xa3, 0xb1, 0xc7, 0x5c, 0x91, 0xc3, 0x61, 0xb4, 0x73, 0x11, 0xf1, 0x1c, 0x2f, 0x12, 0x4c, + 0x74, 0x1c, 0x79, 0x8b, 0x95, 0x62, 0x01, 0xb5, 0xbf, 0x94, 0x87, 0xb3, 0x99, 0x1e, 0x43, 0xb4, + 0xeb, 0x4d, 0xdf, 0x73, 0x23, 0x5f, 0x59, 0x6f, 0xf0, 0x68, 0x5b, 0xa4, 0xb5, 0xbd, 0x2a, 0xca, + 0xb1, 0xc2, 0x40, 0x97, 0x61, 0x90, 0x49, 0xae, 0x3a, 0x32, 0x05, 0x2e, 0x96, 0x78, 0xf8, 0x15, + 0x0e, 0xee, 0x3b, 0x0b, 0xeb, 0x25, 0x18, 0x68, 0xf9, 0x7e, 0x23, 0x79, 0x68, 0xd1, 0xee, 0xfa, + 0x7e, 0x03, 0x33, 0x20, 0xfa, 0x88, 0x18, 0xaf, 0x84, 0xb9, 0x02, 0x76, 0xea, 0x7e, 0xa8, 0x0d, + 0xda, 0x13, 0x30, 0xbc, 0x43, 0xf6, 0x02, 0xd7, 0xdb, 0x4a, 0x9a, 0xb1, 0xdc, 0xe0, 0xc5, 0x58, + 0xc2, 0xcd, 0xa4, 0x4f, 0xc3, 0x47, 0x9d, 0x3e, 0xb5, 0xd0, 0xf3, 0x0a, 0xfc, 0x81, 0x3c, 0x4c, + 0xe0, 0xc5, 0xd2, 0x07, 0x13, 0x71, 0xbb, 0x73, 0x22, 0x8e, 0x3a, 0x7d, 0x6a, 0xef, 0xd9, 0xf8, + 0x05, 0x0b, 0x26, 0x58, 0x6e, 0x05, 0x11, 0xa7, 0xc9, 0xf5, 0xbd, 0x13, 0x60, 0xf1, 0x2e, 0xc1, + 0x60, 0x40, 0x1b, 0x4d, 0xa6, 0x08, 0x64, 0x3d, 0xc1, 0x1c, 0x86, 0xce, 0xc1, 0x00, 0xeb, 0x02, + 0x9d, 0xbc, 0x51, 0x9e, 0x5d, 0xa9, 0xe4, 0x44, 0x0e, 0x66, 0xa5, 0x2c, 0xf8, 0x08, 0x26, 0xad, + 0x86, 0xcb, 0x3b, 0x1d, 0xeb, 0x15, 0xdf, 0x1f, 0xbe, 0xc4, 0xa9, 0x5d, 0x7b, 0x6f, 0xc1, 0x47, + 0xd2, 0x49, 0x76, 0x7f, 0x3e, 0xfd, 0x51, 0x0e, 0x2e, 0xa4, 0xd6, 0xeb, 0x3b, 0xf8, 0x48, 0xf7, + 0xda, 0xc7, 0x19, 0x83, 0x3f, 0x7f, 0x82, 0x46, 0x82, 0x03, 0xfd, 0x72, 0x98, 0x83, 0x7d, 0xc4, + 0x04, 0x49, 0x1d, 0xb2, 0xf7, 0x49, 0x4c, 0x90, 0xd4, 0xbe, 0x65, 0x3c, 0xff, 0xfe, 0x22, 0x97, + 0xf1, 0x2d, 0xec, 0x21, 0x78, 0x85, 0x9e, 0x33, 0x0c, 0x18, 0x0a, 0x8e, 0x79, 0x94, 0x9f, 0x31, + 0xbc, 0x0c, 0x2b, 0x28, 0x5a, 0x80, 0x89, 0xa6, 0xeb, 0xd1, 0xc3, 0x67, 0xcf, 0x64, 0xfc, 0x54, + 0xc8, 0xa6, 0x55, 0x13, 0x8c, 0x93, 0xf8, 0xc8, 0xd5, 0xe2, 0x85, 0xe4, 0xb2, 0x93, 0x6e, 0x67, + 0xf6, 0x76, 0xde, 0xd4, 0xb9, 0xaa, 0x51, 0x4c, 0x89, 0x1d, 0xb2, 0xaa, 0xbd, 0xff, 0xf3, 0xfd, + 0xbf, 0xff, 0x47, 0xd3, 0xdf, 0xfe, 0xb3, 0xaf, 0xc0, 0xd8, 0x43, 0x0b, 0x7c, 0xed, 0xaf, 0xe6, + 0xe1, 0xd1, 0x2e, 0xdb, 0x9e, 0x9f, 0xf5, 0xc6, 0x1c, 0x68, 0x67, 0x7d, 0xc7, 0x3c, 0x54, 0xe0, + 0xd4, 0x66, 0xbb, 0xd1, 0xd8, 0x63, 0x76, 0xf8, 0xa4, 0x2e, 0x31, 0x04, 0x4f, 0x79, 0x4e, 0xe6, + 0xb3, 0x5a, 0x49, 0xc1, 0xc1, 0xa9, 0x35, 0x29, 0x43, 0x4f, 0x6f, 0x92, 0x3d, 0x45, 0x2a, 0xc1, + 0xd0, 0x63, 0x1d, 0x88, 0x4d, 0x5c, 0x74, 0x0d, 0xa6, 0x9c, 0x5d, 0xc7, 0xe5, 0x41, 0x57, 0x25, + 0x01, 0xce, 0xd1, 0x2b, 0x39, 0xdd, 0x42, 0x12, 0x01, 0x77, 0xd6, 0x41, 0xaf, 0x03, 0xf2, 0x45, + 0xee, 0xff, 0x6b, 0xc4, 0x13, 0xaa, 0x31, 0x36, 0x77, 0xf9, 0xf8, 0x48, 0xb8, 0xd5, 0x81, 0x81, + 0x53, 0x6a, 0x25, 0xe2, 0x6f, 0x0c, 0x65, 0xc7, 0xdf, 0xe8, 0x7e, 0x2e, 0xf6, 0x4c, 0xff, 0xf0, + 0x9f, 0x2c, 0x7a, 0x7d, 0x71, 0x26, 0xdf, 0x0c, 0x23, 0xf7, 0x0a, 0x33, 0x6d, 0xe3, 0x32, 0x3c, + 0x2d, 0x6a, 0xc4, 0x69, 0xcd, 0xb4, 0x2d, 0x06, 0x62, 0x13, 0x97, 0x2f, 0x88, 0x30, 0x76, 0x56, + 0x34, 0x58, 0x7c, 0x11, 0xeb, 0x46, 0x61, 0xa0, 0x4f, 0xc3, 0x70, 0xdd, 0xdd, 0x75, 0x43, 0x3f, + 0x10, 0x2b, 0xfd, 0x90, 0xea, 0x82, 0xf8, 0x1c, 0x2c, 0x71, 0x32, 0x58, 0xd2, 0xb3, 0x7f, 0x20, + 0x07, 0x63, 0xb2, 0xc5, 0x37, 0xda, 0x7e, 0xe4, 0x9c, 0xc0, 0xb5, 0x7c, 0xcd, 0xb8, 0x96, 0x3f, + 0xd2, 0x2d, 0xe0, 0x0f, 0xeb, 0x52, 0xe6, 0x75, 0x7c, 0x2b, 0x71, 0x1d, 0x3f, 0xde, 0x9b, 0x54, + 0xf7, 0x6b, 0xf8, 0x1f, 0x59, 0x30, 0x65, 0xe0, 0x9f, 0xc0, 0x6d, 0xb0, 0x62, 0xde, 0x06, 0x8f, + 0xf5, 0xfc, 0x86, 0x8c, 0x5b, 0xe0, 0xbb, 0xf3, 0x89, 0xbe, 0xb3, 0xd3, 0xff, 0x1d, 0x18, 0xd8, + 0x76, 0x82, 0x7a, 0xb7, 0x00, 0xe7, 0x1d, 0x95, 0xe6, 0xaf, 0x3b, 0x81, 0xd0, 0x0d, 0x3e, 0xa5, + 0x72, 0x5e, 0x3b, 0x41, 0x6f, 0xbd, 0x20, 0x6b, 0x0a, 0xbd, 0x04, 0x43, 0x61, 0xcd, 0x6f, 0x29, + 0xcb, 0xf9, 0x8b, 0x3c, 0x1f, 0x36, 0x2d, 0x39, 0xd8, 0x9f, 0x43, 0x66, 0x73, 0xb4, 0x18, 0x0b, + 0x7c, 0xf4, 0x26, 0x8c, 0xb1, 0x5f, 0xca, 0x50, 0x27, 0x9f, 0x9d, 0x0c, 0xa9, 0xaa, 0x23, 0x72, + 0x2b, 0x36, 0xa3, 0x08, 0x9b, 0xa4, 0x66, 0xb7, 0xa0, 0xa8, 0x3e, 0xeb, 0x58, 0xf5, 0x71, 0xff, + 0x2e, 0x0f, 0xd3, 0x29, 0x6b, 0x0e, 0x85, 0xc6, 0x4c, 0x3c, 0xd3, 0xe7, 0x52, 0x7d, 0x8f, 0x73, + 0x11, 0xb2, 0xd7, 0x50, 0x5d, 0xac, 0xad, 0xbe, 0x1b, 0xbd, 0x1d, 0x92, 0x64, 0xa3, 0xb4, 0xa8, + 0x77, 0xa3, 0xb4, 0xb1, 0x13, 0x1b, 0x6a, 0xda, 0x90, 0xea, 0xe9, 0xb1, 0xce, 0xe9, 0x9f, 0xe6, + 0xe1, 0x54, 0x5a, 0x0c, 0x32, 0xf4, 0xf9, 0x44, 0x62, 0xbc, 0xe7, 0xfb, 0x8d, 0x5e, 0xc6, 0xb3, + 0xe5, 0x71, 0x19, 0xf0, 0xe2, 0xbc, 0x99, 0x2a, 0xaf, 0xe7, 0x30, 0x8b, 0x36, 0x99, 0x23, 0x7e, + 0xc0, 0x13, 0x1a, 0xca, 0xe3, 0xe3, 0xe3, 0x7d, 0x77, 0x40, 0x64, 0x42, 0x0c, 0x13, 0x46, 0x00, + 0xb2, 0xb8, 0xb7, 0x11, 0x80, 0x6c, 0x79, 0xd6, 0x85, 0x11, 0xed, 0x6b, 0x8e, 0x75, 0xc6, 0x77, + 0xe8, 0x6d, 0xa5, 0xf5, 0xfb, 0x58, 0x67, 0xfd, 0x47, 0x2d, 0x48, 0xd8, 0x85, 0x2b, 0xb1, 0x98, + 0x95, 0x29, 0x16, 0xbb, 0x08, 0x03, 0x81, 0xdf, 0x20, 0xc9, 0x3c, 0x74, 0xd8, 0x6f, 0x10, 0xcc, + 0x20, 0x14, 0x23, 0x8a, 0x85, 0x1d, 0xa3, 0xfa, 0x43, 0x4e, 0x3c, 0xd1, 0x2e, 0xc1, 0x60, 0x83, + 0xec, 0x92, 0x46, 0x32, 0x5d, 0xc8, 0x4d, 0x5a, 0x88, 0x39, 0xcc, 0xfe, 0x85, 0x01, 0x38, 0xdf, + 0x35, 0x94, 0x05, 0x7d, 0x0e, 0x6d, 0x39, 0x11, 0xb9, 0xe7, 0xec, 0x25, 0xe3, 0xfa, 0x5f, 0xe3, + 0xc5, 0x58, 0xc2, 0x99, 0xe7, 0x0e, 0x0f, 0xcf, 0x9b, 0x10, 0x22, 0x8a, 0xa8, 0xbc, 0x02, 0x6a, + 0x0a, 0xa5, 0xf2, 0x47, 0x21, 0x94, 0x7a, 0x16, 0x20, 0x0c, 0x1b, 0xdc, 0x7a, 0xa6, 0x2e, 0x5c, + 0x82, 0xe2, 0x30, 0xce, 0xd5, 0x9b, 0x02, 0x82, 0x35, 0x2c, 0x54, 0x82, 0xc9, 0x56, 0xe0, 0x47, + 0x5c, 0x26, 0x5b, 0xe2, 0x06, 0x66, 0x83, 0x66, 0x14, 0x81, 0x4a, 0x02, 0x8e, 0x3b, 0x6a, 0xa0, + 0x17, 0x60, 0x44, 0x44, 0x16, 0xa8, 0xf8, 0x7e, 0x43, 0x88, 0x81, 0x94, 0xcd, 0x55, 0x35, 0x06, + 0x61, 0x1d, 0x4f, 0xab, 0xc6, 0x04, 0xbd, 0xc3, 0xa9, 0xd5, 0xb8, 0xb0, 0x57, 0xc3, 0x4b, 0xc4, + 0x23, 0x2c, 0xf4, 0x15, 0x8f, 0x30, 0x16, 0x8c, 0x15, 0xfb, 0xd6, 0x6d, 0x41, 0x4f, 0x51, 0xd2, + 0xcf, 0x0e, 0xc0, 0xb4, 0x58, 0x38, 0xc7, 0xbd, 0x5c, 0x6e, 0x77, 0x2e, 0x97, 0xa3, 0x10, 0x9d, + 0x7d, 0xb0, 0x66, 0x4e, 0x7a, 0xcd, 0xfc, 0xa0, 0x05, 0x26, 0x7b, 0x85, 0xfe, 0xdf, 0xcc, 0xc4, + 0x28, 0x2f, 0x64, 0xb2, 0x6b, 0x75, 0x79, 0x81, 0xbc, 0xc7, 0x14, 0x29, 0xf6, 0x7f, 0xb0, 0xe0, + 0xb1, 0x9e, 0x14, 0xd1, 0x32, 0x14, 0x19, 0x0f, 0xa8, 0xbd, 0xce, 0x1e, 0x57, 0x06, 0xa8, 0x12, + 0x90, 0xc1, 0x92, 0xc6, 0x35, 0xd1, 0x72, 0x47, 0x06, 0x9a, 0x27, 0x52, 0x32, 0xd0, 0x9c, 0x36, + 0x86, 0xe7, 0x21, 0x53, 0xd0, 0x7c, 0x3f, 0xbd, 0x71, 0x0c, 0xe7, 0x0f, 0xf4, 0x71, 0x43, 0xec, + 0x67, 0x27, 0xc4, 0x7e, 0xc8, 0xc4, 0xd6, 0xee, 0x90, 0x4f, 0xc2, 0x24, 0x0b, 0x39, 0xc4, 0xcc, + 0xa1, 0x85, 0x5b, 0x4a, 0x2e, 0x36, 0x79, 0xbc, 0x99, 0x80, 0xe1, 0x0e, 0x6c, 0xfb, 0x0f, 0xf3, + 0x30, 0xc4, 0xb7, 0xdf, 0x09, 0xbc, 0x09, 0x9f, 0x84, 0xa2, 0xdb, 0x6c, 0xb6, 0x79, 0x52, 0x91, + 0x41, 0xee, 0x8b, 0x4a, 0xe7, 0xa9, 0x2c, 0x0b, 0x71, 0x0c, 0x47, 0x2b, 0x42, 0xe2, 0xdc, 0x25, + 0xaa, 0x21, 0xef, 0xf8, 0x7c, 0xc9, 0x89, 0x1c, 0xce, 0xe0, 0xa8, 0x7b, 0x36, 0x96, 0x4d, 0xa3, + 0xcf, 0x02, 0x84, 0x51, 0xe0, 0x7a, 0x5b, 0xb4, 0x4c, 0x04, 0xf1, 0xfc, 0x68, 0x17, 0x6a, 0x55, + 0x85, 0xcc, 0x69, 0xc6, 0x67, 0x8e, 0x02, 0x60, 0x8d, 0x22, 0x9a, 0x37, 0x6e, 0xfa, 0xd9, 0xc4, + 0xdc, 0x01, 0xa7, 0x1a, 0xcf, 0xd9, 0xec, 0x8b, 0x50, 0x54, 0xc4, 0x7b, 0xc9, 0x9f, 0x46, 0x75, + 0xb6, 0xe8, 0x13, 0x30, 0x91, 0xe8, 0xdb, 0xa1, 0xc4, 0x57, 0xbf, 0x68, 0xc1, 0x04, 0xef, 0xcc, + 0xb2, 0xb7, 0x2b, 0x6e, 0x83, 0x77, 0xe1, 0x54, 0x23, 0xe5, 0x54, 0x16, 0xd3, 0xdf, 0xff, 0x29, + 0xae, 0xc4, 0x55, 0x69, 0x50, 0x9c, 0xda, 0x06, 0xba, 0x42, 0x77, 0x1c, 0x3d, 0x75, 0x9d, 0x86, + 0x70, 0x4d, 0x1d, 0xe5, 0xbb, 0x8d, 0x97, 0x61, 0x05, 0xb5, 0x7f, 0xc7, 0x82, 0x29, 0xde, 0xf3, + 0x1b, 0x64, 0x4f, 0x9d, 0x4d, 0x5f, 0xcf, 0xbe, 0x8b, 0x74, 0x56, 0xb9, 0x8c, 0x74, 0x56, 0xfa, + 0xa7, 0xe5, 0xbb, 0x7e, 0xda, 0xcf, 0x58, 0x20, 0x56, 0xc8, 0x09, 0x08, 0x21, 0xbe, 0xd5, 0x14, + 0x42, 0xcc, 0x66, 0x6f, 0x82, 0x0c, 0xe9, 0xc3, 0x9f, 0x5b, 0x30, 0xc9, 0x11, 0x62, 0x6d, 0xf9, + 0xd7, 0x75, 0x1e, 0xfa, 0x49, 0x7a, 0x7b, 0x83, 0xec, 0xad, 0xfb, 0x15, 0x27, 0xda, 0x4e, 0xff, + 0x28, 0x63, 0xb2, 0x06, 0xba, 0x4e, 0x56, 0x5d, 0x6e, 0xa0, 0x43, 0x64, 0xd2, 0x3e, 0x74, 0xb6, + 0x07, 0xfb, 0x6b, 0x16, 0x20, 0xde, 0x8c, 0xc1, 0xb8, 0x51, 0x76, 0x88, 0x95, 0x6a, 0x17, 0x5d, + 0x7c, 0x34, 0x29, 0x08, 0xd6, 0xb0, 0x8e, 0x64, 0x78, 0x12, 0x26, 0x0f, 0xf9, 0xde, 0x26, 0x0f, + 0x87, 0x18, 0xd1, 0x7f, 0x35, 0x04, 0x49, 0x07, 0x18, 0x74, 0x07, 0x46, 0x6b, 0x4e, 0xcb, 0xd9, + 0x70, 0x1b, 0x6e, 0xe4, 0x92, 0xb0, 0x9b, 0xad, 0xd4, 0x92, 0x86, 0x27, 0x94, 0xd4, 0x5a, 0x09, + 0x36, 0xe8, 0xa0, 0x79, 0x80, 0x56, 0xe0, 0xee, 0xba, 0x0d, 0xb2, 0xc5, 0x64, 0x25, 0xcc, 0x19, + 0x9e, 0x1b, 0x00, 0xc9, 0x52, 0xac, 0x61, 0xa4, 0x78, 0x1b, 0xe7, 0x8f, 0xd9, 0xdb, 0x18, 0x4e, + 0xcc, 0xdb, 0x78, 0xe0, 0x50, 0xde, 0xc6, 0x85, 0x43, 0x7b, 0x1b, 0x0f, 0xf6, 0xe5, 0x6d, 0x8c, + 0xe1, 0x8c, 0xe4, 0x3d, 0xe9, 0xff, 0x15, 0xb7, 0x41, 0xc4, 0x83, 0x83, 0x7b, 0xf0, 0xcf, 0x3e, + 0xd8, 0x9f, 0x3b, 0x83, 0x53, 0x31, 0x70, 0x46, 0x4d, 0xf4, 0x29, 0x98, 0x71, 0x1a, 0x0d, 0xff, + 0x9e, 0x9a, 0xd4, 0xe5, 0xb0, 0xe6, 0x34, 0xb8, 0x12, 0x62, 0x98, 0x51, 0x3d, 0xf7, 0x60, 0x7f, + 0x6e, 0x66, 0x21, 0x03, 0x07, 0x67, 0xd6, 0x46, 0xaf, 0x42, 0xb1, 0x15, 0xf8, 0xb5, 0x55, 0xcd, + 0x4b, 0xef, 0x02, 0x1d, 0xc0, 0x8a, 0x2c, 0x3c, 0xd8, 0x9f, 0x1b, 0x53, 0x7f, 0xd8, 0x85, 0x1f, + 0x57, 0x48, 0x71, 0x1f, 0x1e, 0x39, 0x52, 0xf7, 0xe1, 0x1d, 0x98, 0xae, 0x92, 0xc0, 0x65, 0x79, + 0xb7, 0xeb, 0xf1, 0xf9, 0xb4, 0x0e, 0xc5, 0x20, 0x71, 0x22, 0xf7, 0x15, 0x69, 0x50, 0x0b, 0xbb, + 0x2f, 0x4f, 0xe0, 0x98, 0x90, 0xfd, 0xbf, 0x2c, 0x18, 0x16, 0x0e, 0x2f, 0x27, 0xc0, 0x35, 0x2e, + 0x18, 0x9a, 0x84, 0xb9, 0xf4, 0x01, 0x63, 0x9d, 0xc9, 0xd4, 0x21, 0x94, 0x13, 0x3a, 0x84, 0xc7, + 0xba, 0x11, 0xe9, 0xae, 0x3d, 0xf8, 0x6b, 0x79, 0xca, 0xbd, 0x1b, 0xae, 0x97, 0xc7, 0x3f, 0x04, + 0x6b, 0x30, 0x1c, 0x0a, 0xd7, 0xbf, 0x5c, 0xb6, 0xad, 0x7a, 0x72, 0x12, 0x63, 0x3b, 0x36, 0xe1, + 0xec, 0x27, 0x89, 0xa4, 0xfa, 0x14, 0xe6, 0x8f, 0xd1, 0xa7, 0xb0, 0x97, 0x73, 0xea, 0xc0, 0x51, + 0x38, 0xa7, 0xda, 0x5f, 0x61, 0x37, 0xa7, 0x5e, 0x7e, 0x02, 0x4c, 0xd5, 0x35, 0xf3, 0x8e, 0xb5, + 0xbb, 0xac, 0x2c, 0xd1, 0xa9, 0x0c, 0xe6, 0xea, 0xe7, 0x2d, 0x38, 0x9f, 0xf2, 0x55, 0x1a, 0xa7, + 0xf5, 0x14, 0x14, 0x9c, 0x76, 0xdd, 0x55, 0x7b, 0x59, 0xd3, 0x27, 0x2e, 0x88, 0x72, 0xac, 0x30, + 0xd0, 0x12, 0x4c, 0x91, 0xfb, 0x2d, 0x97, 0xab, 0x52, 0x75, 0x63, 0xd3, 0x3c, 0xf7, 0x92, 0x5a, + 0x4e, 0x02, 0x71, 0x27, 0xbe, 0x0a, 0x08, 0x92, 0xcf, 0x0c, 0x08, 0xf2, 0x77, 0x2d, 0x18, 0x51, + 0xce, 0x6f, 0xc7, 0x3e, 0xda, 0x9f, 0x34, 0x47, 0xfb, 0xd1, 0x2e, 0xa3, 0x9d, 0x31, 0xcc, 0xbf, + 0x9d, 0x53, 0xfd, 0xad, 0xf8, 0x41, 0xd4, 0x07, 0x07, 0xf7, 0x12, 0x14, 0x5a, 0x81, 0x1f, 0xf9, + 0x35, 0xbf, 0x21, 0x18, 0xb8, 0x73, 0x71, 0x64, 0x1c, 0x5e, 0x7e, 0xa0, 0xfd, 0xc6, 0x0a, 0x9b, + 0xf2, 0x4e, 0x4e, 0xab, 0x25, 0x01, 0xd2, 0x06, 0x8d, 0xc5, 0x8d, 0x8d, 0x8b, 0xb1, 0x8e, 0xc3, + 0x06, 0xdc, 0x0f, 0x22, 0xc1, 0x67, 0xc5, 0x03, 0xee, 0x07, 0x11, 0x66, 0x10, 0x54, 0x07, 0x88, + 0x9c, 0x60, 0x8b, 0x44, 0xb4, 0x4c, 0x04, 0xef, 0xca, 0x3e, 0x6f, 0xda, 0x91, 0xdb, 0x98, 0x77, + 0xbd, 0x28, 0x8c, 0x82, 0xf9, 0xb2, 0x17, 0xdd, 0x0a, 0xf8, 0x13, 0x52, 0x8b, 0x8e, 0xa3, 0x68, + 0x61, 0x8d, 0xae, 0x74, 0xf4, 0x66, 0x6d, 0x0c, 0x9a, 0xc6, 0x0c, 0x6b, 0xa2, 0x1c, 0x2b, 0x0c, + 0xfb, 0x45, 0x76, 0xfb, 0xb0, 0x31, 0x3d, 0x5c, 0x38, 0x99, 0x5f, 0x2e, 0xaa, 0xd9, 0x60, 0x9a, + 0xcc, 0x92, 0x1e, 0xb4, 0xa6, 0xfb, 0x61, 0x4f, 0x1b, 0xd6, 0xfd, 0xb5, 0xe2, 0xc8, 0x36, 0xe8, + 0xdb, 0x3a, 0x0c, 0x54, 0x9e, 0xee, 0x71, 0x6b, 0x1c, 0xc2, 0x24, 0x85, 0x25, 0x91, 0x60, 0x21, + 0xf6, 0xcb, 0x15, 0xb1, 0x2f, 0xb4, 0x24, 0x12, 0x02, 0x80, 0x63, 0x1c, 0x74, 0x55, 0x08, 0x08, + 0xb8, 0x9c, 0xff, 0xd1, 0x84, 0x80, 0x40, 0x7e, 0xbe, 0x26, 0xd5, 0x79, 0x06, 0x46, 0x54, 0x5e, + 0xd8, 0x0a, 0x4f, 0x37, 0x2a, 0x96, 0xcd, 0x72, 0x5c, 0x8c, 0x75, 0x1c, 0xb4, 0x0e, 0x13, 0x21, + 0x97, 0x9b, 0xa9, 0x88, 0xb5, 0x5c, 0xfe, 0xf8, 0x51, 0x69, 0xd5, 0x53, 0x35, 0xc1, 0x07, 0xac, + 0x88, 0x9f, 0x36, 0xd2, 0xb9, 0x3a, 0x49, 0x02, 0xbd, 0x06, 0xe3, 0x0d, 0xdf, 0xa9, 0x2f, 0x3a, + 0x0d, 0xc7, 0xab, 0xb1, 0xef, 0x2d, 0x98, 0xe9, 0x05, 0x6f, 0x1a, 0x50, 0x9c, 0xc0, 0xa6, 0xcc, + 0x98, 0x5e, 0x22, 0xa2, 0x2c, 0x3b, 0xde, 0x16, 0x09, 0x45, 0x96, 0x4f, 0xc6, 0x8c, 0xdd, 0xcc, + 0xc0, 0xc1, 0x99, 0xb5, 0xd1, 0x4b, 0x30, 0x2a, 0x3f, 0x5f, 0x8b, 0x45, 0x10, 0x3b, 0x32, 0x68, + 0x30, 0x6c, 0x60, 0xa2, 0x7b, 0x70, 0x5a, 0xfe, 0x5f, 0x0f, 0x9c, 0xcd, 0x4d, 0xb7, 0x26, 0x1c, + 0x74, 0xb9, 0x97, 0xe1, 0x82, 0x74, 0x85, 0x5b, 0x4e, 0x43, 0x3a, 0xd8, 0x9f, 0xbb, 0x28, 0x46, + 0x2d, 0x15, 0xce, 0x26, 0x31, 0x9d, 0x3e, 0x5a, 0x85, 0xe9, 0x6d, 0xe2, 0x34, 0xa2, 0xed, 0xa5, + 0x6d, 0x52, 0xdb, 0x91, 0x9b, 0x88, 0x45, 0x38, 0xd0, 0xcc, 0xff, 0xaf, 0x77, 0xa2, 0xe0, 0xb4, + 0x7a, 0xe8, 0x2d, 0x98, 0x69, 0xb5, 0x37, 0x1a, 0x6e, 0xb8, 0xbd, 0xe6, 0x47, 0xcc, 0xb4, 0x47, + 0xa5, 0x99, 0x15, 0xa1, 0x10, 0x54, 0x0c, 0x89, 0x4a, 0x06, 0x1e, 0xce, 0xa4, 0x80, 0xde, 0x85, + 0xd3, 0x89, 0xc5, 0x20, 0x9c, 0xc1, 0xc7, 0xb3, 0x63, 0xd6, 0x57, 0xd3, 0x2a, 0x88, 0xb8, 0x0a, + 0x69, 0x20, 0x9c, 0xde, 0x04, 0x7a, 0x1e, 0x0a, 0x6e, 0x6b, 0xc5, 0x69, 0xba, 0x8d, 0x3d, 0x16, + 0x74, 0xbf, 0xc8, 0x02, 0xd1, 0x17, 0xca, 0x15, 0x5e, 0x76, 0xa0, 0xfd, 0xc6, 0x0a, 0x93, 0x3e, + 0x41, 0xb4, 0xd0, 0xa2, 0xe1, 0xcc, 0x64, 0x6c, 0xb9, 0xac, 0xc5, 0x1f, 0x0d, 0xb1, 0x81, 0xf5, + 0xde, 0x0c, 0xc2, 0xde, 0xa1, 0x95, 0x35, 0x9e, 0x11, 0x7d, 0x0e, 0x46, 0xf5, 0x15, 0x2b, 0xee, + 0xbf, 0xcb, 0xe9, 0x2c, 0x95, 0xb6, 0xb2, 0x39, 0xc7, 0xa9, 0x56, 0xaf, 0x0e, 0xc3, 0x06, 0x45, + 0x9b, 0x40, 0xfa, 0x58, 0xa2, 0x9b, 0x50, 0xa8, 0x35, 0x5c, 0xe2, 0x45, 0xe5, 0x4a, 0xb7, 0xa8, + 0x58, 0x4b, 0x02, 0x47, 0x4c, 0x8e, 0x08, 0x28, 0xce, 0xcb, 0xb0, 0xa2, 0x60, 0xff, 0x5a, 0x0e, + 0xe6, 0x7a, 0x44, 0xa7, 0x4f, 0xe8, 0x2d, 0xac, 0xbe, 0xf4, 0x16, 0x0b, 0x32, 0x41, 0xef, 0x5a, + 0x42, 0x24, 0x92, 0x48, 0xbe, 0x1b, 0x0b, 0x46, 0x92, 0xf8, 0x7d, 0xdb, 0x91, 0xeb, 0xaa, 0x8f, + 0x81, 0x9e, 0x9e, 0x10, 0x86, 0xca, 0x73, 0xb0, 0xff, 0x77, 0x52, 0xa6, 0xfa, 0xca, 0xfe, 0x4a, + 0x0e, 0x4e, 0xab, 0x21, 0xfc, 0xe6, 0x1d, 0xb8, 0xdb, 0x9d, 0x03, 0x77, 0x04, 0xca, 0x3f, 0xfb, + 0x16, 0x0c, 0xf1, 0x30, 0x5f, 0x7d, 0xf0, 0x67, 0x97, 0xcc, 0x88, 0x98, 0x8a, 0x25, 0x30, 0xa2, + 0x62, 0x7e, 0xaf, 0x05, 0x13, 0xeb, 0x4b, 0x95, 0xaa, 0x5f, 0xdb, 0x21, 0xd1, 0x02, 0xe7, 0xa7, + 0xb1, 0xe0, 0xb5, 0xac, 0x87, 0xe4, 0xa1, 0xd2, 0xb8, 0xb3, 0x8b, 0x30, 0xb0, 0xed, 0x87, 0x51, + 0xd2, 0x32, 0xe0, 0xba, 0x1f, 0x46, 0x98, 0x41, 0xec, 0xdf, 0xb5, 0x60, 0x90, 0xa5, 0xa4, 0x97, + 0x52, 0x64, 0x2b, 0x43, 0x8a, 0xdc, 0xcf, 0x77, 0xa1, 0x17, 0x60, 0x88, 0x6c, 0x6e, 0x92, 0x5a, + 0x24, 0x66, 0x55, 0xba, 0x72, 0x0f, 0x2d, 0xb3, 0x52, 0xca, 0x60, 0xb0, 0xc6, 0xf8, 0x5f, 0x2c, + 0x90, 0xd1, 0x5d, 0x28, 0x46, 0x6e, 0x93, 0x2c, 0xd4, 0xeb, 0x42, 0xb7, 0xfa, 0x10, 0xee, 0xe8, + 0xeb, 0x92, 0x00, 0x8e, 0x69, 0xd9, 0x5f, 0xca, 0x01, 0xc4, 0xf1, 0x51, 0x7a, 0x7d, 0xe2, 0x62, + 0x87, 0xd6, 0xed, 0x72, 0x8a, 0xd6, 0x0d, 0xc5, 0x04, 0x53, 0x54, 0x6e, 0x6a, 0x98, 0xf2, 0x7d, + 0x0d, 0xd3, 0xc0, 0x61, 0x86, 0x69, 0x09, 0xa6, 0xe2, 0xf8, 0x2e, 0x66, 0x78, 0x2b, 0xf6, 0x86, + 0x5a, 0x4f, 0x02, 0x71, 0x27, 0xbe, 0x4d, 0xe0, 0xa2, 0x0a, 0x73, 0x21, 0xee, 0x1a, 0x66, 0xba, + 0xab, 0x6b, 0x31, 0x7b, 0x8c, 0x53, 0xac, 0x56, 0xcc, 0x65, 0xaa, 0x15, 0x7f, 0xc2, 0x82, 0x53, + 0xc9, 0x76, 0x98, 0x2f, 0xe5, 0x17, 0x2d, 0x38, 0xcd, 0x94, 0xab, 0xac, 0xd5, 0x4e, 0x55, 0xee, + 0xf3, 0x5d, 0x43, 0x77, 0x64, 0xf4, 0x38, 0x8e, 0x19, 0xb0, 0x9a, 0x46, 0x1a, 0xa7, 0xb7, 0x68, + 0xff, 0xfb, 0x1c, 0xcc, 0x64, 0xc5, 0xfc, 0x60, 0x96, 0xfd, 0xce, 0xfd, 0xea, 0x0e, 0xb9, 0x27, + 0xec, 0xa7, 0x63, 0xcb, 0x7e, 0x5e, 0x8c, 0x25, 0x3c, 0x19, 0x70, 0x3c, 0xd7, 0x5f, 0xc0, 0x71, + 0xb4, 0x0d, 0x53, 0xf7, 0xb6, 0x89, 0x77, 0xdb, 0x0b, 0x9d, 0xc8, 0x0d, 0x37, 0x5d, 0xa6, 0x88, + 0xe4, 0xeb, 0xe6, 0x65, 0x69, 0xe5, 0x7c, 0x37, 0x89, 0x70, 0xb0, 0x3f, 0x77, 0xde, 0x28, 0x88, + 0xbb, 0xcc, 0x0f, 0x12, 0xdc, 0x49, 0xb4, 0x33, 0x5e, 0xfb, 0xc0, 0x31, 0xc6, 0x6b, 0xb7, 0xbf, + 0x68, 0xc1, 0xd9, 0xcc, 0x24, 0x91, 0xe8, 0x0a, 0x14, 0x9c, 0x96, 0xcb, 0x65, 0xb9, 0xe2, 0x18, + 0x65, 0x32, 0x83, 0x4a, 0x99, 0x4b, 0x72, 0x15, 0x54, 0x25, 0xaf, 0xce, 0x65, 0x26, 0xaf, 0xee, + 0x99, 0x8b, 0xda, 0xfe, 0x1e, 0x0b, 0x84, 0x57, 0x62, 0x1f, 0x67, 0xf7, 0x9b, 0x32, 0xf7, 0xbf, + 0x91, 0xd3, 0xe5, 0x62, 0xb6, 0x9b, 0xa6, 0xc8, 0xe4, 0xa2, 0x78, 0x25, 0x23, 0x7f, 0x8b, 0x41, + 0xcb, 0xae, 0x83, 0x80, 0x96, 0x08, 0x93, 0x54, 0xf6, 0xee, 0xcd, 0xb3, 0x00, 0x75, 0x86, 0xab, + 0x65, 0x00, 0x57, 0x37, 0x73, 0x49, 0x41, 0xb0, 0x86, 0x65, 0xff, 0x9b, 0x1c, 0x8c, 0xc8, 0x1c, + 0x22, 0x6d, 0xaf, 0x1f, 0x79, 0xc2, 0xa1, 0x92, 0x0a, 0xb2, 0x94, 0xf9, 0x94, 0x70, 0x25, 0x16, + 0xc3, 0xc4, 0x29, 0xf3, 0x25, 0x00, 0xc7, 0x38, 0x74, 0x17, 0x85, 0xed, 0x0d, 0x86, 0x9e, 0xf0, + 0xa1, 0xab, 0xf2, 0x62, 0x2c, 0xe1, 0xe8, 0x53, 0x30, 0xc9, 0xeb, 0x05, 0x7e, 0xcb, 0xd9, 0xe2, + 0x42, 0xf2, 0x41, 0xe5, 0xfc, 0x3e, 0xb9, 0x9a, 0x80, 0x1d, 0xec, 0xcf, 0x9d, 0x4a, 0x96, 0x31, + 0xed, 0x4f, 0x07, 0x15, 0x66, 0x0b, 0xc3, 0x1b, 0xa1, 0xbb, 0xbf, 0xc3, 0x84, 0x26, 0x06, 0x61, + 0x1d, 0xcf, 0xfe, 0x1c, 0xa0, 0xce, 0x6c, 0x2a, 0xe8, 0x75, 0x6e, 0x00, 0xe9, 0x06, 0xa4, 0xde, + 0x4d, 0x1b, 0xa4, 0xbb, 0x78, 0x4b, 0xf7, 0x17, 0x5e, 0x0b, 0xab, 0xfa, 0xf6, 0x5f, 0xca, 0xc3, + 0x64, 0xd2, 0xe1, 0x17, 0x5d, 0x87, 0x21, 0xce, 0x7a, 0x08, 0xf2, 0x5d, 0x8c, 0x0d, 0x34, 0x37, + 0x61, 0x76, 0x08, 0x0b, 0xee, 0x45, 0xd4, 0x47, 0x6f, 0xc1, 0x48, 0xdd, 0xbf, 0xe7, 0xdd, 0x73, + 0x82, 0xfa, 0x42, 0xa5, 0x2c, 0x96, 0x73, 0xea, 0x6b, 0xa9, 0x14, 0xa3, 0xe9, 0xae, 0xc7, 0x4c, + 0xb1, 0x16, 0x83, 0xb0, 0x4e, 0x0e, 0xad, 0xb3, 0xe0, 0xcf, 0x9b, 0xee, 0xd6, 0xaa, 0xd3, 0xea, + 0x66, 0x0d, 0xbf, 0x24, 0x91, 0x34, 0xca, 0x63, 0x22, 0x42, 0x34, 0x07, 0xe0, 0x98, 0x10, 0xfa, + 0x3c, 0x4c, 0x87, 0x19, 0x32, 0xd9, 0xac, 0xe4, 0x5a, 0xdd, 0xc4, 0x94, 0x8b, 0x8f, 0xd0, 0x77, + 0x6c, 0x9a, 0xf4, 0x36, 0xad, 0x19, 0xfb, 0x47, 0x4e, 0x81, 0xb1, 0x89, 0x8d, 0x5c, 0x8b, 0xd6, + 0x11, 0xe5, 0x5a, 0xc4, 0x50, 0x20, 0xcd, 0x56, 0xb4, 0x57, 0x72, 0x83, 0x6e, 0x19, 0x88, 0x97, + 0x05, 0x4e, 0x27, 0x4d, 0x09, 0xc1, 0x8a, 0x4e, 0x7a, 0x42, 0xcc, 0xfc, 0xd7, 0x31, 0x21, 0xe6, + 0xc0, 0x09, 0x26, 0xc4, 0x5c, 0x83, 0xe1, 0x2d, 0x37, 0xc2, 0xa4, 0xe5, 0x0b, 0xa6, 0x3f, 0x75, + 0x1d, 0x5e, 0xe3, 0x28, 0x9d, 0xa9, 0xd7, 0x04, 0x00, 0x4b, 0x22, 0xe8, 0x75, 0xb5, 0x03, 0x87, + 0xb2, 0xdf, 0xcc, 0x9d, 0x5a, 0xf1, 0xd4, 0x3d, 0x28, 0xd2, 0x5e, 0x0e, 0x3f, 0x6c, 0xda, 0xcb, + 0x15, 0x99, 0xac, 0xb2, 0x90, 0xed, 0xba, 0xc2, 0x72, 0x51, 0xf6, 0x48, 0x51, 0x79, 0x47, 0x4f, + 0xf0, 0x59, 0xcc, 0x3e, 0x09, 0x54, 0xee, 0xce, 0x3e, 0xd3, 0x7a, 0x7e, 0x8f, 0x05, 0xa7, 0x5b, + 0x69, 0xb9, 0x6e, 0x85, 0x02, 0xf9, 0x85, 0xbe, 0xd3, 0xe9, 0x1a, 0x0d, 0x32, 0x41, 0x4d, 0x2a, + 0x1a, 0x4e, 0x6f, 0x8e, 0x0e, 0x74, 0xb0, 0x51, 0x17, 0x8a, 0xcc, 0x4b, 0x19, 0xf9, 0x41, 0xbb, + 0x64, 0x05, 0x5d, 0x4f, 0xc9, 0x45, 0xf9, 0xe1, 0xac, 0x5c, 0x94, 0x7d, 0x67, 0xa0, 0x7c, 0x5d, + 0x65, 0x06, 0x1d, 0xcb, 0x5e, 0x4a, 0x3c, 0xef, 0x67, 0xcf, 0x7c, 0xa0, 0xaf, 0xab, 0x7c, 0xa0, + 0x5d, 0x22, 0x7b, 0xf2, 0x6c, 0x9f, 0x3d, 0xb3, 0x80, 0x6a, 0x99, 0x3c, 0x27, 0x8e, 0x26, 0x93, + 0xa7, 0x71, 0xd5, 0xf0, 0x64, 0x92, 0x4f, 0xf6, 0xb8, 0x6a, 0x0c, 0xba, 0xdd, 0x2f, 0x1b, 0x9e, + 0xb5, 0x74, 0xea, 0xa1, 0xb2, 0x96, 0xde, 0xd1, 0xb3, 0x80, 0xa2, 0x1e, 0x69, 0x2e, 0x29, 0x52, + 0x9f, 0xb9, 0x3f, 0xef, 0xe8, 0x17, 0xe0, 0x74, 0x36, 0x5d, 0x75, 0xcf, 0x75, 0xd2, 0x4d, 0xbd, + 0x02, 0x3b, 0x72, 0x8a, 0x9e, 0x3a, 0x99, 0x9c, 0xa2, 0xa7, 0x8f, 0x3c, 0xa7, 0xe8, 0x99, 0x13, + 0xc8, 0x29, 0xfa, 0xc8, 0x09, 0xe6, 0x14, 0xbd, 0xc3, 0xac, 0x2e, 0x78, 0x6c, 0x17, 0x11, 0x89, + 0x34, 0x3d, 0xea, 0x65, 0x5a, 0x00, 0x18, 0xfe, 0x71, 0x0a, 0x84, 0x63, 0x52, 0x29, 0xb9, 0x4a, + 0x67, 0x8e, 0x21, 0x57, 0xe9, 0x5a, 0x9c, 0xab, 0xf4, 0x6c, 0xf6, 0x54, 0xa7, 0xd8, 0xe9, 0x67, + 0x64, 0x28, 0xbd, 0xa3, 0x67, 0x16, 0x7d, 0xb4, 0x8b, 0x28, 0x3e, 0x4d, 0xf0, 0xd8, 0x25, 0x9f, + 0xe8, 0x6b, 0x3c, 0x9f, 0xe8, 0xb9, 0xec, 0x93, 0x3c, 0x79, 0xdd, 0x19, 0x59, 0x44, 0x69, 0xbf, + 0x54, 0xcc, 0x3b, 0x16, 0x73, 0x35, 0xa3, 0x5f, 0x2a, 0x68, 0x5e, 0x67, 0xbf, 0x14, 0x08, 0xc7, + 0xa4, 0xec, 0xef, 0xcb, 0xc1, 0x85, 0xee, 0xfb, 0x2d, 0x96, 0xa6, 0x56, 0x62, 0x4d, 0x63, 0x42, + 0x9a, 0xca, 0xdf, 0x6c, 0x31, 0x56, 0xdf, 0xe1, 0xc4, 0xae, 0xc1, 0x94, 0x32, 0xf0, 0x6f, 0xb8, + 0xb5, 0xbd, 0xb5, 0xf8, 0xe5, 0xab, 0x9c, 0xa2, 0xab, 0x49, 0x04, 0xdc, 0x59, 0x07, 0x2d, 0xc0, + 0x84, 0x51, 0x58, 0x2e, 0x89, 0xb7, 0x99, 0x12, 0xdf, 0x56, 0x4d, 0x30, 0x4e, 0xe2, 0xdb, 0x5f, + 0xb6, 0xe0, 0x91, 0x8c, 0x34, 0x60, 0x7d, 0x47, 0xcb, 0xda, 0x84, 0x89, 0x96, 0x59, 0xb5, 0x47, + 0x50, 0x3d, 0x23, 0xd9, 0x98, 0xea, 0x6b, 0x02, 0x80, 0x93, 0x44, 0xed, 0x3f, 0xb3, 0xe0, 0x7c, + 0x57, 0x8b, 0x35, 0x84, 0xe1, 0xcc, 0x56, 0x33, 0x74, 0x96, 0x02, 0x52, 0x27, 0x5e, 0xe4, 0x3a, + 0x8d, 0x6a, 0x8b, 0xd4, 0x34, 0x79, 0x38, 0x33, 0xfd, 0xba, 0xb6, 0x5a, 0x5d, 0xe8, 0xc4, 0xc0, + 0x19, 0x35, 0xd1, 0x0a, 0xa0, 0x4e, 0x88, 0x98, 0x61, 0x16, 0xbd, 0xb7, 0x93, 0x1e, 0x4e, 0xa9, + 0x81, 0x5e, 0x84, 0x31, 0x65, 0x09, 0xa7, 0xcd, 0x38, 0x3b, 0xd8, 0xb1, 0x0e, 0xc0, 0x26, 0xde, + 0xe2, 0x95, 0xdf, 0xf8, 0xfd, 0x0b, 0x1f, 0xfa, 0xad, 0xdf, 0xbf, 0xf0, 0xa1, 0xdf, 0xf9, 0xfd, + 0x0b, 0x1f, 0xfa, 0x8e, 0x07, 0x17, 0xac, 0xdf, 0x78, 0x70, 0xc1, 0xfa, 0xad, 0x07, 0x17, 0xac, + 0xdf, 0x79, 0x70, 0xc1, 0xfa, 0xbd, 0x07, 0x17, 0xac, 0x2f, 0xfd, 0xc1, 0x85, 0x0f, 0xbd, 0x99, + 0xdb, 0x7d, 0xe6, 0xff, 0x06, 0x00, 0x00, 0xff, 0xff, 0xaf, 0x9f, 0xac, 0x23, 0x24, 0x01, 0x01, + 0x00, } func (m *AWSElasticBlockStoreVolumeSource) Marshal() (dAtA []byte, err error) { @@ -7889,6 +7993,16 @@ func (m *ConfigMap) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.Immutable != nil { + i-- + if *m.Immutable { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x20 + } if len(m.BinaryData) > 0 { keysForBinaryData := make([]string, 0, len(m.BinaryData)) for k := range m.BinaryData { @@ -9132,6 +9246,13 @@ func (m *EndpointPort) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.AppProtocol != nil { + i -= len(*m.AppProtocol) + copy(dAtA[i:], *m.AppProtocol) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.AppProtocol))) + i-- + dAtA[i] = 0x22 + } i -= len(m.Protocol) copy(dAtA[i:], m.Protocol) i = encodeVarintGenerated(dAtA, i, uint64(len(m.Protocol))) @@ -9809,6 +9930,49 @@ func (m *EphemeralContainers) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } +func (m *EphemeralVolumeSource) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *EphemeralVolumeSource) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *EphemeralVolumeSource) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i-- + if m.ReadOnly { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x10 + if m.VolumeClaimTemplate != nil { + { + size, err := m.VolumeClaimTemplate.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil +} + func (m *Event) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -10016,11 +10180,6 @@ func (m *EventSeries) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l - i -= len(m.State) - copy(dAtA[i:], m.State) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.State))) - i-- - dAtA[i] = 0x1a { size, err := m.LastObservedTime.MarshalToSizedBuffer(dAtA[:i]) if err != nil { @@ -13170,6 +13329,49 @@ func (m *PersistentVolumeClaimStatus) MarshalToSizedBuffer(dAtA []byte) (int, er return len(dAtA) - i, nil } +func (m *PersistentVolumeClaimTemplate) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *PersistentVolumeClaimTemplate) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *PersistentVolumeClaimTemplate) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + { + size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + func (m *PersistentVolumeClaimVolumeSource) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -14500,6 +14702,25 @@ func (m *PodSecurityContext) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.SeccompProfile != nil { + { + size, err := m.SeccompProfile.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x52 + } + if m.FSGroupChangePolicy != nil { + i -= len(*m.FSGroupChangePolicy) + copy(dAtA[i:], *m.FSGroupChangePolicy) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.FSGroupChangePolicy))) + i-- + dAtA[i] = 0x4a + } if m.WindowsOptions != nil { { size, err := m.WindowsOptions.MarshalToSizedBuffer(dAtA[:i]) @@ -14628,6 +14849,18 @@ func (m *PodSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.SetHostnameAsFQDN != nil { + i-- + if *m.SetHostnameAsFQDN { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x2 + i-- + dAtA[i] = 0x98 + } if len(m.EphemeralContainers) > 0 { for iNdEx := len(m.EphemeralContainers) - 1; iNdEx >= 0; iNdEx-- { { @@ -16777,6 +17010,41 @@ func (m *ScopedResourceSelectorRequirement) MarshalToSizedBuffer(dAtA []byte) (i return len(dAtA) - i, nil } +func (m *SeccompProfile) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *SeccompProfile) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *SeccompProfile) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.LocalhostProfile != nil { + i -= len(*m.LocalhostProfile) + copy(dAtA[i:], *m.LocalhostProfile) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.LocalhostProfile))) + i-- + dAtA[i] = 0x12 + } + i -= len(m.Type) + copy(dAtA[i:], m.Type) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Type))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + func (m *Secret) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -16797,6 +17065,16 @@ func (m *Secret) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.Immutable != nil { + i-- + if *m.Immutable { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x28 + } if len(m.StringData) > 0 { keysForStringData := make([]string, 0, len(m.StringData)) for k := range m.StringData { @@ -17170,6 +17448,18 @@ func (m *SecurityContext) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.SeccompProfile != nil { + { + size, err := m.SeccompProfile.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x5a + } if m.WindowsOptions != nil { { size, err := m.WindowsOptions.MarshalToSizedBuffer(dAtA[:i]) @@ -17575,6 +17865,13 @@ func (m *ServicePort) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.AppProtocol != nil { + i -= len(*m.AppProtocol) + copy(dAtA[i:], *m.AppProtocol) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.AppProtocol))) + i-- + dAtA[i] = 0x32 + } i = encodeVarintGenerated(dAtA, i, uint64(m.NodePort)) i-- dAtA[i] = 0x28 @@ -18552,6 +18849,20 @@ func (m *VolumeSource) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.Ephemeral != nil { + { + size, err := m.Ephemeral.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1 + i-- + dAtA[i] = 0xea + } if m.CSI != nil { { size, err := m.CSI.MarshalToSizedBuffer(dAtA[:i]) @@ -19458,6 +19769,9 @@ func (m *ConfigMap) Size() (n int) { n += mapEntrySize + 1 + sovGenerated(uint64(mapEntrySize)) } } + if m.Immutable != nil { + n += 2 + } return n } @@ -19895,6 +20209,10 @@ func (m *EndpointPort) Size() (n int) { n += 1 + sovGenerated(uint64(m.Port)) l = len(m.Protocol) n += 1 + l + sovGenerated(uint64(l)) + if m.AppProtocol != nil { + l = len(*m.AppProtocol) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -20138,6 +20456,20 @@ func (m *EphemeralContainers) Size() (n int) { return n } +func (m *EphemeralVolumeSource) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.VolumeClaimTemplate != nil { + l = m.VolumeClaimTemplate.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + n += 2 + return n +} + func (m *Event) Size() (n int) { if m == nil { return 0 @@ -20206,8 +20538,6 @@ func (m *EventSeries) Size() (n int) { n += 1 + sovGenerated(uint64(m.Count)) l = m.LastObservedTime.Size() n += 1 + l + sovGenerated(uint64(l)) - l = len(m.State) - n += 1 + l + sovGenerated(uint64(l)) return n } @@ -21369,6 +21699,19 @@ func (m *PersistentVolumeClaimStatus) Size() (n int) { return n } +func (m *PersistentVolumeClaimTemplate) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Spec.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + func (m *PersistentVolumeClaimVolumeSource) Size() (n int) { if m == nil { return 0 @@ -21877,6 +22220,14 @@ func (m *PodSecurityContext) Size() (n int) { l = m.WindowsOptions.Size() n += 1 + l + sovGenerated(uint64(l)) } + if m.FSGroupChangePolicy != nil { + l = len(*m.FSGroupChangePolicy) + n += 1 + l + sovGenerated(uint64(l)) + } + if m.SeccompProfile != nil { + l = m.SeccompProfile.Size() + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -22029,6 +22380,9 @@ func (m *PodSpec) Size() (n int) { n += 2 + l + sovGenerated(uint64(l)) } } + if m.SetHostnameAsFQDN != nil { + n += 3 + } return n } @@ -22666,6 +23020,21 @@ func (m *ScopedResourceSelectorRequirement) Size() (n int) { return n } +func (m *SeccompProfile) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Type) + n += 1 + l + sovGenerated(uint64(l)) + if m.LocalhostProfile != nil { + l = len(*m.LocalhostProfile) + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + func (m *Secret) Size() (n int) { if m == nil { return 0 @@ -22696,6 +23065,9 @@ func (m *Secret) Size() (n int) { n += mapEntrySize + 1 + sovGenerated(uint64(mapEntrySize)) } } + if m.Immutable != nil { + n += 2 + } return n } @@ -22842,6 +23214,10 @@ func (m *SecurityContext) Size() (n int) { l = m.WindowsOptions.Size() n += 1 + l + sovGenerated(uint64(l)) } + if m.SeccompProfile != nil { + l = m.SeccompProfile.Size() + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -22961,6 +23337,10 @@ func (m *ServicePort) Size() (n int) { l = m.TargetPort.Size() n += 1 + l + sovGenerated(uint64(l)) n += 1 + sovGenerated(uint64(m.NodePort)) + if m.AppProtocol != nil { + l = len(*m.AppProtocol) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -23438,6 +23818,10 @@ func (m *VolumeSource) Size() (n int) { l = m.CSI.Size() n += 2 + l + sovGenerated(uint64(l)) } + if m.Ephemeral != nil { + l = m.Ephemeral.Size() + n += 2 + l + sovGenerated(uint64(l)) + } return n } @@ -23801,6 +24185,7 @@ func (this *ConfigMap) String() string { `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, `Data:` + mapStringForData + `,`, `BinaryData:` + mapStringForBinaryData + `,`, + `Immutable:` + valueToStringGenerated(this.Immutable) + `,`, `}`, }, "") return s @@ -24127,6 +24512,7 @@ func (this *EndpointPort) String() string { `Name:` + fmt.Sprintf("%v", this.Name) + `,`, `Port:` + fmt.Sprintf("%v", this.Port) + `,`, `Protocol:` + fmt.Sprintf("%v", this.Protocol) + `,`, + `AppProtocol:` + valueToStringGenerated(this.AppProtocol) + `,`, `}`, }, "") return s @@ -24310,6 +24696,17 @@ func (this *EphemeralContainers) String() string { }, "") return s } +func (this *EphemeralVolumeSource) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&EphemeralVolumeSource{`, + `VolumeClaimTemplate:` + strings.Replace(this.VolumeClaimTemplate.String(), "PersistentVolumeClaimTemplate", "PersistentVolumeClaimTemplate", 1) + `,`, + `ReadOnly:` + fmt.Sprintf("%v", this.ReadOnly) + `,`, + `}`, + }, "") + return s +} func (this *Event) String() string { if this == nil { return "nil" @@ -24357,7 +24754,6 @@ func (this *EventSeries) String() string { s := strings.Join([]string{`&EventSeries{`, `Count:` + fmt.Sprintf("%v", this.Count) + `,`, `LastObservedTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastObservedTime), "MicroTime", "v1.MicroTime", 1), `&`, ``, 1) + `,`, - `State:` + fmt.Sprintf("%v", this.State) + `,`, `}`, }, "") return s @@ -25290,6 +25686,17 @@ func (this *PersistentVolumeClaimStatus) String() string { }, "") return s } +func (this *PersistentVolumeClaimTemplate) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&PersistentVolumeClaimTemplate{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "PersistentVolumeClaimSpec", "PersistentVolumeClaimSpec", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} func (this *PersistentVolumeClaimVolumeSource) String() string { if this == nil { return "nil" @@ -25629,6 +26036,8 @@ func (this *PodSecurityContext) String() string { `RunAsGroup:` + valueToStringGenerated(this.RunAsGroup) + `,`, `Sysctls:` + repeatedStringForSysctls + `,`, `WindowsOptions:` + strings.Replace(this.WindowsOptions.String(), "WindowsSecurityContextOptions", "WindowsSecurityContextOptions", 1) + `,`, + `FSGroupChangePolicy:` + valueToStringGenerated(this.FSGroupChangePolicy) + `,`, + `SeccompProfile:` + strings.Replace(this.SeccompProfile.String(), "SeccompProfile", "SeccompProfile", 1) + `,`, `}`, }, "") return s @@ -25747,6 +26156,7 @@ func (this *PodSpec) String() string { `Overhead:` + mapStringForOverhead + `,`, `TopologySpreadConstraints:` + repeatedStringForTopologySpreadConstraints + `,`, `EphemeralContainers:` + repeatedStringForEphemeralContainers + `,`, + `SetHostnameAsFQDN:` + valueToStringGenerated(this.SetHostnameAsFQDN) + `,`, `}`, }, "") return s @@ -26272,6 +26682,17 @@ func (this *ScopedResourceSelectorRequirement) String() string { }, "") return s } +func (this *SeccompProfile) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&SeccompProfile{`, + `Type:` + fmt.Sprintf("%v", this.Type) + `,`, + `LocalhostProfile:` + valueToStringGenerated(this.LocalhostProfile) + `,`, + `}`, + }, "") + return s +} func (this *Secret) String() string { if this == nil { return "nil" @@ -26301,6 +26722,7 @@ func (this *Secret) String() string { `Data:` + mapStringForData + `,`, `Type:` + fmt.Sprintf("%v", this.Type) + `,`, `StringData:` + mapStringForStringData + `,`, + `Immutable:` + valueToStringGenerated(this.Immutable) + `,`, `}`, }, "") return s @@ -26405,6 +26827,7 @@ func (this *SecurityContext) String() string { `RunAsGroup:` + valueToStringGenerated(this.RunAsGroup) + `,`, `ProcMount:` + valueToStringGenerated(this.ProcMount) + `,`, `WindowsOptions:` + strings.Replace(this.WindowsOptions.String(), "WindowsSecurityContextOptions", "WindowsSecurityContextOptions", 1) + `,`, + `SeccompProfile:` + strings.Replace(this.SeccompProfile.String(), "SeccompProfile", "SeccompProfile", 1) + `,`, `}`, }, "") return s @@ -26508,6 +26931,7 @@ func (this *ServicePort) String() string { `Port:` + fmt.Sprintf("%v", this.Port) + `,`, `TargetPort:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.TargetPort), "IntOrString", "intstr.IntOrString", 1), `&`, ``, 1) + `,`, `NodePort:` + fmt.Sprintf("%v", this.NodePort) + `,`, + `AppProtocol:` + valueToStringGenerated(this.AppProtocol) + `,`, `}`, }, "") return s @@ -26802,6 +27226,7 @@ func (this *VolumeSource) String() string { `Projected:` + strings.Replace(this.Projected.String(), "ProjectedVolumeSource", "ProjectedVolumeSource", 1) + `,`, `StorageOS:` + strings.Replace(this.StorageOS.String(), "StorageOSVolumeSource", "StorageOSVolumeSource", 1) + `,`, `CSI:` + strings.Replace(this.CSI.String(), "CSIVolumeSource", "CSIVolumeSource", 1) + `,`, + `Ephemeral:` + strings.Replace(this.Ephemeral.String(), "EphemeralVolumeSource", "EphemeralVolumeSource", 1) + `,`, `}`, }, "") return s @@ -30524,6 +30949,27 @@ func (m *ConfigMap) Unmarshal(dAtA []byte) error { } m.BinaryData[mapkey] = mapvalue iNdEx = postIndex + case 4: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field Immutable", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + b := bool(v != 0) + m.Immutable = &b default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -34258,6 +34704,39 @@ func (m *EndpointPort) Unmarshal(dAtA []byte) error { } m.Protocol = Protocol(dAtA[iNdEx:postIndex]) iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field AppProtocol", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.AppProtocol = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -36174,7 +36653,7 @@ func (m *EphemeralContainers) Unmarshal(dAtA []byte) error { } return nil } -func (m *Event) Unmarshal(dAtA []byte) error { +func (m *EphemeralVolumeSource) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -36197,15 +36676,15 @@ func (m *Event) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: Event: wiretype end group for non-group") + return fmt.Errorf("proto: EphemeralVolumeSource: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: Event: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: EphemeralVolumeSource: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field VolumeClaimTemplate", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -36232,15 +36711,18 @@ func (m *Event) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if m.VolumeClaimTemplate == nil { + m.VolumeClaimTemplate = &PersistentVolumeClaimTemplate{} + } + if err := m.VolumeClaimTemplate.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field InvolvedObject", wireType) + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field ReadOnly", wireType) } - var msglen int + var v int for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -36250,62 +36732,70 @@ func (m *Event) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - msglen |= int(b&0x7F) << shift + v |= int(b&0x7F) << shift if b < 0x80 { break } } - if msglen < 0 { + m.ReadOnly = bool(v != 0) + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { return ErrInvalidLengthGenerated } - postIndex := iNdEx + msglen - if postIndex < 0 { + if (iNdEx + skippy) < 0 { return ErrInvalidLengthGenerated } - if postIndex > l { + if (iNdEx + skippy) > l { return io.ErrUnexpectedEOF } - if err := m.InvolvedObject.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Reason", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *Event) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated } - if postIndex > l { + if iNdEx >= l { return io.ErrUnexpectedEOF } - m.Reason = string(dAtA[iNdEx:postIndex]) - iNdEx = postIndex - case 4: + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: Event: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: Event: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Message", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) } - var stringLen uint64 + var msglen int for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -36315,27 +36805,125 @@ func (m *Event) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - stringLen |= uint64(b&0x7F) << shift + msglen |= int(b&0x7F) << shift if b < 0x80 { break } } - intStringLen := int(stringLen) - if intStringLen < 0 { + if msglen < 0 { return ErrInvalidLengthGenerated } - postIndex := iNdEx + intStringLen + postIndex := iNdEx + msglen if postIndex < 0 { return ErrInvalidLengthGenerated } if postIndex > l { return io.ErrUnexpectedEOF } - m.Message = string(dAtA[iNdEx:postIndex]) + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } iNdEx = postIndex - case 5: + case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Source", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field InvolvedObject", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.InvolvedObject.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Reason", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Reason = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Message", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Message = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Source", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -36909,38 +37497,6 @@ func (m *EventSeries) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field State", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.State = EventSeriesState(dAtA[iNdEx:postIndex]) - iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -47449,7 +48005,7 @@ func (m *PersistentVolumeClaimStatus) Unmarshal(dAtA []byte) error { } return nil } -func (m *PersistentVolumeClaimVolumeSource) Unmarshal(dAtA []byte) error { +func (m *PersistentVolumeClaimTemplate) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -47472,17 +48028,17 @@ func (m *PersistentVolumeClaimVolumeSource) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: PersistentVolumeClaimVolumeSource: wiretype end group for non-group") + return fmt.Errorf("proto: PersistentVolumeClaimTemplate: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: PersistentVolumeClaimVolumeSource: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: PersistentVolumeClaimTemplate: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ClaimName", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) } - var stringLen uint64 + var msglen int for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -47492,29 +48048,30 @@ func (m *PersistentVolumeClaimVolumeSource) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - stringLen |= uint64(b&0x7F) << shift + msglen |= int(b&0x7F) << shift if b < 0x80 { break } } - intStringLen := int(stringLen) - if intStringLen < 0 { + if msglen < 0 { return ErrInvalidLengthGenerated } - postIndex := iNdEx + intStringLen + postIndex := iNdEx + msglen if postIndex < 0 { return ErrInvalidLengthGenerated } if postIndex > l { return io.ErrUnexpectedEOF } - m.ClaimName = string(dAtA[iNdEx:postIndex]) + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } iNdEx = postIndex case 2: - if wireType != 0 { - return fmt.Errorf("proto: wrong wireType = %d for field ReadOnly", wireType) + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) } - var v int + var msglen int for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -47524,12 +48081,25 @@ func (m *PersistentVolumeClaimVolumeSource) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - v |= int(b&0x7F) << shift + msglen |= int(b&0x7F) << shift if b < 0x80 { break } } - m.ReadOnly = bool(v != 0) + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -47554,7 +48124,7 @@ func (m *PersistentVolumeClaimVolumeSource) Unmarshal(dAtA []byte) error { } return nil } -func (m *PersistentVolumeList) Unmarshal(dAtA []byte) error { +func (m *PersistentVolumeClaimVolumeSource) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -47577,17 +48147,17 @@ func (m *PersistentVolumeList) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: PersistentVolumeList: wiretype end group for non-group") + return fmt.Errorf("proto: PersistentVolumeClaimVolumeSource: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: PersistentVolumeList: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: PersistentVolumeClaimVolumeSource: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ClaimName", wireType) } - var msglen int + var stringLen uint64 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -47597,30 +48167,29 @@ func (m *PersistentVolumeList) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - msglen |= int(b&0x7F) << shift + stringLen |= uint64(b&0x7F) << shift if b < 0x80 { break } } - if msglen < 0 { + intStringLen := int(stringLen) + if intStringLen < 0 { return ErrInvalidLengthGenerated } - postIndex := iNdEx + msglen + postIndex := iNdEx + intStringLen if postIndex < 0 { return ErrInvalidLengthGenerated } if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } + m.ClaimName = string(dAtA[iNdEx:postIndex]) iNdEx = postIndex case 2: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field ReadOnly", wireType) } - var msglen int + var v int for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -47630,26 +48199,12 @@ func (m *PersistentVolumeList) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - msglen |= int(b&0x7F) << shift + v |= int(b&0x7F) << shift if b < 0x80 { break } } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.Items = append(m.Items, PersistentVolume{}) - if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex + m.ReadOnly = bool(v != 0) default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -47674,7 +48229,7 @@ func (m *PersistentVolumeList) Unmarshal(dAtA []byte) error { } return nil } -func (m *PersistentVolumeSource) Unmarshal(dAtA []byte) error { +func (m *PersistentVolumeList) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -47697,15 +48252,15 @@ func (m *PersistentVolumeSource) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: PersistentVolumeSource: wiretype end group for non-group") + return fmt.Errorf("proto: PersistentVolumeList: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: PersistentVolumeSource: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: PersistentVolumeList: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field GCEPersistentDisk", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -47732,52 +48287,172 @@ func (m *PersistentVolumeSource) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if m.GCEPersistentDisk == nil { - m.GCEPersistentDisk = &GCEPersistentDiskVolumeSource{} - } - if err := m.GCEPersistentDisk.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field AWSElasticBlockStore", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if m.AWSElasticBlockStore == nil { - m.AWSElasticBlockStore = &AWSElasticBlockStoreVolumeSource{} - } - if err := m.AWSElasticBlockStore.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field HostPath", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Items = append(m.Items, PersistentVolume{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *PersistentVolumeSource) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: PersistentVolumeSource: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: PersistentVolumeSource: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field GCEPersistentDisk", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.GCEPersistentDisk == nil { + m.GCEPersistentDisk = &GCEPersistentDiskVolumeSource{} + } + if err := m.GCEPersistentDisk.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field AWSElasticBlockStore", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.AWSElasticBlockStore == nil { + m.AWSElasticBlockStore = &AWSElasticBlockStoreVolumeSource{} + } + if err := m.AWSElasticBlockStore.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field HostPath", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -51715,6 +52390,75 @@ func (m *PodSecurityContext) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 9: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field FSGroupChangePolicy", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := PodFSGroupChangePolicy(dAtA[iNdEx:postIndex]) + m.FSGroupChangePolicy = &s + iNdEx = postIndex + case 10: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field SeccompProfile", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.SeccompProfile == nil { + m.SeccompProfile = &SeccompProfile{} + } + if err := m.SeccompProfile.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -53064,6 +53808,27 @@ func (m *PodSpec) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 35: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field SetHostnameAsFQDN", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + b := bool(v != 0) + m.SetHostnameAsFQDN = &b default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -59174,6 +59939,124 @@ func (m *ScopedResourceSelectorRequirement) Unmarshal(dAtA []byte) error { } return nil } +func (m *SeccompProfile) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: SeccompProfile: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: SeccompProfile: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Type", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Type = SeccompProfileType(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field LocalhostProfile", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.LocalhostProfile = &s + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} func (m *Secret) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 @@ -59523,6 +60406,27 @@ func (m *Secret) Unmarshal(dAtA []byte) error { } m.StringData[mapkey] = mapvalue iNdEx = postIndex + case 5: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field Immutable", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + b := bool(v != 0) + m.Immutable = &b default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -60625,6 +61529,42 @@ func (m *SecurityContext) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 11: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field SeccompProfile", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.SeccompProfile == nil { + m.SeccompProfile = &SeccompProfile{} + } + if err := m.SeccompProfile.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -61603,6 +62543,39 @@ func (m *ServicePort) Unmarshal(dAtA []byte) error { break } } + case 6: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field AppProtocol", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.AppProtocol = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -65846,6 +66819,42 @@ func (m *VolumeSource) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 29: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Ephemeral", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Ephemeral == nil { + m.Ephemeral = &EphemeralVolumeSource{} + } + if err := m.Ephemeral.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -66311,6 +67320,7 @@ func (m *WindowsSecurityContextOptions) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -66342,10 +67352,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -66366,55 +67374,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/core/v1/generated.proto b/vendor/k8s.io/api/core/v1/generated.proto index c05e2351..916e2601 100644 --- a/vendor/k8s.io/api/core/v1/generated.proto +++ b/vendor/k8s.io/api/core/v1/generated.proto @@ -424,6 +424,7 @@ message ComponentCondition { } // ComponentStatus (and ComponentStatusList) holds the cluster validation info. +// Deprecated: This API is deprecated in v1.19+ message ComponentStatus { // Standard object's metadata. // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata @@ -438,6 +439,7 @@ message ComponentStatus { } // Status of all the conditions for the component as a list of ComponentStatus objects. +// Deprecated: This API is deprecated in v1.19+ message ComponentStatusList { // Standard list metadata. // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds @@ -455,6 +457,14 @@ message ConfigMap { // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + // Immutable, if set to true, ensures that data stored in the ConfigMap cannot + // be updated (only object metadata can be modified). + // If not set to true, the field can be modified at any time. + // Defaulted to nil. + // This is a beta field enabled by ImmutableEphemeralVolumes feature gate. + // +optional + optional bool immutable = 4; + // Data contains the configuration data. // Each key must consist of alphanumeric characters, '-', '_' or '.'. // Values with non-UTF-8 byte sequences must use the BinaryData field. @@ -580,8 +590,10 @@ message ConfigMapVolumeSource { // +optional repeated KeyToPath items = 2; - // Optional: mode bits to use on created files by default. Must be a - // value between 0 and 0777. Defaults to 0644. + // Optional: mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // Defaults to 0644. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -681,7 +693,6 @@ message Container { repeated VolumeMount volumeMounts = 9; // volumeDevices is the list of block devices to be used by the container. - // This is a beta feature. // +patchMergeKey=devicePath // +patchStrategy=merge // +optional @@ -707,7 +718,7 @@ message Container { // This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, // when it might take a long time to load data or warm a cache, than during steady-state operation. // This cannot be updated. - // This is an alpha feature enabled by the StartupProbe feature flag. + // This is a beta feature enabled by the StartupProbe feature flag. // More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes // +optional optional Probe startupProbe = 22; @@ -950,8 +961,10 @@ message DownwardAPIVolumeFile { // +optional optional ResourceFieldSelector resourceFieldRef = 3; - // Optional: mode bits to use on this file, must be a value between 0 - // and 0777. If not specified, the volume defaultMode will be used. + // Optional: mode bits used to set permissions on this file, must be an octal value + // between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // If not specified, the volume defaultMode will be used. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. // +optional @@ -966,7 +979,10 @@ message DownwardAPIVolumeSource { repeated DownwardAPIVolumeFile items = 1; // Optional: mode bits to use on created files by default. Must be a - // value between 0 and 0777. Defaults to 0644. + // Optional: mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // Defaults to 0644. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -1034,6 +1050,17 @@ message EndpointPort { // Default is TCP. // +optional optional string protocol = 3; + + // The application protocol for this port. + // This field follows standard Kubernetes label syntax. + // Un-prefixed names are reserved for IANA standard service names (as per + // RFC-6335 and http://www.iana.org/assignments/service-names). + // Non-standard protocols should use prefixed names such as + // mycompany.com/my-custom-protocol. + // This is a beta field that is guarded by the ServiceAppProtocol feature + // gate and enabled by default. + // +optional + optional string appProtocol = 4; } // EndpointSubset is a group of addresses with a common set of ports. The @@ -1141,7 +1168,7 @@ message EnvVar { // EnvVarSource represents a source for the value of an EnvVar. message EnvVarSource { - // Selects a field of the pod: supports metadata.name, metadata.namespace, metadata.labels, metadata.annotations, + // Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['']`, `metadata.annotations['']`, // spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs. // +optional optional ObjectFieldSelector fieldRef = 1; @@ -1258,7 +1285,6 @@ message EphemeralContainerCommon { repeated VolumeMount volumeMounts = 9; // volumeDevices is the list of block devices to be used by the container. - // This is a beta feature. // +patchMergeKey=devicePath // +patchStrategy=merge // +optional @@ -1347,6 +1373,37 @@ message EphemeralContainers { repeated EphemeralContainer ephemeralContainers = 2; } +// Represents an ephemeral volume that is handled by a normal storage driver. +message EphemeralVolumeSource { + // Will be used to create a stand-alone PVC to provision the volume. + // The pod in which this EphemeralVolumeSource is embedded will be the + // owner of the PVC, i.e. the PVC will be deleted together with the + // pod. The name of the PVC will be `-` where + // `` is the name from the `PodSpec.Volumes` array + // entry. Pod validation will reject the pod if the concatenated name + // is not valid for a PVC (for example, too long). + // + // An existing PVC with that name that is not owned by the pod + // will *not* be used for the pod to avoid using an unrelated + // volume by mistake. Starting the pod is then blocked until + // the unrelated PVC is removed. If such a pre-created PVC is + // meant to be used by the pod, the PVC has to updated with an + // owner reference to the pod once the pod exists. Normally + // this should not be necessary, but it may be useful when + // manually reconstructing a broken cluster. + // + // This field is read-only and no changes will be made by Kubernetes + // to the PVC after it has been created. + // + // Required, must not be nil. + optional PersistentVolumeClaimTemplate volumeClaimTemplate = 1; + + // Specifies a read-only configuration for the volume. + // Defaults to false (read/write). + // +optional + optional bool readOnly = 2; +} + // Event is a report of an event somewhere in the cluster. message Event { // Standard object's metadata. @@ -1431,10 +1488,6 @@ message EventSeries { // Time of the last occurrence observed optional k8s.io.apimachinery.pkg.apis.meta.v1.MicroTime lastObservedTime = 2; - - // State of this Series: Ongoing or Finished - // Deprecated. Planned removal for 1.18 - optional string state = 3; } // EventSource contains information for an event. @@ -1864,8 +1917,10 @@ message KeyToPath { // May not start with the string '..'. optional string path = 2; - // Optional: mode bits to use on this file, must be a value between 0 - // and 0777. If not specified, the volume defaultMode will be used. + // Optional: mode bits used to set permissions on this file. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // If not specified, the volume defaultMode will be used. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. // +optional @@ -1913,7 +1968,6 @@ message LimitRange { // LimitRangeItem defines a min/max usage limit for any resource that matches on kind. message LimitRangeItem { // Type of resource that this limit applies to. - // +optional optional string type = 1; // Max usage constraints on this kind by resource name. @@ -2262,7 +2316,7 @@ message NodeProxyOptions { } // NodeResources is an object for conveying resource information about a node. -// see http://releases.k8s.io/HEAD/docs/design/resources.md for more details. +// see https://kubernetes.io/docs/concepts/architecture/nodes/#capacity for more details. message NodeResources { // Capacity represents the available resources of a node map capacity = 1; @@ -2416,7 +2470,7 @@ message NodeSystemInfo { // SystemUUID reported by the node. For unique machine identification // MachineID is preferred. This field is specific to Red Hat hosts - // https://access.redhat.com/documentation/en-US/Red_Hat_Subscription_Management/1/html/RHSM/getting-system-uuid.html + // https://access.redhat.com/documentation/en-us/red_hat_subscription_management/1/html/rhsm/uuid optional string systemUUID = 2; // Boot ID reported by the node. @@ -2455,6 +2509,20 @@ message ObjectFieldSelector { } // ObjectReference contains enough information to let you inspect or modify the referred object. +// --- +// New uses of this type are discouraged because of difficulty describing its usage when embedded in APIs. +// 1. Ignored fields. It includes many fields which are not generally honored. For instance, ResourceVersion and FieldPath are both very rarely valid in actual usage. +// 2. Invalid usage help. It is impossible to add specific help for individual usage. In most embedded usages, there are particular +// restrictions like, "must refer only to types A and B" or "UID not honored" or "name must be restricted". +// Those cannot be well described when embedded. +// 3. Inconsistent validation. Because the usages are different, the validation rules are different by usage, which makes it hard for users to predict what will happen. +// 4. The fields are both imprecise and overly precise. Kind is not a precise mapping to a URL. This can produce ambiguity +// during interpretation and require a REST mapping. In most cases, the dependency is on the group,resource tuple +// and the version of the actual struct is irrelevant. +// 5. We cannot easily change it. Because this type is embedded in many locations, updates to this type +// will affect numerous schemas. Don't make new APIs embed an underspecified API type they do not control. +// Instead of using this type, create a locally provided and used type that is well-focused on your reference. +// For example, ServiceReferences for admission registration: https://github.com/kubernetes/api/blob/release-1.17/admissionregistration/v1/types.go#L533 . // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object message ObjectReference { // Kind of the referent. @@ -2605,15 +2673,18 @@ message PersistentVolumeClaimSpec { // volumeMode defines what type of volume is required by the claim. // Value of Filesystem is implied when not included in claim spec. - // This is a beta feature. // +optional optional string volumeMode = 6; - // This field requires the VolumeSnapshotDataSource alpha feature gate to be - // enabled and currently VolumeSnapshot is the only supported data source. - // If the provisioner can support VolumeSnapshot data source, it will create - // a new volume and data will be restored to the volume at the same time. - // If the provisioner does not support VolumeSnapshot data source, volume will + // This field can be used to specify either: + // * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot - Beta) + // * An existing PVC (PersistentVolumeClaim) + // * An existing custom resource/object that implements data population (Alpha) + // In order to use VolumeSnapshot object types, the appropriate feature gate + // must be enabled (VolumeSnapshotDataSource or AnyVolumeDataSource) + // If the provisioner or an external controller can support the specified data source, + // it will create a new volume based on the contents of the specified data source. + // If the specified data source is not supported, the volume will // not be created and the failure will be reported as an event. // In the future, we plan to support more data source types and the behavior // of the provisioner may change. @@ -2644,6 +2715,23 @@ message PersistentVolumeClaimStatus { repeated PersistentVolumeClaimCondition conditions = 4; } +// PersistentVolumeClaimTemplate is used to produce +// PersistentVolumeClaim objects as part of an EphemeralVolumeSource. +message PersistentVolumeClaimTemplate { + // May contain labels and annotations that will be copied into the PVC + // when creating it. No other fields are allowed and will be rejected during + // validation. + // + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // The specification for the PersistentVolumeClaim. The entire content is + // copied unchanged into the PVC that gets created from this + // template. The same fields as in a PersistentVolumeClaim + // are also valid here. + optional PersistentVolumeClaimSpec spec = 2; +} + // PersistentVolumeClaimVolumeSource references the user's PVC in the same namespace. // This volume finds the bound PV and mounts that volume for the pod. A // PersistentVolumeClaimVolumeSource is, essentially, a wrapper around another @@ -2821,7 +2909,6 @@ message PersistentVolumeSpec { // volumeMode defines if a volume is intended to be used with a formatted filesystem // or to remain in raw block state. Value of Filesystem is implied when not included in spec. - // This is a beta feature. // +optional optional string volumeMode = 8; @@ -3247,6 +3334,19 @@ message PodSecurityContext { // sysctls (by the container runtime) might fail to launch. // +optional repeated Sysctl sysctls = 7; + + // fsGroupChangePolicy defines behavior of changing ownership and permission of the volume + // before being exposed inside Pod. This field will only apply to + // volume types which support fsGroup based ownership(and permissions). + // It will have no effect on ephemeral volume types such as: secret, configmaps + // and emptydir. + // Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always". + // +optional + optional string fsGroupChangePolicy = 9; + + // The seccomp options to use by the containers in this pod. + // +optional + optional SeccompProfile seccompProfile = 10; } // Describes the class of pods that should avoid this node. @@ -3480,7 +3580,7 @@ message PodSpec { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional optional string preemptionPolicy = 31; @@ -3497,8 +3597,6 @@ message PodSpec { // TopologySpreadConstraints describes how a group of pods ought to spread across topology // domains. Scheduler will schedule pods in a way which abides by the constraints. - // This field is alpha-level and is only honored by clusters that enables the EvenPodsSpread - // feature. // All topologySpreadConstraints are ANDed. // +optional // +patchMergeKey=topologyKey @@ -3507,6 +3605,14 @@ message PodSpec { // +listMapKey=topologyKey // +listMapKey=whenUnsatisfiable repeated TopologySpreadConstraint topologySpreadConstraints = 33; + + // If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). + // In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). + // In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. + // If a pod does not have FQDN, this has no effect. + // Default to false. + // +optional + optional bool setHostnameAsFQDN = 35; } // PodStatus represents information about the status of a pod. Status may trail the actual @@ -3749,8 +3855,9 @@ message ProjectedVolumeSource { // list of volume projections repeated VolumeProjection sources = 1; - // Mode bits to use on created files by default. Must be a value between - // 0 and 0777. + // Mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -4248,6 +4355,27 @@ message ScopedResourceSelectorRequirement { repeated string values = 3; } +// SeccompProfile defines a pod/container's seccomp profile settings. +// Only one profile source may be set. +// +union +message SeccompProfile { + // type indicates which kind of seccomp profile will be applied. + // Valid options are: + // + // Localhost - a profile defined in a file on the node should be used. + // RuntimeDefault - the container runtime default profile should be used. + // Unconfined - no profile should be applied. + // +unionDiscriminator + optional string type = 1; + + // localhostProfile indicates a profile defined in a file on the node should be used. + // The profile must be preconfigured on the node to work. + // Must be a descending path, relative to the kubelet's configured seccomp profile location. + // Must only be set if type is "Localhost". + // +optional + optional string localhostProfile = 2; +} + // Secret holds secret data of a certain type. The total bytes of the values in // the Data field must be less than MaxSecretSize bytes. message Secret { @@ -4256,6 +4384,14 @@ message Secret { // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + // Immutable, if set to true, ensures that data stored in the Secret cannot + // be updated (only object metadata can be modified). + // If not set to true, the field can be modified at any time. + // Defaulted to nil. + // This is a beta field enabled by ImmutableEphemeralVolumes feature gate. + // +optional + optional bool immutable = 5; + // Data contains the secret data. Each key must consist of alphanumeric // characters, '-', '_' or '.'. The serialized form of the secret data is a // base64 encoded string, representing the arbitrary (possibly non-string) @@ -4372,8 +4508,10 @@ message SecretVolumeSource { // +optional repeated KeyToPath items = 2; - // Optional: mode bits to use on created files by default. Must be a - // value between 0 and 0777. Defaults to 0644. + // Optional: mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values + // for mode bits. Defaults to 0644. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -4456,6 +4594,12 @@ message SecurityContext { // This requires the ProcMountType feature flag to be enabled. // +optional optional string procMount = 9; + + // The seccomp options to use by this container. If seccomp options are + // provided at both the pod & container level, the container options + // override the pod options. + // +optional + optional SeccompProfile seccompProfile = 11; } // SerializedReference is a reference to serialized object. @@ -4581,6 +4725,17 @@ message ServicePort { // +optional optional string protocol = 2; + // The application protocol for this port. + // This field follows standard Kubernetes label syntax. + // Un-prefixed names are reserved for IANA standard service names (as per + // RFC-6335 and http://www.iana.org/assignments/service-names). + // Non-standard protocols should use prefixed names such as + // mycompany.com/my-custom-protocol. + // This is a beta field that is guarded by the ServiceAppProtocol feature + // gate and enabled by default. + // +optional + optional string appProtocol = 6; + // The port that will be exposed by this service. optional int32 port = 3; @@ -4719,12 +4874,14 @@ message ServiceSpec { // +optional optional int32 healthCheckNodePort = 12; - // publishNotReadyAddresses, when set to true, indicates that DNS implementations - // must publish the notReadyAddresses of subsets for the Endpoints associated with - // the Service. The default value is false. - // The primary use case for setting this field is to use a StatefulSet's Headless Service - // to propagate SRV records for its Pods without respect to their readiness for purpose - // of peer discovery. + // publishNotReadyAddresses indicates that any agent which deals with endpoints for this + // Service should disregard any indications of ready/not-ready. + // The primary use case for setting this field is for a StatefulSet's Headless Service to + // propagate SRV DNS records for its Pods for the purpose of peer discovery. + // The Kubernetes controllers that generate Endpoints and EndpointSlice resources for + // Services interpret this to mean that all endpoints are considered "ready" even if the + // Pods themselves are not. Agents which consume only Kubernetes generated endpoints + // through the Endpoints or EndpointSlice resources can safely assume this behavior. // +optional optional bool publishNotReadyAddresses = 13; @@ -4732,13 +4889,21 @@ message ServiceSpec { // +optional optional SessionAffinityConfig sessionAffinityConfig = 14; - // ipFamily specifies whether this Service has a preference for a particular IP family (e.g. IPv4 vs. - // IPv6). If a specific IP family is requested, the clusterIP field will be allocated from that family, if it is - // available in the cluster. If no IP family is requested, the cluster's primary IP family will be used. - // Other IP fields (loadBalancerIP, loadBalancerSourceRanges, externalIPs) and controllers which - // allocate external load-balancers should use the same IP family. Endpoints for this Service will be of - // this family. This field is immutable after creation. Assigning a ServiceIPFamily not available in the - // cluster (e.g. IPv6 in IPv4 only cluster) is an error condition and will fail during clusterIP assignment. + // ipFamily specifies whether this Service has a preference for a particular IP family (e.g. + // IPv4 vs. IPv6) when the IPv6DualStack feature gate is enabled. In a dual-stack cluster, + // you can specify ipFamily when creating a ClusterIP Service to determine whether the + // controller will allocate an IPv4 or IPv6 IP for it, and you can specify ipFamily when + // creating a headless Service to determine whether it will have IPv4 or IPv6 Endpoints. In + // either case, if you do not specify an ipFamily explicitly, it will default to the + // cluster's primary IP family. + // This field is part of an alpha feature, and you should not make any assumptions about its + // semantics other than those described above. In particular, you should not assume that it + // can (or cannot) be changed after creation time; that it can only have the values "IPv4" + // and "IPv6"; or that its current value on a given Service correctly reflects the current + // state of that Service. (For ClusterIP Services, look at clusterIP to see if the Service + // is IPv4 or IPv6. For headless Services, look at the endpoints, which may be dual-stack in + // the future. For ExternalName Services, ipFamily has no meaning, but it may be set to an + // irrelevant value anyway.) // +optional optional string ipFamily = 15; @@ -4864,7 +5029,7 @@ message Taint { // Required. The taint key to be applied to a node. optional string key = 1; - // Required. The taint value corresponding to the taint key. + // The taint value corresponding to the taint key. // +optional optional string value = 2; @@ -4937,8 +5102,8 @@ message TopologySelectorTerm { // TopologySpreadConstraint specifies how to spread matching pods among the given topology. message TopologySpreadConstraint { // MaxSkew describes the degree to which pods may be unevenly distributed. - // It's the maximum permitted difference between the number of matching pods in - // any two topology domains of a given topology type. + // When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference + // between the number of matching pods in the target topology and the global minimum. // For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same // labelSelector spread as 1/1/0: // +-------+-------+-------+ @@ -4950,6 +5115,8 @@ message TopologySpreadConstraint { // scheduling it onto zone1(zone2) would make the ActualSkew(2-0) on zone1(zone2) // violate MaxSkew(1). // - if MaxSkew is 2, incoming pod can be scheduled onto any zone. + // When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence + // to topologies that satisfy it. // It's a required field. Default value is 1 and 0 is not allowed. optional int32 maxSkew = 1; @@ -4962,10 +5129,13 @@ message TopologySpreadConstraint { // WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy // the spread constraint. - // - DoNotSchedule (default) tells the scheduler not to schedule it - // - ScheduleAnyway tells the scheduler to still schedule it - // It's considered as "Unsatisfiable" if and only if placing incoming pod on any - // topology violates "MaxSkew". + // - DoNotSchedule (default) tells the scheduler not to schedule it. + // - ScheduleAnyway tells the scheduler to schedule the pod in any location, + // but giving higher precedence to topologies that would help reduce the + // skew. + // A constraint is considered "Unsatisfiable" for an incoming pod + // if and only if every possible node assigment for that pod would violate + // "MaxSkew" on some topology. // For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same // labelSelector spread as 3/1/1: // +-------+-------+-------+ @@ -5218,9 +5388,37 @@ message VolumeSource { // +optional optional StorageOSVolumeSource storageos = 27; - // CSI (Container Storage Interface) represents storage that is handled by an external CSI driver (Alpha feature). + // CSI (Container Storage Interface) represents ephemeral storage that is handled by certain external CSI drivers (Beta feature). // +optional optional CSIVolumeSource csi = 28; + + // Ephemeral represents a volume that is handled by a cluster storage driver (Alpha feature). + // The volume's lifecycle is tied to the pod that defines it - it will be created before the pod starts, + // and deleted when the pod is removed. + // + // Use this if: + // a) the volume is only needed while the pod runs, + // b) features of normal volumes like restoring from snapshot or capacity + // tracking are needed, + // c) the storage driver is specified through a storage class, and + // d) the storage driver supports dynamic volume provisioning through + // a PersistentVolumeClaim (see EphemeralVolumeSource for more + // information on the connection between this volume type + // and PersistentVolumeClaim). + // + // Use PersistentVolumeClaim or one of the vendor-specific + // APIs for volumes that persist for longer than the lifecycle + // of an individual pod. + // + // Use CSI for light-weight local ephemeral volumes if the CSI driver is meant to + // be used that way - see the documentation of the driver for + // more information. + // + // A pod can use both types of ephemeral volumes and + // persistent volumes at the same time. + // + // +optional + optional EphemeralVolumeSource ephemeral = 29; } // Represents a vSphere volume resource. @@ -5256,14 +5454,12 @@ message WeightedPodAffinityTerm { // WindowsSecurityContextOptions contain Windows-specific options and credentials. message WindowsSecurityContextOptions { // GMSACredentialSpecName is the name of the GMSA credential spec to use. - // This field is alpha-level and is only honored by servers that enable the WindowsGMSA feature flag. // +optional optional string gmsaCredentialSpecName = 1; // GMSACredentialSpec is where the GMSA admission webhook // (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the // GMSA credential spec named by the GMSACredentialSpecName field. - // This field is alpha-level and is only honored by servers that enable the WindowsGMSA feature flag. // +optional optional string gmsaCredentialSpec = 2; @@ -5271,7 +5467,6 @@ message WindowsSecurityContextOptions { // Defaults to the user specified in image metadata if unspecified. // May also be set in PodSecurityContext. If set in both SecurityContext and // PodSecurityContext, the value specified in SecurityContext takes precedence. - // This field is beta-level and may be disabled with the WindowsRunAsUserName feature flag. // +optional optional string runAsUserName = 3; } diff --git a/vendor/k8s.io/api/core/v1/lifecycle.go b/vendor/k8s.io/api/core/v1/lifecycle.go new file mode 100644 index 00000000..21ca90e8 --- /dev/null +++ b/vendor/k8s.io/api/core/v1/lifecycle.go @@ -0,0 +1,37 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package v1 + +// APILifecycleIntroduced returns the release in which the API struct was introduced as int versions of major and minor for comparison. +func (in *ComponentStatus) APILifecycleIntroduced() (major, minor int) { + return 1, 0 +} + +// APILifecycleDeprecated returns the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +func (in *ComponentStatus) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleIntroduced returns the release in which the API struct was introduced as int versions of major and minor for comparison. +func (in *ComponentStatusList) APILifecycleIntroduced() (major, minor int) { + return 1, 0 +} + +// APILifecycleDeprecated returns the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +func (in *ComponentStatusList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} diff --git a/vendor/k8s.io/api/core/v1/resource.go b/vendor/k8s.io/api/core/v1/resource.go index bb804125..5bc9cd5b 100644 --- a/vendor/k8s.io/api/core/v1/resource.go +++ b/vendor/k8s.io/api/core/v1/resource.go @@ -41,6 +41,14 @@ func (self *ResourceList) Memory() *resource.Quantity { return &resource.Quantity{Format: resource.BinarySI} } +// Returns the Storage limit if specified. +func (self *ResourceList) Storage() *resource.Quantity { + if val, ok := (*self)[ResourceStorage]; ok { + return &val + } + return &resource.Quantity{Format: resource.BinarySI} +} + func (self *ResourceList) Pods() *resource.Quantity { if val, ok := (*self)[ResourcePods]; ok { return &val diff --git a/vendor/k8s.io/api/core/v1/types.go b/vendor/k8s.io/api/core/v1/types.go index 47a40271..f3ec52e7 100644 --- a/vendor/k8s.io/api/core/v1/types.go +++ b/vendor/k8s.io/api/core/v1/types.go @@ -153,9 +153,36 @@ type VolumeSource struct { // StorageOS represents a StorageOS volume attached and mounted on Kubernetes nodes. // +optional StorageOS *StorageOSVolumeSource `json:"storageos,omitempty" protobuf:"bytes,27,opt,name=storageos"` - // CSI (Container Storage Interface) represents storage that is handled by an external CSI driver (Alpha feature). + // CSI (Container Storage Interface) represents ephemeral storage that is handled by certain external CSI drivers (Beta feature). // +optional CSI *CSIVolumeSource `json:"csi,omitempty" protobuf:"bytes,28,opt,name=csi"` + // Ephemeral represents a volume that is handled by a cluster storage driver (Alpha feature). + // The volume's lifecycle is tied to the pod that defines it - it will be created before the pod starts, + // and deleted when the pod is removed. + // + // Use this if: + // a) the volume is only needed while the pod runs, + // b) features of normal volumes like restoring from snapshot or capacity + // tracking are needed, + // c) the storage driver is specified through a storage class, and + // d) the storage driver supports dynamic volume provisioning through + // a PersistentVolumeClaim (see EphemeralVolumeSource for more + // information on the connection between this volume type + // and PersistentVolumeClaim). + // + // Use PersistentVolumeClaim or one of the vendor-specific + // APIs for volumes that persist for longer than the lifecycle + // of an individual pod. + // + // Use CSI for light-weight local ephemeral volumes if the CSI driver is meant to + // be used that way - see the documentation of the driver for + // more information. + // + // A pod can use both types of ephemeral volumes and + // persistent volumes at the same time. + // + // +optional + Ephemeral *EphemeralVolumeSource `json:"ephemeral,omitempty" protobuf:"bytes,29,opt,name=ephemeral"` } // PersistentVolumeClaimVolumeSource references the user's PVC in the same namespace. @@ -331,7 +358,6 @@ type PersistentVolumeSpec struct { MountOptions []string `json:"mountOptions,omitempty" protobuf:"bytes,7,opt,name=mountOptions"` // volumeMode defines if a volume is intended to be used with a formatted filesystem // or to remain in raw block state. Value of Filesystem is implied when not included in spec. - // This is a beta feature. // +optional VolumeMode *PersistentVolumeMode `json:"volumeMode,omitempty" protobuf:"bytes,8,opt,name=volumeMode,casttype=PersistentVolumeMode"` // NodeAffinity defines constraints that limit what nodes this volume can be accessed from. @@ -460,14 +486,17 @@ type PersistentVolumeClaimSpec struct { StorageClassName *string `json:"storageClassName,omitempty" protobuf:"bytes,5,opt,name=storageClassName"` // volumeMode defines what type of volume is required by the claim. // Value of Filesystem is implied when not included in claim spec. - // This is a beta feature. // +optional VolumeMode *PersistentVolumeMode `json:"volumeMode,omitempty" protobuf:"bytes,6,opt,name=volumeMode,casttype=PersistentVolumeMode"` - // This field requires the VolumeSnapshotDataSource alpha feature gate to be - // enabled and currently VolumeSnapshot is the only supported data source. - // If the provisioner can support VolumeSnapshot data source, it will create - // a new volume and data will be restored to the volume at the same time. - // If the provisioner does not support VolumeSnapshot data source, volume will + // This field can be used to specify either: + // * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot - Beta) + // * An existing PVC (PersistentVolumeClaim) + // * An existing custom resource/object that implements data population (Alpha) + // In order to use VolumeSnapshot object types, the appropriate feature gate + // must be enabled (VolumeSnapshotDataSource or AnyVolumeDataSource) + // If the provisioner or an external controller can support the specified data source, + // it will create a new volume based on the contents of the specified data source. + // If the specified data source is not supported, the volume will // not be created and the failure will be reported as an event. // In the future, we plan to support more data source types and the behavior // of the provisioner may change. @@ -887,9 +916,10 @@ type FlockerVolumeSource struct { type StorageMedium string const ( - StorageMediumDefault StorageMedium = "" // use whatever the default is for the node, assume anything we don't explicitly handle is this - StorageMediumMemory StorageMedium = "Memory" // use memory (e.g. tmpfs on linux) - StorageMediumHugePages StorageMedium = "HugePages" // use hugepages + StorageMediumDefault StorageMedium = "" // use whatever the default is for the node, assume anything we don't explicitly handle is this + StorageMediumMemory StorageMedium = "Memory" // use memory (e.g. tmpfs on linux) + StorageMediumHugePages StorageMedium = "HugePages" // use hugepages + StorageMediumHugePagesPrefix StorageMedium = "HugePages-" // prefix for full medium notation HugePages- ) // Protocol defines network protocols supported for things like container ports. @@ -1089,8 +1119,10 @@ type SecretVolumeSource struct { // relative and may not contain the '..' path or start with '..'. // +optional Items []KeyToPath `json:"items,omitempty" protobuf:"bytes,2,rep,name=items"` - // Optional: mode bits to use on created files by default. Must be a - // value between 0 and 0777. Defaults to 0644. + // Optional: mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values + // for mode bits. Defaults to 0644. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -1515,8 +1547,10 @@ type ConfigMapVolumeSource struct { // relative and may not contain the '..' path or start with '..'. // +optional Items []KeyToPath `json:"items,omitempty" protobuf:"bytes,2,rep,name=items"` - // Optional: mode bits to use on created files by default. Must be a - // value between 0 and 0777. Defaults to 0644. + // Optional: mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // Defaults to 0644. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -1582,8 +1616,9 @@ type ServiceAccountTokenProjection struct { type ProjectedVolumeSource struct { // list of volume projections Sources []VolumeProjection `json:"sources" protobuf:"bytes,1,rep,name=sources"` - // Mode bits to use on created files by default. Must be a value between - // 0 and 0777. + // Mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -1623,8 +1658,10 @@ type KeyToPath struct { // May not contain the path element '..'. // May not start with the string '..'. Path string `json:"path" protobuf:"bytes,2,opt,name=path"` - // Optional: mode bits to use on this file, must be a value between 0 - // and 0777. If not specified, the volume defaultMode will be used. + // Optional: mode bits used to set permissions on this file. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // If not specified, the volume defaultMode will be used. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. // +optional @@ -1736,6 +1773,54 @@ type CSIVolumeSource struct { NodePublishSecretRef *LocalObjectReference `json:"nodePublishSecretRef,omitempty" protobuf:"bytes,5,opt,name=nodePublishSecretRef"` } +// Represents an ephemeral volume that is handled by a normal storage driver. +type EphemeralVolumeSource struct { + // Will be used to create a stand-alone PVC to provision the volume. + // The pod in which this EphemeralVolumeSource is embedded will be the + // owner of the PVC, i.e. the PVC will be deleted together with the + // pod. The name of the PVC will be `-` where + // `` is the name from the `PodSpec.Volumes` array + // entry. Pod validation will reject the pod if the concatenated name + // is not valid for a PVC (for example, too long). + // + // An existing PVC with that name that is not owned by the pod + // will *not* be used for the pod to avoid using an unrelated + // volume by mistake. Starting the pod is then blocked until + // the unrelated PVC is removed. If such a pre-created PVC is + // meant to be used by the pod, the PVC has to updated with an + // owner reference to the pod once the pod exists. Normally + // this should not be necessary, but it may be useful when + // manually reconstructing a broken cluster. + // + // This field is read-only and no changes will be made by Kubernetes + // to the PVC after it has been created. + // + // Required, must not be nil. + VolumeClaimTemplate *PersistentVolumeClaimTemplate `json:"volumeClaimTemplate,omitempty" protobuf:"bytes,1,opt,name=volumeClaimTemplate"` + + // Specifies a read-only configuration for the volume. + // Defaults to false (read/write). + // +optional + ReadOnly bool `json:"readOnly,omitempty" protobuf:"varint,2,opt,name=readOnly"` +} + +// PersistentVolumeClaimTemplate is used to produce +// PersistentVolumeClaim objects as part of an EphemeralVolumeSource. +type PersistentVolumeClaimTemplate struct { + // May contain labels and annotations that will be copied into the PVC + // when creating it. No other fields are allowed and will be rejected during + // validation. + // + // +optional + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // The specification for the PersistentVolumeClaim. The entire content is + // copied unchanged into the PVC that gets created from this + // template. The same fields as in a PersistentVolumeClaim + // are also valid here. + Spec PersistentVolumeClaimSpec `json:"spec" protobuf:"bytes,2,name=spec"` +} + // ContainerPort represents a network port in a single container. type ContainerPort struct { // If specified, this must be an IANA_SVC_NAME and unique within the pod. Each @@ -1847,7 +1932,7 @@ type EnvVar struct { // EnvVarSource represents a source for the value of an EnvVar. type EnvVarSource struct { - // Selects a field of the pod: supports metadata.name, metadata.namespace, metadata.labels, metadata.annotations, + // Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['']`, `metadata.annotations['']`, // spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs. // +optional FieldRef *ObjectFieldSelector `json:"fieldRef,omitempty" protobuf:"bytes,1,opt,name=fieldRef"` @@ -2180,7 +2265,6 @@ type Container struct { // +patchStrategy=merge VolumeMounts []VolumeMount `json:"volumeMounts,omitempty" patchStrategy:"merge" patchMergeKey:"mountPath" protobuf:"bytes,9,rep,name=volumeMounts"` // volumeDevices is the list of block devices to be used by the container. - // This is a beta feature. // +patchMergeKey=devicePath // +patchStrategy=merge // +optional @@ -2203,7 +2287,7 @@ type Container struct { // This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, // when it might take a long time to load data or warm a cache, than during steady-state operation. // This cannot be updated. - // This is an alpha feature enabled by the StartupProbe feature flag. + // This is a beta feature enabled by the StartupProbe feature flag. // More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes // +optional StartupProbe *Probe `json:"startupProbe,omitempty" protobuf:"bytes,22,opt,name=startupProbe"` @@ -2750,7 +2834,7 @@ type PreferredSchedulingTerm struct { type Taint struct { // Required. The taint key to be applied to a node. Key string `json:"key" protobuf:"bytes,1,opt,name=key"` - // Required. The taint value corresponding to the taint key. + // The taint value corresponding to the taint key. // +optional Value string `json:"value,omitempty" protobuf:"bytes,2,opt,name=value"` // Required. The effect of the taint on pods @@ -3023,7 +3107,7 @@ type PodSpec struct { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional PreemptionPolicy *PreemptionPolicy `json:"preemptionPolicy,omitempty" protobuf:"bytes,31,opt,name=preemptionPolicy"` // Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. @@ -3038,8 +3122,6 @@ type PodSpec struct { Overhead ResourceList `json:"overhead,omitempty" protobuf:"bytes,32,opt,name=overhead"` // TopologySpreadConstraints describes how a group of pods ought to spread across topology // domains. Scheduler will schedule pods in a way which abides by the constraints. - // This field is alpha-level and is only honored by clusters that enables the EvenPodsSpread - // feature. // All topologySpreadConstraints are ANDed. // +optional // +patchMergeKey=topologyKey @@ -3048,6 +3130,13 @@ type PodSpec struct { // +listMapKey=topologyKey // +listMapKey=whenUnsatisfiable TopologySpreadConstraints []TopologySpreadConstraint `json:"topologySpreadConstraints,omitempty" patchStrategy:"merge" patchMergeKey:"topologyKey" protobuf:"bytes,33,opt,name=topologySpreadConstraints"` + // If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). + // In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). + // In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. + // If a pod does not have FQDN, this has no effect. + // Default to false. + // +optional + SetHostnameAsFQDN *bool `json:"setHostnameAsFQDN,omitempty" protobuf:"varint,35,opt,name=setHostnameAsFQDN"` } type UnsatisfiableConstraintAction string @@ -3064,8 +3153,8 @@ const ( // TopologySpreadConstraint specifies how to spread matching pods among the given topology. type TopologySpreadConstraint struct { // MaxSkew describes the degree to which pods may be unevenly distributed. - // It's the maximum permitted difference between the number of matching pods in - // any two topology domains of a given topology type. + // When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference + // between the number of matching pods in the target topology and the global minimum. // For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same // labelSelector spread as 1/1/0: // +-------+-------+-------+ @@ -3077,6 +3166,8 @@ type TopologySpreadConstraint struct { // scheduling it onto zone1(zone2) would make the ActualSkew(2-0) on zone1(zone2) // violate MaxSkew(1). // - if MaxSkew is 2, incoming pod can be scheduled onto any zone. + // When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence + // to topologies that satisfy it. // It's a required field. Default value is 1 and 0 is not allowed. MaxSkew int32 `json:"maxSkew" protobuf:"varint,1,opt,name=maxSkew"` // TopologyKey is the key of node labels. Nodes that have a label with this key @@ -3087,10 +3178,13 @@ type TopologySpreadConstraint struct { TopologyKey string `json:"topologyKey" protobuf:"bytes,2,opt,name=topologyKey"` // WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy // the spread constraint. - // - DoNotSchedule (default) tells the scheduler not to schedule it - // - ScheduleAnyway tells the scheduler to still schedule it - // It's considered as "Unsatisfiable" if and only if placing incoming pod on any - // topology violates "MaxSkew". + // - DoNotSchedule (default) tells the scheduler not to schedule it. + // - ScheduleAnyway tells the scheduler to schedule the pod in any location, + // but giving higher precedence to topologies that would help reduce the + // skew. + // A constraint is considered "Unsatisfiable" for an incoming pod + // if and only if every possible node assigment for that pod would violate + // "MaxSkew" on some topology. // For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same // labelSelector spread as 3/1/1: // +-------+-------+-------+ @@ -3125,6 +3219,22 @@ type HostAlias struct { Hostnames []string `json:"hostnames,omitempty" protobuf:"bytes,2,rep,name=hostnames"` } +// PodFSGroupChangePolicy holds policies that will be used for applying fsGroup to a volume +// when volume is mounted. +type PodFSGroupChangePolicy string + +const ( + // FSGroupChangeOnRootMismatch indicates that volume's ownership and permissions will be changed + // only when permission and ownership of root directory does not match with expected + // permissions on the volume. This can help shorten the time it takes to change + // ownership and permissions of a volume. + FSGroupChangeOnRootMismatch PodFSGroupChangePolicy = "OnRootMismatch" + // FSGroupChangeAlways indicates that volume's ownership and permissions + // should always be changed whenever volume is mounted inside a Pod. This the default + // behavior. + FSGroupChangeAlways PodFSGroupChangePolicy = "Always" +) + // PodSecurityContext holds pod-level security attributes and common container settings. // Some fields are also present in container.securityContext. Field values of // container.securityContext take precedence over field values of PodSecurityContext. @@ -3183,8 +3293,53 @@ type PodSecurityContext struct { // sysctls (by the container runtime) might fail to launch. // +optional Sysctls []Sysctl `json:"sysctls,omitempty" protobuf:"bytes,7,rep,name=sysctls"` + // fsGroupChangePolicy defines behavior of changing ownership and permission of the volume + // before being exposed inside Pod. This field will only apply to + // volume types which support fsGroup based ownership(and permissions). + // It will have no effect on ephemeral volume types such as: secret, configmaps + // and emptydir. + // Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always". + // +optional + FSGroupChangePolicy *PodFSGroupChangePolicy `json:"fsGroupChangePolicy,omitempty" protobuf:"bytes,9,opt,name=fsGroupChangePolicy"` + // The seccomp options to use by the containers in this pod. + // +optional + SeccompProfile *SeccompProfile `json:"seccompProfile,omitempty" protobuf:"bytes,10,opt,name=seccompProfile"` +} + +// SeccompProfile defines a pod/container's seccomp profile settings. +// Only one profile source may be set. +// +union +type SeccompProfile struct { + // type indicates which kind of seccomp profile will be applied. + // Valid options are: + // + // Localhost - a profile defined in a file on the node should be used. + // RuntimeDefault - the container runtime default profile should be used. + // Unconfined - no profile should be applied. + // +unionDiscriminator + Type SeccompProfileType `json:"type" protobuf:"bytes,1,opt,name=type,casttype=SeccompProfileType"` + // localhostProfile indicates a profile defined in a file on the node should be used. + // The profile must be preconfigured on the node to work. + // Must be a descending path, relative to the kubelet's configured seccomp profile location. + // Must only be set if type is "Localhost". + // +optional + LocalhostProfile *string `json:"localhostProfile,omitempty" protobuf:"bytes,2,opt,name=localhostProfile"` } +// SeccompProfileType defines the supported seccomp profile types. +type SeccompProfileType string + +const ( + // SeccompProfileTypeUnconfined indicates no seccomp profile is applied (A.K.A. unconfined). + SeccompProfileTypeUnconfined SeccompProfileType = "Unconfined" + // SeccompProfileTypeRuntimeDefault represents the default container runtime seccomp profile. + SeccompProfileTypeRuntimeDefault SeccompProfileType = "RuntimeDefault" + // SeccompProfileTypeLocalhost indicates a profile defined in a file on the node should be used. + // The file's location is based off the kubelet's deprecated flag --seccomp-profile-root. + // Once the flag support is removed the location will be /seccomp. + SeccompProfileTypeLocalhost SeccompProfileType = "Localhost" +) + // PodQOSClass defines the supported qos classes of Pods. type PodQOSClass string @@ -3298,7 +3453,6 @@ type EphemeralContainerCommon struct { // +patchStrategy=merge VolumeMounts []VolumeMount `json:"volumeMounts,omitempty" patchStrategy:"merge" patchMergeKey:"mountPath" protobuf:"bytes,9,rep,name=volumeMounts"` // volumeDevices is the list of block devices to be used by the container. - // This is a beta feature. // +patchMergeKey=devicePath // +patchStrategy=merge // +optional @@ -3936,12 +4090,14 @@ type ServiceSpec struct { // +optional HealthCheckNodePort int32 `json:"healthCheckNodePort,omitempty" protobuf:"bytes,12,opt,name=healthCheckNodePort"` - // publishNotReadyAddresses, when set to true, indicates that DNS implementations - // must publish the notReadyAddresses of subsets for the Endpoints associated with - // the Service. The default value is false. - // The primary use case for setting this field is to use a StatefulSet's Headless Service - // to propagate SRV records for its Pods without respect to their readiness for purpose - // of peer discovery. + // publishNotReadyAddresses indicates that any agent which deals with endpoints for this + // Service should disregard any indications of ready/not-ready. + // The primary use case for setting this field is for a StatefulSet's Headless Service to + // propagate SRV DNS records for its Pods for the purpose of peer discovery. + // The Kubernetes controllers that generate Endpoints and EndpointSlice resources for + // Services interpret this to mean that all endpoints are considered "ready" even if the + // Pods themselves are not. Agents which consume only Kubernetes generated endpoints + // through the Endpoints or EndpointSlice resources can safely assume this behavior. // +optional PublishNotReadyAddresses bool `json:"publishNotReadyAddresses,omitempty" protobuf:"varint,13,opt,name=publishNotReadyAddresses"` @@ -3949,13 +4105,21 @@ type ServiceSpec struct { // +optional SessionAffinityConfig *SessionAffinityConfig `json:"sessionAffinityConfig,omitempty" protobuf:"bytes,14,opt,name=sessionAffinityConfig"` - // ipFamily specifies whether this Service has a preference for a particular IP family (e.g. IPv4 vs. - // IPv6). If a specific IP family is requested, the clusterIP field will be allocated from that family, if it is - // available in the cluster. If no IP family is requested, the cluster's primary IP family will be used. - // Other IP fields (loadBalancerIP, loadBalancerSourceRanges, externalIPs) and controllers which - // allocate external load-balancers should use the same IP family. Endpoints for this Service will be of - // this family. This field is immutable after creation. Assigning a ServiceIPFamily not available in the - // cluster (e.g. IPv6 in IPv4 only cluster) is an error condition and will fail during clusterIP assignment. + // ipFamily specifies whether this Service has a preference for a particular IP family (e.g. + // IPv4 vs. IPv6) when the IPv6DualStack feature gate is enabled. In a dual-stack cluster, + // you can specify ipFamily when creating a ClusterIP Service to determine whether the + // controller will allocate an IPv4 or IPv6 IP for it, and you can specify ipFamily when + // creating a headless Service to determine whether it will have IPv4 or IPv6 Endpoints. In + // either case, if you do not specify an ipFamily explicitly, it will default to the + // cluster's primary IP family. + // This field is part of an alpha feature, and you should not make any assumptions about its + // semantics other than those described above. In particular, you should not assume that it + // can (or cannot) be changed after creation time; that it can only have the values "IPv4" + // and "IPv6"; or that its current value on a given Service correctly reflects the current + // state of that Service. (For ClusterIP Services, look at clusterIP to see if the Service + // is IPv4 or IPv6. For headless Services, look at the endpoints, which may be dual-stack in + // the future. For ExternalName Services, ipFamily has no meaning, but it may be set to an + // irrelevant value anyway.) // +optional IPFamily *IPFamily `json:"ipFamily,omitempty" protobuf:"bytes,15,opt,name=ipFamily,Configcasttype=IPFamily"` @@ -3990,6 +4154,17 @@ type ServicePort struct { // +optional Protocol Protocol `json:"protocol,omitempty" protobuf:"bytes,2,opt,name=protocol,casttype=Protocol"` + // The application protocol for this port. + // This field follows standard Kubernetes label syntax. + // Un-prefixed names are reserved for IANA standard service names (as per + // RFC-6335 and http://www.iana.org/assignments/service-names). + // Non-standard protocols should use prefixed names such as + // mycompany.com/my-custom-protocol. + // This is a beta field that is guarded by the ServiceAppProtocol feature + // gate and enabled by default. + // +optional + AppProtocol *string `json:"appProtocol,omitempty" protobuf:"bytes,6,opt,name=appProtocol"` + // The port that will be exposed by this service. Port int32 `json:"port" protobuf:"varint,3,opt,name=port"` @@ -4061,6 +4236,7 @@ type ServiceList struct { } // +genclient +// +genclient:method=CreateToken,verb=create,subresource=token,input=k8s.io/api/authentication/v1.TokenRequest,result=k8s.io/api/authentication/v1.TokenRequest // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // ServiceAccount binds together: @@ -4204,6 +4380,17 @@ type EndpointPort struct { // Default is TCP. // +optional Protocol Protocol `json:"protocol,omitempty" protobuf:"bytes,3,opt,name=protocol,casttype=Protocol"` + + // The application protocol for this port. + // This field follows standard Kubernetes label syntax. + // Un-prefixed names are reserved for IANA standard service names (as per + // RFC-6335 and http://www.iana.org/assignments/service-names). + // Non-standard protocols should use prefixed names such as + // mycompany.com/my-custom-protocol. + // This is a beta field that is guarded by the ServiceAppProtocol feature + // gate and enabled by default. + // +optional + AppProtocol *string `json:"appProtocol,omitempty" protobuf:"bytes,4,opt,name=appProtocol"` } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object @@ -4323,7 +4510,7 @@ type NodeSystemInfo struct { MachineID string `json:"machineID" protobuf:"bytes,1,opt,name=machineID"` // SystemUUID reported by the node. For unique machine identification // MachineID is preferred. This field is specific to Red Hat hosts - // https://access.redhat.com/documentation/en-US/Red_Hat_Subscription_Management/1/html/RHSM/getting-system-uuid.html + // https://access.redhat.com/documentation/en-us/red_hat_subscription_management/1/html/rhsm/uuid SystemUUID string `json:"systemUUID" protobuf:"bytes,2,opt,name=systemUUID"` // Boot ID reported by the node. BootID string `json:"bootID" protobuf:"bytes,3,opt,name=bootID"` @@ -4981,6 +5168,20 @@ type ServiceProxyOptions struct { } // ObjectReference contains enough information to let you inspect or modify the referred object. +// --- +// New uses of this type are discouraged because of difficulty describing its usage when embedded in APIs. +// 1. Ignored fields. It includes many fields which are not generally honored. For instance, ResourceVersion and FieldPath are both very rarely valid in actual usage. +// 2. Invalid usage help. It is impossible to add specific help for individual usage. In most embedded usages, there are particular +// restrictions like, "must refer only to types A and B" or "UID not honored" or "name must be restricted". +// Those cannot be well described when embedded. +// 3. Inconsistent validation. Because the usages are different, the validation rules are different by usage, which makes it hard for users to predict what will happen. +// 4. The fields are both imprecise and overly precise. Kind is not a precise mapping to a URL. This can produce ambiguity +// during interpretation and require a REST mapping. In most cases, the dependency is on the group,resource tuple +// and the version of the actual struct is irrelevant. +// 5. We cannot easily change it. Because this type is embedded in many locations, updates to this type +// will affect numerous schemas. Don't make new APIs embed an underspecified API type they do not control. +// Instead of using this type, create a locally provided and used type that is well-focused on your reference. +// For example, ServiceReferences for admission registration: https://github.com/kubernetes/api/blob/release-1.17/admissionregistration/v1/types.go#L533 . // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object type ObjectReference struct { // Kind of the referent. @@ -5147,18 +5348,9 @@ type EventSeries struct { Count int32 `json:"count,omitempty" protobuf:"varint,1,name=count"` // Time of the last occurrence observed LastObservedTime metav1.MicroTime `json:"lastObservedTime,omitempty" protobuf:"bytes,2,name=lastObservedTime"` - // State of this Series: Ongoing or Finished - // Deprecated. Planned removal for 1.18 - State EventSeriesState `json:"state,omitempty" protobuf:"bytes,3,name=state"` -} -type EventSeriesState string - -const ( - EventSeriesStateOngoing EventSeriesState = "Ongoing" - EventSeriesStateFinished EventSeriesState = "Finished" - EventSeriesStateUnknown EventSeriesState = "Unknown" -) + // +k8s:deprecated=state,protobuf=3 +} // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object @@ -5194,8 +5386,7 @@ const ( // LimitRangeItem defines a min/max usage limit for any resource that matches on kind. type LimitRangeItem struct { // Type of resource that this limit applies to. - // +optional - Type LimitType `json:"type,omitempty" protobuf:"bytes,1,opt,name=type,casttype=LimitType"` + Type LimitType `json:"type" protobuf:"bytes,1,opt,name=type,casttype=LimitType"` // Max usage constraints on this kind by resource name. // +optional Max ResourceList `json:"max,omitempty" protobuf:"bytes,2,rep,name=max,casttype=ResourceList,castkey=ResourceName"` @@ -5424,6 +5615,14 @@ type Secret struct { // +optional metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + // Immutable, if set to true, ensures that data stored in the Secret cannot + // be updated (only object metadata can be modified). + // If not set to true, the field can be modified at any time. + // Defaulted to nil. + // This is a beta field enabled by ImmutableEphemeralVolumes feature gate. + // +optional + Immutable *bool `json:"immutable,omitempty" protobuf:"varint,5,opt,name=immutable"` + // Data contains the secret data. Each key must consist of alphanumeric // characters, '-', '_' or '.'. The serialized form of the secret data is a // base64 encoded string, representing the arbitrary (possibly non-string) @@ -5557,6 +5756,14 @@ type ConfigMap struct { // +optional metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + // Immutable, if set to true, ensures that data stored in the ConfigMap cannot + // be updated (only object metadata can be modified). + // If not set to true, the field can be modified at any time. + // Defaulted to nil. + // This is a beta field enabled by ImmutableEphemeralVolumes feature gate. + // +optional + Immutable *bool `json:"immutable,omitempty" protobuf:"varint,4,opt,name=immutable"` + // Data contains the configuration data. // Each key must consist of alphanumeric characters, '-', '_' or '.'. // Values with non-UTF-8 byte sequences must use the BinaryData field. @@ -5621,6 +5828,7 @@ type ComponentCondition struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // ComponentStatus (and ComponentStatusList) holds the cluster validation info. +// Deprecated: This API is deprecated in v1.19+ type ComponentStatus struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -5638,6 +5846,7 @@ type ComponentStatus struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // Status of all the conditions for the component as a list of ComponentStatus objects. +// Deprecated: This API is deprecated in v1.19+ type ComponentStatusList struct { metav1.TypeMeta `json:",inline"` // Standard list metadata. @@ -5656,7 +5865,10 @@ type DownwardAPIVolumeSource struct { // +optional Items []DownwardAPIVolumeFile `json:"items,omitempty" protobuf:"bytes,1,rep,name=items"` // Optional: mode bits to use on created files by default. Must be a - // value between 0 and 0777. Defaults to 0644. + // Optional: mode bits used to set permissions on created files by default. + // Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // Defaults to 0644. // Directories within the path are not affected by this setting. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. @@ -5679,8 +5891,10 @@ type DownwardAPIVolumeFile struct { // (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported. // +optional ResourceFieldRef *ResourceFieldSelector `json:"resourceFieldRef,omitempty" protobuf:"bytes,3,opt,name=resourceFieldRef"` - // Optional: mode bits to use on this file, must be a value between 0 - // and 0777. If not specified, the volume defaultMode will be used. + // Optional: mode bits used to set permissions on this file, must be an octal value + // between 0000 and 0777 or a decimal value between 0 and 511. + // YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. + // If not specified, the volume defaultMode will be used. // This might be in conflict with other options that affect the file // mode, like fsGroup, and the result can be other mode bits set. // +optional @@ -5758,6 +5972,11 @@ type SecurityContext struct { // This requires the ProcMountType feature flag to be enabled. // +optional ProcMount *ProcMountType `json:"procMount,omitempty" protobuf:"bytes,9,opt,name=procMount"` + // The seccomp options to use by this container. If seccomp options are + // provided at both the pod & container level, the container options + // override the pod options. + // +optional + SeccompProfile *SeccompProfile `json:"seccompProfile,omitempty" protobuf:"bytes,11,opt,name=seccompProfile"` } type ProcMountType string @@ -5793,14 +6012,12 @@ type SELinuxOptions struct { // WindowsSecurityContextOptions contain Windows-specific options and credentials. type WindowsSecurityContextOptions struct { // GMSACredentialSpecName is the name of the GMSA credential spec to use. - // This field is alpha-level and is only honored by servers that enable the WindowsGMSA feature flag. // +optional GMSACredentialSpecName *string `json:"gmsaCredentialSpecName,omitempty" protobuf:"bytes,1,opt,name=gmsaCredentialSpecName"` // GMSACredentialSpec is where the GMSA admission webhook // (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the // GMSA credential spec named by the GMSACredentialSpecName field. - // This field is alpha-level and is only honored by servers that enable the WindowsGMSA feature flag. // +optional GMSACredentialSpec *string `json:"gmsaCredentialSpec,omitempty" protobuf:"bytes,2,opt,name=gmsaCredentialSpec"` @@ -5808,7 +6025,6 @@ type WindowsSecurityContextOptions struct { // Defaults to the user specified in image metadata if unspecified. // May also be set in PodSecurityContext. If set in both SecurityContext and // PodSecurityContext, the value specified in SecurityContext takes precedence. - // This field is beta-level and may be disabled with the WindowsRunAsUserName feature flag. // +optional RunAsUserName *string `json:"runAsUserName,omitempty" protobuf:"bytes,3,opt,name=runAsUserName"` } @@ -5849,7 +6065,7 @@ type Sysctl struct { } // NodeResources is an object for conveying resource information about a node. -// see http://releases.k8s.io/HEAD/docs/design/resources.md for more details. +// see https://kubernetes.io/docs/concepts/architecture/nodes/#capacity for more details. type NodeResources struct { // Capacity represents the available resources of a node Capacity ResourceList `protobuf:"bytes,1,rep,name=capacity,casttype=ResourceList,castkey=ResourceName"` diff --git a/vendor/k8s.io/api/core/v1/types_swagger_doc_generated.go b/vendor/k8s.io/api/core/v1/types_swagger_doc_generated.go index 441d3e10..61832b81 100644 --- a/vendor/k8s.io/api/core/v1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/core/v1/types_swagger_doc_generated.go @@ -230,7 +230,7 @@ func (ComponentCondition) SwaggerDoc() map[string]string { } var map_ComponentStatus = map[string]string{ - "": "ComponentStatus (and ComponentStatusList) holds the cluster validation info.", + "": "ComponentStatus (and ComponentStatusList) holds the cluster validation info. Deprecated: This API is deprecated in v1.19+", "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", "conditions": "List of component conditions observed", } @@ -240,7 +240,7 @@ func (ComponentStatus) SwaggerDoc() map[string]string { } var map_ComponentStatusList = map[string]string{ - "": "Status of all the conditions for the component as a list of ComponentStatus objects.", + "": "Status of all the conditions for the component as a list of ComponentStatus objects. Deprecated: This API is deprecated in v1.19+", "metadata": "Standard list metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds", "items": "List of ComponentStatus objects.", } @@ -252,6 +252,7 @@ func (ComponentStatusList) SwaggerDoc() map[string]string { var map_ConfigMap = map[string]string{ "": "ConfigMap holds configuration data for pods to consume.", "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "immutable": "Immutable, if set to true, ensures that data stored in the ConfigMap cannot be updated (only object metadata can be modified). If not set to true, the field can be modified at any time. Defaulted to nil. This is a beta field enabled by ImmutableEphemeralVolumes feature gate.", "data": "Data contains the configuration data. Each key must consist of alphanumeric characters, '-', '_' or '.'. Values with non-UTF-8 byte sequences must use the BinaryData field. The keys stored in Data must not overlap with the keys in the BinaryData field, this is enforced during validation process.", "binaryData": "BinaryData contains the binary data. Each key must consist of alphanumeric characters, '-', '_' or '.'. BinaryData can contain byte sequences that are not in the UTF-8 range. The keys stored in BinaryData must not overlap with the ones in the Data field, this is enforced during validation process. Using this field will require 1.10+ apiserver and kubelet.", } @@ -315,7 +316,7 @@ func (ConfigMapProjection) SwaggerDoc() map[string]string { var map_ConfigMapVolumeSource = map[string]string{ "": "Adapts a ConfigMap into a volume.\n\nThe contents of the target ConfigMap's Data field will be presented in a volume as files using the keys in the Data field as the file names, unless the items element is populated with specific mappings of keys to paths. ConfigMap volumes support ownership management and SELinux relabeling.", "items": "If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", - "defaultMode": "Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", + "defaultMode": "Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", "optional": "Specify whether the ConfigMap or its keys must be defined", } @@ -335,10 +336,10 @@ var map_Container = map[string]string{ "env": "List of environment variables to set in the container. Cannot be updated.", "resources": "Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/", "volumeMounts": "Pod volumes to mount into the container's filesystem. Cannot be updated.", - "volumeDevices": "volumeDevices is the list of block devices to be used by the container. This is a beta feature.", + "volumeDevices": "volumeDevices is the list of block devices to be used by the container.", "livenessProbe": "Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes", "readinessProbe": "Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes", - "startupProbe": "StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. This is an alpha feature enabled by the StartupProbe feature flag. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes", + "startupProbe": "StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. This is a beta feature enabled by the StartupProbe feature flag. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes", "lifecycle": "Actions that the management system should take in response to container lifecycle events. Cannot be updated.", "terminationMessagePath": "Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.", "terminationMessagePolicy": "Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated.", @@ -461,7 +462,7 @@ var map_DownwardAPIVolumeFile = map[string]string{ "path": "Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'", "fieldRef": "Required: Selects a field of the pod: only annotations, labels, name and namespace are supported.", "resourceFieldRef": "Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported.", - "mode": "Optional: mode bits to use on this file, must be a value between 0 and 0777. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", + "mode": "Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", } func (DownwardAPIVolumeFile) SwaggerDoc() map[string]string { @@ -471,7 +472,7 @@ func (DownwardAPIVolumeFile) SwaggerDoc() map[string]string { var map_DownwardAPIVolumeSource = map[string]string{ "": "DownwardAPIVolumeSource represents a volume containing downward API info. Downward API volumes support ownership management and SELinux relabeling.", "items": "Items is a list of downward API volume file", - "defaultMode": "Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", + "defaultMode": "Optional: mode bits to use on created files by default. Must be a Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", } func (DownwardAPIVolumeSource) SwaggerDoc() map[string]string { @@ -501,10 +502,11 @@ func (EndpointAddress) SwaggerDoc() map[string]string { } var map_EndpointPort = map[string]string{ - "": "EndpointPort is a tuple that describes a single port.", - "name": "The name of this port. This must match the 'name' field in the corresponding ServicePort. Must be a DNS_LABEL. Optional only if one port is defined.", - "port": "The port number of the endpoint.", - "protocol": "The IP protocol for this port. Must be UDP, TCP, or SCTP. Default is TCP.", + "": "EndpointPort is a tuple that describes a single port.", + "name": "The name of this port. This must match the 'name' field in the corresponding ServicePort. Must be a DNS_LABEL. Optional only if one port is defined.", + "port": "The port number of the endpoint.", + "protocol": "The IP protocol for this port. Must be UDP, TCP, or SCTP. Default is TCP.", + "appProtocol": "The application protocol for this port. This field follows standard Kubernetes label syntax. Un-prefixed names are reserved for IANA standard service names (as per RFC-6335 and http://www.iana.org/assignments/service-names). Non-standard protocols should use prefixed names such as mycompany.com/my-custom-protocol. This is a beta field that is guarded by the ServiceAppProtocol feature gate and enabled by default.", } func (EndpointPort) SwaggerDoc() map[string]string { @@ -566,7 +568,7 @@ func (EnvVar) SwaggerDoc() map[string]string { var map_EnvVarSource = map[string]string{ "": "EnvVarSource represents a source for the value of an EnvVar.", - "fieldRef": "Selects a field of the pod: supports metadata.name, metadata.namespace, metadata.labels, metadata.annotations, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.", + "fieldRef": "Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['']`, `metadata.annotations['']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.", "resourceFieldRef": "Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported.", "configMapKeyRef": "Selects a key of a ConfigMap.", "secretKeyRef": "Selects a key of a secret in the pod's namespace", @@ -597,7 +599,7 @@ var map_EphemeralContainerCommon = map[string]string{ "env": "List of environment variables to set in the container. Cannot be updated.", "resources": "Resources are not allowed for ephemeral containers. Ephemeral containers use spare resources already allocated to the pod.", "volumeMounts": "Pod volumes to mount into the container's filesystem. Cannot be updated.", - "volumeDevices": "volumeDevices is the list of block devices to be used by the container. This is a beta feature.", + "volumeDevices": "volumeDevices is the list of block devices to be used by the container.", "livenessProbe": "Probes are not allowed for ephemeral containers.", "readinessProbe": "Probes are not allowed for ephemeral containers.", "startupProbe": "Probes are not allowed for ephemeral containers.", @@ -624,6 +626,16 @@ func (EphemeralContainers) SwaggerDoc() map[string]string { return map_EphemeralContainers } +var map_EphemeralVolumeSource = map[string]string{ + "": "Represents an ephemeral volume that is handled by a normal storage driver.", + "volumeClaimTemplate": "Will be used to create a stand-alone PVC to provision the volume. The pod in which this EphemeralVolumeSource is embedded will be the owner of the PVC, i.e. the PVC will be deleted together with the pod. The name of the PVC will be `-` where `` is the name from the `PodSpec.Volumes` array entry. Pod validation will reject the pod if the concatenated name is not valid for a PVC (for example, too long).\n\nAn existing PVC with that name that is not owned by the pod will *not* be used for the pod to avoid using an unrelated volume by mistake. Starting the pod is then blocked until the unrelated PVC is removed. If such a pre-created PVC is meant to be used by the pod, the PVC has to updated with an owner reference to the pod once the pod exists. Normally this should not be necessary, but it may be useful when manually reconstructing a broken cluster.\n\nThis field is read-only and no changes will be made by Kubernetes to the PVC after it has been created.\n\nRequired, must not be nil.", + "readOnly": "Specifies a read-only configuration for the volume. Defaults to false (read/write).", +} + +func (EphemeralVolumeSource) SwaggerDoc() map[string]string { + return map_EphemeralVolumeSource +} + var map_Event = map[string]string{ "": "Event is a report of an event somewhere in the cluster.", "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", @@ -661,7 +673,6 @@ var map_EventSeries = map[string]string{ "": "EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time.", "count": "Number of occurrences in this series up to the last heartbeat time", "lastObservedTime": "Time of the last occurrence observed", - "state": "State of this Series: Ongoing or Finished Deprecated. Planned removal for 1.18", } func (EventSeries) SwaggerDoc() map[string]string { @@ -878,7 +889,7 @@ var map_KeyToPath = map[string]string{ "": "Maps a string key to a path within a volume.", "key": "The key to project.", "path": "The relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.", - "mode": "Optional: mode bits to use on this file, must be a value between 0 and 0777. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", + "mode": "Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", } func (KeyToPath) SwaggerDoc() map[string]string { @@ -1132,7 +1143,7 @@ func (NodeProxyOptions) SwaggerDoc() map[string]string { } var map_NodeResources = map[string]string{ - "": "NodeResources is an object for conveying resource information about a node. see http://releases.k8s.io/HEAD/docs/design/resources.md for more details.", + "": "NodeResources is an object for conveying resource information about a node. see https://kubernetes.io/docs/concepts/architecture/nodes/#capacity for more details.", "Capacity": "Capacity represents the available resources of a node", } @@ -1207,7 +1218,7 @@ func (NodeStatus) SwaggerDoc() map[string]string { var map_NodeSystemInfo = map[string]string{ "": "NodeSystemInfo is a set of ids/uuids to uniquely identify the node.", "machineID": "MachineID reported by the node. For unique machine identification in the cluster this field is preferred. Learn more from man(5) machine-id: http://man7.org/linux/man-pages/man5/machine-id.5.html", - "systemUUID": "SystemUUID reported by the node. For unique machine identification MachineID is preferred. This field is specific to Red Hat hosts https://access.redhat.com/documentation/en-US/Red_Hat_Subscription_Management/1/html/RHSM/getting-system-uuid.html", + "systemUUID": "SystemUUID reported by the node. For unique machine identification MachineID is preferred. This field is specific to Red Hat hosts https://access.redhat.com/documentation/en-us/red_hat_subscription_management/1/html/rhsm/uuid", "bootID": "Boot ID reported by the node.", "kernelVersion": "Kernel Version reported by the node from 'uname -r' (e.g. 3.16.0-0.bpo.4-amd64).", "osImage": "OS Image reported by the node from /etc/os-release (e.g. Debian GNU/Linux 7 (wheezy)).", @@ -1298,8 +1309,8 @@ var map_PersistentVolumeClaimSpec = map[string]string{ "resources": "Resources represents the minimum resources the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources", "volumeName": "VolumeName is the binding reference to the PersistentVolume backing this claim.", "storageClassName": "Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1", - "volumeMode": "volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec. This is a beta feature.", - "dataSource": "This field requires the VolumeSnapshotDataSource alpha feature gate to be enabled and currently VolumeSnapshot is the only supported data source. If the provisioner can support VolumeSnapshot data source, it will create a new volume and data will be restored to the volume at the same time. If the provisioner does not support VolumeSnapshot data source, volume will not be created and the failure will be reported as an event. In the future, we plan to support more data source types and the behavior of the provisioner may change.", + "volumeMode": "volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec.", + "dataSource": "This field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot - Beta) * An existing PVC (PersistentVolumeClaim) * An existing custom resource/object that implements data population (Alpha) In order to use VolumeSnapshot object types, the appropriate feature gate must be enabled (VolumeSnapshotDataSource or AnyVolumeDataSource) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. If the specified data source is not supported, the volume will not be created and the failure will be reported as an event. In the future, we plan to support more data source types and the behavior of the provisioner may change.", } func (PersistentVolumeClaimSpec) SwaggerDoc() map[string]string { @@ -1318,6 +1329,16 @@ func (PersistentVolumeClaimStatus) SwaggerDoc() map[string]string { return map_PersistentVolumeClaimStatus } +var map_PersistentVolumeClaimTemplate = map[string]string{ + "": "PersistentVolumeClaimTemplate is used to produce PersistentVolumeClaim objects as part of an EphemeralVolumeSource.", + "metadata": "May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation.", + "spec": "The specification for the PersistentVolumeClaim. The entire content is copied unchanged into the PVC that gets created from this template. The same fields as in a PersistentVolumeClaim are also valid here.", +} + +func (PersistentVolumeClaimTemplate) SwaggerDoc() map[string]string { + return map_PersistentVolumeClaimTemplate +} + var map_PersistentVolumeClaimVolumeSource = map[string]string{ "": "PersistentVolumeClaimVolumeSource references the user's PVC in the same namespace. This volume finds the bound PV and mounts that volume for the pod. A PersistentVolumeClaimVolumeSource is, essentially, a wrapper around another type of volume that is owned by someone else (the system).", "claimName": "ClaimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims", @@ -1376,7 +1397,7 @@ var map_PersistentVolumeSpec = map[string]string{ "persistentVolumeReclaimPolicy": "What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming", "storageClassName": "Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.", "mountOptions": "A list of mount options, e.g. [\"ro\", \"soft\"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options", - "volumeMode": "volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec. This is a beta feature.", + "volumeMode": "volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.", "nodeAffinity": "NodeAffinity defines constraints that limit what nodes this volume can be accessed from. This field influences the scheduling of pods that use this volume.", } @@ -1572,15 +1593,17 @@ func (PodReadinessGate) SwaggerDoc() map[string]string { } var map_PodSecurityContext = map[string]string{ - "": "PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.", - "seLinuxOptions": "The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.", - "windowsOptions": "The Windows specific settings applied to all containers. If unspecified, the options within a container's SecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.", - "runAsUser": "The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.", - "runAsGroup": "The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.", - "runAsNonRoot": "Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.", - "supplementalGroups": "A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", - "fsGroup": "A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw ", - "sysctls": "Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.", + "": "PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.", + "seLinuxOptions": "The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.", + "windowsOptions": "The Windows specific settings applied to all containers. If unspecified, the options within a container's SecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.", + "runAsUser": "The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.", + "runAsGroup": "The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.", + "runAsNonRoot": "Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.", + "supplementalGroups": "A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", + "fsGroup": "A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw ", + "sysctls": "Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.", + "fsGroupChangePolicy": "fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are \"OnRootMismatch\" and \"Always\". If not specified defaults to \"Always\".", + "seccompProfile": "The seccomp options to use by the containers in this pod.", } func (PodSecurityContext) SwaggerDoc() map[string]string { @@ -1629,9 +1652,10 @@ var map_PodSpec = map[string]string{ "readinessGates": "If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to \"True\" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md", "runtimeClassName": "RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the \"legacy\" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.", "enableServiceLinks": "EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", - "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature.", + "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.", "overhead": "Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.", - "topologySpreadConstraints": "TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. This field is alpha-level and is only honored by clusters that enables the EvenPodsSpread feature. All topologySpreadConstraints are ANDed.", + "topologySpreadConstraints": "TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.", + "setHostnameAsFQDN": "If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", } func (PodSpec) SwaggerDoc() map[string]string { @@ -1757,7 +1781,7 @@ func (Probe) SwaggerDoc() map[string]string { var map_ProjectedVolumeSource = map[string]string{ "": "Represents a projected volume source", "sources": "list of volume projections", - "defaultMode": "Mode bits to use on created files by default. Must be a value between 0 and 0777. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", + "defaultMode": "Mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", } func (ProjectedVolumeSource) SwaggerDoc() map[string]string { @@ -2012,9 +2036,20 @@ func (ScopedResourceSelectorRequirement) SwaggerDoc() map[string]string { return map_ScopedResourceSelectorRequirement } +var map_SeccompProfile = map[string]string{ + "": "SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set.", + "type": "type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.", + "localhostProfile": "localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is \"Localhost\".", +} + +func (SeccompProfile) SwaggerDoc() map[string]string { + return map_SeccompProfile +} + var map_Secret = map[string]string{ "": "Secret holds secret data of a certain type. The total bytes of the values in the Data field must be less than MaxSecretSize bytes.", "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "immutable": "Immutable, if set to true, ensures that data stored in the Secret cannot be updated (only object metadata can be modified). If not set to true, the field can be modified at any time. Defaulted to nil. This is a beta field enabled by ImmutableEphemeralVolumes feature gate.", "data": "Data contains the secret data. Each key must consist of alphanumeric characters, '-', '_' or '.'. The serialized form of the secret data is a base64 encoded string, representing the arbitrary (possibly non-string) data value here. Described in https://tools.ietf.org/html/rfc4648#section-4", "stringData": "stringData allows specifying non-binary secret data in string form. It is provided as a write-only convenience method. All keys and values are merged into the data field on write, overwriting any existing values. It is never output when reading from the API.", "type": "Used to facilitate programmatic handling of secret data.", @@ -2077,7 +2112,7 @@ var map_SecretVolumeSource = map[string]string{ "": "Adapts a Secret into a volume.\n\nThe contents of the target Secret's Data field will be presented in a volume as files using the keys in the Data field as the file names. Secret volumes support ownership management and SELinux relabeling.", "secretName": "Name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret", "items": "If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", - "defaultMode": "Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", + "defaultMode": "Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.", "optional": "Specify whether the Secret or its keys must be defined", } @@ -2097,6 +2132,7 @@ var map_SecurityContext = map[string]string{ "readOnlyRootFilesystem": "Whether this container has a read-only root filesystem. Default is false.", "allowPrivilegeEscalation": "AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN", "procMount": "procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled.", + "seccompProfile": "The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options.", } func (SecurityContext) SwaggerDoc() map[string]string { @@ -2167,12 +2203,13 @@ func (ServiceList) SwaggerDoc() map[string]string { } var map_ServicePort = map[string]string{ - "": "ServicePort contains information on service's port.", - "name": "The name of this port within the service. This must be a DNS_LABEL. All ports within a ServiceSpec must have unique names. When considering the endpoints for a Service, this must match the 'name' field in the EndpointPort. Optional if only one ServicePort is defined on this service.", - "protocol": "The IP protocol for this port. Supports \"TCP\", \"UDP\", and \"SCTP\". Default is TCP.", - "port": "The port that will be exposed by this service.", - "targetPort": "Number or name of the port to access on the pods targeted by the service. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. If this is a string, it will be looked up as a named port in the target Pod's container ports. If this is not specified, the value of the 'port' field is used (an identity map). This field is ignored for services with clusterIP=None, and should be omitted or set equal to the 'port' field. More info: https://kubernetes.io/docs/concepts/services-networking/service/#defining-a-service", - "nodePort": "The port on each node on which this service is exposed when type=NodePort or LoadBalancer. Usually assigned by the system. If specified, it will be allocated to the service if unused or else creation of the service will fail. Default is to auto-allocate a port if the ServiceType of this Service requires one. More info: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport", + "": "ServicePort contains information on service's port.", + "name": "The name of this port within the service. This must be a DNS_LABEL. All ports within a ServiceSpec must have unique names. When considering the endpoints for a Service, this must match the 'name' field in the EndpointPort. Optional if only one ServicePort is defined on this service.", + "protocol": "The IP protocol for this port. Supports \"TCP\", \"UDP\", and \"SCTP\". Default is TCP.", + "appProtocol": "The application protocol for this port. This field follows standard Kubernetes label syntax. Un-prefixed names are reserved for IANA standard service names (as per RFC-6335 and http://www.iana.org/assignments/service-names). Non-standard protocols should use prefixed names such as mycompany.com/my-custom-protocol. This is a beta field that is guarded by the ServiceAppProtocol feature gate and enabled by default.", + "port": "The port that will be exposed by this service.", + "targetPort": "Number or name of the port to access on the pods targeted by the service. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. If this is a string, it will be looked up as a named port in the target Pod's container ports. If this is not specified, the value of the 'port' field is used (an identity map). This field is ignored for services with clusterIP=None, and should be omitted or set equal to the 'port' field. More info: https://kubernetes.io/docs/concepts/services-networking/service/#defining-a-service", + "nodePort": "The port on each node on which this service is exposed when type=NodePort or LoadBalancer. Usually assigned by the system. If specified, it will be allocated to the service if unused or else creation of the service will fail. Default is to auto-allocate a port if the ServiceType of this Service requires one. More info: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport", } func (ServicePort) SwaggerDoc() map[string]string { @@ -2201,9 +2238,9 @@ var map_ServiceSpec = map[string]string{ "externalName": "externalName is the external reference that kubedns or equivalent will return as a CNAME record for this service. No proxying will be involved. Must be a valid RFC-1123 hostname (https://tools.ietf.org/html/rfc1123) and requires Type to be ExternalName.", "externalTrafficPolicy": "externalTrafficPolicy denotes if this Service desires to route external traffic to node-local or cluster-wide endpoints. \"Local\" preserves the client source IP and avoids a second hop for LoadBalancer and Nodeport type services, but risks potentially imbalanced traffic spreading. \"Cluster\" obscures the client source IP and may cause a second hop to another node, but should have good overall load-spreading.", "healthCheckNodePort": "healthCheckNodePort specifies the healthcheck nodePort for the service. If not specified, HealthCheckNodePort is created by the service api backend with the allocated nodePort. Will use user-specified nodePort value if specified by the client. Only effects when Type is set to LoadBalancer and ExternalTrafficPolicy is set to Local.", - "publishNotReadyAddresses": "publishNotReadyAddresses, when set to true, indicates that DNS implementations must publish the notReadyAddresses of subsets for the Endpoints associated with the Service. The default value is false. The primary use case for setting this field is to use a StatefulSet's Headless Service to propagate SRV records for its Pods without respect to their readiness for purpose of peer discovery.", + "publishNotReadyAddresses": "publishNotReadyAddresses indicates that any agent which deals with endpoints for this Service should disregard any indications of ready/not-ready. The primary use case for setting this field is for a StatefulSet's Headless Service to propagate SRV DNS records for its Pods for the purpose of peer discovery. The Kubernetes controllers that generate Endpoints and EndpointSlice resources for Services interpret this to mean that all endpoints are considered \"ready\" even if the Pods themselves are not. Agents which consume only Kubernetes generated endpoints through the Endpoints or EndpointSlice resources can safely assume this behavior.", "sessionAffinityConfig": "sessionAffinityConfig contains the configurations of session affinity.", - "ipFamily": "ipFamily specifies whether this Service has a preference for a particular IP family (e.g. IPv4 vs. IPv6). If a specific IP family is requested, the clusterIP field will be allocated from that family, if it is available in the cluster. If no IP family is requested, the cluster's primary IP family will be used. Other IP fields (loadBalancerIP, loadBalancerSourceRanges, externalIPs) and controllers which allocate external load-balancers should use the same IP family. Endpoints for this Service will be of this family. This field is immutable after creation. Assigning a ServiceIPFamily not available in the cluster (e.g. IPv6 in IPv4 only cluster) is an error condition and will fail during clusterIP assignment.", + "ipFamily": "ipFamily specifies whether this Service has a preference for a particular IP family (e.g. IPv4 vs. IPv6) when the IPv6DualStack feature gate is enabled. In a dual-stack cluster, you can specify ipFamily when creating a ClusterIP Service to determine whether the controller will allocate an IPv4 or IPv6 IP for it, and you can specify ipFamily when creating a headless Service to determine whether it will have IPv4 or IPv6 Endpoints. In either case, if you do not specify an ipFamily explicitly, it will default to the cluster's primary IP family. This field is part of an alpha feature, and you should not make any assumptions about its semantics other than those described above. In particular, you should not assume that it can (or cannot) be changed after creation time; that it can only have the values \"IPv4\" and \"IPv6\"; or that its current value on a given Service correctly reflects the current state of that Service. (For ClusterIP Services, look at clusterIP to see if the Service is IPv4 or IPv6. For headless Services, look at the endpoints, which may be dual-stack in the future. For ExternalName Services, ipFamily has no meaning, but it may be set to an irrelevant value anyway.)", "topologyKeys": "topologyKeys is a preference-order list of topology keys which implementations of services should use to preferentially sort endpoints when accessing this Service, it can not be used at the same time as externalTrafficPolicy=Local. Topology keys must be valid label keys and at most 16 keys may be specified. Endpoints are chosen based on the first topology key with available backends. If this field is specified and all entries have no backends that match the topology of the client, the service has no backends for that client and connections should fail. The special value \"*\" may be used to mean \"any topology\". This catch-all value, if used, only makes sense as the last value in the list. If this is not specified or empty, no topology constraints will be applied.", } @@ -2278,7 +2315,7 @@ func (TCPSocketAction) SwaggerDoc() map[string]string { var map_Taint = map[string]string{ "": "The node this Taint is attached to has the \"effect\" on any pod that does not tolerate the Taint.", "key": "Required. The taint key to be applied to a node.", - "value": "Required. The taint value corresponding to the taint key.", + "value": "The taint value corresponding to the taint key.", "effect": "Required. The effect of the taint on pods that do not tolerate the taint. Valid effects are NoSchedule, PreferNoSchedule and NoExecute.", "timeAdded": "TimeAdded represents the time at which the taint was added. It is only written for NoExecute taints.", } @@ -2321,9 +2358,9 @@ func (TopologySelectorTerm) SwaggerDoc() map[string]string { var map_TopologySpreadConstraint = map[string]string{ "": "TopologySpreadConstraint specifies how to spread matching pods among the given topology.", - "maxSkew": "MaxSkew describes the degree to which pods may be unevenly distributed. It's the maximum permitted difference between the number of matching pods in any two topology domains of a given topology type. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 1/1/0: ", + "maxSkew": "MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 1/1/0: ", "topologyKey": "TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each as a \"bucket\", and try to put balanced number of pods into each bucket. It's a required field.", - "whenUnsatisfiable": "WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it - ScheduleAnyway tells the scheduler to still schedule it It's considered as \"Unsatisfiable\" if and only if placing incoming pod on any topology violates \"MaxSkew\". For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: ", + "whenUnsatisfiable": "WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location,\n but giving higher precedence to topologies that would help reduce the\n skew.\nA constraint is considered \"Unsatisfiable\" for an incoming pod if and only if every possible node assigment for that pod would violate \"MaxSkew\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: ", "labelSelector": "LabelSelector is used to find matching pods. Pods that match this label selector are counted to determine the number of pods in their corresponding topology domain.", } @@ -2425,7 +2462,8 @@ var map_VolumeSource = map[string]string{ "portworxVolume": "PortworxVolume represents a portworx volume attached and mounted on kubelets host machine", "scaleIO": "ScaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes.", "storageos": "StorageOS represents a StorageOS volume attached and mounted on Kubernetes nodes.", - "csi": "CSI (Container Storage Interface) represents storage that is handled by an external CSI driver (Alpha feature).", + "csi": "CSI (Container Storage Interface) represents ephemeral storage that is handled by certain external CSI drivers (Beta feature).", + "ephemeral": "Ephemeral represents a volume that is handled by a cluster storage driver (Alpha feature). The volume's lifecycle is tied to the pod that defines it - it will be created before the pod starts, and deleted when the pod is removed.\n\nUse this if: a) the volume is only needed while the pod runs, b) features of normal volumes like restoring from snapshot or capacity\n tracking are needed,\nc) the storage driver is specified through a storage class, and d) the storage driver supports dynamic volume provisioning through\n a PersistentVolumeClaim (see EphemeralVolumeSource for more\n information on the connection between this volume type\n and PersistentVolumeClaim).\n\nUse PersistentVolumeClaim or one of the vendor-specific APIs for volumes that persist for longer than the lifecycle of an individual pod.\n\nUse CSI for light-weight local ephemeral volumes if the CSI driver is meant to be used that way - see the documentation of the driver for more information.\n\nA pod can use both types of ephemeral volumes and persistent volumes at the same time.", } func (VolumeSource) SwaggerDoc() map[string]string { @@ -2456,9 +2494,9 @@ func (WeightedPodAffinityTerm) SwaggerDoc() map[string]string { var map_WindowsSecurityContextOptions = map[string]string{ "": "WindowsSecurityContextOptions contain Windows-specific options and credentials.", - "gmsaCredentialSpecName": "GMSACredentialSpecName is the name of the GMSA credential spec to use. This field is alpha-level and is only honored by servers that enable the WindowsGMSA feature flag.", - "gmsaCredentialSpec": "GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field. This field is alpha-level and is only honored by servers that enable the WindowsGMSA feature flag.", - "runAsUserName": "The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. This field is beta-level and may be disabled with the WindowsRunAsUserName feature flag.", + "gmsaCredentialSpecName": "GMSACredentialSpecName is the name of the GMSA credential spec to use.", + "gmsaCredentialSpec": "GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.", + "runAsUserName": "The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.", } func (WindowsSecurityContextOptions) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/core/v1/well_known_taints.go b/vendor/k8s.io/api/core/v1/well_known_taints.go index e3905192..e1a8f629 100644 --- a/vendor/k8s.io/api/core/v1/well_known_taints.go +++ b/vendor/k8s.io/api/core/v1/well_known_taints.go @@ -18,38 +18,31 @@ package v1 const ( // TaintNodeNotReady will be added when node is not ready - // and feature-gate for TaintBasedEvictions flag is enabled, // and removed when node becomes ready. TaintNodeNotReady = "node.kubernetes.io/not-ready" // TaintNodeUnreachable will be added when node becomes unreachable // (corresponding to NodeReady status ConditionUnknown) - // and feature-gate for TaintBasedEvictions flag is enabled, // and removed when node becomes reachable (NodeReady status ConditionTrue). TaintNodeUnreachable = "node.kubernetes.io/unreachable" // TaintNodeUnschedulable will be added when node becomes unschedulable - // and feature-gate for TaintNodesByCondition flag is enabled, // and removed when node becomes scheduable. TaintNodeUnschedulable = "node.kubernetes.io/unschedulable" // TaintNodeMemoryPressure will be added when node has memory pressure - // and feature-gate for TaintNodesByCondition flag is enabled, // and removed when node has enough memory. TaintNodeMemoryPressure = "node.kubernetes.io/memory-pressure" // TaintNodeDiskPressure will be added when node has disk pressure - // and feature-gate for TaintNodesByCondition flag is enabled, // and removed when node has enough disk. TaintNodeDiskPressure = "node.kubernetes.io/disk-pressure" // TaintNodeNetworkUnavailable will be added when node's network is unavailable - // and feature-gate for TaintNodesByCondition flag is enabled, // and removed when network becomes ready. TaintNodeNetworkUnavailable = "node.kubernetes.io/network-unavailable" // TaintNodePIDPressure will be added when node has pid pressure - // and feature-gate for TaintNodesByCondition flag is enabled, // and removed when node has enough disk. TaintNodePIDPressure = "node.kubernetes.io/pid-pressure" ) diff --git a/vendor/k8s.io/api/core/v1/zz_generated.deepcopy.go b/vendor/k8s.io/api/core/v1/zz_generated.deepcopy.go index ac4855ab..445c7c04 100644 --- a/vendor/k8s.io/api/core/v1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/core/v1/zz_generated.deepcopy.go @@ -519,6 +519,11 @@ func (in *ConfigMap) DeepCopyInto(out *ConfigMap) { *out = *in out.TypeMeta = in.TypeMeta in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + if in.Immutable != nil { + in, out := &in.Immutable, &out.Immutable + *out = new(bool) + **out = **in + } if in.Data != nil { in, out := &in.Data, &out.Data *out = make(map[string]string, len(*in)) @@ -1091,6 +1096,11 @@ func (in *EndpointAddress) DeepCopy() *EndpointAddress { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *EndpointPort) DeepCopyInto(out *EndpointPort) { *out = *in + if in.AppProtocol != nil { + in, out := &in.AppProtocol, &out.AppProtocol + *out = new(string) + **out = **in + } return } @@ -1124,7 +1134,9 @@ func (in *EndpointSubset) DeepCopyInto(out *EndpointSubset) { if in.Ports != nil { in, out := &in.Ports, &out.Ports *out = make([]EndpointPort, len(*in)) - copy(*out, *in) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } } return } @@ -1421,6 +1433,27 @@ func (in *EphemeralContainers) DeepCopyObject() runtime.Object { return nil } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EphemeralVolumeSource) DeepCopyInto(out *EphemeralVolumeSource) { + *out = *in + if in.VolumeClaimTemplate != nil { + in, out := &in.VolumeClaimTemplate, &out.VolumeClaimTemplate + *out = new(PersistentVolumeClaimTemplate) + (*in).DeepCopyInto(*out) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EphemeralVolumeSource. +func (in *EphemeralVolumeSource) DeepCopy() *EphemeralVolumeSource { + if in == nil { + return nil + } + out := new(EphemeralVolumeSource) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *Event) DeepCopyInto(out *Event) { *out = *in @@ -2973,6 +3006,24 @@ func (in *PersistentVolumeClaimStatus) DeepCopy() *PersistentVolumeClaimStatus { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *PersistentVolumeClaimTemplate) DeepCopyInto(out *PersistentVolumeClaimTemplate) { + *out = *in + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + in.Spec.DeepCopyInto(&out.Spec) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PersistentVolumeClaimTemplate. +func (in *PersistentVolumeClaimTemplate) DeepCopy() *PersistentVolumeClaimTemplate { + if in == nil { + return nil + } + out := new(PersistentVolumeClaimTemplate) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *PersistentVolumeClaimVolumeSource) DeepCopyInto(out *PersistentVolumeClaimVolumeSource) { *out = *in @@ -3677,6 +3728,16 @@ func (in *PodSecurityContext) DeepCopyInto(out *PodSecurityContext) { *out = make([]Sysctl, len(*in)) copy(*out, *in) } + if in.FSGroupChangePolicy != nil { + in, out := &in.FSGroupChangePolicy, &out.FSGroupChangePolicy + *out = new(PodFSGroupChangePolicy) + **out = **in + } + if in.SeccompProfile != nil { + in, out := &in.SeccompProfile, &out.SeccompProfile + *out = new(SeccompProfile) + (*in).DeepCopyInto(*out) + } return } @@ -3842,6 +3903,11 @@ func (in *PodSpec) DeepCopyInto(out *PodSpec) { (*in)[i].DeepCopyInto(&(*out)[i]) } } + if in.SetHostnameAsFQDN != nil { + in, out := &in.SetHostnameAsFQDN, &out.SetHostnameAsFQDN + *out = new(bool) + **out = **in + } return } @@ -4658,11 +4724,37 @@ func (in *ScopedResourceSelectorRequirement) DeepCopy() *ScopedResourceSelectorR return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *SeccompProfile) DeepCopyInto(out *SeccompProfile) { + *out = *in + if in.LocalhostProfile != nil { + in, out := &in.LocalhostProfile, &out.LocalhostProfile + *out = new(string) + **out = **in + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new SeccompProfile. +func (in *SeccompProfile) DeepCopy() *SeccompProfile { + if in == nil { + return nil + } + out := new(SeccompProfile) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *Secret) DeepCopyInto(out *Secret) { *out = *in out.TypeMeta = in.TypeMeta in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + if in.Immutable != nil { + in, out := &in.Immutable, &out.Immutable + *out = new(bool) + **out = **in + } if in.Data != nil { in, out := &in.Data, &out.Data *out = make(map[string][]byte, len(*in)) @@ -4914,6 +5006,11 @@ func (in *SecurityContext) DeepCopyInto(out *SecurityContext) { *out = new(ProcMountType) **out = **in } + if in.SeccompProfile != nil { + in, out := &in.SeccompProfile, &out.SeccompProfile + *out = new(SeccompProfile) + (*in).DeepCopyInto(*out) + } return } @@ -5112,6 +5209,11 @@ func (in *ServiceList) DeepCopyObject() runtime.Object { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *ServicePort) DeepCopyInto(out *ServicePort) { *out = *in + if in.AppProtocol != nil { + in, out := &in.AppProtocol, &out.AppProtocol + *out = new(string) + **out = **in + } out.TargetPort = in.TargetPort return } @@ -5157,7 +5259,9 @@ func (in *ServiceSpec) DeepCopyInto(out *ServiceSpec) { if in.Ports != nil { in, out := &in.Ports, &out.Ports *out = make([]ServicePort, len(*in)) - copy(*out, *in) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } } if in.Selector != nil { in, out := &in.Selector, &out.Selector @@ -5698,6 +5802,11 @@ func (in *VolumeSource) DeepCopyInto(out *VolumeSource) { *out = new(CSIVolumeSource) (*in).DeepCopyInto(*out) } + if in.Ephemeral != nil { + in, out := &in.Ephemeral, &out.Ephemeral + *out = new(EphemeralVolumeSource) + (*in).DeepCopyInto(*out) + } return } diff --git a/vendor/k8s.io/api/discovery/v1alpha1/generated.pb.go b/vendor/k8s.io/api/discovery/v1alpha1/generated.pb.go index fa4d3ac5..45c4382c 100644 --- a/vendor/k8s.io/api/discovery/v1alpha1/generated.pb.go +++ b/vendor/k8s.io/api/discovery/v1alpha1/generated.pb.go @@ -44,7 +44,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Endpoint) Reset() { *m = Endpoint{} } func (*Endpoint) ProtoMessage() {} @@ -1621,6 +1621,7 @@ func (m *EndpointSliceList) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1652,10 +1653,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1676,55 +1675,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/discovery/v1alpha1/generated.proto b/vendor/k8s.io/api/discovery/v1alpha1/generated.proto index 62074e7a..2cbbdcdb 100644 --- a/vendor/k8s.io/api/discovery/v1alpha1/generated.proto +++ b/vendor/k8s.io/api/discovery/v1alpha1/generated.proto @@ -151,7 +151,6 @@ message EndpointSliceList { optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; // List of endpoint slices - // +listType=set repeated EndpointSlice items = 2; } diff --git a/vendor/k8s.io/api/discovery/v1alpha1/types.go b/vendor/k8s.io/api/discovery/v1alpha1/types.go index fff30b5c..cf50b501 100644 --- a/vendor/k8s.io/api/discovery/v1alpha1/types.go +++ b/vendor/k8s.io/api/discovery/v1alpha1/types.go @@ -157,6 +157,5 @@ type EndpointSliceList struct { // +optional metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` // List of endpoint slices - // +listType=set Items []EndpointSlice `json:"items" protobuf:"bytes,2,rep,name=items"` } diff --git a/vendor/k8s.io/api/discovery/v1beta1/doc.go b/vendor/k8s.io/api/discovery/v1beta1/doc.go index 9b54d1b9..7d708480 100644 --- a/vendor/k8s.io/api/discovery/v1beta1/doc.go +++ b/vendor/k8s.io/api/discovery/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=discovery.k8s.io package v1beta1 // import "k8s.io/api/discovery/v1beta1" diff --git a/vendor/k8s.io/api/discovery/v1beta1/generated.pb.go b/vendor/k8s.io/api/discovery/v1beta1/generated.pb.go index 2283d12d..ce0046c5 100644 --- a/vendor/k8s.io/api/discovery/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/discovery/v1beta1/generated.pb.go @@ -44,7 +44,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Endpoint) Reset() { *m = Endpoint{} } func (*Endpoint) ProtoMessage() {} @@ -1621,6 +1621,7 @@ func (m *EndpointSliceList) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1652,10 +1653,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1676,55 +1675,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/discovery/v1beta1/generated.proto b/vendor/k8s.io/api/discovery/v1beta1/generated.proto index cce6f970..adf10c0e 100644 --- a/vendor/k8s.io/api/discovery/v1beta1/generated.proto +++ b/vendor/k8s.io/api/discovery/v1beta1/generated.proto @@ -107,8 +107,9 @@ message EndpointPort { // This field follows standard Kubernetes label syntax. // Un-prefixed names are reserved for IANA standard service names (as per // RFC-6335 and http://www.iana.org/assignments/service-names). - // Non-standard protocols should use prefixed names. - // Default is empty string. + // Non-standard protocols should use prefixed names such as + // mycompany.com/my-custom-protocol. + // +optional optional string appProtocol = 4; } @@ -151,7 +152,6 @@ message EndpointSliceList { optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; // List of endpoint slices - // +listType=set repeated EndpointSlice items = 2; } diff --git a/vendor/k8s.io/api/discovery/v1beta1/types.go b/vendor/k8s.io/api/discovery/v1beta1/types.go index e3dc5653..5cafea74 100644 --- a/vendor/k8s.io/api/discovery/v1beta1/types.go +++ b/vendor/k8s.io/api/discovery/v1beta1/types.go @@ -23,6 +23,8 @@ import ( // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.16 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // EndpointSlice represents a subset of the endpoints that implement a service. // For a given service there may be multiple EndpointSlice objects, selected by @@ -143,12 +145,15 @@ type EndpointPort struct { // This field follows standard Kubernetes label syntax. // Un-prefixed names are reserved for IANA standard service names (as per // RFC-6335 and http://www.iana.org/assignments/service-names). - // Non-standard protocols should use prefixed names. - // Default is empty string. + // Non-standard protocols should use prefixed names such as + // mycompany.com/my-custom-protocol. + // +optional AppProtocol *string `json:"appProtocol,omitempty" protobuf:"bytes,4,name=appProtocol"` } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.16 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // EndpointSliceList represents a list of endpoint slices type EndpointSliceList struct { @@ -157,6 +162,5 @@ type EndpointSliceList struct { // +optional metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` // List of endpoint slices - // +listType=set Items []EndpointSlice `json:"items" protobuf:"bytes,2,rep,name=items"` } diff --git a/vendor/k8s.io/api/discovery/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/discovery/v1beta1/types_swagger_doc_generated.go index 9dd3a035..d67cc721 100644 --- a/vendor/k8s.io/api/discovery/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/discovery/v1beta1/types_swagger_doc_generated.go @@ -54,7 +54,7 @@ var map_EndpointPort = map[string]string{ "name": "The name of this port. All ports in an EndpointSlice must have a unique name. If the EndpointSlice is dervied from a Kubernetes service, this corresponds to the Service.ports[].name. Name must either be an empty string or pass DNS_LABEL validation: * must be no more than 63 characters long. * must consist of lower case alphanumeric characters or '-'. * must start and end with an alphanumeric character. Default is empty string.", "protocol": "The IP protocol for this port. Must be UDP, TCP, or SCTP. Default is TCP.", "port": "The port number of the endpoint. If this is not specified, ports are not restricted and must be interpreted in the context of the specific consumer.", - "appProtocol": "The application protocol for this port. This field follows standard Kubernetes label syntax. Un-prefixed names are reserved for IANA standard service names (as per RFC-6335 and http://www.iana.org/assignments/service-names). Non-standard protocols should use prefixed names. Default is empty string.", + "appProtocol": "The application protocol for this port. This field follows standard Kubernetes label syntax. Un-prefixed names are reserved for IANA standard service names (as per RFC-6335 and http://www.iana.org/assignments/service-names). Non-standard protocols should use prefixed names such as mycompany.com/my-custom-protocol.", } func (EndpointPort) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/discovery/v1beta1/well_known_labels.go b/vendor/k8s.io/api/discovery/v1beta1/well_known_labels.go index b0caa3c6..6ca64360 100644 --- a/vendor/k8s.io/api/discovery/v1beta1/well_known_labels.go +++ b/vendor/k8s.io/api/discovery/v1beta1/well_known_labels.go @@ -25,4 +25,8 @@ const ( // same cluster. It is highly recommended to configure this label for all // EndpointSlices. LabelManagedBy = "endpointslice.kubernetes.io/managed-by" + // LabelSkipMirror can be set to true on an Endpoints resource to indicate + // that the EndpointSliceMirroring controller should not mirror this + // resource with EndpointSlices. + LabelSkipMirror = "endpointslice.kubernetes.io/skip-mirror" ) diff --git a/vendor/k8s.io/api/discovery/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/discovery/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..09e94d0e --- /dev/null +++ b/vendor/k8s.io/api/discovery/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,57 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *EndpointSlice) APILifecycleIntroduced() (major, minor int) { + return 1, 16 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *EndpointSlice) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *EndpointSlice) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *EndpointSliceList) APILifecycleIntroduced() (major, minor int) { + return 1, 16 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *EndpointSliceList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *EndpointSliceList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} diff --git a/vendor/k8s.io/api/events/v1/doc.go b/vendor/k8s.io/api/events/v1/doc.go new file mode 100644 index 00000000..6e320e06 --- /dev/null +++ b/vendor/k8s.io/api/events/v1/doc.go @@ -0,0 +1,23 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// +k8s:deepcopy-gen=package +// +k8s:protobuf-gen=package +// +k8s:openapi-gen=true + +// +groupName=events.k8s.io + +package v1 // import "k8s.io/api/events/v1" diff --git a/vendor/k8s.io/api/events/v1/generated.pb.go b/vendor/k8s.io/api/events/v1/generated.pb.go new file mode 100644 index 00000000..717137cf --- /dev/null +++ b/vendor/k8s.io/api/events/v1/generated.pb.go @@ -0,0 +1,1406 @@ +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by protoc-gen-gogo. DO NOT EDIT. +// source: k8s.io/kubernetes/vendor/k8s.io/api/events/v1/generated.proto + +package v1 + +import ( + fmt "fmt" + + io "io" + + proto "github.com/gogo/protobuf/proto" + v11 "k8s.io/api/core/v1" + + math "math" + math_bits "math/bits" + reflect "reflect" + strings "strings" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package + +func (m *Event) Reset() { *m = Event{} } +func (*Event) ProtoMessage() {} +func (*Event) Descriptor() ([]byte, []int) { + return fileDescriptor_ee2600587b650fac, []int{0} +} +func (m *Event) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *Event) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *Event) XXX_Merge(src proto.Message) { + xxx_messageInfo_Event.Merge(m, src) +} +func (m *Event) XXX_Size() int { + return m.Size() +} +func (m *Event) XXX_DiscardUnknown() { + xxx_messageInfo_Event.DiscardUnknown(m) +} + +var xxx_messageInfo_Event proto.InternalMessageInfo + +func (m *EventList) Reset() { *m = EventList{} } +func (*EventList) ProtoMessage() {} +func (*EventList) Descriptor() ([]byte, []int) { + return fileDescriptor_ee2600587b650fac, []int{1} +} +func (m *EventList) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *EventList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *EventList) XXX_Merge(src proto.Message) { + xxx_messageInfo_EventList.Merge(m, src) +} +func (m *EventList) XXX_Size() int { + return m.Size() +} +func (m *EventList) XXX_DiscardUnknown() { + xxx_messageInfo_EventList.DiscardUnknown(m) +} + +var xxx_messageInfo_EventList proto.InternalMessageInfo + +func (m *EventSeries) Reset() { *m = EventSeries{} } +func (*EventSeries) ProtoMessage() {} +func (*EventSeries) Descriptor() ([]byte, []int) { + return fileDescriptor_ee2600587b650fac, []int{2} +} +func (m *EventSeries) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *EventSeries) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *EventSeries) XXX_Merge(src proto.Message) { + xxx_messageInfo_EventSeries.Merge(m, src) +} +func (m *EventSeries) XXX_Size() int { + return m.Size() +} +func (m *EventSeries) XXX_DiscardUnknown() { + xxx_messageInfo_EventSeries.DiscardUnknown(m) +} + +var xxx_messageInfo_EventSeries proto.InternalMessageInfo + +func init() { + proto.RegisterType((*Event)(nil), "k8s.io.api.events.v1.Event") + proto.RegisterType((*EventList)(nil), "k8s.io.api.events.v1.EventList") + proto.RegisterType((*EventSeries)(nil), "k8s.io.api.events.v1.EventSeries") +} + +func init() { + proto.RegisterFile("k8s.io/kubernetes/vendor/k8s.io/api/events/v1/generated.proto", fileDescriptor_ee2600587b650fac) +} + +var fileDescriptor_ee2600587b650fac = []byte{ + // 772 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x54, 0x4f, 0x6f, 0xe3, 0x44, + 0x14, 0x8f, 0xbb, 0x4d, 0xda, 0x4c, 0x76, 0xb7, 0xe9, 0x2c, 0x52, 0x87, 0xae, 0xe4, 0x84, 0xac, + 0x84, 0x22, 0x24, 0x6c, 0xb2, 0x42, 0x88, 0x0b, 0x12, 0xeb, 0xa6, 0xa0, 0xa2, 0x96, 0x4a, 0xd3, + 0x9e, 0x10, 0x87, 0x4e, 0x9c, 0x57, 0xd7, 0x24, 0x9e, 0xb1, 0x66, 0x26, 0x91, 0x7a, 0xe3, 0x82, + 0xc4, 0x91, 0x2f, 0xc0, 0x07, 0x40, 0x7c, 0x91, 0x1e, 0x7b, 0xec, 0x29, 0xa2, 0xe6, 0x8b, 0x20, + 0x8f, 0x9d, 0x38, 0xcd, 0x1f, 0x08, 0xda, 0x9b, 0xe7, 0xbd, 0xdf, 0x9f, 0xf7, 0x66, 0x9e, 0x1f, + 0xfa, 0x6a, 0xf0, 0xa5, 0x72, 0x42, 0xe1, 0x0e, 0x46, 0x3d, 0x90, 0x1c, 0x34, 0x28, 0x77, 0x0c, + 0xbc, 0x2f, 0xa4, 0x9b, 0x27, 0x58, 0x1c, 0xba, 0x30, 0x06, 0xae, 0x95, 0x3b, 0xee, 0xb8, 0x01, + 0x70, 0x90, 0x4c, 0x43, 0xdf, 0x89, 0xa5, 0xd0, 0x02, 0x7f, 0x90, 0xa1, 0x1c, 0x16, 0x87, 0x4e, + 0x86, 0x72, 0xc6, 0x9d, 0xc3, 0x4f, 0x83, 0x50, 0xdf, 0x8c, 0x7a, 0x8e, 0x2f, 0x22, 0x37, 0x10, + 0x81, 0x70, 0x0d, 0xb8, 0x37, 0xba, 0x36, 0x27, 0x73, 0x30, 0x5f, 0x99, 0xc8, 0x61, 0x6b, 0xce, + 0xca, 0x17, 0x12, 0x56, 0x18, 0x1d, 0x7e, 0x5e, 0x60, 0x22, 0xe6, 0xdf, 0x84, 0x1c, 0xe4, 0xad, + 0x1b, 0x0f, 0x82, 0x34, 0xa0, 0xdc, 0x08, 0x34, 0x5b, 0xc5, 0x72, 0xd7, 0xb1, 0xe4, 0x88, 0xeb, + 0x30, 0x82, 0x25, 0xc2, 0x17, 0xff, 0x45, 0x50, 0xfe, 0x0d, 0x44, 0x6c, 0x91, 0xd7, 0xfa, 0xbd, + 0x8a, 0xca, 0xc7, 0x69, 0xff, 0xf8, 0x0a, 0xed, 0xa6, 0xd5, 0xf4, 0x99, 0x66, 0xc4, 0x6a, 0x5a, + 0xed, 0xda, 0xdb, 0xcf, 0x9c, 0xe2, 0x92, 0x66, 0xa2, 0x4e, 0x3c, 0x08, 0xd2, 0x80, 0x72, 0x52, + 0xb4, 0x33, 0xee, 0x38, 0xe7, 0xbd, 0x9f, 0xc0, 0xd7, 0x67, 0xa0, 0x99, 0x87, 0xef, 0x26, 0x8d, + 0x52, 0x32, 0x69, 0xa0, 0x22, 0x46, 0x67, 0xaa, 0xf8, 0x0a, 0x55, 0xcd, 0x55, 0x5f, 0x86, 0x11, + 0x90, 0x2d, 0x63, 0xe1, 0x6e, 0x66, 0x71, 0x16, 0xfa, 0x52, 0xa4, 0x34, 0x6f, 0x3f, 0x77, 0xa8, + 0x1e, 0x4f, 0x95, 0x68, 0x21, 0x8a, 0x8f, 0x51, 0x45, 0x81, 0x0c, 0x41, 0x91, 0x67, 0x46, 0xfe, + 0x23, 0x67, 0xd5, 0x33, 0x3b, 0x86, 0x7b, 0x61, 0x80, 0x1e, 0x4a, 0x26, 0x8d, 0x4a, 0xf6, 0x4d, + 0x73, 0x32, 0x3e, 0x43, 0xaf, 0x24, 0xc4, 0x42, 0xea, 0x90, 0x07, 0x47, 0x82, 0x6b, 0x29, 0x86, + 0x43, 0x90, 0x64, 0xbb, 0x69, 0xb5, 0xab, 0xde, 0xeb, 0xbc, 0x82, 0x57, 0x74, 0x19, 0x42, 0x57, + 0xf1, 0xf0, 0xb7, 0x68, 0x7f, 0x16, 0x3e, 0xe1, 0x4a, 0x33, 0xee, 0x03, 0x29, 0x1b, 0xb1, 0x0f, + 0x73, 0xb1, 0x7d, 0xba, 0x08, 0xa0, 0xcb, 0x1c, 0xfc, 0x31, 0xaa, 0x30, 0x5f, 0x87, 0x82, 0x93, + 0x8a, 0x61, 0xbf, 0xcc, 0xd9, 0x95, 0x77, 0x26, 0x4a, 0xf3, 0x6c, 0x8a, 0x93, 0xc0, 0x94, 0xe0, + 0x64, 0xe7, 0x29, 0x8e, 0x9a, 0x28, 0xcd, 0xb3, 0xf8, 0x12, 0x55, 0x25, 0x04, 0x4c, 0xf6, 0x43, + 0x1e, 0x90, 0x5d, 0x73, 0x63, 0x6f, 0xe6, 0x6f, 0x2c, 0x9d, 0xe9, 0xe2, 0x85, 0x29, 0x5c, 0x83, + 0x04, 0xee, 0xcf, 0x3d, 0x02, 0x9d, 0xb2, 0x69, 0x21, 0x84, 0xbf, 0x43, 0x3b, 0x12, 0x86, 0xe9, + 0x8c, 0x91, 0xea, 0xe6, 0x9a, 0xb5, 0x64, 0xd2, 0xd8, 0xa1, 0x19, 0x8f, 0x4e, 0x05, 0x70, 0x13, + 0x6d, 0x73, 0xa1, 0x81, 0x20, 0xd3, 0xc7, 0xf3, 0xdc, 0x77, 0xfb, 0x7b, 0xa1, 0x81, 0x9a, 0x4c, + 0x8a, 0xd0, 0xb7, 0x31, 0x90, 0xda, 0x53, 0xc4, 0xe5, 0x6d, 0x0c, 0xd4, 0x64, 0x30, 0xa0, 0x7a, + 0x1f, 0x62, 0x09, 0x7e, 0xaa, 0x78, 0x21, 0x46, 0xd2, 0x07, 0xf2, 0xdc, 0x14, 0xd6, 0x58, 0x55, + 0x58, 0x36, 0x1c, 0x06, 0xe6, 0x91, 0x5c, 0xae, 0xde, 0x5d, 0x10, 0xa0, 0x4b, 0x92, 0xf8, 0x57, + 0x0b, 0x91, 0x22, 0xf8, 0x4d, 0x28, 0x95, 0x99, 0x49, 0xa5, 0x59, 0x14, 0x93, 0x17, 0xc6, 0xef, + 0x93, 0xcd, 0xa6, 0xdd, 0x0c, 0x7a, 0x33, 0xb7, 0x26, 0xdd, 0x35, 0x9a, 0x74, 0xad, 0x1b, 0xfe, + 0xc5, 0x42, 0x07, 0x45, 0xf2, 0x94, 0xcd, 0x57, 0xf2, 0xf2, 0x7f, 0x57, 0xd2, 0xc8, 0x2b, 0x39, + 0xe8, 0xae, 0x96, 0xa4, 0xeb, 0xbc, 0xf0, 0x3b, 0xb4, 0x57, 0xa4, 0x8e, 0xc4, 0x88, 0x6b, 0xb2, + 0xd7, 0xb4, 0xda, 0x65, 0xef, 0x20, 0x97, 0xdc, 0xeb, 0x3e, 0x4d, 0xd3, 0x45, 0x7c, 0xeb, 0x4f, + 0x0b, 0x65, 0xbf, 0xfa, 0x69, 0xa8, 0x34, 0xfe, 0x71, 0x69, 0x47, 0x39, 0x9b, 0x35, 0x92, 0xb2, + 0xcd, 0x86, 0xaa, 0xe7, 0xce, 0xbb, 0xd3, 0xc8, 0xdc, 0x7e, 0xfa, 0x1a, 0x95, 0x43, 0x0d, 0x91, + 0x22, 0x5b, 0xcd, 0x67, 0xed, 0xda, 0xdb, 0xd7, 0xff, 0xb2, 0x3c, 0xbc, 0x17, 0xb9, 0x4e, 0xf9, + 0x24, 0x65, 0xd0, 0x8c, 0xd8, 0xfa, 0xc3, 0x42, 0xb5, 0xb9, 0xe5, 0x82, 0xdf, 0xa0, 0xb2, 0x6f, + 0xda, 0xb6, 0x4c, 0xdb, 0x33, 0x52, 0xd6, 0x6c, 0x96, 0xc3, 0x23, 0x54, 0x1f, 0x32, 0xa5, 0xcf, + 0x7b, 0x0a, 0xe4, 0x18, 0xfa, 0xef, 0xb3, 0x1d, 0x67, 0xf3, 0x7a, 0xba, 0x20, 0x48, 0x97, 0x2c, + 0xbc, 0xf6, 0xdd, 0xa3, 0x5d, 0xba, 0x7f, 0xb4, 0x4b, 0x0f, 0x8f, 0x76, 0xe9, 0xe7, 0xc4, 0xb6, + 0xee, 0x12, 0xdb, 0xba, 0x4f, 0x6c, 0xeb, 0x21, 0xb1, 0xad, 0xbf, 0x12, 0xdb, 0xfa, 0xed, 0x6f, + 0xbb, 0xf4, 0xc3, 0xd6, 0xb8, 0xf3, 0x4f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x6b, 0xcb, 0x1e, 0x6e, + 0x6b, 0x07, 0x00, 0x00, +} + +func (m *Event) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *Event) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *Event) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i = encodeVarintGenerated(dAtA, i, uint64(m.DeprecatedCount)) + i-- + dAtA[i] = 0x78 + { + size, err := m.DeprecatedLastTimestamp.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x72 + { + size, err := m.DeprecatedFirstTimestamp.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x6a + { + size, err := m.DeprecatedSource.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x62 + i -= len(m.Type) + copy(dAtA[i:], m.Type) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Type))) + i-- + dAtA[i] = 0x5a + i -= len(m.Note) + copy(dAtA[i:], m.Note) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Note))) + i-- + dAtA[i] = 0x52 + if m.Related != nil { + { + size, err := m.Related.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x4a + } + { + size, err := m.Regarding.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x42 + i -= len(m.Reason) + copy(dAtA[i:], m.Reason) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Reason))) + i-- + dAtA[i] = 0x3a + i -= len(m.Action) + copy(dAtA[i:], m.Action) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Action))) + i-- + dAtA[i] = 0x32 + i -= len(m.ReportingInstance) + copy(dAtA[i:], m.ReportingInstance) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.ReportingInstance))) + i-- + dAtA[i] = 0x2a + i -= len(m.ReportingController) + copy(dAtA[i:], m.ReportingController) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.ReportingController))) + i-- + dAtA[i] = 0x22 + if m.Series != nil { + { + size, err := m.Series.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1a + } + { + size, err := m.EventTime.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *EventList) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *EventList) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *EventList) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + { + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *EventSeries) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *EventSeries) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *EventSeries) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + { + size, err := m.LastObservedTime.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + i = encodeVarintGenerated(dAtA, i, uint64(m.Count)) + i-- + dAtA[i] = 0x8 + return len(dAtA) - i, nil +} + +func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { + offset -= sovGenerated(v) + base := offset + for v >= 1<<7 { + dAtA[offset] = uint8(v&0x7f | 0x80) + v >>= 7 + offset++ + } + dAtA[offset] = uint8(v) + return base +} +func (m *Event) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.EventTime.Size() + n += 1 + l + sovGenerated(uint64(l)) + if m.Series != nil { + l = m.Series.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + l = len(m.ReportingController) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.ReportingInstance) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Action) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Reason) + n += 1 + l + sovGenerated(uint64(l)) + l = m.Regarding.Size() + n += 1 + l + sovGenerated(uint64(l)) + if m.Related != nil { + l = m.Related.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + l = len(m.Note) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Type) + n += 1 + l + sovGenerated(uint64(l)) + l = m.DeprecatedSource.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.DeprecatedFirstTimestamp.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.DeprecatedLastTimestamp.Size() + n += 1 + l + sovGenerated(uint64(l)) + n += 1 + sovGenerated(uint64(m.DeprecatedCount)) + return n +} + +func (m *EventList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *EventSeries) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + n += 1 + sovGenerated(uint64(m.Count)) + l = m.LastObservedTime.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func sovGenerated(x uint64) (n int) { + return (math_bits.Len64(x|1) + 6) / 7 +} +func sozGenerated(x uint64) (n int) { + return sovGenerated(uint64((x << 1) ^ uint64((int64(x) >> 63)))) +} +func (this *Event) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&Event{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `EventTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.EventTime), "MicroTime", "v1.MicroTime", 1), `&`, ``, 1) + `,`, + `Series:` + strings.Replace(this.Series.String(), "EventSeries", "EventSeries", 1) + `,`, + `ReportingController:` + fmt.Sprintf("%v", this.ReportingController) + `,`, + `ReportingInstance:` + fmt.Sprintf("%v", this.ReportingInstance) + `,`, + `Action:` + fmt.Sprintf("%v", this.Action) + `,`, + `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`, + `Regarding:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.Regarding), "ObjectReference", "v11.ObjectReference", 1), `&`, ``, 1) + `,`, + `Related:` + strings.Replace(fmt.Sprintf("%v", this.Related), "ObjectReference", "v11.ObjectReference", 1) + `,`, + `Note:` + fmt.Sprintf("%v", this.Note) + `,`, + `Type:` + fmt.Sprintf("%v", this.Type) + `,`, + `DeprecatedSource:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.DeprecatedSource), "EventSource", "v11.EventSource", 1), `&`, ``, 1) + `,`, + `DeprecatedFirstTimestamp:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.DeprecatedFirstTimestamp), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `DeprecatedLastTimestamp:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.DeprecatedLastTimestamp), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `DeprecatedCount:` + fmt.Sprintf("%v", this.DeprecatedCount) + `,`, + `}`, + }, "") + return s +} +func (this *EventList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]Event{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "Event", "Event", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&EventList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} +func (this *EventSeries) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&EventSeries{`, + `Count:` + fmt.Sprintf("%v", this.Count) + `,`, + `LastObservedTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastObservedTime), "MicroTime", "v1.MicroTime", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func valueToStringGenerated(v interface{}) string { + rv := reflect.ValueOf(v) + if rv.IsNil() { + return "nil" + } + pv := reflect.Indirect(rv).Interface() + return fmt.Sprintf("*%v", pv) +} +func (m *Event) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: Event: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: Event: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field EventTime", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.EventTime.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Series", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Series == nil { + m.Series = &EventSeries{} + } + if err := m.Series.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ReportingController", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.ReportingController = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ReportingInstance", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.ReportingInstance = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 6: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Action", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Action = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 7: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Reason", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Reason = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 8: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Regarding", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Regarding.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 9: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Related", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Related == nil { + m.Related = &v11.ObjectReference{} + } + if err := m.Related.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 10: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Note", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Note = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 11: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Type", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Type = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 12: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field DeprecatedSource", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.DeprecatedSource.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 13: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field DeprecatedFirstTimestamp", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.DeprecatedFirstTimestamp.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 14: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field DeprecatedLastTimestamp", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.DeprecatedLastTimestamp.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 15: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field DeprecatedCount", wireType) + } + m.DeprecatedCount = 0 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + m.DeprecatedCount |= int32(b&0x7F) << shift + if b < 0x80 { + break + } + } + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *EventList) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: EventList: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: EventList: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Items = append(m.Items, Event{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *EventSeries) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: EventSeries: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: EventSeries: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field Count", wireType) + } + m.Count = 0 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + m.Count |= int32(b&0x7F) << shift + if b < 0x80 { + break + } + } + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field LastObservedTime", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.LastObservedTime.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func skipGenerated(dAtA []byte) (n int, err error) { + l := len(dAtA) + iNdEx := 0 + depth := 0 + for iNdEx < l { + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return 0, ErrIntOverflowGenerated + } + if iNdEx >= l { + return 0, io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= (uint64(b) & 0x7F) << shift + if b < 0x80 { + break + } + } + wireType := int(wire & 0x7) + switch wireType { + case 0: + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return 0, ErrIntOverflowGenerated + } + if iNdEx >= l { + return 0, io.ErrUnexpectedEOF + } + iNdEx++ + if dAtA[iNdEx-1] < 0x80 { + break + } + } + case 1: + iNdEx += 8 + case 2: + var length int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return 0, ErrIntOverflowGenerated + } + if iNdEx >= l { + return 0, io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + length |= (int(b) & 0x7F) << shift + if b < 0x80 { + break + } + } + if length < 0 { + return 0, ErrInvalidLengthGenerated + } + iNdEx += length + case 3: + depth++ + case 4: + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- + case 5: + iNdEx += 4 + default: + return 0, fmt.Errorf("proto: illegal wireType %d", wireType) + } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } + } + return 0, io.ErrUnexpectedEOF +} + +var ( + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") +) diff --git a/vendor/k8s.io/api/events/v1/generated.proto b/vendor/k8s.io/api/events/v1/generated.proto new file mode 100644 index 00000000..18e3d018 --- /dev/null +++ b/vendor/k8s.io/api/events/v1/generated.proto @@ -0,0 +1,125 @@ +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + + +// This file was autogenerated by go-to-protobuf. Do not edit it manually! + +syntax = 'proto2'; + +package k8s.io.api.events.v1; + +import "k8s.io/api/core/v1/generated.proto"; +import "k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto"; +import "k8s.io/apimachinery/pkg/runtime/generated.proto"; +import "k8s.io/apimachinery/pkg/runtime/schema/generated.proto"; + +// Package-wide variables from generator "generated". +option go_package = "v1"; + +// Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system. +message Event { + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // eventTime is the time when this Event was first observed. It is required. + optional k8s.io.apimachinery.pkg.apis.meta.v1.MicroTime eventTime = 2; + + // series is data about the Event series this event represents or nil if it's a singleton Event. + // +optional + optional EventSeries series = 3; + + // reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. + // This field cannot be empty for new Events. + // +optional + optional string reportingController = 4; + + // reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. + // This field cannot be empty for new Events and it can have at most 128 characters. + // +optional + optional string reportingInstance = 5; + + // action is what action was taken/failed regarding to the regarding object. It is machine-readable. + // This field can have at most 128 characters. + // +optional + optional string action = 6; + + // reason is why the action was taken. It is human-readable. + // This field can have at most 128 characters. + // +optional + optional string reason = 7; + + // regarding contains the object this Event is about. In most cases it's an Object reporting controller + // implements, e.g. ReplicaSetController implements ReplicaSets and this event is emitted because + // it acts on some changes in a ReplicaSet object. + // +optional + optional k8s.io.api.core.v1.ObjectReference regarding = 8; + + // related is the optional secondary object for more complex actions. E.g. when regarding object triggers + // a creation or deletion of related object. + // +optional + optional k8s.io.api.core.v1.ObjectReference related = 9; + + // note is a human-readable description of the status of this operation. + // Maximal length of the note is 1kB, but libraries should be prepared to + // handle values up to 64kB. + // +optional + optional string note = 10; + + // type is the type of this event (Normal, Warning), new types could be added in the future. + // It is machine-readable. + // +optional + optional string type = 11; + + // deprecatedSource is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + optional k8s.io.api.core.v1.EventSource deprecatedSource = 12; + + // deprecatedFirstTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.Time deprecatedFirstTimestamp = 13; + + // deprecatedLastTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.Time deprecatedLastTimestamp = 14; + + // deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + optional int32 deprecatedCount = 15; +} + +// EventList is a list of Event objects. +message EventList { + // Standard list metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; + + // items is a list of schema objects. + repeated Event items = 2; +} + +// EventSeries contain information on series of events, i.e. thing that was/is happening +// continuously for some time. How often to update the EventSeries is up to the event reporters. +// The default event reporter in "k8s.io/client-go/tools/events/event_broadcaster.go" shows +// how this struct is updated on heartbeats and can guide customized reporter implementations. +message EventSeries { + // count is the number of occurrences in this series up to the last heartbeat time. + optional int32 count = 1; + + // lastObservedTime is the time when last Event from the series was seen before last heartbeat. + optional k8s.io.apimachinery.pkg.apis.meta.v1.MicroTime lastObservedTime = 2; +} + diff --git a/vendor/k8s.io/api/auditregistration/v1alpha1/register.go b/vendor/k8s.io/api/events/v1/register.go similarity index 73% rename from vendor/k8s.io/api/auditregistration/v1alpha1/register.go rename to vendor/k8s.io/api/events/v1/register.go index d6271608..ca90e6cb 100644 --- a/vendor/k8s.io/api/auditregistration/v1alpha1/register.go +++ b/vendor/k8s.io/api/events/v1/register.go @@ -1,5 +1,5 @@ /* -Copyright 2018 The Kubernetes Authors. +Copyright 2020 The Kubernetes Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. @@ -14,7 +14,7 @@ See the License for the specific language governing permissions and limitations under the License. */ -package v1alpha1 +package v1 import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" @@ -23,10 +23,10 @@ import ( ) // GroupName is the group name use in this package -const GroupName = "auditregistration.k8s.io" +const GroupName = "events.k8s.io" // SchemeGroupVersion is group version used to register these objects -var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: "v1alpha1"} +var SchemeGroupVersion = schema.GroupVersion{Group: GroupName, Version: "v1"} // Resource takes an unqualified resource and returns a Group qualified GroupResource func Resource(resource string) schema.GroupResource { @@ -34,23 +34,20 @@ func Resource(resource string) schema.GroupResource { } var ( - SchemeBuilder runtime.SchemeBuilder + // TODO: move SchemeBuilder with zz_generated.deepcopy.go to k8s.io/api. + // localSchemeBuilder and AddToScheme will stay in k8s.io/kubernetes. + SchemeBuilder = runtime.NewSchemeBuilder(addKnownTypes) localSchemeBuilder = &SchemeBuilder AddToScheme = localSchemeBuilder.AddToScheme ) -func init() { - // We only register manually written functions here. The registration of the - // generated functions takes place in the generated files. The separation - // makes the code compile even when the generated files are missing. - localSchemeBuilder.Register(addKnownTypes) -} - +// Adds the list of known types to api.Scheme. func addKnownTypes(scheme *runtime.Scheme) error { scheme.AddKnownTypes(SchemeGroupVersion, - &AuditSink{}, - &AuditSinkList{}, + &Event{}, + &EventList{}, ) + metav1.AddToGroupVersion(scheme, SchemeGroupVersion) return nil } diff --git a/vendor/k8s.io/api/events/v1/types.go b/vendor/k8s.io/api/events/v1/types.go new file mode 100644 index 00000000..07ede554 --- /dev/null +++ b/vendor/k8s.io/api/events/v1/types.go @@ -0,0 +1,119 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package v1 + +import ( + corev1 "k8s.io/api/core/v1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" +) + +// +genclient +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system. +type Event struct { + metav1.TypeMeta `json:",inline"` + // +optional + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // eventTime is the time when this Event was first observed. It is required. + EventTime metav1.MicroTime `json:"eventTime" protobuf:"bytes,2,opt,name=eventTime"` + + // series is data about the Event series this event represents or nil if it's a singleton Event. + // +optional + Series *EventSeries `json:"series,omitempty" protobuf:"bytes,3,opt,name=series"` + + // reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. + // This field cannot be empty for new Events. + // +optional + ReportingController string `json:"reportingController,omitempty" protobuf:"bytes,4,opt,name=reportingController"` + + // reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. + // This field cannot be empty for new Events and it can have at most 128 characters. + // +optional + ReportingInstance string `json:"reportingInstance,omitempty" protobuf:"bytes,5,opt,name=reportingInstance"` + + // action is what action was taken/failed regarding to the regarding object. It is machine-readable. + // This field can have at most 128 characters. + // +optional + Action string `json:"action,omitempty" protobuf:"bytes,6,name=action"` + + // reason is why the action was taken. It is human-readable. + // This field can have at most 128 characters. + // +optional + Reason string `json:"reason,omitempty" protobuf:"bytes,7,name=reason"` + + // regarding contains the object this Event is about. In most cases it's an Object reporting controller + // implements, e.g. ReplicaSetController implements ReplicaSets and this event is emitted because + // it acts on some changes in a ReplicaSet object. + // +optional + Regarding corev1.ObjectReference `json:"regarding,omitempty" protobuf:"bytes,8,opt,name=regarding"` + + // related is the optional secondary object for more complex actions. E.g. when regarding object triggers + // a creation or deletion of related object. + // +optional + Related *corev1.ObjectReference `json:"related,omitempty" protobuf:"bytes,9,opt,name=related"` + + // note is a human-readable description of the status of this operation. + // Maximal length of the note is 1kB, but libraries should be prepared to + // handle values up to 64kB. + // +optional + Note string `json:"note,omitempty" protobuf:"bytes,10,opt,name=note"` + + // type is the type of this event (Normal, Warning), new types could be added in the future. + // It is machine-readable. + // +optional + Type string `json:"type,omitempty" protobuf:"bytes,11,opt,name=type"` + + // deprecatedSource is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + DeprecatedSource corev1.EventSource `json:"deprecatedSource,omitempty" protobuf:"bytes,12,opt,name=deprecatedSource"` + // deprecatedFirstTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + DeprecatedFirstTimestamp metav1.Time `json:"deprecatedFirstTimestamp,omitempty" protobuf:"bytes,13,opt,name=deprecatedFirstTimestamp"` + // deprecatedLastTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + DeprecatedLastTimestamp metav1.Time `json:"deprecatedLastTimestamp,omitempty" protobuf:"bytes,14,opt,name=deprecatedLastTimestamp"` + // deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type. + // +optional + DeprecatedCount int32 `json:"deprecatedCount,omitempty" protobuf:"varint,15,opt,name=deprecatedCount"` +} + +// EventSeries contain information on series of events, i.e. thing that was/is happening +// continuously for some time. How often to update the EventSeries is up to the event reporters. +// The default event reporter in "k8s.io/client-go/tools/events/event_broadcaster.go" shows +// how this struct is updated on heartbeats and can guide customized reporter implementations. +type EventSeries struct { + // count is the number of occurrences in this series up to the last heartbeat time. + Count int32 `json:"count" protobuf:"varint,1,opt,name=count"` + // lastObservedTime is the time when last Event from the series was seen before last heartbeat. + LastObservedTime metav1.MicroTime `json:"lastObservedTime" protobuf:"bytes,2,opt,name=lastObservedTime"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// EventList is a list of Event objects. +type EventList struct { + metav1.TypeMeta `json:",inline"` + // Standard list metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // items is a list of schema objects. + Items []Event `json:"items" protobuf:"bytes,2,rep,name=items"` +} diff --git a/vendor/k8s.io/api/events/v1/types_swagger_doc_generated.go b/vendor/k8s.io/api/events/v1/types_swagger_doc_generated.go new file mode 100644 index 00000000..e0467436 --- /dev/null +++ b/vendor/k8s.io/api/events/v1/types_swagger_doc_generated.go @@ -0,0 +1,72 @@ +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package v1 + +// This file contains a collection of methods that can be used from go-restful to +// generate Swagger API documentation for its models. Please read this PR for more +// information on the implementation: https://github.com/emicklei/go-restful/pull/215 +// +// TODOs are ignored from the parser (e.g. TODO(andronat):... || TODO:...) if and only if +// they are on one line! For multiple line or blocks that you want to ignore use ---. +// Any context after a --- is ignored. +// +// Those methods can be generated by using hack/update-generated-swagger-docs.sh + +// AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. +var map_Event = map[string]string{ + "": "Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system.", + "eventTime": "eventTime is the time when this Event was first observed. It is required.", + "series": "series is data about the Event series this event represents or nil if it's a singleton Event.", + "reportingController": "reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. This field cannot be empty for new Events.", + "reportingInstance": "reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. This field cannot be empty for new Events and it can have at most 128 characters.", + "action": "action is what action was taken/failed regarding to the regarding object. It is machine-readable. This field can have at most 128 characters.", + "reason": "reason is why the action was taken. It is human-readable. This field can have at most 128 characters.", + "regarding": "regarding contains the object this Event is about. In most cases it's an Object reporting controller implements, e.g. ReplicaSetController implements ReplicaSets and this event is emitted because it acts on some changes in a ReplicaSet object.", + "related": "related is the optional secondary object for more complex actions. E.g. when regarding object triggers a creation or deletion of related object.", + "note": "note is a human-readable description of the status of this operation. Maximal length of the note is 1kB, but libraries should be prepared to handle values up to 64kB.", + "type": "type is the type of this event (Normal, Warning), new types could be added in the future. It is machine-readable.", + "deprecatedSource": "deprecatedSource is the deprecated field assuring backward compatibility with core.v1 Event type.", + "deprecatedFirstTimestamp": "deprecatedFirstTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type.", + "deprecatedLastTimestamp": "deprecatedLastTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type.", + "deprecatedCount": "deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type.", +} + +func (Event) SwaggerDoc() map[string]string { + return map_Event +} + +var map_EventList = map[string]string{ + "": "EventList is a list of Event objects.", + "metadata": "Standard list metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "items": "items is a list of schema objects.", +} + +func (EventList) SwaggerDoc() map[string]string { + return map_EventList +} + +var map_EventSeries = map[string]string{ + "": "EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time. How often to update the EventSeries is up to the event reporters. The default event reporter in \"k8s.io/client-go/tools/events/event_broadcaster.go\" shows how this struct is updated on heartbeats and can guide customized reporter implementations.", + "count": "count is the number of occurrences in this series up to the last heartbeat time.", + "lastObservedTime": "lastObservedTime is the time when last Event from the series was seen before last heartbeat.", +} + +func (EventSeries) SwaggerDoc() map[string]string { + return map_EventSeries +} + +// AUTO-GENERATED FUNCTIONS END HERE diff --git a/vendor/k8s.io/api/events/v1/zz_generated.deepcopy.go b/vendor/k8s.io/api/events/v1/zz_generated.deepcopy.go new file mode 100644 index 00000000..19a3c5f7 --- /dev/null +++ b/vendor/k8s.io/api/events/v1/zz_generated.deepcopy.go @@ -0,0 +1,117 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by deepcopy-gen. DO NOT EDIT. + +package v1 + +import ( + corev1 "k8s.io/api/core/v1" + runtime "k8s.io/apimachinery/pkg/runtime" +) + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *Event) DeepCopyInto(out *Event) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + in.EventTime.DeepCopyInto(&out.EventTime) + if in.Series != nil { + in, out := &in.Series, &out.Series + *out = new(EventSeries) + (*in).DeepCopyInto(*out) + } + out.Regarding = in.Regarding + if in.Related != nil { + in, out := &in.Related, &out.Related + *out = new(corev1.ObjectReference) + **out = **in + } + out.DeprecatedSource = in.DeprecatedSource + in.DeprecatedFirstTimestamp.DeepCopyInto(&out.DeprecatedFirstTimestamp) + in.DeprecatedLastTimestamp.DeepCopyInto(&out.DeprecatedLastTimestamp) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Event. +func (in *Event) DeepCopy() *Event { + if in == nil { + return nil + } + out := new(Event) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *Event) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EventList) DeepCopyInto(out *EventList) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ListMeta.DeepCopyInto(&out.ListMeta) + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]Event, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventList. +func (in *EventList) DeepCopy() *EventList { + if in == nil { + return nil + } + out := new(EventList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *EventList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *EventSeries) DeepCopyInto(out *EventSeries) { + *out = *in + in.LastObservedTime.DeepCopyInto(&out.LastObservedTime) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new EventSeries. +func (in *EventSeries) DeepCopy() *EventSeries { + if in == nil { + return nil + } + out := new(EventSeries) + in.DeepCopyInto(out) + return out +} diff --git a/vendor/k8s.io/api/events/v1beta1/doc.go b/vendor/k8s.io/api/events/v1beta1/doc.go index 9bec7b3c..46048a65 100644 --- a/vendor/k8s.io/api/events/v1beta1/doc.go +++ b/vendor/k8s.io/api/events/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=events.k8s.io diff --git a/vendor/k8s.io/api/events/v1beta1/generated.pb.go b/vendor/k8s.io/api/events/v1beta1/generated.pb.go index 0e9a8e78..3709ef63 100644 --- a/vendor/k8s.io/api/events/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/events/v1beta1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Event) Reset() { *m = Event{} } func (*Event) ProtoMessage() {} @@ -139,58 +139,56 @@ func init() { } var fileDescriptor_4f97f691c32a5ac8 = []byte{ - // 801 bytes of a gzipped FileDescriptorProto + // 774 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x54, 0xcd, 0x6e, 0xdb, 0x46, - 0x10, 0x16, 0x13, 0x4b, 0xb2, 0x56, 0x49, 0x2c, 0x6f, 0x0e, 0xde, 0xb8, 0x00, 0xa5, 0x2a, 0x40, - 0x20, 0x14, 0x08, 0x59, 0x07, 0x45, 0xdb, 0x6b, 0x18, 0xb9, 0x45, 0x02, 0xbb, 0x01, 0xd6, 0x3e, + 0x10, 0x16, 0x13, 0x4b, 0xb2, 0x56, 0x49, 0x2c, 0x6f, 0x0e, 0xde, 0xb8, 0x00, 0x25, 0x28, 0x40, + 0x20, 0x14, 0x08, 0x59, 0x07, 0x45, 0xd1, 0x6b, 0x18, 0xb9, 0x45, 0x02, 0xbb, 0x01, 0x36, 0x3e, 0x15, 0x3d, 0x64, 0x45, 0x4d, 0x68, 0x56, 0xe2, 0x2e, 0xb1, 0xbb, 0x12, 0xe0, 0x5b, 0x2f, 0x05, - 0x7a, 0xec, 0x33, 0xf4, 0x09, 0xfa, 0x18, 0x3e, 0xe6, 0x98, 0x93, 0x50, 0xb3, 0x6f, 0xd1, 0x53, - 0xc1, 0xe5, 0x4a, 0x94, 0xf5, 0x83, 0xa8, 0xe8, 0x4d, 0x9c, 0xf9, 0x7e, 0x66, 0x66, 0x47, 0x83, - 0x82, 0xd1, 0xb7, 0xca, 0x8b, 0x85, 0x3f, 0x9a, 0x0c, 0x40, 0x72, 0xd0, 0xa0, 0xfc, 0x29, 0xf0, - 0xa1, 0x90, 0xbe, 0x4d, 0xb0, 0x34, 0xf6, 0x61, 0x0a, 0x5c, 0x2b, 0x7f, 0x7a, 0x32, 0x00, 0xcd, - 0x4e, 0xfc, 0x08, 0x38, 0x48, 0xa6, 0x61, 0xe8, 0xa5, 0x52, 0x68, 0x81, 0x9f, 0x14, 0x50, 0x8f, - 0xa5, 0xb1, 0x57, 0x40, 0x3d, 0x0b, 0x3d, 0x7e, 0x1e, 0xc5, 0xfa, 0x6a, 0x32, 0xf0, 0x42, 0x91, - 0xf8, 0x91, 0x88, 0x84, 0x6f, 0x18, 0x83, 0xc9, 0x7b, 0xf3, 0x65, 0x3e, 0xcc, 0xaf, 0x42, 0xe9, - 0xb8, 0xbb, 0x64, 0x1a, 0x0a, 0x09, 0xfe, 0x74, 0xcd, 0xed, 0xf8, 0xab, 0x12, 0x93, 0xb0, 0xf0, - 0x2a, 0xe6, 0x20, 0xaf, 0xfd, 0x74, 0x14, 0xe5, 0x01, 0xe5, 0x27, 0xa0, 0xd9, 0x26, 0x96, 0xbf, - 0x8d, 0x25, 0x27, 0x5c, 0xc7, 0x09, 0xac, 0x11, 0xbe, 0xfe, 0x14, 0x41, 0x85, 0x57, 0x90, 0xb0, - 0x55, 0x5e, 0xf7, 0x8f, 0x06, 0xaa, 0x9e, 0xe6, 0x43, 0xc0, 0xef, 0xd0, 0x7e, 0x5e, 0xcd, 0x90, - 0x69, 0x46, 0x9c, 0x8e, 0xd3, 0x6b, 0xbe, 0xf8, 0xd2, 0x2b, 0x27, 0xb5, 0x10, 0xf5, 0xd2, 0x51, - 0x94, 0x07, 0x94, 0x97, 0xa3, 0xbd, 0xe9, 0x89, 0xf7, 0x76, 0xf0, 0x33, 0x84, 0xfa, 0x1c, 0x34, - 0x0b, 0xf0, 0xcd, 0xac, 0x5d, 0xc9, 0x66, 0x6d, 0x54, 0xc6, 0xe8, 0x42, 0x15, 0xbf, 0x43, 0x0d, - 0x33, 0xef, 0xcb, 0x38, 0x01, 0x72, 0xcf, 0x58, 0xf8, 0xbb, 0x59, 0x9c, 0xc7, 0xa1, 0x14, 0x39, - 0x2d, 0x38, 0xb4, 0x0e, 0x8d, 0xd3, 0xb9, 0x12, 0x2d, 0x45, 0xf1, 0x1b, 0x54, 0x53, 0x20, 0x63, - 0x50, 0xe4, 0xbe, 0x91, 0x7f, 0xe6, 0x6d, 0x7d, 0x6b, 0xcf, 0x08, 0x5c, 0x18, 0x74, 0x80, 0xb2, - 0x59, 0xbb, 0x56, 0xfc, 0xa6, 0x56, 0x01, 0x9f, 0xa3, 0xc7, 0x12, 0x52, 0x21, 0x75, 0xcc, 0xa3, - 0x57, 0x82, 0x6b, 0x29, 0xc6, 0x63, 0x90, 0x64, 0xaf, 0xe3, 0xf4, 0x1a, 0xc1, 0x67, 0xb6, 0x8c, - 0xc7, 0x74, 0x1d, 0x42, 0x37, 0xf1, 0xf0, 0xf7, 0xe8, 0x70, 0x11, 0x7e, 0xcd, 0x95, 0x66, 0x3c, - 0x04, 0x52, 0x35, 0x62, 0x4f, 0xac, 0xd8, 0x21, 0x5d, 0x05, 0xd0, 0x75, 0x0e, 0x7e, 0x86, 0x6a, - 0x2c, 0xd4, 0xb1, 0xe0, 0xa4, 0x66, 0xd8, 0x8f, 0x2c, 0xbb, 0xf6, 0xd2, 0x44, 0xa9, 0xcd, 0xe6, - 0x38, 0x09, 0x4c, 0x09, 0x4e, 0xea, 0x77, 0x71, 0xd4, 0x44, 0xa9, 0xcd, 0xe2, 0x4b, 0xd4, 0x90, - 0x10, 0x31, 0x39, 0x8c, 0x79, 0x44, 0xf6, 0xcd, 0xd8, 0x9e, 0x2e, 0x8f, 0x2d, 0x5f, 0xec, 0xf2, - 0x99, 0x29, 0xbc, 0x07, 0x09, 0x3c, 0x5c, 0x7a, 0x09, 0x3a, 0x67, 0xd3, 0x52, 0x08, 0xbf, 0x41, - 0x75, 0x09, 0xe3, 0x7c, 0xd1, 0x48, 0x63, 0x77, 0xcd, 0x66, 0x36, 0x6b, 0xd7, 0x69, 0xc1, 0xa3, - 0x73, 0x01, 0xdc, 0x41, 0x7b, 0x5c, 0x68, 0x20, 0xc8, 0xf4, 0xf1, 0xc0, 0xfa, 0xee, 0xfd, 0x20, - 0x34, 0x50, 0x93, 0xc9, 0x11, 0xfa, 0x3a, 0x05, 0xd2, 0xbc, 0x8b, 0xb8, 0xbc, 0x4e, 0x81, 0x9a, - 0x0c, 0x06, 0xd4, 0x1a, 0x42, 0x2a, 0x21, 0xcc, 0x15, 0x2f, 0xc4, 0x44, 0x86, 0x40, 0x1e, 0x98, - 0xc2, 0xda, 0x9b, 0x0a, 0x2b, 0x96, 0xc3, 0xc0, 0x02, 0x62, 0xe5, 0x5a, 0xfd, 0x15, 0x01, 0xba, - 0x26, 0x89, 0x7f, 0x73, 0x10, 0x29, 0x83, 0xdf, 0xc5, 0x52, 0x99, 0xc5, 0x54, 0x9a, 0x25, 0x29, - 0x79, 0x68, 0xfc, 0xbe, 0xd8, 0x6d, 0xe5, 0xcd, 0xb6, 0x77, 0xac, 0x35, 0xe9, 0x6f, 0xd1, 0xa4, - 0x5b, 0xdd, 0xf0, 0xaf, 0x0e, 0x3a, 0x2a, 0x93, 0x67, 0x6c, 0xb9, 0x92, 0x47, 0xff, 0xb9, 0x92, - 0xb6, 0xad, 0xe4, 0xa8, 0xbf, 0x59, 0x92, 0x6e, 0xf3, 0xc2, 0x2f, 0xd1, 0x41, 0x99, 0x7a, 0x25, - 0x26, 0x5c, 0x93, 0x83, 0x8e, 0xd3, 0xab, 0x06, 0x47, 0x56, 0xf2, 0xa0, 0x7f, 0x37, 0x4d, 0x57, - 0xf1, 0xdd, 0x3f, 0x1d, 0x54, 0xfc, 0xdf, 0xcf, 0x62, 0xa5, 0xf1, 0x4f, 0x6b, 0x87, 0xca, 0xdb, - 0xad, 0x91, 0x9c, 0x6d, 0xce, 0x54, 0xcb, 0x3a, 0xef, 0xcf, 0x23, 0x4b, 0x47, 0xea, 0x14, 0x55, - 0x63, 0x0d, 0x89, 0x22, 0xf7, 0x3a, 0xf7, 0x7b, 0xcd, 0x17, 0x9d, 0x4f, 0x5d, 0x90, 0xe0, 0xa1, - 0x15, 0xab, 0xbe, 0xce, 0x69, 0xb4, 0x60, 0x77, 0x33, 0x07, 0x35, 0x97, 0x2e, 0x0c, 0x7e, 0x8a, - 0xaa, 0xa1, 0xe9, 0xdd, 0x31, 0xbd, 0x2f, 0x48, 0x45, 0xc7, 0x45, 0x0e, 0x4f, 0x50, 0x6b, 0xcc, - 0x94, 0x7e, 0x3b, 0x50, 0x20, 0xa7, 0x30, 0xfc, 0x3f, 0x77, 0x72, 0xb1, 0xb4, 0x67, 0x2b, 0x82, - 0x74, 0xcd, 0x02, 0x7f, 0x83, 0xaa, 0x4a, 0x33, 0x0d, 0xe6, 0x68, 0x36, 0x82, 0xcf, 0xe7, 0xb5, - 0x5d, 0xe4, 0xc1, 0x7f, 0x66, 0xed, 0xd6, 0x52, 0x23, 0x26, 0x46, 0x0b, 0x7c, 0xf0, 0xfc, 0xe6, - 0xd6, 0xad, 0x7c, 0xb8, 0x75, 0x2b, 0x1f, 0x6f, 0xdd, 0xca, 0x2f, 0x99, 0xeb, 0xdc, 0x64, 0xae, - 0xf3, 0x21, 0x73, 0x9d, 0x8f, 0x99, 0xeb, 0xfc, 0x95, 0xb9, 0xce, 0xef, 0x7f, 0xbb, 0x95, 0x1f, - 0xeb, 0x76, 0x5e, 0xff, 0x06, 0x00, 0x00, 0xff, 0xff, 0x25, 0x9b, 0x14, 0x4d, 0xbd, 0x07, 0x00, - 0x00, + 0x7a, 0xec, 0x33, 0xf4, 0xd6, 0x5b, 0x1f, 0xc3, 0xc7, 0x1c, 0x7d, 0x12, 0x6a, 0xf6, 0x45, 0x0a, + 0x2e, 0x57, 0xa2, 0xac, 0x1f, 0x58, 0x45, 0x6f, 0xe2, 0xcc, 0xf7, 0x33, 0x33, 0x3b, 0x1a, 0x14, + 0x8c, 0xbe, 0x55, 0x5e, 0x2c, 0xfc, 0xd1, 0x64, 0x00, 0x92, 0x83, 0x06, 0xe5, 0x4f, 0x81, 0x0f, + 0x85, 0xf4, 0x6d, 0x82, 0xa5, 0xb1, 0x0f, 0x53, 0xe0, 0x5a, 0xf9, 0xd3, 0x93, 0x01, 0x68, 0x76, + 0xe2, 0x47, 0xc0, 0x41, 0x32, 0x0d, 0x43, 0x2f, 0x95, 0x42, 0x0b, 0xfc, 0xac, 0x80, 0x7a, 0x2c, + 0x8d, 0xbd, 0x02, 0xea, 0x59, 0xe8, 0xf1, 0xcb, 0x28, 0xd6, 0x97, 0x93, 0x81, 0x17, 0x8a, 0xc4, + 0x8f, 0x44, 0x24, 0x7c, 0xc3, 0x18, 0x4c, 0x3e, 0x99, 0x2f, 0xf3, 0x61, 0x7e, 0x15, 0x4a, 0xc7, + 0xdd, 0x25, 0xd3, 0x50, 0x48, 0xf0, 0xa7, 0x6b, 0x6e, 0xc7, 0x5f, 0x97, 0x98, 0x84, 0x85, 0x97, + 0x31, 0x07, 0x79, 0xe5, 0xa7, 0xa3, 0x28, 0x0f, 0x28, 0x3f, 0x01, 0xcd, 0x36, 0xb1, 0xfc, 0x6d, + 0x2c, 0x39, 0xe1, 0x3a, 0x4e, 0x60, 0x8d, 0xf0, 0xcd, 0x7d, 0x04, 0x15, 0x5e, 0x42, 0xc2, 0x56, + 0x79, 0xdd, 0x3f, 0x1a, 0xa8, 0x7a, 0x9a, 0x0f, 0x01, 0x7f, 0x44, 0xfb, 0x79, 0x35, 0x43, 0xa6, + 0x19, 0x71, 0x3a, 0x4e, 0xaf, 0xf9, 0xea, 0x2b, 0xaf, 0x9c, 0xd4, 0x42, 0xd4, 0x4b, 0x47, 0x51, + 0x1e, 0x50, 0x5e, 0x8e, 0xf6, 0xa6, 0x27, 0xde, 0xfb, 0xc1, 0xcf, 0x10, 0xea, 0x73, 0xd0, 0x2c, + 0xc0, 0xd7, 0xb3, 0x76, 0x25, 0x9b, 0xb5, 0x51, 0x19, 0xa3, 0x0b, 0x55, 0xfc, 0x11, 0x35, 0xcc, + 0xbc, 0x2f, 0xe2, 0x04, 0xc8, 0x03, 0x63, 0xe1, 0xef, 0x66, 0x71, 0x1e, 0x87, 0x52, 0xe4, 0xb4, + 0xe0, 0xd0, 0x3a, 0x34, 0x4e, 0xe7, 0x4a, 0xb4, 0x14, 0xc5, 0xef, 0x50, 0x4d, 0x81, 0x8c, 0x41, + 0x91, 0x87, 0x46, 0xfe, 0x85, 0xb7, 0xf5, 0xad, 0x3d, 0x23, 0xf0, 0xc1, 0xa0, 0x03, 0x94, 0xcd, + 0xda, 0xb5, 0xe2, 0x37, 0xb5, 0x0a, 0xf8, 0x1c, 0x3d, 0x95, 0x90, 0x0a, 0xa9, 0x63, 0x1e, 0xbd, + 0x11, 0x5c, 0x4b, 0x31, 0x1e, 0x83, 0x24, 0x7b, 0x1d, 0xa7, 0xd7, 0x08, 0xbe, 0xb0, 0x65, 0x3c, + 0xa5, 0xeb, 0x10, 0xba, 0x89, 0x87, 0xbf, 0x47, 0x87, 0x8b, 0xf0, 0x5b, 0xae, 0x34, 0xe3, 0x21, + 0x90, 0xaa, 0x11, 0x7b, 0x66, 0xc5, 0x0e, 0xe9, 0x2a, 0x80, 0xae, 0x73, 0xf0, 0x0b, 0x54, 0x63, + 0xa1, 0x8e, 0x05, 0x27, 0x35, 0xc3, 0x7e, 0x62, 0xd9, 0xb5, 0xd7, 0x26, 0x4a, 0x6d, 0x36, 0xc7, + 0x49, 0x60, 0x4a, 0x70, 0x52, 0xbf, 0x8b, 0xa3, 0x26, 0x4a, 0x6d, 0x16, 0x5f, 0xa0, 0x86, 0x84, + 0x88, 0xc9, 0x61, 0xcc, 0x23, 0xb2, 0x6f, 0xc6, 0xf6, 0x7c, 0x79, 0x6c, 0xf9, 0x62, 0x97, 0xcf, + 0x4c, 0xe1, 0x13, 0x48, 0xe0, 0xe1, 0xd2, 0x4b, 0xd0, 0x39, 0x9b, 0x96, 0x42, 0xf8, 0x1d, 0xaa, + 0x4b, 0x18, 0xe7, 0x8b, 0x46, 0x1a, 0xbb, 0x6b, 0x36, 0xb3, 0x59, 0xbb, 0x4e, 0x0b, 0x1e, 0x9d, + 0x0b, 0xe0, 0x0e, 0xda, 0xe3, 0x42, 0x03, 0x41, 0xa6, 0x8f, 0x47, 0xd6, 0x77, 0xef, 0x07, 0xa1, + 0x81, 0x9a, 0x4c, 0x8e, 0xd0, 0x57, 0x29, 0x90, 0xe6, 0x5d, 0xc4, 0xc5, 0x55, 0x0a, 0xd4, 0x64, + 0x30, 0xa0, 0xd6, 0x10, 0x52, 0x09, 0x61, 0xae, 0xf8, 0x41, 0x4c, 0x64, 0x08, 0xe4, 0x91, 0x29, + 0xac, 0xbd, 0xa9, 0xb0, 0x62, 0x39, 0x0c, 0x2c, 0x20, 0x56, 0xae, 0xd5, 0x5f, 0x11, 0xa0, 0x6b, + 0x92, 0xf8, 0x37, 0x07, 0x91, 0x32, 0xf8, 0x5d, 0x2c, 0x95, 0x59, 0x4c, 0xa5, 0x59, 0x92, 0x92, + 0xc7, 0xc6, 0xef, 0xcb, 0xdd, 0x56, 0xde, 0x6c, 0x7b, 0xc7, 0x5a, 0x93, 0xfe, 0x16, 0x4d, 0xba, + 0xd5, 0x0d, 0xff, 0xea, 0xa0, 0xa3, 0x32, 0x79, 0xc6, 0x96, 0x2b, 0x79, 0xf2, 0x9f, 0x2b, 0x69, + 0xdb, 0x4a, 0x8e, 0xfa, 0x9b, 0x25, 0xe9, 0x36, 0x2f, 0xfc, 0x1a, 0x1d, 0x94, 0xa9, 0x37, 0x62, + 0xc2, 0x35, 0x39, 0xe8, 0x38, 0xbd, 0x6a, 0x70, 0x64, 0x25, 0x0f, 0xfa, 0x77, 0xd3, 0x74, 0x15, + 0xdf, 0xfd, 0xcb, 0x41, 0xc5, 0xff, 0xfd, 0x2c, 0x56, 0x1a, 0xff, 0xb4, 0x76, 0xa8, 0xbc, 0xdd, + 0x1a, 0xc9, 0xd9, 0xe6, 0x4c, 0xb5, 0xac, 0xf3, 0xfe, 0x3c, 0xb2, 0x74, 0xa4, 0x4e, 0x51, 0x35, + 0xd6, 0x90, 0x28, 0xf2, 0xa0, 0xf3, 0xb0, 0xd7, 0x7c, 0xd5, 0xb9, 0xef, 0x82, 0x04, 0x8f, 0xad, + 0x58, 0xf5, 0x6d, 0x4e, 0xa3, 0x05, 0xbb, 0xfb, 0xa7, 0x83, 0x9a, 0x4b, 0x17, 0x06, 0x3f, 0x47, + 0xd5, 0xd0, 0xf4, 0xee, 0x98, 0xde, 0x17, 0xa4, 0xa2, 0xe3, 0x22, 0x87, 0x27, 0xa8, 0x35, 0x66, + 0x4a, 0xbf, 0x1f, 0x28, 0x90, 0x53, 0x18, 0xfe, 0x9f, 0x3b, 0xb9, 0x58, 0xda, 0xb3, 0x15, 0x41, + 0xba, 0x66, 0x11, 0xbc, 0xbc, 0xbe, 0x75, 0x2b, 0x9f, 0x6f, 0xdd, 0xca, 0xcd, 0xad, 0x5b, 0xf9, + 0x25, 0x73, 0x9d, 0xeb, 0xcc, 0x75, 0x3e, 0x67, 0xae, 0x73, 0x93, 0xb9, 0xce, 0xdf, 0x99, 0xeb, + 0xfc, 0xfe, 0x8f, 0x5b, 0xf9, 0xb1, 0x6e, 0xdb, 0xfe, 0x37, 0x00, 0x00, 0xff, 0xff, 0xfc, 0xc9, + 0x09, 0x14, 0x84, 0x07, 0x00, 0x00, } func (m *Event) Marshal() (dAtA []byte, err error) { @@ -400,11 +398,6 @@ func (m *EventSeries) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l - i -= len(m.State) - copy(dAtA[i:], m.State) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.State))) - i-- - dAtA[i] = 0x1a { size, err := m.LastObservedTime.MarshalToSizedBuffer(dAtA[:i]) if err != nil { @@ -500,8 +493,6 @@ func (m *EventSeries) Size() (n int) { n += 1 + sovGenerated(uint64(m.Count)) l = m.LastObservedTime.Size() n += 1 + l + sovGenerated(uint64(l)) - l = len(m.State) - n += 1 + l + sovGenerated(uint64(l)) return n } @@ -558,7 +549,6 @@ func (this *EventSeries) String() string { s := strings.Join([]string{`&EventSeries{`, `Count:` + fmt.Sprintf("%v", this.Count) + `,`, `LastObservedTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastObservedTime), "MicroTime", "v1.MicroTime", 1), `&`, ``, 1) + `,`, - `State:` + fmt.Sprintf("%v", this.State) + `,`, `}`, }, "") return s @@ -1306,38 +1296,6 @@ func (m *EventSeries) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field State", wireType) - } - var stringLen uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - stringLen |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - intStringLen := int(stringLen) - if intStringLen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + intStringLen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - m.State = EventSeriesState(dAtA[iNdEx:postIndex]) - iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -1365,6 +1323,7 @@ func (m *EventSeries) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1396,10 +1355,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1420,55 +1377,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/events/v1beta1/generated.proto b/vendor/k8s.io/api/events/v1beta1/generated.proto index 58f5aa42..79bde87c 100644 --- a/vendor/k8s.io/api/events/v1beta1/generated.proto +++ b/vendor/k8s.io/api/events/v1beta1/generated.proto @@ -34,63 +34,68 @@ message Event { // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; - // Required. Time when this Event was first observed. + // eventTime is the time when this Event was first observed. It is required. optional k8s.io.apimachinery.pkg.apis.meta.v1.MicroTime eventTime = 2; - // Data about the Event series this event represents or nil if it's a singleton Event. + // series is data about the Event series this event represents or nil if it's a singleton Event. // +optional optional EventSeries series = 3; - // Name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. + // reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. + // This field cannot be empty for new Events. // +optional optional string reportingController = 4; - // ID of the controller instance, e.g. `kubelet-xyzf`. + // reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. + // This field cannot be empty for new Events and it can have at most 128 characters. // +optional optional string reportingInstance = 5; - // What action was taken/failed regarding to the regarding object. + // action is what action was taken/failed regarding to the regarding object. It is machine-readable. + // This field can have at most 128 characters. // +optional optional string action = 6; - // Why the action was taken. + // reason is why the action was taken. It is human-readable. + // This field can have at most 128 characters. + // +optional optional string reason = 7; - // The object this Event is about. In most cases it's an Object reporting controller implements. - // E.g. ReplicaSetController implements ReplicaSets and this event is emitted because + // regarding contains the object this Event is about. In most cases it's an Object reporting controller + // implements, e.g. ReplicaSetController implements ReplicaSets and this event is emitted because // it acts on some changes in a ReplicaSet object. // +optional optional k8s.io.api.core.v1.ObjectReference regarding = 8; - // Optional secondary object for more complex actions. E.g. when regarding object triggers + // related is the optional secondary object for more complex actions. E.g. when regarding object triggers // a creation or deletion of related object. // +optional optional k8s.io.api.core.v1.ObjectReference related = 9; - // Optional. A human-readable description of the status of this operation. + // note is a human-readable description of the status of this operation. // Maximal length of the note is 1kB, but libraries should be prepared to // handle values up to 64kB. // +optional optional string note = 10; - // Type of this event (Normal, Warning), new types could be added in the - // future. + // type is the type of this event (Normal, Warning), new types could be added in the future. + // It is machine-readable. // +optional optional string type = 11; - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedSource is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional optional k8s.io.api.core.v1.EventSource deprecatedSource = 12; - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedFirstTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.Time deprecatedFirstTimestamp = 13; - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedLastTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.Time deprecatedLastTimestamp = 14; - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional optional int32 deprecatedCount = 15; } @@ -102,21 +107,17 @@ message EventList { // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; - // Items is a list of schema objects. + // items is a list of schema objects. repeated Event items = 2; } // EventSeries contain information on series of events, i.e. thing that was/is happening // continuously for some time. message EventSeries { - // Number of occurrences in this series up to the last heartbeat time + // count is the number of occurrences in this series up to the last heartbeat time. optional int32 count = 1; - // Time when last Event from the series was seen before last heartbeat. + // lastObservedTime is the time when last Event from the series was seen before last heartbeat. optional k8s.io.apimachinery.pkg.apis.meta.v1.MicroTime lastObservedTime = 2; - - // Information whether this series is ongoing or finished. - // Deprecated. Planned removal for 1.18 - optional string state = 3; } diff --git a/vendor/k8s.io/api/events/v1beta1/types.go b/vendor/k8s.io/api/events/v1beta1/types.go index 0571fbb2..e2ed214b 100644 --- a/vendor/k8s.io/api/events/v1beta1/types.go +++ b/vendor/k8s.io/api/events/v1beta1/types.go @@ -23,6 +23,8 @@ import ( // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system. type Event struct { @@ -30,60 +32,65 @@ type Event struct { // +optional metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` - // Required. Time when this Event was first observed. + // eventTime is the time when this Event was first observed. It is required. EventTime metav1.MicroTime `json:"eventTime" protobuf:"bytes,2,opt,name=eventTime"` - // Data about the Event series this event represents or nil if it's a singleton Event. + // series is data about the Event series this event represents or nil if it's a singleton Event. // +optional Series *EventSeries `json:"series,omitempty" protobuf:"bytes,3,opt,name=series"` - // Name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. + // reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. + // This field cannot be empty for new Events. // +optional ReportingController string `json:"reportingController,omitempty" protobuf:"bytes,4,opt,name=reportingController"` - // ID of the controller instance, e.g. `kubelet-xyzf`. + // reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. + // This field cannot be empty for new Events and it can have at most 128 characters. // +optional ReportingInstance string `json:"reportingInstance,omitempty" protobuf:"bytes,5,opt,name=reportingInstance"` - // What action was taken/failed regarding to the regarding object. + // action is what action was taken/failed regarding to the regarding object. It is machine-readable. + // This field can have at most 128 characters. // +optional Action string `json:"action,omitempty" protobuf:"bytes,6,name=action"` - // Why the action was taken. + // reason is why the action was taken. It is human-readable. + // This field can have at most 128 characters. + // +optional Reason string `json:"reason,omitempty" protobuf:"bytes,7,name=reason"` - // The object this Event is about. In most cases it's an Object reporting controller implements. - // E.g. ReplicaSetController implements ReplicaSets and this event is emitted because + // regarding contains the object this Event is about. In most cases it's an Object reporting controller + // implements, e.g. ReplicaSetController implements ReplicaSets and this event is emitted because // it acts on some changes in a ReplicaSet object. // +optional Regarding corev1.ObjectReference `json:"regarding,omitempty" protobuf:"bytes,8,opt,name=regarding"` - // Optional secondary object for more complex actions. E.g. when regarding object triggers + // related is the optional secondary object for more complex actions. E.g. when regarding object triggers // a creation or deletion of related object. // +optional Related *corev1.ObjectReference `json:"related,omitempty" protobuf:"bytes,9,opt,name=related"` - // Optional. A human-readable description of the status of this operation. + // note is a human-readable description of the status of this operation. // Maximal length of the note is 1kB, but libraries should be prepared to // handle values up to 64kB. // +optional Note string `json:"note,omitempty" protobuf:"bytes,10,opt,name=note"` - // Type of this event (Normal, Warning), new types could be added in the - // future. + // type is the type of this event (Normal, Warning), new types could be added in the future. + // It is machine-readable. // +optional Type string `json:"type,omitempty" protobuf:"bytes,11,opt,name=type"` - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedSource is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional DeprecatedSource corev1.EventSource `json:"deprecatedSource,omitempty" protobuf:"bytes,12,opt,name=deprecatedSource"` - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedFirstTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional DeprecatedFirstTimestamp metav1.Time `json:"deprecatedFirstTimestamp,omitempty" protobuf:"bytes,13,opt,name=deprecatedFirstTimestamp"` - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedLastTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional DeprecatedLastTimestamp metav1.Time `json:"deprecatedLastTimestamp,omitempty" protobuf:"bytes,14,opt,name=deprecatedLastTimestamp"` - // Deprecated field assuring backward compatibility with core.v1 Event type + // deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type. // +optional DeprecatedCount int32 `json:"deprecatedCount,omitempty" protobuf:"varint,15,opt,name=deprecatedCount"` } @@ -91,24 +98,17 @@ type Event struct { // EventSeries contain information on series of events, i.e. thing that was/is happening // continuously for some time. type EventSeries struct { - // Number of occurrences in this series up to the last heartbeat time + // count is the number of occurrences in this series up to the last heartbeat time. Count int32 `json:"count" protobuf:"varint,1,opt,name=count"` - // Time when last Event from the series was seen before last heartbeat. + // lastObservedTime is the time when last Event from the series was seen before last heartbeat. LastObservedTime metav1.MicroTime `json:"lastObservedTime" protobuf:"bytes,2,opt,name=lastObservedTime"` - // Information whether this series is ongoing or finished. - // Deprecated. Planned removal for 1.18 - State EventSeriesState `json:"state" protobuf:"bytes,3,opt,name=state"` -} - -type EventSeriesState string -const ( - EventSeriesStateOngoing EventSeriesState = "Ongoing" - EventSeriesStateFinished EventSeriesState = "Finished" - EventSeriesStateUnknown EventSeriesState = "Unknown" -) + // +k8s:deprecated=state,protobuf=3 +} // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.8 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // EventList is a list of Event objects. type EventList struct { @@ -118,6 +118,6 @@ type EventList struct { // +optional metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` - // Items is a list of schema objects. + // items is a list of schema objects. Items []Event `json:"items" protobuf:"bytes,2,rep,name=items"` } diff --git a/vendor/k8s.io/api/events/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/events/v1beta1/types_swagger_doc_generated.go index 639daca6..8c987f89 100644 --- a/vendor/k8s.io/api/events/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/events/v1beta1/types_swagger_doc_generated.go @@ -29,20 +29,20 @@ package v1beta1 // AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. var map_Event = map[string]string{ "": "Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system.", - "eventTime": "Required. Time when this Event was first observed.", - "series": "Data about the Event series this event represents or nil if it's a singleton Event.", - "reportingController": "Name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`.", - "reportingInstance": "ID of the controller instance, e.g. `kubelet-xyzf`.", - "action": "What action was taken/failed regarding to the regarding object.", - "reason": "Why the action was taken.", - "regarding": "The object this Event is about. In most cases it's an Object reporting controller implements. E.g. ReplicaSetController implements ReplicaSets and this event is emitted because it acts on some changes in a ReplicaSet object.", - "related": "Optional secondary object for more complex actions. E.g. when regarding object triggers a creation or deletion of related object.", - "note": "Optional. A human-readable description of the status of this operation. Maximal length of the note is 1kB, but libraries should be prepared to handle values up to 64kB.", - "type": "Type of this event (Normal, Warning), new types could be added in the future.", - "deprecatedSource": "Deprecated field assuring backward compatibility with core.v1 Event type", - "deprecatedFirstTimestamp": "Deprecated field assuring backward compatibility with core.v1 Event type", - "deprecatedLastTimestamp": "Deprecated field assuring backward compatibility with core.v1 Event type", - "deprecatedCount": "Deprecated field assuring backward compatibility with core.v1 Event type", + "eventTime": "eventTime is the time when this Event was first observed. It is required.", + "series": "series is data about the Event series this event represents or nil if it's a singleton Event.", + "reportingController": "reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. This field cannot be empty for new Events.", + "reportingInstance": "reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. This field cannot be empty for new Events and it can have at most 128 characters.", + "action": "action is what action was taken/failed regarding to the regarding object. It is machine-readable. This field can have at most 128 characters.", + "reason": "reason is why the action was taken. It is human-readable. This field can have at most 128 characters.", + "regarding": "regarding contains the object this Event is about. In most cases it's an Object reporting controller implements, e.g. ReplicaSetController implements ReplicaSets and this event is emitted because it acts on some changes in a ReplicaSet object.", + "related": "related is the optional secondary object for more complex actions. E.g. when regarding object triggers a creation or deletion of related object.", + "note": "note is a human-readable description of the status of this operation. Maximal length of the note is 1kB, but libraries should be prepared to handle values up to 64kB.", + "type": "type is the type of this event (Normal, Warning), new types could be added in the future. It is machine-readable.", + "deprecatedSource": "deprecatedSource is the deprecated field assuring backward compatibility with core.v1 Event type.", + "deprecatedFirstTimestamp": "deprecatedFirstTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type.", + "deprecatedLastTimestamp": "deprecatedLastTimestamp is the deprecated field assuring backward compatibility with core.v1 Event type.", + "deprecatedCount": "deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type.", } func (Event) SwaggerDoc() map[string]string { @@ -52,7 +52,7 @@ func (Event) SwaggerDoc() map[string]string { var map_EventList = map[string]string{ "": "EventList is a list of Event objects.", "metadata": "Standard list metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", - "items": "Items is a list of schema objects.", + "items": "items is a list of schema objects.", } func (EventList) SwaggerDoc() map[string]string { @@ -61,9 +61,8 @@ func (EventList) SwaggerDoc() map[string]string { var map_EventSeries = map[string]string{ "": "EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time.", - "count": "Number of occurrences in this series up to the last heartbeat time", - "lastObservedTime": "Time when last Event from the series was seen before last heartbeat.", - "state": "Information whether this series is ongoing or finished. Deprecated. Planned removal for 1.18", + "count": "count is the number of occurrences in this series up to the last heartbeat time.", + "lastObservedTime": "lastObservedTime is the time when last Event from the series was seen before last heartbeat.", } func (EventSeries) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/events/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/events/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..2ab7b412 --- /dev/null +++ b/vendor/k8s.io/api/events/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,57 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Event) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Event) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Event) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *EventList) APILifecycleIntroduced() (major, minor int) { + return 1, 8 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *EventList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *EventList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} diff --git a/vendor/k8s.io/api/extensions/v1beta1/doc.go b/vendor/k8s.io/api/extensions/v1beta1/doc.go index fa799f30..c9af49d5 100644 --- a/vendor/k8s.io/api/extensions/v1beta1/doc.go +++ b/vendor/k8s.io/api/extensions/v1beta1/doc.go @@ -17,5 +17,6 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v1beta1 // import "k8s.io/api/extensions/v1beta1" diff --git a/vendor/k8s.io/api/extensions/v1beta1/generated.pb.go b/vendor/k8s.io/api/extensions/v1beta1/generated.pb.go index 65b47eab..bd37f432 100644 --- a/vendor/k8s.io/api/extensions/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/extensions/v1beta1/generated.pb.go @@ -47,7 +47,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *AllowedCSIDriver) Reset() { *m = AllowedCSIDriver{} } func (*AllowedCSIDriver) ProtoMessage() {} @@ -1309,38 +1309,10 @@ func (m *ReplicaSetStatus) XXX_DiscardUnknown() { var xxx_messageInfo_ReplicaSetStatus proto.InternalMessageInfo -func (m *ReplicationControllerDummy) Reset() { *m = ReplicationControllerDummy{} } -func (*ReplicationControllerDummy) ProtoMessage() {} -func (*ReplicationControllerDummy) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{45} -} -func (m *ReplicationControllerDummy) XXX_Unmarshal(b []byte) error { - return m.Unmarshal(b) -} -func (m *ReplicationControllerDummy) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - b = b[:cap(b)] - n, err := m.MarshalToSizedBuffer(b) - if err != nil { - return nil, err - } - return b[:n], nil -} -func (m *ReplicationControllerDummy) XXX_Merge(src proto.Message) { - xxx_messageInfo_ReplicationControllerDummy.Merge(m, src) -} -func (m *ReplicationControllerDummy) XXX_Size() int { - return m.Size() -} -func (m *ReplicationControllerDummy) XXX_DiscardUnknown() { - xxx_messageInfo_ReplicationControllerDummy.DiscardUnknown(m) -} - -var xxx_messageInfo_ReplicationControllerDummy proto.InternalMessageInfo - func (m *RollbackConfig) Reset() { *m = RollbackConfig{} } func (*RollbackConfig) ProtoMessage() {} func (*RollbackConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{46} + return fileDescriptor_cdc93917efc28165, []int{45} } func (m *RollbackConfig) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1368,7 +1340,7 @@ var xxx_messageInfo_RollbackConfig proto.InternalMessageInfo func (m *RollingUpdateDaemonSet) Reset() { *m = RollingUpdateDaemonSet{} } func (*RollingUpdateDaemonSet) ProtoMessage() {} func (*RollingUpdateDaemonSet) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{47} + return fileDescriptor_cdc93917efc28165, []int{46} } func (m *RollingUpdateDaemonSet) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1396,7 +1368,7 @@ var xxx_messageInfo_RollingUpdateDaemonSet proto.InternalMessageInfo func (m *RollingUpdateDeployment) Reset() { *m = RollingUpdateDeployment{} } func (*RollingUpdateDeployment) ProtoMessage() {} func (*RollingUpdateDeployment) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{48} + return fileDescriptor_cdc93917efc28165, []int{47} } func (m *RollingUpdateDeployment) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1424,7 +1396,7 @@ var xxx_messageInfo_RollingUpdateDeployment proto.InternalMessageInfo func (m *RunAsGroupStrategyOptions) Reset() { *m = RunAsGroupStrategyOptions{} } func (*RunAsGroupStrategyOptions) ProtoMessage() {} func (*RunAsGroupStrategyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{49} + return fileDescriptor_cdc93917efc28165, []int{48} } func (m *RunAsGroupStrategyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1452,7 +1424,7 @@ var xxx_messageInfo_RunAsGroupStrategyOptions proto.InternalMessageInfo func (m *RunAsUserStrategyOptions) Reset() { *m = RunAsUserStrategyOptions{} } func (*RunAsUserStrategyOptions) ProtoMessage() {} func (*RunAsUserStrategyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{50} + return fileDescriptor_cdc93917efc28165, []int{49} } func (m *RunAsUserStrategyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1480,7 +1452,7 @@ var xxx_messageInfo_RunAsUserStrategyOptions proto.InternalMessageInfo func (m *RuntimeClassStrategyOptions) Reset() { *m = RuntimeClassStrategyOptions{} } func (*RuntimeClassStrategyOptions) ProtoMessage() {} func (*RuntimeClassStrategyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{51} + return fileDescriptor_cdc93917efc28165, []int{50} } func (m *RuntimeClassStrategyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1508,7 +1480,7 @@ var xxx_messageInfo_RuntimeClassStrategyOptions proto.InternalMessageInfo func (m *SELinuxStrategyOptions) Reset() { *m = SELinuxStrategyOptions{} } func (*SELinuxStrategyOptions) ProtoMessage() {} func (*SELinuxStrategyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{52} + return fileDescriptor_cdc93917efc28165, []int{51} } func (m *SELinuxStrategyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1536,7 +1508,7 @@ var xxx_messageInfo_SELinuxStrategyOptions proto.InternalMessageInfo func (m *Scale) Reset() { *m = Scale{} } func (*Scale) ProtoMessage() {} func (*Scale) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{53} + return fileDescriptor_cdc93917efc28165, []int{52} } func (m *Scale) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1564,7 +1536,7 @@ var xxx_messageInfo_Scale proto.InternalMessageInfo func (m *ScaleSpec) Reset() { *m = ScaleSpec{} } func (*ScaleSpec) ProtoMessage() {} func (*ScaleSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{54} + return fileDescriptor_cdc93917efc28165, []int{53} } func (m *ScaleSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1592,7 +1564,7 @@ var xxx_messageInfo_ScaleSpec proto.InternalMessageInfo func (m *ScaleStatus) Reset() { *m = ScaleStatus{} } func (*ScaleStatus) ProtoMessage() {} func (*ScaleStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{55} + return fileDescriptor_cdc93917efc28165, []int{54} } func (m *ScaleStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1620,7 +1592,7 @@ var xxx_messageInfo_ScaleStatus proto.InternalMessageInfo func (m *SupplementalGroupsStrategyOptions) Reset() { *m = SupplementalGroupsStrategyOptions{} } func (*SupplementalGroupsStrategyOptions) ProtoMessage() {} func (*SupplementalGroupsStrategyOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cdc93917efc28165, []int{56} + return fileDescriptor_cdc93917efc28165, []int{55} } func (m *SupplementalGroupsStrategyOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1692,7 +1664,6 @@ func init() { proto.RegisterType((*ReplicaSetList)(nil), "k8s.io.api.extensions.v1beta1.ReplicaSetList") proto.RegisterType((*ReplicaSetSpec)(nil), "k8s.io.api.extensions.v1beta1.ReplicaSetSpec") proto.RegisterType((*ReplicaSetStatus)(nil), "k8s.io.api.extensions.v1beta1.ReplicaSetStatus") - proto.RegisterType((*ReplicationControllerDummy)(nil), "k8s.io.api.extensions.v1beta1.ReplicationControllerDummy") proto.RegisterType((*RollbackConfig)(nil), "k8s.io.api.extensions.v1beta1.RollbackConfig") proto.RegisterType((*RollingUpdateDaemonSet)(nil), "k8s.io.api.extensions.v1beta1.RollingUpdateDaemonSet") proto.RegisterType((*RollingUpdateDeployment)(nil), "k8s.io.api.extensions.v1beta1.RollingUpdateDeployment") @@ -1712,238 +1683,241 @@ func init() { } var fileDescriptor_cdc93917efc28165 = []byte{ - // 3684 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xec, 0x5b, 0x4f, 0x6c, 0x1b, 0x47, - 0x77, 0xf7, 0x92, 0x94, 0x48, 0x3d, 0xfd, 0x1f, 0xc9, 0x12, 0x3f, 0x3b, 0x16, 0xfd, 0x6d, 0x00, - 0xd7, 0x49, 0x6d, 0x32, 0x76, 0x6c, 0x7f, 0xae, 0x8d, 0x7e, 0x89, 0x28, 0x59, 0xb6, 0x52, 0xfd, - 0x61, 0x86, 0x92, 0x1b, 0x04, 0x4d, 0x9a, 0x15, 0x39, 0xa2, 0xd6, 0x5a, 0xee, 0x6e, 0x76, 0x87, - 0x8a, 0x08, 0xf4, 0xd0, 0x43, 0x51, 0xa0, 0x40, 0x8b, 0xf6, 0x92, 0xb6, 0xc7, 0x06, 0x05, 0x7a, - 0x6a, 0xd1, 0xde, 0xda, 0x43, 0x10, 0xa0, 0x40, 0x0a, 0x18, 0x45, 0x5a, 0xe4, 0xd6, 0x9c, 0x84, - 0x46, 0x39, 0x15, 0x3d, 0xf5, 0x56, 0xf8, 0x50, 0x14, 0x33, 0x3b, 0xfb, 0x7f, 0x57, 0x5c, 0x29, - 0xb6, 0xd0, 0x00, 0xbd, 0x89, 0xf3, 0xde, 0xfb, 0xbd, 0x37, 0x33, 0x6f, 0xde, 0x7b, 0x33, 0xfb, - 0x04, 0x2b, 0xfb, 0xf7, 0xed, 0xaa, 0x6a, 0xd4, 0xf6, 0x7b, 0x3b, 0xc4, 0xd2, 0x09, 0x25, 0x76, - 0xed, 0x80, 0xe8, 0x6d, 0xc3, 0xaa, 0x09, 0x82, 0x62, 0xaa, 0x35, 0x72, 0x48, 0x89, 0x6e, 0xab, - 0x86, 0x6e, 0xd7, 0x0e, 0x6e, 0xed, 0x10, 0xaa, 0xdc, 0xaa, 0x75, 0x88, 0x4e, 0x2c, 0x85, 0x92, - 0x76, 0xd5, 0xb4, 0x0c, 0x6a, 0xa0, 0x2b, 0x0e, 0x7b, 0x55, 0x31, 0xd5, 0xaa, 0xcf, 0x5e, 0x15, - 0xec, 0x97, 0x6e, 0x76, 0x54, 0xba, 0xd7, 0xdb, 0xa9, 0xb6, 0x8c, 0x6e, 0xad, 0x63, 0x74, 0x8c, - 0x1a, 0x97, 0xda, 0xe9, 0xed, 0xf2, 0x5f, 0xfc, 0x07, 0xff, 0xcb, 0x41, 0xbb, 0x24, 0x07, 0x94, - 0xb7, 0x0c, 0x8b, 0xd4, 0x0e, 0x62, 0x1a, 0x2f, 0xdd, 0xf1, 0x79, 0xba, 0x4a, 0x6b, 0x4f, 0xd5, - 0x89, 0xd5, 0xaf, 0x99, 0xfb, 0x1d, 0x36, 0x60, 0xd7, 0xba, 0x84, 0x2a, 0x49, 0x52, 0xb5, 0x34, - 0x29, 0xab, 0xa7, 0x53, 0xb5, 0x4b, 0x62, 0x02, 0xf7, 0x06, 0x09, 0xd8, 0xad, 0x3d, 0xd2, 0x55, - 0x62, 0x72, 0x6f, 0xa7, 0xc9, 0xf5, 0xa8, 0xaa, 0xd5, 0x54, 0x9d, 0xda, 0xd4, 0x8a, 0x0a, 0xc9, - 0x77, 0x60, 0x6a, 0x51, 0xd3, 0x8c, 0xcf, 0x48, 0x7b, 0xa9, 0xb9, 0xba, 0x6c, 0xa9, 0x07, 0xc4, - 0x42, 0x57, 0xa1, 0xa0, 0x2b, 0x5d, 0x52, 0x96, 0xae, 0x4a, 0xd7, 0x47, 0xea, 0x63, 0xcf, 0x8f, - 0x2a, 0x17, 0x8e, 0x8f, 0x2a, 0x85, 0x0d, 0xa5, 0x4b, 0x30, 0xa7, 0xc8, 0x0f, 0x61, 0x5a, 0x48, - 0xad, 0x68, 0xe4, 0xf0, 0xa9, 0xa1, 0xf5, 0xba, 0x04, 0x5d, 0x83, 0xe1, 0x36, 0x07, 0x10, 0x82, - 0x13, 0x42, 0x70, 0xd8, 0x81, 0xc5, 0x82, 0x2a, 0xdb, 0x30, 0x29, 0x84, 0x9f, 0x18, 0x36, 0x6d, - 0x28, 0x74, 0x0f, 0xdd, 0x06, 0x30, 0x15, 0xba, 0xd7, 0xb0, 0xc8, 0xae, 0x7a, 0x28, 0xc4, 0x91, - 0x10, 0x87, 0x86, 0x47, 0xc1, 0x01, 0x2e, 0x74, 0x03, 0x4a, 0x16, 0x51, 0xda, 0x9b, 0xba, 0xd6, - 0x2f, 0xe7, 0xae, 0x4a, 0xd7, 0x4b, 0xf5, 0x29, 0x21, 0x51, 0xc2, 0x62, 0x1c, 0x7b, 0x1c, 0xf2, - 0xe7, 0x39, 0x18, 0x59, 0x56, 0x48, 0xd7, 0xd0, 0x9b, 0x84, 0xa2, 0x4f, 0xa0, 0xc4, 0xb6, 0xab, - 0xad, 0x50, 0x85, 0x6b, 0x1b, 0xbd, 0xfd, 0x56, 0xd5, 0x77, 0x27, 0x6f, 0xf5, 0xaa, 0xe6, 0x7e, - 0x87, 0x0d, 0xd8, 0x55, 0xc6, 0x5d, 0x3d, 0xb8, 0x55, 0xdd, 0xdc, 0x79, 0x46, 0x5a, 0x74, 0x9d, - 0x50, 0xc5, 0xb7, 0xcf, 0x1f, 0xc3, 0x1e, 0x2a, 0xda, 0x80, 0x82, 0x6d, 0x92, 0x16, 0xb7, 0x6c, - 0xf4, 0xf6, 0x8d, 0xea, 0x89, 0xce, 0x5a, 0xf5, 0x2c, 0x6b, 0x9a, 0xa4, 0xe5, 0xaf, 0x38, 0xfb, - 0x85, 0x39, 0x0e, 0x7a, 0x0a, 0xc3, 0x36, 0x55, 0x68, 0xcf, 0x2e, 0xe7, 0x39, 0x62, 0x35, 0x33, - 0x22, 0x97, 0xf2, 0x37, 0xc3, 0xf9, 0x8d, 0x05, 0x9a, 0xfc, 0x1f, 0x39, 0x40, 0x1e, 0xef, 0x92, - 0xa1, 0xb7, 0x55, 0xaa, 0x1a, 0x3a, 0x7a, 0x00, 0x05, 0xda, 0x37, 0x5d, 0x17, 0xb8, 0xe6, 0x1a, - 0xb4, 0xd5, 0x37, 0xc9, 0x8b, 0xa3, 0xca, 0x5c, 0x5c, 0x82, 0x51, 0x30, 0x97, 0x41, 0x6b, 0x9e, - 0xa9, 0x39, 0x2e, 0x7d, 0x27, 0xac, 0xfa, 0xc5, 0x51, 0x25, 0xe1, 0xb0, 0x55, 0x3d, 0xa4, 0xb0, - 0x81, 0xe8, 0x00, 0x90, 0xa6, 0xd8, 0x74, 0xcb, 0x52, 0x74, 0xdb, 0xd1, 0xa4, 0x76, 0x89, 0x58, - 0x84, 0x37, 0xb3, 0x6d, 0x1a, 0x93, 0xa8, 0x5f, 0x12, 0x56, 0xa0, 0xb5, 0x18, 0x1a, 0x4e, 0xd0, - 0xc0, 0xbc, 0xd9, 0x22, 0x8a, 0x6d, 0xe8, 0xe5, 0x42, 0xd8, 0x9b, 0x31, 0x1f, 0xc5, 0x82, 0x8a, - 0xde, 0x80, 0x62, 0x97, 0xd8, 0xb6, 0xd2, 0x21, 0xe5, 0x21, 0xce, 0x38, 0x29, 0x18, 0x8b, 0xeb, - 0xce, 0x30, 0x76, 0xe9, 0xf2, 0x97, 0x12, 0x8c, 0x7b, 0x2b, 0xb7, 0xa6, 0xda, 0x14, 0xfd, 0x56, - 0xcc, 0x0f, 0xab, 0xd9, 0xa6, 0xc4, 0xa4, 0xb9, 0x17, 0x7a, 0x3e, 0xef, 0x8e, 0x04, 0x7c, 0x70, - 0x1d, 0x86, 0x54, 0x4a, 0xba, 0x6c, 0x1f, 0xf2, 0xd7, 0x47, 0x6f, 0x5f, 0xcf, 0xea, 0x32, 0xf5, - 0x71, 0x01, 0x3a, 0xb4, 0xca, 0xc4, 0xb1, 0x83, 0x22, 0xff, 0x69, 0x21, 0x60, 0x3e, 0x73, 0x4d, - 0xf4, 0x11, 0x94, 0x6c, 0xa2, 0x91, 0x16, 0x35, 0x2c, 0x61, 0xfe, 0xdb, 0x19, 0xcd, 0x57, 0x76, - 0x88, 0xd6, 0x14, 0xa2, 0xf5, 0x31, 0x66, 0xbf, 0xfb, 0x0b, 0x7b, 0x90, 0xe8, 0x7d, 0x28, 0x51, - 0xd2, 0x35, 0x35, 0x85, 0x12, 0x71, 0x8e, 0x5e, 0x0f, 0x4e, 0x81, 0x79, 0x0e, 0x03, 0x6b, 0x18, - 0xed, 0x2d, 0xc1, 0xc6, 0x8f, 0x8f, 0xb7, 0x24, 0xee, 0x28, 0xf6, 0x60, 0xd0, 0x01, 0x4c, 0xf4, - 0xcc, 0x36, 0xe3, 0xa4, 0x2c, 0x0a, 0x76, 0xfa, 0xc2, 0x93, 0xee, 0x65, 0x5d, 0x9b, 0xed, 0x90, - 0x74, 0x7d, 0x4e, 0xe8, 0x9a, 0x08, 0x8f, 0xe3, 0x88, 0x16, 0xb4, 0x08, 0x93, 0x5d, 0x55, 0x67, - 0x71, 0xa9, 0xdf, 0x24, 0x2d, 0x43, 0x6f, 0xdb, 0xdc, 0xad, 0x86, 0xea, 0xf3, 0x02, 0x60, 0x72, - 0x3d, 0x4c, 0xc6, 0x51, 0x7e, 0xf4, 0x1e, 0x20, 0x77, 0x1a, 0x8f, 0x9d, 0x20, 0xae, 0x1a, 0x3a, - 0xf7, 0xb9, 0xbc, 0xef, 0xdc, 0x5b, 0x31, 0x0e, 0x9c, 0x20, 0x85, 0xd6, 0x60, 0xd6, 0x22, 0x07, - 0x2a, 0x9b, 0xe3, 0x13, 0xd5, 0xa6, 0x86, 0xd5, 0x5f, 0x53, 0xbb, 0x2a, 0x2d, 0x0f, 0x73, 0x9b, - 0xca, 0xc7, 0x47, 0x95, 0x59, 0x9c, 0x40, 0xc7, 0x89, 0x52, 0xf2, 0x9f, 0x0d, 0xc3, 0x64, 0x24, - 0xde, 0xa0, 0xa7, 0x30, 0xd7, 0xea, 0x59, 0x16, 0xd1, 0xe9, 0x46, 0xaf, 0xbb, 0x43, 0xac, 0x66, - 0x6b, 0x8f, 0xb4, 0x7b, 0x1a, 0x69, 0x73, 0x47, 0x19, 0xaa, 0x2f, 0x08, 0x8b, 0xe7, 0x96, 0x12, - 0xb9, 0x70, 0x8a, 0x34, 0x5b, 0x05, 0x9d, 0x0f, 0xad, 0xab, 0xb6, 0xed, 0x61, 0xe6, 0x38, 0xa6, - 0xb7, 0x0a, 0x1b, 0x31, 0x0e, 0x9c, 0x20, 0xc5, 0x6c, 0x6c, 0x13, 0x5b, 0xb5, 0x48, 0x3b, 0x6a, - 0x63, 0x3e, 0x6c, 0xe3, 0x72, 0x22, 0x17, 0x4e, 0x91, 0x46, 0x77, 0x61, 0xd4, 0xd1, 0xc6, 0xf7, - 0x4f, 0x6c, 0xf4, 0x8c, 0x00, 0x1b, 0xdd, 0xf0, 0x49, 0x38, 0xc8, 0xc7, 0xa6, 0x66, 0xec, 0xd8, - 0xc4, 0x3a, 0x20, 0xed, 0xf4, 0x0d, 0xde, 0x8c, 0x71, 0xe0, 0x04, 0x29, 0x36, 0x35, 0xc7, 0x03, - 0x63, 0x53, 0x1b, 0x0e, 0x4f, 0x6d, 0x3b, 0x91, 0x0b, 0xa7, 0x48, 0x33, 0x3f, 0x76, 0x4c, 0x5e, - 0x3c, 0x50, 0x54, 0x4d, 0xd9, 0xd1, 0x48, 0xb9, 0x18, 0xf6, 0xe3, 0x8d, 0x30, 0x19, 0x47, 0xf9, - 0xd1, 0x63, 0x98, 0x76, 0x86, 0xb6, 0x75, 0xc5, 0x03, 0x29, 0x71, 0x90, 0x9f, 0x09, 0x90, 0xe9, - 0x8d, 0x28, 0x03, 0x8e, 0xcb, 0xa0, 0x07, 0x30, 0xd1, 0x32, 0x34, 0x8d, 0xfb, 0xe3, 0x92, 0xd1, - 0xd3, 0x69, 0x79, 0x84, 0xa3, 0x20, 0x76, 0x1e, 0x97, 0x42, 0x14, 0x1c, 0xe1, 0x44, 0x04, 0xa0, - 0xe5, 0x26, 0x1c, 0xbb, 0x0c, 0x3c, 0x3e, 0xde, 0xca, 0x1a, 0x03, 0xbc, 0x54, 0xe5, 0xd7, 0x00, - 0xde, 0x90, 0x8d, 0x03, 0xc0, 0xf2, 0x3f, 0x4b, 0x30, 0x9f, 0x12, 0x3a, 0xd0, 0x3b, 0xa1, 0x14, - 0xfb, 0xab, 0x91, 0x14, 0x7b, 0x39, 0x45, 0x2c, 0x90, 0x67, 0x75, 0x18, 0xb7, 0xd8, 0xac, 0xf4, - 0x8e, 0xc3, 0x22, 0x62, 0xe4, 0xdd, 0x01, 0xd3, 0xc0, 0x41, 0x19, 0x3f, 0xe6, 0x4f, 0x1f, 0x1f, - 0x55, 0xc6, 0x43, 0x34, 0x1c, 0x86, 0x97, 0xff, 0x3c, 0x07, 0xb0, 0x4c, 0x4c, 0xcd, 0xe8, 0x77, - 0x89, 0x7e, 0x1e, 0x35, 0xd4, 0x66, 0xa8, 0x86, 0xba, 0x39, 0x68, 0x7b, 0x3c, 0xd3, 0x52, 0x8b, - 0xa8, 0xdf, 0x8c, 0x14, 0x51, 0xb5, 0xec, 0x90, 0x27, 0x57, 0x51, 0xff, 0x96, 0x87, 0x19, 0x9f, - 0xd9, 0x2f, 0xa3, 0x1e, 0x86, 0xf6, 0xf8, 0x57, 0x22, 0x7b, 0x3c, 0x9f, 0x20, 0xf2, 0xca, 0xea, - 0xa8, 0x67, 0x30, 0xc1, 0xaa, 0x1c, 0x67, 0x2f, 0x79, 0x0d, 0x35, 0x7c, 0xea, 0x1a, 0xca, 0xcb, - 0x76, 0x6b, 0x21, 0x24, 0x1c, 0x41, 0x4e, 0xa9, 0xd9, 0x8a, 0x3f, 0xc5, 0x9a, 0xed, 0x2b, 0x09, - 0x26, 0xfc, 0x6d, 0x3a, 0x87, 0xa2, 0x6d, 0x23, 0x5c, 0xb4, 0xbd, 0x91, 0xd9, 0x45, 0x53, 0xaa, - 0xb6, 0xff, 0x66, 0x05, 0xbe, 0xc7, 0xc4, 0x0e, 0xf8, 0x8e, 0xd2, 0xda, 0x1f, 0x7c, 0xc7, 0x43, - 0x9f, 0x4b, 0x80, 0x44, 0x16, 0x58, 0xd4, 0x75, 0x83, 0x2a, 0x4e, 0xac, 0x74, 0xcc, 0x5a, 0xcd, - 0x6c, 0x96, 0xab, 0xb1, 0xba, 0x1d, 0xc3, 0x7a, 0xa4, 0x53, 0xab, 0xef, 0x6f, 0x72, 0x9c, 0x01, - 0x27, 0x18, 0x80, 0x14, 0x00, 0x4b, 0x60, 0x6e, 0x19, 0xe2, 0x20, 0xdf, 0xcc, 0x10, 0xf3, 0x98, - 0xc0, 0x92, 0xa1, 0xef, 0xaa, 0x1d, 0x3f, 0xec, 0x60, 0x0f, 0x08, 0x07, 0x40, 0x2f, 0x3d, 0x82, - 0xf9, 0x14, 0x6b, 0xd1, 0x14, 0xe4, 0xf7, 0x49, 0xdf, 0x59, 0x36, 0xcc, 0xfe, 0x44, 0xb3, 0x30, - 0x74, 0xa0, 0x68, 0x3d, 0x27, 0xfc, 0x8e, 0x60, 0xe7, 0xc7, 0x83, 0xdc, 0x7d, 0x49, 0xfe, 0x72, - 0x28, 0xe8, 0x3b, 0xbc, 0x62, 0xbe, 0xce, 0x2e, 0xad, 0xa6, 0xa6, 0xb6, 0x14, 0x5b, 0x14, 0x42, - 0x63, 0xce, 0x85, 0xd5, 0x19, 0xc3, 0x1e, 0x35, 0x54, 0x5b, 0xe7, 0x5e, 0x6d, 0x6d, 0x9d, 0x7f, - 0x39, 0xb5, 0xf5, 0x6f, 0x43, 0xc9, 0x76, 0xab, 0xea, 0x02, 0x87, 0xbc, 0x75, 0x8a, 0xf8, 0x2a, - 0x0a, 0x6a, 0x4f, 0x81, 0x57, 0x4a, 0x7b, 0xa0, 0x49, 0x45, 0xf4, 0xd0, 0x29, 0x8b, 0xe8, 0x97, - 0x5a, 0xf8, 0xb2, 0x78, 0x63, 0x2a, 0x3d, 0x9b, 0xb4, 0x79, 0x6c, 0x2b, 0xf9, 0xf1, 0xa6, 0xc1, - 0x47, 0xb1, 0xa0, 0xa2, 0x8f, 0x42, 0x2e, 0x5b, 0x3a, 0x8b, 0xcb, 0x4e, 0xa4, 0xbb, 0x2b, 0xda, - 0x86, 0x79, 0xd3, 0x32, 0x3a, 0x16, 0xb1, 0xed, 0x65, 0xa2, 0xb4, 0x35, 0x55, 0x27, 0xee, 0xfa, - 0x38, 0x15, 0xd1, 0xe5, 0xe3, 0xa3, 0xca, 0x7c, 0x23, 0x99, 0x05, 0xa7, 0xc9, 0xca, 0xcf, 0x0b, - 0x30, 0x15, 0xcd, 0x80, 0x29, 0x45, 0xaa, 0x74, 0xa6, 0x22, 0xf5, 0x46, 0xe0, 0x30, 0x38, 0x15, - 0x7c, 0xe0, 0x05, 0x27, 0x76, 0x20, 0x16, 0x61, 0x52, 0x44, 0x03, 0x97, 0x28, 0xca, 0x74, 0x6f, - 0xf7, 0xb7, 0xc3, 0x64, 0x1c, 0xe5, 0x47, 0x0f, 0x61, 0xdc, 0xe2, 0x75, 0xb7, 0x0b, 0xe0, 0xd4, - 0xae, 0x17, 0x05, 0xc0, 0x38, 0x0e, 0x12, 0x71, 0x98, 0x97, 0xd5, 0xad, 0x7e, 0x39, 0xea, 0x02, - 0x14, 0xc2, 0x75, 0xeb, 0x62, 0x94, 0x01, 0xc7, 0x65, 0xd0, 0x3a, 0xcc, 0xf4, 0xf4, 0x38, 0x94, - 0xe3, 0xca, 0x97, 0x05, 0xd4, 0xcc, 0x76, 0x9c, 0x05, 0x27, 0xc9, 0xa1, 0xdd, 0x50, 0x29, 0x3b, - 0xcc, 0xc3, 0xf3, 0xed, 0xcc, 0x07, 0x2f, 0x73, 0x2d, 0x9b, 0x50, 0x6e, 0x97, 0xb2, 0x96, 0xdb, - 0xf2, 0x3f, 0x4a, 0xc1, 0x24, 0xe4, 0x95, 0xc0, 0x83, 0x5e, 0x99, 0x62, 0x12, 0x81, 0xea, 0xc8, - 0x48, 0xae, 0x7e, 0xef, 0x9d, 0xaa, 0xfa, 0xf5, 0x93, 0xe7, 0xe0, 0xf2, 0xf7, 0x0b, 0x09, 0xe6, - 0x56, 0x9a, 0x8f, 0x2d, 0xa3, 0x67, 0xba, 0xe6, 0x6c, 0x9a, 0xce, 0xd2, 0xfc, 0x02, 0x0a, 0x56, - 0x4f, 0x73, 0xe7, 0xf1, 0xba, 0x3b, 0x0f, 0xdc, 0xd3, 0xd8, 0x3c, 0x66, 0x22, 0x52, 0xce, 0x24, - 0x98, 0x00, 0xda, 0x80, 0x61, 0x4b, 0xd1, 0x3b, 0xc4, 0x4d, 0xab, 0xd7, 0x06, 0x58, 0xbf, 0xba, - 0x8c, 0x19, 0x7b, 0xa0, 0xb0, 0xe1, 0xd2, 0x58, 0xa0, 0xc8, 0x7f, 0x24, 0xc1, 0xe4, 0x93, 0xad, - 0xad, 0xc6, 0xaa, 0xce, 0x4f, 0x34, 0x7f, 0x5b, 0xbd, 0x0a, 0x05, 0x53, 0xa1, 0x7b, 0xd1, 0x4c, - 0xcf, 0x68, 0x98, 0x53, 0xd0, 0x07, 0x50, 0x64, 0x91, 0x84, 0xe8, 0xed, 0x8c, 0xa5, 0xb6, 0x80, - 0xaf, 0x3b, 0x42, 0x7e, 0xf5, 0x24, 0x06, 0xb0, 0x0b, 0x27, 0xef, 0xc3, 0x6c, 0xc0, 0x1c, 0xb6, - 0x1e, 0x4f, 0x59, 0x76, 0x44, 0x4d, 0x18, 0x62, 0x9a, 0x59, 0x0e, 0xcc, 0x67, 0x78, 0xcc, 0x8c, - 0x4c, 0xc9, 0xaf, 0x74, 0xd8, 0x2f, 0x1b, 0x3b, 0x58, 0xf2, 0x3a, 0x8c, 0xf3, 0x07, 0x65, 0xc3, - 0xa2, 0x7c, 0x59, 0xd0, 0x15, 0xc8, 0x77, 0x55, 0x5d, 0xe4, 0xd9, 0x51, 0x21, 0x93, 0x67, 0x39, - 0x82, 0x8d, 0x73, 0xb2, 0x72, 0x28, 0x22, 0x8f, 0x4f, 0x56, 0x0e, 0x31, 0x1b, 0x97, 0x1f, 0x43, - 0x51, 0x2c, 0x77, 0x10, 0x28, 0x7f, 0x32, 0x50, 0x3e, 0x01, 0x68, 0x13, 0x8a, 0xab, 0x8d, 0xba, - 0x66, 0x38, 0x55, 0x57, 0x4b, 0x6d, 0x5b, 0xd1, 0xbd, 0x58, 0x5a, 0x5d, 0xc6, 0x98, 0x53, 0x90, - 0x0c, 0xc3, 0xe4, 0xb0, 0x45, 0x4c, 0xca, 0x3d, 0x62, 0xa4, 0x0e, 0x6c, 0x97, 0x1f, 0xf1, 0x11, - 0x2c, 0x28, 0xf2, 0x1f, 0xe7, 0xa0, 0x28, 0x96, 0xe3, 0x1c, 0x6e, 0x61, 0x6b, 0xa1, 0x5b, 0xd8, - 0x9b, 0xd9, 0x5c, 0x23, 0xf5, 0x0a, 0xb6, 0x15, 0xb9, 0x82, 0xdd, 0xc8, 0x88, 0x77, 0xf2, 0xfd, - 0xeb, 0xef, 0x24, 0x98, 0x08, 0x3b, 0x25, 0xba, 0x0b, 0xa3, 0x2c, 0xe1, 0xa8, 0x2d, 0xb2, 0xe1, - 0xd7, 0xb9, 0xde, 0x23, 0x4c, 0xd3, 0x27, 0xe1, 0x20, 0x1f, 0xea, 0x78, 0x62, 0xcc, 0x8f, 0xc4, - 0xa4, 0xd3, 0x97, 0xb4, 0x47, 0x55, 0xad, 0xea, 0x7c, 0x5a, 0xa9, 0xae, 0xea, 0x74, 0xd3, 0x6a, - 0x52, 0x4b, 0xd5, 0x3b, 0x31, 0x45, 0xdc, 0x29, 0x83, 0xc8, 0xf2, 0x3f, 0x48, 0x30, 0x2a, 0x4c, - 0x3e, 0x87, 0x5b, 0xc5, 0x6f, 0x84, 0x6f, 0x15, 0xd7, 0x32, 0x1e, 0xf0, 0xe4, 0x2b, 0xc5, 0x5f, - 0xf9, 0xa6, 0xb3, 0x23, 0xcd, 0xbc, 0x7a, 0xcf, 0xb0, 0x69, 0xd4, 0xab, 0xd9, 0x61, 0xc4, 0x9c, - 0x82, 0x7a, 0x30, 0xa5, 0x46, 0x62, 0x80, 0x58, 0xda, 0x5a, 0x36, 0x4b, 0x3c, 0xb1, 0x7a, 0x59, - 0xc0, 0x4f, 0x45, 0x29, 0x38, 0xa6, 0x42, 0x26, 0x10, 0xe3, 0x42, 0xef, 0x43, 0x61, 0x8f, 0x52, - 0x33, 0xe1, 0xbd, 0x7a, 0x40, 0xe4, 0xf1, 0x4d, 0x28, 0xf1, 0xd9, 0x6d, 0x6d, 0x35, 0x30, 0x87, - 0x92, 0xff, 0xc7, 0x5f, 0x8f, 0xa6, 0xe3, 0xe3, 0x5e, 0x3c, 0x95, 0xce, 0x12, 0x4f, 0x47, 0x93, - 0x62, 0x29, 0x7a, 0x02, 0x79, 0xaa, 0x65, 0xbd, 0x16, 0x0a, 0xc4, 0xad, 0xb5, 0xa6, 0x1f, 0x90, - 0xb6, 0xd6, 0x9a, 0x98, 0x41, 0xa0, 0x4d, 0x18, 0x62, 0xd9, 0x87, 0x1d, 0xc1, 0x7c, 0xf6, 0x23, - 0xcd, 0xe6, 0xef, 0x3b, 0x04, 0xfb, 0x65, 0x63, 0x07, 0x47, 0xfe, 0x14, 0xc6, 0x43, 0xe7, 0x14, - 0x7d, 0x02, 0x63, 0x9a, 0xa1, 0xb4, 0xeb, 0x8a, 0xa6, 0xe8, 0x2d, 0xe2, 0x7e, 0x1c, 0xb8, 0x96, - 0x74, 0xc3, 0x58, 0x0b, 0xf0, 0x89, 0x53, 0x3e, 0x2b, 0x94, 0x8c, 0x05, 0x69, 0x38, 0x84, 0x28, - 0x2b, 0x00, 0xfe, 0x1c, 0x51, 0x05, 0x86, 0x98, 0x9f, 0x39, 0xf9, 0x64, 0xa4, 0x3e, 0xc2, 0x2c, - 0x64, 0xee, 0x67, 0x63, 0x67, 0x1c, 0xdd, 0x06, 0xb0, 0x49, 0xcb, 0x22, 0x94, 0x07, 0x83, 0x5c, - 0xf8, 0x03, 0x63, 0xd3, 0xa3, 0xe0, 0x00, 0x97, 0xfc, 0x4f, 0x12, 0x8c, 0x6f, 0x10, 0xfa, 0x99, - 0x61, 0xed, 0x37, 0x0c, 0x4d, 0x6d, 0xf5, 0xcf, 0x21, 0xd8, 0xe2, 0x50, 0xb0, 0x7d, 0x6b, 0xc0, - 0xce, 0x84, 0xac, 0x4b, 0x0b, 0xb9, 0xf2, 0x57, 0x12, 0xcc, 0x87, 0x38, 0x1f, 0xf9, 0x47, 0x77, - 0x1b, 0x86, 0x4c, 0xc3, 0xa2, 0x6e, 0x22, 0x3e, 0x95, 0x42, 0x16, 0xc6, 0x02, 0xa9, 0x98, 0xc1, - 0x60, 0x07, 0x0d, 0xad, 0x41, 0x8e, 0x1a, 0xc2, 0x55, 0x4f, 0x87, 0x49, 0x88, 0x55, 0x07, 0x81, - 0x99, 0xdb, 0x32, 0x70, 0x8e, 0x1a, 0x6c, 0x23, 0xca, 0x21, 0xae, 0x60, 0xf0, 0x79, 0x45, 0x33, - 0xc0, 0x50, 0xd8, 0xb5, 0x8c, 0xee, 0x99, 0xe7, 0xe0, 0x6d, 0xc4, 0x8a, 0x65, 0x74, 0x31, 0xc7, - 0x92, 0xbf, 0x96, 0x60, 0x3a, 0xc4, 0x79, 0x0e, 0x81, 0xff, 0xfd, 0x70, 0xe0, 0xbf, 0x71, 0x9a, - 0x89, 0xa4, 0x84, 0xff, 0xaf, 0x73, 0x91, 0x69, 0xb0, 0x09, 0xa3, 0x5d, 0x18, 0x35, 0x8d, 0x76, - 0xf3, 0x25, 0x7c, 0x0e, 0x9c, 0x64, 0x79, 0xb3, 0xe1, 0x63, 0xe1, 0x20, 0x30, 0x3a, 0x84, 0x69, - 0x5d, 0xe9, 0x12, 0xdb, 0x54, 0x5a, 0xa4, 0xf9, 0x12, 0x1e, 0x48, 0x2e, 0xf2, 0xef, 0x0d, 0x51, - 0x44, 0x1c, 0x57, 0x82, 0xd6, 0xa1, 0xa8, 0x9a, 0xbc, 0x8e, 0x13, 0xb5, 0xcb, 0xc0, 0x2c, 0xea, - 0x54, 0x7d, 0x4e, 0x3c, 0x17, 0x3f, 0xb0, 0x8b, 0x21, 0xff, 0x75, 0xd4, 0x1b, 0x98, 0xff, 0xa1, - 0xc7, 0x50, 0xe2, 0x8d, 0x19, 0x2d, 0x43, 0x73, 0xbf, 0x0c, 0xb0, 0x9d, 0x6d, 0x88, 0xb1, 0x17, - 0x47, 0x95, 0xcb, 0x09, 0x8f, 0xbe, 0x2e, 0x19, 0x7b, 0xc2, 0x68, 0x03, 0x0a, 0xe6, 0x8f, 0xa9, - 0x60, 0x78, 0x92, 0xe3, 0x65, 0x0b, 0xc7, 0x91, 0x7f, 0x2f, 0x1f, 0x31, 0x97, 0xa7, 0xba, 0x67, - 0x2f, 0x6d, 0xd7, 0xbd, 0x8a, 0x29, 0x75, 0xe7, 0x77, 0xa0, 0x28, 0x32, 0xbc, 0x70, 0xe6, 0x5f, - 0x9c, 0xc6, 0x99, 0x83, 0x59, 0xcc, 0xbb, 0xb0, 0xb8, 0x83, 0x2e, 0x30, 0xfa, 0x18, 0x86, 0x89, - 0xa3, 0xc2, 0xc9, 0x8d, 0xf7, 0x4e, 0xa3, 0xc2, 0x8f, 0xab, 0x7e, 0xa1, 0x2a, 0xc6, 0x04, 0x2a, - 0x7a, 0x87, 0xad, 0x17, 0xe3, 0x65, 0x97, 0x40, 0xbb, 0x5c, 0xe0, 0xe9, 0xea, 0x8a, 0x33, 0x6d, - 0x6f, 0xf8, 0xc5, 0x51, 0x05, 0xfc, 0x9f, 0x38, 0x28, 0x21, 0xff, 0x8b, 0x04, 0xd3, 0x7c, 0x85, - 0x5a, 0x3d, 0x4b, 0xa5, 0xfd, 0x73, 0x4b, 0x4c, 0x4f, 0x43, 0x89, 0xe9, 0xce, 0x80, 0x65, 0x89, - 0x59, 0x98, 0x9a, 0x9c, 0xbe, 0x91, 0xe0, 0x62, 0x8c, 0xfb, 0x1c, 0xe2, 0xe2, 0x76, 0x38, 0x2e, - 0xbe, 0x75, 0xda, 0x09, 0xa5, 0xc4, 0xc6, 0xff, 0x9a, 0x4e, 0x98, 0x0e, 0x3f, 0x29, 0xb7, 0x01, - 0x4c, 0x4b, 0x3d, 0x50, 0x35, 0xd2, 0x11, 0x1f, 0xc1, 0x4b, 0x81, 0x16, 0x27, 0x8f, 0x82, 0x03, - 0x5c, 0xc8, 0x86, 0xb9, 0x36, 0xd9, 0x55, 0x7a, 0x1a, 0x5d, 0x6c, 0xb7, 0x97, 0x14, 0x53, 0xd9, - 0x51, 0x35, 0x95, 0xaa, 0xe2, 0xb9, 0x60, 0xa4, 0xfe, 0xd0, 0xf9, 0x38, 0x9d, 0xc4, 0xf1, 0xe2, - 0xa8, 0x72, 0x25, 0xe9, 0xeb, 0x90, 0xcb, 0xd2, 0xc7, 0x29, 0xd0, 0xa8, 0x0f, 0x65, 0x8b, 0x7c, - 0xda, 0x53, 0x2d, 0xd2, 0x5e, 0xb6, 0x0c, 0x33, 0xa4, 0x36, 0xcf, 0xd5, 0xfe, 0xfa, 0xf1, 0x51, - 0xa5, 0x8c, 0x53, 0x78, 0x06, 0x2b, 0x4e, 0x85, 0x47, 0xcf, 0x60, 0x46, 0x11, 0xcd, 0x68, 0x41, - 0xad, 0xce, 0x29, 0xb9, 0x7f, 0x7c, 0x54, 0x99, 0x59, 0x8c, 0x93, 0x07, 0x2b, 0x4c, 0x02, 0x45, - 0x35, 0x28, 0x1e, 0xf0, 0xbe, 0x35, 0xbb, 0x3c, 0xc4, 0xf1, 0x59, 0x22, 0x28, 0x3a, 0xad, 0x6c, - 0x0c, 0x73, 0x78, 0xa5, 0xc9, 0x4f, 0x9f, 0xcb, 0xc5, 0x2e, 0x94, 0xac, 0x96, 0x14, 0x27, 0x9e, - 0xbf, 0x18, 0x97, 0xfc, 0xa8, 0xf5, 0xc4, 0x27, 0xe1, 0x20, 0x1f, 0xfa, 0x08, 0x46, 0xf6, 0xc4, - 0xab, 0x84, 0x5d, 0x2e, 0x66, 0x4a, 0xc2, 0xa1, 0x57, 0x8c, 0xfa, 0xb4, 0x50, 0x31, 0xe2, 0x0e, - 0xdb, 0xd8, 0x47, 0x44, 0x6f, 0x40, 0x91, 0xff, 0x58, 0x5d, 0xe6, 0xcf, 0x71, 0x25, 0x3f, 0xb6, - 0x3d, 0x71, 0x86, 0xb1, 0x4b, 0x77, 0x59, 0x57, 0x1b, 0x4b, 0xfc, 0x59, 0x38, 0xc2, 0xba, 0xda, - 0x58, 0xc2, 0x2e, 0x1d, 0x7d, 0x02, 0x45, 0x9b, 0xac, 0xa9, 0x7a, 0xef, 0xb0, 0x0c, 0x99, 0x3e, - 0x2a, 0x37, 0x1f, 0x71, 0xee, 0xc8, 0xc3, 0x98, 0xaf, 0x41, 0xd0, 0xb1, 0x0b, 0x8b, 0xf6, 0x60, - 0xc4, 0xea, 0xe9, 0x8b, 0xf6, 0xb6, 0x4d, 0xac, 0xf2, 0x28, 0xd7, 0x31, 0x28, 0x9c, 0x63, 0x97, - 0x3f, 0xaa, 0xc5, 0x5b, 0x21, 0x8f, 0x03, 0xfb, 0xe0, 0x68, 0x0f, 0x80, 0xff, 0xe0, 0x6f, 0x70, - 0xe5, 0x39, 0xae, 0xea, 0x7e, 0x16, 0x55, 0x49, 0x4f, 0x7d, 0xe2, 0x1d, 0xde, 0x23, 0xe3, 0x00, - 0x36, 0xfa, 0x43, 0x09, 0x90, 0xdd, 0x33, 0x4d, 0x8d, 0x74, 0x89, 0x4e, 0x15, 0x8d, 0x8f, 0xda, - 0xe5, 0x31, 0xae, 0xf2, 0xdd, 0x41, 0x2b, 0x18, 0x13, 0x8c, 0xaa, 0xf6, 0x9e, 0xd7, 0xe3, 0xac, - 0x38, 0x41, 0x2f, 0xdb, 0xc4, 0x5d, 0x31, 0xeb, 0xf1, 0x4c, 0x9b, 0x98, 0xfc, 0xba, 0xe9, 0x6f, - 0xa2, 0xa0, 0x63, 0x17, 0x16, 0x3d, 0x85, 0x39, 0xb7, 0xc1, 0x12, 0x1b, 0x06, 0x5d, 0x51, 0x35, - 0x62, 0xf7, 0x6d, 0x4a, 0xba, 0xe5, 0x09, 0xee, 0x60, 0x5e, 0x97, 0x09, 0x4e, 0xe4, 0xc2, 0x29, - 0xd2, 0xa8, 0x0b, 0x15, 0x37, 0x38, 0xb1, 0x93, 0xeb, 0x45, 0xc7, 0x47, 0x76, 0x4b, 0xd1, 0x9c, - 0x2f, 0x0e, 0x93, 0x5c, 0xc1, 0xeb, 0xc7, 0x47, 0x95, 0xca, 0xf2, 0xc9, 0xac, 0x78, 0x10, 0x16, - 0xfa, 0x00, 0xca, 0x4a, 0x9a, 0x9e, 0x29, 0xae, 0xe7, 0x35, 0x16, 0xf1, 0x52, 0x15, 0xa4, 0x4a, - 0x23, 0x0a, 0x53, 0x4a, 0xb8, 0xd5, 0xd5, 0x2e, 0x4f, 0x67, 0x7a, 0xf2, 0x8c, 0x74, 0xc8, 0xfa, - 0xcf, 0x1e, 0x11, 0x82, 0x8d, 0x63, 0x1a, 0xd0, 0xef, 0x00, 0x52, 0xa2, 0xdd, 0xb9, 0x76, 0x19, - 0x65, 0x4a, 0x74, 0xb1, 0xb6, 0x5e, 0xdf, 0xed, 0x62, 0x24, 0x1b, 0x27, 0xe8, 0x61, 0x05, 0xba, - 0x12, 0xe9, 0x28, 0xb6, 0xcb, 0xf3, 0x5c, 0x79, 0x2d, 0x9b, 0x72, 0x4f, 0x2e, 0xf0, 0x61, 0x25, - 0x8a, 0x88, 0xe3, 0x4a, 0xd0, 0x1a, 0xcc, 0x8a, 0xc1, 0x6d, 0xdd, 0x56, 0x76, 0x49, 0xb3, 0x6f, - 0xb7, 0xa8, 0x66, 0x97, 0x67, 0x78, 0x7c, 0xe7, 0x1f, 0xf7, 0x16, 0x13, 0xe8, 0x38, 0x51, 0x0a, - 0xbd, 0x0b, 0x53, 0xbb, 0x86, 0xb5, 0xa3, 0xb6, 0xdb, 0x44, 0x77, 0x91, 0x66, 0x39, 0xd2, 0x2c, - 0xdb, 0x87, 0x95, 0x08, 0x0d, 0xc7, 0xb8, 0x91, 0x0d, 0x17, 0x05, 0x72, 0xc3, 0x32, 0x5a, 0xeb, - 0x46, 0x4f, 0xa7, 0x4e, 0xd9, 0x77, 0xd1, 0x4b, 0xa3, 0x17, 0x17, 0x93, 0x18, 0x5e, 0x1c, 0x55, - 0xae, 0x26, 0x57, 0xf9, 0x3e, 0x13, 0x4e, 0xc6, 0x46, 0x26, 0x8c, 0x89, 0x3e, 0xf1, 0x25, 0x4d, - 0xb1, 0xed, 0x72, 0x99, 0x1f, 0xfd, 0x07, 0x83, 0x03, 0x9e, 0x27, 0x12, 0x3d, 0xff, 0x53, 0xc7, - 0x47, 0x95, 0xb1, 0x20, 0x03, 0x0e, 0x69, 0xe0, 0x7d, 0x41, 0xe2, 0x6b, 0xd4, 0xf9, 0xf4, 0x56, - 0x9f, 0xae, 0x2f, 0xc8, 0x37, 0xed, 0xa5, 0xf5, 0x05, 0x05, 0x20, 0x4f, 0x7e, 0x97, 0xfe, 0xcf, - 0x1c, 0xcc, 0xf8, 0xcc, 0x99, 0xfb, 0x82, 0x12, 0x44, 0xfe, 0xbf, 0xbf, 0x3a, 0x5b, 0xaf, 0x8e, - 0xbf, 0x74, 0xff, 0xf7, 0x7a, 0x75, 0x7c, 0xdb, 0x52, 0x6e, 0x0f, 0x7f, 0x9b, 0x0b, 0x4e, 0xe0, - 0x94, 0x0d, 0x23, 0x2f, 0xa1, 0xc5, 0xf8, 0x27, 0xd7, 0x73, 0x22, 0x7f, 0x93, 0x87, 0xa9, 0xe8, - 0x69, 0x0c, 0xf5, 0x15, 0x48, 0x03, 0xfb, 0x0a, 0x1a, 0x30, 0xbb, 0xdb, 0xd3, 0xb4, 0x3e, 0x9f, - 0x43, 0xa0, 0xb9, 0xc0, 0xf9, 0x2e, 0xf8, 0x9a, 0x90, 0x9c, 0x5d, 0x49, 0xe0, 0xc1, 0x89, 0x92, - 0xf1, 0x36, 0x83, 0xc2, 0x8f, 0x6d, 0x33, 0x18, 0x3a, 0x43, 0x9b, 0x41, 0x72, 0xa7, 0x46, 0xfe, - 0x4c, 0x9d, 0x1a, 0x67, 0xe9, 0x31, 0x48, 0x08, 0x62, 0x03, 0xfb, 0x65, 0x5f, 0x83, 0x4b, 0x42, - 0x8c, 0xf2, 0xde, 0x01, 0x9d, 0x5a, 0x86, 0xa6, 0x11, 0x6b, 0xb9, 0xd7, 0xed, 0xf6, 0xe5, 0x5f, - 0xc2, 0x44, 0xb8, 0x2b, 0xc6, 0xd9, 0x69, 0xa7, 0x31, 0x47, 0x7c, 0x9d, 0x0d, 0xec, 0xb4, 0x33, - 0x8e, 0x3d, 0x0e, 0xf9, 0xf7, 0x25, 0x98, 0x4b, 0xee, 0x7e, 0x45, 0x1a, 0x4c, 0x74, 0x95, 0xc3, - 0x60, 0x47, 0xb2, 0x74, 0xc6, 0x77, 0x33, 0xde, 0x0e, 0xb1, 0x1e, 0xc2, 0xc2, 0x11, 0x6c, 0xf9, - 0x07, 0x09, 0xe6, 0x53, 0x1a, 0x11, 0xce, 0xd7, 0x12, 0xf4, 0x21, 0x94, 0xba, 0xca, 0x61, 0xb3, - 0x67, 0x75, 0xc8, 0x99, 0x5f, 0x0a, 0xf9, 0x71, 0x5f, 0x17, 0x28, 0xd8, 0xc3, 0x93, 0xff, 0x52, - 0x82, 0x9f, 0xa5, 0x5e, 0xa4, 0xd0, 0xbd, 0x50, 0xcf, 0x84, 0x1c, 0xe9, 0x99, 0x40, 0x71, 0xc1, - 0x57, 0xd4, 0x32, 0xf1, 0x85, 0x04, 0xe5, 0xb4, 0x9b, 0x25, 0xba, 0x1b, 0x32, 0xf2, 0xe7, 0x11, - 0x23, 0xa7, 0x63, 0x72, 0xaf, 0xc8, 0xc6, 0x7f, 0x95, 0xe0, 0xf2, 0x09, 0x15, 0x9a, 0x77, 0x81, - 0x21, 0xed, 0x20, 0x17, 0x7f, 0xd4, 0x16, 0x5f, 0xc4, 0xfc, 0x0b, 0x4c, 0x02, 0x0f, 0x4e, 0x95, - 0x46, 0xdb, 0x30, 0x2f, 0x6e, 0x4f, 0x51, 0x9a, 0x28, 0x3e, 0x78, 0x6b, 0xd9, 0x72, 0x32, 0x0b, - 0x4e, 0x93, 0x95, 0xff, 0x46, 0x82, 0xb9, 0xe4, 0x27, 0x03, 0xf4, 0x76, 0x68, 0xc9, 0x2b, 0x91, - 0x25, 0x9f, 0x8c, 0x48, 0x89, 0x05, 0xff, 0x18, 0x26, 0xc4, 0xc3, 0x82, 0x80, 0x11, 0xce, 0x2c, - 0x27, 0xe5, 0x17, 0x01, 0xe1, 0x96, 0xb7, 0xfc, 0x98, 0x84, 0xc7, 0x70, 0x04, 0x4d, 0xfe, 0x83, - 0x1c, 0x0c, 0x35, 0x5b, 0x8a, 0x46, 0xce, 0xa1, 0xba, 0x7d, 0x2f, 0x54, 0xdd, 0x0e, 0xfa, 0xa7, - 0x2d, 0x6e, 0x55, 0x6a, 0x61, 0x8b, 0x23, 0x85, 0xed, 0x9b, 0x99, 0xd0, 0x4e, 0xae, 0x69, 0x7f, - 0x0d, 0x46, 0x3c, 0xa5, 0xa7, 0x4b, 0xb5, 0xf2, 0x5f, 0xe4, 0x60, 0x34, 0xa0, 0xe2, 0x94, 0x89, - 0x7a, 0x37, 0x54, 0x9d, 0xe4, 0x33, 0x3c, 0xe3, 0x04, 0x74, 0x55, 0xdd, 0x7a, 0xc4, 0x69, 0x3a, - 0xf6, 0xdb, 0x4c, 0xe3, 0x65, 0xca, 0x2f, 0x61, 0x82, 0x2a, 0x56, 0x87, 0x50, 0xef, 0xb3, 0x46, - 0x9e, 0xfb, 0xa2, 0xd7, 0xfd, 0xbe, 0x15, 0xa2, 0xe2, 0x08, 0xf7, 0xa5, 0x87, 0x30, 0x1e, 0x52, - 0x76, 0xaa, 0x9e, 0xe1, 0xbf, 0x97, 0xe0, 0xe7, 0x03, 0x9f, 0x82, 0x50, 0x3d, 0x74, 0x48, 0xaa, - 0x91, 0x43, 0xb2, 0x90, 0x0e, 0xf0, 0xea, 0x7a, 0xcf, 0xea, 0x37, 0x9f, 0x7f, 0xbf, 0x70, 0xe1, - 0xdb, 0xef, 0x17, 0x2e, 0x7c, 0xf7, 0xfd, 0xc2, 0x85, 0xdf, 0x3d, 0x5e, 0x90, 0x9e, 0x1f, 0x2f, - 0x48, 0xdf, 0x1e, 0x2f, 0x48, 0xdf, 0x1d, 0x2f, 0x48, 0xff, 0x7e, 0xbc, 0x20, 0xfd, 0xc9, 0x0f, - 0x0b, 0x17, 0x3e, 0x2c, 0x0a, 0xb8, 0xff, 0x0d, 0x00, 0x00, 0xff, 0xff, 0xa0, 0x62, 0xda, 0xf9, - 0x07, 0x3e, 0x00, 0x00, + // 3743 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xec, 0x5b, 0x4d, 0x6c, 0x1c, 0xc7, + 0x72, 0xd6, 0xec, 0x2e, 0xb9, 0xcb, 0xe2, 0x7f, 0x93, 0x22, 0xf7, 0x49, 0x4f, 0x5c, 0xbd, 0x31, + 0xa0, 0xc8, 0x8e, 0xb4, 0x6b, 0xc9, 0x92, 0x9e, 0x22, 0x21, 0xef, 0x99, 0x4b, 0x8a, 0x12, 0x5f, + 0xf8, 0xb3, 0xee, 0x25, 0x65, 0xc3, 0x88, 0x1d, 0x0f, 0x77, 0x9b, 0xcb, 0x11, 0x67, 0x67, 0xc6, + 0xd3, 0xb3, 0x34, 0x17, 0xc8, 0x21, 0x87, 0x20, 0x80, 0x81, 0x00, 0xc9, 0xc5, 0x49, 0x8e, 0x31, + 0x02, 0xe4, 0x94, 0x20, 0xc7, 0xe4, 0x60, 0x18, 0x09, 0xe2, 0x00, 0x42, 0xe0, 0x04, 0xbe, 0xc5, + 0x27, 0x22, 0xa6, 0x4f, 0x41, 0x4e, 0xb9, 0x05, 0x3a, 0x05, 0xdd, 0xd3, 0xf3, 0x3f, 0xc3, 0x1d, + 0xd2, 0x12, 0x11, 0x03, 0xef, 0x24, 0x6e, 0x57, 0xd5, 0x57, 0xd5, 0xdd, 0xd5, 0x55, 0xd5, 0x3d, + 0x25, 0x58, 0xd9, 0xbf, 0x4f, 0xab, 0xaa, 0x51, 0xdb, 0xef, 0xed, 0x10, 0x4b, 0x27, 0x36, 0xa1, + 0xb5, 0x03, 0xa2, 0xb7, 0x0d, 0xab, 0x26, 0x08, 0x8a, 0xa9, 0xd6, 0xc8, 0xa1, 0x4d, 0x74, 0xaa, + 0x1a, 0x3a, 0xad, 0x1d, 0xdc, 0xda, 0x21, 0xb6, 0x72, 0xab, 0xd6, 0x21, 0x3a, 0xb1, 0x14, 0x9b, + 0xb4, 0xab, 0xa6, 0x65, 0xd8, 0x06, 0xba, 0xe2, 0xb0, 0x57, 0x15, 0x53, 0xad, 0xfa, 0xec, 0x55, + 0xc1, 0x7e, 0xe9, 0x66, 0x47, 0xb5, 0xf7, 0x7a, 0x3b, 0xd5, 0x96, 0xd1, 0xad, 0x75, 0x8c, 0x8e, + 0x51, 0xe3, 0x52, 0x3b, 0xbd, 0x5d, 0xfe, 0x8b, 0xff, 0xe0, 0x7f, 0x39, 0x68, 0x97, 0xe4, 0x80, + 0xf2, 0x96, 0x61, 0x91, 0xda, 0x41, 0x4c, 0xe3, 0xa5, 0x3b, 0x3e, 0x4f, 0x57, 0x69, 0xed, 0xa9, + 0x3a, 0xb1, 0xfa, 0x35, 0x73, 0xbf, 0xc3, 0x06, 0x68, 0xad, 0x4b, 0x6c, 0x25, 0x49, 0xaa, 0x96, + 0x26, 0x65, 0xf5, 0x74, 0x5b, 0xed, 0x92, 0x98, 0xc0, 0xbd, 0x41, 0x02, 0xb4, 0xb5, 0x47, 0xba, + 0x4a, 0x4c, 0xee, 0xad, 0x34, 0xb9, 0x9e, 0xad, 0x6a, 0x35, 0x55, 0xb7, 0xa9, 0x6d, 0x45, 0x85, + 0xe4, 0x3b, 0x30, 0xb5, 0xa8, 0x69, 0xc6, 0x27, 0xa4, 0xbd, 0xd4, 0x5c, 0x5d, 0xb6, 0xd4, 0x03, + 0x62, 0xa1, 0xab, 0x50, 0xd0, 0x95, 0x2e, 0x29, 0x4b, 0x57, 0xa5, 0xeb, 0x23, 0xf5, 0xb1, 0xe7, + 0x47, 0x95, 0x0b, 0xc7, 0x47, 0x95, 0xc2, 0x86, 0xd2, 0x25, 0x98, 0x53, 0xe4, 0x87, 0x30, 0x2d, + 0xa4, 0x56, 0x34, 0x72, 0xf8, 0xd4, 0xd0, 0x7a, 0x5d, 0x82, 0xae, 0xc1, 0x70, 0x9b, 0x03, 0x08, + 0xc1, 0x09, 0x21, 0x38, 0xec, 0xc0, 0x62, 0x41, 0x95, 0x29, 0x4c, 0x0a, 0xe1, 0x27, 0x06, 0xb5, + 0x1b, 0x8a, 0xbd, 0x87, 0x6e, 0x03, 0x98, 0x8a, 0xbd, 0xd7, 0xb0, 0xc8, 0xae, 0x7a, 0x28, 0xc4, + 0x91, 0x10, 0x87, 0x86, 0x47, 0xc1, 0x01, 0x2e, 0x74, 0x03, 0x4a, 0x16, 0x51, 0xda, 0x9b, 0xba, + 0xd6, 0x2f, 0xe7, 0xae, 0x4a, 0xd7, 0x4b, 0xf5, 0x29, 0x21, 0x51, 0xc2, 0x62, 0x1c, 0x7b, 0x1c, + 0xf2, 0x67, 0x39, 0x18, 0x59, 0x56, 0x48, 0xd7, 0xd0, 0x9b, 0xc4, 0x46, 0x1f, 0x41, 0x89, 0x6d, + 0x57, 0x5b, 0xb1, 0x15, 0xae, 0x6d, 0xf4, 0xf6, 0x9b, 0x55, 0xdf, 0x9d, 0xbc, 0xd5, 0xab, 0x9a, + 0xfb, 0x1d, 0x36, 0x40, 0xab, 0x8c, 0xbb, 0x7a, 0x70, 0xab, 0xba, 0xb9, 0xf3, 0x8c, 0xb4, 0xec, + 0x75, 0x62, 0x2b, 0xbe, 0x7d, 0xfe, 0x18, 0xf6, 0x50, 0xd1, 0x06, 0x14, 0xa8, 0x49, 0x5a, 0xdc, + 0xb2, 0xd1, 0xdb, 0x37, 0xaa, 0x27, 0x3a, 0x6b, 0xd5, 0xb3, 0xac, 0x69, 0x92, 0x96, 0xbf, 0xe2, + 0xec, 0x17, 0xe6, 0x38, 0xe8, 0x29, 0x0c, 0x53, 0x5b, 0xb1, 0x7b, 0xb4, 0x9c, 0xe7, 0x88, 0xd5, + 0xcc, 0x88, 0x5c, 0xca, 0xdf, 0x0c, 0xe7, 0x37, 0x16, 0x68, 0xf2, 0x7f, 0xe5, 0x00, 0x79, 0xbc, + 0x4b, 0x86, 0xde, 0x56, 0x6d, 0xd5, 0xd0, 0xd1, 0x03, 0x28, 0xd8, 0x7d, 0xd3, 0x75, 0x81, 0x6b, + 0xae, 0x41, 0x5b, 0x7d, 0x93, 0xbc, 0x38, 0xaa, 0xcc, 0xc5, 0x25, 0x18, 0x05, 0x73, 0x19, 0xb4, + 0xe6, 0x99, 0x9a, 0xe3, 0xd2, 0x77, 0xc2, 0xaa, 0x5f, 0x1c, 0x55, 0x12, 0x0e, 0x5b, 0xd5, 0x43, + 0x0a, 0x1b, 0x88, 0x0e, 0x00, 0x69, 0x0a, 0xb5, 0xb7, 0x2c, 0x45, 0xa7, 0x8e, 0x26, 0xb5, 0x4b, + 0xc4, 0x22, 0xbc, 0x91, 0x6d, 0xd3, 0x98, 0x44, 0xfd, 0x92, 0xb0, 0x02, 0xad, 0xc5, 0xd0, 0x70, + 0x82, 0x06, 0xe6, 0xcd, 0x16, 0x51, 0xa8, 0xa1, 0x97, 0x0b, 0x61, 0x6f, 0xc6, 0x7c, 0x14, 0x0b, + 0x2a, 0x7a, 0x1d, 0x8a, 0x5d, 0x42, 0xa9, 0xd2, 0x21, 0xe5, 0x21, 0xce, 0x38, 0x29, 0x18, 0x8b, + 0xeb, 0xce, 0x30, 0x76, 0xe9, 0xf2, 0x17, 0x12, 0x8c, 0x7b, 0x2b, 0xb7, 0xa6, 0x52, 0x1b, 0xfd, + 0x6e, 0xcc, 0x0f, 0xab, 0xd9, 0xa6, 0xc4, 0xa4, 0xb9, 0x17, 0x7a, 0x3e, 0xef, 0x8e, 0x04, 0x7c, + 0x70, 0x1d, 0x86, 0x54, 0x9b, 0x74, 0xd9, 0x3e, 0xe4, 0xaf, 0x8f, 0xde, 0xbe, 0x9e, 0xd5, 0x65, + 0xea, 0xe3, 0x02, 0x74, 0x68, 0x95, 0x89, 0x63, 0x07, 0x45, 0xfe, 0xb3, 0x42, 0xc0, 0x7c, 0xe6, + 0x9a, 0xe8, 0x03, 0x28, 0x51, 0xa2, 0x91, 0x96, 0x6d, 0x58, 0xc2, 0xfc, 0xb7, 0x32, 0x9a, 0xaf, + 0xec, 0x10, 0xad, 0x29, 0x44, 0xeb, 0x63, 0xcc, 0x7e, 0xf7, 0x17, 0xf6, 0x20, 0xd1, 0x3b, 0x50, + 0xb2, 0x49, 0xd7, 0xd4, 0x14, 0x9b, 0x88, 0x73, 0xf4, 0x5a, 0x70, 0x0a, 0xcc, 0x73, 0x18, 0x58, + 0xc3, 0x68, 0x6f, 0x09, 0x36, 0x7e, 0x7c, 0xbc, 0x25, 0x71, 0x47, 0xb1, 0x07, 0x83, 0x0e, 0x60, + 0xa2, 0x67, 0xb6, 0x19, 0xa7, 0xcd, 0xa2, 0x60, 0xa7, 0x2f, 0x3c, 0xe9, 0x5e, 0xd6, 0xb5, 0xd9, + 0x0e, 0x49, 0xd7, 0xe7, 0x84, 0xae, 0x89, 0xf0, 0x38, 0x8e, 0x68, 0x41, 0x8b, 0x30, 0xd9, 0x55, + 0x75, 0x16, 0x97, 0xfa, 0x4d, 0xd2, 0x32, 0xf4, 0x36, 0xe5, 0x6e, 0x35, 0x54, 0x9f, 0x17, 0x00, + 0x93, 0xeb, 0x61, 0x32, 0x8e, 0xf2, 0xa3, 0x5f, 0x01, 0x72, 0xa7, 0xf1, 0xd8, 0x09, 0xe2, 0xaa, + 0xa1, 0x73, 0x9f, 0xcb, 0xfb, 0xce, 0xbd, 0x15, 0xe3, 0xc0, 0x09, 0x52, 0x68, 0x0d, 0x66, 0x2d, + 0x72, 0xa0, 0xb2, 0x39, 0x3e, 0x51, 0xa9, 0x6d, 0x58, 0xfd, 0x35, 0xb5, 0xab, 0xda, 0xe5, 0x61, + 0x6e, 0x53, 0xf9, 0xf8, 0xa8, 0x32, 0x8b, 0x13, 0xe8, 0x38, 0x51, 0x4a, 0xfe, 0xf3, 0x61, 0x98, + 0x8c, 0xc4, 0x1b, 0xf4, 0x14, 0xe6, 0x5a, 0x3d, 0xcb, 0x22, 0xba, 0xbd, 0xd1, 0xeb, 0xee, 0x10, + 0xab, 0xd9, 0xda, 0x23, 0xed, 0x9e, 0x46, 0xda, 0xdc, 0x51, 0x86, 0xea, 0x0b, 0xc2, 0xe2, 0xb9, + 0xa5, 0x44, 0x2e, 0x9c, 0x22, 0xcd, 0x56, 0x41, 0xe7, 0x43, 0xeb, 0x2a, 0xa5, 0x1e, 0x66, 0x8e, + 0x63, 0x7a, 0xab, 0xb0, 0x11, 0xe3, 0xc0, 0x09, 0x52, 0xcc, 0xc6, 0x36, 0xa1, 0xaa, 0x45, 0xda, + 0x51, 0x1b, 0xf3, 0x61, 0x1b, 0x97, 0x13, 0xb9, 0x70, 0x8a, 0x34, 0xba, 0x0b, 0xa3, 0x8e, 0x36, + 0xbe, 0x7f, 0x62, 0xa3, 0x67, 0x04, 0xd8, 0xe8, 0x86, 0x4f, 0xc2, 0x41, 0x3e, 0x36, 0x35, 0x63, + 0x87, 0x12, 0xeb, 0x80, 0xb4, 0xd3, 0x37, 0x78, 0x33, 0xc6, 0x81, 0x13, 0xa4, 0xd8, 0xd4, 0x1c, + 0x0f, 0x8c, 0x4d, 0x6d, 0x38, 0x3c, 0xb5, 0xed, 0x44, 0x2e, 0x9c, 0x22, 0xcd, 0xfc, 0xd8, 0x31, + 0x79, 0xf1, 0x40, 0x51, 0x35, 0x65, 0x47, 0x23, 0xe5, 0x62, 0xd8, 0x8f, 0x37, 0xc2, 0x64, 0x1c, + 0xe5, 0x47, 0x8f, 0x61, 0xda, 0x19, 0xda, 0xd6, 0x15, 0x0f, 0xa4, 0xc4, 0x41, 0x7e, 0x22, 0x40, + 0xa6, 0x37, 0xa2, 0x0c, 0x38, 0x2e, 0x83, 0x1e, 0xc0, 0x44, 0xcb, 0xd0, 0x34, 0xee, 0x8f, 0x4b, + 0x46, 0x4f, 0xb7, 0xcb, 0x23, 0x1c, 0x05, 0xb1, 0xf3, 0xb8, 0x14, 0xa2, 0xe0, 0x08, 0x27, 0x22, + 0x00, 0x2d, 0x37, 0xe1, 0xd0, 0x32, 0xf0, 0xf8, 0x78, 0x2b, 0x6b, 0x0c, 0xf0, 0x52, 0x95, 0x5f, + 0x03, 0x78, 0x43, 0x14, 0x07, 0x80, 0xe5, 0x7f, 0x95, 0x60, 0x3e, 0x25, 0x74, 0xa0, 0x5f, 0x86, + 0x52, 0xec, 0x6f, 0x46, 0x52, 0xec, 0xe5, 0x14, 0xb1, 0x40, 0x9e, 0xd5, 0x61, 0xdc, 0x62, 0xb3, + 0xd2, 0x3b, 0x0e, 0x8b, 0x88, 0x91, 0x77, 0x07, 0x4c, 0x03, 0x07, 0x65, 0xfc, 0x98, 0x3f, 0x7d, + 0x7c, 0x54, 0x19, 0x0f, 0xd1, 0x70, 0x18, 0x5e, 0xfe, 0x8b, 0x1c, 0xc0, 0x32, 0x31, 0x35, 0xa3, + 0xdf, 0x25, 0xfa, 0x79, 0xd4, 0x50, 0x9b, 0xa1, 0x1a, 0xea, 0xe6, 0xa0, 0xed, 0xf1, 0x4c, 0x4b, + 0x2d, 0xa2, 0xde, 0x8d, 0x14, 0x51, 0xb5, 0xec, 0x90, 0x27, 0x57, 0x51, 0xff, 0x91, 0x87, 0x19, + 0x9f, 0xd9, 0x2f, 0xa3, 0x1e, 0x86, 0xf6, 0xf8, 0x37, 0x22, 0x7b, 0x3c, 0x9f, 0x20, 0xf2, 0xca, + 0xea, 0xa8, 0x67, 0x30, 0xc1, 0xaa, 0x1c, 0x67, 0x2f, 0x79, 0x0d, 0x35, 0x7c, 0xea, 0x1a, 0xca, + 0xcb, 0x76, 0x6b, 0x21, 0x24, 0x1c, 0x41, 0x4e, 0xa9, 0xd9, 0x8a, 0x3f, 0xc6, 0x9a, 0xed, 0x4b, + 0x09, 0x26, 0xfc, 0x6d, 0x3a, 0x87, 0xa2, 0x6d, 0x23, 0x5c, 0xb4, 0xbd, 0x9e, 0xd9, 0x45, 0x53, + 0xaa, 0xb6, 0xff, 0x65, 0x05, 0xbe, 0xc7, 0xc4, 0x0e, 0xf8, 0x8e, 0xd2, 0xda, 0x1f, 0x7c, 0xc7, + 0x43, 0x9f, 0x49, 0x80, 0x44, 0x16, 0x58, 0xd4, 0x75, 0xc3, 0x56, 0x9c, 0x58, 0xe9, 0x98, 0xb5, + 0x9a, 0xd9, 0x2c, 0x57, 0x63, 0x75, 0x3b, 0x86, 0xf5, 0x48, 0xb7, 0xad, 0xbe, 0xbf, 0xc9, 0x71, + 0x06, 0x9c, 0x60, 0x00, 0x52, 0x00, 0x2c, 0x81, 0xb9, 0x65, 0x88, 0x83, 0x7c, 0x33, 0x43, 0xcc, + 0x63, 0x02, 0x4b, 0x86, 0xbe, 0xab, 0x76, 0xfc, 0xb0, 0x83, 0x3d, 0x20, 0x1c, 0x00, 0xbd, 0xf4, + 0x08, 0xe6, 0x53, 0xac, 0x45, 0x53, 0x90, 0xdf, 0x27, 0x7d, 0x67, 0xd9, 0x30, 0xfb, 0x13, 0xcd, + 0xc2, 0xd0, 0x81, 0xa2, 0xf5, 0x9c, 0xf0, 0x3b, 0x82, 0x9d, 0x1f, 0x0f, 0x72, 0xf7, 0x25, 0xf9, + 0x8b, 0xa1, 0xa0, 0xef, 0xf0, 0x8a, 0xf9, 0x3a, 0xbb, 0xb4, 0x9a, 0x9a, 0xda, 0x52, 0xa8, 0x28, + 0x84, 0xc6, 0x9c, 0x0b, 0xab, 0x33, 0x86, 0x3d, 0x6a, 0xa8, 0xb6, 0xce, 0xbd, 0xda, 0xda, 0x3a, + 0xff, 0x72, 0x6a, 0xeb, 0xdf, 0x83, 0x12, 0x75, 0xab, 0xea, 0x02, 0x87, 0xbc, 0x75, 0x8a, 0xf8, + 0x2a, 0x0a, 0x6a, 0x4f, 0x81, 0x57, 0x4a, 0x7b, 0xa0, 0x49, 0x45, 0xf4, 0xd0, 0x29, 0x8b, 0xe8, + 0x97, 0x5a, 0xf8, 0xb2, 0x78, 0x63, 0x2a, 0x3d, 0x4a, 0xda, 0x3c, 0xb6, 0x95, 0xfc, 0x78, 0xd3, + 0xe0, 0xa3, 0x58, 0x50, 0xd1, 0x07, 0x21, 0x97, 0x2d, 0x9d, 0xc5, 0x65, 0x27, 0xd2, 0xdd, 0x15, + 0x6d, 0xc3, 0xbc, 0x69, 0x19, 0x1d, 0x8b, 0x50, 0xba, 0x4c, 0x94, 0xb6, 0xa6, 0xea, 0xc4, 0x5d, + 0x1f, 0xa7, 0x22, 0xba, 0x7c, 0x7c, 0x54, 0x99, 0x6f, 0x24, 0xb3, 0xe0, 0x34, 0x59, 0xf9, 0x79, + 0x01, 0xa6, 0xa2, 0x19, 0x30, 0xa5, 0x48, 0x95, 0xce, 0x54, 0xa4, 0xde, 0x08, 0x1c, 0x06, 0xa7, + 0x82, 0x0f, 0xbc, 0xe0, 0xc4, 0x0e, 0xc4, 0x22, 0x4c, 0x8a, 0x68, 0xe0, 0x12, 0x45, 0x99, 0xee, + 0xed, 0xfe, 0x76, 0x98, 0x8c, 0xa3, 0xfc, 0xe8, 0x21, 0x8c, 0x5b, 0xbc, 0xee, 0x76, 0x01, 0x9c, + 0xda, 0xf5, 0xa2, 0x00, 0x18, 0xc7, 0x41, 0x22, 0x0e, 0xf3, 0xb2, 0xba, 0xd5, 0x2f, 0x47, 0x5d, + 0x80, 0x42, 0xb8, 0x6e, 0x5d, 0x8c, 0x32, 0xe0, 0xb8, 0x0c, 0x5a, 0x87, 0x99, 0x9e, 0x1e, 0x87, + 0x72, 0x5c, 0xf9, 0xb2, 0x80, 0x9a, 0xd9, 0x8e, 0xb3, 0xe0, 0x24, 0x39, 0xb4, 0x1b, 0x2a, 0x65, + 0x87, 0x79, 0x78, 0xbe, 0x9d, 0xf9, 0xe0, 0x65, 0xae, 0x65, 0x13, 0xca, 0xed, 0x52, 0xd6, 0x72, + 0x5b, 0xfe, 0x27, 0x29, 0x98, 0x84, 0xbc, 0x12, 0x78, 0xd0, 0x2b, 0x53, 0x4c, 0x22, 0x50, 0x1d, + 0x19, 0xc9, 0xd5, 0xef, 0xbd, 0x53, 0x55, 0xbf, 0x7e, 0xf2, 0x1c, 0x5c, 0xfe, 0x7e, 0x2e, 0xc1, + 0xdc, 0x4a, 0xf3, 0xb1, 0x65, 0xf4, 0x4c, 0xd7, 0x9c, 0x4d, 0xd3, 0x59, 0x9a, 0x9f, 0x43, 0xc1, + 0xea, 0x69, 0xee, 0x3c, 0x5e, 0x73, 0xe7, 0x81, 0x7b, 0x1a, 0x9b, 0xc7, 0x4c, 0x44, 0xca, 0x99, + 0x04, 0x13, 0x40, 0x1b, 0x30, 0x6c, 0x29, 0x7a, 0x87, 0xb8, 0x69, 0xf5, 0xda, 0x00, 0xeb, 0x57, + 0x97, 0x31, 0x63, 0x0f, 0x14, 0x36, 0x5c, 0x1a, 0x0b, 0x14, 0xf9, 0x9f, 0x25, 0x98, 0x7c, 0xb2, + 0xb5, 0xd5, 0x58, 0xd5, 0xf9, 0x89, 0xe6, 0x6f, 0xab, 0x57, 0xa1, 0x60, 0x2a, 0xf6, 0x5e, 0x34, + 0xd3, 0x33, 0x1a, 0xe6, 0x14, 0x74, 0x07, 0x4a, 0xec, 0x5f, 0x66, 0x17, 0x3f, 0x52, 0x23, 0x3c, + 0x10, 0x96, 0x1a, 0x62, 0xec, 0x45, 0xe0, 0x6f, 0xec, 0x71, 0xa2, 0xf7, 0xa0, 0xc8, 0xe2, 0x0f, + 0xd1, 0xdb, 0x19, 0x0b, 0x74, 0x61, 0x54, 0xdd, 0x11, 0xf2, 0x6b, 0x2e, 0x31, 0x80, 0x5d, 0x38, + 0x79, 0x1f, 0x66, 0x03, 0x93, 0x60, 0xab, 0xf8, 0x94, 0xe5, 0x54, 0xd4, 0x84, 0x21, 0xa6, 0x9d, + 0x65, 0xce, 0x7c, 0x86, 0x27, 0xd0, 0xc8, 0x42, 0xf8, 0xf5, 0x11, 0xfb, 0x45, 0xb1, 0x83, 0x25, + 0xaf, 0xc3, 0x38, 0x7f, 0x86, 0x36, 0x2c, 0x9b, 0x2f, 0x26, 0xba, 0x02, 0xf9, 0xae, 0xaa, 0x8b, + 0xec, 0x3c, 0x2a, 0x64, 0xf2, 0x2c, 0xb3, 0xb0, 0x71, 0x4e, 0x56, 0x0e, 0x45, 0xbc, 0xf2, 0xc9, + 0xca, 0x21, 0x66, 0xe3, 0xf2, 0x63, 0x28, 0x8a, 0x4d, 0x0a, 0x02, 0xe5, 0x4f, 0x06, 0xca, 0x27, + 0x00, 0x6d, 0x42, 0x71, 0xb5, 0x51, 0xd7, 0x0c, 0xa7, 0x56, 0x6b, 0xa9, 0x6d, 0x2b, 0xba, 0x83, + 0x4b, 0xab, 0xcb, 0x18, 0x73, 0x0a, 0x92, 0x61, 0x98, 0x1c, 0xb6, 0x88, 0x69, 0x73, 0x3f, 0x1a, + 0xa9, 0x03, 0xf3, 0x8d, 0x47, 0x7c, 0x04, 0x0b, 0x8a, 0xfc, 0x27, 0x39, 0x28, 0x8a, 0xe5, 0x38, + 0x87, 0xbb, 0xdb, 0x5a, 0xe8, 0xee, 0xf6, 0x46, 0x36, 0xd7, 0x48, 0xbd, 0xb8, 0x6d, 0x45, 0x2e, + 0x6e, 0x37, 0x32, 0xe2, 0x9d, 0x7c, 0x6b, 0xfb, 0x34, 0x07, 0x13, 0x61, 0xa7, 0x44, 0x77, 0x61, + 0x94, 0xa5, 0x29, 0xb5, 0x45, 0x36, 0xfc, 0xea, 0xd8, 0x7b, 0xba, 0x69, 0xfa, 0x24, 0x1c, 0xe4, + 0x43, 0x1d, 0x4f, 0x8c, 0xf9, 0x91, 0x98, 0x74, 0xfa, 0x92, 0xf6, 0x6c, 0x55, 0xab, 0x3a, 0x1f, + 0x64, 0xaa, 0xab, 0xba, 0xbd, 0x69, 0x35, 0x6d, 0x4b, 0xd5, 0x3b, 0x31, 0x45, 0xdc, 0x29, 0x83, + 0xc8, 0xe8, 0x5d, 0x96, 0x32, 0xa9, 0xd1, 0xb3, 0x5a, 0x24, 0xa9, 0xf4, 0x75, 0xcb, 0x36, 0x76, + 0x40, 0xdb, 0x6b, 0x46, 0x4b, 0xd1, 0x9c, 0xcd, 0xc1, 0x64, 0x97, 0x58, 0x44, 0x6f, 0x11, 0xb7, + 0xdc, 0x74, 0x20, 0xb0, 0x07, 0x26, 0xff, 0x83, 0x04, 0xa3, 0x62, 0x2d, 0xce, 0xe1, 0x92, 0xf3, + 0x3b, 0xe1, 0x4b, 0xce, 0xb5, 0x8c, 0x91, 0x23, 0xf9, 0x86, 0xf3, 0xd7, 0xbe, 0xe9, 0x2c, 0x56, + 0xb0, 0xe3, 0xb2, 0x67, 0x50, 0x3b, 0x7a, 0x5c, 0xd8, 0x29, 0xc7, 0x9c, 0x82, 0x7a, 0x30, 0xa5, + 0x46, 0x82, 0x8b, 0xd8, 0xb3, 0x5a, 0x36, 0x4b, 0x3c, 0xb1, 0x7a, 0x59, 0xc0, 0x4f, 0x45, 0x29, + 0x38, 0xa6, 0x42, 0x26, 0x10, 0xe3, 0x42, 0xef, 0x40, 0x61, 0xcf, 0xb6, 0xcd, 0x84, 0xe7, 0xf3, + 0x01, 0x21, 0xcd, 0x37, 0xa1, 0xc4, 0x67, 0xb7, 0xb5, 0xd5, 0xc0, 0x1c, 0x4a, 0xfe, 0xc7, 0x9c, + 0xb7, 0x1e, 0xfc, 0xce, 0xf1, 0xb6, 0x37, 0xdb, 0x25, 0x4d, 0xa1, 0x94, 0x3b, 0xb6, 0x73, 0x3f, + 0x9e, 0x0d, 0x18, 0xee, 0xd1, 0x70, 0x8c, 0x1b, 0x6d, 0xf9, 0xa1, 0x5e, 0x3a, 0x4b, 0xa8, 0x1f, + 0x4d, 0x0a, 0xf3, 0xe8, 0x09, 0xe4, 0x6d, 0x2d, 0xeb, 0x3d, 0x57, 0x20, 0x6e, 0xad, 0x35, 0xfd, + 0x58, 0xb9, 0xb5, 0xd6, 0xc4, 0x0c, 0x02, 0x6d, 0xc2, 0x10, 0x4b, 0xa7, 0x2c, 0x3a, 0xe4, 0xb3, + 0x47, 0x1b, 0xb6, 0x82, 0xbe, 0x4b, 0xb1, 0x5f, 0x14, 0x3b, 0x38, 0xf2, 0xc7, 0x30, 0x1e, 0x0a, + 0x21, 0xe8, 0x23, 0x18, 0xd3, 0x0c, 0xa5, 0x5d, 0x57, 0x34, 0x45, 0x6f, 0x11, 0xf7, 0x6b, 0xc7, + 0xb5, 0xa4, 0xb3, 0xb7, 0x16, 0xe0, 0x13, 0x01, 0x68, 0x56, 0x28, 0x19, 0x0b, 0xd2, 0x70, 0x08, + 0x51, 0x56, 0x00, 0xfc, 0x39, 0xa2, 0x0a, 0x0c, 0x31, 0x4f, 0x75, 0x52, 0xdd, 0x48, 0x7d, 0x84, + 0x59, 0xc8, 0x1c, 0x98, 0x62, 0x67, 0x1c, 0xdd, 0x06, 0xa0, 0xa4, 0x65, 0x11, 0x9b, 0x6f, 0x67, + 0x2e, 0xfc, 0xc5, 0xb4, 0xe9, 0x51, 0x70, 0x80, 0x4b, 0xfe, 0x17, 0x09, 0xc6, 0x37, 0x88, 0xfd, + 0x89, 0x61, 0xed, 0x37, 0x0c, 0x4d, 0x6d, 0xf5, 0xcf, 0x21, 0x0f, 0xe0, 0x50, 0x1e, 0x78, 0x73, + 0xc0, 0xce, 0x84, 0xac, 0x4b, 0xcb, 0x06, 0xf2, 0x97, 0x12, 0xcc, 0x87, 0x38, 0x1f, 0xf9, 0x87, + 0x7f, 0x1b, 0x86, 0x4c, 0xc3, 0xb2, 0xdd, 0x1a, 0xe1, 0x54, 0x0a, 0x59, 0x84, 0x0d, 0x54, 0x09, + 0x0c, 0x06, 0x3b, 0x68, 0x68, 0x0d, 0x72, 0xb6, 0x21, 0x5c, 0xf5, 0x74, 0x98, 0x84, 0x58, 0x75, + 0x10, 0x98, 0xb9, 0x2d, 0x03, 0xe7, 0x6c, 0x83, 0x6d, 0x44, 0x39, 0xc4, 0x15, 0x0c, 0x5f, 0xaf, + 0x68, 0x06, 0x18, 0x0a, 0xbb, 0x96, 0xd1, 0x3d, 0xf3, 0x1c, 0xbc, 0x8d, 0x58, 0xb1, 0x8c, 0x2e, + 0xe6, 0x58, 0xf2, 0x57, 0x12, 0x4c, 0x87, 0x38, 0xcf, 0x21, 0x75, 0xbc, 0x13, 0x4e, 0x1d, 0x37, + 0x4e, 0x33, 0x91, 0x94, 0x04, 0xf2, 0x55, 0x2e, 0x32, 0x0d, 0x36, 0x61, 0xb4, 0x0b, 0xa3, 0xa6, + 0xd1, 0x6e, 0xbe, 0x84, 0xef, 0x9b, 0x93, 0x2c, 0xa5, 0x37, 0x7c, 0x2c, 0x1c, 0x04, 0x46, 0x87, + 0x30, 0xad, 0x2b, 0x5d, 0x42, 0x4d, 0xa5, 0x45, 0x9a, 0x2f, 0xe1, 0xc5, 0xe7, 0x22, 0xff, 0x80, + 0x12, 0x45, 0xc4, 0x71, 0x25, 0x68, 0x1d, 0x8a, 0xaa, 0xc9, 0x4b, 0x4c, 0x51, 0x4b, 0x0c, 0xcc, + 0xc3, 0x4e, 0x41, 0xea, 0xc4, 0x73, 0xf1, 0x03, 0xbb, 0x18, 0xf2, 0xdf, 0x44, 0xbd, 0x81, 0x57, + 0x2c, 0x8f, 0xa1, 0xc4, 0x3b, 0x4d, 0x5a, 0x86, 0xe6, 0x7e, 0xea, 0xe0, 0x97, 0x0b, 0x31, 0xf6, + 0xe2, 0xa8, 0x72, 0x39, 0xe1, 0x15, 0xdb, 0x25, 0x63, 0x4f, 0x18, 0x6d, 0x40, 0xc1, 0xfc, 0x21, + 0xc5, 0x15, 0x4f, 0x93, 0xbc, 0xa2, 0xe2, 0x38, 0xf2, 0x1f, 0xe6, 0x23, 0xe6, 0xf2, 0x64, 0xf9, + 0xec, 0xa5, 0xed, 0xba, 0x57, 0xcc, 0xa5, 0xee, 0xfc, 0x0e, 0x14, 0x45, 0xaa, 0x15, 0xce, 0xfc, + 0xf3, 0xd3, 0x38, 0x73, 0x30, 0x8b, 0x79, 0x77, 0x29, 0x77, 0xd0, 0x05, 0x46, 0x1f, 0xc2, 0x30, + 0x71, 0x54, 0x38, 0xb9, 0xf1, 0xde, 0x69, 0x54, 0xf8, 0x71, 0xd5, 0xaf, 0xa1, 0xc5, 0x98, 0x40, + 0x45, 0xbf, 0x64, 0xeb, 0xc5, 0x78, 0x59, 0xc9, 0x49, 0xcb, 0x05, 0x9e, 0xae, 0xae, 0x38, 0xd3, + 0xf6, 0x86, 0x5f, 0x1c, 0x55, 0xc0, 0xff, 0x89, 0x83, 0x12, 0xf2, 0xbf, 0x49, 0x30, 0xcd, 0x57, + 0xa8, 0xd5, 0xb3, 0x54, 0xbb, 0x7f, 0x6e, 0x89, 0xe9, 0x69, 0x28, 0x31, 0xdd, 0x19, 0xb0, 0x2c, + 0x31, 0x0b, 0x53, 0x93, 0xd3, 0xd7, 0x12, 0x5c, 0x8c, 0x71, 0x9f, 0x43, 0x5c, 0xdc, 0x0e, 0xc7, + 0xc5, 0x37, 0x4f, 0x3b, 0xa1, 0x94, 0xd8, 0xf8, 0x3f, 0xd3, 0x09, 0xd3, 0xe1, 0x27, 0xe5, 0x36, + 0x80, 0x69, 0xa9, 0x07, 0xaa, 0x46, 0x3a, 0xe2, 0xab, 0x7e, 0x29, 0xd0, 0xb3, 0xe5, 0x51, 0x70, + 0x80, 0x0b, 0x51, 0x98, 0x6b, 0x93, 0x5d, 0xa5, 0xa7, 0xd9, 0x8b, 0xed, 0xf6, 0x92, 0x62, 0x2a, + 0x3b, 0xaa, 0xa6, 0xda, 0xaa, 0x78, 0xff, 0x18, 0xa9, 0x3f, 0x74, 0xbe, 0xb6, 0x27, 0x71, 0xbc, + 0x38, 0xaa, 0x5c, 0x49, 0xfa, 0xdc, 0xe5, 0xb2, 0xf4, 0x71, 0x0a, 0x34, 0xea, 0x43, 0xd9, 0x22, + 0x1f, 0xf7, 0x54, 0x8b, 0xb4, 0x97, 0x2d, 0xc3, 0x0c, 0xa9, 0xcd, 0x73, 0xb5, 0xbf, 0x7d, 0x7c, + 0x54, 0x29, 0xe3, 0x14, 0x9e, 0xc1, 0x8a, 0x53, 0xe1, 0xd1, 0x33, 0x98, 0x51, 0x44, 0x77, 0x5d, + 0x50, 0xab, 0x73, 0x4a, 0xee, 0x1f, 0x1f, 0x55, 0x66, 0x16, 0xe3, 0xe4, 0xc1, 0x0a, 0x93, 0x40, + 0x51, 0x0d, 0x8a, 0x07, 0xbc, 0x11, 0x8f, 0x96, 0x87, 0x38, 0x3e, 0x4b, 0x04, 0x45, 0xa7, 0x37, + 0x8f, 0x61, 0x0e, 0xaf, 0x34, 0xf9, 0xe9, 0x73, 0xb9, 0xd8, 0x5d, 0x97, 0xd5, 0x92, 0xe2, 0xc4, + 0xf3, 0x27, 0xf0, 0x92, 0x1f, 0xb5, 0x9e, 0xf8, 0x24, 0x1c, 0xe4, 0x43, 0x1f, 0xc0, 0xc8, 0x9e, + 0x78, 0x30, 0xa1, 0xe5, 0x62, 0xa6, 0x24, 0x1c, 0x7a, 0x60, 0xa9, 0x4f, 0x0b, 0x15, 0x23, 0xee, + 0x30, 0xc5, 0x3e, 0x22, 0x7a, 0x1d, 0x8a, 0xfc, 0xc7, 0xea, 0x32, 0x7f, 0x5f, 0x2c, 0xf9, 0xb1, + 0xed, 0x89, 0x33, 0x8c, 0x5d, 0xba, 0xcb, 0xba, 0xda, 0x58, 0xe2, 0xef, 0xdc, 0x11, 0xd6, 0xd5, + 0xc6, 0x12, 0x76, 0xe9, 0xe8, 0x23, 0x28, 0x52, 0xb2, 0xa6, 0xea, 0xbd, 0xc3, 0x32, 0x64, 0xfa, + 0x4a, 0xde, 0x7c, 0xc4, 0xb9, 0x23, 0x2f, 0x7d, 0xbe, 0x06, 0x41, 0xc7, 0x2e, 0x2c, 0xda, 0x83, + 0x11, 0xab, 0xa7, 0x2f, 0xd2, 0x6d, 0x4a, 0xac, 0xf2, 0x28, 0xd7, 0x31, 0x28, 0x9c, 0x63, 0x97, + 0x3f, 0xaa, 0xc5, 0x5b, 0x21, 0x8f, 0x03, 0xfb, 0xe0, 0x68, 0x0f, 0x80, 0xff, 0xe0, 0x8f, 0x8a, + 0xe5, 0x39, 0xae, 0xea, 0x7e, 0x16, 0x55, 0x49, 0x6f, 0x97, 0xe2, 0xc3, 0x82, 0x47, 0xc6, 0x01, + 0x6c, 0xf4, 0xc7, 0x12, 0x20, 0xda, 0x33, 0x4d, 0x8d, 0x74, 0x89, 0x6e, 0x2b, 0x1a, 0x1f, 0xa5, + 0xe5, 0x31, 0xae, 0xf2, 0xed, 0x41, 0x2b, 0x18, 0x13, 0x8c, 0xaa, 0xf6, 0xbe, 0x17, 0xc4, 0x59, + 0x71, 0x82, 0x5e, 0xb6, 0x89, 0xbb, 0x62, 0xd6, 0xe3, 0x99, 0x36, 0x31, 0xf9, 0xb9, 0xd6, 0xdf, + 0x44, 0x41, 0xc7, 0x2e, 0x2c, 0x7a, 0x0a, 0x73, 0x6e, 0xc7, 0x28, 0x36, 0x0c, 0x7b, 0x45, 0xd5, + 0x08, 0xed, 0x53, 0x9b, 0x74, 0xcb, 0x13, 0xdc, 0xc1, 0xbc, 0xb6, 0x19, 0x9c, 0xc8, 0x85, 0x53, + 0xa4, 0x51, 0x17, 0x2a, 0x6e, 0x70, 0x62, 0x27, 0xd7, 0x8b, 0x8e, 0x8f, 0x68, 0x4b, 0xd1, 0x9c, + 0x4f, 0x28, 0x93, 0x5c, 0xc1, 0x6b, 0xc7, 0x47, 0x95, 0xca, 0xf2, 0xc9, 0xac, 0x78, 0x10, 0x16, + 0x7a, 0x0f, 0xca, 0x4a, 0x9a, 0x9e, 0x29, 0xae, 0xe7, 0xa7, 0x2c, 0xe2, 0xa5, 0x2a, 0x48, 0x95, + 0x46, 0x36, 0x4c, 0x29, 0xe1, 0xde, 0x5d, 0x5a, 0x9e, 0xce, 0xf4, 0x1a, 0x1b, 0x69, 0xf9, 0xf5, + 0x1f, 0x4e, 0x22, 0x04, 0x8a, 0x63, 0x1a, 0xd0, 0xef, 0x03, 0x52, 0xa2, 0xed, 0xc6, 0xb4, 0x8c, + 0x32, 0x25, 0xba, 0x58, 0x9f, 0xb2, 0xef, 0x76, 0x31, 0x12, 0xc5, 0x09, 0x7a, 0x58, 0x81, 0xae, + 0x44, 0x5a, 0xa4, 0x69, 0x79, 0x9e, 0x2b, 0xaf, 0x65, 0x53, 0xee, 0xc9, 0x05, 0xbe, 0x14, 0x45, + 0x11, 0x71, 0x5c, 0x09, 0x5a, 0x83, 0x59, 0x31, 0xb8, 0xad, 0x53, 0x65, 0x97, 0x34, 0xfb, 0xb4, + 0x65, 0x6b, 0xb4, 0x3c, 0xc3, 0xe3, 0x3b, 0xff, 0x5a, 0xb9, 0x98, 0x40, 0xc7, 0x89, 0x52, 0xe8, + 0x6d, 0x98, 0xda, 0x35, 0xac, 0x1d, 0xb5, 0xdd, 0x26, 0xba, 0x8b, 0x34, 0xcb, 0x91, 0xf8, 0x3b, + 0xd0, 0x4a, 0x84, 0x86, 0x63, 0xdc, 0x88, 0xc2, 0x45, 0x81, 0xdc, 0xb0, 0x8c, 0xd6, 0xba, 0xd1, + 0xd3, 0x6d, 0xa7, 0xec, 0xbb, 0xe8, 0xa5, 0xd1, 0x8b, 0x8b, 0x49, 0x0c, 0x2f, 0x8e, 0x2a, 0x57, + 0x93, 0xab, 0x7c, 0x9f, 0x09, 0x27, 0x63, 0x23, 0x13, 0xc6, 0x44, 0xe3, 0x3b, 0x7f, 0x90, 0x2a, + 0x97, 0xf9, 0xd1, 0x7f, 0x30, 0x38, 0xe0, 0x79, 0x22, 0xd1, 0xf3, 0x3f, 0x75, 0x7c, 0x54, 0x19, + 0x0b, 0x32, 0xe0, 0x90, 0x06, 0xde, 0xe8, 0x24, 0x3e, 0xaf, 0x9d, 0x4f, 0xb3, 0xf8, 0xe9, 0x1a, + 0x9d, 0x7c, 0xd3, 0x5e, 0x5a, 0xa3, 0x53, 0x00, 0xf2, 0xe4, 0x27, 0xf3, 0xff, 0xce, 0xc1, 0x8c, + 0xcf, 0x9c, 0xb9, 0xd1, 0x29, 0x41, 0xe4, 0xd7, 0x0d, 0xe3, 0xd9, 0x9a, 0x8f, 0xfc, 0xa5, 0xfb, + 0xff, 0xd7, 0x7c, 0xe4, 0xdb, 0x96, 0x72, 0x7b, 0xf8, 0xbb, 0x5c, 0x70, 0x02, 0xa7, 0xec, 0x80, + 0x79, 0x09, 0x3d, 0xd3, 0x3f, 0xba, 0x26, 0x1a, 0xf9, 0xeb, 0x3c, 0x4c, 0x45, 0x4f, 0x63, 0xa8, + 0x51, 0x42, 0x1a, 0xd8, 0x28, 0xd1, 0x80, 0xd9, 0xdd, 0x9e, 0xa6, 0xf5, 0xf9, 0x1c, 0x02, 0xdd, + 0x12, 0xce, 0x27, 0xcb, 0x9f, 0x0a, 0xc9, 0xd9, 0x95, 0x04, 0x1e, 0x9c, 0x28, 0x19, 0xef, 0x9b, + 0x28, 0xfc, 0xd0, 0xbe, 0x89, 0xa1, 0x33, 0xf4, 0x4d, 0x24, 0xb7, 0x9e, 0xe4, 0xcf, 0xd4, 0x7a, + 0x72, 0x96, 0xa6, 0x89, 0x84, 0x20, 0x36, 0xb0, 0x01, 0xf8, 0x17, 0x30, 0x11, 0x6e, 0xe4, 0x71, + 0xf6, 0xd2, 0xe9, 0x25, 0x12, 0x9f, 0x86, 0x03, 0x7b, 0xe9, 0x8c, 0x63, 0x8f, 0x43, 0xfe, 0x23, + 0x09, 0xe6, 0x92, 0x1b, 0x76, 0x91, 0x06, 0x13, 0x5d, 0xe5, 0x30, 0xd8, 0x44, 0x2d, 0x9d, 0xf1, + 0x65, 0x8c, 0x77, 0x70, 0xac, 0x87, 0xb0, 0x70, 0x04, 0x5b, 0xfe, 0x5e, 0x82, 0xf9, 0x94, 0xde, + 0x89, 0xf3, 0xb5, 0x04, 0xbd, 0x0f, 0xa5, 0xae, 0x72, 0xd8, 0xec, 0x59, 0x1d, 0x72, 0xe6, 0xb7, + 0x40, 0x7e, 0xa0, 0xd7, 0x05, 0x0a, 0xf6, 0xf0, 0xe4, 0xbf, 0x92, 0xe0, 0x27, 0xa9, 0x57, 0x25, + 0x74, 0x2f, 0xd4, 0xe6, 0x21, 0x47, 0xda, 0x3c, 0x50, 0x5c, 0xf0, 0x15, 0x75, 0x79, 0x7c, 0x2e, + 0x41, 0x39, 0xed, 0xee, 0x88, 0xee, 0x86, 0x8c, 0xfc, 0x59, 0xc4, 0xc8, 0xe9, 0x98, 0xdc, 0x2b, + 0xb2, 0xf1, 0xdf, 0x25, 0xb8, 0x7c, 0x42, 0x0d, 0xe6, 0x5d, 0x51, 0x48, 0x3b, 0xc8, 0xc5, 0x9f, + 0xad, 0xc5, 0x37, 0x2f, 0xff, 0x8a, 0x92, 0xc0, 0x83, 0x53, 0xa5, 0xd1, 0x36, 0xcc, 0x8b, 0xfb, + 0x51, 0x94, 0x26, 0xca, 0x0b, 0xde, 0x0d, 0xb7, 0x9c, 0xcc, 0x82, 0xd3, 0x64, 0xe5, 0xbf, 0x95, + 0x60, 0x2e, 0xf9, 0x51, 0x00, 0xbd, 0x15, 0x5a, 0xf2, 0x4a, 0x64, 0xc9, 0x27, 0x23, 0x52, 0x62, + 0xc1, 0x3f, 0x84, 0x09, 0xf1, 0x74, 0x20, 0x60, 0x84, 0x33, 0xcb, 0x49, 0x19, 0x44, 0x40, 0xb8, + 0x05, 0x2c, 0x3f, 0x26, 0xe1, 0x31, 0x1c, 0x41, 0x93, 0x3f, 0xcd, 0xc1, 0x50, 0xb3, 0xa5, 0x68, + 0xe4, 0x1c, 0xea, 0xd7, 0x5f, 0x85, 0xea, 0xd7, 0x41, 0xff, 0xcf, 0x8c, 0x5b, 0x95, 0x5a, 0xba, + 0xe2, 0x48, 0xe9, 0xfa, 0x46, 0x26, 0xb4, 0x93, 0xab, 0xd6, 0xdf, 0x82, 0x11, 0x4f, 0xe9, 0xe9, + 0x92, 0xa9, 0xfc, 0x97, 0x39, 0x18, 0x0d, 0xa8, 0x38, 0x65, 0x2a, 0xde, 0x0d, 0xd5, 0x1f, 0xf9, + 0x0c, 0x0f, 0x35, 0x01, 0x5d, 0x55, 0xb7, 0xe2, 0x70, 0xfa, 0xa4, 0xfd, 0xce, 0xd8, 0x78, 0x21, + 0xf2, 0x0b, 0x98, 0xb0, 0x15, 0xab, 0x43, 0x6c, 0xef, 0xc3, 0x85, 0xd3, 0xc7, 0xe5, 0x35, 0xec, + 0x6f, 0x85, 0xa8, 0x38, 0xc2, 0x7d, 0xe9, 0x21, 0x8c, 0x87, 0x94, 0x9d, 0xaa, 0xcd, 0xf9, 0xef, + 0x25, 0xf8, 0xd9, 0xc0, 0xc7, 0x1e, 0x54, 0x0f, 0x1d, 0x92, 0x6a, 0xe4, 0x90, 0x2c, 0xa4, 0x03, + 0xbc, 0xba, 0x76, 0xb9, 0xfa, 0xcd, 0xe7, 0xdf, 0x2d, 0x5c, 0xf8, 0xe6, 0xbb, 0x85, 0x0b, 0xdf, + 0x7e, 0xb7, 0x70, 0xe1, 0x0f, 0x8e, 0x17, 0xa4, 0xe7, 0xc7, 0x0b, 0xd2, 0x37, 0xc7, 0x0b, 0xd2, + 0xb7, 0xc7, 0x0b, 0xd2, 0x7f, 0x1e, 0x2f, 0x48, 0x7f, 0xfa, 0xfd, 0xc2, 0x85, 0xf7, 0x8b, 0x02, + 0xee, 0xff, 0x02, 0x00, 0x00, 0xff, 0xff, 0x98, 0xf4, 0xad, 0x8a, 0xba, 0x3e, 0x00, 0x00, } func (m *AllowedCSIDriver) Marshal() (dAtA []byte, err error) { @@ -2843,6 +2817,13 @@ func (m *HTTPIngressPath) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.PathType != nil { + i -= len(*m.PathType) + copy(dAtA[i:], *m.PathType) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.PathType))) + i-- + dAtA[i] = 0x1a + } { size, err := m.Backend.MarshalToSizedBuffer(dAtA[:i]) if err != nil { @@ -3066,6 +3047,18 @@ func (m *IngressBackend) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.Resource != nil { + { + size, err := m.Resource.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1a + } { size, err := m.ServicePort.MarshalToSizedBuffer(dAtA[:i]) if err != nil { @@ -3224,6 +3217,13 @@ func (m *IngressSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.IngressClassName != nil { + i -= len(*m.IngressClassName) + copy(dAtA[i:], *m.IngressClassName) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.IngressClassName))) + i-- + dAtA[i] = 0x22 + } if len(m.Rules) > 0 { for iNdEx := len(m.Rules) - 1; iNdEx >= 0; iNdEx-- { { @@ -4332,29 +4332,6 @@ func (m *ReplicaSetStatus) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } -func (m *ReplicationControllerDummy) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) - if err != nil { - return nil, err - } - return dAtA[:n], nil -} - -func (m *ReplicationControllerDummy) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) -} - -func (m *ReplicationControllerDummy) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - return len(dAtA) - i, nil -} - func (m *RollbackConfig) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -5133,6 +5110,10 @@ func (m *HTTPIngressPath) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) l = m.Backend.Size() n += 1 + l + sovGenerated(uint64(l)) + if m.PathType != nil { + l = len(*m.PathType) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -5215,6 +5196,10 @@ func (m *IngressBackend) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) l = m.ServicePort.Size() n += 1 + l + sovGenerated(uint64(l)) + if m.Resource != nil { + l = m.Resource.Size() + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -5283,6 +5268,10 @@ func (m *IngressSpec) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) } } + if m.IngressClassName != nil { + l = len(*m.IngressClassName) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -5675,15 +5664,6 @@ func (m *ReplicaSetStatus) Size() (n int) { return n } -func (m *ReplicationControllerDummy) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - return n -} - func (m *RollbackConfig) Size() (n int) { if m == nil { return 0 @@ -6122,6 +6102,7 @@ func (this *HTTPIngressPath) String() string { s := strings.Join([]string{`&HTTPIngressPath{`, `Path:` + fmt.Sprintf("%v", this.Path) + `,`, `Backend:` + strings.Replace(strings.Replace(this.Backend.String(), "IngressBackend", "IngressBackend", 1), `&`, ``, 1) + `,`, + `PathType:` + valueToStringGenerated(this.PathType) + `,`, `}`, }, "") return s @@ -6193,6 +6174,7 @@ func (this *IngressBackend) String() string { s := strings.Join([]string{`&IngressBackend{`, `ServiceName:` + fmt.Sprintf("%v", this.ServiceName) + `,`, `ServicePort:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ServicePort), "IntOrString", "intstr.IntOrString", 1), `&`, ``, 1) + `,`, + `Resource:` + strings.Replace(fmt.Sprintf("%v", this.Resource), "TypedLocalObjectReference", "v11.TypedLocalObjectReference", 1) + `,`, `}`, }, "") return s @@ -6252,6 +6234,7 @@ func (this *IngressSpec) String() string { `Backend:` + strings.Replace(this.Backend.String(), "IngressBackend", "IngressBackend", 1) + `,`, `TLS:` + repeatedStringForTLS + `,`, `Rules:` + repeatedStringForRules + `,`, + `IngressClassName:` + valueToStringGenerated(this.IngressClassName) + `,`, `}`, }, "") return s @@ -6547,15 +6530,6 @@ func (this *ReplicaSetStatus) String() string { }, "") return s } -func (this *ReplicationControllerDummy) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&ReplicationControllerDummy{`, - `}`, - }, "") - return s -} func (this *RollbackConfig) String() string { if this == nil { return "nil" @@ -9672,6 +9646,39 @@ func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field PathType", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := PathType(dAtA[iNdEx:postIndex]) + m.PathType = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -10328,6 +10335,42 @@ func (m *IngressBackend) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Resource", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Resource == nil { + m.Resource = &v11.TypedLocalObjectReference{} + } + if err := m.Resource.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -10812,6 +10855,39 @@ func (m *IngressSpec) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field IngressClassName", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.IngressClassName = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -13816,59 +13892,6 @@ func (m *ReplicaSetStatus) Unmarshal(dAtA []byte) error { } return nil } -func (m *ReplicationControllerDummy) Unmarshal(dAtA []byte) error { - l := len(dAtA) - iNdEx := 0 - for iNdEx < l { - preIndex := iNdEx - var wire uint64 - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - wire |= uint64(b&0x7F) << shift - if b < 0x80 { - break - } - } - fieldNum := int32(wire >> 3) - wireType := int(wire & 0x7) - if wireType == 4 { - return fmt.Errorf("proto: ReplicationControllerDummy: wiretype end group for non-group") - } - if fieldNum <= 0 { - return fmt.Errorf("proto: ReplicationControllerDummy: illegal tag %d (wire type %d)", fieldNum, wire) - } - switch fieldNum { - default: - iNdEx = preIndex - skippy, err := skipGenerated(dAtA[iNdEx:]) - if err != nil { - return err - } - if skippy < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) < 0 { - return ErrInvalidLengthGenerated - } - if (iNdEx + skippy) > l { - return io.ErrUnexpectedEOF - } - iNdEx += skippy - } - } - - if iNdEx > l { - return io.ErrUnexpectedEOF - } - return nil -} func (m *RollbackConfig) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 @@ -15209,6 +15232,7 @@ func (m *SupplementalGroupsStrategyOptions) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -15240,10 +15264,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -15264,55 +15286,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/extensions/v1beta1/generated.proto b/vendor/k8s.io/api/extensions/v1beta1/generated.proto index 6c90cb3c..a81cce68 100644 --- a/vendor/k8s.io/api/extensions/v1beta1/generated.proto +++ b/vendor/k8s.io/api/extensions/v1beta1/generated.proto @@ -316,7 +316,7 @@ message DeploymentSpec { // The number of old ReplicaSets to retain to allow rollback. // This is a pointer to distinguish between explicit zero and not specified. // This is set to the max value of int32 (i.e. 2147483647) by default, which - // means "retaining all old RelicaSets". + // means "retaining all old ReplicaSets". // +optional optional int32 revisionHistoryLimit = 6; @@ -408,19 +408,33 @@ message FSGroupStrategyOptions { repeated IDRange ranges = 2; } -// HTTPIngressPath associates a path regex with a backend. Incoming urls matching -// the path are forwarded to the backend. +// HTTPIngressPath associates a path with a backend. Incoming urls matching the +// path are forwarded to the backend. message HTTPIngressPath { - // Path is an extended POSIX regex as defined by IEEE Std 1003.1, - // (i.e this follows the egrep/unix syntax, not the perl syntax) - // matched against the path of an incoming request. Currently it can - // contain characters disallowed from the conventional "path" - // part of a URL as defined by RFC 3986. Paths must begin with - // a '/'. If unspecified, the path defaults to a catch all sending - // traffic to the backend. + // Path is matched against the path of an incoming request. Currently it can + // contain characters disallowed from the conventional "path" part of a URL + // as defined by RFC 3986. Paths must begin with a '/'. When unspecified, + // all paths from incoming requests are matched. // +optional optional string path = 1; + // PathType determines the interpretation of the Path matching. PathType can + // be one of the following values: + // * Exact: Matches the URL path exactly. + // * Prefix: Matches based on a URL path prefix split by '/'. Matching is + // done on a path element by element basis. A path element refers is the + // list of labels in the path split by the '/' separator. A request is a + // match for path p if every p is an element-wise prefix of p of the + // request path. Note that if the last element of the path is a substring + // of the last element in request path, it is not a match (e.g. /foo/bar + // matches /foo/bar/baz, but does not match /foo/barbaz). + // * ImplementationSpecific: Interpretation of the Path matching is up to + // the IngressClass. Implementations can treat this as a separate PathType + // or treat it identically to Prefix or Exact path types. + // Implementations are required to support all path types. + // Defaults to ImplementationSpecific. + optional string pathType = 3; + // Backend defines the referenced service endpoint to which the traffic // will be forwarded to. optional IngressBackend backend = 2; @@ -458,16 +472,16 @@ message IDRange { } // DEPRECATED 1.9 - This group version of IPBlock is deprecated by networking/v1/IPBlock. -// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24") that is allowed to the pods -// matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should -// not be included within this rule. +// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24","2001:db9::/64") that is allowed +// to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs +// that should not be included within this rule. message IPBlock { // CIDR is a string representing the IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" optional string cidr = 1; // Except is a slice of CIDRs that should not be included within an IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" // Except values will be rejected if they are outside the CIDR range // +optional repeated string except = 2; @@ -498,10 +512,18 @@ message Ingress { // IngressBackend describes all endpoints for a given service and port. message IngressBackend { // Specifies the name of the referenced service. + // +optional optional string serviceName = 1; // Specifies the port of the referenced service. + // +optional optional k8s.io.apimachinery.pkg.util.intstr.IntOrString servicePort = 2; + + // Resource is an ObjectRef to another Kubernetes resource in the namespace + // of the Ingress object. If resource is specified, serviceName and servicePort + // must not be specified. + // +optional + optional k8s.io.api.core.v1.TypedLocalObjectReference resource = 3; } // IngressList is a collection of Ingress. @@ -519,18 +541,28 @@ message IngressList { // the related backend services. Incoming requests are first evaluated for a host // match, then routed to the backend associated with the matching IngressRuleValue. message IngressRule { - // Host is the fully qualified domain name of a network host, as defined - // by RFC 3986. Note the following deviations from the "host" part of the - // URI as defined in the RFC: - // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the - // IP in the Spec of the parent Ingress. + // Host is the fully qualified domain name of a network host, as defined by RFC 3986. + // Note the following deviations from the "host" part of the + // URI as defined in RFC 3986: + // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to + // the IP in the Spec of the parent Ingress. // 2. The `:` delimiter is not respected because ports are not allowed. // Currently the port of an Ingress is implicitly :80 for http and // :443 for https. // Both these may change in the future. - // Incoming requests are matched against the host before the IngressRuleValue. - // If the host is unspecified, the Ingress routes all traffic based on the - // specified IngressRuleValue. + // Incoming requests are matched against the host before the + // IngressRuleValue. If the host is unspecified, the Ingress routes all + // traffic based on the specified IngressRuleValue. + // + // Host can be "precise" which is a domain name without the terminating dot of + // a network host (e.g. "foo.bar.com") or "wildcard", which is a domain name + // prefixed with a single wildcard label (e.g. "*.foo.com"). + // The wildcard character '*' must appear by itself as the first DNS label and + // matches only a single label. You cannot have a wildcard label by itself (e.g. Host == "*"). + // Requests will be matched against the Host field in the following way: + // 1. If Host is precise, the request matches this rule if the http host header is equal to Host. + // 2. If Host is a wildcard, then the request matches this rule if the http host header + // is to equal to the suffix (removing the first label) of the wildcard rule. // +optional optional string host = 1; @@ -554,6 +586,19 @@ message IngressRuleValue { // IngressSpec describes the Ingress the user wishes to exist. message IngressSpec { + // IngressClassName is the name of the IngressClass cluster resource. The + // associated IngressClass defines which controller will implement the + // resource. This replaces the deprecated `kubernetes.io/ingress.class` + // annotation. For backwards compatibility, when that annotation is set, it + // must be given precedence over this field. The controller may emit a + // warning if the field and annotation have different values. + // Implementations of this API should ignore Ingresses without a class + // specified. An IngressClass resource may be marked as default, which can + // be used to set a default value for this field. For more information, + // refer to the IngressClass documentation. + // +optional + optional string ingressClassName = 4; + // A default backend capable of servicing requests that don't match any // rule. At least one of 'backend' or 'rules' must be specified. This field // is optional to allow the loadbalancer controller or defaulting logic to @@ -803,7 +848,7 @@ message PodSecurityPolicySpec { // +optional repeated string allowedCapabilities = 4; - // volumes is a white list of allowed volume plugins. Empty indicates that + // volumes is an allowlist of volume plugins. Empty indicates that // no volumes may be used. To allow all volumes you may use '*'. // +optional repeated string volumes = 5; @@ -860,18 +905,18 @@ message PodSecurityPolicySpec { // +optional optional bool allowPrivilegeEscalation = 16; - // allowedHostPaths is a white list of allowed host paths. Empty indicates + // allowedHostPaths is an allowlist of host paths. Empty indicates // that all host paths may be used. // +optional repeated AllowedHostPath allowedHostPaths = 17; - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all + // allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all // Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes // is allowed in the "volumes" field. // +optional repeated AllowedFlexVolume allowedFlexVolumes = 18; - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. + // AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. // An empty value indicates that any CSI driver can be used for inline ephemeral volumes. // +optional repeated AllowedCSIDriver allowedCSIDrivers = 23; @@ -879,7 +924,7 @@ message PodSecurityPolicySpec { // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. // Each entry is either a plain sysctl name or ends in "*" in which case it is considered // as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. - // Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. + // Kubelet has to allowlist all unsafe sysctls explicitly to avoid rejection. // // Examples: // e.g. "foo/*" allows "foo/bar", "foo/baz", etc. @@ -897,7 +942,7 @@ message PodSecurityPolicySpec { // +optional repeated string forbiddenSysctls = 20; - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. + // AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. // Empty or nil indicates that only the DefaultProcMountType may be used. // This requires the ProcMountType feature flag to be enabled. // +optional @@ -1025,10 +1070,6 @@ message ReplicaSetStatus { repeated ReplicaSetCondition conditions = 6; } -// Dummy definition -message ReplicationControllerDummy { -} - // DEPRECATED. message RollbackConfig { // The revision to rollback to. If set to 0, rollback to the last revision. @@ -1113,7 +1154,7 @@ message RunAsUserStrategyOptions { // RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses // for a pod. message RuntimeClassStrategyOptions { - // allowedRuntimeClassNames is a whitelist of RuntimeClass names that may be specified on a pod. + // allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. // A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the // list. An empty list requires the RuntimeClassName field to be unset. repeated string allowedRuntimeClassNames = 1; diff --git a/vendor/k8s.io/api/extensions/v1beta1/register.go b/vendor/k8s.io/api/extensions/v1beta1/register.go index 7625f678..c69eff0b 100644 --- a/vendor/k8s.io/api/extensions/v1beta1/register.go +++ b/vendor/k8s.io/api/extensions/v1beta1/register.go @@ -47,7 +47,6 @@ func addKnownTypes(scheme *runtime.Scheme) error { &Deployment{}, &DeploymentList{}, &DeploymentRollback{}, - &ReplicationControllerDummy{}, &Scale{}, &DaemonSetList{}, &DaemonSet{}, diff --git a/vendor/k8s.io/api/extensions/v1beta1/types.go b/vendor/k8s.io/api/extensions/v1beta1/types.go index eb255341..bd75c51b 100644 --- a/vendor/k8s.io/api/extensions/v1beta1/types.go +++ b/vendor/k8s.io/api/extensions/v1beta1/types.go @@ -50,6 +50,9 @@ type ScaleStatus struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.1 +// +k8s:prerelease-lifecycle-gen:deprecated=1.2 +// +k8s:prerelease-lifecycle-gen:removed=1.16 // represents a scaling request for a resource. type Scale struct { @@ -67,17 +70,14 @@ type Scale struct { Status ScaleStatus `json:"status,omitempty" protobuf:"bytes,3,opt,name=status"` } -// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object - -// Dummy definition -type ReplicationControllerDummy struct { - metav1.TypeMeta `json:",inline"` -} - // +genclient // +genclient:method=GetScale,verb=get,subresource=scale,result=Scale // +genclient:method=UpdateScale,verb=update,subresource=scale,input=Scale,result=Scale // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.1 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,Deployment // DEPRECATED - This group version of Deployment is deprecated by apps/v1beta2/Deployment. See the release notes for // more information. @@ -126,7 +126,7 @@ type DeploymentSpec struct { // The number of old ReplicaSets to retain to allow rollback. // This is a pointer to distinguish between explicit zero and not specified. // This is set to the max value of int32 (i.e. 2147483647) by default, which - // means "retaining all old RelicaSets". + // means "retaining all old ReplicaSets". // +optional RevisionHistoryLimit *int32 `json:"revisionHistoryLimit,omitempty" protobuf:"varint,6,opt,name=revisionHistoryLimit"` @@ -151,6 +151,9 @@ type DeploymentSpec struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 // DEPRECATED. // DeploymentRollback stores the information required to rollback a deployment. @@ -308,6 +311,10 @@ type DeploymentCondition struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.1 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DeploymentList // DeploymentList is a list of Deployments. type DeploymentList struct { @@ -482,6 +489,10 @@ type DaemonSetCondition struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.1 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DaemonSet // DEPRECATED - This group version of DaemonSet is deprecated by apps/v1beta2/DaemonSet. See the release notes for // more information. @@ -521,6 +532,10 @@ const ( ) // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.1 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,DaemonSetList // DaemonSetList is a collection of daemon sets. type DaemonSetList struct { @@ -536,6 +551,10 @@ type DaemonSetList struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.1 +// +k8s:prerelease-lifecycle-gen:deprecated=1.14 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,Ingress // Ingress is a collection of rules that allow inbound connections to reach the // endpoints defined by a backend. An Ingress can be configured to give services @@ -561,6 +580,10 @@ type Ingress struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.1 +// +k8s:prerelease-lifecycle-gen:deprecated=1.14 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,IngressList // IngressList is a collection of Ingress. type IngressList struct { @@ -576,6 +599,19 @@ type IngressList struct { // IngressSpec describes the Ingress the user wishes to exist. type IngressSpec struct { + // IngressClassName is the name of the IngressClass cluster resource. The + // associated IngressClass defines which controller will implement the + // resource. This replaces the deprecated `kubernetes.io/ingress.class` + // annotation. For backwards compatibility, when that annotation is set, it + // must be given precedence over this field. The controller may emit a + // warning if the field and annotation have different values. + // Implementations of this API should ignore Ingresses without a class + // specified. An IngressClass resource may be marked as default, which can + // be used to set a default value for this field. For more information, + // refer to the IngressClass documentation. + // +optional + IngressClassName *string `json:"ingressClassName,omitempty" protobuf:"bytes,4,opt,name=ingressClassName"` + // A default backend capable of servicing requests that don't match any // rule. At least one of 'backend' or 'rules' must be specified. This field // is optional to allow the loadbalancer controller or defaulting logic to @@ -627,18 +663,28 @@ type IngressStatus struct { // the related backend services. Incoming requests are first evaluated for a host // match, then routed to the backend associated with the matching IngressRuleValue. type IngressRule struct { - // Host is the fully qualified domain name of a network host, as defined - // by RFC 3986. Note the following deviations from the "host" part of the - // URI as defined in the RFC: - // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the - // IP in the Spec of the parent Ingress. + // Host is the fully qualified domain name of a network host, as defined by RFC 3986. + // Note the following deviations from the "host" part of the + // URI as defined in RFC 3986: + // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to + // the IP in the Spec of the parent Ingress. // 2. The `:` delimiter is not respected because ports are not allowed. // Currently the port of an Ingress is implicitly :80 for http and // :443 for https. // Both these may change in the future. - // Incoming requests are matched against the host before the IngressRuleValue. - // If the host is unspecified, the Ingress routes all traffic based on the - // specified IngressRuleValue. + // Incoming requests are matched against the host before the + // IngressRuleValue. If the host is unspecified, the Ingress routes all + // traffic based on the specified IngressRuleValue. + // + // Host can be "precise" which is a domain name without the terminating dot of + // a network host (e.g. "foo.bar.com") or "wildcard", which is a domain name + // prefixed with a single wildcard label (e.g. "*.foo.com"). + // The wildcard character '*' must appear by itself as the first DNS label and + // matches only a single label. You cannot have a wildcard label by itself (e.g. Host == "*"). + // Requests will be matched against the Host field in the following way: + // 1. If Host is precise, the request matches this rule if the http host header is equal to Host. + // 2. If Host is a wildcard, then the request matches this rule if the http host header + // is to equal to the suffix (removing the first label) of the wildcard rule. // +optional Host string `json:"host,omitempty" protobuf:"bytes,1,opt,name=host"` // IngressRuleValue represents a rule to route requests for this IngressRule. @@ -677,19 +723,63 @@ type HTTPIngressRuleValue struct { // options usable by a loadbalancer, like http keep-alive. } -// HTTPIngressPath associates a path regex with a backend. Incoming urls matching -// the path are forwarded to the backend. +// PathType represents the type of path referred to by a HTTPIngressPath. +type PathType string + +const ( + // PathTypeExact matches the URL path exactly and with case sensitivity. + PathTypeExact = PathType("Exact") + + // PathTypePrefix matches based on a URL path prefix split by '/'. Matching + // is case sensitive and done on a path element by element basis. A path + // element refers to the list of labels in the path split by the '/' + // separator. A request is a match for path p if every p is an element-wise + // prefix of p of the request path. Note that if the last element of the + // path is a substring of the last element in request path, it is not a + // match (e.g. /foo/bar matches /foo/bar/baz, but does not match + // /foo/barbaz). If multiple matching paths exist in an Ingress spec, the + // longest matching path is given priority. + // Examples: + // - /foo/bar does not match requests to /foo/barbaz + // - /foo/bar matches request to /foo/bar and /foo/bar/baz + // - /foo and /foo/ both match requests to /foo and /foo/. If both paths are + // present in an Ingress spec, the longest matching path (/foo/) is given + // priority. + PathTypePrefix = PathType("Prefix") + + // PathTypeImplementationSpecific matching is up to the IngressClass. + // Implementations can treat this as a separate PathType or treat it + // identically to Prefix or Exact path types. + PathTypeImplementationSpecific = PathType("ImplementationSpecific") +) + +// HTTPIngressPath associates a path with a backend. Incoming urls matching the +// path are forwarded to the backend. type HTTPIngressPath struct { - // Path is an extended POSIX regex as defined by IEEE Std 1003.1, - // (i.e this follows the egrep/unix syntax, not the perl syntax) - // matched against the path of an incoming request. Currently it can - // contain characters disallowed from the conventional "path" - // part of a URL as defined by RFC 3986. Paths must begin with - // a '/'. If unspecified, the path defaults to a catch all sending - // traffic to the backend. + // Path is matched against the path of an incoming request. Currently it can + // contain characters disallowed from the conventional "path" part of a URL + // as defined by RFC 3986. Paths must begin with a '/'. When unspecified, + // all paths from incoming requests are matched. // +optional Path string `json:"path,omitempty" protobuf:"bytes,1,opt,name=path"` + // PathType determines the interpretation of the Path matching. PathType can + // be one of the following values: + // * Exact: Matches the URL path exactly. + // * Prefix: Matches based on a URL path prefix split by '/'. Matching is + // done on a path element by element basis. A path element refers is the + // list of labels in the path split by the '/' separator. A request is a + // match for path p if every p is an element-wise prefix of p of the + // request path. Note that if the last element of the path is a substring + // of the last element in request path, it is not a match (e.g. /foo/bar + // matches /foo/bar/baz, but does not match /foo/barbaz). + // * ImplementationSpecific: Interpretation of the Path matching is up to + // the IngressClass. Implementations can treat this as a separate PathType + // or treat it identically to Prefix or Exact path types. + // Implementations are required to support all path types. + // Defaults to ImplementationSpecific. + PathType *PathType `json:"pathType,omitempty" protobuf:"bytes,3,opt,name=pathType"` + // Backend defines the referenced service endpoint to which the traffic // will be forwarded to. Backend IngressBackend `json:"backend" protobuf:"bytes,2,opt,name=backend"` @@ -698,16 +788,28 @@ type HTTPIngressPath struct { // IngressBackend describes all endpoints for a given service and port. type IngressBackend struct { // Specifies the name of the referenced service. - ServiceName string `json:"serviceName" protobuf:"bytes,1,opt,name=serviceName"` + // +optional + ServiceName string `json:"serviceName,omitempty" protobuf:"bytes,1,opt,name=serviceName"` // Specifies the port of the referenced service. - ServicePort intstr.IntOrString `json:"servicePort" protobuf:"bytes,2,opt,name=servicePort"` + // +optional + ServicePort intstr.IntOrString `json:"servicePort,omitempty" protobuf:"bytes,2,opt,name=servicePort"` + + // Resource is an ObjectRef to another Kubernetes resource in the namespace + // of the Ingress object. If resource is specified, serviceName and servicePort + // must not be specified. + // +optional + Resource *v1.TypedLocalObjectReference `json:"resource,omitempty" protobuf:"bytes,3,opt,name=resource"` } // +genclient // +genclient:method=GetScale,verb=get,subresource=scale,result=Scale // +genclient:method=UpdateScale,verb=update,subresource=scale,input=Scale,result=Scale // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ReplicaSet // DEPRECATED - This group version of ReplicaSet is deprecated by apps/v1beta2/ReplicaSet. See the release notes for // more information. @@ -736,6 +838,10 @@ type ReplicaSet struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.8 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=apps,v1,ReplicaSetList // ReplicaSetList is a collection of ReplicaSets. type ReplicaSetList struct { @@ -838,6 +944,10 @@ type ReplicaSetCondition struct { // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.11 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=policy,v1beta1,PodSecurityPolicy // PodSecurityPolicy governs the ability to make requests that affect the Security Context // that will be applied to a pod and container. @@ -875,7 +985,7 @@ type PodSecurityPolicySpec struct { // You must not list a capability in both allowedCapabilities and requiredDropCapabilities. // +optional AllowedCapabilities []v1.Capability `json:"allowedCapabilities,omitempty" protobuf:"bytes,4,rep,name=allowedCapabilities,casttype=k8s.io/api/core/v1.Capability"` - // volumes is a white list of allowed volume plugins. Empty indicates that + // volumes is an allowlist of volume plugins. Empty indicates that // no volumes may be used. To allow all volumes you may use '*'. // +optional Volumes []FSType `json:"volumes,omitempty" protobuf:"bytes,5,rep,name=volumes,casttype=FSType"` @@ -919,23 +1029,23 @@ type PodSecurityPolicySpec struct { // privilege escalation. If unspecified, defaults to true. // +optional AllowPrivilegeEscalation *bool `json:"allowPrivilegeEscalation,omitempty" protobuf:"varint,16,opt,name=allowPrivilegeEscalation"` - // allowedHostPaths is a white list of allowed host paths. Empty indicates + // allowedHostPaths is an allowlist of host paths. Empty indicates // that all host paths may be used. // +optional AllowedHostPaths []AllowedHostPath `json:"allowedHostPaths,omitempty" protobuf:"bytes,17,rep,name=allowedHostPaths"` - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all + // allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all // Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes // is allowed in the "volumes" field. // +optional AllowedFlexVolumes []AllowedFlexVolume `json:"allowedFlexVolumes,omitempty" protobuf:"bytes,18,rep,name=allowedFlexVolumes"` - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. + // AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. // An empty value indicates that any CSI driver can be used for inline ephemeral volumes. // +optional AllowedCSIDrivers []AllowedCSIDriver `json:"allowedCSIDrivers,omitempty" protobuf:"bytes,23,rep,name=allowedCSIDrivers"` // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. // Each entry is either a plain sysctl name or ends in "*" in which case it is considered // as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. - // Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. + // Kubelet has to allowlist all unsafe sysctls explicitly to avoid rejection. // // Examples: // e.g. "foo/*" allows "foo/bar", "foo/baz", etc. @@ -951,7 +1061,7 @@ type PodSecurityPolicySpec struct { // e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. // +optional ForbiddenSysctls []string `json:"forbiddenSysctls,omitempty" protobuf:"bytes,20,rep,name=forbiddenSysctls"` - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. + // AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. // Empty or nil indicates that only the DefaultProcMountType may be used. // This requires the ProcMountType feature flag to be enabled. // +optional @@ -1179,7 +1289,7 @@ const ( // RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses // for a pod. type RuntimeClassStrategyOptions struct { - // allowedRuntimeClassNames is a whitelist of RuntimeClass names that may be specified on a pod. + // allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. // A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the // list. An empty list requires the RuntimeClassName field to be unset. AllowedRuntimeClassNames []string `json:"allowedRuntimeClassNames" protobuf:"bytes,1,rep,name=allowedRuntimeClassNames"` @@ -1196,6 +1306,10 @@ type RuntimeClassStrategyOptions struct { const AllowAllRuntimeClassNames = "*" // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.2 +// +k8s:prerelease-lifecycle-gen:deprecated=1.11 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=policy,v1beta1,PodSecurityPolicyList // PodSecurityPolicyList is a list of PodSecurityPolicy objects. // Deprecated: use PodSecurityPolicyList from policy API Group instead. @@ -1212,6 +1326,10 @@ type PodSecurityPolicyList struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.3 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,NetworkPolicy // DEPRECATED 1.9 - This group version of NetworkPolicy is deprecated by networking/v1/NetworkPolicy. // NetworkPolicy describes what network traffic is allowed for a set of Pods @@ -1341,15 +1459,15 @@ type NetworkPolicyPort struct { } // DEPRECATED 1.9 - This group version of IPBlock is deprecated by networking/v1/IPBlock. -// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24") that is allowed to the pods -// matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should -// not be included within this rule. +// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24","2001:db9::/64") that is allowed +// to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs +// that should not be included within this rule. type IPBlock struct { // CIDR is a string representing the IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" CIDR string `json:"cidr" protobuf:"bytes,1,name=cidr"` // Except is a slice of CIDRs that should not be included within an IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" // Except values will be rejected if they are outside the CIDR range // +optional Except []string `json:"except,omitempty" protobuf:"bytes,2,rep,name=except"` @@ -1382,6 +1500,10 @@ type NetworkPolicyPeer struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.3 +// +k8s:prerelease-lifecycle-gen:deprecated=1.9 +// +k8s:prerelease-lifecycle-gen:removed=1.16 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,NetworkPolicyList // DEPRECATED 1.9 - This group version of NetworkPolicyList is deprecated by networking/v1/NetworkPolicyList. // Network Policy List is a list of NetworkPolicy objects. diff --git a/vendor/k8s.io/api/extensions/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/extensions/v1beta1/types_swagger_doc_generated.go index a7eb2ec9..0ef3c005 100644 --- a/vendor/k8s.io/api/extensions/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/extensions/v1beta1/types_swagger_doc_generated.go @@ -183,7 +183,7 @@ var map_DeploymentSpec = map[string]string{ "template": "Template describes the pods that will be created.", "strategy": "The deployment strategy to use to replace existing pods with new ones.", "minReadySeconds": "Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready)", - "revisionHistoryLimit": "The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. This is set to the max value of int32 (i.e. 2147483647) by default, which means \"retaining all old RelicaSets\".", + "revisionHistoryLimit": "The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. This is set to the max value of int32 (i.e. 2147483647) by default, which means \"retaining all old ReplicaSets\".", "paused": "Indicates that the deployment is paused and will not be processed by the deployment controller.", "rollbackTo": "DEPRECATED. The config this deployment is rolling back to. Will be cleared after rollback is done.", "progressDeadlineSeconds": "The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. This is set to the max value of int32 (i.e. 2147483647) by default, which means \"no deadline\".", @@ -230,9 +230,10 @@ func (FSGroupStrategyOptions) SwaggerDoc() map[string]string { } var map_HTTPIngressPath = map[string]string{ - "": "HTTPIngressPath associates a path regex with a backend. Incoming urls matching the path are forwarded to the backend.", - "path": "Path is an extended POSIX regex as defined by IEEE Std 1003.1, (i.e this follows the egrep/unix syntax, not the perl syntax) matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional \"path\" part of a URL as defined by RFC 3986. Paths must begin with a '/'. If unspecified, the path defaults to a catch all sending traffic to the backend.", - "backend": "Backend defines the referenced service endpoint to which the traffic will be forwarded to.", + "": "HTTPIngressPath associates a path with a backend. Incoming urls matching the path are forwarded to the backend.", + "path": "Path is matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional \"path\" part of a URL as defined by RFC 3986. Paths must begin with a '/'. When unspecified, all paths from incoming requests are matched.", + "pathType": "PathType determines the interpretation of the Path matching. PathType can be one of the following values: * Exact: Matches the URL path exactly. * Prefix: Matches based on a URL path prefix split by '/'. Matching is\n done on a path element by element basis. A path element refers is the\n list of labels in the path split by the '/' separator. A request is a\n match for path p if every p is an element-wise prefix of p of the\n request path. Note that if the last element of the path is a substring\n of the last element in request path, it is not a match (e.g. /foo/bar\n matches /foo/bar/baz, but does not match /foo/barbaz).\n* ImplementationSpecific: Interpretation of the Path matching is up to\n the IngressClass. Implementations can treat this as a separate PathType\n or treat it identically to Prefix or Exact path types.\nImplementations are required to support all path types. Defaults to ImplementationSpecific.", + "backend": "Backend defines the referenced service endpoint to which the traffic will be forwarded to.", } func (HTTPIngressPath) SwaggerDoc() map[string]string { @@ -269,9 +270,9 @@ func (IDRange) SwaggerDoc() map[string]string { } var map_IPBlock = map[string]string{ - "": "DEPRECATED 1.9 - This group version of IPBlock is deprecated by networking/v1/IPBlock. IPBlock describes a particular CIDR (Ex. \"192.168.1.1/24\") that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule.", - "cidr": "CIDR is a string representing the IP Block Valid examples are \"192.168.1.1/24\"", - "except": "Except is a slice of CIDRs that should not be included within an IP Block Valid examples are \"192.168.1.1/24\" Except values will be rejected if they are outside the CIDR range", + "": "DEPRECATED 1.9 - This group version of IPBlock is deprecated by networking/v1/IPBlock. IPBlock describes a particular CIDR (Ex. \"192.168.1.1/24\",\"2001:db9::/64\") that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule.", + "cidr": "CIDR is a string representing the IP Block Valid examples are \"192.168.1.1/24\" or \"2001:db9::/64\"", + "except": "Except is a slice of CIDRs that should not be included within an IP Block Valid examples are \"192.168.1.1/24\" or \"2001:db9::/64\" Except values will be rejected if they are outside the CIDR range", } func (IPBlock) SwaggerDoc() map[string]string { @@ -293,6 +294,7 @@ var map_IngressBackend = map[string]string{ "": "IngressBackend describes all endpoints for a given service and port.", "serviceName": "Specifies the name of the referenced service.", "servicePort": "Specifies the port of the referenced service.", + "resource": "Resource is an ObjectRef to another Kubernetes resource in the namespace of the Ingress object. If resource is specified, serviceName and servicePort must not be specified.", } func (IngressBackend) SwaggerDoc() map[string]string { @@ -311,7 +313,7 @@ func (IngressList) SwaggerDoc() map[string]string { var map_IngressRule = map[string]string{ "": "IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue.", - "host": "Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the \"host\" part of the URI as defined in the RFC: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the\n\t IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.", + "host": "Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the \"host\" part of the URI as defined in RFC 3986: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to\n the IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.\n\nHost can be \"precise\" which is a domain name without the terminating dot of a network host (e.g. \"foo.bar.com\") or \"wildcard\", which is a domain name prefixed with a single wildcard label (e.g. \"*.foo.com\"). The wildcard character '*' must appear by itself as the first DNS label and matches only a single label. You cannot have a wildcard label by itself (e.g. Host == \"*\"). Requests will be matched against the Host field in the following way: 1. If Host is precise, the request matches this rule if the http host header is equal to Host. 2. If Host is a wildcard, then the request matches this rule if the http host header is to equal to the suffix (removing the first label) of the wildcard rule.", } func (IngressRule) SwaggerDoc() map[string]string { @@ -327,10 +329,11 @@ func (IngressRuleValue) SwaggerDoc() map[string]string { } var map_IngressSpec = map[string]string{ - "": "IngressSpec describes the Ingress the user wishes to exist.", - "backend": "A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default.", - "tls": "TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.", - "rules": "A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.", + "": "IngressSpec describes the Ingress the user wishes to exist.", + "ingressClassName": "IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.", + "backend": "A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default.", + "tls": "TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.", + "rules": "A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.", } func (IngressSpec) SwaggerDoc() map[string]string { @@ -455,7 +458,7 @@ var map_PodSecurityPolicySpec = map[string]string{ "defaultAddCapabilities": "defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list.", "requiredDropCapabilities": "requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added.", "allowedCapabilities": "allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities.", - "volumes": "volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.", + "volumes": "volumes is an allowlist of volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.", "hostNetwork": "hostNetwork determines if the policy allows the use of HostNetwork in the pod spec.", "hostPorts": "hostPorts determines which host port ranges are allowed to be exposed.", "hostPID": "hostPID determines if the policy allows the use of HostPID in the pod spec.", @@ -468,12 +471,12 @@ var map_PodSecurityPolicySpec = map[string]string{ "readOnlyRootFilesystem": "readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to.", "defaultAllowPrivilegeEscalation": "defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process.", "allowPrivilegeEscalation": "allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true.", - "allowedHostPaths": "allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used.", - "allowedFlexVolumes": "allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the \"volumes\" field.", - "allowedCSIDrivers": "AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes.", - "allowedUnsafeSysctls": "allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in \"*\" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. \"foo/*\" allows \"foo/bar\", \"foo/baz\", etc. e.g. \"foo.*\" allows \"foo.bar\", \"foo.baz\", etc.", + "allowedHostPaths": "allowedHostPaths is an allowlist of host paths. Empty indicates that all host paths may be used.", + "allowedFlexVolumes": "allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the \"volumes\" field.", + "allowedCSIDrivers": "AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes.", + "allowedUnsafeSysctls": "allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in \"*\" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to allowlist all unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. \"foo/*\" allows \"foo/bar\", \"foo/baz\", etc. e.g. \"foo.*\" allows \"foo.bar\", \"foo.baz\", etc.", "forbiddenSysctls": "forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in \"*\" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden.\n\nExamples: e.g. \"foo/*\" forbids \"foo/bar\", \"foo/baz\", etc. e.g. \"foo.*\" forbids \"foo.bar\", \"foo.baz\", etc.", - "allowedProcMountTypes": "AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.", + "allowedProcMountTypes": "AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.", "runtimeClass": "runtimeClass is the strategy that will dictate the allowable RuntimeClasses for a pod. If this field is omitted, the pod's runtimeClassName field is unrestricted. Enforcement of this field depends on the RuntimeClass feature gate being enabled.", } @@ -541,14 +544,6 @@ func (ReplicaSetStatus) SwaggerDoc() map[string]string { return map_ReplicaSetStatus } -var map_ReplicationControllerDummy = map[string]string{ - "": "Dummy definition", -} - -func (ReplicationControllerDummy) SwaggerDoc() map[string]string { - return map_ReplicationControllerDummy -} - var map_RollbackConfig = map[string]string{ "": "DEPRECATED.", "revision": "The revision to rollback to. If set to 0, rollback to the last revision.", @@ -599,7 +594,7 @@ func (RunAsUserStrategyOptions) SwaggerDoc() map[string]string { var map_RuntimeClassStrategyOptions = map[string]string{ "": "RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses for a pod.", - "allowedRuntimeClassNames": "allowedRuntimeClassNames is a whitelist of RuntimeClass names that may be specified on a pod. A value of \"*\" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.", + "allowedRuntimeClassNames": "allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of \"*\" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.", "defaultRuntimeClassName": "defaultRuntimeClassName is the default RuntimeClassName to set on the pod. The default MUST be allowed by the allowedRuntimeClassNames list. A value of nil does not mutate the Pod.", } diff --git a/vendor/k8s.io/api/extensions/v1beta1/zz_generated.deepcopy.go b/vendor/k8s.io/api/extensions/v1beta1/zz_generated.deepcopy.go index cb610179..913f4851 100644 --- a/vendor/k8s.io/api/extensions/v1beta1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/extensions/v1beta1/zz_generated.deepcopy.go @@ -458,7 +458,12 @@ func (in *FSGroupStrategyOptions) DeepCopy() *FSGroupStrategyOptions { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *HTTPIngressPath) DeepCopyInto(out *HTTPIngressPath) { *out = *in - out.Backend = in.Backend + if in.PathType != nil { + in, out := &in.PathType, &out.PathType + *out = new(PathType) + **out = **in + } + in.Backend.DeepCopyInto(&out.Backend) return } @@ -478,7 +483,9 @@ func (in *HTTPIngressRuleValue) DeepCopyInto(out *HTTPIngressRuleValue) { if in.Paths != nil { in, out := &in.Paths, &out.Paths *out = make([]HTTPIngressPath, len(*in)) - copy(*out, *in) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } } return } @@ -578,6 +585,11 @@ func (in *Ingress) DeepCopyObject() runtime.Object { func (in *IngressBackend) DeepCopyInto(out *IngressBackend) { *out = *in out.ServicePort = in.ServicePort + if in.Resource != nil { + in, out := &in.Resource, &out.Resource + *out = new(corev1.TypedLocalObjectReference) + (*in).DeepCopyInto(*out) + } return } @@ -665,10 +677,15 @@ func (in *IngressRuleValue) DeepCopy() *IngressRuleValue { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *IngressSpec) DeepCopyInto(out *IngressSpec) { *out = *in + if in.IngressClassName != nil { + in, out := &in.IngressClassName, &out.IngressClassName + *out = new(string) + **out = **in + } if in.Backend != nil { in, out := &in.Backend, &out.Backend *out = new(IngressBackend) - **out = **in + (*in).DeepCopyInto(*out) } if in.TLS != nil { in, out := &in.TLS, &out.TLS @@ -1231,31 +1248,6 @@ func (in *ReplicaSetStatus) DeepCopy() *ReplicaSetStatus { return out } -// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. -func (in *ReplicationControllerDummy) DeepCopyInto(out *ReplicationControllerDummy) { - *out = *in - out.TypeMeta = in.TypeMeta - return -} - -// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ReplicationControllerDummy. -func (in *ReplicationControllerDummy) DeepCopy() *ReplicationControllerDummy { - if in == nil { - return nil - } - out := new(ReplicationControllerDummy) - in.DeepCopyInto(out) - return out -} - -// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. -func (in *ReplicationControllerDummy) DeepCopyObject() runtime.Object { - if c := in.DeepCopy(); c != nil { - return c - } - return nil -} - // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *RollbackConfig) DeepCopyInto(out *RollbackConfig) { *out = *in diff --git a/vendor/k8s.io/api/extensions/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/extensions/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..5023dd31 --- /dev/null +++ b/vendor/k8s.io/api/extensions/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,349 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DaemonSet) APILifecycleIntroduced() (major, minor int) { + return 1, 1 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DaemonSet) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DaemonSet) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DaemonSet"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DaemonSet) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DaemonSetList) APILifecycleIntroduced() (major, minor int) { + return 1, 1 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DaemonSetList) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DaemonSetList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DaemonSetList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DaemonSetList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Deployment) APILifecycleIntroduced() (major, minor int) { + return 1, 1 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Deployment) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Deployment) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "Deployment"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Deployment) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DeploymentList) APILifecycleIntroduced() (major, minor int) { + return 1, 1 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DeploymentList) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *DeploymentList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "DeploymentList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DeploymentList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *DeploymentRollback) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *DeploymentRollback) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *DeploymentRollback) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Ingress) APILifecycleIntroduced() (major, minor int) { + return 1, 1 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Ingress) APILifecycleDeprecated() (major, minor int) { + return 1, 14 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Ingress) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "Ingress"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Ingress) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *IngressList) APILifecycleIntroduced() (major, minor int) { + return 1, 1 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *IngressList) APILifecycleDeprecated() (major, minor int) { + return 1, 14 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *IngressList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "IngressList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *IngressList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *NetworkPolicy) APILifecycleIntroduced() (major, minor int) { + return 1, 3 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *NetworkPolicy) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *NetworkPolicy) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "NetworkPolicy"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *NetworkPolicy) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *NetworkPolicyList) APILifecycleIntroduced() (major, minor int) { + return 1, 3 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *NetworkPolicyList) APILifecycleDeprecated() (major, minor int) { + return 1, 9 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *NetworkPolicyList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "NetworkPolicyList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *NetworkPolicyList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PodSecurityPolicy) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PodSecurityPolicy) APILifecycleDeprecated() (major, minor int) { + return 1, 11 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *PodSecurityPolicy) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "policy", Version: "v1beta1", Kind: "PodSecurityPolicy"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PodSecurityPolicy) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PodSecurityPolicyList) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PodSecurityPolicyList) APILifecycleDeprecated() (major, minor int) { + return 1, 11 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *PodSecurityPolicyList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "policy", Version: "v1beta1", Kind: "PodSecurityPolicyList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PodSecurityPolicyList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ReplicaSet) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ReplicaSet) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ReplicaSet) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ReplicaSet"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ReplicaSet) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ReplicaSetList) APILifecycleIntroduced() (major, minor int) { + return 1, 2 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ReplicaSetList) APILifecycleDeprecated() (major, minor int) { + return 1, 8 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ReplicaSetList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apps", Version: "v1", Kind: "ReplicaSetList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ReplicaSetList) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Scale) APILifecycleIntroduced() (major, minor int) { + return 1, 1 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Scale) APILifecycleDeprecated() (major, minor int) { + return 1, 2 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Scale) APILifecycleRemoved() (major, minor int) { + return 1, 16 +} diff --git a/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.pb.go b/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.pb.go index d44ec3c9..86c86120 100644 --- a/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.pb.go +++ b/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.pb.go @@ -41,7 +41,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *FlowDistinguisherMethod) Reset() { *m = FlowDistinguisherMethod{} } func (*FlowDistinguisherMethod) ProtoMessage() {} @@ -5350,6 +5350,7 @@ func (m *UserSubject) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -5381,10 +5382,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -5405,55 +5404,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.proto b/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.proto index 6134b5e6..0801dd6c 100644 --- a/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.proto +++ b/vendor/k8s.io/api/flowcontrol/v1alpha1/generated.proto @@ -40,17 +40,17 @@ message FlowDistinguisherMethod { // similar attributes and is identified by a pair of strings: the name of the FlowSchema and a "flow distinguisher". message FlowSchema { // `metadata` is the standard object's metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; // `spec` is the specification of the desired behavior of a FlowSchema. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional optional FlowSchemaSpec spec = 2; // `status` is the current status of a FlowSchema. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional optional FlowSchemaStatus status = 3; } @@ -79,12 +79,11 @@ message FlowSchemaCondition { // FlowSchemaList is a list of FlowSchema objects. message FlowSchemaList { // `metadata` is the standard list metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; // `items` is a list of FlowSchemas. - // +listType=set repeated FlowSchema items = 2; } @@ -97,8 +96,8 @@ message FlowSchemaSpec { // `matchingPrecedence` is used to choose among the FlowSchemas that match a given request. The chosen // FlowSchema is among those with the numerically lowest (which we take to be logically highest) - // MatchingPrecedence. Each MatchingPrecedence value must be non-negative. - // Note that if the precedence is not specified or zero, it will be set to 1000 as default. + // MatchingPrecedence. Each MatchingPrecedence value must be ranged in [1,10000]. + // Note that if the precedence is not specified, it will be set to 1000 as default. // +optional optional int32 matchingPrecedence = 2; @@ -110,7 +109,7 @@ message FlowSchemaSpec { // `rules` describes which requests will match this flow schema. This FlowSchema matches a request if and only if // at least one member of rules matches the request. // if it is an empty slice, there will be no requests matching the FlowSchema. - // +listType=set + // +listType=atomic // +optional repeated PolicyRulesWithSubjects rules = 4; } @@ -210,20 +209,20 @@ message PolicyRulesWithSubjects { // subjects is the list of normal user, serviceaccount, or group that this rule cares about. // There must be at least one member in this slice. // A slice that includes both the system:authenticated and system:unauthenticated user groups matches every request. - // +listType=set + // +listType=atomic // Required. repeated Subject subjects = 1; // `resourceRules` is a slice of ResourcePolicyRules that identify matching requests according to their verb and the // target resource. // At least one of `resourceRules` and `nonResourceRules` has to be non-empty. - // +listType=set + // +listType=atomic // +optional repeated ResourcePolicyRule resourceRules = 2; // `nonResourceRules` is a list of NonResourcePolicyRules that identify matching requests according to their verb // and the target non-resource URL. - // +listType=set + // +listType=atomic // +optional repeated NonResourcePolicyRule nonResourceRules = 3; } @@ -231,17 +230,17 @@ message PolicyRulesWithSubjects { // PriorityLevelConfiguration represents the configuration of a priority level. message PriorityLevelConfiguration { // `metadata` is the standard object's metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; // `spec` is the specification of the desired behavior of a "request-priority". - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional optional PriorityLevelConfigurationSpec spec = 2; // `status` is the current status of a "request-priority". - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional optional PriorityLevelConfigurationStatus status = 3; } @@ -270,12 +269,11 @@ message PriorityLevelConfigurationCondition { // PriorityLevelConfigurationList is a list of PriorityLevelConfiguration objects. message PriorityLevelConfigurationList { // `metadata` is the standard object's metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; // `items` is a list of request-priorities. - // +listType=set repeated PriorityLevelConfiguration items = 2; } diff --git a/vendor/k8s.io/api/flowcontrol/v1alpha1/types.go b/vendor/k8s.io/api/flowcontrol/v1alpha1/types.go index 41073bdc..a67c6dd0 100644 --- a/vendor/k8s.io/api/flowcontrol/v1alpha1/types.go +++ b/vendor/k8s.io/api/flowcontrol/v1alpha1/types.go @@ -33,7 +33,10 @@ const ( // System preset priority level names const ( - PriorityLevelConfigurationNameExempt = "exempt" + PriorityLevelConfigurationNameExempt = "exempt" + PriorityLevelConfigurationNameCatchAll = "catch-all" + FlowSchemaNameExempt = "exempt" + FlowSchemaNameCatchAll = "catch-all" ) // Conditions @@ -43,6 +46,11 @@ const ( PriorityLevelConfigurationConditionConcurrencyShared = "ConcurrencyShared" ) +// Constants used by api validation. +const ( + FlowSchemaMaxMatchingPrecedence int32 = 10000 +) + // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object @@ -52,15 +60,15 @@ const ( type FlowSchema struct { metav1.TypeMeta `json:",inline"` // `metadata` is the standard object's metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` // `spec` is the specification of the desired behavior of a FlowSchema. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional Spec FlowSchemaSpec `json:"spec,omitempty" protobuf:"bytes,2,opt,name=spec"` // `status` is the current status of a FlowSchema. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional Status FlowSchemaStatus `json:"status,omitempty" protobuf:"bytes,3,opt,name=status"` } @@ -71,12 +79,11 @@ type FlowSchema struct { type FlowSchemaList struct { metav1.TypeMeta `json:",inline"` // `metadata` is the standard list metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` // `items` is a list of FlowSchemas. - // +listType=set Items []FlowSchema `json:"items" protobuf:"bytes,2,rep,name=items"` } @@ -88,8 +95,8 @@ type FlowSchemaSpec struct { PriorityLevelConfiguration PriorityLevelConfigurationReference `json:"priorityLevelConfiguration" protobuf:"bytes,1,opt,name=priorityLevelConfiguration"` // `matchingPrecedence` is used to choose among the FlowSchemas that match a given request. The chosen // FlowSchema is among those with the numerically lowest (which we take to be logically highest) - // MatchingPrecedence. Each MatchingPrecedence value must be non-negative. - // Note that if the precedence is not specified or zero, it will be set to 1000 as default. + // MatchingPrecedence. Each MatchingPrecedence value must be ranged in [1,10000]. + // Note that if the precedence is not specified, it will be set to 1000 as default. // +optional MatchingPrecedence int32 `json:"matchingPrecedence" protobuf:"varint,2,opt,name=matchingPrecedence"` // `distinguisherMethod` defines how to compute the flow distinguisher for requests that match this schema. @@ -99,7 +106,7 @@ type FlowSchemaSpec struct { // `rules` describes which requests will match this flow schema. This FlowSchema matches a request if and only if // at least one member of rules matches the request. // if it is an empty slice, there will be no requests matching the FlowSchema. - // +listType=set + // +listType=atomic // +optional Rules []PolicyRulesWithSubjects `json:"rules,omitempty" protobuf:"bytes,4,rep,name=rules"` } @@ -144,18 +151,18 @@ type PolicyRulesWithSubjects struct { // subjects is the list of normal user, serviceaccount, or group that this rule cares about. // There must be at least one member in this slice. // A slice that includes both the system:authenticated and system:unauthenticated user groups matches every request. - // +listType=set + // +listType=atomic // Required. Subjects []Subject `json:"subjects" protobuf:"bytes,1,rep,name=subjects"` // `resourceRules` is a slice of ResourcePolicyRules that identify matching requests according to their verb and the // target resource. // At least one of `resourceRules` and `nonResourceRules` has to be non-empty. - // +listType=set + // +listType=atomic // +optional ResourceRules []ResourcePolicyRule `json:"resourceRules,omitempty" protobuf:"bytes,2,opt,name=resourceRules"` // `nonResourceRules` is a list of NonResourcePolicyRules that identify matching requests according to their verb // and the target non-resource URL. - // +listType=set + // +listType=atomic // +optional NonResourceRules []NonResourcePolicyRule `json:"nonResourceRules,omitempty" protobuf:"bytes,3,opt,name=nonResourceRules"` } @@ -319,15 +326,15 @@ type FlowSchemaConditionType string type PriorityLevelConfiguration struct { metav1.TypeMeta `json:",inline"` // `metadata` is the standard object's metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` // `spec` is the specification of the desired behavior of a "request-priority". - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional Spec PriorityLevelConfigurationSpec `json:"spec,omitempty" protobuf:"bytes,2,opt,name=spec"` // `status` is the current status of a "request-priority". - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status // +optional Status PriorityLevelConfigurationStatus `json:"status,omitempty" protobuf:"bytes,3,opt,name=status"` } @@ -338,11 +345,10 @@ type PriorityLevelConfiguration struct { type PriorityLevelConfigurationList struct { metav1.TypeMeta `json:",inline"` // `metadata` is the standard object's metadata. - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata // +optional metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` // `items` is a list of request-priorities. - // +listType=set Items []PriorityLevelConfiguration `json:"items" protobuf:"bytes,2,rep,name=items"` } diff --git a/vendor/k8s.io/api/flowcontrol/v1alpha1/types_swagger_doc_generated.go b/vendor/k8s.io/api/flowcontrol/v1alpha1/types_swagger_doc_generated.go index 08380a1b..211d55e5 100644 --- a/vendor/k8s.io/api/flowcontrol/v1alpha1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/flowcontrol/v1alpha1/types_swagger_doc_generated.go @@ -38,9 +38,9 @@ func (FlowDistinguisherMethod) SwaggerDoc() map[string]string { var map_FlowSchema = map[string]string{ "": "FlowSchema defines the schema of a group of flows. Note that a flow is made up of a set of inbound API requests with similar attributes and is identified by a pair of strings: the name of the FlowSchema and a \"flow distinguisher\".", - "metadata": "`metadata` is the standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata", - "spec": "`spec` is the specification of the desired behavior of a FlowSchema. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status", - "status": "`status` is the current status of a FlowSchema. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status", + "metadata": "`metadata` is the standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "spec": "`spec` is the specification of the desired behavior of a FlowSchema. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", + "status": "`status` is the current status of a FlowSchema. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", } func (FlowSchema) SwaggerDoc() map[string]string { @@ -62,7 +62,7 @@ func (FlowSchemaCondition) SwaggerDoc() map[string]string { var map_FlowSchemaList = map[string]string{ "": "FlowSchemaList is a list of FlowSchema objects.", - "metadata": "`metadata` is the standard list metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata", + "metadata": "`metadata` is the standard list metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", "items": "`items` is a list of FlowSchemas.", } @@ -73,7 +73,7 @@ func (FlowSchemaList) SwaggerDoc() map[string]string { var map_FlowSchemaSpec = map[string]string{ "": "FlowSchemaSpec describes how the FlowSchema's specification looks like.", "priorityLevelConfiguration": "`priorityLevelConfiguration` should reference a PriorityLevelConfiguration in the cluster. If the reference cannot be resolved, the FlowSchema will be ignored and marked as invalid in its status. Required.", - "matchingPrecedence": "`matchingPrecedence` is used to choose among the FlowSchemas that match a given request. The chosen FlowSchema is among those with the numerically lowest (which we take to be logically highest) MatchingPrecedence. Each MatchingPrecedence value must be non-negative. Note that if the precedence is not specified or zero, it will be set to 1000 as default.", + "matchingPrecedence": "`matchingPrecedence` is used to choose among the FlowSchemas that match a given request. The chosen FlowSchema is among those with the numerically lowest (which we take to be logically highest) MatchingPrecedence. Each MatchingPrecedence value must be ranged in [1,10000]. Note that if the precedence is not specified, it will be set to 1000 as default.", "distinguisherMethod": "`distinguisherMethod` defines how to compute the flow distinguisher for requests that match this schema. `nil` specifies that the distinguisher is disabled and thus will always be the empty string.", "rules": "`rules` describes which requests will match this flow schema. This FlowSchema matches a request if and only if at least one member of rules matches the request. if it is an empty slice, there will be no requests matching the FlowSchema.", } @@ -143,9 +143,9 @@ func (PolicyRulesWithSubjects) SwaggerDoc() map[string]string { var map_PriorityLevelConfiguration = map[string]string{ "": "PriorityLevelConfiguration represents the configuration of a priority level.", - "metadata": "`metadata` is the standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata", - "spec": "`spec` is the specification of the desired behavior of a \"request-priority\". More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status", - "status": "`status` is the current status of a \"request-priority\". More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status", + "metadata": "`metadata` is the standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "spec": "`spec` is the specification of the desired behavior of a \"request-priority\". More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", + "status": "`status` is the current status of a \"request-priority\". More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", } func (PriorityLevelConfiguration) SwaggerDoc() map[string]string { @@ -167,7 +167,7 @@ func (PriorityLevelConfigurationCondition) SwaggerDoc() map[string]string { var map_PriorityLevelConfigurationList = map[string]string{ "": "PriorityLevelConfigurationList is a list of PriorityLevelConfiguration objects.", - "metadata": "`metadata` is the standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata", + "metadata": "`metadata` is the standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", "items": "`items` is a list of request-priorities.", } diff --git a/vendor/k8s.io/api/networking/v1/generated.pb.go b/vendor/k8s.io/api/networking/v1/generated.pb.go index c9b22fc7..4e03b543 100644 --- a/vendor/k8s.io/api/networking/v1/generated.pb.go +++ b/vendor/k8s.io/api/networking/v1/generated.pb.go @@ -25,8 +25,8 @@ import ( io "io" proto "github.com/gogo/protobuf/proto" - k8s_io_api_core_v1 "k8s.io/api/core/v1" + v11 "k8s.io/api/core/v1" v1 "k8s.io/apimachinery/pkg/apis/meta/v1" math "math" @@ -46,12 +46,68 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package + +func (m *HTTPIngressPath) Reset() { *m = HTTPIngressPath{} } +func (*HTTPIngressPath) ProtoMessage() {} +func (*HTTPIngressPath) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{0} +} +func (m *HTTPIngressPath) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *HTTPIngressPath) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *HTTPIngressPath) XXX_Merge(src proto.Message) { + xxx_messageInfo_HTTPIngressPath.Merge(m, src) +} +func (m *HTTPIngressPath) XXX_Size() int { + return m.Size() +} +func (m *HTTPIngressPath) XXX_DiscardUnknown() { + xxx_messageInfo_HTTPIngressPath.DiscardUnknown(m) +} + +var xxx_messageInfo_HTTPIngressPath proto.InternalMessageInfo + +func (m *HTTPIngressRuleValue) Reset() { *m = HTTPIngressRuleValue{} } +func (*HTTPIngressRuleValue) ProtoMessage() {} +func (*HTTPIngressRuleValue) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{1} +} +func (m *HTTPIngressRuleValue) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *HTTPIngressRuleValue) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *HTTPIngressRuleValue) XXX_Merge(src proto.Message) { + xxx_messageInfo_HTTPIngressRuleValue.Merge(m, src) +} +func (m *HTTPIngressRuleValue) XXX_Size() int { + return m.Size() +} +func (m *HTTPIngressRuleValue) XXX_DiscardUnknown() { + xxx_messageInfo_HTTPIngressRuleValue.DiscardUnknown(m) +} + +var xxx_messageInfo_HTTPIngressRuleValue proto.InternalMessageInfo func (m *IPBlock) Reset() { *m = IPBlock{} } func (*IPBlock) ProtoMessage() {} func (*IPBlock) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{0} + return fileDescriptor_1c72867a70a7cc90, []int{2} } func (m *IPBlock) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -76,10 +132,346 @@ func (m *IPBlock) XXX_DiscardUnknown() { var xxx_messageInfo_IPBlock proto.InternalMessageInfo +func (m *Ingress) Reset() { *m = Ingress{} } +func (*Ingress) ProtoMessage() {} +func (*Ingress) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{3} +} +func (m *Ingress) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *Ingress) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *Ingress) XXX_Merge(src proto.Message) { + xxx_messageInfo_Ingress.Merge(m, src) +} +func (m *Ingress) XXX_Size() int { + return m.Size() +} +func (m *Ingress) XXX_DiscardUnknown() { + xxx_messageInfo_Ingress.DiscardUnknown(m) +} + +var xxx_messageInfo_Ingress proto.InternalMessageInfo + +func (m *IngressBackend) Reset() { *m = IngressBackend{} } +func (*IngressBackend) ProtoMessage() {} +func (*IngressBackend) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{4} +} +func (m *IngressBackend) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressBackend) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressBackend) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressBackend.Merge(m, src) +} +func (m *IngressBackend) XXX_Size() int { + return m.Size() +} +func (m *IngressBackend) XXX_DiscardUnknown() { + xxx_messageInfo_IngressBackend.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressBackend proto.InternalMessageInfo + +func (m *IngressClass) Reset() { *m = IngressClass{} } +func (*IngressClass) ProtoMessage() {} +func (*IngressClass) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{5} +} +func (m *IngressClass) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressClass) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressClass) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressClass.Merge(m, src) +} +func (m *IngressClass) XXX_Size() int { + return m.Size() +} +func (m *IngressClass) XXX_DiscardUnknown() { + xxx_messageInfo_IngressClass.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressClass proto.InternalMessageInfo + +func (m *IngressClassList) Reset() { *m = IngressClassList{} } +func (*IngressClassList) ProtoMessage() {} +func (*IngressClassList) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{6} +} +func (m *IngressClassList) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressClassList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressClassList) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressClassList.Merge(m, src) +} +func (m *IngressClassList) XXX_Size() int { + return m.Size() +} +func (m *IngressClassList) XXX_DiscardUnknown() { + xxx_messageInfo_IngressClassList.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressClassList proto.InternalMessageInfo + +func (m *IngressClassSpec) Reset() { *m = IngressClassSpec{} } +func (*IngressClassSpec) ProtoMessage() {} +func (*IngressClassSpec) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{7} +} +func (m *IngressClassSpec) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressClassSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressClassSpec) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressClassSpec.Merge(m, src) +} +func (m *IngressClassSpec) XXX_Size() int { + return m.Size() +} +func (m *IngressClassSpec) XXX_DiscardUnknown() { + xxx_messageInfo_IngressClassSpec.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressClassSpec proto.InternalMessageInfo + +func (m *IngressList) Reset() { *m = IngressList{} } +func (*IngressList) ProtoMessage() {} +func (*IngressList) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{8} +} +func (m *IngressList) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressList) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressList.Merge(m, src) +} +func (m *IngressList) XXX_Size() int { + return m.Size() +} +func (m *IngressList) XXX_DiscardUnknown() { + xxx_messageInfo_IngressList.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressList proto.InternalMessageInfo + +func (m *IngressRule) Reset() { *m = IngressRule{} } +func (*IngressRule) ProtoMessage() {} +func (*IngressRule) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{9} +} +func (m *IngressRule) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressRule) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressRule) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressRule.Merge(m, src) +} +func (m *IngressRule) XXX_Size() int { + return m.Size() +} +func (m *IngressRule) XXX_DiscardUnknown() { + xxx_messageInfo_IngressRule.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressRule proto.InternalMessageInfo + +func (m *IngressRuleValue) Reset() { *m = IngressRuleValue{} } +func (*IngressRuleValue) ProtoMessage() {} +func (*IngressRuleValue) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{10} +} +func (m *IngressRuleValue) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressRuleValue) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressRuleValue) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressRuleValue.Merge(m, src) +} +func (m *IngressRuleValue) XXX_Size() int { + return m.Size() +} +func (m *IngressRuleValue) XXX_DiscardUnknown() { + xxx_messageInfo_IngressRuleValue.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressRuleValue proto.InternalMessageInfo + +func (m *IngressServiceBackend) Reset() { *m = IngressServiceBackend{} } +func (*IngressServiceBackend) ProtoMessage() {} +func (*IngressServiceBackend) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{11} +} +func (m *IngressServiceBackend) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressServiceBackend) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressServiceBackend) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressServiceBackend.Merge(m, src) +} +func (m *IngressServiceBackend) XXX_Size() int { + return m.Size() +} +func (m *IngressServiceBackend) XXX_DiscardUnknown() { + xxx_messageInfo_IngressServiceBackend.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressServiceBackend proto.InternalMessageInfo + +func (m *IngressSpec) Reset() { *m = IngressSpec{} } +func (*IngressSpec) ProtoMessage() {} +func (*IngressSpec) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{12} +} +func (m *IngressSpec) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressSpec) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressSpec.Merge(m, src) +} +func (m *IngressSpec) XXX_Size() int { + return m.Size() +} +func (m *IngressSpec) XXX_DiscardUnknown() { + xxx_messageInfo_IngressSpec.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressSpec proto.InternalMessageInfo + +func (m *IngressStatus) Reset() { *m = IngressStatus{} } +func (*IngressStatus) ProtoMessage() {} +func (*IngressStatus) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{13} +} +func (m *IngressStatus) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressStatus) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressStatus.Merge(m, src) +} +func (m *IngressStatus) XXX_Size() int { + return m.Size() +} +func (m *IngressStatus) XXX_DiscardUnknown() { + xxx_messageInfo_IngressStatus.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressStatus proto.InternalMessageInfo + +func (m *IngressTLS) Reset() { *m = IngressTLS{} } +func (*IngressTLS) ProtoMessage() {} +func (*IngressTLS) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{14} +} +func (m *IngressTLS) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressTLS) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressTLS) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressTLS.Merge(m, src) +} +func (m *IngressTLS) XXX_Size() int { + return m.Size() +} +func (m *IngressTLS) XXX_DiscardUnknown() { + xxx_messageInfo_IngressTLS.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressTLS proto.InternalMessageInfo + func (m *NetworkPolicy) Reset() { *m = NetworkPolicy{} } func (*NetworkPolicy) ProtoMessage() {} func (*NetworkPolicy) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{1} + return fileDescriptor_1c72867a70a7cc90, []int{15} } func (m *NetworkPolicy) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -107,7 +499,7 @@ var xxx_messageInfo_NetworkPolicy proto.InternalMessageInfo func (m *NetworkPolicyEgressRule) Reset() { *m = NetworkPolicyEgressRule{} } func (*NetworkPolicyEgressRule) ProtoMessage() {} func (*NetworkPolicyEgressRule) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{2} + return fileDescriptor_1c72867a70a7cc90, []int{16} } func (m *NetworkPolicyEgressRule) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -135,7 +527,7 @@ var xxx_messageInfo_NetworkPolicyEgressRule proto.InternalMessageInfo func (m *NetworkPolicyIngressRule) Reset() { *m = NetworkPolicyIngressRule{} } func (*NetworkPolicyIngressRule) ProtoMessage() {} func (*NetworkPolicyIngressRule) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{3} + return fileDescriptor_1c72867a70a7cc90, []int{17} } func (m *NetworkPolicyIngressRule) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -163,7 +555,7 @@ var xxx_messageInfo_NetworkPolicyIngressRule proto.InternalMessageInfo func (m *NetworkPolicyList) Reset() { *m = NetworkPolicyList{} } func (*NetworkPolicyList) ProtoMessage() {} func (*NetworkPolicyList) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{4} + return fileDescriptor_1c72867a70a7cc90, []int{18} } func (m *NetworkPolicyList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -191,7 +583,7 @@ var xxx_messageInfo_NetworkPolicyList proto.InternalMessageInfo func (m *NetworkPolicyPeer) Reset() { *m = NetworkPolicyPeer{} } func (*NetworkPolicyPeer) ProtoMessage() {} func (*NetworkPolicyPeer) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{5} + return fileDescriptor_1c72867a70a7cc90, []int{19} } func (m *NetworkPolicyPeer) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -219,7 +611,7 @@ var xxx_messageInfo_NetworkPolicyPeer proto.InternalMessageInfo func (m *NetworkPolicyPort) Reset() { *m = NetworkPolicyPort{} } func (*NetworkPolicyPort) ProtoMessage() {} func (*NetworkPolicyPort) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{6} + return fileDescriptor_1c72867a70a7cc90, []int{20} } func (m *NetworkPolicyPort) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -247,7 +639,7 @@ var xxx_messageInfo_NetworkPolicyPort proto.InternalMessageInfo func (m *NetworkPolicySpec) Reset() { *m = NetworkPolicySpec{} } func (*NetworkPolicySpec) ProtoMessage() {} func (*NetworkPolicySpec) Descriptor() ([]byte, []int) { - return fileDescriptor_1c72867a70a7cc90, []int{7} + return fileDescriptor_1c72867a70a7cc90, []int{21} } func (m *NetworkPolicySpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -272,114 +664,160 @@ func (m *NetworkPolicySpec) XXX_DiscardUnknown() { var xxx_messageInfo_NetworkPolicySpec proto.InternalMessageInfo -func init() { - proto.RegisterType((*IPBlock)(nil), "k8s.io.api.networking.v1.IPBlock") - proto.RegisterType((*NetworkPolicy)(nil), "k8s.io.api.networking.v1.NetworkPolicy") - proto.RegisterType((*NetworkPolicyEgressRule)(nil), "k8s.io.api.networking.v1.NetworkPolicyEgressRule") - proto.RegisterType((*NetworkPolicyIngressRule)(nil), "k8s.io.api.networking.v1.NetworkPolicyIngressRule") - proto.RegisterType((*NetworkPolicyList)(nil), "k8s.io.api.networking.v1.NetworkPolicyList") - proto.RegisterType((*NetworkPolicyPeer)(nil), "k8s.io.api.networking.v1.NetworkPolicyPeer") - proto.RegisterType((*NetworkPolicyPort)(nil), "k8s.io.api.networking.v1.NetworkPolicyPort") - proto.RegisterType((*NetworkPolicySpec)(nil), "k8s.io.api.networking.v1.NetworkPolicySpec") +func (m *ServiceBackendPort) Reset() { *m = ServiceBackendPort{} } +func (*ServiceBackendPort) ProtoMessage() {} +func (*ServiceBackendPort) Descriptor() ([]byte, []int) { + return fileDescriptor_1c72867a70a7cc90, []int{22} } - -func init() { - proto.RegisterFile("k8s.io/kubernetes/vendor/k8s.io/api/networking/v1/generated.proto", fileDescriptor_1c72867a70a7cc90) +func (m *ServiceBackendPort) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) } - -var fileDescriptor_1c72867a70a7cc90 = []byte{ - // 804 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x55, 0xcf, 0x8f, 0xdb, 0x44, - 0x14, 0x8e, 0x9d, 0x6c, 0x92, 0x4e, 0x28, 0x65, 0x07, 0x21, 0xac, 0x45, 0xd8, 0xc1, 0x17, 0x56, - 0xaa, 0x18, 0x93, 0x16, 0x21, 0x6e, 0x08, 0x43, 0x29, 0x91, 0xba, 0xbb, 0xd1, 0x6c, 0x2f, 0x20, - 0x90, 0x70, 0x9c, 0x59, 0xef, 0x34, 0xb1, 0xc7, 0x1a, 0x4f, 0x42, 0xf7, 0xc6, 0x9f, 0xc0, 0x1f, - 0xc2, 0x91, 0x1b, 0x87, 0x72, 0xdc, 0x63, 0x8f, 0x3d, 0x59, 0xac, 0xf9, 0x2f, 0xf6, 0x84, 0x66, - 0x3c, 0x89, 0xf3, 0xa3, 0x11, 0xd9, 0x15, 0xbd, 0x65, 0xde, 0xbc, 0xef, 0x7b, 0xf3, 0xde, 0xfb, - 0xf2, 0x19, 0x7c, 0x35, 0xfe, 0x22, 0x43, 0x94, 0x79, 0xe3, 0xe9, 0x90, 0xf0, 0x84, 0x08, 0x92, - 0x79, 0x33, 0x92, 0x8c, 0x18, 0xf7, 0xf4, 0x45, 0x90, 0x52, 0x2f, 0x21, 0xe2, 0x17, 0xc6, 0xc7, - 0x34, 0x89, 0xbc, 0x59, 0xcf, 0x8b, 0x48, 0x42, 0x78, 0x20, 0xc8, 0x08, 0xa5, 0x9c, 0x09, 0x06, - 0xad, 0x32, 0x13, 0x05, 0x29, 0x45, 0x55, 0x26, 0x9a, 0xf5, 0x0e, 0x3e, 0x89, 0xa8, 0x38, 0x9f, - 0x0e, 0x51, 0xc8, 0x62, 0x2f, 0x62, 0x11, 0xf3, 0x14, 0x60, 0x38, 0x3d, 0x53, 0x27, 0x75, 0x50, - 0xbf, 0x4a, 0xa2, 0x03, 0x77, 0xa9, 0x64, 0xc8, 0x38, 0x79, 0x4d, 0xb1, 0x83, 0xcf, 0xaa, 0x9c, - 0x38, 0x08, 0xcf, 0x69, 0x42, 0xf8, 0x85, 0x97, 0x8e, 0x23, 0x19, 0xc8, 0xbc, 0x98, 0x88, 0xe0, - 0x75, 0x28, 0x6f, 0x1b, 0x8a, 0x4f, 0x13, 0x41, 0x63, 0xb2, 0x01, 0xf8, 0xfc, 0xbf, 0x00, 0x59, - 0x78, 0x4e, 0xe2, 0x60, 0x03, 0xf7, 0x70, 0x1b, 0x6e, 0x2a, 0xe8, 0xc4, 0xa3, 0x89, 0xc8, 0x04, - 0x5f, 0x07, 0xb9, 0x27, 0xa0, 0xd5, 0x1f, 0xf8, 0x13, 0x16, 0x8e, 0x61, 0x17, 0x34, 0x42, 0x3a, - 0xe2, 0x96, 0xd1, 0x35, 0x0e, 0xef, 0xf8, 0x6f, 0x5d, 0xe6, 0x4e, 0xad, 0xc8, 0x9d, 0xc6, 0xd7, - 0xfd, 0x6f, 0x30, 0x56, 0x37, 0xd0, 0x05, 0x4d, 0xf2, 0x3c, 0x24, 0xa9, 0xb0, 0xcc, 0x6e, 0xfd, - 0xf0, 0x8e, 0x0f, 0x8a, 0xdc, 0x69, 0x3e, 0x52, 0x11, 0xac, 0x6f, 0xdc, 0xbf, 0x0c, 0x70, 0xf7, - 0xb8, 0xdc, 0xc4, 0x80, 0x4d, 0x68, 0x78, 0x01, 0x7f, 0x06, 0x6d, 0x39, 0x9b, 0x51, 0x20, 0x02, - 0xc5, 0xdd, 0x79, 0xf0, 0x29, 0xaa, 0xd6, 0xb6, 0x78, 0x2a, 0x4a, 0xc7, 0x91, 0x0c, 0x64, 0x48, - 0x66, 0xa3, 0x59, 0x0f, 0x9d, 0x0c, 0x9f, 0x91, 0x50, 0x1c, 0x11, 0x11, 0xf8, 0x50, 0xbf, 0x06, - 0x54, 0x31, 0xbc, 0x60, 0x85, 0x47, 0xa0, 0x91, 0xa5, 0x24, 0xb4, 0x4c, 0xc5, 0x7e, 0x1f, 0x6d, - 0x13, 0x05, 0x5a, 0x79, 0xd8, 0x69, 0x4a, 0xc2, 0xaa, 0x4d, 0x79, 0xc2, 0x8a, 0xc6, 0xfd, 0xc3, - 0x00, 0xef, 0xaf, 0x64, 0x3e, 0x8a, 0x38, 0xc9, 0x32, 0x3c, 0x9d, 0x10, 0x38, 0x00, 0x7b, 0x29, - 0xe3, 0x22, 0xb3, 0x8c, 0x6e, 0xfd, 0x06, 0xb5, 0x06, 0x8c, 0x0b, 0xff, 0xae, 0xae, 0xb5, 0x27, - 0x4f, 0x19, 0x2e, 0x89, 0xe0, 0x63, 0x60, 0x0a, 0xa6, 0x06, 0x7a, 0x03, 0x3a, 0x42, 0xb8, 0x0f, - 0x34, 0x9d, 0xf9, 0x94, 0x61, 0x53, 0x30, 0xf7, 0x4f, 0x03, 0x58, 0x2b, 0x59, 0xfd, 0xe4, 0x4d, - 0xbe, 0xfb, 0x08, 0x34, 0xce, 0x38, 0x8b, 0x6f, 0xf3, 0xf2, 0xc5, 0xd0, 0xbf, 0xe5, 0x2c, 0xc6, - 0x8a, 0xc6, 0x7d, 0x61, 0x80, 0xfd, 0x95, 0xcc, 0x27, 0x34, 0x13, 0xf0, 0xc7, 0x0d, 0xed, 0xa0, - 0xdd, 0xb4, 0x23, 0xd1, 0x4a, 0x39, 0xef, 0xe8, 0x5a, 0xed, 0x79, 0x64, 0x49, 0x37, 0x4f, 0xc0, - 0x1e, 0x15, 0x24, 0xce, 0x74, 0x0f, 0x1f, 0xef, 0xd8, 0x43, 0x35, 0x90, 0xbe, 0x44, 0xe3, 0x92, - 0xc4, 0x7d, 0x61, 0xae, 0x75, 0x20, 0x7b, 0x85, 0x67, 0xa0, 0x93, 0xb2, 0xd1, 0x29, 0x99, 0x90, - 0x50, 0x30, 0xae, 0x9b, 0x78, 0xb8, 0x63, 0x13, 0xc1, 0x90, 0x4c, 0xe6, 0x50, 0xff, 0x5e, 0x91, - 0x3b, 0x9d, 0x41, 0xc5, 0x85, 0x97, 0x89, 0xe1, 0x73, 0xb0, 0x9f, 0x04, 0x31, 0xc9, 0xd2, 0x20, - 0x24, 0x8b, 0x6a, 0xe6, 0xed, 0xab, 0xbd, 0x57, 0xe4, 0xce, 0xfe, 0xf1, 0x3a, 0x23, 0xde, 0x2c, - 0x02, 0xbf, 0x03, 0x2d, 0x9a, 0x2a, 0x0b, 0xb1, 0xea, 0xaa, 0xde, 0x47, 0xdb, 0xe7, 0xa8, 0xbd, - 0xc6, 0xef, 0x14, 0xb9, 0x33, 0x37, 0x1e, 0x3c, 0x87, 0xbb, 0xbf, 0xaf, 0x6b, 0x40, 0x0a, 0x0e, - 0x3e, 0x06, 0x6d, 0xe5, 0x55, 0x21, 0x9b, 0x68, 0x6f, 0xba, 0x2f, 0xf7, 0x39, 0xd0, 0xb1, 0xeb, - 0xdc, 0xf9, 0x60, 0xd3, 0xbc, 0xd1, 0xfc, 0x1a, 0x2f, 0xc0, 0xf0, 0x18, 0x34, 0xa4, 0x74, 0xf5, - 0x54, 0xb6, 0x9b, 0x90, 0xf4, 0x4b, 0x54, 0xfa, 0x25, 0xea, 0x27, 0xe2, 0x84, 0x9f, 0x0a, 0x4e, - 0x93, 0xc8, 0x6f, 0x4b, 0xc9, 0xca, 0x27, 0x61, 0xc5, 0xe3, 0x5e, 0xaf, 0x2f, 0x5c, 0x7a, 0x08, - 0x7c, 0xf6, 0xbf, 0x2d, 0xfc, 0x5d, 0x2d, 0xb3, 0xed, 0x4b, 0xff, 0x09, 0xb4, 0x68, 0xf9, 0x27, - 0xd7, 0x12, 0x7e, 0xb0, 0xa3, 0x84, 0x97, 0xac, 0xc1, 0xbf, 0xa7, 0xcb, 0xb4, 0xe6, 0xc1, 0x39, - 0x27, 0xfc, 0x1e, 0x34, 0x49, 0xc9, 0x5e, 0x57, 0xec, 0xbd, 0x1d, 0xd9, 0x2b, 0xbf, 0xf4, 0xdf, - 0xd6, 0xe4, 0x4d, 0x1d, 0xd3, 0x84, 0xf0, 0x4b, 0x39, 0x25, 0x99, 0xfb, 0xf4, 0x22, 0x25, 0x99, - 0xd5, 0x50, 0xdf, 0x93, 0x0f, 0xcb, 0x66, 0x17, 0xe1, 0xeb, 0xdc, 0x01, 0xd5, 0x11, 0x2f, 0x23, - 0xfc, 0xc3, 0xcb, 0x2b, 0xbb, 0xf6, 0xf2, 0xca, 0xae, 0xbd, 0xba, 0xb2, 0x6b, 0xbf, 0x16, 0xb6, - 0x71, 0x59, 0xd8, 0xc6, 0xcb, 0xc2, 0x36, 0x5e, 0x15, 0xb6, 0xf1, 0x77, 0x61, 0x1b, 0xbf, 0xfd, - 0x63, 0xd7, 0x7e, 0x30, 0x67, 0xbd, 0x7f, 0x03, 0x00, 0x00, 0xff, 0xff, 0xf7, 0x7b, 0xc9, 0x59, - 0x67, 0x08, 0x00, 0x00, -} - -func (m *IPBlock) Marshal() (dAtA []byte, err error) { - size := m.Size() - dAtA = make([]byte, size) - n, err := m.MarshalToSizedBuffer(dAtA[:size]) +func (m *ServiceBackendPort) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) if err != nil { return nil, err } - return dAtA[:n], nil + return b[:n], nil +} +func (m *ServiceBackendPort) XXX_Merge(src proto.Message) { + xxx_messageInfo_ServiceBackendPort.Merge(m, src) +} +func (m *ServiceBackendPort) XXX_Size() int { + return m.Size() +} +func (m *ServiceBackendPort) XXX_DiscardUnknown() { + xxx_messageInfo_ServiceBackendPort.DiscardUnknown(m) } -func (m *IPBlock) MarshalTo(dAtA []byte) (int, error) { - size := m.Size() - return m.MarshalToSizedBuffer(dAtA[:size]) +var xxx_messageInfo_ServiceBackendPort proto.InternalMessageInfo + +func init() { + proto.RegisterType((*HTTPIngressPath)(nil), "k8s.io.api.networking.v1.HTTPIngressPath") + proto.RegisterType((*HTTPIngressRuleValue)(nil), "k8s.io.api.networking.v1.HTTPIngressRuleValue") + proto.RegisterType((*IPBlock)(nil), "k8s.io.api.networking.v1.IPBlock") + proto.RegisterType((*Ingress)(nil), "k8s.io.api.networking.v1.Ingress") + proto.RegisterType((*IngressBackend)(nil), "k8s.io.api.networking.v1.IngressBackend") + proto.RegisterType((*IngressClass)(nil), "k8s.io.api.networking.v1.IngressClass") + proto.RegisterType((*IngressClassList)(nil), "k8s.io.api.networking.v1.IngressClassList") + proto.RegisterType((*IngressClassSpec)(nil), "k8s.io.api.networking.v1.IngressClassSpec") + proto.RegisterType((*IngressList)(nil), "k8s.io.api.networking.v1.IngressList") + proto.RegisterType((*IngressRule)(nil), "k8s.io.api.networking.v1.IngressRule") + proto.RegisterType((*IngressRuleValue)(nil), "k8s.io.api.networking.v1.IngressRuleValue") + proto.RegisterType((*IngressServiceBackend)(nil), "k8s.io.api.networking.v1.IngressServiceBackend") + proto.RegisterType((*IngressSpec)(nil), "k8s.io.api.networking.v1.IngressSpec") + proto.RegisterType((*IngressStatus)(nil), "k8s.io.api.networking.v1.IngressStatus") + proto.RegisterType((*IngressTLS)(nil), "k8s.io.api.networking.v1.IngressTLS") + proto.RegisterType((*NetworkPolicy)(nil), "k8s.io.api.networking.v1.NetworkPolicy") + proto.RegisterType((*NetworkPolicyEgressRule)(nil), "k8s.io.api.networking.v1.NetworkPolicyEgressRule") + proto.RegisterType((*NetworkPolicyIngressRule)(nil), "k8s.io.api.networking.v1.NetworkPolicyIngressRule") + proto.RegisterType((*NetworkPolicyList)(nil), "k8s.io.api.networking.v1.NetworkPolicyList") + proto.RegisterType((*NetworkPolicyPeer)(nil), "k8s.io.api.networking.v1.NetworkPolicyPeer") + proto.RegisterType((*NetworkPolicyPort)(nil), "k8s.io.api.networking.v1.NetworkPolicyPort") + proto.RegisterType((*NetworkPolicySpec)(nil), "k8s.io.api.networking.v1.NetworkPolicySpec") + proto.RegisterType((*ServiceBackendPort)(nil), "k8s.io.api.networking.v1.ServiceBackendPort") } -func (m *IPBlock) MarshalToSizedBuffer(dAtA []byte) (int, error) { - i := len(dAtA) - _ = i - var l int - _ = l - if len(m.Except) > 0 { - for iNdEx := len(m.Except) - 1; iNdEx >= 0; iNdEx-- { - i -= len(m.Except[iNdEx]) - copy(dAtA[i:], m.Except[iNdEx]) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.Except[iNdEx]))) - i-- - dAtA[i] = 0x12 - } - } - i -= len(m.CIDR) - copy(dAtA[i:], m.CIDR) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.CIDR))) - i-- - dAtA[i] = 0xa - return len(dAtA) - i, nil +func init() { + proto.RegisterFile("k8s.io/kubernetes/vendor/k8s.io/api/networking/v1/generated.proto", fileDescriptor_1c72867a70a7cc90) } -func (m *NetworkPolicy) Marshal() (dAtA []byte, err error) { +var fileDescriptor_1c72867a70a7cc90 = []byte{ + // 1441 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x57, 0xcd, 0x6f, 0x1b, 0x45, + 0x1b, 0xcf, 0x3a, 0x71, 0xec, 0x8c, 0xd3, 0x34, 0x9d, 0xb7, 0xd5, 0x6b, 0xf5, 0xd5, 0x6b, 0xe7, + 0x5d, 0xbd, 0xb4, 0x81, 0xd2, 0x35, 0x71, 0x2b, 0xc4, 0x0d, 0xd8, 0xf4, 0x2b, 0xe0, 0x26, 0xd6, + 0xd8, 0x2a, 0x02, 0x51, 0xd4, 0xf1, 0x7a, 0x62, 0x6f, 0xbd, 0xde, 0x59, 0x66, 0xc7, 0xa1, 0xbd, + 0x71, 0xe1, 0xc0, 0x8d, 0x7f, 0x81, 0x03, 0x37, 0x6e, 0x70, 0x43, 0x50, 0xb8, 0xa0, 0x1e, 0x7b, + 0xec, 0xc9, 0xa2, 0xe6, 0xbf, 0xc8, 0x09, 0xcd, 0xec, 0xec, 0xa7, 0x63, 0x6c, 0xaa, 0x2a, 0x27, + 0x7b, 0x9f, 0x8f, 0xdf, 0xf3, 0x39, 0xcf, 0x33, 0x03, 0xde, 0x1f, 0xbc, 0xe3, 0x1b, 0x36, 0xad, + 0x0d, 0x46, 0x1d, 0xc2, 0x5c, 0xc2, 0x89, 0x5f, 0x3b, 0x22, 0x6e, 0x97, 0xb2, 0x9a, 0x62, 0x60, + 0xcf, 0xae, 0xb9, 0x84, 0x7f, 0x41, 0xd9, 0xc0, 0x76, 0x7b, 0xb5, 0xa3, 0x9d, 0x5a, 0x8f, 0xb8, + 0x84, 0x61, 0x4e, 0xba, 0x86, 0xc7, 0x28, 0xa7, 0xb0, 0x1c, 0x48, 0x1a, 0xd8, 0xb3, 0x8d, 0x58, + 0xd2, 0x38, 0xda, 0xb9, 0x78, 0xb5, 0x67, 0xf3, 0xfe, 0xa8, 0x63, 0x58, 0x74, 0x58, 0xeb, 0xd1, + 0x1e, 0xad, 0x49, 0x85, 0xce, 0xe8, 0x50, 0x7e, 0xc9, 0x0f, 0xf9, 0x2f, 0x00, 0xba, 0xa8, 0x27, + 0x4c, 0x5a, 0x94, 0x91, 0x13, 0x8c, 0x5d, 0xbc, 0x1e, 0xcb, 0x0c, 0xb1, 0xd5, 0xb7, 0x5d, 0xc2, + 0x1e, 0xd7, 0xbc, 0x41, 0x4f, 0x10, 0xfc, 0xda, 0x90, 0x70, 0x7c, 0x92, 0x56, 0x6d, 0x96, 0x16, + 0x1b, 0xb9, 0xdc, 0x1e, 0x92, 0x29, 0x85, 0xb7, 0xe7, 0x29, 0xf8, 0x56, 0x9f, 0x0c, 0xf1, 0x94, + 0xde, 0xb5, 0x59, 0x7a, 0x23, 0x6e, 0x3b, 0x35, 0xdb, 0xe5, 0x3e, 0x67, 0x59, 0x25, 0xfd, 0x17, + 0x0d, 0x9c, 0xbd, 0xd3, 0x6e, 0x37, 0xf7, 0xdc, 0x1e, 0x23, 0xbe, 0xdf, 0xc4, 0xbc, 0x0f, 0xb7, + 0xc0, 0x8a, 0x87, 0x79, 0xbf, 0xac, 0x6d, 0x69, 0xdb, 0x6b, 0xe6, 0xfa, 0xd3, 0x71, 0x75, 0x69, + 0x32, 0xae, 0xae, 0x08, 0x1e, 0x92, 0x1c, 0x78, 0x1d, 0x14, 0xc5, 0x6f, 0xfb, 0xb1, 0x47, 0xca, + 0xcb, 0x52, 0xaa, 0x3c, 0x19, 0x57, 0x8b, 0x4d, 0x45, 0x3b, 0x4e, 0xfc, 0x47, 0x91, 0x24, 0x6c, + 0x81, 0x42, 0x07, 0x5b, 0x03, 0xe2, 0x76, 0xcb, 0xb9, 0x2d, 0x6d, 0xbb, 0x54, 0xdf, 0x36, 0x66, + 0x95, 0xcf, 0x50, 0xfe, 0x98, 0x81, 0xbc, 0x79, 0x56, 0x39, 0x51, 0x50, 0x04, 0x14, 0x22, 0xe9, + 0x87, 0xe0, 0x7c, 0xc2, 0x7f, 0x34, 0x72, 0xc8, 0x3d, 0xec, 0x8c, 0x08, 0xdc, 0x07, 0x79, 0x61, + 0xd8, 0x2f, 0x6b, 0x5b, 0xcb, 0xdb, 0xa5, 0xfa, 0xeb, 0xb3, 0x4d, 0x65, 0xc2, 0x37, 0xcf, 0x28, + 0x5b, 0x79, 0xf1, 0xe5, 0xa3, 0x00, 0x46, 0x3f, 0x00, 0x85, 0xbd, 0xa6, 0xe9, 0x50, 0x6b, 0x20, + 0xf2, 0x63, 0xd9, 0x5d, 0x96, 0xcd, 0xcf, 0xee, 0xde, 0x0d, 0x84, 0x24, 0x07, 0xea, 0x60, 0x95, + 0x3c, 0xb2, 0x88, 0xc7, 0xcb, 0xb9, 0xad, 0xe5, 0xed, 0x35, 0x13, 0x4c, 0xc6, 0xd5, 0xd5, 0x9b, + 0x92, 0x82, 0x14, 0x47, 0xff, 0x2a, 0x07, 0x0a, 0xca, 0x2c, 0x7c, 0x00, 0x8a, 0xa2, 0x7d, 0xba, + 0x98, 0x63, 0x89, 0x5a, 0xaa, 0xbf, 0x95, 0xf0, 0x37, 0xaa, 0xa6, 0xe1, 0x0d, 0x7a, 0x82, 0xe0, + 0x1b, 0x42, 0x5a, 0xf8, 0x7e, 0xd0, 0x79, 0x48, 0x2c, 0x7e, 0x97, 0x70, 0x6c, 0x42, 0xe5, 0x07, + 0x88, 0x69, 0x28, 0x42, 0x85, 0xb7, 0xc1, 0x8a, 0xef, 0x11, 0x4b, 0x25, 0xfe, 0xb5, 0xb9, 0x89, + 0x6f, 0x79, 0xc4, 0x8a, 0x43, 0x13, 0x5f, 0x48, 0x02, 0xc0, 0x03, 0xb0, 0xea, 0x73, 0xcc, 0x47, + 0xbe, 0x2c, 0x7c, 0xa9, 0x7e, 0x79, 0x3e, 0x94, 0x14, 0x37, 0x37, 0x14, 0xd8, 0x6a, 0xf0, 0x8d, + 0x14, 0x8c, 0xfe, 0x9b, 0x06, 0x36, 0xd2, 0xd5, 0x86, 0xf7, 0x40, 0xc1, 0x27, 0xec, 0xc8, 0xb6, + 0x48, 0x79, 0x45, 0x1a, 0xa9, 0xcd, 0x37, 0x12, 0xc8, 0x87, 0xfd, 0x52, 0x12, 0xbd, 0xa2, 0x68, + 0x28, 0x04, 0x83, 0x1f, 0x81, 0x22, 0x23, 0x3e, 0x1d, 0x31, 0x8b, 0x28, 0xef, 0xaf, 0x26, 0x81, + 0xc5, 0xb9, 0x17, 0x90, 0xa2, 0x59, 0xbb, 0x0d, 0x6a, 0x61, 0x27, 0x48, 0x25, 0x22, 0x87, 0x84, + 0x11, 0xd7, 0x22, 0xe6, 0xba, 0xe8, 0x72, 0xa4, 0x20, 0x50, 0x04, 0x26, 0x4e, 0xd1, 0xba, 0x72, + 0x64, 0xd7, 0xc1, 0xa7, 0x52, 0xd0, 0x46, 0xaa, 0xa0, 0x6f, 0xcc, 0x4d, 0x90, 0xf4, 0x6b, 0x56, + 0x55, 0xf5, 0x9f, 0x35, 0xb0, 0x99, 0x14, 0x6c, 0xd8, 0x3e, 0x87, 0x9f, 0x4e, 0x05, 0x61, 0x2c, + 0x16, 0x84, 0xd0, 0x96, 0x21, 0x6c, 0x2a, 0x53, 0xc5, 0x90, 0x92, 0x08, 0xe0, 0x43, 0x90, 0xb7, + 0x39, 0x19, 0xfa, 0xf2, 0x88, 0x94, 0xea, 0x97, 0x16, 0x8b, 0x20, 0x3e, 0x9d, 0x7b, 0x42, 0x19, + 0x05, 0x18, 0xfa, 0x77, 0x19, 0xff, 0x45, 0x68, 0xb0, 0x0e, 0x80, 0x45, 0x5d, 0xce, 0xa8, 0xe3, + 0x90, 0xf0, 0xb4, 0x46, 0x49, 0xdd, 0x8d, 0x38, 0x28, 0x21, 0x05, 0xef, 0x03, 0xe0, 0x61, 0x86, + 0x87, 0x84, 0x13, 0xe6, 0xab, 0xe4, 0xfe, 0xc3, 0x26, 0xd9, 0x10, 0xf0, 0xcd, 0x08, 0x04, 0x25, + 0x00, 0xf5, 0x1f, 0x34, 0x50, 0x52, 0x7e, 0x9e, 0x42, 0x8a, 0x6f, 0xa5, 0x53, 0xfc, 0xbf, 0xf9, + 0xe3, 0xf6, 0xe4, 0xec, 0x7e, 0x1b, 0x7b, 0x2d, 0x06, 0xac, 0x18, 0x80, 0x7d, 0xea, 0xf3, 0xec, + 0x00, 0xbc, 0x43, 0x7d, 0x8e, 0x24, 0x07, 0x7a, 0x60, 0xd3, 0xce, 0x4c, 0xe4, 0x85, 0x3b, 0x35, + 0xd2, 0x30, 0xcb, 0x0a, 0x79, 0x33, 0xcb, 0x41, 0x53, 0xe8, 0xfa, 0x03, 0x30, 0x25, 0x25, 0xce, + 0x48, 0x9f, 0x73, 0xef, 0x84, 0xcc, 0xce, 0x5e, 0x01, 0xb1, 0xf5, 0xa2, 0x8c, 0xa9, 0xdd, 0x6e, + 0x22, 0x89, 0xa2, 0x7f, 0xad, 0x81, 0x0b, 0x27, 0x4e, 0x1b, 0x91, 0x0f, 0x17, 0x0f, 0x49, 0x36, + 0x1f, 0xfb, 0x78, 0x48, 0x90, 0xe4, 0xc0, 0x7d, 0xb0, 0xe2, 0x51, 0xc6, 0x55, 0x0e, 0xde, 0x9c, + 0xed, 0x49, 0x1a, 0xb9, 0x49, 0x19, 0x4f, 0x2c, 0x60, 0xca, 0x38, 0x92, 0x38, 0xfa, 0xef, 0xb9, + 0xa8, 0x22, 0xb2, 0xd5, 0xdf, 0x8b, 0xf2, 0x2d, 0xdb, 0x5f, 0x58, 0x96, 0xa3, 0x73, 0xcd, 0x3c, + 0x9f, 0xc8, 0x5f, 0xc4, 0x43, 0x53, 0xd2, 0xb0, 0x0b, 0x36, 0xba, 0xe4, 0x10, 0x8f, 0x1c, 0xae, + 0x6c, 0xab, 0xac, 0x2d, 0xbe, 0xa3, 0xe1, 0x64, 0x5c, 0xdd, 0xb8, 0x91, 0xc2, 0x40, 0x19, 0x4c, + 0xb8, 0x0b, 0x96, 0xb9, 0x13, 0xf6, 0xe3, 0xff, 0xe7, 0x42, 0xb7, 0x1b, 0x2d, 0xb3, 0xa4, 0xc2, + 0x5f, 0x6e, 0x37, 0x5a, 0x48, 0x68, 0xc3, 0x0f, 0x40, 0x9e, 0x8d, 0x1c, 0x22, 0x36, 0xd0, 0xf2, + 0x42, 0xcb, 0x4c, 0xd4, 0x34, 0x6e, 0x6d, 0xf1, 0xe5, 0xa3, 0x00, 0x42, 0xff, 0x1c, 0x9c, 0x49, + 0xad, 0x29, 0xf8, 0x00, 0xac, 0x3b, 0x14, 0x77, 0x4d, 0xec, 0x60, 0xd7, 0x52, 0x63, 0x23, 0x33, + 0x9d, 0xc2, 0x11, 0xd0, 0x48, 0xc8, 0xa9, 0x25, 0x77, 0x5e, 0x19, 0x59, 0x4f, 0xf2, 0x50, 0x0a, + 0x51, 0xc7, 0x00, 0xc4, 0xe1, 0xc1, 0x2a, 0xc8, 0x8b, 0x13, 0x13, 0xdc, 0x53, 0xd6, 0xcc, 0x35, + 0xe1, 0xa1, 0x38, 0x48, 0x3e, 0x0a, 0xe8, 0x62, 0x8a, 0xf9, 0xc4, 0x62, 0x84, 0xcb, 0xa2, 0xe6, + 0xd2, 0x53, 0xac, 0x15, 0x71, 0x50, 0x42, 0x4a, 0xff, 0x55, 0x03, 0x67, 0xf6, 0x83, 0x4c, 0x34, + 0xa9, 0x63, 0x5b, 0x8f, 0x4f, 0x61, 0x21, 0xdd, 0x4d, 0x2d, 0xa4, 0x2b, 0xb3, 0x8b, 0x92, 0x72, + 0x6c, 0xe6, 0x46, 0xfa, 0x51, 0x03, 0xff, 0x4e, 0x49, 0xde, 0x8c, 0xe7, 0x4f, 0x13, 0xe4, 0xc5, + 0x29, 0x08, 0xef, 0x76, 0x8b, 0xda, 0x92, 0xa7, 0x29, 0xbe, 0xdd, 0x09, 0x04, 0x14, 0x00, 0xc1, + 0xdb, 0x20, 0xc7, 0xa9, 0x6a, 0xcb, 0x85, 0xe1, 0x08, 0x61, 0x26, 0x50, 0x70, 0xb9, 0x36, 0x45, + 0x39, 0x4e, 0xf5, 0x9f, 0x34, 0x50, 0x4e, 0x49, 0x25, 0xe7, 0xe6, 0xab, 0xf7, 0xfb, 0x2e, 0x58, + 0x39, 0x64, 0x74, 0xf8, 0x32, 0x9e, 0x47, 0x49, 0xbf, 0xc5, 0xe8, 0x10, 0x49, 0x18, 0xfd, 0x89, + 0x06, 0xce, 0xa5, 0x24, 0x4f, 0x61, 0x49, 0x35, 0xd2, 0x4b, 0xea, 0xf2, 0x82, 0x31, 0xcc, 0x58, + 0x55, 0x4f, 0x72, 0x99, 0x08, 0x44, 0xac, 0xf0, 0x10, 0x94, 0x3c, 0xda, 0x6d, 0x11, 0x87, 0x58, + 0x9c, 0x86, 0x67, 0xfa, 0xda, 0x82, 0x41, 0xe0, 0x0e, 0x71, 0x42, 0x55, 0xf3, 0xec, 0x64, 0x5c, + 0x2d, 0x35, 0x63, 0x2c, 0x94, 0x04, 0x86, 0x8f, 0xc0, 0x39, 0x31, 0xee, 0x7d, 0x0f, 0x5b, 0x24, + 0xb2, 0x96, 0x7b, 0x79, 0x6b, 0x17, 0x26, 0xe3, 0xea, 0xb9, 0xfd, 0x2c, 0x22, 0x9a, 0x36, 0x02, + 0xef, 0x80, 0x82, 0xed, 0xc9, 0xe7, 0x89, 0xba, 0xd9, 0xfe, 0xdd, 0xb2, 0x0f, 0xde, 0x31, 0xc1, + 0x25, 0x59, 0x7d, 0xa0, 0x50, 0x5d, 0xff, 0x3e, 0xdb, 0x03, 0xa2, 0xe1, 0xe0, 0x6d, 0x50, 0x94, + 0x0f, 0x46, 0x8b, 0x3a, 0x6a, 0xcd, 0x5d, 0x91, 0x2f, 0x3e, 0x45, 0x3b, 0x1e, 0x57, 0xff, 0x33, + 0xfd, 0x82, 0x36, 0x42, 0x36, 0x8a, 0x94, 0x33, 0x9b, 0x70, 0xf6, 0x10, 0x12, 0x8f, 0x56, 0x23, + 0x78, 0xb4, 0x1a, 0x7b, 0x2e, 0x3f, 0x60, 0x2d, 0xce, 0x6c, 0xb7, 0x17, 0x6c, 0xe5, 0xc4, 0x26, + 0x3c, 0xce, 0x16, 0x5c, 0xee, 0xc3, 0x87, 0xaf, 0xac, 0xe0, 0xff, 0x52, 0x6d, 0x36, 0xbb, 0xe8, + 0xf7, 0x41, 0x41, 0x6d, 0x53, 0xd5, 0xc2, 0xf5, 0x05, 0x5b, 0x38, 0xb9, 0x9d, 0xa2, 0x07, 0x6e, + 0x48, 0x0c, 0x31, 0xe1, 0xc7, 0x60, 0x95, 0x04, 0xe8, 0xc1, 0xba, 0xdb, 0x59, 0x10, 0x3d, 0x9e, + 0x97, 0xf1, 0xd3, 0x4b, 0xd1, 0x14, 0x20, 0x7c, 0x57, 0x64, 0x49, 0xc8, 0x8a, 0xcb, 0xac, 0x5f, + 0x5e, 0x91, 0x1b, 0xe8, 0xbf, 0x41, 0xb0, 0x11, 0xf9, 0x58, 0xdc, 0x66, 0xa3, 0x4f, 0x94, 0xd4, + 0xd0, 0x3f, 0x03, 0x70, 0xfa, 0xc2, 0xb2, 0xc0, 0x75, 0xe8, 0x12, 0x58, 0x75, 0x47, 0xc3, 0x0e, + 0x09, 0x0e, 0x47, 0x3e, 0x76, 0x70, 0x5f, 0x52, 0x91, 0xe2, 0x9a, 0xdb, 0x4f, 0x5f, 0x54, 0x96, + 0x9e, 0xbd, 0xa8, 0x2c, 0x3d, 0x7f, 0x51, 0x59, 0xfa, 0x72, 0x52, 0xd1, 0x9e, 0x4e, 0x2a, 0xda, + 0xb3, 0x49, 0x45, 0x7b, 0x3e, 0xa9, 0x68, 0x7f, 0x4c, 0x2a, 0xda, 0x37, 0x7f, 0x56, 0x96, 0x3e, + 0xc9, 0x1d, 0xed, 0xfc, 0x15, 0x00, 0x00, 0xff, 0xff, 0xc5, 0x87, 0xf6, 0x28, 0x4c, 0x12, 0x00, + 0x00, +} + +func (m *HTTPIngressPath) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) n, err := m.MarshalToSizedBuffer(dAtA[:size]) @@ -389,18 +827,25 @@ func (m *NetworkPolicy) Marshal() (dAtA []byte, err error) { return dAtA[:n], nil } -func (m *NetworkPolicy) MarshalTo(dAtA []byte) (int, error) { +func (m *HTTPIngressPath) MarshalTo(dAtA []byte) (int, error) { size := m.Size() return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicy) MarshalToSizedBuffer(dAtA []byte) (int, error) { +func (m *HTTPIngressPath) MarshalToSizedBuffer(dAtA []byte) (int, error) { i := len(dAtA) _ = i var l int _ = l + if m.PathType != nil { + i -= len(*m.PathType) + copy(dAtA[i:], *m.PathType) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.PathType))) + i-- + dAtA[i] = 0x1a + } { - size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + size, err := m.Backend.MarshalToSizedBuffer(dAtA[:i]) if err != nil { return 0, err } @@ -409,20 +854,15 @@ func (m *NetworkPolicy) MarshalToSizedBuffer(dAtA []byte) (int, error) { } i-- dAtA[i] = 0x12 - { - size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } + i -= len(m.Path) + copy(dAtA[i:], m.Path) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Path))) i-- dAtA[i] = 0xa return len(dAtA) - i, nil } -func (m *NetworkPolicyEgressRule) Marshal() (dAtA []byte, err error) { +func (m *HTTPIngressRuleValue) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) n, err := m.MarshalToSizedBuffer(dAtA[:size]) @@ -432,34 +872,20 @@ func (m *NetworkPolicyEgressRule) Marshal() (dAtA []byte, err error) { return dAtA[:n], nil } -func (m *NetworkPolicyEgressRule) MarshalTo(dAtA []byte) (int, error) { +func (m *HTTPIngressRuleValue) MarshalTo(dAtA []byte) (int, error) { size := m.Size() return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicyEgressRule) MarshalToSizedBuffer(dAtA []byte) (int, error) { +func (m *HTTPIngressRuleValue) MarshalToSizedBuffer(dAtA []byte) (int, error) { i := len(dAtA) _ = i var l int _ = l - if len(m.To) > 0 { - for iNdEx := len(m.To) - 1; iNdEx >= 0; iNdEx-- { - { - size, err := m.To[iNdEx].MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x12 - } - } - if len(m.Ports) > 0 { - for iNdEx := len(m.Ports) - 1; iNdEx >= 0; iNdEx-- { + if len(m.Paths) > 0 { + for iNdEx := len(m.Paths) - 1; iNdEx >= 0; iNdEx-- { { - size, err := m.Ports[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + size, err := m.Paths[iNdEx].MarshalToSizedBuffer(dAtA[:i]) if err != nil { return 0, err } @@ -473,7 +899,7 @@ func (m *NetworkPolicyEgressRule) MarshalToSizedBuffer(dAtA []byte) (int, error) return len(dAtA) - i, nil } -func (m *NetworkPolicyIngressRule) Marshal() (dAtA []byte, err error) { +func (m *IPBlock) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) n, err := m.MarshalToSizedBuffer(dAtA[:size]) @@ -483,48 +909,34 @@ func (m *NetworkPolicyIngressRule) Marshal() (dAtA []byte, err error) { return dAtA[:n], nil } -func (m *NetworkPolicyIngressRule) MarshalTo(dAtA []byte) (int, error) { +func (m *IPBlock) MarshalTo(dAtA []byte) (int, error) { size := m.Size() return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicyIngressRule) MarshalToSizedBuffer(dAtA []byte) (int, error) { +func (m *IPBlock) MarshalToSizedBuffer(dAtA []byte) (int, error) { i := len(dAtA) _ = i var l int _ = l - if len(m.From) > 0 { - for iNdEx := len(m.From) - 1; iNdEx >= 0; iNdEx-- { - { - size, err := m.From[iNdEx].MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } + if len(m.Except) > 0 { + for iNdEx := len(m.Except) - 1; iNdEx >= 0; iNdEx-- { + i -= len(m.Except[iNdEx]) + copy(dAtA[i:], m.Except[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Except[iNdEx]))) i-- dAtA[i] = 0x12 } } - if len(m.Ports) > 0 { - for iNdEx := len(m.Ports) - 1; iNdEx >= 0; iNdEx-- { - { - size, err := m.Ports[iNdEx].MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0xa - } - } + i -= len(m.CIDR) + copy(dAtA[i:], m.CIDR) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.CIDR))) + i-- + dAtA[i] = 0xa return len(dAtA) - i, nil } -func (m *NetworkPolicyList) Marshal() (dAtA []byte, err error) { +func (m *Ingress) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) n, err := m.MarshalToSizedBuffer(dAtA[:size]) @@ -534,32 +946,38 @@ func (m *NetworkPolicyList) Marshal() (dAtA []byte, err error) { return dAtA[:n], nil } -func (m *NetworkPolicyList) MarshalTo(dAtA []byte) (int, error) { +func (m *Ingress) MarshalTo(dAtA []byte) (int, error) { size := m.Size() return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicyList) MarshalToSizedBuffer(dAtA []byte) (int, error) { +func (m *Ingress) MarshalToSizedBuffer(dAtA []byte) (int, error) { i := len(dAtA) _ = i var l int _ = l - if len(m.Items) > 0 { - for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { - { - size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x12 + { + size, err := m.Status.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } + i-- + dAtA[i] = 0x1a { - size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) if err != nil { return 0, err } @@ -571,7 +989,7 @@ func (m *NetworkPolicyList) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } -func (m *NetworkPolicyPeer) Marshal() (dAtA []byte, err error) { +func (m *IngressBackend) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) n, err := m.MarshalToSizedBuffer(dAtA[:size]) @@ -581,31 +999,19 @@ func (m *NetworkPolicyPeer) Marshal() (dAtA []byte, err error) { return dAtA[:n], nil } -func (m *NetworkPolicyPeer) MarshalTo(dAtA []byte) (int, error) { +func (m *IngressBackend) MarshalTo(dAtA []byte) (int, error) { size := m.Size() return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicyPeer) MarshalToSizedBuffer(dAtA []byte) (int, error) { +func (m *IngressBackend) MarshalToSizedBuffer(dAtA []byte) (int, error) { i := len(dAtA) _ = i var l int _ = l - if m.IPBlock != nil { - { - size, err := m.IPBlock.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x1a - } - if m.NamespaceSelector != nil { + if m.Service != nil { { - size, err := m.NamespaceSelector.MarshalToSizedBuffer(dAtA[:i]) + size, err := m.Service.MarshalToSizedBuffer(dAtA[:i]) if err != nil { return 0, err } @@ -613,11 +1019,11 @@ func (m *NetworkPolicyPeer) MarshalToSizedBuffer(dAtA []byte) (int, error) { i = encodeVarintGenerated(dAtA, i, uint64(size)) } i-- - dAtA[i] = 0x12 + dAtA[i] = 0x22 } - if m.PodSelector != nil { + if m.Resource != nil { { - size, err := m.PodSelector.MarshalToSizedBuffer(dAtA[:i]) + size, err := m.Resource.MarshalToSizedBuffer(dAtA[:i]) if err != nil { return 0, err } @@ -625,12 +1031,12 @@ func (m *NetworkPolicyPeer) MarshalToSizedBuffer(dAtA []byte) (int, error) { i = encodeVarintGenerated(dAtA, i, uint64(size)) } i-- - dAtA[i] = 0xa + dAtA[i] = 0x1a } return len(dAtA) - i, nil } -func (m *NetworkPolicyPort) Marshal() (dAtA []byte, err error) { +func (m *IngressClass) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) n, err := m.MarshalToSizedBuffer(dAtA[:size]) @@ -640,39 +1046,40 @@ func (m *NetworkPolicyPort) Marshal() (dAtA []byte, err error) { return dAtA[:n], nil } -func (m *NetworkPolicyPort) MarshalTo(dAtA []byte) (int, error) { +func (m *IngressClass) MarshalTo(dAtA []byte) (int, error) { size := m.Size() return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicyPort) MarshalToSizedBuffer(dAtA []byte) (int, error) { +func (m *IngressClass) MarshalToSizedBuffer(dAtA []byte) (int, error) { i := len(dAtA) _ = i var l int _ = l - if m.Port != nil { - { - size, err := m.Port.MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) + { + size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err } - i-- - dAtA[i] = 0x12 + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } - if m.Protocol != nil { - i -= len(*m.Protocol) - copy(dAtA[i:], *m.Protocol) - i = encodeVarintGenerated(dAtA, i, uint64(len(*m.Protocol))) - i-- - dAtA[i] = 0xa + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } + i-- + dAtA[i] = 0xa return len(dAtA) - i, nil } -func (m *NetworkPolicySpec) Marshal() (dAtA []byte, err error) { +func (m *IngressClassList) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) n, err := m.MarshalToSizedBuffer(dAtA[:size]) @@ -682,43 +1089,20 @@ func (m *NetworkPolicySpec) Marshal() (dAtA []byte, err error) { return dAtA[:n], nil } -func (m *NetworkPolicySpec) MarshalTo(dAtA []byte) (int, error) { +func (m *IngressClassList) MarshalTo(dAtA []byte) (int, error) { size := m.Size() return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicySpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { +func (m *IngressClassList) MarshalToSizedBuffer(dAtA []byte) (int, error) { i := len(dAtA) _ = i var l int _ = l - if len(m.PolicyTypes) > 0 { - for iNdEx := len(m.PolicyTypes) - 1; iNdEx >= 0; iNdEx-- { - i -= len(m.PolicyTypes[iNdEx]) - copy(dAtA[i:], m.PolicyTypes[iNdEx]) - i = encodeVarintGenerated(dAtA, i, uint64(len(m.PolicyTypes[iNdEx]))) - i-- - dAtA[i] = 0x22 - } - } - if len(m.Egress) > 0 { - for iNdEx := len(m.Egress) - 1; iNdEx >= 0; iNdEx-- { - { - size, err := m.Egress[iNdEx].MarshalToSizedBuffer(dAtA[:i]) - if err != nil { - return 0, err - } - i -= size - i = encodeVarintGenerated(dAtA, i, uint64(size)) - } - i-- - dAtA[i] = 0x1a - } - } - if len(m.Ingress) > 0 { - for iNdEx := len(m.Ingress) - 1; iNdEx >= 0; iNdEx-- { + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { { - size, err := m.Ingress[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) if err != nil { return 0, err } @@ -730,7 +1114,7 @@ func (m *NetworkPolicySpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { } } { - size, err := m.PodSelector.MarshalToSizedBuffer(dAtA[:i]) + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) if err != nil { return 0, err } @@ -742,314 +1126,3184 @@ func (m *NetworkPolicySpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } -func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { - offset -= sovGenerated(v) - base := offset - for v >= 1<<7 { - dAtA[offset] = uint8(v&0x7f | 0x80) - v >>= 7 - offset++ +func (m *IngressClassSpec) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - dAtA[offset] = uint8(v) - return base + return dAtA[:n], nil } -func (m *IPBlock) Size() (n int) { - if m == nil { - return 0 - } + +func (m *IngressClassSpec) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressClassSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i var l int _ = l - l = len(m.CIDR) - n += 1 + l + sovGenerated(uint64(l)) - if len(m.Except) > 0 { - for _, s := range m.Except { - l = len(s) - n += 1 + l + sovGenerated(uint64(l)) + if m.Parameters != nil { + { + size, err := m.Parameters.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } + i-- + dAtA[i] = 0x12 } - return n + i -= len(m.Controller) + copy(dAtA[i:], m.Controller) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Controller))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil } -func (m *NetworkPolicy) Size() (n int) { - if m == nil { - return 0 +func (m *IngressList) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - var l int - _ = l - l = m.ObjectMeta.Size() - n += 1 + l + sovGenerated(uint64(l)) - l = m.Spec.Size() - n += 1 + l + sovGenerated(uint64(l)) - return n + return dAtA[:n], nil } -func (m *NetworkPolicyEgressRule) Size() (n int) { - if m == nil { - return 0 - } +func (m *IngressList) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressList) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i var l int _ = l - if len(m.Ports) > 0 { - for _, e := range m.Ports { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 } } - if len(m.To) > 0 { - for _, e := range m.To { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) + { + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } - return n + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil } -func (m *NetworkPolicyIngressRule) Size() (n int) { - if m == nil { - return 0 +func (m *IngressRule) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } + return dAtA[:n], nil +} + +func (m *IngressRule) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressRule) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i var l int _ = l - if len(m.Ports) > 0 { - for _, e := range m.Ports { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - } - if len(m.From) > 0 { - for _, e := range m.From { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) + { + size, err := m.IngressRuleValue.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } - return n + i-- + dAtA[i] = 0x12 + i -= len(m.Host) + copy(dAtA[i:], m.Host) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Host))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil } -func (m *NetworkPolicyList) Size() (n int) { - if m == nil { - return 0 +func (m *IngressRuleValue) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } + return dAtA[:n], nil +} + +func (m *IngressRuleValue) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressRuleValue) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i var l int _ = l - l = m.ListMeta.Size() - n += 1 + l + sovGenerated(uint64(l)) - if len(m.Items) > 0 { - for _, e := range m.Items { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) + if m.HTTP != nil { + { + size, err := m.HTTP.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } + i-- + dAtA[i] = 0xa } - return n + return len(dAtA) - i, nil } -func (m *NetworkPolicyPeer) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - if m.PodSelector != nil { - l = m.PodSelector.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - if m.NamespaceSelector != nil { - l = m.NamespaceSelector.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - if m.IPBlock != nil { - l = m.IPBlock.Size() - n += 1 + l + sovGenerated(uint64(l)) +func (m *IngressServiceBackend) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - return n + return dAtA[:n], nil } -func (m *NetworkPolicyPort) Size() (n int) { - if m == nil { - return 0 - } - var l int - _ = l - if m.Protocol != nil { - l = len(*m.Protocol) - n += 1 + l + sovGenerated(uint64(l)) - } - if m.Port != nil { - l = m.Port.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - return n +func (m *IngressServiceBackend) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) } -func (m *NetworkPolicySpec) Size() (n int) { - if m == nil { - return 0 - } +func (m *IngressServiceBackend) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i var l int _ = l - l = m.PodSelector.Size() - n += 1 + l + sovGenerated(uint64(l)) - if len(m.Ingress) > 0 { - for _, e := range m.Ingress { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) - } - } - if len(m.Egress) > 0 { - for _, e := range m.Egress { - l = e.Size() - n += 1 + l + sovGenerated(uint64(l)) + { + size, err := m.Port.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } - if len(m.PolicyTypes) > 0 { - for _, s := range m.PolicyTypes { - l = len(s) - n += 1 + l + sovGenerated(uint64(l)) - } + i-- + dAtA[i] = 0x12 + i -= len(m.Name) + copy(dAtA[i:], m.Name) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Name))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *IngressSpec) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - return n + return dAtA[:n], nil } -func sovGenerated(x uint64) (n int) { - return (math_bits.Len64(x|1) + 6) / 7 +func (m *IngressSpec) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) } -func sozGenerated(x uint64) (n int) { - return sovGenerated(uint64((x << 1) ^ uint64((int64(x) >> 63)))) + +func (m *IngressSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.IngressClassName != nil { + i -= len(*m.IngressClassName) + copy(dAtA[i:], *m.IngressClassName) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.IngressClassName))) + i-- + dAtA[i] = 0x22 + } + if len(m.Rules) > 0 { + for iNdEx := len(m.Rules) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Rules[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1a + } + } + if len(m.TLS) > 0 { + for iNdEx := len(m.TLS) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.TLS[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + if m.DefaultBackend != nil { + { + size, err := m.DefaultBackend.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil } -func (this *IPBlock) String() string { - if this == nil { - return "nil" + +func (m *IngressStatus) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - s := strings.Join([]string{`&IPBlock{`, - `CIDR:` + fmt.Sprintf("%v", this.CIDR) + `,`, - `Except:` + fmt.Sprintf("%v", this.Except) + `,`, - `}`, - }, "") - return s + return dAtA[:n], nil } -func (this *NetworkPolicy) String() string { - if this == nil { - return "nil" + +func (m *IngressStatus) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressStatus) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + { + size, err := m.LoadBalancer.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } - s := strings.Join([]string{`&NetworkPolicy{`, - `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, - `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "NetworkPolicySpec", "NetworkPolicySpec", 1), `&`, ``, 1) + `,`, - `}`, - }, "") - return s + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil } -func (this *NetworkPolicyEgressRule) String() string { - if this == nil { - return "nil" + +func (m *IngressTLS) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - repeatedStringForPorts := "[]NetworkPolicyPort{" - for _, f := range this.Ports { - repeatedStringForPorts += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPort", "NetworkPolicyPort", 1), `&`, ``, 1) + "," + return dAtA[:n], nil +} + +func (m *IngressTLS) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressTLS) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i -= len(m.SecretName) + copy(dAtA[i:], m.SecretName) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.SecretName))) + i-- + dAtA[i] = 0x12 + if len(m.Hosts) > 0 { + for iNdEx := len(m.Hosts) - 1; iNdEx >= 0; iNdEx-- { + i -= len(m.Hosts[iNdEx]) + copy(dAtA[i:], m.Hosts[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Hosts[iNdEx]))) + i-- + dAtA[i] = 0xa + } } - repeatedStringForPorts += "}" - repeatedStringForTo := "[]NetworkPolicyPeer{" - for _, f := range this.To { - repeatedStringForTo += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPeer", "NetworkPolicyPeer", 1), `&`, ``, 1) + "," + return len(dAtA) - i, nil +} + +func (m *NetworkPolicy) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - repeatedStringForTo += "}" - s := strings.Join([]string{`&NetworkPolicyEgressRule{`, - `Ports:` + repeatedStringForPorts + `,`, - `To:` + repeatedStringForTo + `,`, - `}`, - }, "") - return s + return dAtA[:n], nil } -func (this *NetworkPolicyIngressRule) String() string { - if this == nil { - return "nil" + +func (m *NetworkPolicy) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *NetworkPolicy) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + { + size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } - repeatedStringForPorts := "[]NetworkPolicyPort{" - for _, f := range this.Ports { - repeatedStringForPorts += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPort", "NetworkPolicyPort", 1), `&`, ``, 1) + "," + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) } - repeatedStringForPorts += "}" - repeatedStringForFrom := "[]NetworkPolicyPeer{" - for _, f := range this.From { - repeatedStringForFrom += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPeer", "NetworkPolicyPeer", 1), `&`, ``, 1) + "," + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *NetworkPolicyEgressRule) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err } - repeatedStringForFrom += "}" - s := strings.Join([]string{`&NetworkPolicyIngressRule{`, - `Ports:` + repeatedStringForPorts + `,`, - `From:` + repeatedStringForFrom + `,`, - `}`, - }, "") - return s + return dAtA[:n], nil } -func (this *NetworkPolicyList) String() string { - if this == nil { - return "nil" + +func (m *NetworkPolicyEgressRule) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *NetworkPolicyEgressRule) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.To) > 0 { + for iNdEx := len(m.To) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.To[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } } - repeatedStringForItems := "[]NetworkPolicy{" - for _, f := range this.Items { - repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "NetworkPolicy", "NetworkPolicy", 1), `&`, ``, 1) + "," + if len(m.Ports) > 0 { + for iNdEx := len(m.Ports) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Ports[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + } } - repeatedStringForItems += "}" - s := strings.Join([]string{`&NetworkPolicyList{`, - `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, - `Items:` + repeatedStringForItems + `,`, - `}`, - }, "") - return s + return len(dAtA) - i, nil } -func (this *NetworkPolicyPeer) String() string { - if this == nil { - return "nil" + +func (m *NetworkPolicyIngressRule) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *NetworkPolicyIngressRule) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *NetworkPolicyIngressRule) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.From) > 0 { + for iNdEx := len(m.From) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.From[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + if len(m.Ports) > 0 { + for iNdEx := len(m.Ports) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Ports[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + } + } + return len(dAtA) - i, nil +} + +func (m *NetworkPolicyList) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *NetworkPolicyList) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *NetworkPolicyList) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + { + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *NetworkPolicyPeer) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *NetworkPolicyPeer) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *NetworkPolicyPeer) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.IPBlock != nil { + { + size, err := m.IPBlock.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1a + } + if m.NamespaceSelector != nil { + { + size, err := m.NamespaceSelector.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + if m.PodSelector != nil { + { + size, err := m.PodSelector.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil +} + +func (m *NetworkPolicyPort) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *NetworkPolicyPort) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *NetworkPolicyPort) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.Port != nil { + { + size, err := m.Port.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + if m.Protocol != nil { + i -= len(*m.Protocol) + copy(dAtA[i:], *m.Protocol) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.Protocol))) + i-- + dAtA[i] = 0xa + } + return len(dAtA) - i, nil +} + +func (m *NetworkPolicySpec) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *NetworkPolicySpec) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *NetworkPolicySpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.PolicyTypes) > 0 { + for iNdEx := len(m.PolicyTypes) - 1; iNdEx >= 0; iNdEx-- { + i -= len(m.PolicyTypes[iNdEx]) + copy(dAtA[i:], m.PolicyTypes[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.PolicyTypes[iNdEx]))) + i-- + dAtA[i] = 0x22 + } + } + if len(m.Egress) > 0 { + for iNdEx := len(m.Egress) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Egress[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1a + } + } + if len(m.Ingress) > 0 { + for iNdEx := len(m.Ingress) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Ingress[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + { + size, err := m.PodSelector.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *ServiceBackendPort) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *ServiceBackendPort) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *ServiceBackendPort) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i = encodeVarintGenerated(dAtA, i, uint64(m.Number)) + i-- + dAtA[i] = 0x10 + i -= len(m.Name) + copy(dAtA[i:], m.Name) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Name))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { + offset -= sovGenerated(v) + base := offset + for v >= 1<<7 { + dAtA[offset] = uint8(v&0x7f | 0x80) + v >>= 7 + offset++ + } + dAtA[offset] = uint8(v) + return base +} +func (m *HTTPIngressPath) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Path) + n += 1 + l + sovGenerated(uint64(l)) + l = m.Backend.Size() + n += 1 + l + sovGenerated(uint64(l)) + if m.PathType != nil { + l = len(*m.PathType) + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *HTTPIngressRuleValue) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if len(m.Paths) > 0 { + for _, e := range m.Paths { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *IPBlock) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.CIDR) + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Except) > 0 { + for _, s := range m.Except { + l = len(s) + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *Ingress) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Spec.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Status.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *IngressBackend) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.Resource != nil { + l = m.Resource.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + if m.Service != nil { + l = m.Service.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *IngressClass) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Spec.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *IngressClassList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *IngressClassSpec) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Controller) + n += 1 + l + sovGenerated(uint64(l)) + if m.Parameters != nil { + l = m.Parameters.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *IngressList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *IngressRule) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Host) + n += 1 + l + sovGenerated(uint64(l)) + l = m.IngressRuleValue.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *IngressRuleValue) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.HTTP != nil { + l = m.HTTP.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *IngressServiceBackend) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Name) + n += 1 + l + sovGenerated(uint64(l)) + l = m.Port.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *IngressSpec) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.DefaultBackend != nil { + l = m.DefaultBackend.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + if len(m.TLS) > 0 { + for _, e := range m.TLS { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if len(m.Rules) > 0 { + for _, e := range m.Rules { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if m.IngressClassName != nil { + l = len(*m.IngressClassName) + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *IngressStatus) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.LoadBalancer.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *IngressTLS) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if len(m.Hosts) > 0 { + for _, s := range m.Hosts { + l = len(s) + n += 1 + l + sovGenerated(uint64(l)) + } + } + l = len(m.SecretName) + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *NetworkPolicy) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Spec.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *NetworkPolicyEgressRule) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if len(m.Ports) > 0 { + for _, e := range m.Ports { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if len(m.To) > 0 { + for _, e := range m.To { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *NetworkPolicyIngressRule) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if len(m.Ports) > 0 { + for _, e := range m.Ports { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if len(m.From) > 0 { + for _, e := range m.From { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *NetworkPolicyList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *NetworkPolicyPeer) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.PodSelector != nil { + l = m.PodSelector.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + if m.NamespaceSelector != nil { + l = m.NamespaceSelector.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + if m.IPBlock != nil { + l = m.IPBlock.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *NetworkPolicyPort) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.Protocol != nil { + l = len(*m.Protocol) + n += 1 + l + sovGenerated(uint64(l)) + } + if m.Port != nil { + l = m.Port.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *NetworkPolicySpec) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.PodSelector.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Ingress) > 0 { + for _, e := range m.Ingress { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if len(m.Egress) > 0 { + for _, e := range m.Egress { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + if len(m.PolicyTypes) > 0 { + for _, s := range m.PolicyTypes { + l = len(s) + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *ServiceBackendPort) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Name) + n += 1 + l + sovGenerated(uint64(l)) + n += 1 + sovGenerated(uint64(m.Number)) + return n +} + +func sovGenerated(x uint64) (n int) { + return (math_bits.Len64(x|1) + 6) / 7 +} +func sozGenerated(x uint64) (n int) { + return sovGenerated(uint64((x << 1) ^ uint64((int64(x) >> 63)))) +} +func (this *HTTPIngressPath) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&HTTPIngressPath{`, + `Path:` + fmt.Sprintf("%v", this.Path) + `,`, + `Backend:` + strings.Replace(strings.Replace(this.Backend.String(), "IngressBackend", "IngressBackend", 1), `&`, ``, 1) + `,`, + `PathType:` + valueToStringGenerated(this.PathType) + `,`, + `}`, + }, "") + return s +} +func (this *HTTPIngressRuleValue) String() string { + if this == nil { + return "nil" + } + repeatedStringForPaths := "[]HTTPIngressPath{" + for _, f := range this.Paths { + repeatedStringForPaths += strings.Replace(strings.Replace(f.String(), "HTTPIngressPath", "HTTPIngressPath", 1), `&`, ``, 1) + "," + } + repeatedStringForPaths += "}" + s := strings.Join([]string{`&HTTPIngressRuleValue{`, + `Paths:` + repeatedStringForPaths + `,`, + `}`, + }, "") + return s +} +func (this *IPBlock) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IPBlock{`, + `CIDR:` + fmt.Sprintf("%v", this.CIDR) + `,`, + `Except:` + fmt.Sprintf("%v", this.Except) + `,`, + `}`, + }, "") + return s +} +func (this *Ingress) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&Ingress{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "IngressSpec", "IngressSpec", 1), `&`, ``, 1) + `,`, + `Status:` + strings.Replace(strings.Replace(this.Status.String(), "IngressStatus", "IngressStatus", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressBackend) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressBackend{`, + `Resource:` + strings.Replace(fmt.Sprintf("%v", this.Resource), "TypedLocalObjectReference", "v11.TypedLocalObjectReference", 1) + `,`, + `Service:` + strings.Replace(this.Service.String(), "IngressServiceBackend", "IngressServiceBackend", 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressClass) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressClass{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "IngressClassSpec", "IngressClassSpec", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressClassList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]IngressClass{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "IngressClass", "IngressClass", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&IngressClassList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} +func (this *IngressClassSpec) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressClassSpec{`, + `Controller:` + fmt.Sprintf("%v", this.Controller) + `,`, + `Parameters:` + strings.Replace(fmt.Sprintf("%v", this.Parameters), "TypedLocalObjectReference", "v11.TypedLocalObjectReference", 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]Ingress{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "Ingress", "Ingress", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&IngressList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} +func (this *IngressRule) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressRule{`, + `Host:` + fmt.Sprintf("%v", this.Host) + `,`, + `IngressRuleValue:` + strings.Replace(strings.Replace(this.IngressRuleValue.String(), "IngressRuleValue", "IngressRuleValue", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressRuleValue) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressRuleValue{`, + `HTTP:` + strings.Replace(this.HTTP.String(), "HTTPIngressRuleValue", "HTTPIngressRuleValue", 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressServiceBackend) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressServiceBackend{`, + `Name:` + fmt.Sprintf("%v", this.Name) + `,`, + `Port:` + strings.Replace(strings.Replace(this.Port.String(), "ServiceBackendPort", "ServiceBackendPort", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressSpec) String() string { + if this == nil { + return "nil" + } + repeatedStringForTLS := "[]IngressTLS{" + for _, f := range this.TLS { + repeatedStringForTLS += strings.Replace(strings.Replace(f.String(), "IngressTLS", "IngressTLS", 1), `&`, ``, 1) + "," + } + repeatedStringForTLS += "}" + repeatedStringForRules := "[]IngressRule{" + for _, f := range this.Rules { + repeatedStringForRules += strings.Replace(strings.Replace(f.String(), "IngressRule", "IngressRule", 1), `&`, ``, 1) + "," + } + repeatedStringForRules += "}" + s := strings.Join([]string{`&IngressSpec{`, + `DefaultBackend:` + strings.Replace(this.DefaultBackend.String(), "IngressBackend", "IngressBackend", 1) + `,`, + `TLS:` + repeatedStringForTLS + `,`, + `Rules:` + repeatedStringForRules + `,`, + `IngressClassName:` + valueToStringGenerated(this.IngressClassName) + `,`, + `}`, + }, "") + return s +} +func (this *IngressStatus) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressStatus{`, + `LoadBalancer:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LoadBalancer), "LoadBalancerStatus", "v11.LoadBalancerStatus", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressTLS) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressTLS{`, + `Hosts:` + fmt.Sprintf("%v", this.Hosts) + `,`, + `SecretName:` + fmt.Sprintf("%v", this.SecretName) + `,`, + `}`, + }, "") + return s +} +func (this *NetworkPolicy) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&NetworkPolicy{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "NetworkPolicySpec", "NetworkPolicySpec", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *NetworkPolicyEgressRule) String() string { + if this == nil { + return "nil" + } + repeatedStringForPorts := "[]NetworkPolicyPort{" + for _, f := range this.Ports { + repeatedStringForPorts += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPort", "NetworkPolicyPort", 1), `&`, ``, 1) + "," + } + repeatedStringForPorts += "}" + repeatedStringForTo := "[]NetworkPolicyPeer{" + for _, f := range this.To { + repeatedStringForTo += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPeer", "NetworkPolicyPeer", 1), `&`, ``, 1) + "," + } + repeatedStringForTo += "}" + s := strings.Join([]string{`&NetworkPolicyEgressRule{`, + `Ports:` + repeatedStringForPorts + `,`, + `To:` + repeatedStringForTo + `,`, + `}`, + }, "") + return s +} +func (this *NetworkPolicyIngressRule) String() string { + if this == nil { + return "nil" + } + repeatedStringForPorts := "[]NetworkPolicyPort{" + for _, f := range this.Ports { + repeatedStringForPorts += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPort", "NetworkPolicyPort", 1), `&`, ``, 1) + "," + } + repeatedStringForPorts += "}" + repeatedStringForFrom := "[]NetworkPolicyPeer{" + for _, f := range this.From { + repeatedStringForFrom += strings.Replace(strings.Replace(f.String(), "NetworkPolicyPeer", "NetworkPolicyPeer", 1), `&`, ``, 1) + "," + } + repeatedStringForFrom += "}" + s := strings.Join([]string{`&NetworkPolicyIngressRule{`, + `Ports:` + repeatedStringForPorts + `,`, + `From:` + repeatedStringForFrom + `,`, + `}`, + }, "") + return s +} +func (this *NetworkPolicyList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]NetworkPolicy{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "NetworkPolicy", "NetworkPolicy", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&NetworkPolicyList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} +func (this *NetworkPolicyPeer) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&NetworkPolicyPeer{`, + `PodSelector:` + strings.Replace(fmt.Sprintf("%v", this.PodSelector), "LabelSelector", "v1.LabelSelector", 1) + `,`, + `NamespaceSelector:` + strings.Replace(fmt.Sprintf("%v", this.NamespaceSelector), "LabelSelector", "v1.LabelSelector", 1) + `,`, + `IPBlock:` + strings.Replace(this.IPBlock.String(), "IPBlock", "IPBlock", 1) + `,`, + `}`, + }, "") + return s +} +func (this *NetworkPolicyPort) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&NetworkPolicyPort{`, + `Protocol:` + valueToStringGenerated(this.Protocol) + `,`, + `Port:` + strings.Replace(fmt.Sprintf("%v", this.Port), "IntOrString", "intstr.IntOrString", 1) + `,`, + `}`, + }, "") + return s +} +func (this *NetworkPolicySpec) String() string { + if this == nil { + return "nil" + } + repeatedStringForIngress := "[]NetworkPolicyIngressRule{" + for _, f := range this.Ingress { + repeatedStringForIngress += strings.Replace(strings.Replace(f.String(), "NetworkPolicyIngressRule", "NetworkPolicyIngressRule", 1), `&`, ``, 1) + "," + } + repeatedStringForIngress += "}" + repeatedStringForEgress := "[]NetworkPolicyEgressRule{" + for _, f := range this.Egress { + repeatedStringForEgress += strings.Replace(strings.Replace(f.String(), "NetworkPolicyEgressRule", "NetworkPolicyEgressRule", 1), `&`, ``, 1) + "," + } + repeatedStringForEgress += "}" + s := strings.Join([]string{`&NetworkPolicySpec{`, + `PodSelector:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.PodSelector), "LabelSelector", "v1.LabelSelector", 1), `&`, ``, 1) + `,`, + `Ingress:` + repeatedStringForIngress + `,`, + `Egress:` + repeatedStringForEgress + `,`, + `PolicyTypes:` + fmt.Sprintf("%v", this.PolicyTypes) + `,`, + `}`, + }, "") + return s +} +func (this *ServiceBackendPort) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&ServiceBackendPort{`, + `Name:` + fmt.Sprintf("%v", this.Name) + `,`, + `Number:` + fmt.Sprintf("%v", this.Number) + `,`, + `}`, + }, "") + return s +} +func valueToStringGenerated(v interface{}) string { + rv := reflect.ValueOf(v) + if rv.IsNil() { + return "nil" + } + pv := reflect.Indirect(rv).Interface() + return fmt.Sprintf("*%v", pv) +} +func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: HTTPIngressPath: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: HTTPIngressPath: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Path", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Path = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Backend", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Backend.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field PathType", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := PathType(dAtA[iNdEx:postIndex]) + m.PathType = &s + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *HTTPIngressRuleValue) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: HTTPIngressRuleValue: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: HTTPIngressRuleValue: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Paths", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Paths = append(m.Paths, HTTPIngressPath{}) + if err := m.Paths[len(m.Paths)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IPBlock) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IPBlock: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IPBlock: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field CIDR", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.CIDR = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Except", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Except = append(m.Except, string(dAtA[iNdEx:postIndex])) + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *Ingress) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: Ingress: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: Ingress: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressBackend) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressBackend: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressBackend: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Resource", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Resource == nil { + m.Resource = &v11.TypedLocalObjectReference{} + } + if err := m.Resource.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Service", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Service == nil { + m.Service = &IngressServiceBackend{} + } + if err := m.Service.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressClass) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressClass: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressClass: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressClassList) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressClassList: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressClassList: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Items = append(m.Items, IngressClass{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressClassSpec) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressClassSpec: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressClassSpec: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Controller", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Controller = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Parameters", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Parameters == nil { + m.Parameters = &v11.TypedLocalObjectReference{} + } + if err := m.Parameters.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressList) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressList: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressList: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Items = append(m.Items, Ingress{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressRule) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressRule: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressRule: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Host", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Host = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field IngressRuleValue", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.IngressRuleValue.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressRuleValue) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressRuleValue: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressRuleValue: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field HTTP", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.HTTP == nil { + m.HTTP = &HTTPIngressRuleValue{} + } + if err := m.HTTP.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressServiceBackend) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressServiceBackend: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressServiceBackend: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Name", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Name = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Port", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Port.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *IngressSpec) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressSpec: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressSpec: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field DefaultBackend", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.DefaultBackend == nil { + m.DefaultBackend = &IngressBackend{} + } + if err := m.DefaultBackend.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field TLS", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.TLS = append(m.TLS, IngressTLS{}) + if err := m.TLS[len(m.TLS)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Rules", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Rules = append(m.Rules, IngressRule{}) + if err := m.Rules[len(m.Rules)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field IngressClassName", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.IngressClassName = &s + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } } - s := strings.Join([]string{`&NetworkPolicyPeer{`, - `PodSelector:` + strings.Replace(fmt.Sprintf("%v", this.PodSelector), "LabelSelector", "v1.LabelSelector", 1) + `,`, - `NamespaceSelector:` + strings.Replace(fmt.Sprintf("%v", this.NamespaceSelector), "LabelSelector", "v1.LabelSelector", 1) + `,`, - `IPBlock:` + strings.Replace(this.IPBlock.String(), "IPBlock", "IPBlock", 1) + `,`, - `}`, - }, "") - return s -} -func (this *NetworkPolicyPort) String() string { - if this == nil { - return "nil" + + if iNdEx > l { + return io.ErrUnexpectedEOF } - s := strings.Join([]string{`&NetworkPolicyPort{`, - `Protocol:` + valueToStringGenerated(this.Protocol) + `,`, - `Port:` + strings.Replace(fmt.Sprintf("%v", this.Port), "IntOrString", "intstr.IntOrString", 1) + `,`, - `}`, - }, "") - return s + return nil } -func (this *NetworkPolicySpec) String() string { - if this == nil { - return "nil" - } - repeatedStringForIngress := "[]NetworkPolicyIngressRule{" - for _, f := range this.Ingress { - repeatedStringForIngress += strings.Replace(strings.Replace(f.String(), "NetworkPolicyIngressRule", "NetworkPolicyIngressRule", 1), `&`, ``, 1) + "," - } - repeatedStringForIngress += "}" - repeatedStringForEgress := "[]NetworkPolicyEgressRule{" - for _, f := range this.Egress { - repeatedStringForEgress += strings.Replace(strings.Replace(f.String(), "NetworkPolicyEgressRule", "NetworkPolicyEgressRule", 1), `&`, ``, 1) + "," +func (m *IngressStatus) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: IngressStatus: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: IngressStatus: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field LoadBalancer", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.LoadBalancer.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } } - repeatedStringForEgress += "}" - s := strings.Join([]string{`&NetworkPolicySpec{`, - `PodSelector:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.PodSelector), "LabelSelector", "v1.LabelSelector", 1), `&`, ``, 1) + `,`, - `Ingress:` + repeatedStringForIngress + `,`, - `Egress:` + repeatedStringForEgress + `,`, - `PolicyTypes:` + fmt.Sprintf("%v", this.PolicyTypes) + `,`, - `}`, - }, "") - return s -} -func valueToStringGenerated(v interface{}) string { - rv := reflect.ValueOf(v) - if rv.IsNil() { - return "nil" + + if iNdEx > l { + return io.ErrUnexpectedEOF } - pv := reflect.Indirect(rv).Interface() - return fmt.Sprintf("*%v", pv) + return nil } -func (m *IPBlock) Unmarshal(dAtA []byte) error { +func (m *IngressTLS) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -1072,15 +4326,15 @@ func (m *IPBlock) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: IPBlock: wiretype end group for non-group") + return fmt.Errorf("proto: IngressTLS: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: IPBlock: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: IngressTLS: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field CIDR", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Hosts", wireType) } var stringLen uint64 for shift := uint(0); ; shift += 7 { @@ -1108,11 +4362,11 @@ func (m *IPBlock) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - m.CIDR = string(dAtA[iNdEx:postIndex]) + m.Hosts = append(m.Hosts, string(dAtA[iNdEx:postIndex])) iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Except", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field SecretName", wireType) } var stringLen uint64 for shift := uint(0); ; shift += 7 { @@ -1140,7 +4394,7 @@ func (m *IPBlock) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - m.Except = append(m.Except, string(dAtA[iNdEx:postIndex])) + m.SecretName = string(dAtA[iNdEx:postIndex]) iNdEx = postIndex default: iNdEx = preIndex @@ -2116,9 +5370,114 @@ func (m *NetworkPolicySpec) Unmarshal(dAtA []byte) error { } return nil } +func (m *ServiceBackendPort) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: ServiceBackendPort: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: ServiceBackendPort: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Name", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Name = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field Number", wireType) + } + m.Number = 0 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + m.Number |= int32(b&0x7F) << shift + if b < 0x80 { + break + } + } + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -2150,10 +5509,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -2174,55 +5531,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/networking/v1/generated.proto b/vendor/k8s.io/api/networking/v1/generated.proto index 3cb73804..a98ef94c 100644 --- a/vendor/k8s.io/api/networking/v1/generated.proto +++ b/vendor/k8s.io/api/networking/v1/generated.proto @@ -30,21 +30,276 @@ import "k8s.io/apimachinery/pkg/util/intstr/generated.proto"; // Package-wide variables from generator "generated". option go_package = "v1"; -// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24") that is allowed to the pods -// matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should -// not be included within this rule. +// HTTPIngressPath associates a path with a backend. Incoming urls matching the +// path are forwarded to the backend. +message HTTPIngressPath { + // Path is matched against the path of an incoming request. Currently it can + // contain characters disallowed from the conventional "path" part of a URL + // as defined by RFC 3986. Paths must begin with a '/'. When unspecified, + // all paths from incoming requests are matched. + // +optional + optional string path = 1; + + // PathType determines the interpretation of the Path matching. PathType can + // be one of the following values: + // * Exact: Matches the URL path exactly. + // * Prefix: Matches based on a URL path prefix split by '/'. Matching is + // done on a path element by element basis. A path element refers is the + // list of labels in the path split by the '/' separator. A request is a + // match for path p if every p is an element-wise prefix of p of the + // request path. Note that if the last element of the path is a substring + // of the last element in request path, it is not a match (e.g. /foo/bar + // matches /foo/bar/baz, but does not match /foo/barbaz). + // * ImplementationSpecific: Interpretation of the Path matching is up to + // the IngressClass. Implementations can treat this as a separate PathType + // or treat it identically to Prefix or Exact path types. + // Implementations are required to support all path types. + optional string pathType = 3; + + // Backend defines the referenced service endpoint to which the traffic + // will be forwarded to. + optional IngressBackend backend = 2; +} + +// HTTPIngressRuleValue is a list of http selectors pointing to backends. +// In the example: http:///? -> backend where +// where parts of the url correspond to RFC 3986, this resource will be used +// to match against everything after the last '/' and before the first '?' +// or '#'. +message HTTPIngressRuleValue { + // A collection of paths that map requests to backends. + // +listType=atomic + repeated HTTPIngressPath paths = 1; +} + +// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24","2001:db9::/64") that is allowed +// to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs +// that should not be included within this rule. message IPBlock { // CIDR is a string representing the IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" optional string cidr = 1; // Except is a slice of CIDRs that should not be included within an IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" // Except values will be rejected if they are outside the CIDR range // +optional repeated string except = 2; } +// Ingress is a collection of rules that allow inbound connections to reach the +// endpoints defined by a backend. An Ingress can be configured to give services +// externally-reachable urls, load balance traffic, terminate SSL, offer name +// based virtual hosting etc. +message Ingress { + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // Spec is the desired state of the Ingress. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + optional IngressSpec spec = 2; + + // Status is the current state of the Ingress. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + optional IngressStatus status = 3; +} + +// IngressBackend describes all endpoints for a given service and port. +message IngressBackend { + // Service references a Service as a Backend. + // This is a mutually exclusive setting with "Resource". + // +optional + optional IngressServiceBackend service = 4; + + // Resource is an ObjectRef to another Kubernetes resource in the namespace + // of the Ingress object. If resource is specified, a service.Name and + // service.Port must not be specified. + // This is a mutually exclusive setting with "Service". + // +optional + optional k8s.io.api.core.v1.TypedLocalObjectReference resource = 3; +} + +// IngressClass represents the class of the Ingress, referenced by the Ingress +// Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be +// used to indicate that an IngressClass should be considered default. When a +// single IngressClass resource has this annotation set to true, new Ingress +// resources without a class specified will be assigned this default class. +message IngressClass { + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // Spec is the desired state of the IngressClass. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + optional IngressClassSpec spec = 2; +} + +// IngressClassList is a collection of IngressClasses. +message IngressClassList { + // Standard list metadata. + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; + + // Items is the list of IngressClasses. + repeated IngressClass items = 2; +} + +// IngressClassSpec provides information about the class of an Ingress. +message IngressClassSpec { + // Controller refers to the name of the controller that should handle this + // class. This allows for different "flavors" that are controlled by the + // same controller. For example, you may have different Parameters for the + // same implementing controller. This should be specified as a + // domain-prefixed path no more than 250 characters in length, e.g. + // "acme.io/ingress-controller". This field is immutable. + optional string controller = 1; + + // Parameters is a link to a custom resource containing additional + // configuration for the controller. This is optional if the controller does + // not require extra parameters. + // +optional + optional k8s.io.api.core.v1.TypedLocalObjectReference parameters = 2; +} + +// IngressList is a collection of Ingress. +message IngressList { + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; + + // Items is the list of Ingress. + repeated Ingress items = 2; +} + +// IngressRule represents the rules mapping the paths under a specified host to +// the related backend services. Incoming requests are first evaluated for a host +// match, then routed to the backend associated with the matching IngressRuleValue. +message IngressRule { + // Host is the fully qualified domain name of a network host, as defined by RFC 3986. + // Note the following deviations from the "host" part of the + // URI as defined in RFC 3986: + // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to + // the IP in the Spec of the parent Ingress. + // 2. The `:` delimiter is not respected because ports are not allowed. + // Currently the port of an Ingress is implicitly :80 for http and + // :443 for https. + // Both these may change in the future. + // Incoming requests are matched against the host before the + // IngressRuleValue. If the host is unspecified, the Ingress routes all + // traffic based on the specified IngressRuleValue. + // + // Host can be "precise" which is a domain name without the terminating dot of + // a network host (e.g. "foo.bar.com") or "wildcard", which is a domain name + // prefixed with a single wildcard label (e.g. "*.foo.com"). + // The wildcard character '*' must appear by itself as the first DNS label and + // matches only a single label. You cannot have a wildcard label by itself (e.g. Host == "*"). + // Requests will be matched against the Host field in the following way: + // 1. If Host is precise, the request matches this rule if the http host header is equal to Host. + // 2. If Host is a wildcard, then the request matches this rule if the http host header + // is to equal to the suffix (removing the first label) of the wildcard rule. + // +optional + optional string host = 1; + + // IngressRuleValue represents a rule to route requests for this IngressRule. + // If unspecified, the rule defaults to a http catch-all. Whether that sends + // just traffic matching the host to the default backend or all traffic to the + // default backend, is left to the controller fulfilling the Ingress. Http is + // currently the only supported IngressRuleValue. + // +optional + optional IngressRuleValue ingressRuleValue = 2; +} + +// IngressRuleValue represents a rule to apply against incoming requests. If the +// rule is satisfied, the request is routed to the specified backend. Currently +// mixing different types of rules in a single Ingress is disallowed, so exactly +// one of the following must be set. +message IngressRuleValue { + // +optional + optional HTTPIngressRuleValue http = 1; +} + +// IngressServiceBackend references a Kubernetes Service as a Backend. +message IngressServiceBackend { + // Name is the referenced service. The service must exist in + // the same namespace as the Ingress object. + optional string name = 1; + + // Port of the referenced service. A port name or port number + // is required for a IngressServiceBackend. + optional ServiceBackendPort port = 2; +} + +// IngressSpec describes the Ingress the user wishes to exist. +message IngressSpec { + // IngressClassName is the name of the IngressClass cluster resource. The + // associated IngressClass defines which controller will implement the + // resource. This replaces the deprecated `kubernetes.io/ingress.class` + // annotation. For backwards compatibility, when that annotation is set, it + // must be given precedence over this field. The controller may emit a + // warning if the field and annotation have different values. + // Implementations of this API should ignore Ingresses without a class + // specified. An IngressClass resource may be marked as default, which can + // be used to set a default value for this field. For more information, + // refer to the IngressClass documentation. + // +optional + optional string ingressClassName = 4; + + // DefaultBackend is the backend that should handle requests that don't + // match any rule. If Rules are not specified, DefaultBackend must be specified. + // If DefaultBackend is not set, the handling of requests that do not match any + // of the rules will be up to the Ingress controller. + // +optional + optional IngressBackend defaultBackend = 1; + + // TLS configuration. Currently the Ingress only supports a single TLS + // port, 443. If multiple members of this list specify different hosts, they + // will be multiplexed on the same port according to the hostname specified + // through the SNI TLS extension, if the ingress controller fulfilling the + // ingress supports SNI. + // +listType=atomic + // +optional + repeated IngressTLS tls = 2; + + // A list of host rules used to configure the Ingress. If unspecified, or + // no rule matches, all traffic is sent to the default backend. + // +listType=atomic + // +optional + repeated IngressRule rules = 3; +} + +// IngressStatus describe the current state of the Ingress. +message IngressStatus { + // LoadBalancer contains the current status of the load-balancer. + // +optional + optional k8s.io.api.core.v1.LoadBalancerStatus loadBalancer = 1; +} + +// IngressTLS describes the transport layer security associated with an Ingress. +message IngressTLS { + // Hosts are a list of hosts included in the TLS certificate. The values in + // this list must match the name/s used in the tlsSecret. Defaults to the + // wildcard host setting for the loadbalancer controller fulfilling this + // Ingress, if left unspecified. + // +listType=atomic + // +optional + repeated string hosts = 1; + + // SecretName is the name of the secret used to terminate TLS traffic on + // port 443. Field is left optional to allow TLS routing based on SNI + // hostname alone. If the SNI host in a listener conflicts with the "Host" + // header field used by an IngressRule, the SNI host is used for termination + // and value of the Host header is used for routing. + // +optional + optional string secretName = 2; +} + // NetworkPolicy describes what network traffic is allowed for a set of Pods message NetworkPolicy { // Standard object's metadata. @@ -109,7 +364,7 @@ message NetworkPolicyList { repeated NetworkPolicy items = 2; } -// NetworkPolicyPeer describes a peer to allow traffic from. Only certain combinations of +// NetworkPolicyPeer describes a peer to allow traffic to/from. Only certain combinations of // fields are allowed message NetworkPolicyPeer { // This is a label selector which selects Pods. This field follows standard label @@ -193,3 +448,16 @@ message NetworkPolicySpec { repeated string policyTypes = 4; } +// ServiceBackendPort is the service port being referenced. +message ServiceBackendPort { + // Name is the name of the port on the Service. + // This is a mutually exclusive setting with "Number". + // +optional + optional string name = 1; + + // Number is the numerical port number (e.g. 80) on the Service. + // This is a mutually exclusive setting with "Name". + // +optional + optional int32 number = 2; +} + diff --git a/vendor/k8s.io/api/networking/v1/register.go b/vendor/k8s.io/api/networking/v1/register.go index f47f22e9..a200d543 100644 --- a/vendor/k8s.io/api/networking/v1/register.go +++ b/vendor/k8s.io/api/networking/v1/register.go @@ -44,6 +44,10 @@ var ( // Adds the list of known types to the given scheme. func addKnownTypes(scheme *runtime.Scheme) error { scheme.AddKnownTypes(SchemeGroupVersion, + &Ingress{}, + &IngressList{}, + &IngressClass{}, + &IngressClassList{}, &NetworkPolicy{}, &NetworkPolicyList{}, ) diff --git a/vendor/k8s.io/api/networking/v1/types.go b/vendor/k8s.io/api/networking/v1/types.go index 38a640f0..df256908 100644 --- a/vendor/k8s.io/api/networking/v1/types.go +++ b/vendor/k8s.io/api/networking/v1/types.go @@ -147,21 +147,21 @@ type NetworkPolicyPort struct { Port *intstr.IntOrString `json:"port,omitempty" protobuf:"bytes,2,opt,name=port"` } -// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24") that is allowed to the pods -// matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should -// not be included within this rule. +// IPBlock describes a particular CIDR (Ex. "192.168.1.1/24","2001:db9::/64") that is allowed +// to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs +// that should not be included within this rule. type IPBlock struct { // CIDR is a string representing the IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" CIDR string `json:"cidr" protobuf:"bytes,1,name=cidr"` // Except is a slice of CIDRs that should not be included within an IP Block - // Valid examples are "192.168.1.1/24" + // Valid examples are "192.168.1.1/24" or "2001:db9::/64" // Except values will be rejected if they are outside the CIDR range // +optional Except []string `json:"except,omitempty" protobuf:"bytes,2,rep,name=except"` } -// NetworkPolicyPeer describes a peer to allow traffic from. Only certain combinations of +// NetworkPolicyPeer describes a peer to allow traffic to/from. Only certain combinations of // fields are allowed type NetworkPolicyPeer struct { // This is a label selector which selects Pods. This field follows standard label @@ -201,3 +201,314 @@ type NetworkPolicyList struct { // Items is a list of schema objects. Items []NetworkPolicy `json:"items" protobuf:"bytes,2,rep,name=items"` } + +// +genclient +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// Ingress is a collection of rules that allow inbound connections to reach the +// endpoints defined by a backend. An Ingress can be configured to give services +// externally-reachable urls, load balance traffic, terminate SSL, offer name +// based virtual hosting etc. +type Ingress struct { + metav1.TypeMeta `json:",inline"` + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Spec is the desired state of the Ingress. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + Spec IngressSpec `json:"spec,omitempty" protobuf:"bytes,2,opt,name=spec"` + + // Status is the current state of the Ingress. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + Status IngressStatus `json:"status,omitempty" protobuf:"bytes,3,opt,name=status"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// IngressList is a collection of Ingress. +type IngressList struct { + metav1.TypeMeta `json:",inline"` + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Items is the list of Ingress. + Items []Ingress `json:"items" protobuf:"bytes,2,rep,name=items"` +} + +// IngressSpec describes the Ingress the user wishes to exist. +type IngressSpec struct { + // IngressClassName is the name of the IngressClass cluster resource. The + // associated IngressClass defines which controller will implement the + // resource. This replaces the deprecated `kubernetes.io/ingress.class` + // annotation. For backwards compatibility, when that annotation is set, it + // must be given precedence over this field. The controller may emit a + // warning if the field and annotation have different values. + // Implementations of this API should ignore Ingresses without a class + // specified. An IngressClass resource may be marked as default, which can + // be used to set a default value for this field. For more information, + // refer to the IngressClass documentation. + // +optional + IngressClassName *string `json:"ingressClassName,omitempty" protobuf:"bytes,4,opt,name=ingressClassName"` + + // DefaultBackend is the backend that should handle requests that don't + // match any rule. If Rules are not specified, DefaultBackend must be specified. + // If DefaultBackend is not set, the handling of requests that do not match any + // of the rules will be up to the Ingress controller. + // +optional + DefaultBackend *IngressBackend `json:"defaultBackend,omitempty" protobuf:"bytes,1,opt,name=defaultBackend"` + + // TLS configuration. Currently the Ingress only supports a single TLS + // port, 443. If multiple members of this list specify different hosts, they + // will be multiplexed on the same port according to the hostname specified + // through the SNI TLS extension, if the ingress controller fulfilling the + // ingress supports SNI. + // +listType=atomic + // +optional + TLS []IngressTLS `json:"tls,omitempty" protobuf:"bytes,2,rep,name=tls"` + + // A list of host rules used to configure the Ingress. If unspecified, or + // no rule matches, all traffic is sent to the default backend. + // +listType=atomic + // +optional + Rules []IngressRule `json:"rules,omitempty" protobuf:"bytes,3,rep,name=rules"` +} + +// IngressTLS describes the transport layer security associated with an Ingress. +type IngressTLS struct { + // Hosts are a list of hosts included in the TLS certificate. The values in + // this list must match the name/s used in the tlsSecret. Defaults to the + // wildcard host setting for the loadbalancer controller fulfilling this + // Ingress, if left unspecified. + // +listType=atomic + // +optional + Hosts []string `json:"hosts,omitempty" protobuf:"bytes,1,rep,name=hosts"` + // SecretName is the name of the secret used to terminate TLS traffic on + // port 443. Field is left optional to allow TLS routing based on SNI + // hostname alone. If the SNI host in a listener conflicts with the "Host" + // header field used by an IngressRule, the SNI host is used for termination + // and value of the Host header is used for routing. + // +optional + SecretName string `json:"secretName,omitempty" protobuf:"bytes,2,opt,name=secretName"` +} + +// IngressStatus describe the current state of the Ingress. +type IngressStatus struct { + // LoadBalancer contains the current status of the load-balancer. + // +optional + LoadBalancer v1.LoadBalancerStatus `json:"loadBalancer,omitempty" protobuf:"bytes,1,opt,name=loadBalancer"` +} + +// IngressRule represents the rules mapping the paths under a specified host to +// the related backend services. Incoming requests are first evaluated for a host +// match, then routed to the backend associated with the matching IngressRuleValue. +type IngressRule struct { + // Host is the fully qualified domain name of a network host, as defined by RFC 3986. + // Note the following deviations from the "host" part of the + // URI as defined in RFC 3986: + // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to + // the IP in the Spec of the parent Ingress. + // 2. The `:` delimiter is not respected because ports are not allowed. + // Currently the port of an Ingress is implicitly :80 for http and + // :443 for https. + // Both these may change in the future. + // Incoming requests are matched against the host before the + // IngressRuleValue. If the host is unspecified, the Ingress routes all + // traffic based on the specified IngressRuleValue. + // + // Host can be "precise" which is a domain name without the terminating dot of + // a network host (e.g. "foo.bar.com") or "wildcard", which is a domain name + // prefixed with a single wildcard label (e.g. "*.foo.com"). + // The wildcard character '*' must appear by itself as the first DNS label and + // matches only a single label. You cannot have a wildcard label by itself (e.g. Host == "*"). + // Requests will be matched against the Host field in the following way: + // 1. If Host is precise, the request matches this rule if the http host header is equal to Host. + // 2. If Host is a wildcard, then the request matches this rule if the http host header + // is to equal to the suffix (removing the first label) of the wildcard rule. + // +optional + Host string `json:"host,omitempty" protobuf:"bytes,1,opt,name=host"` + // IngressRuleValue represents a rule to route requests for this IngressRule. + // If unspecified, the rule defaults to a http catch-all. Whether that sends + // just traffic matching the host to the default backend or all traffic to the + // default backend, is left to the controller fulfilling the Ingress. Http is + // currently the only supported IngressRuleValue. + // +optional + IngressRuleValue `json:",inline,omitempty" protobuf:"bytes,2,opt,name=ingressRuleValue"` +} + +// IngressRuleValue represents a rule to apply against incoming requests. If the +// rule is satisfied, the request is routed to the specified backend. Currently +// mixing different types of rules in a single Ingress is disallowed, so exactly +// one of the following must be set. +type IngressRuleValue struct { + // +optional + HTTP *HTTPIngressRuleValue `json:"http,omitempty" protobuf:"bytes,1,opt,name=http"` +} + +// HTTPIngressRuleValue is a list of http selectors pointing to backends. +// In the example: http:///? -> backend where +// where parts of the url correspond to RFC 3986, this resource will be used +// to match against everything after the last '/' and before the first '?' +// or '#'. +type HTTPIngressRuleValue struct { + // A collection of paths that map requests to backends. + // +listType=atomic + Paths []HTTPIngressPath `json:"paths" protobuf:"bytes,1,rep,name=paths"` +} + +// PathType represents the type of path referred to by a HTTPIngressPath. +type PathType string + +const ( + // PathTypeExact matches the URL path exactly and with case sensitivity. + PathTypeExact = PathType("Exact") + + // PathTypePrefix matches based on a URL path prefix split by '/'. Matching + // is case sensitive and done on a path element by element basis. A path + // element refers to the list of labels in the path split by the '/' + // separator. A request is a match for path p if every p is an element-wise + // prefix of p of the request path. Note that if the last element of the + // path is a substring of the last element in request path, it is not a + // match (e.g. /foo/bar matches /foo/bar/baz, but does not match + // /foo/barbaz). If multiple matching paths exist in an Ingress spec, the + // longest matching path is given priority. + // Examples: + // - /foo/bar does not match requests to /foo/barbaz + // - /foo/bar matches request to /foo/bar and /foo/bar/baz + // - /foo and /foo/ both match requests to /foo and /foo/. If both paths are + // present in an Ingress spec, the longest matching path (/foo/) is given + // priority. + PathTypePrefix = PathType("Prefix") + + // PathTypeImplementationSpecific matching is up to the IngressClass. + // Implementations can treat this as a separate PathType or treat it + // identically to Prefix or Exact path types. + PathTypeImplementationSpecific = PathType("ImplementationSpecific") +) + +// HTTPIngressPath associates a path with a backend. Incoming urls matching the +// path are forwarded to the backend. +type HTTPIngressPath struct { + // Path is matched against the path of an incoming request. Currently it can + // contain characters disallowed from the conventional "path" part of a URL + // as defined by RFC 3986. Paths must begin with a '/'. When unspecified, + // all paths from incoming requests are matched. + // +optional + Path string `json:"path,omitempty" protobuf:"bytes,1,opt,name=path"` + + // PathType determines the interpretation of the Path matching. PathType can + // be one of the following values: + // * Exact: Matches the URL path exactly. + // * Prefix: Matches based on a URL path prefix split by '/'. Matching is + // done on a path element by element basis. A path element refers is the + // list of labels in the path split by the '/' separator. A request is a + // match for path p if every p is an element-wise prefix of p of the + // request path. Note that if the last element of the path is a substring + // of the last element in request path, it is not a match (e.g. /foo/bar + // matches /foo/bar/baz, but does not match /foo/barbaz). + // * ImplementationSpecific: Interpretation of the Path matching is up to + // the IngressClass. Implementations can treat this as a separate PathType + // or treat it identically to Prefix or Exact path types. + // Implementations are required to support all path types. + PathType *PathType `json:"pathType,omitempty" protobuf:"bytes,3,opt,name=pathType"` + + // Backend defines the referenced service endpoint to which the traffic + // will be forwarded to. + Backend IngressBackend `json:"backend" protobuf:"bytes,2,opt,name=backend"` +} + +// IngressBackend describes all endpoints for a given service and port. +type IngressBackend struct { + // Service references a Service as a Backend. + // This is a mutually exclusive setting with "Resource". + // +optional + Service *IngressServiceBackend `json:"service,omitempty" protobuf:"bytes,4,opt,name=service"` + + // Resource is an ObjectRef to another Kubernetes resource in the namespace + // of the Ingress object. If resource is specified, a service.Name and + // service.Port must not be specified. + // This is a mutually exclusive setting with "Service". + // +optional + Resource *v1.TypedLocalObjectReference `json:"resource,omitempty" protobuf:"bytes,3,opt,name=resource"` +} + +// IngressServiceBackend references a Kubernetes Service as a Backend. +type IngressServiceBackend struct { + // Name is the referenced service. The service must exist in + // the same namespace as the Ingress object. + Name string `json:"name" protobuf:"bytes,1,opt,name=name"` + + // Port of the referenced service. A port name or port number + // is required for a IngressServiceBackend. + Port ServiceBackendPort `json:"port,omitempty" protobuf:"bytes,2,opt,name=port"` +} + +// ServiceBackendPort is the service port being referenced. +type ServiceBackendPort struct { + // Name is the name of the port on the Service. + // This is a mutually exclusive setting with "Number". + // +optional + Name string `json:"name,omitempty" protobuf:"bytes,1,opt,name=name"` + + // Number is the numerical port number (e.g. 80) on the Service. + // This is a mutually exclusive setting with "Name". + // +optional + Number int32 `json:"number,omitempty" protobuf:"bytes,2,opt,name=number"` +} + +// +genclient +// +genclient:nonNamespaced +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// IngressClass represents the class of the Ingress, referenced by the Ingress +// Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be +// used to indicate that an IngressClass should be considered default. When a +// single IngressClass resource has this annotation set to true, new Ingress +// resources without a class specified will be assigned this default class. +type IngressClass struct { + metav1.TypeMeta `json:",inline"` + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Spec is the desired state of the IngressClass. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + Spec IngressClassSpec `json:"spec,omitempty" protobuf:"bytes,2,opt,name=spec"` +} + +// IngressClassSpec provides information about the class of an Ingress. +type IngressClassSpec struct { + // Controller refers to the name of the controller that should handle this + // class. This allows for different "flavors" that are controlled by the + // same controller. For example, you may have different Parameters for the + // same implementing controller. This should be specified as a + // domain-prefixed path no more than 250 characters in length, e.g. + // "acme.io/ingress-controller". This field is immutable. + Controller string `json:"controller,omitempty" protobuf:"bytes,1,opt,name=controller"` + + // Parameters is a link to a custom resource containing additional + // configuration for the controller. This is optional if the controller does + // not require extra parameters. + // +optional + Parameters *v1.TypedLocalObjectReference `json:"parameters,omitempty" protobuf:"bytes,2,opt,name=parameters"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// IngressClassList is a collection of IngressClasses. +type IngressClassList struct { + metav1.TypeMeta `json:",inline"` + // Standard list metadata. + // +optional + metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Items is the list of IngressClasses. + Items []IngressClass `json:"items" protobuf:"bytes,2,rep,name=items"` +} diff --git a/vendor/k8s.io/api/networking/v1/types_swagger_doc_generated.go b/vendor/k8s.io/api/networking/v1/types_swagger_doc_generated.go index 188b72a1..41b9b3fb 100644 --- a/vendor/k8s.io/api/networking/v1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/networking/v1/types_swagger_doc_generated.go @@ -27,16 +27,155 @@ package v1 // Those methods can be generated by using hack/update-generated-swagger-docs.sh // AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. +var map_HTTPIngressPath = map[string]string{ + "": "HTTPIngressPath associates a path with a backend. Incoming urls matching the path are forwarded to the backend.", + "path": "Path is matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional \"path\" part of a URL as defined by RFC 3986. Paths must begin with a '/'. When unspecified, all paths from incoming requests are matched.", + "pathType": "PathType determines the interpretation of the Path matching. PathType can be one of the following values: * Exact: Matches the URL path exactly. * Prefix: Matches based on a URL path prefix split by '/'. Matching is\n done on a path element by element basis. A path element refers is the\n list of labels in the path split by the '/' separator. A request is a\n match for path p if every p is an element-wise prefix of p of the\n request path. Note that if the last element of the path is a substring\n of the last element in request path, it is not a match (e.g. /foo/bar\n matches /foo/bar/baz, but does not match /foo/barbaz).\n* ImplementationSpecific: Interpretation of the Path matching is up to\n the IngressClass. Implementations can treat this as a separate PathType\n or treat it identically to Prefix or Exact path types.\nImplementations are required to support all path types.", + "backend": "Backend defines the referenced service endpoint to which the traffic will be forwarded to.", +} + +func (HTTPIngressPath) SwaggerDoc() map[string]string { + return map_HTTPIngressPath +} + +var map_HTTPIngressRuleValue = map[string]string{ + "": "HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'.", + "paths": "A collection of paths that map requests to backends.", +} + +func (HTTPIngressRuleValue) SwaggerDoc() map[string]string { + return map_HTTPIngressRuleValue +} + var map_IPBlock = map[string]string{ - "": "IPBlock describes a particular CIDR (Ex. \"192.168.1.1/24\") that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule.", - "cidr": "CIDR is a string representing the IP Block Valid examples are \"192.168.1.1/24\"", - "except": "Except is a slice of CIDRs that should not be included within an IP Block Valid examples are \"192.168.1.1/24\" Except values will be rejected if they are outside the CIDR range", + "": "IPBlock describes a particular CIDR (Ex. \"192.168.1.1/24\",\"2001:db9::/64\") that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule.", + "cidr": "CIDR is a string representing the IP Block Valid examples are \"192.168.1.1/24\" or \"2001:db9::/64\"", + "except": "Except is a slice of CIDRs that should not be included within an IP Block Valid examples are \"192.168.1.1/24\" or \"2001:db9::/64\" Except values will be rejected if they are outside the CIDR range", } func (IPBlock) SwaggerDoc() map[string]string { return map_IPBlock } +var map_Ingress = map[string]string{ + "": "Ingress is a collection of rules that allow inbound connections to reach the endpoints defined by a backend. An Ingress can be configured to give services externally-reachable urls, load balance traffic, terminate SSL, offer name based virtual hosting etc.", + "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "spec": "Spec is the desired state of the Ingress. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", + "status": "Status is the current state of the Ingress. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", +} + +func (Ingress) SwaggerDoc() map[string]string { + return map_Ingress +} + +var map_IngressBackend = map[string]string{ + "": "IngressBackend describes all endpoints for a given service and port.", + "service": "Service references a Service as a Backend. This is a mutually exclusive setting with \"Resource\".", + "resource": "Resource is an ObjectRef to another Kubernetes resource in the namespace of the Ingress object. If resource is specified, a service.Name and service.Port must not be specified. This is a mutually exclusive setting with \"Service\".", +} + +func (IngressBackend) SwaggerDoc() map[string]string { + return map_IngressBackend +} + +var map_IngressClass = map[string]string{ + "": "IngressClass represents the class of the Ingress, referenced by the Ingress Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be used to indicate that an IngressClass should be considered default. When a single IngressClass resource has this annotation set to true, new Ingress resources without a class specified will be assigned this default class.", + "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "spec": "Spec is the desired state of the IngressClass. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", +} + +func (IngressClass) SwaggerDoc() map[string]string { + return map_IngressClass +} + +var map_IngressClassList = map[string]string{ + "": "IngressClassList is a collection of IngressClasses.", + "metadata": "Standard list metadata.", + "items": "Items is the list of IngressClasses.", +} + +func (IngressClassList) SwaggerDoc() map[string]string { + return map_IngressClassList +} + +var map_IngressClassSpec = map[string]string{ + "": "IngressClassSpec provides information about the class of an Ingress.", + "controller": "Controller refers to the name of the controller that should handle this class. This allows for different \"flavors\" that are controlled by the same controller. For example, you may have different Parameters for the same implementing controller. This should be specified as a domain-prefixed path no more than 250 characters in length, e.g. \"acme.io/ingress-controller\". This field is immutable.", + "parameters": "Parameters is a link to a custom resource containing additional configuration for the controller. This is optional if the controller does not require extra parameters.", +} + +func (IngressClassSpec) SwaggerDoc() map[string]string { + return map_IngressClassSpec +} + +var map_IngressList = map[string]string{ + "": "IngressList is a collection of Ingress.", + "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "items": "Items is the list of Ingress.", +} + +func (IngressList) SwaggerDoc() map[string]string { + return map_IngressList +} + +var map_IngressRule = map[string]string{ + "": "IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue.", + "host": "Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the \"host\" part of the URI as defined in RFC 3986: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to\n the IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.\n\nHost can be \"precise\" which is a domain name without the terminating dot of a network host (e.g. \"foo.bar.com\") or \"wildcard\", which is a domain name prefixed with a single wildcard label (e.g. \"*.foo.com\"). The wildcard character '*' must appear by itself as the first DNS label and matches only a single label. You cannot have a wildcard label by itself (e.g. Host == \"*\"). Requests will be matched against the Host field in the following way: 1. If Host is precise, the request matches this rule if the http host header is equal to Host. 2. If Host is a wildcard, then the request matches this rule if the http host header is to equal to the suffix (removing the first label) of the wildcard rule.", +} + +func (IngressRule) SwaggerDoc() map[string]string { + return map_IngressRule +} + +var map_IngressRuleValue = map[string]string{ + "": "IngressRuleValue represents a rule to apply against incoming requests. If the rule is satisfied, the request is routed to the specified backend. Currently mixing different types of rules in a single Ingress is disallowed, so exactly one of the following must be set.", +} + +func (IngressRuleValue) SwaggerDoc() map[string]string { + return map_IngressRuleValue +} + +var map_IngressServiceBackend = map[string]string{ + "": "IngressServiceBackend references a Kubernetes Service as a Backend.", + "name": "Name is the referenced service. The service must exist in the same namespace as the Ingress object.", + "port": "Port of the referenced service. A port name or port number is required for a IngressServiceBackend.", +} + +func (IngressServiceBackend) SwaggerDoc() map[string]string { + return map_IngressServiceBackend +} + +var map_IngressSpec = map[string]string{ + "": "IngressSpec describes the Ingress the user wishes to exist.", + "ingressClassName": "IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.", + "defaultBackend": "DefaultBackend is the backend that should handle requests that don't match any rule. If Rules are not specified, DefaultBackend must be specified. If DefaultBackend is not set, the handling of requests that do not match any of the rules will be up to the Ingress controller.", + "tls": "TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.", + "rules": "A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.", +} + +func (IngressSpec) SwaggerDoc() map[string]string { + return map_IngressSpec +} + +var map_IngressStatus = map[string]string{ + "": "IngressStatus describe the current state of the Ingress.", + "loadBalancer": "LoadBalancer contains the current status of the load-balancer.", +} + +func (IngressStatus) SwaggerDoc() map[string]string { + return map_IngressStatus +} + +var map_IngressTLS = map[string]string{ + "": "IngressTLS describes the transport layer security associated with an Ingress.", + "hosts": "Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.", + "secretName": "SecretName is the name of the secret used to terminate TLS traffic on port 443. Field is left optional to allow TLS routing based on SNI hostname alone. If the SNI host in a listener conflicts with the \"Host\" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing.", +} + +func (IngressTLS) SwaggerDoc() map[string]string { + return map_IngressTLS +} + var map_NetworkPolicy = map[string]string{ "": "NetworkPolicy describes what network traffic is allowed for a set of Pods", "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", @@ -78,7 +217,7 @@ func (NetworkPolicyList) SwaggerDoc() map[string]string { } var map_NetworkPolicyPeer = map[string]string{ - "": "NetworkPolicyPeer describes a peer to allow traffic from. Only certain combinations of fields are allowed", + "": "NetworkPolicyPeer describes a peer to allow traffic to/from. Only certain combinations of fields are allowed", "podSelector": "This is a label selector which selects Pods. This field follows standard label selector semantics; if present but empty, it selects all pods.\n\nIf NamespaceSelector is also set, then the NetworkPolicyPeer as a whole selects the Pods matching PodSelector in the Namespaces selected by NamespaceSelector. Otherwise it selects the Pods matching PodSelector in the policy's own Namespace.", "namespaceSelector": "Selects Namespaces using cluster-scoped labels. This field follows standard label selector semantics; if present but empty, it selects all namespaces.\n\nIf PodSelector is also set, then the NetworkPolicyPeer as a whole selects the Pods matching PodSelector in the Namespaces selected by NamespaceSelector. Otherwise it selects all Pods in the Namespaces selected by NamespaceSelector.", "ipBlock": "IPBlock defines policy on a particular IPBlock. If this field is set then neither of the other fields can be.", @@ -110,4 +249,14 @@ func (NetworkPolicySpec) SwaggerDoc() map[string]string { return map_NetworkPolicySpec } +var map_ServiceBackendPort = map[string]string{ + "": "ServiceBackendPort is the service port being referenced.", + "name": "Name is the name of the port on the Service. This is a mutually exclusive setting with \"Number\".", + "number": "Number is the numerical port number (e.g. 80) on the Service. This is a mutually exclusive setting with \"Name\".", +} + +func (ServiceBackendPort) SwaggerDoc() map[string]string { + return map_ServiceBackendPort +} + // AUTO-GENERATED FUNCTIONS END HERE diff --git a/vendor/k8s.io/api/networking/v1/zz_generated.deepcopy.go b/vendor/k8s.io/api/networking/v1/zz_generated.deepcopy.go index 1833e978..b17e7892 100644 --- a/vendor/k8s.io/api/networking/v1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/networking/v1/zz_generated.deepcopy.go @@ -27,6 +27,51 @@ import ( intstr "k8s.io/apimachinery/pkg/util/intstr" ) +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *HTTPIngressPath) DeepCopyInto(out *HTTPIngressPath) { + *out = *in + if in.PathType != nil { + in, out := &in.PathType, &out.PathType + *out = new(PathType) + **out = **in + } + in.Backend.DeepCopyInto(&out.Backend) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HTTPIngressPath. +func (in *HTTPIngressPath) DeepCopy() *HTTPIngressPath { + if in == nil { + return nil + } + out := new(HTTPIngressPath) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *HTTPIngressRuleValue) DeepCopyInto(out *HTTPIngressRuleValue) { + *out = *in + if in.Paths != nil { + in, out := &in.Paths, &out.Paths + *out = make([]HTTPIngressPath, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new HTTPIngressRuleValue. +func (in *HTTPIngressRuleValue) DeepCopy() *HTTPIngressRuleValue { + if in == nil { + return nil + } + out := new(HTTPIngressRuleValue) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *IPBlock) DeepCopyInto(out *IPBlock) { *out = *in @@ -48,6 +93,307 @@ func (in *IPBlock) DeepCopy() *IPBlock { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *Ingress) DeepCopyInto(out *Ingress) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + in.Spec.DeepCopyInto(&out.Spec) + in.Status.DeepCopyInto(&out.Status) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Ingress. +func (in *Ingress) DeepCopy() *Ingress { + if in == nil { + return nil + } + out := new(Ingress) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *Ingress) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressBackend) DeepCopyInto(out *IngressBackend) { + *out = *in + if in.Service != nil { + in, out := &in.Service, &out.Service + *out = new(IngressServiceBackend) + **out = **in + } + if in.Resource != nil { + in, out := &in.Resource, &out.Resource + *out = new(corev1.TypedLocalObjectReference) + (*in).DeepCopyInto(*out) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressBackend. +func (in *IngressBackend) DeepCopy() *IngressBackend { + if in == nil { + return nil + } + out := new(IngressBackend) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressClass) DeepCopyInto(out *IngressClass) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + in.Spec.DeepCopyInto(&out.Spec) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressClass. +func (in *IngressClass) DeepCopy() *IngressClass { + if in == nil { + return nil + } + out := new(IngressClass) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *IngressClass) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressClassList) DeepCopyInto(out *IngressClassList) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ListMeta.DeepCopyInto(&out.ListMeta) + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]IngressClass, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressClassList. +func (in *IngressClassList) DeepCopy() *IngressClassList { + if in == nil { + return nil + } + out := new(IngressClassList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *IngressClassList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressClassSpec) DeepCopyInto(out *IngressClassSpec) { + *out = *in + if in.Parameters != nil { + in, out := &in.Parameters, &out.Parameters + *out = new(corev1.TypedLocalObjectReference) + (*in).DeepCopyInto(*out) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressClassSpec. +func (in *IngressClassSpec) DeepCopy() *IngressClassSpec { + if in == nil { + return nil + } + out := new(IngressClassSpec) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressList) DeepCopyInto(out *IngressList) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ListMeta.DeepCopyInto(&out.ListMeta) + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]Ingress, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressList. +func (in *IngressList) DeepCopy() *IngressList { + if in == nil { + return nil + } + out := new(IngressList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *IngressList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressRule) DeepCopyInto(out *IngressRule) { + *out = *in + in.IngressRuleValue.DeepCopyInto(&out.IngressRuleValue) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressRule. +func (in *IngressRule) DeepCopy() *IngressRule { + if in == nil { + return nil + } + out := new(IngressRule) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressRuleValue) DeepCopyInto(out *IngressRuleValue) { + *out = *in + if in.HTTP != nil { + in, out := &in.HTTP, &out.HTTP + *out = new(HTTPIngressRuleValue) + (*in).DeepCopyInto(*out) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressRuleValue. +func (in *IngressRuleValue) DeepCopy() *IngressRuleValue { + if in == nil { + return nil + } + out := new(IngressRuleValue) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressServiceBackend) DeepCopyInto(out *IngressServiceBackend) { + *out = *in + out.Port = in.Port + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressServiceBackend. +func (in *IngressServiceBackend) DeepCopy() *IngressServiceBackend { + if in == nil { + return nil + } + out := new(IngressServiceBackend) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressSpec) DeepCopyInto(out *IngressSpec) { + *out = *in + if in.IngressClassName != nil { + in, out := &in.IngressClassName, &out.IngressClassName + *out = new(string) + **out = **in + } + if in.DefaultBackend != nil { + in, out := &in.DefaultBackend, &out.DefaultBackend + *out = new(IngressBackend) + (*in).DeepCopyInto(*out) + } + if in.TLS != nil { + in, out := &in.TLS, &out.TLS + *out = make([]IngressTLS, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + if in.Rules != nil { + in, out := &in.Rules, &out.Rules + *out = make([]IngressRule, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressSpec. +func (in *IngressSpec) DeepCopy() *IngressSpec { + if in == nil { + return nil + } + out := new(IngressSpec) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressStatus) DeepCopyInto(out *IngressStatus) { + *out = *in + in.LoadBalancer.DeepCopyInto(&out.LoadBalancer) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressStatus. +func (in *IngressStatus) DeepCopy() *IngressStatus { + if in == nil { + return nil + } + out := new(IngressStatus) + in.DeepCopyInto(out) + return out +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressTLS) DeepCopyInto(out *IngressTLS) { + *out = *in + if in.Hosts != nil { + in, out := &in.Hosts, &out.Hosts + *out = make([]string, len(*in)) + copy(*out, *in) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressTLS. +func (in *IngressTLS) DeepCopy() *IngressTLS { + if in == nil { + return nil + } + out := new(IngressTLS) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *NetworkPolicy) DeepCopyInto(out *NetworkPolicy) { *out = *in @@ -260,3 +606,19 @@ func (in *NetworkPolicySpec) DeepCopy() *NetworkPolicySpec { in.DeepCopyInto(out) return out } + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *ServiceBackendPort) DeepCopyInto(out *ServiceBackendPort) { + *out = *in + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new ServiceBackendPort. +func (in *ServiceBackendPort) DeepCopy() *ServiceBackendPort { + if in == nil { + return nil + } + out := new(ServiceBackendPort) + in.DeepCopyInto(out) + return out +} diff --git a/vendor/k8s.io/api/networking/v1beta1/doc.go b/vendor/k8s.io/api/networking/v1beta1/doc.go index 12d3d4ff..fa6d01ce 100644 --- a/vendor/k8s.io/api/networking/v1beta1/doc.go +++ b/vendor/k8s.io/api/networking/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=networking.k8s.io package v1beta1 // import "k8s.io/api/networking/v1beta1" diff --git a/vendor/k8s.io/api/networking/v1beta1/generated.pb.go b/vendor/k8s.io/api/networking/v1beta1/generated.pb.go index 8ed56009..6f51df86 100644 --- a/vendor/k8s.io/api/networking/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/networking/v1beta1/generated.pb.go @@ -25,6 +25,7 @@ import ( io "io" proto "github.com/gogo/protobuf/proto" + v11 "k8s.io/api/core/v1" math "math" math_bits "math/bits" @@ -41,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *HTTPIngressPath) Reset() { *m = HTTPIngressPath{} } func (*HTTPIngressPath) ProtoMessage() {} @@ -155,10 +156,94 @@ func (m *IngressBackend) XXX_DiscardUnknown() { var xxx_messageInfo_IngressBackend proto.InternalMessageInfo +func (m *IngressClass) Reset() { *m = IngressClass{} } +func (*IngressClass) ProtoMessage() {} +func (*IngressClass) Descriptor() ([]byte, []int) { + return fileDescriptor_5bea11de0ceb8f53, []int{4} +} +func (m *IngressClass) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressClass) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressClass) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressClass.Merge(m, src) +} +func (m *IngressClass) XXX_Size() int { + return m.Size() +} +func (m *IngressClass) XXX_DiscardUnknown() { + xxx_messageInfo_IngressClass.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressClass proto.InternalMessageInfo + +func (m *IngressClassList) Reset() { *m = IngressClassList{} } +func (*IngressClassList) ProtoMessage() {} +func (*IngressClassList) Descriptor() ([]byte, []int) { + return fileDescriptor_5bea11de0ceb8f53, []int{5} +} +func (m *IngressClassList) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressClassList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressClassList) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressClassList.Merge(m, src) +} +func (m *IngressClassList) XXX_Size() int { + return m.Size() +} +func (m *IngressClassList) XXX_DiscardUnknown() { + xxx_messageInfo_IngressClassList.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressClassList proto.InternalMessageInfo + +func (m *IngressClassSpec) Reset() { *m = IngressClassSpec{} } +func (*IngressClassSpec) ProtoMessage() {} +func (*IngressClassSpec) Descriptor() ([]byte, []int) { + return fileDescriptor_5bea11de0ceb8f53, []int{6} +} +func (m *IngressClassSpec) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *IngressClassSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *IngressClassSpec) XXX_Merge(src proto.Message) { + xxx_messageInfo_IngressClassSpec.Merge(m, src) +} +func (m *IngressClassSpec) XXX_Size() int { + return m.Size() +} +func (m *IngressClassSpec) XXX_DiscardUnknown() { + xxx_messageInfo_IngressClassSpec.DiscardUnknown(m) +} + +var xxx_messageInfo_IngressClassSpec proto.InternalMessageInfo + func (m *IngressList) Reset() { *m = IngressList{} } func (*IngressList) ProtoMessage() {} func (*IngressList) Descriptor() ([]byte, []int) { - return fileDescriptor_5bea11de0ceb8f53, []int{4} + return fileDescriptor_5bea11de0ceb8f53, []int{7} } func (m *IngressList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -186,7 +271,7 @@ var xxx_messageInfo_IngressList proto.InternalMessageInfo func (m *IngressRule) Reset() { *m = IngressRule{} } func (*IngressRule) ProtoMessage() {} func (*IngressRule) Descriptor() ([]byte, []int) { - return fileDescriptor_5bea11de0ceb8f53, []int{5} + return fileDescriptor_5bea11de0ceb8f53, []int{8} } func (m *IngressRule) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -214,7 +299,7 @@ var xxx_messageInfo_IngressRule proto.InternalMessageInfo func (m *IngressRuleValue) Reset() { *m = IngressRuleValue{} } func (*IngressRuleValue) ProtoMessage() {} func (*IngressRuleValue) Descriptor() ([]byte, []int) { - return fileDescriptor_5bea11de0ceb8f53, []int{6} + return fileDescriptor_5bea11de0ceb8f53, []int{9} } func (m *IngressRuleValue) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -242,7 +327,7 @@ var xxx_messageInfo_IngressRuleValue proto.InternalMessageInfo func (m *IngressSpec) Reset() { *m = IngressSpec{} } func (*IngressSpec) ProtoMessage() {} func (*IngressSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_5bea11de0ceb8f53, []int{7} + return fileDescriptor_5bea11de0ceb8f53, []int{10} } func (m *IngressSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -270,7 +355,7 @@ var xxx_messageInfo_IngressSpec proto.InternalMessageInfo func (m *IngressStatus) Reset() { *m = IngressStatus{} } func (*IngressStatus) ProtoMessage() {} func (*IngressStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_5bea11de0ceb8f53, []int{8} + return fileDescriptor_5bea11de0ceb8f53, []int{11} } func (m *IngressStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -298,7 +383,7 @@ var xxx_messageInfo_IngressStatus proto.InternalMessageInfo func (m *IngressTLS) Reset() { *m = IngressTLS{} } func (*IngressTLS) ProtoMessage() {} func (*IngressTLS) Descriptor() ([]byte, []int) { - return fileDescriptor_5bea11de0ceb8f53, []int{9} + return fileDescriptor_5bea11de0ceb8f53, []int{12} } func (m *IngressTLS) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -328,6 +413,9 @@ func init() { proto.RegisterType((*HTTPIngressRuleValue)(nil), "k8s.io.api.networking.v1beta1.HTTPIngressRuleValue") proto.RegisterType((*Ingress)(nil), "k8s.io.api.networking.v1beta1.Ingress") proto.RegisterType((*IngressBackend)(nil), "k8s.io.api.networking.v1beta1.IngressBackend") + proto.RegisterType((*IngressClass)(nil), "k8s.io.api.networking.v1beta1.IngressClass") + proto.RegisterType((*IngressClassList)(nil), "k8s.io.api.networking.v1beta1.IngressClassList") + proto.RegisterType((*IngressClassSpec)(nil), "k8s.io.api.networking.v1beta1.IngressClassSpec") proto.RegisterType((*IngressList)(nil), "k8s.io.api.networking.v1beta1.IngressList") proto.RegisterType((*IngressRule)(nil), "k8s.io.api.networking.v1beta1.IngressRule") proto.RegisterType((*IngressRuleValue)(nil), "k8s.io.api.networking.v1beta1.IngressRuleValue") @@ -341,58 +429,69 @@ func init() { } var fileDescriptor_5bea11de0ceb8f53 = []byte{ - // 812 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x55, 0xcf, 0x6e, 0xfb, 0x44, - 0x10, 0x8e, 0xf3, 0xa7, 0x69, 0xd7, 0xfd, 0xa7, 0xa5, 0x87, 0xa8, 0x12, 0x6e, 0xe4, 0x03, 0x2a, - 0x88, 0xae, 0x69, 0x0a, 0x88, 0xb3, 0x0f, 0xa8, 0x15, 0x81, 0x86, 0x75, 0x84, 0x10, 0xe2, 0xd0, - 0x8d, 0xb3, 0x38, 0x26, 0x89, 0x6d, 0x76, 0xd7, 0x41, 0xdc, 0x78, 0x01, 0x04, 0x4f, 0xc1, 0x99, - 0x23, 0x8f, 0xd0, 0x63, 0x8f, 0x3d, 0x55, 0x34, 0xbc, 0x07, 0x42, 0xbb, 0xde, 0xda, 0x4e, 0xd2, - 0xfe, 0x6a, 0xfd, 0x6e, 0xde, 0x9d, 0xf9, 0xbe, 0xd9, 0x99, 0xf9, 0x66, 0x0c, 0x3e, 0x9f, 0x7e, - 0xc6, 0x51, 0x18, 0x3b, 0xd3, 0x74, 0x44, 0x59, 0x44, 0x05, 0xe5, 0xce, 0x82, 0x46, 0xe3, 0x98, - 0x39, 0xda, 0x40, 0x92, 0xd0, 0x89, 0xa8, 0xf8, 0x39, 0x66, 0xd3, 0x30, 0x0a, 0x9c, 0xc5, 0xf9, - 0x88, 0x0a, 0x72, 0xee, 0x04, 0x34, 0xa2, 0x8c, 0x08, 0x3a, 0x46, 0x09, 0x8b, 0x45, 0x0c, 0xdf, - 0xcd, 0xdc, 0x11, 0x49, 0x42, 0x54, 0xb8, 0x23, 0xed, 0x7e, 0x7c, 0x16, 0x84, 0x62, 0x92, 0x8e, - 0x90, 0x1f, 0xcf, 0x9d, 0x20, 0x0e, 0x62, 0x47, 0xa1, 0x46, 0xe9, 0x0f, 0xea, 0xa4, 0x0e, 0xea, - 0x2b, 0x63, 0x3b, 0xb6, 0x4b, 0xc1, 0xfd, 0x98, 0x51, 0x67, 0xb1, 0x11, 0xf1, 0xf8, 0xe3, 0xc2, - 0x67, 0x4e, 0xfc, 0x49, 0x18, 0x51, 0xf6, 0x8b, 0x93, 0x4c, 0x03, 0x79, 0xc1, 0x9d, 0x39, 0x15, - 0xe4, 0x39, 0x94, 0xf3, 0x12, 0x8a, 0xa5, 0x91, 0x08, 0xe7, 0x74, 0x03, 0xf0, 0xe9, 0x6b, 0x00, - 0xee, 0x4f, 0xe8, 0x9c, 0x6c, 0xe0, 0x2e, 0x5e, 0xc2, 0xa5, 0x22, 0x9c, 0x39, 0x61, 0x24, 0xb8, - 0x60, 0xeb, 0x20, 0xfb, 0x37, 0x03, 0x1c, 0x5c, 0x0e, 0x87, 0x83, 0xab, 0x28, 0x60, 0x94, 0xf3, - 0x01, 0x11, 0x13, 0xd8, 0x05, 0xcd, 0x84, 0x88, 0x49, 0xc7, 0xe8, 0x1a, 0xa7, 0x3b, 0xee, 0xee, - 0xed, 0xc3, 0x49, 0x6d, 0xf9, 0x70, 0xd2, 0x94, 0x36, 0xac, 0x2c, 0xf0, 0x5b, 0xd0, 0x1e, 0x11, - 0x7f, 0x4a, 0xa3, 0x71, 0xa7, 0xde, 0x35, 0x4e, 0xcd, 0xde, 0x19, 0x7a, 0x63, 0x37, 0x90, 0xa6, - 0x77, 0x33, 0x90, 0x7b, 0xa0, 0x39, 0xdb, 0xfa, 0x02, 0x3f, 0xd1, 0xd9, 0x53, 0x70, 0x54, 0x7a, - 0x0e, 0x4e, 0x67, 0xf4, 0x1b, 0x32, 0x4b, 0x29, 0xf4, 0x40, 0x4b, 0x46, 0xe6, 0x1d, 0xa3, 0xdb, - 0x38, 0x35, 0x7b, 0xe8, 0x95, 0x78, 0x6b, 0x29, 0xb9, 0x7b, 0x3a, 0x60, 0x4b, 0x9e, 0x38, 0xce, - 0xb8, 0xec, 0xdf, 0xeb, 0xa0, 0xad, 0xbd, 0xe0, 0x0d, 0xd8, 0x96, 0x1d, 0x1c, 0x13, 0x41, 0x54, - 0xe2, 0x66, 0xef, 0xa3, 0x52, 0x8c, 0xbc, 0xa0, 0x28, 0x99, 0x06, 0xf2, 0x82, 0x23, 0xe9, 0x8d, - 0x16, 0xe7, 0xe8, 0x7a, 0xf4, 0x23, 0xf5, 0xc5, 0x97, 0x54, 0x10, 0x17, 0xea, 0x28, 0xa0, 0xb8, - 0xc3, 0x39, 0x2b, 0xec, 0x83, 0x26, 0x4f, 0xa8, 0xaf, 0x2b, 0xf6, 0x41, 0xb5, 0x8a, 0x79, 0x09, - 0xf5, 0x8b, 0x16, 0xc8, 0x13, 0x56, 0x2c, 0x70, 0x08, 0xb6, 0xb8, 0x20, 0x22, 0xe5, 0x9d, 0x86, - 0xe2, 0xfb, 0xb0, 0x22, 0x9f, 0xc2, 0xb8, 0xfb, 0x9a, 0x71, 0x2b, 0x3b, 0x63, 0xcd, 0x65, 0xff, - 0x65, 0x80, 0xfd, 0xd5, 0x5e, 0xc1, 0x4f, 0x80, 0xc9, 0x29, 0x5b, 0x84, 0x3e, 0xfd, 0x8a, 0xcc, - 0xa9, 0x16, 0xc5, 0x3b, 0x1a, 0x6f, 0x7a, 0x85, 0x09, 0x97, 0xfd, 0x60, 0x90, 0xc3, 0x06, 0x31, - 0x13, 0x3a, 0xe9, 0x97, 0x4b, 0x2a, 0x35, 0x8a, 0x32, 0x8d, 0xa2, 0xab, 0x48, 0x5c, 0x33, 0x4f, - 0xb0, 0x30, 0x0a, 0x36, 0x02, 0x49, 0x32, 0x5c, 0x66, 0xb6, 0xff, 0x36, 0x80, 0xa9, 0x9f, 0xdc, - 0x0f, 0xb9, 0x80, 0xdf, 0x6f, 0x34, 0x12, 0x55, 0x6b, 0xa4, 0x44, 0xab, 0x36, 0x1e, 0xea, 0x98, - 0xdb, 0x4f, 0x37, 0xa5, 0x26, 0x7e, 0x01, 0x5a, 0xa1, 0xa0, 0x73, 0xde, 0xa9, 0x2b, 0x1d, 0xbe, - 0x57, 0x51, 0xf7, 0xb9, 0xfe, 0xae, 0x24, 0x18, 0x67, 0x1c, 0xf6, 0x9f, 0xc5, 0xd3, 0xa5, 0xd2, - 0xe5, 0xe0, 0x4d, 0x62, 0x2e, 0xd6, 0x07, 0xef, 0x32, 0xe6, 0x02, 0x2b, 0x0b, 0x4c, 0xc1, 0x61, - 0xb8, 0x36, 0x1a, 0xba, 0xb4, 0x4e, 0xb5, 0x97, 0xe4, 0x30, 0xb7, 0xa3, 0xe9, 0x0f, 0xd7, 0x2d, - 0x78, 0x23, 0x84, 0x4d, 0xc1, 0x86, 0x17, 0xfc, 0x1a, 0x34, 0x27, 0x42, 0x24, 0xba, 0xc6, 0x17, - 0xd5, 0x07, 0xb2, 0x78, 0xc2, 0xb6, 0xca, 0x6e, 0x38, 0x1c, 0x60, 0x45, 0x65, 0xff, 0x57, 0xd4, - 0xc3, 0xcb, 0x34, 0x9e, 0xaf, 0x19, 0xe3, 0x6d, 0xd6, 0x8c, 0xf9, 0xdc, 0x8a, 0x81, 0x97, 0xa0, - 0x21, 0x66, 0x4f, 0x0d, 0x7c, 0xbf, 0x1a, 0xe3, 0xb0, 0xef, 0xb9, 0xa6, 0x2e, 0x58, 0x63, 0xd8, - 0xf7, 0xb0, 0xa4, 0x80, 0xd7, 0xa0, 0xc5, 0xd2, 0x19, 0x95, 0x23, 0xd8, 0xa8, 0x3e, 0xd2, 0x32, - 0xff, 0x42, 0x10, 0xf2, 0xc4, 0x71, 0xc6, 0x63, 0xff, 0x04, 0xf6, 0x56, 0xe6, 0x14, 0xde, 0x80, - 0xdd, 0x59, 0x4c, 0xc6, 0x2e, 0x99, 0x91, 0xc8, 0xa7, 0x4c, 0x97, 0x61, 0x45, 0x75, 0xf2, 0x6f, - 0xa5, 0xe4, 0x5b, 0xf2, 0xd3, 0x53, 0x7e, 0xa4, 0x83, 0xec, 0x96, 0x6d, 0x78, 0x85, 0xd1, 0x26, - 0x00, 0x14, 0x39, 0xc2, 0x13, 0xd0, 0x92, 0x3a, 0xcb, 0xd6, 0xec, 0x8e, 0xbb, 0x23, 0x5f, 0x28, - 0xe5, 0xc7, 0x71, 0x76, 0x0f, 0x7b, 0x00, 0x70, 0xea, 0x33, 0x2a, 0xd4, 0x32, 0xa8, 0x2b, 0xa1, - 0xe6, 0x6b, 0xcf, 0xcb, 0x2d, 0xb8, 0xe4, 0xe5, 0x9e, 0xdd, 0x3e, 0x5a, 0xb5, 0xbb, 0x47, 0xab, - 0x76, 0xff, 0x68, 0xd5, 0x7e, 0x5d, 0x5a, 0xc6, 0xed, 0xd2, 0x32, 0xee, 0x96, 0x96, 0x71, 0xbf, - 0xb4, 0x8c, 0x7f, 0x96, 0x96, 0xf1, 0xc7, 0xbf, 0x56, 0xed, 0xbb, 0xb6, 0x2e, 0xd3, 0xff, 0x01, - 0x00, 0x00, 0xff, 0xff, 0xdb, 0x8a, 0xe4, 0xd8, 0x21, 0x08, 0x00, 0x00, + // 990 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x56, 0x4f, 0x6f, 0xe3, 0x44, + 0x14, 0xaf, 0x93, 0x66, 0x9b, 0x4e, 0xb2, 0xdd, 0x6a, 0xe8, 0x21, 0xaa, 0x84, 0x5b, 0xf9, 0x80, + 0xca, 0x9f, 0xda, 0x34, 0xbb, 0x20, 0x8e, 0xc8, 0x2b, 0xa1, 0x56, 0x04, 0x1a, 0x26, 0x16, 0x20, + 0x04, 0xd2, 0x4e, 0x9c, 0xb7, 0x8e, 0x89, 0x63, 0x9b, 0x99, 0x71, 0xd0, 0xde, 0xb8, 0x72, 0x82, + 0x2f, 0x01, 0x9f, 0x81, 0x23, 0x82, 0x4b, 0x8f, 0x7b, 0xdc, 0x53, 0x45, 0xc3, 0xb7, 0xe0, 0x84, + 0x66, 0x3c, 0xb5, 0x9d, 0xa4, 0xa5, 0x59, 0x0e, 0x7b, 0x8a, 0x67, 0xde, 0x7b, 0xbf, 0x37, 0xef, + 0xf7, 0x7e, 0x33, 0x2f, 0xe8, 0xa3, 0xc9, 0x07, 0xdc, 0x0e, 0x13, 0x67, 0x92, 0x0d, 0x81, 0xc5, + 0x20, 0x80, 0x3b, 0x33, 0x88, 0x47, 0x09, 0x73, 0xb4, 0x81, 0xa6, 0xa1, 0x13, 0x83, 0xf8, 0x3e, + 0x61, 0x93, 0x30, 0x0e, 0x9c, 0xd9, 0xc9, 0x10, 0x04, 0x3d, 0x71, 0x02, 0x88, 0x81, 0x51, 0x01, + 0x23, 0x3b, 0x65, 0x89, 0x48, 0xf0, 0xeb, 0xb9, 0xbb, 0x4d, 0xd3, 0xd0, 0x2e, 0xdd, 0x6d, 0xed, + 0xbe, 0x7f, 0x1c, 0x84, 0x62, 0x9c, 0x0d, 0x6d, 0x3f, 0x99, 0x3a, 0x41, 0x12, 0x24, 0x8e, 0x8a, + 0x1a, 0x66, 0x4f, 0xd5, 0x4a, 0x2d, 0xd4, 0x57, 0x8e, 0xb6, 0x6f, 0x55, 0x92, 0xfb, 0x09, 0x03, + 0x67, 0xb6, 0x92, 0x71, 0xff, 0x51, 0xe9, 0x33, 0xa5, 0xfe, 0x38, 0x8c, 0x81, 0x3d, 0x73, 0xd2, + 0x49, 0x20, 0x37, 0xb8, 0x33, 0x05, 0x41, 0x6f, 0x8a, 0x72, 0x6e, 0x8b, 0x62, 0x59, 0x2c, 0xc2, + 0x29, 0xac, 0x04, 0xbc, 0x7f, 0x57, 0x00, 0xf7, 0xc7, 0x30, 0xa5, 0x2b, 0x71, 0x0f, 0x6f, 0x8b, + 0xcb, 0x44, 0x18, 0x39, 0x61, 0x2c, 0xb8, 0x60, 0xcb, 0x41, 0xd6, 0x9f, 0x06, 0x7a, 0x70, 0xea, + 0x79, 0xfd, 0xb3, 0x38, 0x60, 0xc0, 0x79, 0x9f, 0x8a, 0x31, 0x3e, 0x44, 0x9b, 0x29, 0x15, 0xe3, + 0x8e, 0x71, 0x68, 0x1c, 0x6d, 0xbb, 0xed, 0x8b, 0xcb, 0x83, 0x8d, 0xf9, 0xe5, 0xc1, 0xa6, 0xb4, + 0x11, 0x65, 0xc1, 0x8f, 0x50, 0x53, 0xfe, 0x7a, 0xcf, 0x52, 0xe8, 0xd4, 0x95, 0x57, 0x67, 0x7e, + 0x79, 0xd0, 0xec, 0xeb, 0xbd, 0x7f, 0x2a, 0xdf, 0xa4, 0xf0, 0xc4, 0x5f, 0xa2, 0xad, 0x21, 0xf5, + 0x27, 0x10, 0x8f, 0x3a, 0xb5, 0x43, 0xe3, 0xa8, 0xd5, 0x3d, 0xb6, 0xff, 0xb3, 0x87, 0xb6, 0x3e, + 0x94, 0x9b, 0x07, 0xb9, 0x0f, 0xf4, 0x49, 0xb6, 0xf4, 0x06, 0xb9, 0x86, 0xb3, 0x26, 0x68, 0xaf, + 0x52, 0x04, 0xc9, 0x22, 0xf8, 0x9c, 0x46, 0x19, 0xe0, 0x01, 0x6a, 0xc8, 0xec, 0xbc, 0x63, 0x1c, + 0xd6, 0x8f, 0x5a, 0x5d, 0xfb, 0x8e, 0x7c, 0x4b, 0x44, 0xb8, 0xf7, 0x75, 0xc2, 0x86, 0x5c, 0x71, + 0x92, 0x63, 0x59, 0x3f, 0xd5, 0xd0, 0x96, 0xf6, 0xc2, 0x4f, 0x50, 0x53, 0xf6, 0x7d, 0x44, 0x05, + 0x55, 0x74, 0xb5, 0xba, 0xef, 0x56, 0x72, 0x14, 0x6d, 0xb0, 0xd3, 0x49, 0x20, 0x37, 0xb8, 0x2d, + 0xbd, 0xed, 0xd9, 0x89, 0x7d, 0x3e, 0xfc, 0x16, 0x7c, 0xf1, 0x09, 0x08, 0xea, 0x62, 0x9d, 0x05, + 0x95, 0x7b, 0xa4, 0x40, 0xc5, 0x3d, 0xb4, 0xc9, 0x53, 0xf0, 0x35, 0x63, 0x6f, 0xad, 0xc7, 0xd8, + 0x20, 0x05, 0xbf, 0x6c, 0x9c, 0x5c, 0x11, 0x85, 0x82, 0x3d, 0x74, 0x8f, 0x0b, 0x2a, 0x32, 0xae, + 0xda, 0xd6, 0xea, 0xbe, 0xb3, 0x26, 0x9e, 0x8a, 0x71, 0x77, 0x34, 0xe2, 0xbd, 0x7c, 0x4d, 0x34, + 0x96, 0xf5, 0x63, 0x0d, 0xed, 0x2c, 0xf6, 0x0a, 0xbf, 0x87, 0x5a, 0x1c, 0xd8, 0x2c, 0xf4, 0xe1, + 0x53, 0x3a, 0x05, 0x2d, 0xa5, 0xd7, 0x74, 0x7c, 0x6b, 0x50, 0x9a, 0x48, 0xd5, 0x0f, 0x07, 0x45, + 0x58, 0x3f, 0x61, 0x42, 0x17, 0x7d, 0x3b, 0xa5, 0x52, 0xd9, 0x76, 0xae, 0x6c, 0xfb, 0x2c, 0x16, + 0xe7, 0x6c, 0x20, 0x58, 0x18, 0x07, 0x2b, 0x89, 0x24, 0x18, 0xa9, 0x22, 0xe3, 0x2f, 0x50, 0x93, + 0x01, 0x4f, 0x32, 0xe6, 0x83, 0xa6, 0x62, 0x41, 0x8c, 0xf2, 0x09, 0x90, 0x6d, 0x92, 0xba, 0x1d, + 0xf5, 0x12, 0x9f, 0x46, 0x79, 0x73, 0x08, 0x3c, 0x05, 0x06, 0xb1, 0x0f, 0x6e, 0x5b, 0x0a, 0x9e, + 0x68, 0x08, 0x52, 0x80, 0xc9, 0x0b, 0xd5, 0xd6, 0x5c, 0x3c, 0x8e, 0xe8, 0x2b, 0x91, 0xc8, 0x67, + 0x0b, 0x12, 0x71, 0xd6, 0x6b, 0xa9, 0x3a, 0xdc, 0x6d, 0x3a, 0xb1, 0xfe, 0x30, 0xd0, 0x6e, 0xd5, + 0xb1, 0x17, 0x72, 0x81, 0xbf, 0x5e, 0xa9, 0xc4, 0x5e, 0xaf, 0x12, 0x19, 0xad, 0xea, 0xd8, 0xd5, + 0xa9, 0x9a, 0xd7, 0x3b, 0x95, 0x2a, 0xfa, 0xa8, 0x11, 0x0a, 0x98, 0xf2, 0x4e, 0x4d, 0xdd, 0xd5, + 0xb7, 0x5f, 0xa2, 0x8c, 0xf2, 0xa2, 0x9e, 0x49, 0x04, 0x92, 0x03, 0x59, 0xbf, 0x2c, 0x15, 0x21, + 0xeb, 0xc3, 0x5d, 0x84, 0xfc, 0x24, 0x16, 0x2c, 0x89, 0x22, 0x60, 0x5a, 0x97, 0x05, 0xbd, 0x8f, + 0x0b, 0x0b, 0xa9, 0x78, 0xe1, 0x6f, 0x10, 0x4a, 0x29, 0xa3, 0x53, 0x10, 0xc0, 0xf8, 0x4d, 0x6f, + 0xd7, 0xdd, 0x72, 0xd9, 0x91, 0xf0, 0xfd, 0x02, 0x84, 0x54, 0x00, 0xad, 0xdf, 0x0c, 0xd4, 0xd2, + 0xe7, 0x7c, 0x05, 0x3c, 0x7f, 0xbc, 0xc8, 0xf3, 0x1b, 0x6b, 0xbe, 0xc1, 0x37, 0x53, 0xfc, 0x6b, + 0x79, 0x74, 0xf9, 0xea, 0xca, 0xd1, 0x31, 0x4e, 0xb8, 0x58, 0x1e, 0x1d, 0xa7, 0x09, 0x17, 0x44, + 0x59, 0x70, 0x86, 0x76, 0xc3, 0xa5, 0x67, 0xfa, 0xe5, 0x84, 0x5b, 0x84, 0xb9, 0x1d, 0x0d, 0xbf, + 0xbb, 0x6c, 0x21, 0x2b, 0x29, 0x2c, 0x40, 0x2b, 0x5e, 0xf2, 0xde, 0x8c, 0x85, 0x48, 0x35, 0xc7, + 0x0f, 0xd7, 0x1f, 0x0e, 0xe5, 0x11, 0x9a, 0xaa, 0x3a, 0xcf, 0xeb, 0x13, 0x05, 0x65, 0xfd, 0x5e, + 0x2b, 0xf8, 0x50, 0x6a, 0xfb, 0xb0, 0xa8, 0x56, 0x29, 0x50, 0xbd, 0x85, 0x9b, 0x8a, 0x9b, 0xbd, + 0xca, 0xc1, 0x0b, 0x1b, 0x59, 0xf1, 0xc6, 0x5e, 0x39, 0x34, 0x8d, 0xff, 0x33, 0x34, 0x5b, 0x37, + 0x0d, 0x4c, 0x7c, 0x8a, 0xea, 0x22, 0xba, 0x96, 0xc0, 0x9b, 0xeb, 0x21, 0x7a, 0xbd, 0x81, 0xdb, + 0xd2, 0x94, 0xd7, 0xbd, 0xde, 0x80, 0x48, 0x08, 0x7c, 0x8e, 0x1a, 0x2c, 0x8b, 0x40, 0x0e, 0x94, + 0xfa, 0xfa, 0x03, 0x4a, 0x32, 0x58, 0x4a, 0x4a, 0xae, 0x38, 0xc9, 0x71, 0xac, 0xef, 0xd0, 0xfd, + 0x85, 0xa9, 0x83, 0x9f, 0xa0, 0x76, 0x94, 0xd0, 0x91, 0x4b, 0x23, 0x1a, 0xfb, 0xfa, 0xce, 0x2e, + 0xe9, 0xf6, 0xfa, 0xfe, 0xf5, 0x2a, 0x7e, 0x7a, 0x66, 0xed, 0xe9, 0x24, 0xed, 0xaa, 0x8d, 0x2c, + 0x20, 0x5a, 0x14, 0xa1, 0xb2, 0x46, 0x7c, 0x80, 0x1a, 0x52, 0xa9, 0xf9, 0x9f, 0x86, 0x6d, 0x77, + 0x5b, 0x9e, 0x50, 0x0a, 0x98, 0x93, 0x7c, 0x5f, 0x3e, 0x21, 0x1c, 0x7c, 0x06, 0x42, 0xb5, 0xb3, + 0xb6, 0xf8, 0x84, 0x0c, 0x0a, 0x0b, 0xa9, 0x78, 0xb9, 0xc7, 0x17, 0x57, 0xe6, 0xc6, 0xf3, 0x2b, + 0x73, 0xe3, 0xc5, 0x95, 0xb9, 0xf1, 0xc3, 0xdc, 0x34, 0x2e, 0xe6, 0xa6, 0xf1, 0x7c, 0x6e, 0x1a, + 0x2f, 0xe6, 0xa6, 0xf1, 0xd7, 0xdc, 0x34, 0x7e, 0xfe, 0xdb, 0xdc, 0xf8, 0x6a, 0x4b, 0xd3, 0xf4, + 0x6f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x6e, 0x54, 0x4d, 0x9d, 0x25, 0x0b, 0x00, 0x00, } func (m *HTTPIngressPath) Marshal() (dAtA []byte, err error) { @@ -415,6 +514,13 @@ func (m *HTTPIngressPath) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.PathType != nil { + i -= len(*m.PathType) + copy(dAtA[i:], *m.PathType) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.PathType))) + i-- + dAtA[i] = 0x1a + } { size, err := m.Backend.MarshalToSizedBuffer(dAtA[:i]) if err != nil { @@ -543,6 +649,18 @@ func (m *IngressBackend) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.Resource != nil { + { + size, err := m.Resource.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x1a + } { size, err := m.ServicePort.MarshalToSizedBuffer(dAtA[:i]) if err != nil { @@ -561,6 +679,136 @@ func (m *IngressBackend) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } +func (m *IngressClass) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *IngressClass) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressClass) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + { + size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *IngressClassList) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *IngressClassList) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressClassList) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + { + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *IngressClassSpec) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *IngressClassSpec) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *IngressClassSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.Parameters != nil { + { + size, err := m.Parameters.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + i -= len(m.Controller) + copy(dAtA[i:], m.Controller) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Controller))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + func (m *IngressList) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -701,6 +949,13 @@ func (m *IngressSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.IngressClassName != nil { + i -= len(*m.IngressClassName) + copy(dAtA[i:], *m.IngressClassName) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.IngressClassName))) + i-- + dAtA[i] = 0x22 + } if len(m.Rules) > 0 { for iNdEx := len(m.Rules) - 1; iNdEx >= 0; iNdEx-- { { @@ -835,6 +1090,10 @@ func (m *HTTPIngressPath) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) l = m.Backend.Size() n += 1 + l + sovGenerated(uint64(l)) + if m.PathType != nil { + l = len(*m.PathType) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -878,6 +1137,55 @@ func (m *IngressBackend) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) l = m.ServicePort.Size() n += 1 + l + sovGenerated(uint64(l)) + if m.Resource != nil { + l = m.Resource.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *IngressClass) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Spec.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *IngressClassList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *IngressClassSpec) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Controller) + n += 1 + l + sovGenerated(uint64(l)) + if m.Parameters != nil { + l = m.Parameters.Size() + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -946,6 +1254,10 @@ func (m *IngressSpec) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) } } + if m.IngressClassName != nil { + l = len(*m.IngressClassName) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -990,6 +1302,7 @@ func (this *HTTPIngressPath) String() string { s := strings.Join([]string{`&HTTPIngressPath{`, `Path:` + fmt.Sprintf("%v", this.Path) + `,`, `Backend:` + strings.Replace(strings.Replace(this.Backend.String(), "IngressBackend", "IngressBackend", 1), `&`, ``, 1) + `,`, + `PathType:` + valueToStringGenerated(this.PathType) + `,`, `}`, }, "") return s @@ -1028,46 +1341,85 @@ func (this *IngressBackend) String() string { s := strings.Join([]string{`&IngressBackend{`, `ServiceName:` + fmt.Sprintf("%v", this.ServiceName) + `,`, `ServicePort:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ServicePort), "IntOrString", "intstr.IntOrString", 1), `&`, ``, 1) + `,`, + `Resource:` + strings.Replace(fmt.Sprintf("%v", this.Resource), "TypedLocalObjectReference", "v11.TypedLocalObjectReference", 1) + `,`, `}`, }, "") return s } -func (this *IngressList) String() string { +func (this *IngressClass) String() string { if this == nil { return "nil" } - repeatedStringForItems := "[]Ingress{" + s := strings.Join([]string{`&IngressClass{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "IngressClassSpec", "IngressClassSpec", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressClassList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]IngressClass{" for _, f := range this.Items { - repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "Ingress", "Ingress", 1), `&`, ``, 1) + "," + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "IngressClass", "IngressClass", 1), `&`, ``, 1) + "," } repeatedStringForItems += "}" - s := strings.Join([]string{`&IngressList{`, + s := strings.Join([]string{`&IngressClassList{`, `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, `Items:` + repeatedStringForItems + `,`, `}`, }, "") return s } -func (this *IngressRule) String() string { +func (this *IngressClassSpec) String() string { if this == nil { return "nil" } - s := strings.Join([]string{`&IngressRule{`, - `Host:` + fmt.Sprintf("%v", this.Host) + `,`, - `IngressRuleValue:` + strings.Replace(strings.Replace(this.IngressRuleValue.String(), "IngressRuleValue", "IngressRuleValue", 1), `&`, ``, 1) + `,`, + s := strings.Join([]string{`&IngressClassSpec{`, + `Controller:` + fmt.Sprintf("%v", this.Controller) + `,`, + `Parameters:` + strings.Replace(fmt.Sprintf("%v", this.Parameters), "TypedLocalObjectReference", "v11.TypedLocalObjectReference", 1) + `,`, `}`, }, "") return s } -func (this *IngressRuleValue) String() string { +func (this *IngressList) String() string { if this == nil { return "nil" } - s := strings.Join([]string{`&IngressRuleValue{`, - `HTTP:` + strings.Replace(this.HTTP.String(), "HTTPIngressRuleValue", "HTTPIngressRuleValue", 1) + `,`, - `}`, - }, "") - return s + repeatedStringForItems := "[]Ingress{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "Ingress", "Ingress", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&IngressList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} +func (this *IngressRule) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressRule{`, + `Host:` + fmt.Sprintf("%v", this.Host) + `,`, + `IngressRuleValue:` + strings.Replace(strings.Replace(this.IngressRuleValue.String(), "IngressRuleValue", "IngressRuleValue", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *IngressRuleValue) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&IngressRuleValue{`, + `HTTP:` + strings.Replace(this.HTTP.String(), "HTTPIngressRuleValue", "HTTPIngressRuleValue", 1) + `,`, + `}`, + }, "") + return s } func (this *IngressSpec) String() string { if this == nil { @@ -1087,6 +1439,7 @@ func (this *IngressSpec) String() string { `Backend:` + strings.Replace(this.Backend.String(), "IngressBackend", "IngressBackend", 1) + `,`, `TLS:` + repeatedStringForTLS + `,`, `Rules:` + repeatedStringForRules + `,`, + `IngressClassName:` + valueToStringGenerated(this.IngressClassName) + `,`, `}`, }, "") return s @@ -1105,22 +1458,412 @@ func (this *IngressTLS) String() string { if this == nil { return "nil" } - s := strings.Join([]string{`&IngressTLS{`, - `Hosts:` + fmt.Sprintf("%v", this.Hosts) + `,`, - `SecretName:` + fmt.Sprintf("%v", this.SecretName) + `,`, - `}`, - }, "") - return s -} -func valueToStringGenerated(v interface{}) string { - rv := reflect.ValueOf(v) - if rv.IsNil() { - return "nil" + s := strings.Join([]string{`&IngressTLS{`, + `Hosts:` + fmt.Sprintf("%v", this.Hosts) + `,`, + `SecretName:` + fmt.Sprintf("%v", this.SecretName) + `,`, + `}`, + }, "") + return s +} +func valueToStringGenerated(v interface{}) string { + rv := reflect.ValueOf(v) + if rv.IsNil() { + return "nil" + } + pv := reflect.Indirect(rv).Interface() + return fmt.Sprintf("*%v", pv) +} +func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: HTTPIngressPath: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: HTTPIngressPath: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Path", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Path = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Backend", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Backend.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field PathType", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := PathType(dAtA[iNdEx:postIndex]) + m.PathType = &s + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *HTTPIngressRuleValue) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: HTTPIngressRuleValue: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: HTTPIngressRuleValue: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Paths", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Paths = append(m.Paths, HTTPIngressPath{}) + if err := m.Paths[len(m.Paths)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *Ingress) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: Ingress: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: Ingress: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF } - pv := reflect.Indirect(rv).Interface() - return fmt.Sprintf("*%v", pv) + return nil } -func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { +func (m *IngressBackend) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -1143,15 +1886,15 @@ func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: HTTPIngressPath: wiretype end group for non-group") + return fmt.Errorf("proto: IngressBackend: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: HTTPIngressPath: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: IngressBackend: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Path", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ServiceName", wireType) } var stringLen uint64 for shift := uint(0); ; shift += 7 { @@ -1179,11 +1922,11 @@ func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - m.Path = string(dAtA[iNdEx:postIndex]) + m.ServiceName = string(dAtA[iNdEx:postIndex]) iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Backend", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ServicePort", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -1210,7 +1953,43 @@ func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Backend.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if err := m.ServicePort.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Resource", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if m.Resource == nil { + m.Resource = &v11.TypedLocalObjectReference{} + } + if err := m.Resource.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex @@ -1238,7 +2017,7 @@ func (m *HTTPIngressPath) Unmarshal(dAtA []byte) error { } return nil } -func (m *HTTPIngressRuleValue) Unmarshal(dAtA []byte) error { +func (m *IngressClass) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -1261,15 +2040,15 @@ func (m *HTTPIngressRuleValue) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: HTTPIngressRuleValue: wiretype end group for non-group") + return fmt.Errorf("proto: IngressClass: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: HTTPIngressRuleValue: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: IngressClass: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Paths", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -1296,8 +2075,40 @@ func (m *HTTPIngressRuleValue) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - m.Paths = append(m.Paths, HTTPIngressPath{}) - if err := m.Paths[len(m.Paths)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex @@ -1325,7 +2136,7 @@ func (m *HTTPIngressRuleValue) Unmarshal(dAtA []byte) error { } return nil } -func (m *Ingress) Unmarshal(dAtA []byte) error { +func (m *IngressClassList) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -1348,15 +2159,15 @@ func (m *Ingress) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: Ingress: wiretype end group for non-group") + return fmt.Errorf("proto: IngressClassList: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: Ingress: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: IngressClassList: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -1383,46 +2194,13 @@ func (m *Ingress) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) - } - var msglen int - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return ErrIntOverflowGenerated - } - if iNdEx >= l { - return io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - msglen |= int(b&0x7F) << shift - if b < 0x80 { - break - } - } - if msglen < 0 { - return ErrInvalidLengthGenerated - } - postIndex := iNdEx + msglen - if postIndex < 0 { - return ErrInvalidLengthGenerated - } - if postIndex > l { - return io.ErrUnexpectedEOF - } - if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { - return err - } - iNdEx = postIndex - case 3: - if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -1449,7 +2227,8 @@ func (m *Ingress) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + m.Items = append(m.Items, IngressClass{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex @@ -1477,7 +2256,7 @@ func (m *Ingress) Unmarshal(dAtA []byte) error { } return nil } -func (m *IngressBackend) Unmarshal(dAtA []byte) error { +func (m *IngressClassSpec) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -1500,15 +2279,15 @@ func (m *IngressBackend) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: IngressBackend: wiretype end group for non-group") + return fmt.Errorf("proto: IngressClassSpec: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: IngressBackend: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: IngressClassSpec: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ServiceName", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Controller", wireType) } var stringLen uint64 for shift := uint(0); ; shift += 7 { @@ -1536,11 +2315,11 @@ func (m *IngressBackend) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - m.ServiceName = string(dAtA[iNdEx:postIndex]) + m.Controller = string(dAtA[iNdEx:postIndex]) iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field ServicePort", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field Parameters", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -1567,7 +2346,10 @@ func (m *IngressBackend) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.ServicePort.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if m.Parameters == nil { + m.Parameters = &v11.TypedLocalObjectReference{} + } + if err := m.Parameters.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex @@ -2055,6 +2837,39 @@ func (m *IngressSpec) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field IngressClassName", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.IngressClassName = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -2285,6 +3100,7 @@ func (m *IngressTLS) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -2316,10 +3132,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -2340,55 +3154,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/networking/v1beta1/generated.proto b/vendor/k8s.io/api/networking/v1beta1/generated.proto index a72545c8..a97c318d 100644 --- a/vendor/k8s.io/api/networking/v1beta1/generated.proto +++ b/vendor/k8s.io/api/networking/v1beta1/generated.proto @@ -30,19 +30,33 @@ import "k8s.io/apimachinery/pkg/util/intstr/generated.proto"; // Package-wide variables from generator "generated". option go_package = "v1beta1"; -// HTTPIngressPath associates a path regex with a backend. Incoming urls matching -// the path are forwarded to the backend. +// HTTPIngressPath associates a path with a backend. Incoming urls matching the +// path are forwarded to the backend. message HTTPIngressPath { - // Path is an extended POSIX regex as defined by IEEE Std 1003.1, - // (i.e this follows the egrep/unix syntax, not the perl syntax) - // matched against the path of an incoming request. Currently it can - // contain characters disallowed from the conventional "path" - // part of a URL as defined by RFC 3986. Paths must begin with - // a '/'. If unspecified, the path defaults to a catch all sending - // traffic to the backend. + // Path is matched against the path of an incoming request. Currently it can + // contain characters disallowed from the conventional "path" part of a URL + // as defined by RFC 3986. Paths must begin with a '/'. When unspecified, + // all paths from incoming requests are matched. // +optional optional string path = 1; + // PathType determines the interpretation of the Path matching. PathType can + // be one of the following values: + // * Exact: Matches the URL path exactly. + // * Prefix: Matches based on a URL path prefix split by '/'. Matching is + // done on a path element by element basis. A path element refers is the + // list of labels in the path split by the '/' separator. A request is a + // match for path p if every p is an element-wise prefix of p of the + // request path. Note that if the last element of the path is a substring + // of the last element in request path, it is not a match (e.g. /foo/bar + // matches /foo/bar/baz, but does not match /foo/barbaz). + // * ImplementationSpecific: Interpretation of the Path matching is up to + // the IngressClass. Implementations can treat this as a separate PathType + // or treat it identically to Prefix or Exact path types. + // Implementations are required to support all path types. + // Defaults to ImplementationSpecific. + optional string pathType = 3; + // Backend defines the referenced service endpoint to which the traffic // will be forwarded to. optional IngressBackend backend = 2; @@ -82,10 +96,62 @@ message Ingress { // IngressBackend describes all endpoints for a given service and port. message IngressBackend { // Specifies the name of the referenced service. + // +optional optional string serviceName = 1; // Specifies the port of the referenced service. + // +optional optional k8s.io.apimachinery.pkg.util.intstr.IntOrString servicePort = 2; + + // Resource is an ObjectRef to another Kubernetes resource in the namespace + // of the Ingress object. If resource is specified, serviceName and servicePort + // must not be specified. + // +optional + optional k8s.io.api.core.v1.TypedLocalObjectReference resource = 3; +} + +// IngressClass represents the class of the Ingress, referenced by the Ingress +// Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be +// used to indicate that an IngressClass should be considered default. When a +// single IngressClass resource has this annotation set to true, new Ingress +// resources without a class specified will be assigned this default class. +message IngressClass { + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // Spec is the desired state of the IngressClass. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + optional IngressClassSpec spec = 2; +} + +// IngressClassList is a collection of IngressClasses. +message IngressClassList { + // Standard list metadata. + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; + + // Items is the list of IngressClasses. + repeated IngressClass items = 2; +} + +// IngressClassSpec provides information about the class of an Ingress. +message IngressClassSpec { + // Controller refers to the name of the controller that should handle this + // class. This allows for different "flavors" that are controlled by the + // same controller. For example, you may have different Parameters for the + // same implementing controller. This should be specified as a + // domain-prefixed path no more than 250 characters in length, e.g. + // "acme.io/ingress-controller". This field is immutable. + optional string controller = 1; + + // Parameters is a link to a custom resource containing additional + // configuration for the controller. This is optional if the controller does + // not require extra parameters. + // +optional + optional k8s.io.api.core.v1.TypedLocalObjectReference parameters = 2; } // IngressList is a collection of Ingress. @@ -103,18 +169,28 @@ message IngressList { // the related backend services. Incoming requests are first evaluated for a host // match, then routed to the backend associated with the matching IngressRuleValue. message IngressRule { - // Host is the fully qualified domain name of a network host, as defined - // by RFC 3986. Note the following deviations from the "host" part of the - // URI as defined in the RFC: - // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the - // IP in the Spec of the parent Ingress. + // Host is the fully qualified domain name of a network host, as defined by RFC 3986. + // Note the following deviations from the "host" part of the + // URI as defined in RFC 3986: + // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to + // the IP in the Spec of the parent Ingress. // 2. The `:` delimiter is not respected because ports are not allowed. // Currently the port of an Ingress is implicitly :80 for http and // :443 for https. // Both these may change in the future. - // Incoming requests are matched against the host before the IngressRuleValue. - // If the host is unspecified, the Ingress routes all traffic based on the - // specified IngressRuleValue. + // Incoming requests are matched against the host before the + // IngressRuleValue. If the host is unspecified, the Ingress routes all + // traffic based on the specified IngressRuleValue. + // + // Host can be "precise" which is a domain name without the terminating dot of + // a network host (e.g. "foo.bar.com") or "wildcard", which is a domain name + // prefixed with a single wildcard label (e.g. "*.foo.com"). + // The wildcard character '*' must appear by itself as the first DNS label and + // matches only a single label. You cannot have a wildcard label by itself (e.g. Host == "*"). + // Requests will be matched against the Host field in the following way: + // 1. If Host is precise, the request matches this rule if the http host header is equal to Host. + // 2. If Host is a wildcard, then the request matches this rule if the http host header + // is to equal to the suffix (removing the first label) of the wildcard rule. // +optional optional string host = 1; @@ -138,6 +214,19 @@ message IngressRuleValue { // IngressSpec describes the Ingress the user wishes to exist. message IngressSpec { + // IngressClassName is the name of the IngressClass cluster resource. The + // associated IngressClass defines which controller will implement the + // resource. This replaces the deprecated `kubernetes.io/ingress.class` + // annotation. For backwards compatibility, when that annotation is set, it + // must be given precedence over this field. The controller may emit a + // warning if the field and annotation have different values. + // Implementations of this API should ignore Ingresses without a class + // specified. An IngressClass resource may be marked as default, which can + // be used to set a default value for this field. For more information, + // refer to the IngressClass documentation. + // +optional + optional string ingressClassName = 4; + // A default backend capable of servicing requests that don't match any // rule. At least one of 'backend' or 'rules' must be specified. This field // is optional to allow the loadbalancer controller or defaulting logic to @@ -175,11 +264,11 @@ message IngressTLS { // +optional repeated string hosts = 1; - // SecretName is the name of the secret used to terminate SSL traffic on 443. - // Field is left optional to allow SSL routing based on SNI hostname alone. - // If the SNI host in a listener conflicts with the "Host" header field used - // by an IngressRule, the SNI host is used for termination and value of the - // Host header is used for routing. + // SecretName is the name of the secret used to terminate TLS traffic on + // port 443. Field is left optional to allow TLS routing based on SNI + // hostname alone. If the SNI host in a listener conflicts with the "Host" + // header field used by an IngressRule, the SNI host is used for termination + // and value of the Host header is used for routing. // +optional optional string secretName = 2; } diff --git a/vendor/k8s.io/api/networking/v1beta1/register.go b/vendor/k8s.io/api/networking/v1beta1/register.go index c046c490..04234953 100644 --- a/vendor/k8s.io/api/networking/v1beta1/register.go +++ b/vendor/k8s.io/api/networking/v1beta1/register.go @@ -49,6 +49,8 @@ func addKnownTypes(scheme *runtime.Scheme) error { scheme.AddKnownTypes(SchemeGroupVersion, &Ingress{}, &IngressList{}, + &IngressClass{}, + &IngressClassList{}, ) // Add the watch version that applies metav1.AddToGroupVersion(scheme, SchemeGroupVersion) diff --git a/vendor/k8s.io/api/networking/v1beta1/types.go b/vendor/k8s.io/api/networking/v1beta1/types.go index 37277bf8..ef9bd4d6 100644 --- a/vendor/k8s.io/api/networking/v1beta1/types.go +++ b/vendor/k8s.io/api/networking/v1beta1/types.go @@ -17,13 +17,16 @@ limitations under the License. package v1beta1 import ( - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/util/intstr" ) // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.14 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,Ingress // Ingress is a collection of rules that allow inbound connections to reach the // endpoints defined by a backend. An Ingress can be configured to give services @@ -48,6 +51,9 @@ type Ingress struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.14 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,IngressList // IngressList is a collection of Ingress. type IngressList struct { @@ -63,6 +69,19 @@ type IngressList struct { // IngressSpec describes the Ingress the user wishes to exist. type IngressSpec struct { + // IngressClassName is the name of the IngressClass cluster resource. The + // associated IngressClass defines which controller will implement the + // resource. This replaces the deprecated `kubernetes.io/ingress.class` + // annotation. For backwards compatibility, when that annotation is set, it + // must be given precedence over this field. The controller may emit a + // warning if the field and annotation have different values. + // Implementations of this API should ignore Ingresses without a class + // specified. An IngressClass resource may be marked as default, which can + // be used to set a default value for this field. For more information, + // refer to the IngressClass documentation. + // +optional + IngressClassName *string `json:"ingressClassName,omitempty" protobuf:"bytes,4,opt,name=ingressClassName"` + // A default backend capable of servicing requests that don't match any // rule. At least one of 'backend' or 'rules' must be specified. This field // is optional to allow the loadbalancer controller or defaulting logic to @@ -93,11 +112,11 @@ type IngressTLS struct { // Ingress, if left unspecified. // +optional Hosts []string `json:"hosts,omitempty" protobuf:"bytes,1,rep,name=hosts"` - // SecretName is the name of the secret used to terminate SSL traffic on 443. - // Field is left optional to allow SSL routing based on SNI hostname alone. - // If the SNI host in a listener conflicts with the "Host" header field used - // by an IngressRule, the SNI host is used for termination and value of the - // Host header is used for routing. + // SecretName is the name of the secret used to terminate TLS traffic on + // port 443. Field is left optional to allow TLS routing based on SNI + // hostname alone. If the SNI host in a listener conflicts with the "Host" + // header field used by an IngressRule, the SNI host is used for termination + // and value of the Host header is used for routing. // +optional SecretName string `json:"secretName,omitempty" protobuf:"bytes,2,opt,name=secretName"` // TODO: Consider specifying different modes of termination, protocols etc. @@ -114,18 +133,28 @@ type IngressStatus struct { // the related backend services. Incoming requests are first evaluated for a host // match, then routed to the backend associated with the matching IngressRuleValue. type IngressRule struct { - // Host is the fully qualified domain name of a network host, as defined - // by RFC 3986. Note the following deviations from the "host" part of the - // URI as defined in the RFC: - // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the - // IP in the Spec of the parent Ingress. + // Host is the fully qualified domain name of a network host, as defined by RFC 3986. + // Note the following deviations from the "host" part of the + // URI as defined in RFC 3986: + // 1. IPs are not allowed. Currently an IngressRuleValue can only apply to + // the IP in the Spec of the parent Ingress. // 2. The `:` delimiter is not respected because ports are not allowed. // Currently the port of an Ingress is implicitly :80 for http and // :443 for https. // Both these may change in the future. - // Incoming requests are matched against the host before the IngressRuleValue. - // If the host is unspecified, the Ingress routes all traffic based on the - // specified IngressRuleValue. + // Incoming requests are matched against the host before the + // IngressRuleValue. If the host is unspecified, the Ingress routes all + // traffic based on the specified IngressRuleValue. + // + // Host can be "precise" which is a domain name without the terminating dot of + // a network host (e.g. "foo.bar.com") or "wildcard", which is a domain name + // prefixed with a single wildcard label (e.g. "*.foo.com"). + // The wildcard character '*' must appear by itself as the first DNS label and + // matches only a single label. You cannot have a wildcard label by itself (e.g. Host == "*"). + // Requests will be matched against the Host field in the following way: + // 1. If Host is precise, the request matches this rule if the http host header is equal to Host. + // 2. If Host is a wildcard, then the request matches this rule if the http host header + // is to equal to the suffix (removing the first label) of the wildcard rule. // +optional Host string `json:"host,omitempty" protobuf:"bytes,1,opt,name=host"` // IngressRuleValue represents a rule to route requests for this IngressRule. @@ -164,19 +193,63 @@ type HTTPIngressRuleValue struct { // options usable by a loadbalancer, like http keep-alive. } -// HTTPIngressPath associates a path regex with a backend. Incoming urls matching -// the path are forwarded to the backend. +// PathType represents the type of path referred to by a HTTPIngressPath. +type PathType string + +const ( + // PathTypeExact matches the URL path exactly and with case sensitivity. + PathTypeExact = PathType("Exact") + + // PathTypePrefix matches based on a URL path prefix split by '/'. Matching + // is case sensitive and done on a path element by element basis. A path + // element refers to the list of labels in the path split by the '/' + // separator. A request is a match for path p if every p is an element-wise + // prefix of p of the request path. Note that if the last element of the + // path is a substring of the last element in request path, it is not a + // match (e.g. /foo/bar matches /foo/bar/baz, but does not match + // /foo/barbaz). If multiple matching paths exist in an Ingress spec, the + // longest matching path is given priority. + // Examples: + // - /foo/bar does not match requests to /foo/barbaz + // - /foo/bar matches request to /foo/bar and /foo/bar/baz + // - /foo and /foo/ both match requests to /foo and /foo/. If both paths are + // present in an Ingress spec, the longest matching path (/foo/) is given + // priority. + PathTypePrefix = PathType("Prefix") + + // PathTypeImplementationSpecific matching is up to the IngressClass. + // Implementations can treat this as a separate PathType or treat it + // identically to Prefix or Exact path types. + PathTypeImplementationSpecific = PathType("ImplementationSpecific") +) + +// HTTPIngressPath associates a path with a backend. Incoming urls matching the +// path are forwarded to the backend. type HTTPIngressPath struct { - // Path is an extended POSIX regex as defined by IEEE Std 1003.1, - // (i.e this follows the egrep/unix syntax, not the perl syntax) - // matched against the path of an incoming request. Currently it can - // contain characters disallowed from the conventional "path" - // part of a URL as defined by RFC 3986. Paths must begin with - // a '/'. If unspecified, the path defaults to a catch all sending - // traffic to the backend. + // Path is matched against the path of an incoming request. Currently it can + // contain characters disallowed from the conventional "path" part of a URL + // as defined by RFC 3986. Paths must begin with a '/'. When unspecified, + // all paths from incoming requests are matched. // +optional Path string `json:"path,omitempty" protobuf:"bytes,1,opt,name=path"` + // PathType determines the interpretation of the Path matching. PathType can + // be one of the following values: + // * Exact: Matches the URL path exactly. + // * Prefix: Matches based on a URL path prefix split by '/'. Matching is + // done on a path element by element basis. A path element refers is the + // list of labels in the path split by the '/' separator. A request is a + // match for path p if every p is an element-wise prefix of p of the + // request path. Note that if the last element of the path is a substring + // of the last element in request path, it is not a match (e.g. /foo/bar + // matches /foo/bar/baz, but does not match /foo/barbaz). + // * ImplementationSpecific: Interpretation of the Path matching is up to + // the IngressClass. Implementations can treat this as a separate PathType + // or treat it identically to Prefix or Exact path types. + // Implementations are required to support all path types. + // Defaults to ImplementationSpecific. + PathType *PathType `json:"pathType,omitempty" protobuf:"bytes,3,opt,name=pathType"` + // Backend defines the referenced service endpoint to which the traffic // will be forwarded to. Backend IngressBackend `json:"backend" protobuf:"bytes,2,opt,name=backend"` @@ -185,8 +258,74 @@ type HTTPIngressPath struct { // IngressBackend describes all endpoints for a given service and port. type IngressBackend struct { // Specifies the name of the referenced service. - ServiceName string `json:"serviceName" protobuf:"bytes,1,opt,name=serviceName"` + // +optional + ServiceName string `json:"serviceName,omitempty" protobuf:"bytes,1,opt,name=serviceName"` // Specifies the port of the referenced service. - ServicePort intstr.IntOrString `json:"servicePort" protobuf:"bytes,2,opt,name=servicePort"` + // +optional + ServicePort intstr.IntOrString `json:"servicePort,omitempty" protobuf:"bytes,2,opt,name=servicePort"` + + // Resource is an ObjectRef to another Kubernetes resource in the namespace + // of the Ingress object. If resource is specified, serviceName and servicePort + // must not be specified. + // +optional + Resource *v1.TypedLocalObjectReference `json:"resource,omitempty" protobuf:"bytes,3,opt,name=resource"` +} + +// +genclient +// +genclient:nonNamespaced +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.18 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,IngressClassList + +// IngressClass represents the class of the Ingress, referenced by the Ingress +// Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be +// used to indicate that an IngressClass should be considered default. When a +// single IngressClass resource has this annotation set to true, new Ingress +// resources without a class specified will be assigned this default class. +type IngressClass struct { + metav1.TypeMeta `json:",inline"` + // Standard object's metadata. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Spec is the desired state of the IngressClass. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status + // +optional + Spec IngressClassSpec `json:"spec,omitempty" protobuf:"bytes,2,opt,name=spec"` +} + +// IngressClassSpec provides information about the class of an Ingress. +type IngressClassSpec struct { + // Controller refers to the name of the controller that should handle this + // class. This allows for different "flavors" that are controlled by the + // same controller. For example, you may have different Parameters for the + // same implementing controller. This should be specified as a + // domain-prefixed path no more than 250 characters in length, e.g. + // "acme.io/ingress-controller". This field is immutable. + Controller string `json:"controller,omitempty" protobuf:"bytes,1,opt,name=controller"` + + // Parameters is a link to a custom resource containing additional + // configuration for the controller. This is optional if the controller does + // not require extra parameters. + // +optional + Parameters *v1.TypedLocalObjectReference `json:"parameters,omitempty" protobuf:"bytes,2,opt,name=parameters"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.18 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=networking.k8s.io,v1,IngressClassList + +// IngressClassList is a collection of IngressClasses. +type IngressClassList struct { + metav1.TypeMeta `json:",inline"` + // Standard list metadata. + // +optional + metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Items is the list of IngressClasses. + Items []IngressClass `json:"items" protobuf:"bytes,2,rep,name=items"` } diff --git a/vendor/k8s.io/api/networking/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/networking/v1beta1/types_swagger_doc_generated.go index 4ae5e32d..c774249d 100644 --- a/vendor/k8s.io/api/networking/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/networking/v1beta1/types_swagger_doc_generated.go @@ -28,9 +28,10 @@ package v1beta1 // AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. var map_HTTPIngressPath = map[string]string{ - "": "HTTPIngressPath associates a path regex with a backend. Incoming urls matching the path are forwarded to the backend.", - "path": "Path is an extended POSIX regex as defined by IEEE Std 1003.1, (i.e this follows the egrep/unix syntax, not the perl syntax) matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional \"path\" part of a URL as defined by RFC 3986. Paths must begin with a '/'. If unspecified, the path defaults to a catch all sending traffic to the backend.", - "backend": "Backend defines the referenced service endpoint to which the traffic will be forwarded to.", + "": "HTTPIngressPath associates a path with a backend. Incoming urls matching the path are forwarded to the backend.", + "path": "Path is matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional \"path\" part of a URL as defined by RFC 3986. Paths must begin with a '/'. When unspecified, all paths from incoming requests are matched.", + "pathType": "PathType determines the interpretation of the Path matching. PathType can be one of the following values: * Exact: Matches the URL path exactly. * Prefix: Matches based on a URL path prefix split by '/'. Matching is\n done on a path element by element basis. A path element refers is the\n list of labels in the path split by the '/' separator. A request is a\n match for path p if every p is an element-wise prefix of p of the\n request path. Note that if the last element of the path is a substring\n of the last element in request path, it is not a match (e.g. /foo/bar\n matches /foo/bar/baz, but does not match /foo/barbaz).\n* ImplementationSpecific: Interpretation of the Path matching is up to\n the IngressClass. Implementations can treat this as a separate PathType\n or treat it identically to Prefix or Exact path types.\nImplementations are required to support all path types. Defaults to ImplementationSpecific.", + "backend": "Backend defines the referenced service endpoint to which the traffic will be forwarded to.", } func (HTTPIngressPath) SwaggerDoc() map[string]string { @@ -61,12 +62,43 @@ var map_IngressBackend = map[string]string{ "": "IngressBackend describes all endpoints for a given service and port.", "serviceName": "Specifies the name of the referenced service.", "servicePort": "Specifies the port of the referenced service.", + "resource": "Resource is an ObjectRef to another Kubernetes resource in the namespace of the Ingress object. If resource is specified, serviceName and servicePort must not be specified.", } func (IngressBackend) SwaggerDoc() map[string]string { return map_IngressBackend } +var map_IngressClass = map[string]string{ + "": "IngressClass represents the class of the Ingress, referenced by the Ingress Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be used to indicate that an IngressClass should be considered default. When a single IngressClass resource has this annotation set to true, new Ingress resources without a class specified will be assigned this default class.", + "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "spec": "Spec is the desired state of the IngressClass. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status", +} + +func (IngressClass) SwaggerDoc() map[string]string { + return map_IngressClass +} + +var map_IngressClassList = map[string]string{ + "": "IngressClassList is a collection of IngressClasses.", + "metadata": "Standard list metadata.", + "items": "Items is the list of IngressClasses.", +} + +func (IngressClassList) SwaggerDoc() map[string]string { + return map_IngressClassList +} + +var map_IngressClassSpec = map[string]string{ + "": "IngressClassSpec provides information about the class of an Ingress.", + "controller": "Controller refers to the name of the controller that should handle this class. This allows for different \"flavors\" that are controlled by the same controller. For example, you may have different Parameters for the same implementing controller. This should be specified as a domain-prefixed path no more than 250 characters in length, e.g. \"acme.io/ingress-controller\". This field is immutable.", + "parameters": "Parameters is a link to a custom resource containing additional configuration for the controller. This is optional if the controller does not require extra parameters.", +} + +func (IngressClassSpec) SwaggerDoc() map[string]string { + return map_IngressClassSpec +} + var map_IngressList = map[string]string{ "": "IngressList is a collection of Ingress.", "metadata": "Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", @@ -79,7 +111,7 @@ func (IngressList) SwaggerDoc() map[string]string { var map_IngressRule = map[string]string{ "": "IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue.", - "host": "Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the \"host\" part of the URI as defined in the RFC: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the\n\t IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.", + "host": "Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the \"host\" part of the URI as defined in RFC 3986: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to\n the IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.\n\nHost can be \"precise\" which is a domain name without the terminating dot of a network host (e.g. \"foo.bar.com\") or \"wildcard\", which is a domain name prefixed with a single wildcard label (e.g. \"*.foo.com\"). The wildcard character '*' must appear by itself as the first DNS label and matches only a single label. You cannot have a wildcard label by itself (e.g. Host == \"*\"). Requests will be matched against the Host field in the following way: 1. If Host is precise, the request matches this rule if the http host header is equal to Host. 2. If Host is a wildcard, then the request matches this rule if the http host header is to equal to the suffix (removing the first label) of the wildcard rule.", } func (IngressRule) SwaggerDoc() map[string]string { @@ -95,10 +127,11 @@ func (IngressRuleValue) SwaggerDoc() map[string]string { } var map_IngressSpec = map[string]string{ - "": "IngressSpec describes the Ingress the user wishes to exist.", - "backend": "A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default.", - "tls": "TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.", - "rules": "A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.", + "": "IngressSpec describes the Ingress the user wishes to exist.", + "ingressClassName": "IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.", + "backend": "A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default.", + "tls": "TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.", + "rules": "A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.", } func (IngressSpec) SwaggerDoc() map[string]string { @@ -117,7 +150,7 @@ func (IngressStatus) SwaggerDoc() map[string]string { var map_IngressTLS = map[string]string{ "": "IngressTLS describes the transport layer security associated with an Ingress.", "hosts": "Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.", - "secretName": "SecretName is the name of the secret used to terminate SSL traffic on 443. Field is left optional to allow SSL routing based on SNI hostname alone. If the SNI host in a listener conflicts with the \"Host\" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing.", + "secretName": "SecretName is the name of the secret used to terminate TLS traffic on port 443. Field is left optional to allow TLS routing based on SNI hostname alone. If the SNI host in a listener conflicts with the \"Host\" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing.", } func (IngressTLS) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/networking/v1beta1/well_known_annotations.go b/vendor/k8s.io/api/networking/v1beta1/well_known_annotations.go new file mode 100644 index 00000000..1629b5d5 --- /dev/null +++ b/vendor/k8s.io/api/networking/v1beta1/well_known_annotations.go @@ -0,0 +1,32 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package v1beta1 + +const ( + // AnnotationIsDefaultIngressClass can be used to indicate that an + // IngressClass should be considered default. When a single IngressClass + // resource has this annotation set to true, new Ingress resources without a + // class specified will be assigned this default class. + AnnotationIsDefaultIngressClass = "ingressclass.kubernetes.io/is-default-class" + + // AnnotationIngressClass indicates the class of an Ingress to be used when + // determining which controller should implement the Ingress. Use of this + // annotation is deprecated. The Ingress class field should be used instead + // of this annotation. + // +deprecated + AnnotationIngressClass = "kubernetes.io/ingress.class" +) diff --git a/vendor/k8s.io/api/networking/v1beta1/zz_generated.deepcopy.go b/vendor/k8s.io/api/networking/v1beta1/zz_generated.deepcopy.go index 16ac936a..d55ccde6 100644 --- a/vendor/k8s.io/api/networking/v1beta1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/networking/v1beta1/zz_generated.deepcopy.go @@ -21,13 +21,19 @@ limitations under the License. package v1beta1 import ( + v1 "k8s.io/api/core/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *HTTPIngressPath) DeepCopyInto(out *HTTPIngressPath) { *out = *in - out.Backend = in.Backend + if in.PathType != nil { + in, out := &in.PathType, &out.PathType + *out = new(PathType) + **out = **in + } + in.Backend.DeepCopyInto(&out.Backend) return } @@ -47,7 +53,9 @@ func (in *HTTPIngressRuleValue) DeepCopyInto(out *HTTPIngressRuleValue) { if in.Paths != nil { in, out := &in.Paths, &out.Paths *out = make([]HTTPIngressPath, len(*in)) - copy(*out, *in) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } } return } @@ -94,6 +102,11 @@ func (in *Ingress) DeepCopyObject() runtime.Object { func (in *IngressBackend) DeepCopyInto(out *IngressBackend) { *out = *in out.ServicePort = in.ServicePort + if in.Resource != nil { + in, out := &in.Resource, &out.Resource + *out = new(v1.TypedLocalObjectReference) + (*in).DeepCopyInto(*out) + } return } @@ -107,6 +120,87 @@ func (in *IngressBackend) DeepCopy() *IngressBackend { return out } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressClass) DeepCopyInto(out *IngressClass) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + in.Spec.DeepCopyInto(&out.Spec) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressClass. +func (in *IngressClass) DeepCopy() *IngressClass { + if in == nil { + return nil + } + out := new(IngressClass) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *IngressClass) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressClassList) DeepCopyInto(out *IngressClassList) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ListMeta.DeepCopyInto(&out.ListMeta) + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]IngressClass, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressClassList. +func (in *IngressClassList) DeepCopy() *IngressClassList { + if in == nil { + return nil + } + out := new(IngressClassList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *IngressClassList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *IngressClassSpec) DeepCopyInto(out *IngressClassSpec) { + *out = *in + if in.Parameters != nil { + in, out := &in.Parameters, &out.Parameters + *out = new(v1.TypedLocalObjectReference) + (*in).DeepCopyInto(*out) + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new IngressClassSpec. +func (in *IngressClassSpec) DeepCopy() *IngressClassSpec { + if in == nil { + return nil + } + out := new(IngressClassSpec) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *IngressList) DeepCopyInto(out *IngressList) { *out = *in @@ -181,10 +275,15 @@ func (in *IngressRuleValue) DeepCopy() *IngressRuleValue { // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *IngressSpec) DeepCopyInto(out *IngressSpec) { *out = *in + if in.IngressClassName != nil { + in, out := &in.IngressClassName, &out.IngressClassName + *out = new(string) + **out = **in + } if in.Backend != nil { in, out := &in.Backend, &out.Backend *out = new(IngressBackend) - **out = **in + (*in).DeepCopyInto(*out) } if in.TLS != nil { in, out := &in.TLS, &out.TLS diff --git a/vendor/k8s.io/api/networking/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/networking/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..5e69fd5d --- /dev/null +++ b/vendor/k8s.io/api/networking/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,121 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Ingress) APILifecycleIntroduced() (major, minor int) { + return 1, 14 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Ingress) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Ingress) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "Ingress"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Ingress) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *IngressClass) APILifecycleIntroduced() (major, minor int) { + return 1, 18 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *IngressClass) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *IngressClass) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "IngressClassList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *IngressClass) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *IngressClassList) APILifecycleIntroduced() (major, minor int) { + return 1, 18 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *IngressClassList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *IngressClassList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "IngressClassList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *IngressClassList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *IngressList) APILifecycleIntroduced() (major, minor int) { + return 1, 14 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *IngressList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *IngressList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "networking.k8s.io", Version: "v1", Kind: "IngressList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *IngressList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/node/v1alpha1/generated.pb.go b/vendor/k8s.io/api/node/v1alpha1/generated.pb.go index 34f4dd6d..e6658a96 100644 --- a/vendor/k8s.io/api/node/v1alpha1/generated.pb.go +++ b/vendor/k8s.io/api/node/v1alpha1/generated.pb.go @@ -46,7 +46,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Overhead) Reset() { *m = Overhead{} } func (*Overhead) ProtoMessage() {} @@ -1500,6 +1500,7 @@ func (m *Scheduling) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1531,10 +1532,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1555,55 +1554,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/node/v1beta1/doc.go b/vendor/k8s.io/api/node/v1beta1/doc.go index e87583ce..c76ba89c 100644 --- a/vendor/k8s.io/api/node/v1beta1/doc.go +++ b/vendor/k8s.io/api/node/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=node.k8s.io diff --git a/vendor/k8s.io/api/node/v1beta1/generated.pb.go b/vendor/k8s.io/api/node/v1beta1/generated.pb.go index 63992f43..b85cbd29 100644 --- a/vendor/k8s.io/api/node/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/node/v1beta1/generated.pb.go @@ -46,7 +46,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Overhead) Reset() { *m = Overhead{} } func (*Overhead) ProtoMessage() {} @@ -1329,6 +1329,7 @@ func (m *Scheduling) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1360,10 +1361,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1384,55 +1383,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/node/v1beta1/types.go b/vendor/k8s.io/api/node/v1beta1/types.go index 793a48f6..1d2b9631 100644 --- a/vendor/k8s.io/api/node/v1beta1/types.go +++ b/vendor/k8s.io/api/node/v1beta1/types.go @@ -24,6 +24,8 @@ import ( // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.13 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // RuntimeClass defines a class of container runtime supported in the cluster. // The RuntimeClass is used to determine which container runtime is used to run @@ -92,6 +94,8 @@ type Scheduling struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.13 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // RuntimeClassList is a list of RuntimeClass objects. type RuntimeClassList struct { diff --git a/vendor/k8s.io/api/node/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/node/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..2fda72ef --- /dev/null +++ b/vendor/k8s.io/api/node/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,57 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *RuntimeClass) APILifecycleIntroduced() (major, minor int) { + return 1, 13 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *RuntimeClass) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *RuntimeClass) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *RuntimeClassList) APILifecycleIntroduced() (major, minor int) { + return 1, 13 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *RuntimeClassList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *RuntimeClassList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} diff --git a/vendor/k8s.io/api/policy/v1beta1/doc.go b/vendor/k8s.io/api/policy/v1beta1/doc.go index 05d8332f..9e9c7d13 100644 --- a/vendor/k8s.io/api/policy/v1beta1/doc.go +++ b/vendor/k8s.io/api/policy/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // Package policy is for any kind of policy object. Suitable examples, even if // they aren't all here, are PodDisruptionBudget, PodSecurityPolicy, diff --git a/vendor/k8s.io/api/policy/v1beta1/generated.pb.go b/vendor/k8s.io/api/policy/v1beta1/generated.pb.go index 5b57f699..40ec7ef7 100644 --- a/vendor/k8s.io/api/policy/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/policy/v1beta1/generated.pb.go @@ -47,7 +47,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *AllowedCSIDriver) Reset() { *m = AllowedCSIDriver{} } func (*AllowedCSIDriver) ProtoMessage() {} @@ -609,125 +609,125 @@ func init() { } var fileDescriptor_014060e454a820dc = []byte{ - // 1883 bytes of a gzipped FileDescriptorProto + // 1878 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x58, 0xdd, 0x6e, 0x1b, 0xc7, - 0x15, 0xd6, 0x9a, 0xfa, 0xa1, 0x46, 0x3f, 0x16, 0x47, 0x3f, 0x5e, 0x2b, 0x35, 0xd7, 0xd9, 0x00, + 0x15, 0xd6, 0x9a, 0xfa, 0xa1, 0x46, 0x3f, 0x16, 0x47, 0x3f, 0x5e, 0x2b, 0x0d, 0xd7, 0xd9, 0x00, 0x85, 0x9b, 0x26, 0xcb, 0x58, 0x76, 0x5c, 0xa3, 0x69, 0x8b, 0x68, 0x45, 0xc9, 0x56, 0x60, 0x59, 0xec, 0xd0, 0x0e, 0xda, 0xc2, 0x2d, 0x3a, 0xe4, 0x8e, 0xa8, 0x8d, 0x96, 0xbb, 0xdb, 0x99, 0x59, 0x46, 0xbc, 0xeb, 0x45, 0x2f, 0x7a, 0xd9, 0x17, 0x08, 0xfa, 0x00, 0x45, 0xaf, 0xfa, 0x12, 0x0e, - 0x50, 0x04, 0xb9, 0x0c, 0x7a, 0x41, 0xd4, 0x2c, 0xfa, 0x12, 0xbe, 0x0a, 0x76, 0x38, 0xbb, 0xe4, - 0xfe, 0x91, 0x76, 0x00, 0xfb, 0x8e, 0x3b, 0xe7, 0xfb, 0xbe, 0x33, 0x73, 0xe6, 0xcc, 0x99, 0xc3, - 0x01, 0xe6, 0xc5, 0x7d, 0x66, 0xd8, 0x5e, 0xed, 0x22, 0x68, 0x11, 0xea, 0x12, 0x4e, 0x58, 0xad, - 0x47, 0x5c, 0xcb, 0xa3, 0x35, 0x69, 0xc0, 0xbe, 0x5d, 0xf3, 0x3d, 0xc7, 0x6e, 0xf7, 0x6b, 0xbd, - 0xdb, 0x2d, 0xc2, 0xf1, 0xed, 0x5a, 0x87, 0xb8, 0x84, 0x62, 0x4e, 0x2c, 0xc3, 0xa7, 0x1e, 0xf7, - 0xe0, 0xf5, 0x11, 0xd4, 0xc0, 0xbe, 0x6d, 0x8c, 0xa0, 0x86, 0x84, 0xee, 0x7e, 0xd8, 0xb1, 0xf9, - 0x79, 0xd0, 0x32, 0xda, 0x5e, 0xb7, 0xd6, 0xf1, 0x3a, 0x5e, 0x4d, 0x30, 0x5a, 0xc1, 0x99, 0xf8, - 0x12, 0x1f, 0xe2, 0xd7, 0x48, 0x69, 0x57, 0x9f, 0x70, 0xda, 0xf6, 0x28, 0xa9, 0xf5, 0x32, 0xde, - 0x76, 0xef, 0x8e, 0x31, 0x5d, 0xdc, 0x3e, 0xb7, 0x5d, 0x42, 0xfb, 0x35, 0xff, 0xa2, 0x13, 0x0e, - 0xb0, 0x5a, 0x97, 0x70, 0x9c, 0xc7, 0xaa, 0x15, 0xb1, 0x68, 0xe0, 0x72, 0xbb, 0x4b, 0x32, 0x84, - 0x7b, 0xb3, 0x08, 0xac, 0x7d, 0x4e, 0xba, 0x38, 0xc3, 0xbb, 0x53, 0xc4, 0x0b, 0xb8, 0xed, 0xd4, - 0x6c, 0x97, 0x33, 0x4e, 0xd3, 0x24, 0xfd, 0x2e, 0xd8, 0xd8, 0x77, 0x1c, 0xef, 0x4b, 0x62, 0x1d, - 0x34, 0x8f, 0xeb, 0xd4, 0xee, 0x11, 0x0a, 0x6f, 0x82, 0x79, 0x17, 0x77, 0x89, 0xaa, 0xdc, 0x54, - 0x6e, 0x2d, 0x9b, 0xab, 0xcf, 0x07, 0xda, 0xdc, 0x70, 0xa0, 0xcd, 0x3f, 0xc6, 0x5d, 0x82, 0x84, - 0x45, 0xff, 0x04, 0x54, 0x24, 0xeb, 0xc8, 0x21, 0x97, 0x9f, 0x7b, 0x4e, 0xd0, 0x25, 0xf0, 0xc7, - 0x60, 0xd1, 0x12, 0x02, 0x92, 0xb8, 0x2e, 0x89, 0x8b, 0x23, 0x59, 0x24, 0xad, 0x3a, 0x03, 0x57, - 0x25, 0xf9, 0xa1, 0xc7, 0x78, 0x03, 0xf3, 0x73, 0xb8, 0x07, 0x80, 0x8f, 0xf9, 0x79, 0x83, 0x92, - 0x33, 0xfb, 0x52, 0xd2, 0xa1, 0xa4, 0x83, 0x46, 0x6c, 0x41, 0x13, 0x28, 0xf8, 0x01, 0x28, 0x53, - 0x82, 0xad, 0x53, 0xd7, 0xe9, 0xab, 0x57, 0x6e, 0x2a, 0xb7, 0xca, 0xe6, 0x86, 0x64, 0x94, 0x91, - 0x1c, 0x47, 0x31, 0x42, 0xff, 0x8f, 0x02, 0xca, 0x87, 0x3d, 0xbb, 0xcd, 0x6d, 0xcf, 0x85, 0x7f, - 0x04, 0xe5, 0x70, 0xb7, 0x2c, 0xcc, 0xb1, 0x70, 0xb6, 0xb2, 0xf7, 0x91, 0x31, 0xce, 0xa4, 0x38, - 0x78, 0x86, 0x7f, 0xd1, 0x09, 0x07, 0x98, 0x11, 0xa2, 0x8d, 0xde, 0x6d, 0xe3, 0xb4, 0xf5, 0x05, - 0x69, 0xf3, 0x13, 0xc2, 0xf1, 0x78, 0x7a, 0xe3, 0x31, 0x14, 0xab, 0x42, 0x07, 0xac, 0x59, 0xc4, - 0x21, 0x9c, 0x9c, 0xfa, 0xa1, 0x47, 0x26, 0x66, 0xb8, 0xb2, 0x77, 0xe7, 0xd5, 0xdc, 0xd4, 0x27, - 0xa9, 0x66, 0x65, 0x38, 0xd0, 0xd6, 0x12, 0x43, 0x28, 0x29, 0xae, 0x7f, 0xa5, 0x80, 0x9d, 0xa3, - 0xe6, 0x03, 0xea, 0x05, 0x7e, 0x93, 0x87, 0xbb, 0xdb, 0xe9, 0x4b, 0x13, 0xfc, 0x19, 0x98, 0xa7, - 0x81, 0x13, 0xed, 0xe5, 0x7b, 0xd1, 0x5e, 0xa2, 0xc0, 0x21, 0x2f, 0x07, 0xda, 0x66, 0x8a, 0xf5, - 0xa4, 0xef, 0x13, 0x24, 0x08, 0xf0, 0x33, 0xb0, 0x48, 0xb1, 0xdb, 0x21, 0xe1, 0xd4, 0x4b, 0xb7, - 0x56, 0xf6, 0x74, 0xa3, 0xf0, 0xac, 0x19, 0xc7, 0x75, 0x14, 0x42, 0xc7, 0x3b, 0x2e, 0x3e, 0x19, - 0x92, 0x0a, 0xfa, 0x09, 0x58, 0x13, 0x5b, 0xed, 0x51, 0x2e, 0x2c, 0xf0, 0x06, 0x28, 0x75, 0x6d, - 0x57, 0x4c, 0x6a, 0xc1, 0x5c, 0x91, 0xac, 0xd2, 0x89, 0xed, 0xa2, 0x70, 0x5c, 0x98, 0xf1, 0xa5, - 0x88, 0xd9, 0xa4, 0x19, 0x5f, 0xa2, 0x70, 0x5c, 0x7f, 0x00, 0x96, 0xa4, 0xc7, 0x49, 0xa1, 0xd2, - 0x74, 0xa1, 0x52, 0x8e, 0xd0, 0x3f, 0xae, 0x80, 0xcd, 0x86, 0x67, 0xd5, 0x6d, 0x46, 0x03, 0x11, - 0x2f, 0x33, 0xb0, 0x3a, 0x84, 0xbf, 0x85, 0xfc, 0x78, 0x02, 0xe6, 0x99, 0x4f, 0xda, 0x32, 0x2d, - 0xf6, 0xa6, 0xc4, 0x36, 0x67, 0x7e, 0x4d, 0x9f, 0xb4, 0xc7, 0xc7, 0x32, 0xfc, 0x42, 0x42, 0x0d, - 0x3e, 0x03, 0x8b, 0x8c, 0x63, 0x1e, 0x30, 0xb5, 0x24, 0x74, 0xef, 0xbe, 0xa6, 0xae, 0xe0, 0x8e, - 0x77, 0x71, 0xf4, 0x8d, 0xa4, 0xa6, 0xfe, 0x6f, 0x05, 0x5c, 0xcb, 0x61, 0x3d, 0xb2, 0x19, 0x87, - 0xcf, 0x32, 0x11, 0x33, 0x5e, 0x2d, 0x62, 0x21, 0x5b, 0xc4, 0x2b, 0x3e, 0xbc, 0xd1, 0xc8, 0x44, - 0xb4, 0x9a, 0x60, 0xc1, 0xe6, 0xa4, 0x1b, 0xa5, 0xa2, 0xf1, 0x7a, 0xcb, 0x32, 0xd7, 0xa4, 0xf4, - 0xc2, 0x71, 0x28, 0x82, 0x46, 0x5a, 0xfa, 0x37, 0x57, 0x72, 0x97, 0x13, 0x86, 0x13, 0x9e, 0x81, - 0xd5, 0xae, 0xed, 0xee, 0xf7, 0xb0, 0xed, 0xe0, 0x96, 0x3c, 0x3d, 0xd3, 0x92, 0x20, 0xac, 0xb0, - 0xc6, 0xa8, 0xc2, 0x1a, 0xc7, 0x2e, 0x3f, 0xa5, 0x4d, 0x4e, 0x6d, 0xb7, 0x63, 0x6e, 0x0c, 0x07, - 0xda, 0xea, 0xc9, 0x84, 0x12, 0x4a, 0xe8, 0xc2, 0xdf, 0x83, 0x32, 0x23, 0x0e, 0x69, 0x73, 0x8f, - 0xbe, 0x5e, 0x85, 0x78, 0x84, 0x5b, 0xc4, 0x69, 0x4a, 0xaa, 0xb9, 0x1a, 0xc6, 0x2d, 0xfa, 0x42, - 0xb1, 0x24, 0x74, 0xc0, 0x7a, 0x17, 0x5f, 0x3e, 0x75, 0x71, 0xbc, 0x90, 0xd2, 0x0f, 0x5c, 0x08, - 0x1c, 0x0e, 0xb4, 0xf5, 0x93, 0x84, 0x16, 0x4a, 0x69, 0xeb, 0xff, 0x9f, 0x07, 0xd7, 0x0b, 0xb3, - 0x0a, 0x7e, 0x06, 0xa0, 0xd7, 0x62, 0x84, 0xf6, 0x88, 0xf5, 0x60, 0x74, 0x07, 0xd9, 0x5e, 0x74, - 0x70, 0x77, 0xe5, 0x06, 0xc1, 0xd3, 0x0c, 0x02, 0xe5, 0xb0, 0xe0, 0x5f, 0x14, 0xb0, 0x66, 0x8d, - 0xdc, 0x10, 0xab, 0xe1, 0x59, 0x51, 0x62, 0x3c, 0xf8, 0x21, 0xf9, 0x6e, 0xd4, 0x27, 0x95, 0x0e, - 0x5d, 0x4e, 0xfb, 0xe6, 0xb6, 0x9c, 0xd0, 0x5a, 0xc2, 0x86, 0x92, 0x4e, 0xe1, 0x09, 0x80, 0x56, - 0x2c, 0xc9, 0xe4, 0x9d, 0x26, 0x42, 0xbc, 0x60, 0xde, 0x90, 0x0a, 0xdb, 0x09, 0xbf, 0x11, 0x08, - 0xe5, 0x10, 0xe1, 0xaf, 0xc0, 0x7a, 0x3b, 0xa0, 0x94, 0xb8, 0xfc, 0x21, 0xc1, 0x0e, 0x3f, 0xef, - 0xab, 0xf3, 0x42, 0x6a, 0x47, 0x4a, 0xad, 0x1f, 0x24, 0xac, 0x28, 0x85, 0x0e, 0xf9, 0x16, 0x61, - 0x36, 0x25, 0x56, 0xc4, 0x5f, 0x48, 0xf2, 0xeb, 0x09, 0x2b, 0x4a, 0xa1, 0xe1, 0x7d, 0xb0, 0x4a, - 0x2e, 0x7d, 0xd2, 0x8e, 0x62, 0xba, 0x28, 0xd8, 0x5b, 0x92, 0xbd, 0x7a, 0x38, 0x61, 0x43, 0x09, - 0xe4, 0xae, 0x03, 0x60, 0x36, 0x88, 0x70, 0x03, 0x94, 0x2e, 0x48, 0x7f, 0x74, 0xf3, 0xa0, 0xf0, - 0x27, 0xfc, 0x14, 0x2c, 0xf4, 0xb0, 0x13, 0x10, 0x99, 0xeb, 0xef, 0xbf, 0x5a, 0xae, 0x3f, 0xb1, - 0xbb, 0x04, 0x8d, 0x88, 0x3f, 0xbf, 0x72, 0x5f, 0xd1, 0xbf, 0x56, 0x40, 0xa5, 0xe1, 0x59, 0x4d, - 0xd2, 0x0e, 0xa8, 0xcd, 0xfb, 0x0d, 0xb1, 0xcf, 0x6f, 0xa1, 0x66, 0xa3, 0x44, 0xcd, 0xfe, 0x68, - 0x7a, 0xae, 0x25, 0x67, 0x57, 0x54, 0xb1, 0xf5, 0xe7, 0x0a, 0xd8, 0xce, 0xa0, 0xdf, 0x42, 0x45, - 0xfd, 0x75, 0xb2, 0xa2, 0x7e, 0xf0, 0x3a, 0x8b, 0x29, 0xa8, 0xa7, 0x5f, 0x57, 0x72, 0x96, 0x22, - 0xaa, 0x69, 0xd8, 0xdd, 0x51, 0xbb, 0x67, 0x3b, 0xa4, 0x43, 0x2c, 0xb1, 0x98, 0xf2, 0x44, 0x77, - 0x17, 0x5b, 0xd0, 0x04, 0x0a, 0x32, 0xb0, 0x63, 0x91, 0x33, 0x1c, 0x38, 0x7c, 0xdf, 0xb2, 0x0e, - 0xb0, 0x8f, 0x5b, 0xb6, 0x63, 0x73, 0x5b, 0xb6, 0x23, 0xcb, 0xe6, 0x27, 0xc3, 0x81, 0xb6, 0x53, - 0xcf, 0x45, 0xbc, 0x1c, 0x68, 0x37, 0xb2, 0xdd, 0xbc, 0x11, 0x43, 0xfa, 0xa8, 0x40, 0x1a, 0xf6, - 0x81, 0x4a, 0xc9, 0x9f, 0x82, 0xf0, 0x50, 0xd4, 0xa9, 0xe7, 0x27, 0xdc, 0x96, 0x84, 0xdb, 0x5f, - 0x0e, 0x07, 0x9a, 0x8a, 0x0a, 0x30, 0xb3, 0x1d, 0x17, 0xca, 0xc3, 0x2f, 0xc0, 0x26, 0x96, 0x7d, - 0xf8, 0xa4, 0xd7, 0x79, 0xe1, 0xf5, 0xfe, 0x70, 0xa0, 0x6d, 0xee, 0x67, 0xcd, 0xb3, 0x1d, 0xe6, - 0x89, 0xc2, 0x1a, 0x58, 0xea, 0x89, 0x96, 0x9d, 0xa9, 0x0b, 0x42, 0x7f, 0x7b, 0x38, 0xd0, 0x96, - 0x46, 0x5d, 0x7c, 0xa8, 0xb9, 0x78, 0xd4, 0x14, 0x8d, 0x60, 0x84, 0x82, 0x1f, 0x83, 0x95, 0x73, - 0x8f, 0xf1, 0xc7, 0x84, 0x7f, 0xe9, 0xd1, 0x0b, 0x51, 0x18, 0xca, 0xe6, 0xa6, 0xdc, 0xc1, 0x95, - 0x87, 0x63, 0x13, 0x9a, 0xc4, 0xc1, 0xdf, 0x82, 0xe5, 0x73, 0xd9, 0xf6, 0x31, 0x75, 0x49, 0x24, - 0xda, 0xad, 0x29, 0x89, 0x96, 0x68, 0x11, 0xcd, 0x8a, 0x94, 0x5f, 0x8e, 0x86, 0x19, 0x1a, 0xab, - 0xc1, 0x9f, 0x80, 0x25, 0xf1, 0x71, 0x5c, 0x57, 0xcb, 0x62, 0x36, 0x57, 0x25, 0x7c, 0xe9, 0xe1, - 0x68, 0x18, 0x45, 0xf6, 0x08, 0x7a, 0xdc, 0x38, 0x50, 0x97, 0xb3, 0xd0, 0xe3, 0xc6, 0x01, 0x8a, - 0xec, 0xf0, 0x19, 0x58, 0x62, 0xe4, 0x91, 0xed, 0x06, 0x97, 0x2a, 0x10, 0x47, 0xee, 0xf6, 0x94, - 0xe9, 0x36, 0x0f, 0x05, 0x32, 0xd5, 0x70, 0x8f, 0xd5, 0xa5, 0x1d, 0x45, 0x92, 0xd0, 0x02, 0xcb, - 0x34, 0x70, 0xf7, 0xd9, 0x53, 0x46, 0xa8, 0xba, 0x92, 0xb9, 0xed, 0xd3, 0xfa, 0x28, 0xc2, 0xa6, - 0x3d, 0xc4, 0x91, 0x89, 0x11, 0x68, 0x2c, 0x0c, 0x2d, 0x00, 0xc4, 0x87, 0xe8, 0xeb, 0xd5, 0x9d, - 0x99, 0x7d, 0x20, 0x8a, 0xc1, 0x69, 0x3f, 0xeb, 0xe1, 0xf1, 0x1c, 0x9b, 0xd1, 0x84, 0x2e, 0xfc, - 0xab, 0x02, 0x20, 0x0b, 0x7c, 0xdf, 0x21, 0x5d, 0xe2, 0x72, 0xec, 0x88, 0x51, 0xa6, 0xae, 0x0a, - 0x77, 0xbf, 0x98, 0x16, 0xb5, 0x0c, 0x29, 0xed, 0x36, 0x6e, 0x06, 0xb2, 0x50, 0x94, 0xe3, 0x33, - 0xdc, 0xb4, 0x33, 0xb9, 0xda, 0xb5, 0x99, 0x9b, 0x96, 0xff, 0x2f, 0x69, 0xbc, 0x69, 0xd2, 0x8e, - 0x22, 0x49, 0xf8, 0x39, 0xd8, 0x89, 0xfe, 0x43, 0x22, 0xcf, 0xe3, 0x47, 0xb6, 0x43, 0x58, 0x9f, - 0x71, 0xd2, 0x55, 0xd7, 0x45, 0x32, 0x55, 0x25, 0x73, 0x07, 0xe5, 0xa2, 0x50, 0x01, 0x1b, 0x76, - 0x81, 0x16, 0x15, 0xa1, 0xf0, 0x84, 0xc6, 0x55, 0xf0, 0x90, 0xb5, 0xb1, 0x33, 0xea, 0x8d, 0xae, - 0x0a, 0x07, 0xef, 0x0d, 0x07, 0x9a, 0x56, 0x9f, 0x0e, 0x45, 0xb3, 0xb4, 0xe0, 0x6f, 0x80, 0x8a, - 0x8b, 0xfc, 0x6c, 0x08, 0x3f, 0x3f, 0x0a, 0x2b, 0x5b, 0xa1, 0x83, 0x42, 0x36, 0xf4, 0xc1, 0x06, - 0x4e, 0xfe, 0x9b, 0x67, 0x6a, 0x45, 0x9c, 0xf5, 0xf7, 0xa7, 0xec, 0x43, 0xea, 0x01, 0xc0, 0x54, - 0x65, 0x18, 0x37, 0x52, 0x06, 0x86, 0x32, 0xea, 0xf0, 0x12, 0x40, 0x9c, 0x7e, 0x7c, 0x60, 0x2a, - 0x9c, 0x79, 0x91, 0x65, 0x5e, 0x2c, 0xc6, 0xa9, 0x96, 0x31, 0x31, 0x94, 0xe3, 0x03, 0x72, 0x50, - 0xc1, 0xa9, 0xc7, 0x12, 0xa6, 0x5e, 0x13, 0x8e, 0x7f, 0x3a, 0xdb, 0x71, 0xcc, 0x31, 0xaf, 0x4b, - 0xbf, 0x95, 0xb4, 0x85, 0xa1, 0xac, 0x03, 0xf8, 0x08, 0x6c, 0xc9, 0xc1, 0xa7, 0x2e, 0xc3, 0x67, - 0xa4, 0xd9, 0x67, 0x6d, 0xee, 0x30, 0x75, 0x53, 0xd4, 0x6e, 0x75, 0x38, 0xd0, 0xb6, 0xf6, 0x73, - 0xec, 0x28, 0x97, 0x05, 0x3f, 0x05, 0x1b, 0x67, 0x1e, 0x6d, 0xd9, 0x96, 0x45, 0xdc, 0x48, 0x69, - 0x4b, 0x28, 0x6d, 0x85, 0xf1, 0x3f, 0x4a, 0xd9, 0x50, 0x06, 0x0d, 0x19, 0xd8, 0x96, 0xca, 0x0d, - 0xea, 0xb5, 0x4f, 0xbc, 0xc0, 0xe5, 0xe1, 0x75, 0xc1, 0xd4, 0xed, 0xf8, 0x8a, 0xdc, 0xde, 0xcf, - 0x03, 0xbc, 0x1c, 0x68, 0x37, 0x73, 0xae, 0xab, 0x04, 0x08, 0xe5, 0x6b, 0x43, 0x07, 0xac, 0xca, - 0xe7, 0xaf, 0x03, 0x07, 0x33, 0xa6, 0xaa, 0xe2, 0xa8, 0xdf, 0x9b, 0x5e, 0xd8, 0x62, 0x78, 0xfa, - 0xbc, 0x8b, 0xff, 0x65, 0x93, 0x00, 0x94, 0x50, 0xd7, 0xff, 0xae, 0x80, 0xeb, 0x85, 0x85, 0x11, - 0xde, 0x4b, 0xbc, 0xa9, 0xe8, 0xa9, 0x37, 0x15, 0x98, 0x25, 0xbe, 0x81, 0x27, 0x95, 0xaf, 0x14, - 0xa0, 0x16, 0xdd, 0x10, 0xf0, 0xe3, 0xc4, 0x04, 0xdf, 0x4d, 0x4d, 0xb0, 0x92, 0xe1, 0xbd, 0x81, - 0xf9, 0x7d, 0xa3, 0x80, 0x77, 0xa6, 0xec, 0x40, 0x5c, 0x90, 0x88, 0x35, 0x89, 0x7a, 0x8c, 0xc3, - 0xa3, 0xac, 0x88, 0x3c, 0x1a, 0x17, 0xa4, 0x1c, 0x0c, 0x2a, 0x64, 0xc3, 0xa7, 0xe0, 0x9a, 0xac, - 0x86, 0x69, 0x9b, 0xe8, 0xdc, 0x97, 0xcd, 0x77, 0x86, 0x03, 0xed, 0x5a, 0x3d, 0x1f, 0x82, 0x8a, - 0xb8, 0xfa, 0x3f, 0x15, 0xb0, 0x93, 0x7f, 0xe5, 0xc3, 0x3b, 0x89, 0x70, 0x6b, 0xa9, 0x70, 0x5f, - 0x4d, 0xb1, 0x64, 0xb0, 0xff, 0x00, 0xd6, 0x65, 0x63, 0x90, 0x7c, 0x22, 0x4c, 0x04, 0x3d, 0x3c, - 0x22, 0x61, 0x4f, 0x2f, 0x25, 0xa2, 0xf4, 0x15, 0xff, 0xc6, 0x93, 0x63, 0x28, 0xa5, 0xa6, 0xff, - 0x4b, 0x01, 0xef, 0xce, 0xbc, 0x6c, 0xa1, 0x99, 0x98, 0xba, 0x91, 0x9a, 0x7a, 0xb5, 0x58, 0xe0, - 0xcd, 0xbc, 0x14, 0x9a, 0x1f, 0x3e, 0x7f, 0x51, 0x9d, 0xfb, 0xf6, 0x45, 0x75, 0xee, 0xbb, 0x17, - 0xd5, 0xb9, 0x3f, 0x0f, 0xab, 0xca, 0xf3, 0x61, 0x55, 0xf9, 0x76, 0x58, 0x55, 0xbe, 0x1b, 0x56, - 0x95, 0xff, 0x0e, 0xab, 0xca, 0xdf, 0xfe, 0x57, 0x9d, 0xfb, 0xdd, 0x92, 0x94, 0xfb, 0x3e, 0x00, - 0x00, 0xff, 0xff, 0x48, 0x23, 0x7b, 0x0e, 0x44, 0x18, 0x00, 0x00, + 0x50, 0x04, 0xb9, 0x0c, 0x7a, 0x41, 0xd4, 0xec, 0x5b, 0xf8, 0xaa, 0xd8, 0xe1, 0xec, 0x92, 0xfb, + 0x47, 0x5a, 0x01, 0xec, 0x3b, 0xee, 0x9c, 0xef, 0xfb, 0xce, 0xcc, 0x99, 0x33, 0x67, 0x0e, 0x07, + 0x98, 0x17, 0x0f, 0x98, 0x61, 0x7b, 0xb5, 0x8b, 0xa0, 0x45, 0xa8, 0x4b, 0x38, 0x61, 0xb5, 0x1e, + 0x71, 0x2d, 0x8f, 0xd6, 0xa4, 0x01, 0xfb, 0x76, 0xcd, 0xf7, 0x1c, 0xbb, 0xdd, 0xaf, 0xf5, 0xee, + 0xb4, 0x08, 0xc7, 0x77, 0x6a, 0x1d, 0xe2, 0x12, 0x8a, 0x39, 0xb1, 0x0c, 0x9f, 0x7a, 0xdc, 0x83, + 0x37, 0x47, 0x50, 0x03, 0xfb, 0xb6, 0x31, 0x82, 0x1a, 0x12, 0xba, 0xfb, 0x51, 0xc7, 0xe6, 0xe7, + 0x41, 0xcb, 0x68, 0x7b, 0xdd, 0x5a, 0xc7, 0xeb, 0x78, 0x35, 0xc1, 0x68, 0x05, 0x67, 0xe2, 0x4b, + 0x7c, 0x88, 0x5f, 0x23, 0xa5, 0x5d, 0x7d, 0xc2, 0x69, 0xdb, 0xa3, 0xa4, 0xd6, 0xcb, 0x78, 0xdb, + 0xbd, 0x37, 0xc6, 0x74, 0x71, 0xfb, 0xdc, 0x76, 0x09, 0xed, 0xd7, 0xfc, 0x8b, 0x4e, 0x38, 0xc0, + 0x6a, 0x5d, 0xc2, 0x71, 0x1e, 0xab, 0x56, 0xc4, 0xa2, 0x81, 0xcb, 0xed, 0x2e, 0xc9, 0x10, 0xee, + 0xcf, 0x22, 0xb0, 0xf6, 0x39, 0xe9, 0xe2, 0x0c, 0xef, 0x6e, 0x11, 0x2f, 0xe0, 0xb6, 0x53, 0xb3, + 0x5d, 0xce, 0x38, 0x4d, 0x93, 0xf4, 0x7b, 0x60, 0x63, 0xdf, 0x71, 0xbc, 0xaf, 0x88, 0x75, 0xd0, + 0x3c, 0xae, 0x53, 0xbb, 0x47, 0x28, 0xbc, 0x05, 0xe6, 0x5d, 0xdc, 0x25, 0xaa, 0x72, 0x4b, 0xb9, + 0xbd, 0x6c, 0xae, 0xbe, 0x18, 0x68, 0x73, 0xc3, 0x81, 0x36, 0xff, 0x04, 0x77, 0x09, 0x12, 0x16, + 0xfd, 0x53, 0x50, 0x91, 0xac, 0x23, 0x87, 0x5c, 0x7e, 0xe1, 0x39, 0x41, 0x97, 0xc0, 0x1f, 0x83, + 0x45, 0x4b, 0x08, 0x48, 0xe2, 0xba, 0x24, 0x2e, 0x8e, 0x64, 0x91, 0xb4, 0xea, 0x0c, 0x5c, 0x97, + 0xe4, 0x47, 0x1e, 0xe3, 0x0d, 0xcc, 0xcf, 0xe1, 0x1e, 0x00, 0x3e, 0xe6, 0xe7, 0x0d, 0x4a, 0xce, + 0xec, 0x4b, 0x49, 0x87, 0x92, 0x0e, 0x1a, 0xb1, 0x05, 0x4d, 0xa0, 0xe0, 0x87, 0xa0, 0x4c, 0x09, + 0xb6, 0x4e, 0x5d, 0xa7, 0xaf, 0x5e, 0xbb, 0xa5, 0xdc, 0x2e, 0x9b, 0x1b, 0x92, 0x51, 0x46, 0x72, + 0x1c, 0xc5, 0x08, 0xfd, 0x3f, 0x0a, 0x28, 0x1f, 0xf6, 0xec, 0x36, 0xb7, 0x3d, 0x17, 0xfe, 0x11, + 0x94, 0xc3, 0xdd, 0xb2, 0x30, 0xc7, 0xc2, 0xd9, 0xca, 0xde, 0xc7, 0xc6, 0x38, 0x93, 0xe2, 0xe0, + 0x19, 0xfe, 0x45, 0x27, 0x1c, 0x60, 0x46, 0x88, 0x36, 0x7a, 0x77, 0x8c, 0xd3, 0xd6, 0x97, 0xa4, + 0xcd, 0x4f, 0x08, 0xc7, 0xe3, 0xe9, 0x8d, 0xc7, 0x50, 0xac, 0x0a, 0x1d, 0xb0, 0x66, 0x11, 0x87, + 0x70, 0x72, 0xea, 0x87, 0x1e, 0x99, 0x98, 0xe1, 0xca, 0xde, 0xdd, 0xd7, 0x73, 0x53, 0x9f, 0xa4, + 0x9a, 0x95, 0xe1, 0x40, 0x5b, 0x4b, 0x0c, 0xa1, 0xa4, 0xb8, 0xfe, 0xb5, 0x02, 0x76, 0x8e, 0x9a, + 0x0f, 0xa9, 0x17, 0xf8, 0x4d, 0x1e, 0xee, 0x6e, 0xa7, 0x2f, 0x4d, 0xf0, 0x67, 0x60, 0x9e, 0x06, + 0x4e, 0xb4, 0x97, 0xef, 0x47, 0x7b, 0x89, 0x02, 0x87, 0xbc, 0x1a, 0x68, 0x9b, 0x29, 0xd6, 0xd3, + 0xbe, 0x4f, 0x90, 0x20, 0xc0, 0xcf, 0xc1, 0x22, 0xc5, 0x6e, 0x87, 0x84, 0x53, 0x2f, 0xdd, 0x5e, + 0xd9, 0xd3, 0x8d, 0xc2, 0xb3, 0x66, 0x1c, 0xd7, 0x51, 0x08, 0x1d, 0xef, 0xb8, 0xf8, 0x64, 0x48, + 0x2a, 0xe8, 0x27, 0x60, 0x4d, 0x6c, 0xb5, 0x47, 0xb9, 0xb0, 0xc0, 0x77, 0x41, 0xa9, 0x6b, 0xbb, + 0x62, 0x52, 0x0b, 0xe6, 0x8a, 0x64, 0x95, 0x4e, 0x6c, 0x17, 0x85, 0xe3, 0xc2, 0x8c, 0x2f, 0x45, + 0xcc, 0x26, 0xcd, 0xf8, 0x12, 0x85, 0xe3, 0xfa, 0x43, 0xb0, 0x24, 0x3d, 0x4e, 0x0a, 0x95, 0xa6, + 0x0b, 0x95, 0x72, 0x84, 0xfe, 0x71, 0x0d, 0x6c, 0x36, 0x3c, 0xab, 0x6e, 0x33, 0x1a, 0x88, 0x78, + 0x99, 0x81, 0xd5, 0x21, 0xfc, 0x2d, 0xe4, 0xc7, 0x53, 0x30, 0xcf, 0x7c, 0xd2, 0x96, 0x69, 0xb1, + 0x37, 0x25, 0xb6, 0x39, 0xf3, 0x6b, 0xfa, 0xa4, 0x3d, 0x3e, 0x96, 0xe1, 0x17, 0x12, 0x6a, 0xf0, + 0x39, 0x58, 0x64, 0x1c, 0xf3, 0x80, 0xa9, 0x25, 0xa1, 0x7b, 0xef, 0x8a, 0xba, 0x82, 0x3b, 0xde, + 0xc5, 0xd1, 0x37, 0x92, 0x9a, 0xfa, 0xbf, 0x15, 0x70, 0x23, 0x87, 0xf5, 0xd8, 0x66, 0x1c, 0x3e, + 0xcf, 0x44, 0xcc, 0x78, 0xbd, 0x88, 0x85, 0x6c, 0x11, 0xaf, 0xf8, 0xf0, 0x46, 0x23, 0x13, 0xd1, + 0x6a, 0x82, 0x05, 0x9b, 0x93, 0x6e, 0x94, 0x8a, 0xc6, 0xd5, 0x96, 0x65, 0xae, 0x49, 0xe9, 0x85, + 0xe3, 0x50, 0x04, 0x8d, 0xb4, 0xf4, 0x6f, 0xaf, 0xe5, 0x2e, 0x27, 0x0c, 0x27, 0x3c, 0x03, 0xab, + 0x5d, 0xdb, 0xdd, 0xef, 0x61, 0xdb, 0xc1, 0x2d, 0x79, 0x7a, 0xa6, 0x25, 0x41, 0x58, 0x61, 0x8d, + 0x51, 0x85, 0x35, 0x8e, 0x5d, 0x7e, 0x4a, 0x9b, 0x9c, 0xda, 0x6e, 0xc7, 0xdc, 0x18, 0x0e, 0xb4, + 0xd5, 0x93, 0x09, 0x25, 0x94, 0xd0, 0x85, 0xbf, 0x07, 0x65, 0x46, 0x1c, 0xd2, 0xe6, 0x1e, 0xbd, + 0x5a, 0x85, 0x78, 0x8c, 0x5b, 0xc4, 0x69, 0x4a, 0xaa, 0xb9, 0x1a, 0xc6, 0x2d, 0xfa, 0x42, 0xb1, + 0x24, 0x74, 0xc0, 0x7a, 0x17, 0x5f, 0x3e, 0x73, 0x71, 0xbc, 0x90, 0xd2, 0x0f, 0x5c, 0x08, 0x1c, + 0x0e, 0xb4, 0xf5, 0x93, 0x84, 0x16, 0x4a, 0x69, 0xeb, 0xc3, 0x79, 0x70, 0xb3, 0x30, 0xab, 0xe0, + 0xe7, 0x00, 0x7a, 0x2d, 0x46, 0x68, 0x8f, 0x58, 0x0f, 0x47, 0x77, 0x90, 0xed, 0x45, 0x07, 0x77, + 0x57, 0x6e, 0x10, 0x3c, 0xcd, 0x20, 0x50, 0x0e, 0x0b, 0xfe, 0x45, 0x01, 0x6b, 0xd6, 0xc8, 0x0d, + 0xb1, 0x1a, 0x9e, 0x15, 0x25, 0xc6, 0xc3, 0x1f, 0x92, 0xef, 0x46, 0x7d, 0x52, 0xe9, 0xd0, 0xe5, + 0xb4, 0x6f, 0x6e, 0xcb, 0x09, 0xad, 0x25, 0x6c, 0x28, 0xe9, 0x34, 0x5c, 0x92, 0x15, 0x4b, 0x32, + 0x79, 0xa7, 0x89, 0x10, 0x2f, 0x8c, 0x97, 0x54, 0xcf, 0x20, 0x50, 0x0e, 0x0b, 0xfe, 0x0a, 0xac, + 0xb7, 0x03, 0x4a, 0x89, 0xcb, 0x1f, 0x11, 0xec, 0xf0, 0xf3, 0xbe, 0x3a, 0x2f, 0x74, 0x76, 0xa4, + 0xce, 0xfa, 0x41, 0xc2, 0x8a, 0x52, 0xe8, 0x90, 0x6f, 0x11, 0x66, 0x53, 0x62, 0x45, 0xfc, 0x85, + 0x24, 0xbf, 0x9e, 0xb0, 0xa2, 0x14, 0x1a, 0x3e, 0x00, 0xab, 0xe4, 0xd2, 0x27, 0xed, 0x28, 0xa0, + 0x8b, 0x82, 0xbd, 0x25, 0xd9, 0xab, 0x87, 0x13, 0x36, 0x94, 0x40, 0xee, 0x3a, 0x00, 0x66, 0x23, + 0x08, 0x37, 0x40, 0xe9, 0x82, 0xf4, 0x47, 0xd7, 0x0e, 0x0a, 0x7f, 0xc2, 0xcf, 0xc0, 0x42, 0x0f, + 0x3b, 0x01, 0x91, 0x89, 0xfe, 0xc1, 0xeb, 0x25, 0xfa, 0x53, 0xbb, 0x4b, 0xd0, 0x88, 0xf8, 0xf3, + 0x6b, 0x0f, 0x14, 0xfd, 0x1b, 0x05, 0x54, 0x1a, 0x9e, 0xd5, 0x24, 0xed, 0x80, 0xda, 0xbc, 0xdf, + 0x10, 0x9b, 0xfc, 0x16, 0x0a, 0x36, 0x4a, 0x14, 0xec, 0x8f, 0xa7, 0x27, 0x5a, 0x72, 0x76, 0x45, + 0xe5, 0x5a, 0x7f, 0xa1, 0x80, 0xed, 0x0c, 0xfa, 0x2d, 0x94, 0xd3, 0x5f, 0x27, 0xcb, 0xe9, 0x87, + 0x57, 0x59, 0x4c, 0x41, 0x31, 0xfd, 0xa6, 0x92, 0xb3, 0x14, 0x51, 0x4a, 0xc3, 0xd6, 0x8e, 0xda, + 0x3d, 0xdb, 0x21, 0x1d, 0x62, 0x89, 0xc5, 0x94, 0x27, 0x5a, 0xbb, 0xd8, 0x82, 0x26, 0x50, 0x90, + 0x81, 0x1d, 0x8b, 0x9c, 0xe1, 0xc0, 0xe1, 0xfb, 0x96, 0x75, 0x80, 0x7d, 0xdc, 0xb2, 0x1d, 0x9b, + 0xdb, 0xb2, 0x17, 0x59, 0x36, 0x3f, 0x1d, 0x0e, 0xb4, 0x9d, 0x7a, 0x2e, 0xe2, 0xd5, 0x40, 0x7b, + 0x37, 0xdb, 0xca, 0x1b, 0x31, 0xa4, 0x8f, 0x0a, 0xa4, 0x61, 0x1f, 0xa8, 0x94, 0xfc, 0x29, 0x08, + 0x0f, 0x45, 0x9d, 0x7a, 0x7e, 0xc2, 0x6d, 0x49, 0xb8, 0xfd, 0xe5, 0x70, 0xa0, 0xa9, 0xa8, 0x00, + 0x33, 0xdb, 0x71, 0xa1, 0x3c, 0xfc, 0x12, 0x6c, 0x62, 0xd9, 0x84, 0x4f, 0x7a, 0x9d, 0x17, 0x5e, + 0x1f, 0x0c, 0x07, 0xda, 0xe6, 0x7e, 0xd6, 0x3c, 0xdb, 0x61, 0x9e, 0x28, 0xac, 0x81, 0xa5, 0x9e, + 0xe8, 0xd7, 0x99, 0xba, 0x20, 0xf4, 0xb7, 0x87, 0x03, 0x6d, 0x69, 0xd4, 0xc2, 0x87, 0x9a, 0x8b, + 0x47, 0x4d, 0xd1, 0x05, 0x46, 0x28, 0xf8, 0x09, 0x58, 0x39, 0xf7, 0x18, 0x7f, 0x42, 0xf8, 0x57, + 0x1e, 0xbd, 0x10, 0x85, 0xa1, 0x6c, 0x6e, 0xca, 0x1d, 0x5c, 0x79, 0x34, 0x36, 0xa1, 0x49, 0x1c, + 0xfc, 0x2d, 0x58, 0x3e, 0x97, 0x3d, 0x1f, 0x53, 0x97, 0x44, 0xa2, 0xdd, 0x9e, 0x92, 0x68, 0x89, + 0xfe, 0xd0, 0xac, 0x48, 0xf9, 0xe5, 0x68, 0x98, 0xa1, 0xb1, 0x1a, 0xfc, 0x09, 0x58, 0x12, 0x1f, + 0xc7, 0x75, 0xb5, 0x2c, 0x66, 0x73, 0x5d, 0xc2, 0x97, 0x1e, 0x8d, 0x86, 0x51, 0x64, 0x8f, 0xa0, + 0xc7, 0x8d, 0x03, 0x75, 0x39, 0x0b, 0x3d, 0x6e, 0x1c, 0xa0, 0xc8, 0x0e, 0x9f, 0x83, 0x25, 0x46, + 0x1e, 0xdb, 0x6e, 0x70, 0xa9, 0x02, 0x71, 0xe4, 0xee, 0x4c, 0x99, 0x6e, 0xf3, 0x50, 0x20, 0x53, + 0xdd, 0xf6, 0x58, 0x5d, 0xda, 0x51, 0x24, 0x09, 0x2d, 0xb0, 0x4c, 0x03, 0x77, 0x9f, 0x3d, 0x63, + 0x84, 0xaa, 0x2b, 0x99, 0xab, 0x3e, 0xad, 0x8f, 0x22, 0x6c, 0xda, 0x43, 0x1c, 0x99, 0x18, 0x81, + 0xc6, 0xc2, 0xd0, 0x02, 0x40, 0x7c, 0x88, 0xa6, 0x5e, 0xdd, 0x99, 0xd9, 0x04, 0xa2, 0x18, 0x9c, + 0xf6, 0xb3, 0x1e, 0x1e, 0xcf, 0xb1, 0x19, 0x4d, 0xe8, 0xc2, 0xbf, 0x2a, 0x00, 0xb2, 0xc0, 0xf7, + 0x1d, 0xd2, 0x25, 0x2e, 0xc7, 0x8e, 0x18, 0x65, 0xea, 0xaa, 0x70, 0xf7, 0x8b, 0x69, 0x51, 0xcb, + 0x90, 0xd2, 0x6e, 0xe3, 0x6b, 0x33, 0x0b, 0x45, 0x39, 0x3e, 0xc3, 0x4d, 0x3b, 0x93, 0xab, 0x5d, + 0x9b, 0xb9, 0x69, 0xf9, 0x7f, 0x91, 0xc6, 0x9b, 0x26, 0xed, 0x28, 0x92, 0x84, 0x5f, 0x80, 0x9d, + 0xe8, 0x0f, 0x24, 0xf2, 0x3c, 0x7e, 0x64, 0x3b, 0x84, 0xf5, 0x19, 0x27, 0x5d, 0x75, 0x5d, 0x24, + 0x53, 0x55, 0x32, 0x77, 0x50, 0x2e, 0x0a, 0x15, 0xb0, 0x61, 0x17, 0x68, 0x51, 0x11, 0x0a, 0x4f, + 0x68, 0x5c, 0x05, 0x0f, 0x59, 0x1b, 0x3b, 0xa3, 0xc6, 0xe8, 0xba, 0x70, 0xf0, 0xfe, 0x70, 0xa0, + 0x69, 0xf5, 0xe9, 0x50, 0x34, 0x4b, 0x0b, 0xfe, 0x06, 0xa8, 0xb8, 0xc8, 0xcf, 0x86, 0xf0, 0xf3, + 0xa3, 0xb0, 0xb2, 0x15, 0x3a, 0x28, 0x64, 0x43, 0x1f, 0x6c, 0xe0, 0xe4, 0x5f, 0x79, 0xa6, 0x56, + 0xc4, 0x59, 0xff, 0x60, 0xca, 0x3e, 0xa4, 0xfe, 0xfd, 0x9b, 0xaa, 0x0c, 0xe3, 0x46, 0xca, 0xc0, + 0x50, 0x46, 0x1d, 0x5e, 0x02, 0x88, 0xd3, 0x2f, 0x0f, 0x4c, 0x85, 0x33, 0x2f, 0xb2, 0xcc, 0x73, + 0xc5, 0x38, 0xd5, 0x32, 0x26, 0x86, 0x72, 0x7c, 0x40, 0x0e, 0x2a, 0x38, 0xf5, 0x52, 0xc2, 0xd4, + 0x1b, 0xc2, 0xf1, 0x4f, 0x67, 0x3b, 0x8e, 0x39, 0xe6, 0x4d, 0xe9, 0xb7, 0x92, 0xb6, 0x30, 0x94, + 0x75, 0x00, 0x1f, 0x83, 0x2d, 0x39, 0xf8, 0xcc, 0x65, 0xf8, 0x8c, 0x34, 0xfb, 0xac, 0xcd, 0x1d, + 0xa6, 0x6e, 0x8a, 0xda, 0xad, 0x0e, 0x07, 0xda, 0xd6, 0x7e, 0x8e, 0x1d, 0xe5, 0xb2, 0xe0, 0x67, + 0x60, 0xe3, 0xcc, 0xa3, 0x2d, 0xdb, 0xb2, 0x88, 0x1b, 0x29, 0x6d, 0x09, 0xa5, 0xad, 0x30, 0xfe, + 0x47, 0x29, 0x1b, 0xca, 0xa0, 0x21, 0x03, 0xdb, 0x52, 0xb9, 0x41, 0xbd, 0xf6, 0x89, 0x17, 0xb8, + 0x3c, 0xbc, 0x2e, 0x98, 0xba, 0x1d, 0x5f, 0x91, 0xdb, 0xfb, 0x79, 0x80, 0x57, 0x03, 0xed, 0x56, + 0xce, 0x75, 0x95, 0x00, 0xa1, 0x7c, 0x6d, 0xe8, 0x80, 0x55, 0xf9, 0xf6, 0x75, 0xe0, 0x60, 0xc6, + 0x54, 0x55, 0x1c, 0xf5, 0xfb, 0xd3, 0x0b, 0x5b, 0x0c, 0x4f, 0x9f, 0x77, 0xf1, 0xa7, 0x6c, 0x12, + 0x80, 0x12, 0xea, 0xfa, 0xdf, 0x15, 0x70, 0xb3, 0xb0, 0x30, 0xc2, 0xfb, 0x89, 0x07, 0x15, 0x3d, + 0xf5, 0xa0, 0x02, 0xb3, 0xc4, 0x37, 0xf0, 0x9e, 0xf2, 0xb5, 0x02, 0xd4, 0xa2, 0x1b, 0x02, 0x7e, + 0x92, 0x98, 0xe0, 0x7b, 0xa9, 0x09, 0x56, 0x32, 0xbc, 0x37, 0x30, 0xbf, 0x6f, 0x15, 0xf0, 0xce, + 0x94, 0x1d, 0x88, 0x0b, 0x12, 0xb1, 0x26, 0x51, 0x4f, 0x70, 0x78, 0x94, 0x15, 0x91, 0x47, 0xe3, + 0x82, 0x94, 0x83, 0x41, 0x85, 0x6c, 0xf8, 0x0c, 0xdc, 0x90, 0xd5, 0x30, 0x6d, 0x13, 0x9d, 0xfb, + 0xb2, 0xf9, 0xce, 0x70, 0xa0, 0xdd, 0xa8, 0xe7, 0x43, 0x50, 0x11, 0x57, 0xff, 0xa7, 0x02, 0x76, + 0xf2, 0xaf, 0x7c, 0x78, 0x37, 0x11, 0x6e, 0x2d, 0x15, 0xee, 0xeb, 0x29, 0x96, 0x0c, 0xf6, 0x1f, + 0xc0, 0xba, 0x6c, 0x0c, 0x92, 0xef, 0x83, 0x89, 0xa0, 0x87, 0x47, 0x24, 0xec, 0xe9, 0xa5, 0x44, + 0x94, 0xbe, 0xe2, 0xaf, 0x78, 0x72, 0x0c, 0xa5, 0xd4, 0xf4, 0x7f, 0x29, 0xe0, 0xbd, 0x99, 0x97, + 0x2d, 0x34, 0x13, 0x53, 0x37, 0x52, 0x53, 0xaf, 0x16, 0x0b, 0xbc, 0x99, 0x67, 0x42, 0xf3, 0xa3, + 0x17, 0x2f, 0xab, 0x73, 0xdf, 0xbd, 0xac, 0xce, 0x7d, 0xff, 0xb2, 0x3a, 0xf7, 0xe7, 0x61, 0x55, + 0x79, 0x31, 0xac, 0x2a, 0xdf, 0x0d, 0xab, 0xca, 0xf7, 0xc3, 0xaa, 0xf2, 0xdf, 0x61, 0x55, 0xf9, + 0xdb, 0xff, 0xaa, 0x73, 0xbf, 0x5b, 0x92, 0x72, 0xff, 0x0f, 0x00, 0x00, 0xff, 0xff, 0x5d, 0xe0, + 0x55, 0x1c, 0x41, 0x18, 0x00, 0x00, } func (m *AllowedCSIDriver) Marshal() (dAtA []byte, err error) { @@ -1155,7 +1155,7 @@ func (m *PodDisruptionBudgetStatus) MarshalToSizedBuffer(dAtA []byte) (int, erro i = encodeVarintGenerated(dAtA, i, uint64(m.CurrentHealthy)) i-- dAtA[i] = 0x20 - i = encodeVarintGenerated(dAtA, i, uint64(m.PodDisruptionsAllowed)) + i = encodeVarintGenerated(dAtA, i, uint64(m.DisruptionsAllowed)) i-- dAtA[i] = 0x18 if len(m.DisruptedPods) > 0 { @@ -1940,7 +1940,7 @@ func (m *PodDisruptionBudgetStatus) Size() (n int) { n += mapEntrySize + 1 + sovGenerated(uint64(mapEntrySize)) } } - n += 1 + sovGenerated(uint64(m.PodDisruptionsAllowed)) + n += 1 + sovGenerated(uint64(m.DisruptionsAllowed)) n += 1 + sovGenerated(uint64(m.CurrentHealthy)) n += 1 + sovGenerated(uint64(m.DesiredHealthy)) n += 1 + sovGenerated(uint64(m.ExpectedPods)) @@ -2307,7 +2307,7 @@ func (this *PodDisruptionBudgetStatus) String() string { s := strings.Join([]string{`&PodDisruptionBudgetStatus{`, `ObservedGeneration:` + fmt.Sprintf("%v", this.ObservedGeneration) + `,`, `DisruptedPods:` + mapStringForDisruptedPods + `,`, - `PodDisruptionsAllowed:` + fmt.Sprintf("%v", this.PodDisruptionsAllowed) + `,`, + `DisruptionsAllowed:` + fmt.Sprintf("%v", this.DisruptionsAllowed) + `,`, `CurrentHealthy:` + fmt.Sprintf("%v", this.CurrentHealthy) + `,`, `DesiredHealthy:` + fmt.Sprintf("%v", this.DesiredHealthy) + `,`, `ExpectedPods:` + fmt.Sprintf("%v", this.ExpectedPods) + `,`, @@ -3783,9 +3783,9 @@ func (m *PodDisruptionBudgetStatus) Unmarshal(dAtA []byte) error { iNdEx = postIndex case 3: if wireType != 0 { - return fmt.Errorf("proto: wrong wireType = %d for field PodDisruptionsAllowed", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field DisruptionsAllowed", wireType) } - m.PodDisruptionsAllowed = 0 + m.DisruptionsAllowed = 0 for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -3795,7 +3795,7 @@ func (m *PodDisruptionBudgetStatus) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - m.PodDisruptionsAllowed |= int32(b&0x7F) << shift + m.DisruptionsAllowed |= int32(b&0x7F) << shift if b < 0x80 { break } @@ -5478,6 +5478,7 @@ func (m *SupplementalGroupsStrategyOptions) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -5509,10 +5510,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -5533,55 +5532,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/policy/v1beta1/generated.proto b/vendor/k8s.io/api/policy/v1beta1/generated.proto index 9679dafc..30db726f 100644 --- a/vendor/k8s.io/api/policy/v1beta1/generated.proto +++ b/vendor/k8s.io/api/policy/v1beta1/generated.proto @@ -150,7 +150,7 @@ message PodDisruptionBudgetSpec { // PodDisruptionBudgetStatus represents information about the status of a // PodDisruptionBudget. Status may trail the actual state of a system. message PodDisruptionBudgetStatus { - // Most recent generation observed when updating this PDB status. PodDisruptionsAllowed and other + // Most recent generation observed when updating this PDB status. DisruptionsAllowed and other // status information is valid only if observedGeneration equals to PDB's object generation. // +optional optional int64 observedGeneration = 1; @@ -230,7 +230,7 @@ message PodSecurityPolicySpec { // +optional repeated string allowedCapabilities = 4; - // volumes is a white list of allowed volume plugins. Empty indicates that + // volumes is an allowlist of volume plugins. Empty indicates that // no volumes may be used. To allow all volumes you may use '*'. // +optional repeated string volumes = 5; @@ -287,27 +287,27 @@ message PodSecurityPolicySpec { // +optional optional bool allowPrivilegeEscalation = 16; - // allowedHostPaths is a white list of allowed host paths. Empty indicates + // allowedHostPaths is an allowlist of host paths. Empty indicates // that all host paths may be used. // +optional repeated AllowedHostPath allowedHostPaths = 17; - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all + // allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all // Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes // is allowed in the "volumes" field. // +optional repeated AllowedFlexVolume allowedFlexVolumes = 18; - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. + // AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. // An empty value indicates that any CSI driver can be used for inline ephemeral volumes. - // This is an alpha field, and is only honored if the API server enables the CSIInlineVolume feature gate. + // This is a beta field, and is only honored if the API server enables the CSIInlineVolume feature gate. // +optional repeated AllowedCSIDriver allowedCSIDrivers = 23; // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. // Each entry is either a plain sysctl name or ends in "*" in which case it is considered // as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. - // Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. + // Kubelet has to allowlist all allowed unsafe sysctls explicitly to avoid rejection. // // Examples: // e.g. "foo/*" allows "foo/bar", "foo/baz", etc. @@ -325,7 +325,7 @@ message PodSecurityPolicySpec { // +optional repeated string forbiddenSysctls = 20; - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. + // AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. // Empty or nil indicates that only the DefaultProcMountType may be used. // This requires the ProcMountType feature flag to be enabled. // +optional @@ -363,7 +363,7 @@ message RunAsUserStrategyOptions { // RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses // for a pod. message RuntimeClassStrategyOptions { - // allowedRuntimeClassNames is a whitelist of RuntimeClass names that may be specified on a pod. + // allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. // A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the // list. An empty list requires the RuntimeClassName field to be unset. repeated string allowedRuntimeClassNames = 1; diff --git a/vendor/k8s.io/api/policy/v1beta1/types.go b/vendor/k8s.io/api/policy/v1beta1/types.go index d8e417ab..711afc80 100644 --- a/vendor/k8s.io/api/policy/v1beta1/types.go +++ b/vendor/k8s.io/api/policy/v1beta1/types.go @@ -47,7 +47,7 @@ type PodDisruptionBudgetSpec struct { // PodDisruptionBudgetStatus represents information about the status of a // PodDisruptionBudget. Status may trail the actual state of a system. type PodDisruptionBudgetStatus struct { - // Most recent generation observed when updating this PDB status. PodDisruptionsAllowed and other + // Most recent generation observed when updating this PDB status. DisruptionsAllowed and other // status information is valid only if observedGeneration equals to PDB's object generation. // +optional ObservedGeneration int64 `json:"observedGeneration,omitempty" protobuf:"varint,1,opt,name=observedGeneration"` @@ -67,7 +67,7 @@ type PodDisruptionBudgetStatus struct { DisruptedPods map[string]metav1.Time `json:"disruptedPods,omitempty" protobuf:"bytes,2,rep,name=disruptedPods"` // Number of pod disruptions that are currently allowed. - PodDisruptionsAllowed int32 `json:"disruptionsAllowed" protobuf:"varint,3,opt,name=disruptionsAllowed"` + DisruptionsAllowed int32 `json:"disruptionsAllowed" protobuf:"varint,3,opt,name=disruptionsAllowed"` // current number of healthy pods CurrentHealthy int32 `json:"currentHealthy" protobuf:"varint,4,opt,name=currentHealthy"` @@ -81,6 +81,8 @@ type PodDisruptionBudgetStatus struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.5 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // PodDisruptionBudget is an object to define the max disruption that can be caused to a collection of pods type PodDisruptionBudget struct { @@ -97,6 +99,8 @@ type PodDisruptionBudget struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.5 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // PodDisruptionBudgetList is a collection of PodDisruptionBudgets. type PodDisruptionBudgetList struct { @@ -109,6 +113,8 @@ type PodDisruptionBudgetList struct { // +genclient // +genclient:noVerbs // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.5 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // Eviction evicts a pod from its node subject to certain policies and safety constraints. // This is a subresource of Pod. A request to cause such an eviction is @@ -128,6 +134,8 @@ type Eviction struct { // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.10 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // PodSecurityPolicy governs the ability to make requests that affect the Security Context // that will be applied to a pod and container. @@ -163,7 +171,7 @@ type PodSecurityPolicySpec struct { // You must not list a capability in both allowedCapabilities and requiredDropCapabilities. // +optional AllowedCapabilities []v1.Capability `json:"allowedCapabilities,omitempty" protobuf:"bytes,4,rep,name=allowedCapabilities,casttype=k8s.io/api/core/v1.Capability"` - // volumes is a white list of allowed volume plugins. Empty indicates that + // volumes is an allowlist of volume plugins. Empty indicates that // no volumes may be used. To allow all volumes you may use '*'. // +optional Volumes []FSType `json:"volumes,omitempty" protobuf:"bytes,5,rep,name=volumes,casttype=FSType"` @@ -207,24 +215,24 @@ type PodSecurityPolicySpec struct { // privilege escalation. If unspecified, defaults to true. // +optional AllowPrivilegeEscalation *bool `json:"allowPrivilegeEscalation,omitempty" protobuf:"varint,16,opt,name=allowPrivilegeEscalation"` - // allowedHostPaths is a white list of allowed host paths. Empty indicates + // allowedHostPaths is an allowlist of host paths. Empty indicates // that all host paths may be used. // +optional AllowedHostPaths []AllowedHostPath `json:"allowedHostPaths,omitempty" protobuf:"bytes,17,rep,name=allowedHostPaths"` - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all + // allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all // Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes // is allowed in the "volumes" field. // +optional AllowedFlexVolumes []AllowedFlexVolume `json:"allowedFlexVolumes,omitempty" protobuf:"bytes,18,rep,name=allowedFlexVolumes"` - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. + // AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. // An empty value indicates that any CSI driver can be used for inline ephemeral volumes. - // This is an alpha field, and is only honored if the API server enables the CSIInlineVolume feature gate. + // This is a beta field, and is only honored if the API server enables the CSIInlineVolume feature gate. // +optional AllowedCSIDrivers []AllowedCSIDriver `json:"allowedCSIDrivers,omitempty" protobuf:"bytes,23,rep,name=allowedCSIDrivers"` // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. // Each entry is either a plain sysctl name or ends in "*" in which case it is considered // as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. - // Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. + // Kubelet has to allowlist all allowed unsafe sysctls explicitly to avoid rejection. // // Examples: // e.g. "foo/*" allows "foo/bar", "foo/baz", etc. @@ -240,7 +248,7 @@ type PodSecurityPolicySpec struct { // e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. // +optional ForbiddenSysctls []string `json:"forbiddenSysctls,omitempty" protobuf:"bytes,20,rep,name=forbiddenSysctls"` - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. + // AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. // Empty or nil indicates that only the DefaultProcMountType may be used. // This requires the ProcMountType feature flag to be enabled. // +optional @@ -305,6 +313,7 @@ const ( PortworxVolume FSType = "portworxVolume" ScaleIO FSType = "scaleIO" CSI FSType = "csi" + Ephemeral FSType = "ephemeral" All FSType = "*" ) @@ -458,7 +467,7 @@ const ( // RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses // for a pod. type RuntimeClassStrategyOptions struct { - // allowedRuntimeClassNames is a whitelist of RuntimeClass names that may be specified on a pod. + // allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. // A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the // list. An empty list requires the RuntimeClassName field to be unset. AllowedRuntimeClassNames []string `json:"allowedRuntimeClassNames" protobuf:"bytes,1,rep,name=allowedRuntimeClassNames"` @@ -475,6 +484,8 @@ type RuntimeClassStrategyOptions struct { const AllowAllRuntimeClassNames = "*" // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.10 +// +k8s:prerelease-lifecycle-gen:deprecated=1.22 // PodSecurityPolicyList is a list of PodSecurityPolicy objects. type PodSecurityPolicyList struct { diff --git a/vendor/k8s.io/api/policy/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/policy/v1beta1/types_swagger_doc_generated.go index 40a951c4..05a50366 100644 --- a/vendor/k8s.io/api/policy/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/policy/v1beta1/types_swagger_doc_generated.go @@ -126,7 +126,7 @@ func (PodDisruptionBudgetSpec) SwaggerDoc() map[string]string { var map_PodDisruptionBudgetStatus = map[string]string{ "": "PodDisruptionBudgetStatus represents information about the status of a PodDisruptionBudget. Status may trail the actual state of a system.", - "observedGeneration": "Most recent generation observed when updating this PDB status. PodDisruptionsAllowed and other status information is valid only if observedGeneration equals to PDB's object generation.", + "observedGeneration": "Most recent generation observed when updating this PDB status. DisruptionsAllowed and other status information is valid only if observedGeneration equals to PDB's object generation.", "disruptedPods": "DisruptedPods contains information about pods whose eviction was processed by the API server eviction subresource handler but has not yet been observed by the PodDisruptionBudget controller. A pod will be in this map from the time when the API server processed the eviction request to the time when the pod is seen by PDB controller as having been marked for deletion (or after a timeout). The key in the map is the name of the pod and the value is the time when the API server processed the eviction request. If the deletion didn't occur and a pod is still there it will be removed from the list automatically by PodDisruptionBudget controller after some time. If everything goes smooth this map should be empty for the most of the time. Large number of entries in the map may indicate problems with pod deletions.", "disruptionsAllowed": "Number of pod disruptions that are currently allowed.", "currentHealthy": "current number of healthy pods", @@ -164,7 +164,7 @@ var map_PodSecurityPolicySpec = map[string]string{ "defaultAddCapabilities": "defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list.", "requiredDropCapabilities": "requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added.", "allowedCapabilities": "allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities.", - "volumes": "volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.", + "volumes": "volumes is an allowlist of volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.", "hostNetwork": "hostNetwork determines if the policy allows the use of HostNetwork in the pod spec.", "hostPorts": "hostPorts determines which host port ranges are allowed to be exposed.", "hostPID": "hostPID determines if the policy allows the use of HostPID in the pod spec.", @@ -177,12 +177,12 @@ var map_PodSecurityPolicySpec = map[string]string{ "readOnlyRootFilesystem": "readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to.", "defaultAllowPrivilegeEscalation": "defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process.", "allowPrivilegeEscalation": "allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true.", - "allowedHostPaths": "allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used.", - "allowedFlexVolumes": "allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the \"volumes\" field.", - "allowedCSIDrivers": "AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes. This is an alpha field, and is only honored if the API server enables the CSIInlineVolume feature gate.", - "allowedUnsafeSysctls": "allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in \"*\" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. \"foo/*\" allows \"foo/bar\", \"foo/baz\", etc. e.g. \"foo.*\" allows \"foo.bar\", \"foo.baz\", etc.", + "allowedHostPaths": "allowedHostPaths is an allowlist of host paths. Empty indicates that all host paths may be used.", + "allowedFlexVolumes": "allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the \"volumes\" field.", + "allowedCSIDrivers": "AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes. This is a beta field, and is only honored if the API server enables the CSIInlineVolume feature gate.", + "allowedUnsafeSysctls": "allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in \"*\" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to allowlist all allowed unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. \"foo/*\" allows \"foo/bar\", \"foo/baz\", etc. e.g. \"foo.*\" allows \"foo.bar\", \"foo.baz\", etc.", "forbiddenSysctls": "forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in \"*\" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden.\n\nExamples: e.g. \"foo/*\" forbids \"foo/bar\", \"foo/baz\", etc. e.g. \"foo.*\" forbids \"foo.bar\", \"foo.baz\", etc.", - "allowedProcMountTypes": "AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.", + "allowedProcMountTypes": "AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.", "runtimeClass": "runtimeClass is the strategy that will dictate the allowable RuntimeClasses for a pod. If this field is omitted, the pod's runtimeClassName field is unrestricted. Enforcement of this field depends on the RuntimeClass feature gate being enabled.", } @@ -212,7 +212,7 @@ func (RunAsUserStrategyOptions) SwaggerDoc() map[string]string { var map_RuntimeClassStrategyOptions = map[string]string{ "": "RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses for a pod.", - "allowedRuntimeClassNames": "allowedRuntimeClassNames is a whitelist of RuntimeClass names that may be specified on a pod. A value of \"*\" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.", + "allowedRuntimeClassNames": "allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of \"*\" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.", "defaultRuntimeClassName": "defaultRuntimeClassName is the default RuntimeClassName to set on the pod. The default MUST be allowed by the allowedRuntimeClassNames list. A value of nil does not mutate the Pod.", } diff --git a/vendor/k8s.io/api/policy/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/policy/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..fca0a2a2 --- /dev/null +++ b/vendor/k8s.io/api/policy/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,111 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Eviction) APILifecycleIntroduced() (major, minor int) { + return 1, 5 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Eviction) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Eviction) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PodDisruptionBudget) APILifecycleIntroduced() (major, minor int) { + return 1, 5 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PodDisruptionBudget) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PodDisruptionBudget) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PodDisruptionBudgetList) APILifecycleIntroduced() (major, minor int) { + return 1, 5 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PodDisruptionBudgetList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PodDisruptionBudgetList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PodSecurityPolicy) APILifecycleIntroduced() (major, minor int) { + return 1, 10 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PodSecurityPolicy) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PodSecurityPolicy) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PodSecurityPolicyList) APILifecycleIntroduced() (major, minor int) { + return 1, 10 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PodSecurityPolicyList) APILifecycleDeprecated() (major, minor int) { + return 1, 22 +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PodSecurityPolicyList) APILifecycleRemoved() (major, minor int) { + return 1, 25 +} diff --git a/vendor/k8s.io/api/rbac/v1/generated.pb.go b/vendor/k8s.io/api/rbac/v1/generated.pb.go index 9bb48fc3..ba6872d6 100644 --- a/vendor/k8s.io/api/rbac/v1/generated.pb.go +++ b/vendor/k8s.io/api/rbac/v1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *AggregationRule) Reset() { *m = AggregationRule{} } func (*AggregationRule) ProtoMessage() {} @@ -3183,6 +3183,7 @@ func (m *Subject) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -3214,10 +3215,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -3238,55 +3237,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/rbac/v1alpha1/generated.pb.go b/vendor/k8s.io/api/rbac/v1alpha1/generated.pb.go index 19336247..3b12526d 100644 --- a/vendor/k8s.io/api/rbac/v1alpha1/generated.pb.go +++ b/vendor/k8s.io/api/rbac/v1alpha1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *AggregationRule) Reset() { *m = AggregationRule{} } func (*AggregationRule) ProtoMessage() {} @@ -3184,6 +3184,7 @@ func (m *Subject) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -3215,10 +3216,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -3239,55 +3238,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/rbac/v1alpha1/generated.proto b/vendor/k8s.io/api/rbac/v1alpha1/generated.proto index 5c50a87e..60354e6a 100644 --- a/vendor/k8s.io/api/rbac/v1alpha1/generated.proto +++ b/vendor/k8s.io/api/rbac/v1alpha1/generated.proto @@ -37,7 +37,7 @@ message AggregationRule { } // ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22. message ClusterRole { // Standard object's metadata. // +optional @@ -56,7 +56,7 @@ message ClusterRole { // ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, // and adds who information via Subject. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22. message ClusterRoleBinding { // Standard object's metadata. // +optional @@ -72,7 +72,7 @@ message ClusterRoleBinding { } // ClusterRoleBindingList is a collection of ClusterRoleBindings. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindings, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindings, and will no longer be served in v1.22. message ClusterRoleBindingList { // Standard object's metadata. // +optional @@ -83,7 +83,7 @@ message ClusterRoleBindingList { } // ClusterRoleList is a collection of ClusterRoles. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.22. message ClusterRoleList { // Standard object's metadata. // +optional @@ -113,7 +113,6 @@ message PolicyRule { repeated string resourceNames = 5; // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path - // This name is intentionally different than the internal type so that the DefaultConvert works nicely and because the ordering may be different. // Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. // Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both. // +optional @@ -121,7 +120,7 @@ message PolicyRule { } // Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22. message Role { // Standard object's metadata. // +optional @@ -135,7 +134,7 @@ message Role { // RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. // It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given // namespace only have effect in that namespace. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22. message RoleBinding { // Standard object's metadata. // +optional @@ -151,7 +150,7 @@ message RoleBinding { } // RoleBindingList is a collection of RoleBindings -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.22. message RoleBindingList { // Standard object's metadata. // +optional @@ -162,7 +161,7 @@ message RoleBindingList { } // RoleList is a collection of Roles. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.22. message RoleList { // Standard object's metadata. // +optional diff --git a/vendor/k8s.io/api/rbac/v1alpha1/types.go b/vendor/k8s.io/api/rbac/v1alpha1/types.go index a5d3e38f..538ae4c9 100644 --- a/vendor/k8s.io/api/rbac/v1alpha1/types.go +++ b/vendor/k8s.io/api/rbac/v1alpha1/types.go @@ -62,7 +62,6 @@ type PolicyRule struct { ResourceNames []string `json:"resourceNames,omitempty" protobuf:"bytes,5,rep,name=resourceNames"` // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path - // This name is intentionally different than the internal type so that the DefaultConvert works nicely and because the ordering may be different. // Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. // Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both. // +optional @@ -103,7 +102,7 @@ type RoleRef struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22. type Role struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -121,7 +120,7 @@ type Role struct { // RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. // It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given // namespace only have effect in that namespace. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22. type RoleBinding struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -140,7 +139,7 @@ type RoleBinding struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // RoleBindingList is a collection of RoleBindings -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.22. type RoleBindingList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -154,7 +153,7 @@ type RoleBindingList struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // RoleList is a collection of Roles. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.22. type RoleList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -170,7 +169,7 @@ type RoleList struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22. type ClusterRole struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -202,7 +201,7 @@ type AggregationRule struct { // ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, // and adds who information via Subject. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22. type ClusterRoleBinding struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -221,7 +220,7 @@ type ClusterRoleBinding struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // ClusterRoleBindingList is a collection of ClusterRoleBindings. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindings, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindings, and will no longer be served in v1.22. type ClusterRoleBindingList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -235,7 +234,7 @@ type ClusterRoleBindingList struct { // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object // ClusterRoleList is a collection of ClusterRoles. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.22. type ClusterRoleList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. diff --git a/vendor/k8s.io/api/rbac/v1alpha1/types_swagger_doc_generated.go b/vendor/k8s.io/api/rbac/v1alpha1/types_swagger_doc_generated.go index 8238de21..acb84478 100644 --- a/vendor/k8s.io/api/rbac/v1alpha1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/rbac/v1alpha1/types_swagger_doc_generated.go @@ -37,7 +37,7 @@ func (AggregationRule) SwaggerDoc() map[string]string { } var map_ClusterRole = map[string]string{ - "": "ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.20.", + "": "ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "rules": "Rules holds all the PolicyRules for this ClusterRole", "aggregationRule": "AggregationRule is an optional field that describes how to build the Rules for this ClusterRole. If AggregationRule is set, then the Rules are controller managed and direct changes to Rules will be stomped by the controller.", @@ -48,7 +48,7 @@ func (ClusterRole) SwaggerDoc() map[string]string { } var map_ClusterRoleBinding = map[string]string{ - "": "ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.20.", + "": "ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "subjects": "Subjects holds references to the objects the role applies to.", "roleRef": "RoleRef can only reference a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error.", @@ -59,7 +59,7 @@ func (ClusterRoleBinding) SwaggerDoc() map[string]string { } var map_ClusterRoleBindingList = map[string]string{ - "": "ClusterRoleBindingList is a collection of ClusterRoleBindings. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindings, and will no longer be served in v1.20.", + "": "ClusterRoleBindingList is a collection of ClusterRoleBindings. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindings, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of ClusterRoleBindings", } @@ -69,7 +69,7 @@ func (ClusterRoleBindingList) SwaggerDoc() map[string]string { } var map_ClusterRoleList = map[string]string{ - "": "ClusterRoleList is a collection of ClusterRoles. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.20.", + "": "ClusterRoleList is a collection of ClusterRoles. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of ClusterRoles", } @@ -84,7 +84,7 @@ var map_PolicyRule = map[string]string{ "apiGroups": "APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed.", "resources": "Resources is a list of resources this rule applies to. ResourceAll represents all resources.", "resourceNames": "ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed.", - "nonResourceURLs": "NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path This name is intentionally different than the internal type so that the DefaultConvert works nicely and because the ordering may be different. Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as \"pods\" or \"secrets\") or non-resource URL paths (such as \"/api\"), but not both.", + "nonResourceURLs": "NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as \"pods\" or \"secrets\") or non-resource URL paths (such as \"/api\"), but not both.", } func (PolicyRule) SwaggerDoc() map[string]string { @@ -92,7 +92,7 @@ func (PolicyRule) SwaggerDoc() map[string]string { } var map_Role = map[string]string{ - "": "Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.20.", + "": "Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "rules": "Rules holds all the PolicyRules for this Role", } @@ -102,7 +102,7 @@ func (Role) SwaggerDoc() map[string]string { } var map_RoleBinding = map[string]string{ - "": "RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.20.", + "": "RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "subjects": "Subjects holds references to the objects the role applies to.", "roleRef": "RoleRef can reference a Role in the current namespace or a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error.", @@ -113,7 +113,7 @@ func (RoleBinding) SwaggerDoc() map[string]string { } var map_RoleBindingList = map[string]string{ - "": "RoleBindingList is a collection of RoleBindings Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.20.", + "": "RoleBindingList is a collection of RoleBindings Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of RoleBindings", } @@ -123,7 +123,7 @@ func (RoleBindingList) SwaggerDoc() map[string]string { } var map_RoleList = map[string]string{ - "": "RoleList is a collection of Roles. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.20.", + "": "RoleList is a collection of Roles. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of Roles", } diff --git a/vendor/k8s.io/api/rbac/v1beta1/doc.go b/vendor/k8s.io/api/rbac/v1beta1/doc.go index fe7aae97..156f273e 100644 --- a/vendor/k8s.io/api/rbac/v1beta1/doc.go +++ b/vendor/k8s.io/api/rbac/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=rbac.authorization.k8s.io diff --git a/vendor/k8s.io/api/rbac/v1beta1/generated.pb.go b/vendor/k8s.io/api/rbac/v1beta1/generated.pb.go index 6c80f52f..53d36320 100644 --- a/vendor/k8s.io/api/rbac/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/rbac/v1beta1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *AggregationRule) Reset() { *m = AggregationRule{} } func (*AggregationRule) ProtoMessage() {} @@ -3183,6 +3183,7 @@ func (m *Subject) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -3214,10 +3215,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -3238,55 +3237,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/rbac/v1beta1/generated.proto b/vendor/k8s.io/api/rbac/v1beta1/generated.proto index 87e0dbdf..44cd6c24 100644 --- a/vendor/k8s.io/api/rbac/v1beta1/generated.proto +++ b/vendor/k8s.io/api/rbac/v1beta1/generated.proto @@ -37,7 +37,7 @@ message AggregationRule { } // ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22. message ClusterRole { // Standard object's metadata. // +optional @@ -56,7 +56,7 @@ message ClusterRole { // ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, // and adds who information via Subject. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22. message ClusterRoleBinding { // Standard object's metadata. // +optional @@ -72,7 +72,7 @@ message ClusterRoleBinding { } // ClusterRoleBindingList is a collection of ClusterRoleBindings. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindingList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindingList, and will no longer be served in v1.22. message ClusterRoleBindingList { // Standard object's metadata. // +optional @@ -83,7 +83,7 @@ message ClusterRoleBindingList { } // ClusterRoleList is a collection of ClusterRoles. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.22. message ClusterRoleList { // Standard object's metadata. // +optional @@ -121,7 +121,7 @@ message PolicyRule { } // Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22. message Role { // Standard object's metadata. // +optional @@ -135,7 +135,7 @@ message Role { // RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. // It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given // namespace only have effect in that namespace. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22. message RoleBinding { // Standard object's metadata. // +optional @@ -151,7 +151,7 @@ message RoleBinding { } // RoleBindingList is a collection of RoleBindings -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.22. message RoleBindingList { // Standard object's metadata. // +optional @@ -162,7 +162,7 @@ message RoleBindingList { } // RoleList is a collection of Roles -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.22. message RoleList { // Standard object's metadata. // +optional diff --git a/vendor/k8s.io/api/rbac/v1beta1/types.go b/vendor/k8s.io/api/rbac/v1beta1/types.go index 74c70936..f1678881 100644 --- a/vendor/k8s.io/api/rbac/v1beta1/types.go +++ b/vendor/k8s.io/api/rbac/v1beta1/types.go @@ -100,9 +100,13 @@ type RoleRef struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,Role // Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22. type Role struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -116,11 +120,15 @@ type Role struct { // +genclient // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,RoleBinding // RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. // It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given // namespace only have effect in that namespace. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22. type RoleBinding struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -137,9 +145,13 @@ type RoleBinding struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,RoleBindingList // RoleBindingList is a collection of RoleBindings -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.22. type RoleBindingList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -151,9 +163,13 @@ type RoleBindingList struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,RoleList // RoleList is a collection of Roles -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.22. type RoleList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -167,9 +183,13 @@ type RoleList struct { // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,ClusterRole // ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22. type ClusterRole struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -197,10 +217,14 @@ type AggregationRule struct { // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,ClusterRoleBinding // ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, // and adds who information via Subject. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22. type ClusterRoleBinding struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -217,9 +241,13 @@ type ClusterRoleBinding struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,ClusterRoleBindingList // ClusterRoleBindingList is a collection of ClusterRoleBindings. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindingList, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindingList, and will no longer be served in v1.22. type ClusterRoleBindingList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. @@ -231,9 +259,13 @@ type ClusterRoleBindingList struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.6 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=rbac.authorization.k8s.io,v1,ClusterRoleList // ClusterRoleList is a collection of ClusterRoles. -// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.20. +// Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.22. type ClusterRoleList struct { metav1.TypeMeta `json:",inline"` // Standard object's metadata. diff --git a/vendor/k8s.io/api/rbac/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/rbac/v1beta1/types_swagger_doc_generated.go index 8e9d7ace..0512301f 100644 --- a/vendor/k8s.io/api/rbac/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/rbac/v1beta1/types_swagger_doc_generated.go @@ -37,7 +37,7 @@ func (AggregationRule) SwaggerDoc() map[string]string { } var map_ClusterRole = map[string]string{ - "": "ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.20.", + "": "ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "rules": "Rules holds all the PolicyRules for this ClusterRole", "aggregationRule": "AggregationRule is an optional field that describes how to build the Rules for this ClusterRole. If AggregationRule is set, then the Rules are controller managed and direct changes to Rules will be stomped by the controller.", @@ -48,7 +48,7 @@ func (ClusterRole) SwaggerDoc() map[string]string { } var map_ClusterRoleBinding = map[string]string{ - "": "ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.20.", + "": "ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "subjects": "Subjects holds references to the objects the role applies to.", "roleRef": "RoleRef can only reference a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error.", @@ -59,7 +59,7 @@ func (ClusterRoleBinding) SwaggerDoc() map[string]string { } var map_ClusterRoleBindingList = map[string]string{ - "": "ClusterRoleBindingList is a collection of ClusterRoleBindings. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindingList, and will no longer be served in v1.20.", + "": "ClusterRoleBindingList is a collection of ClusterRoleBindings. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBindingList, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of ClusterRoleBindings", } @@ -69,7 +69,7 @@ func (ClusterRoleBindingList) SwaggerDoc() map[string]string { } var map_ClusterRoleList = map[string]string{ - "": "ClusterRoleList is a collection of ClusterRoles. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.20.", + "": "ClusterRoleList is a collection of ClusterRoles. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoles, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of ClusterRoles", } @@ -92,7 +92,7 @@ func (PolicyRule) SwaggerDoc() map[string]string { } var map_Role = map[string]string{ - "": "Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.20.", + "": "Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "rules": "Rules holds all the PolicyRules for this Role", } @@ -102,7 +102,7 @@ func (Role) SwaggerDoc() map[string]string { } var map_RoleBinding = map[string]string{ - "": "RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.20.", + "": "RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "subjects": "Subjects holds references to the objects the role applies to.", "roleRef": "RoleRef can reference a Role in the current namespace or a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error.", @@ -113,7 +113,7 @@ func (RoleBinding) SwaggerDoc() map[string]string { } var map_RoleBindingList = map[string]string{ - "": "RoleBindingList is a collection of RoleBindings Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.20.", + "": "RoleBindingList is a collection of RoleBindings Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBindingList, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of RoleBindings", } @@ -123,7 +123,7 @@ func (RoleBindingList) SwaggerDoc() map[string]string { } var map_RoleList = map[string]string{ - "": "RoleList is a collection of Roles Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.20.", + "": "RoleList is a collection of Roles Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleList, and will no longer be served in v1.22.", "metadata": "Standard object's metadata.", "items": "Items is a list of Roles", } diff --git a/vendor/k8s.io/api/rbac/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/rbac/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..a4d99b35 --- /dev/null +++ b/vendor/k8s.io/api/rbac/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,217 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ClusterRole) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ClusterRole) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ClusterRole) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "ClusterRole"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ClusterRole) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ClusterRoleBinding) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ClusterRoleBinding) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ClusterRoleBinding) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "ClusterRoleBinding"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ClusterRoleBinding) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ClusterRoleBindingList) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ClusterRoleBindingList) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ClusterRoleBindingList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "ClusterRoleBindingList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ClusterRoleBindingList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ClusterRoleList) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ClusterRoleList) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ClusterRoleList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "ClusterRoleList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ClusterRoleList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *Role) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *Role) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *Role) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "Role"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *Role) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *RoleBinding) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *RoleBinding) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *RoleBinding) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "RoleBinding"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *RoleBinding) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *RoleBindingList) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *RoleBindingList) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *RoleBindingList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "RoleBindingList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *RoleBindingList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *RoleList) APILifecycleIntroduced() (major, minor int) { + return 1, 6 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *RoleList) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *RoleList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "rbac.authorization.k8s.io", Version: "v1", Kind: "RoleList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *RoleList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/scheduling/v1/generated.pb.go b/vendor/k8s.io/api/scheduling/v1/generated.pb.go index 7e9764f1..efc3102e 100644 --- a/vendor/k8s.io/api/scheduling/v1/generated.pb.go +++ b/vendor/k8s.io/api/scheduling/v1/generated.pb.go @@ -43,7 +43,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *PriorityClass) Reset() { *m = PriorityClass{} } func (*PriorityClass) ProtoMessage() {} @@ -652,6 +652,7 @@ func (m *PriorityClassList) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -683,10 +684,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -707,55 +706,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/scheduling/v1/generated.proto b/vendor/k8s.io/api/scheduling/v1/generated.proto index 82f6e0a2..e7489f53 100644 --- a/vendor/k8s.io/api/scheduling/v1/generated.proto +++ b/vendor/k8s.io/api/scheduling/v1/generated.proto @@ -57,7 +57,7 @@ message PriorityClass { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional optional string preemptionPolicy = 5; } diff --git a/vendor/k8s.io/api/scheduling/v1/types.go b/vendor/k8s.io/api/scheduling/v1/types.go index 087ee10d..b4ff3476 100644 --- a/vendor/k8s.io/api/scheduling/v1/types.go +++ b/vendor/k8s.io/api/scheduling/v1/types.go @@ -54,7 +54,7 @@ type PriorityClass struct { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional PreemptionPolicy *apiv1.PreemptionPolicy `json:"preemptionPolicy,omitempty" protobuf:"bytes,5,opt,name=preemptionPolicy"` } diff --git a/vendor/k8s.io/api/scheduling/v1/types_swagger_doc_generated.go b/vendor/k8s.io/api/scheduling/v1/types_swagger_doc_generated.go index 4cfb9d3e..7524adf9 100644 --- a/vendor/k8s.io/api/scheduling/v1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/scheduling/v1/types_swagger_doc_generated.go @@ -33,7 +33,7 @@ var map_PriorityClass = map[string]string{ "value": "The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec.", "globalDefault": "globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority.", "description": "description is an arbitrary string that usually provides guidelines on when this priority class should be used.", - "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature.", + "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.", } func (PriorityClass) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/scheduling/v1alpha1/generated.pb.go b/vendor/k8s.io/api/scheduling/v1alpha1/generated.pb.go index 05bff0ff..8a62104d 100644 --- a/vendor/k8s.io/api/scheduling/v1alpha1/generated.pb.go +++ b/vendor/k8s.io/api/scheduling/v1alpha1/generated.pb.go @@ -43,7 +43,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *PriorityClass) Reset() { *m = PriorityClass{} } func (*PriorityClass) ProtoMessage() {} @@ -652,6 +652,7 @@ func (m *PriorityClassList) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -683,10 +684,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -707,55 +706,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/scheduling/v1alpha1/generated.proto b/vendor/k8s.io/api/scheduling/v1alpha1/generated.proto index 682fb873..8c4a2c46 100644 --- a/vendor/k8s.io/api/scheduling/v1alpha1/generated.proto +++ b/vendor/k8s.io/api/scheduling/v1alpha1/generated.proto @@ -58,7 +58,7 @@ message PriorityClass { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional optional string preemptionPolicy = 5; } diff --git a/vendor/k8s.io/api/scheduling/v1alpha1/types.go b/vendor/k8s.io/api/scheduling/v1alpha1/types.go index 86a2c513..4d8462c7 100644 --- a/vendor/k8s.io/api/scheduling/v1alpha1/types.go +++ b/vendor/k8s.io/api/scheduling/v1alpha1/types.go @@ -55,7 +55,7 @@ type PriorityClass struct { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional PreemptionPolicy *apiv1.PreemptionPolicy `json:"preemptionPolicy,omitempty" protobuf:"bytes,5,opt,name=preemptionPolicy"` } diff --git a/vendor/k8s.io/api/scheduling/v1alpha1/types_swagger_doc_generated.go b/vendor/k8s.io/api/scheduling/v1alpha1/types_swagger_doc_generated.go index 63a9a353..940c39a0 100644 --- a/vendor/k8s.io/api/scheduling/v1alpha1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/scheduling/v1alpha1/types_swagger_doc_generated.go @@ -33,7 +33,7 @@ var map_PriorityClass = map[string]string{ "value": "The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec.", "globalDefault": "globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority.", "description": "description is an arbitrary string that usually provides guidelines on when this priority class should be used.", - "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature.", + "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.", } func (PriorityClass) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/scheduling/v1beta1/doc.go b/vendor/k8s.io/api/scheduling/v1beta1/doc.go index e6619689..152d241f 100644 --- a/vendor/k8s.io/api/scheduling/v1beta1/doc.go +++ b/vendor/k8s.io/api/scheduling/v1beta1/doc.go @@ -17,6 +17,7 @@ limitations under the License. // +k8s:deepcopy-gen=package // +k8s:protobuf-gen=package // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=scheduling.k8s.io diff --git a/vendor/k8s.io/api/scheduling/v1beta1/generated.pb.go b/vendor/k8s.io/api/scheduling/v1beta1/generated.pb.go index 198fcd02..b89af56b 100644 --- a/vendor/k8s.io/api/scheduling/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/scheduling/v1beta1/generated.pb.go @@ -43,7 +43,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *PriorityClass) Reset() { *m = PriorityClass{} } func (*PriorityClass) ProtoMessage() {} @@ -652,6 +652,7 @@ func (m *PriorityClassList) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -683,10 +684,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -707,55 +706,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/scheduling/v1beta1/generated.proto b/vendor/k8s.io/api/scheduling/v1beta1/generated.proto index 2582891b..eae3c01f 100644 --- a/vendor/k8s.io/api/scheduling/v1beta1/generated.proto +++ b/vendor/k8s.io/api/scheduling/v1beta1/generated.proto @@ -58,7 +58,7 @@ message PriorityClass { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional optional string preemptionPolicy = 5; } diff --git a/vendor/k8s.io/api/scheduling/v1beta1/types.go b/vendor/k8s.io/api/scheduling/v1beta1/types.go index f806ecd4..d68b4b31 100644 --- a/vendor/k8s.io/api/scheduling/v1beta1/types.go +++ b/vendor/k8s.io/api/scheduling/v1beta1/types.go @@ -24,6 +24,10 @@ import ( // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.11 +// +k8s:prerelease-lifecycle-gen:deprecated=1.14 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=scheduling.k8s.io,v1,PriorityClass // DEPRECATED - This group version of PriorityClass is deprecated by scheduling.k8s.io/v1/PriorityClass. // PriorityClass defines mapping from a priority class name to the priority @@ -55,12 +59,16 @@ type PriorityClass struct { // PreemptionPolicy is the Policy for preempting pods with lower priority. // One of Never, PreemptLowerPriority. // Defaults to PreemptLowerPriority if unset. - // This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature. + // This field is beta-level, gated by the NonPreemptingPriority feature-gate. // +optional PreemptionPolicy *apiv1.PreemptionPolicy `json:"preemptionPolicy,omitempty" protobuf:"bytes,5,opt,name=preemptionPolicy"` } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.11 +// +k8s:prerelease-lifecycle-gen:deprecated=1.14 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=scheduling.k8s.io,v1,PriorityClassList // PriorityClassList is a collection of priority classes. type PriorityClassList struct { diff --git a/vendor/k8s.io/api/scheduling/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/scheduling/v1beta1/types_swagger_doc_generated.go index ffded9df..d576c840 100644 --- a/vendor/k8s.io/api/scheduling/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/scheduling/v1beta1/types_swagger_doc_generated.go @@ -33,7 +33,7 @@ var map_PriorityClass = map[string]string{ "value": "The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec.", "globalDefault": "globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority.", "description": "description is an arbitrary string that usually provides guidelines on when this priority class should be used.", - "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is alpha-level and is only honored by servers that enable the NonPreemptingPriority feature.", + "preemptionPolicy": "PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.", } func (PriorityClass) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/scheduling/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/scheduling/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..45969d15 --- /dev/null +++ b/vendor/k8s.io/api/scheduling/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,73 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PriorityClass) APILifecycleIntroduced() (major, minor int) { + return 1, 11 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PriorityClass) APILifecycleDeprecated() (major, minor int) { + return 1, 14 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *PriorityClass) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "scheduling.k8s.io", Version: "v1", Kind: "PriorityClass"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PriorityClass) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *PriorityClassList) APILifecycleIntroduced() (major, minor int) { + return 1, 11 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *PriorityClassList) APILifecycleDeprecated() (major, minor int) { + return 1, 14 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *PriorityClassList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "scheduling.k8s.io", Version: "v1", Kind: "PriorityClassList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *PriorityClassList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/api/settings/v1alpha1/generated.pb.go b/vendor/k8s.io/api/settings/v1alpha1/generated.pb.go index 7469f74a..7ed066d3 100644 --- a/vendor/k8s.io/api/settings/v1alpha1/generated.pb.go +++ b/vendor/k8s.io/api/settings/v1alpha1/generated.pb.go @@ -42,7 +42,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *PodPreset) Reset() { *m = PodPreset{} } func (*PodPreset) ProtoMessage() {} @@ -970,6 +970,7 @@ func (m *PodPresetSpec) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1001,10 +1002,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1025,55 +1024,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/storage/v1/generated.pb.go b/vendor/k8s.io/api/storage/v1/generated.pb.go index 3d09ee7e..2c7088c3 100644 --- a/vendor/k8s.io/api/storage/v1/generated.pb.go +++ b/vendor/k8s.io/api/storage/v1/generated.pb.go @@ -44,12 +44,96 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package + +func (m *CSIDriver) Reset() { *m = CSIDriver{} } +func (*CSIDriver) ProtoMessage() {} +func (*CSIDriver) Descriptor() ([]byte, []int) { + return fileDescriptor_3b530c1983504d8d, []int{0} +} +func (m *CSIDriver) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CSIDriver) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CSIDriver) XXX_Merge(src proto.Message) { + xxx_messageInfo_CSIDriver.Merge(m, src) +} +func (m *CSIDriver) XXX_Size() int { + return m.Size() +} +func (m *CSIDriver) XXX_DiscardUnknown() { + xxx_messageInfo_CSIDriver.DiscardUnknown(m) +} + +var xxx_messageInfo_CSIDriver proto.InternalMessageInfo + +func (m *CSIDriverList) Reset() { *m = CSIDriverList{} } +func (*CSIDriverList) ProtoMessage() {} +func (*CSIDriverList) Descriptor() ([]byte, []int) { + return fileDescriptor_3b530c1983504d8d, []int{1} +} +func (m *CSIDriverList) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CSIDriverList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CSIDriverList) XXX_Merge(src proto.Message) { + xxx_messageInfo_CSIDriverList.Merge(m, src) +} +func (m *CSIDriverList) XXX_Size() int { + return m.Size() +} +func (m *CSIDriverList) XXX_DiscardUnknown() { + xxx_messageInfo_CSIDriverList.DiscardUnknown(m) +} + +var xxx_messageInfo_CSIDriverList proto.InternalMessageInfo + +func (m *CSIDriverSpec) Reset() { *m = CSIDriverSpec{} } +func (*CSIDriverSpec) ProtoMessage() {} +func (*CSIDriverSpec) Descriptor() ([]byte, []int) { + return fileDescriptor_3b530c1983504d8d, []int{2} +} +func (m *CSIDriverSpec) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CSIDriverSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CSIDriverSpec) XXX_Merge(src proto.Message) { + xxx_messageInfo_CSIDriverSpec.Merge(m, src) +} +func (m *CSIDriverSpec) XXX_Size() int { + return m.Size() +} +func (m *CSIDriverSpec) XXX_DiscardUnknown() { + xxx_messageInfo_CSIDriverSpec.DiscardUnknown(m) +} + +var xxx_messageInfo_CSIDriverSpec proto.InternalMessageInfo func (m *CSINode) Reset() { *m = CSINode{} } func (*CSINode) ProtoMessage() {} func (*CSINode) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{0} + return fileDescriptor_3b530c1983504d8d, []int{3} } func (m *CSINode) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -77,7 +161,7 @@ var xxx_messageInfo_CSINode proto.InternalMessageInfo func (m *CSINodeDriver) Reset() { *m = CSINodeDriver{} } func (*CSINodeDriver) ProtoMessage() {} func (*CSINodeDriver) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{1} + return fileDescriptor_3b530c1983504d8d, []int{4} } func (m *CSINodeDriver) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -105,7 +189,7 @@ var xxx_messageInfo_CSINodeDriver proto.InternalMessageInfo func (m *CSINodeList) Reset() { *m = CSINodeList{} } func (*CSINodeList) ProtoMessage() {} func (*CSINodeList) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{2} + return fileDescriptor_3b530c1983504d8d, []int{5} } func (m *CSINodeList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -133,7 +217,7 @@ var xxx_messageInfo_CSINodeList proto.InternalMessageInfo func (m *CSINodeSpec) Reset() { *m = CSINodeSpec{} } func (*CSINodeSpec) ProtoMessage() {} func (*CSINodeSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{3} + return fileDescriptor_3b530c1983504d8d, []int{6} } func (m *CSINodeSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -161,7 +245,7 @@ var xxx_messageInfo_CSINodeSpec proto.InternalMessageInfo func (m *StorageClass) Reset() { *m = StorageClass{} } func (*StorageClass) ProtoMessage() {} func (*StorageClass) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{4} + return fileDescriptor_3b530c1983504d8d, []int{7} } func (m *StorageClass) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -189,7 +273,7 @@ var xxx_messageInfo_StorageClass proto.InternalMessageInfo func (m *StorageClassList) Reset() { *m = StorageClassList{} } func (*StorageClassList) ProtoMessage() {} func (*StorageClassList) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{5} + return fileDescriptor_3b530c1983504d8d, []int{8} } func (m *StorageClassList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -217,7 +301,7 @@ var xxx_messageInfo_StorageClassList proto.InternalMessageInfo func (m *VolumeAttachment) Reset() { *m = VolumeAttachment{} } func (*VolumeAttachment) ProtoMessage() {} func (*VolumeAttachment) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{6} + return fileDescriptor_3b530c1983504d8d, []int{9} } func (m *VolumeAttachment) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -245,7 +329,7 @@ var xxx_messageInfo_VolumeAttachment proto.InternalMessageInfo func (m *VolumeAttachmentList) Reset() { *m = VolumeAttachmentList{} } func (*VolumeAttachmentList) ProtoMessage() {} func (*VolumeAttachmentList) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{7} + return fileDescriptor_3b530c1983504d8d, []int{10} } func (m *VolumeAttachmentList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -273,7 +357,7 @@ var xxx_messageInfo_VolumeAttachmentList proto.InternalMessageInfo func (m *VolumeAttachmentSource) Reset() { *m = VolumeAttachmentSource{} } func (*VolumeAttachmentSource) ProtoMessage() {} func (*VolumeAttachmentSource) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{8} + return fileDescriptor_3b530c1983504d8d, []int{11} } func (m *VolumeAttachmentSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -301,7 +385,7 @@ var xxx_messageInfo_VolumeAttachmentSource proto.InternalMessageInfo func (m *VolumeAttachmentSpec) Reset() { *m = VolumeAttachmentSpec{} } func (*VolumeAttachmentSpec) ProtoMessage() {} func (*VolumeAttachmentSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{9} + return fileDescriptor_3b530c1983504d8d, []int{12} } func (m *VolumeAttachmentSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -329,7 +413,7 @@ var xxx_messageInfo_VolumeAttachmentSpec proto.InternalMessageInfo func (m *VolumeAttachmentStatus) Reset() { *m = VolumeAttachmentStatus{} } func (*VolumeAttachmentStatus) ProtoMessage() {} func (*VolumeAttachmentStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{10} + return fileDescriptor_3b530c1983504d8d, []int{13} } func (m *VolumeAttachmentStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -357,7 +441,7 @@ var xxx_messageInfo_VolumeAttachmentStatus proto.InternalMessageInfo func (m *VolumeError) Reset() { *m = VolumeError{} } func (*VolumeError) ProtoMessage() {} func (*VolumeError) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{11} + return fileDescriptor_3b530c1983504d8d, []int{14} } func (m *VolumeError) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -385,7 +469,7 @@ var xxx_messageInfo_VolumeError proto.InternalMessageInfo func (m *VolumeNodeResources) Reset() { *m = VolumeNodeResources{} } func (*VolumeNodeResources) ProtoMessage() {} func (*VolumeNodeResources) Descriptor() ([]byte, []int) { - return fileDescriptor_3b530c1983504d8d, []int{12} + return fileDescriptor_3b530c1983504d8d, []int{15} } func (m *VolumeNodeResources) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -411,6 +495,9 @@ func (m *VolumeNodeResources) XXX_DiscardUnknown() { var xxx_messageInfo_VolumeNodeResources proto.InternalMessageInfo func init() { + proto.RegisterType((*CSIDriver)(nil), "k8s.io.api.storage.v1.CSIDriver") + proto.RegisterType((*CSIDriverList)(nil), "k8s.io.api.storage.v1.CSIDriverList") + proto.RegisterType((*CSIDriverSpec)(nil), "k8s.io.api.storage.v1.CSIDriverSpec") proto.RegisterType((*CSINode)(nil), "k8s.io.api.storage.v1.CSINode") proto.RegisterType((*CSINodeDriver)(nil), "k8s.io.api.storage.v1.CSINodeDriver") proto.RegisterType((*CSINodeList)(nil), "k8s.io.api.storage.v1.CSINodeList") @@ -433,83 +520,254 @@ func init() { } var fileDescriptor_3b530c1983504d8d = []byte{ - // 1212 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x57, 0x41, 0x6f, 0xe3, 0x44, - 0x14, 0xae, 0x9b, 0xa4, 0x4d, 0x27, 0x2d, 0x9b, 0xce, 0x16, 0x08, 0x39, 0x24, 0x95, 0x41, 0x10, - 0x0a, 0xeb, 0x6c, 0x97, 0x65, 0xb5, 0x42, 0x02, 0x29, 0x6e, 0x23, 0x51, 0xd1, 0xb4, 0xd5, 0xb4, - 0xac, 0x10, 0x02, 0xc4, 0xd4, 0x1e, 0x52, 0x6f, 0x62, 0x8f, 0xf1, 0x4c, 0x02, 0xb9, 0x71, 0xe2, - 0x86, 0x04, 0x57, 0x7e, 0x05, 0x5c, 0x39, 0x72, 0x2a, 0xb7, 0x15, 0xa7, 0x3d, 0x45, 0xd4, 0x9c, - 0xe1, 0x07, 0xf4, 0x84, 0x66, 0x3c, 0x8d, 0x9d, 0xc4, 0x29, 0xe9, 0xa5, 0xb7, 0xcc, 0x9b, 0xf7, - 0x7d, 0xef, 0xbd, 0xf9, 0xde, 0xbc, 0x71, 0xc0, 0x07, 0x9d, 0xc7, 0xcc, 0x70, 0x68, 0xbd, 0xd3, - 0x3b, 0x25, 0x81, 0x47, 0x38, 0x61, 0xf5, 0x3e, 0xf1, 0x6c, 0x1a, 0xd4, 0xd5, 0x06, 0xf6, 0x9d, - 0x3a, 0xe3, 0x34, 0xc0, 0x6d, 0x52, 0xef, 0x6f, 0xd7, 0xdb, 0xc4, 0x23, 0x01, 0xe6, 0xc4, 0x36, - 0xfc, 0x80, 0x72, 0x0a, 0x5f, 0x8c, 0xdc, 0x0c, 0xec, 0x3b, 0x86, 0x72, 0x33, 0xfa, 0xdb, 0xe5, - 0x7b, 0x6d, 0x87, 0x9f, 0xf5, 0x4e, 0x0d, 0x8b, 0xba, 0xf5, 0x36, 0x6d, 0xd3, 0xba, 0xf4, 0x3e, - 0xed, 0x7d, 0x25, 0x57, 0x72, 0x21, 0x7f, 0x45, 0x2c, 0x65, 0x3d, 0x11, 0xcc, 0xa2, 0x41, 0x5a, - 0xa4, 0xf2, 0xc3, 0xd8, 0xc7, 0xc5, 0xd6, 0x99, 0xe3, 0x91, 0x60, 0x50, 0xf7, 0x3b, 0x6d, 0x61, - 0x60, 0x75, 0x97, 0x70, 0x9c, 0x86, 0xaa, 0xcf, 0x42, 0x05, 0x3d, 0x8f, 0x3b, 0x2e, 0x99, 0x02, - 0x3c, 0xfa, 0x3f, 0x00, 0xb3, 0xce, 0x88, 0x8b, 0x27, 0x71, 0xfa, 0xaf, 0x1a, 0x58, 0xde, 0x39, - 0xde, 0x3b, 0xa0, 0x36, 0x81, 0x5f, 0x82, 0xbc, 0xc8, 0xc7, 0xc6, 0x1c, 0x97, 0xb4, 0x4d, 0xad, - 0x56, 0x78, 0x70, 0xdf, 0x88, 0xcf, 0x69, 0x44, 0x6b, 0xf8, 0x9d, 0xb6, 0x30, 0x30, 0x43, 0x78, - 0x1b, 0xfd, 0x6d, 0xe3, 0xf0, 0xf4, 0x29, 0xb1, 0x78, 0x8b, 0x70, 0x6c, 0xc2, 0xf3, 0x61, 0x75, - 0x21, 0x1c, 0x56, 0x41, 0x6c, 0x43, 0x23, 0x56, 0xb8, 0x0b, 0xb2, 0xcc, 0x27, 0x56, 0x69, 0x51, - 0xb2, 0xeb, 0x46, 0xaa, 0x0a, 0x86, 0xca, 0xe7, 0xd8, 0x27, 0x96, 0xb9, 0xaa, 0xf8, 0xb2, 0x62, - 0x85, 0x24, 0x5a, 0xff, 0x57, 0x03, 0x6b, 0xca, 0x67, 0x37, 0x70, 0xfa, 0x24, 0x80, 0x9b, 0x20, - 0xeb, 0x61, 0x97, 0xc8, 0xac, 0x57, 0x62, 0xcc, 0x01, 0x76, 0x09, 0x92, 0x3b, 0xf0, 0x75, 0xb0, - 0xe4, 0x51, 0x9b, 0xec, 0xed, 0xca, 0xd8, 0x2b, 0xe6, 0x0b, 0xca, 0x67, 0xe9, 0x40, 0x5a, 0x91, - 0xda, 0x85, 0x0f, 0xc1, 0x2a, 0xa7, 0x3e, 0xed, 0xd2, 0xf6, 0xe0, 0x23, 0x32, 0x60, 0xa5, 0xcc, - 0x66, 0xa6, 0xb6, 0x62, 0x16, 0xc3, 0x61, 0x75, 0xf5, 0x24, 0x61, 0x47, 0x63, 0x5e, 0xf0, 0x73, - 0x50, 0xc0, 0xdd, 0x2e, 0xb5, 0x30, 0xc7, 0xa7, 0x5d, 0x52, 0xca, 0xca, 0xf2, 0xb6, 0x66, 0x94, - 0xf7, 0x84, 0x76, 0x7b, 0x2e, 0x11, 0x71, 0x11, 0x61, 0xb4, 0x17, 0x58, 0x84, 0x99, 0x77, 0xc2, - 0x61, 0xb5, 0xd0, 0x88, 0x29, 0x50, 0x92, 0x4f, 0xff, 0x45, 0x03, 0x05, 0x55, 0xf0, 0xbe, 0xc3, - 0x38, 0xfc, 0x6c, 0x4a, 0x28, 0x63, 0x3e, 0xa1, 0x04, 0x5a, 0xca, 0x54, 0x54, 0xe5, 0xe7, 0xaf, - 0x2c, 0x09, 0x91, 0x76, 0x40, 0xce, 0xe1, 0xc4, 0x65, 0xa5, 0xc5, 0xcd, 0x4c, 0xad, 0xf0, 0xa0, - 0x72, 0xbd, 0x4a, 0xe6, 0x9a, 0xa2, 0xca, 0xed, 0x09, 0x10, 0x8a, 0xb0, 0xfa, 0x17, 0xa3, 0x8c, - 0x85, 0x70, 0xf0, 0x10, 0x2c, 0xdb, 0x52, 0x2a, 0x56, 0xd2, 0x24, 0xeb, 0x6b, 0xd7, 0xb3, 0x46, - 0xba, 0x9a, 0x77, 0x14, 0xf7, 0x72, 0xb4, 0x66, 0xe8, 0x8a, 0x45, 0xff, 0x61, 0x09, 0xac, 0x1e, - 0x47, 0xb0, 0x9d, 0x2e, 0x66, 0xec, 0x16, 0x9a, 0xf7, 0x5d, 0x50, 0xf0, 0x03, 0xda, 0x77, 0x98, - 0x43, 0x3d, 0x12, 0xa8, 0x3e, 0xba, 0xab, 0x20, 0x85, 0xa3, 0x78, 0x0b, 0x25, 0xfd, 0x60, 0x1b, - 0x00, 0x1f, 0x07, 0xd8, 0x25, 0x5c, 0x54, 0x9f, 0x91, 0xd5, 0xbf, 0x33, 0xa3, 0xfa, 0x64, 0x45, - 0xc6, 0xd1, 0x08, 0xd5, 0xf4, 0x78, 0x30, 0x88, 0xb3, 0x8b, 0x37, 0x50, 0x82, 0x1a, 0x76, 0xc0, - 0x5a, 0x40, 0xac, 0x2e, 0x76, 0xdc, 0x23, 0xda, 0x75, 0xac, 0x81, 0x6c, 0xc3, 0x15, 0xb3, 0x19, - 0x0e, 0xab, 0x6b, 0x28, 0xb9, 0x71, 0x39, 0xac, 0xde, 0x9f, 0x9e, 0x5c, 0xc6, 0x11, 0x09, 0x98, - 0xc3, 0x38, 0xf1, 0x78, 0xd4, 0xa1, 0x63, 0x18, 0x34, 0xce, 0x2d, 0xee, 0x89, 0x4b, 0x7b, 0x1e, - 0x3f, 0xf4, 0xb9, 0x43, 0x3d, 0x56, 0xca, 0xc5, 0xf7, 0xa4, 0x95, 0xb0, 0xa3, 0x31, 0x2f, 0xb8, - 0x0f, 0x36, 0x44, 0x5f, 0x7f, 0x13, 0x05, 0x68, 0x7e, 0xeb, 0x63, 0x4f, 0x9c, 0x52, 0x69, 0x69, - 0x53, 0xab, 0xe5, 0xcd, 0x52, 0x38, 0xac, 0x6e, 0x34, 0x52, 0xf6, 0x51, 0x2a, 0x0a, 0x7e, 0x02, - 0xd6, 0xfb, 0xd2, 0x64, 0x3a, 0x9e, 0xed, 0x78, 0xed, 0x16, 0xb5, 0x49, 0x69, 0x59, 0x16, 0xbd, - 0x15, 0x0e, 0xab, 0xeb, 0x4f, 0x26, 0x37, 0x2f, 0xd3, 0x8c, 0x68, 0x9a, 0x04, 0x7e, 0x0d, 0xd6, - 0x65, 0x44, 0x62, 0xab, 0x4b, 0xef, 0x10, 0x56, 0xca, 0x4b, 0xe9, 0x6a, 0x49, 0xe9, 0xc4, 0xd1, - 0x09, 0xdd, 0xae, 0x46, 0xc3, 0x31, 0xe9, 0x12, 0x8b, 0xd3, 0xe0, 0x84, 0x04, 0xae, 0xf9, 0x8a, - 0xd2, 0x6b, 0xbd, 0x31, 0x49, 0x85, 0xa6, 0xd9, 0xcb, 0xef, 0x83, 0x3b, 0x13, 0x82, 0xc3, 0x22, - 0xc8, 0x74, 0xc8, 0x20, 0x1a, 0x6a, 0x48, 0xfc, 0x84, 0x1b, 0x20, 0xd7, 0xc7, 0xdd, 0x1e, 0x89, - 0x9a, 0x0f, 0x45, 0x8b, 0xf7, 0x16, 0x1f, 0x6b, 0xfa, 0x6f, 0x1a, 0x28, 0x26, 0xbb, 0xe7, 0x16, - 0xe6, 0xc4, 0x87, 0xe3, 0x73, 0xe2, 0xd5, 0x39, 0x7a, 0x7a, 0xc6, 0xb0, 0xf8, 0x79, 0x11, 0x14, - 0x23, 0x5d, 0x1a, 0x9c, 0x63, 0xeb, 0xcc, 0x25, 0x1e, 0xbf, 0x85, 0x0b, 0xdd, 0x1a, 0x7b, 0x8d, - 0xde, 0xba, 0x76, 0x5c, 0xc7, 0x89, 0xcd, 0x7a, 0x96, 0xe0, 0xc7, 0x60, 0x89, 0x71, 0xcc, 0x7b, - 0xe2, 0x92, 0x0b, 0xc2, 0x7b, 0xf3, 0x12, 0x4a, 0x50, 0xfc, 0x22, 0x45, 0x6b, 0xa4, 0xc8, 0xf4, - 0xdf, 0x35, 0xb0, 0x31, 0x09, 0xb9, 0x05, 0x75, 0xf7, 0xc7, 0xd5, 0x7d, 0x63, 0xce, 0x62, 0x66, - 0x28, 0xfc, 0xa7, 0x06, 0x5e, 0x9a, 0xaa, 0x5b, 0xbe, 0x7d, 0x62, 0x26, 0xf8, 0x13, 0x93, 0xe7, - 0x20, 0x7e, 0xcb, 0xe5, 0x4c, 0x38, 0x4a, 0xd9, 0x47, 0xa9, 0x28, 0xf8, 0x14, 0x14, 0x1d, 0xaf, - 0xeb, 0x78, 0x24, 0xb2, 0x1d, 0xc7, 0xfa, 0xa6, 0x5e, 0xdc, 0x49, 0x66, 0x29, 0xee, 0x46, 0x38, - 0xac, 0x16, 0xf7, 0x26, 0x58, 0xd0, 0x14, 0xaf, 0xfe, 0x47, 0x8a, 0x32, 0xf2, 0xb5, 0x7b, 0x1b, - 0xe4, 0xb1, 0xb4, 0x90, 0x40, 0x95, 0x31, 0x3a, 0xe9, 0x86, 0xb2, 0xa3, 0x91, 0x87, 0xec, 0x1b, - 0x79, 0x14, 0x2a, 0xd1, 0xb9, 0xfb, 0x46, 0x82, 0x12, 0x7d, 0x23, 0xd7, 0x48, 0x91, 0x89, 0x24, - 0xc4, 0x37, 0x8d, 0x3c, 0xcb, 0xcc, 0x78, 0x12, 0x07, 0xca, 0x8e, 0x46, 0x1e, 0xfa, 0x3f, 0x99, - 0x14, 0x81, 0x64, 0x03, 0x26, 0xaa, 0xb1, 0x65, 0x35, 0xf9, 0xa9, 0x6a, 0xec, 0x51, 0x35, 0x36, - 0xfc, 0x49, 0x03, 0x10, 0x8f, 0x28, 0x5a, 0x57, 0x0d, 0x1a, 0x75, 0x51, 0xf3, 0x46, 0x57, 0xc2, - 0x68, 0x4c, 0xf1, 0x44, 0x2f, 0x61, 0x59, 0xc5, 0x87, 0xd3, 0x0e, 0x28, 0x25, 0x38, 0xb4, 0x41, - 0x21, 0xb2, 0x36, 0x83, 0x80, 0x06, 0xea, 0x7a, 0xea, 0xd7, 0xe6, 0x22, 0x3d, 0xcd, 0x8a, 0xfc, - 0x2c, 0x8b, 0xa1, 0x97, 0xc3, 0x6a, 0x21, 0xb1, 0x8f, 0x92, 0xb4, 0x22, 0x8a, 0x4d, 0xe2, 0x28, - 0xd9, 0x9b, 0x45, 0xd9, 0x25, 0xb3, 0xa3, 0x24, 0x68, 0xcb, 0x4d, 0xf0, 0xf2, 0x8c, 0x63, 0xb9, - 0xd1, 0x7b, 0xf1, 0xbd, 0x06, 0x92, 0x31, 0xe0, 0x3e, 0xc8, 0x8a, 0xbf, 0x09, 0x6a, 0x90, 0x6c, - 0xcd, 0x37, 0x48, 0x4e, 0x1c, 0x97, 0xc4, 0xa3, 0x50, 0xac, 0x90, 0x64, 0x81, 0x6f, 0x82, 0x65, - 0x97, 0x30, 0x86, 0xdb, 0x2a, 0x72, 0xfc, 0x21, 0xd7, 0x8a, 0xcc, 0xe8, 0x6a, 0x5f, 0x7f, 0x04, - 0xee, 0xa6, 0x7c, 0x10, 0xc3, 0x2a, 0xc8, 0x59, 0xe2, 0xcb, 0x41, 0x26, 0x94, 0x33, 0x57, 0xc4, - 0x44, 0xd9, 0x11, 0x06, 0x14, 0xd9, 0xcd, 0xda, 0xf9, 0x45, 0x65, 0xe1, 0xd9, 0x45, 0x65, 0xe1, - 0xf9, 0x45, 0x65, 0xe1, 0xbb, 0xb0, 0xa2, 0x9d, 0x87, 0x15, 0xed, 0x59, 0x58, 0xd1, 0x9e, 0x87, - 0x15, 0xed, 0xaf, 0xb0, 0xa2, 0xfd, 0xf8, 0x77, 0x65, 0xe1, 0xd3, 0xc5, 0xfe, 0xf6, 0x7f, 0x01, - 0x00, 0x00, 0xff, 0xff, 0x5c, 0x59, 0x23, 0xb9, 0x2c, 0x0e, 0x00, 0x00, + // 1395 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x57, 0xcf, 0x6f, 0x1b, 0xc5, + 0x17, 0xcf, 0xc6, 0x76, 0x7e, 0x8c, 0x93, 0xc6, 0x99, 0xe4, 0xfb, 0xfd, 0xfa, 0x9b, 0x83, 0x37, + 0x5a, 0x2a, 0x08, 0x85, 0xae, 0x9b, 0x52, 0xaa, 0xaa, 0x52, 0x91, 0xe2, 0xc4, 0xa5, 0x11, 0x71, + 0x12, 0x8d, 0x4b, 0x85, 0x10, 0x20, 0x26, 0xbb, 0x13, 0x67, 0x1b, 0xef, 0xce, 0x76, 0x77, 0x6c, + 0xf0, 0x8d, 0x13, 0x37, 0x24, 0xb8, 0xf2, 0x57, 0x80, 0x04, 0x17, 0x8e, 0x9c, 0xca, 0xad, 0xe2, + 0xd4, 0xd3, 0x8a, 0x2e, 0x67, 0xb8, 0x71, 0xc9, 0x09, 0xcd, 0xec, 0xd8, 0xfb, 0xc3, 0xeb, 0x34, + 0xbd, 0xe4, 0xe6, 0x79, 0xef, 0x7d, 0x3e, 0xef, 0xbd, 0x79, 0x3f, 0x66, 0x0d, 0xde, 0x3b, 0xbd, + 0xe3, 0xeb, 0x16, 0xad, 0x9f, 0xf6, 0x8e, 0x88, 0xe7, 0x10, 0x46, 0xfc, 0x7a, 0x9f, 0x38, 0x26, + 0xf5, 0xea, 0x52, 0x81, 0x5d, 0xab, 0xee, 0x33, 0xea, 0xe1, 0x0e, 0xa9, 0xf7, 0x37, 0xeb, 0x1d, + 0xe2, 0x10, 0x0f, 0x33, 0x62, 0xea, 0xae, 0x47, 0x19, 0x85, 0xff, 0x89, 0xcc, 0x74, 0xec, 0x5a, + 0xba, 0x34, 0xd3, 0xfb, 0x9b, 0x6b, 0xd7, 0x3b, 0x16, 0x3b, 0xe9, 0x1d, 0xe9, 0x06, 0xb5, 0xeb, + 0x1d, 0xda, 0xa1, 0x75, 0x61, 0x7d, 0xd4, 0x3b, 0x16, 0x27, 0x71, 0x10, 0xbf, 0x22, 0x96, 0x35, + 0x2d, 0xe1, 0xcc, 0xa0, 0x5e, 0x9e, 0xa7, 0xb5, 0x5b, 0xb1, 0x8d, 0x8d, 0x8d, 0x13, 0xcb, 0x21, + 0xde, 0xa0, 0xee, 0x9e, 0x76, 0xb8, 0xc0, 0xaf, 0xdb, 0x84, 0xe1, 0x3c, 0x54, 0x7d, 0x12, 0xca, + 0xeb, 0x39, 0xcc, 0xb2, 0xc9, 0x18, 0xe0, 0xf6, 0xcb, 0x00, 0xbe, 0x71, 0x42, 0x6c, 0x9c, 0xc5, + 0x69, 0x3f, 0x2b, 0x60, 0x7e, 0xbb, 0xbd, 0xbb, 0xe3, 0x59, 0x7d, 0xe2, 0xc1, 0xcf, 0xc1, 0x1c, + 0x8f, 0xc8, 0xc4, 0x0c, 0x57, 0x95, 0x75, 0x65, 0xa3, 0x7c, 0xf3, 0x86, 0x1e, 0xdf, 0xd4, 0x88, + 0x58, 0x77, 0x4f, 0x3b, 0x5c, 0xe0, 0xeb, 0xdc, 0x5a, 0xef, 0x6f, 0xea, 0x07, 0x47, 0x8f, 0x89, + 0xc1, 0x5a, 0x84, 0xe1, 0x06, 0x7c, 0x1a, 0xa8, 0x53, 0x61, 0xa0, 0x82, 0x58, 0x86, 0x46, 0xac, + 0xf0, 0x3e, 0x28, 0xfa, 0x2e, 0x31, 0xaa, 0xd3, 0x82, 0xfd, 0xaa, 0x9e, 0x5b, 0x07, 0x7d, 0x14, + 0x51, 0xdb, 0x25, 0x46, 0x63, 0x41, 0x32, 0x16, 0xf9, 0x09, 0x09, 0xbc, 0xf6, 0x93, 0x02, 0x16, + 0x47, 0x56, 0x7b, 0x96, 0xcf, 0xe0, 0x27, 0x63, 0xb1, 0xeb, 0x17, 0x8b, 0x9d, 0xa3, 0x45, 0xe4, + 0x15, 0xe9, 0x67, 0x6e, 0x28, 0x49, 0xc4, 0xdd, 0x04, 0x25, 0x8b, 0x11, 0xdb, 0xaf, 0x4e, 0xaf, + 0x17, 0x36, 0xca, 0x37, 0xd7, 0x5f, 0x16, 0x78, 0x63, 0x51, 0x92, 0x95, 0x76, 0x39, 0x0c, 0x45, + 0x68, 0xed, 0x9f, 0xe9, 0x44, 0xd8, 0x3c, 0x1d, 0x78, 0x17, 0x5c, 0xc1, 0x8c, 0x61, 0xe3, 0x04, + 0x91, 0x27, 0x3d, 0xcb, 0x23, 0xa6, 0x08, 0x7e, 0xae, 0x01, 0xc3, 0x40, 0xbd, 0xb2, 0x95, 0xd2, + 0xa0, 0x8c, 0x25, 0xc7, 0xba, 0xd4, 0xdc, 0x75, 0x8e, 0xe9, 0x81, 0xd3, 0xa2, 0x3d, 0x87, 0x89, + 0x6b, 0x95, 0xd8, 0xc3, 0x94, 0x06, 0x65, 0x2c, 0xa1, 0x01, 0x56, 0xfb, 0xb4, 0xdb, 0xb3, 0xc9, + 0x9e, 0x75, 0x4c, 0x8c, 0x81, 0xd1, 0x25, 0x2d, 0x6a, 0x12, 0xbf, 0x5a, 0x58, 0x2f, 0x6c, 0xcc, + 0x37, 0xea, 0x61, 0xa0, 0xae, 0x3e, 0xca, 0xd1, 0x9f, 0x05, 0xea, 0x4a, 0x8e, 0x1c, 0xe5, 0x92, + 0xc1, 0x7b, 0x60, 0x49, 0x5e, 0xce, 0x36, 0x76, 0xb1, 0x61, 0xb1, 0x41, 0xb5, 0x28, 0x22, 0x5c, + 0x09, 0x03, 0x75, 0xa9, 0x9d, 0x56, 0xa1, 0xac, 0x2d, 0x7c, 0x00, 0x16, 0x8f, 0xfd, 0xf7, 0x3d, + 0xda, 0x73, 0x0f, 0x69, 0xd7, 0x32, 0x06, 0xd5, 0xd2, 0xba, 0xb2, 0x31, 0xdf, 0xd0, 0xc2, 0x40, + 0x5d, 0xbc, 0xdf, 0x4e, 0x28, 0xce, 0xb2, 0x02, 0x94, 0x06, 0x6a, 0x3f, 0x2a, 0x60, 0x76, 0xbb, + 0xbd, 0xbb, 0x4f, 0x4d, 0x72, 0x09, 0x4d, 0xbe, 0x93, 0x6a, 0x72, 0x6d, 0x72, 0xaf, 0xf0, 0x78, + 0x26, 0xb6, 0xf8, 0xdf, 0x51, 0x8b, 0x73, 0x1b, 0x39, 0x9e, 0xeb, 0xa0, 0xe8, 0x60, 0x9b, 0x88, + 0xa8, 0xe7, 0x63, 0xcc, 0x3e, 0xb6, 0x09, 0x12, 0x1a, 0xf8, 0x3a, 0x98, 0x71, 0xa8, 0x49, 0x76, + 0x77, 0x84, 0xef, 0xf9, 0xc6, 0x15, 0x69, 0x33, 0xb3, 0x2f, 0xa4, 0x48, 0x6a, 0xe1, 0x2d, 0xb0, + 0xc0, 0xa8, 0x4b, 0xbb, 0xb4, 0x33, 0xf8, 0x80, 0x0c, 0x86, 0x55, 0xaf, 0x84, 0x81, 0xba, 0xf0, + 0x30, 0x21, 0x47, 0x29, 0x2b, 0xf8, 0x29, 0x28, 0xe3, 0x6e, 0x97, 0x1a, 0x98, 0xe1, 0xa3, 0x2e, + 0x11, 0xa5, 0x2c, 0xdf, 0xbc, 0x36, 0x21, 0xbd, 0xa8, 0x4b, 0xb8, 0x5f, 0x44, 0x7c, 0xda, 0xf3, + 0x0c, 0xe2, 0x37, 0x96, 0xc2, 0x40, 0x2d, 0x6f, 0xc5, 0x14, 0x28, 0xc9, 0xa7, 0xfd, 0xa0, 0x80, + 0xb2, 0x4c, 0xf8, 0x12, 0x26, 0x7a, 0x3b, 0x3d, 0xd1, 0xb5, 0xf3, 0xab, 0x34, 0x61, 0x9e, 0x3f, + 0x1b, 0x45, 0x2c, 0x86, 0xf9, 0x00, 0xcc, 0x9a, 0xa2, 0x54, 0x7e, 0x55, 0x11, 0xac, 0x57, 0xcf, + 0x67, 0x95, 0xbb, 0x62, 0x49, 0x72, 0xcf, 0x46, 0x67, 0x1f, 0x0d, 0x59, 0xb4, 0x6f, 0x66, 0xc0, + 0xc2, 0x70, 0x4c, 0xba, 0xd8, 0xf7, 0x2f, 0xa1, 0x79, 0xdf, 0x05, 0x65, 0xd7, 0xa3, 0x7d, 0xcb, + 0xb7, 0xa8, 0x43, 0x3c, 0xd9, 0x47, 0x2b, 0x12, 0x52, 0x3e, 0x8c, 0x55, 0x28, 0x69, 0x07, 0x3b, + 0x00, 0xb8, 0xd8, 0xc3, 0x36, 0x61, 0x3c, 0xfb, 0x82, 0xc8, 0xfe, 0x9d, 0x09, 0xd9, 0x27, 0x33, + 0xd2, 0x0f, 0x47, 0xa8, 0xa6, 0xc3, 0xbc, 0x41, 0x1c, 0x5d, 0xac, 0x40, 0x09, 0x6a, 0x78, 0x0a, + 0x16, 0x3d, 0x62, 0x74, 0xb1, 0x65, 0xcb, 0xa5, 0x50, 0x14, 0x11, 0x36, 0xf9, 0x52, 0x40, 0x49, + 0xc5, 0x59, 0xa0, 0xde, 0x18, 0x7f, 0xa0, 0xf5, 0x43, 0xe2, 0xf9, 0x96, 0xcf, 0x88, 0xc3, 0xa2, + 0x0e, 0x4d, 0x61, 0x50, 0x9a, 0x9b, 0xcf, 0x89, 0xcd, 0xd7, 0xe5, 0x81, 0xcb, 0x2c, 0xea, 0xf8, + 0xd5, 0x52, 0x3c, 0x27, 0xad, 0x84, 0x1c, 0xa5, 0xac, 0xe0, 0x1e, 0x58, 0xe5, 0x7d, 0xfd, 0x45, + 0xe4, 0xa0, 0xf9, 0xa5, 0x8b, 0x1d, 0x7e, 0x4b, 0xd5, 0x19, 0xb1, 0xfb, 0xaa, 0x7c, 0xb7, 0x6e, + 0xe5, 0xe8, 0x51, 0x2e, 0x0a, 0x7e, 0x04, 0x96, 0xa3, 0xe5, 0xda, 0xb0, 0x1c, 0xd3, 0x72, 0x3a, + 0x7c, 0xb5, 0x56, 0x67, 0x45, 0xd2, 0xd7, 0xc2, 0x40, 0x5d, 0x7e, 0x94, 0x55, 0x9e, 0xe5, 0x09, + 0xd1, 0x38, 0x09, 0x7c, 0x02, 0x96, 0x85, 0x47, 0x62, 0xca, 0xa1, 0xb7, 0x88, 0x5f, 0x9d, 0x13, + 0xa5, 0xdb, 0x48, 0x96, 0x8e, 0x5f, 0x1d, 0xaf, 0xdb, 0x70, 0x35, 0xb4, 0x49, 0x97, 0x18, 0x8c, + 0x7a, 0x0f, 0x89, 0x67, 0x37, 0xfe, 0x2f, 0xeb, 0xb5, 0xbc, 0x95, 0xa5, 0x42, 0xe3, 0xec, 0x6b, + 0xf7, 0xc0, 0x52, 0xa6, 0xe0, 0xb0, 0x02, 0x0a, 0xa7, 0x64, 0x10, 0x2d, 0x35, 0xc4, 0x7f, 0xc2, + 0x55, 0x50, 0xea, 0xe3, 0x6e, 0x8f, 0x44, 0xcd, 0x87, 0xa2, 0xc3, 0xdd, 0xe9, 0x3b, 0x8a, 0xf6, + 0x8b, 0x02, 0x2a, 0xc9, 0xee, 0xb9, 0x84, 0x3d, 0xf1, 0x20, 0xbd, 0x27, 0x5e, 0xbb, 0x40, 0x4f, + 0x4f, 0x58, 0x16, 0xdf, 0x4f, 0x83, 0x4a, 0x54, 0x97, 0xe8, 0x5d, 0xb7, 0x89, 0xc3, 0x2e, 0x61, + 0xa0, 0x5b, 0xa9, 0xd7, 0xe8, 0xad, 0x73, 0xd7, 0x75, 0x1c, 0xd8, 0xa4, 0x67, 0x09, 0x7e, 0x08, + 0x66, 0x7c, 0x86, 0x59, 0x8f, 0x0f, 0x39, 0x27, 0xbc, 0x7e, 0x51, 0x42, 0x01, 0x8a, 0x5f, 0xa4, + 0xe8, 0x8c, 0x24, 0x99, 0xf6, 0xab, 0x02, 0x56, 0xb3, 0x90, 0x4b, 0xa8, 0xee, 0x5e, 0xba, 0xba, + 0x6f, 0x5c, 0x30, 0x99, 0x09, 0x15, 0xfe, 0x5d, 0x01, 0xff, 0x1d, 0xcb, 0x5b, 0xbc, 0x7d, 0x7c, + 0x27, 0xb8, 0x99, 0xcd, 0xb3, 0x1f, 0xbf, 0xe5, 0x62, 0x27, 0x1c, 0xe6, 0xe8, 0x51, 0x2e, 0x0a, + 0x3e, 0x06, 0x15, 0xcb, 0xe9, 0x5a, 0x0e, 0x89, 0x64, 0xed, 0xb8, 0xbe, 0xb9, 0x83, 0x9b, 0x65, + 0x16, 0xc5, 0x5d, 0x0d, 0x03, 0xb5, 0xb2, 0x9b, 0x61, 0x41, 0x63, 0xbc, 0xda, 0x6f, 0x39, 0x95, + 0x11, 0xaf, 0xdd, 0xdb, 0x60, 0x2e, 0xfa, 0x20, 0x25, 0x9e, 0x4c, 0x63, 0x74, 0xd3, 0x5b, 0x52, + 0x8e, 0x46, 0x16, 0xa2, 0x6f, 0xc4, 0x55, 0xc8, 0x40, 0x2f, 0xdc, 0x37, 0x02, 0x94, 0xe8, 0x1b, + 0x71, 0x46, 0x92, 0x8c, 0x07, 0xc1, 0xbf, 0x69, 0xc4, 0x5d, 0x16, 0xd2, 0x41, 0xec, 0x4b, 0x39, + 0x1a, 0x59, 0x68, 0x7f, 0x15, 0x72, 0x0a, 0x24, 0x1a, 0x30, 0x91, 0xcd, 0xf0, 0x13, 0x3c, 0x9b, + 0x8d, 0x39, 0xca, 0xc6, 0x84, 0xdf, 0x29, 0x00, 0xe2, 0x11, 0x45, 0x6b, 0xd8, 0xa0, 0x51, 0x17, + 0x35, 0x5f, 0x69, 0x24, 0xf4, 0xad, 0x31, 0x9e, 0xe8, 0x25, 0x5c, 0x93, 0xfe, 0xe1, 0xb8, 0x01, + 0xca, 0x71, 0x0e, 0x4d, 0x50, 0x8e, 0xa4, 0x4d, 0xcf, 0xa3, 0x9e, 0x1c, 0x4f, 0xed, 0xdc, 0x58, + 0x84, 0x65, 0xa3, 0x26, 0x3e, 0xcb, 0x62, 0xe8, 0x59, 0xa0, 0x96, 0x13, 0x7a, 0x94, 0xa4, 0xe5, + 0x5e, 0x4c, 0x12, 0x7b, 0x29, 0xbe, 0x9a, 0x97, 0x1d, 0x32, 0xd9, 0x4b, 0x82, 0x76, 0xad, 0x09, + 0xfe, 0x37, 0xe1, 0x5a, 0x5e, 0xe9, 0xbd, 0xf8, 0x5a, 0x01, 0x49, 0x1f, 0x70, 0x0f, 0x14, 0xf9, + 0xbf, 0x61, 0xb9, 0x48, 0xae, 0x5d, 0x6c, 0x91, 0x3c, 0xb4, 0x6c, 0x12, 0xaf, 0x42, 0x7e, 0x42, + 0x82, 0x05, 0xbe, 0x09, 0x66, 0x6d, 0xe2, 0xfb, 0xb8, 0x23, 0x3d, 0xc7, 0x1f, 0x72, 0xad, 0x48, + 0x8c, 0x86, 0x7a, 0xed, 0x36, 0x58, 0xc9, 0xf9, 0x20, 0x86, 0x2a, 0x28, 0x19, 0xe2, 0x8f, 0x1b, + 0x0f, 0xa8, 0xd4, 0x98, 0xe7, 0x1b, 0x65, 0x5b, 0xfc, 0x5f, 0x8b, 0xe4, 0x8d, 0x8d, 0xa7, 0x2f, + 0x6a, 0x53, 0xcf, 0x5e, 0xd4, 0xa6, 0x9e, 0xbf, 0xa8, 0x4d, 0x7d, 0x15, 0xd6, 0x94, 0xa7, 0x61, + 0x4d, 0x79, 0x16, 0xd6, 0x94, 0xe7, 0x61, 0x4d, 0xf9, 0x23, 0xac, 0x29, 0xdf, 0xfe, 0x59, 0x9b, + 0xfa, 0x78, 0xba, 0xbf, 0xf9, 0x6f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x9e, 0x83, 0x24, 0x44, 0x13, + 0x11, 0x00, 0x00, +} + +func (m *CSIDriver) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CSIDriver) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CSIDriver) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + { + size, err := m.Spec.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *CSIDriverList) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CSIDriverList) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CSIDriverList) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + { + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *CSIDriverSpec) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CSIDriverSpec) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CSIDriverSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.FSGroupPolicy != nil { + i -= len(*m.FSGroupPolicy) + copy(dAtA[i:], *m.FSGroupPolicy) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.FSGroupPolicy))) + i-- + dAtA[i] = 0x2a + } + if m.StorageCapacity != nil { + i-- + if *m.StorageCapacity { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x20 + } + if len(m.VolumeLifecycleModes) > 0 { + for iNdEx := len(m.VolumeLifecycleModes) - 1; iNdEx >= 0; iNdEx-- { + i -= len(m.VolumeLifecycleModes[iNdEx]) + copy(dAtA[i:], m.VolumeLifecycleModes[iNdEx]) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.VolumeLifecycleModes[iNdEx]))) + i-- + dAtA[i] = 0x1a + } + } + if m.PodInfoOnMount != nil { + i-- + if *m.PodInfoOnMount { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x10 + } + if m.AttachRequired != nil { + i-- + if *m.AttachRequired { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x8 + } + return len(dAtA) - i, nil } func (m *CSINode) Marshal() (dAtA []byte, err error) { @@ -1190,6 +1448,64 @@ func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { dAtA[offset] = uint8(v) return base } +func (m *CSIDriver) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = m.Spec.Size() + n += 1 + l + sovGenerated(uint64(l)) + return n +} + +func (m *CSIDriverList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + +func (m *CSIDriverSpec) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + if m.AttachRequired != nil { + n += 2 + } + if m.PodInfoOnMount != nil { + n += 2 + } + if len(m.VolumeLifecycleModes) > 0 { + for _, s := range m.VolumeLifecycleModes { + l = len(s) + n += 1 + l + sovGenerated(uint64(l)) + } + } + if m.StorageCapacity != nil { + n += 2 + } + if m.FSGroupPolicy != nil { + l = len(*m.FSGroupPolicy) + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + func (m *CSINode) Size() (n int) { if m == nil { return 0 @@ -1440,11 +1756,52 @@ func sovGenerated(x uint64) (n int) { func sozGenerated(x uint64) (n int) { return sovGenerated(uint64((x << 1) ^ uint64((int64(x) >> 63)))) } -func (this *CSINode) String() string { +func (this *CSIDriver) String() string { if this == nil { return "nil" } - s := strings.Join([]string{`&CSINode{`, + s := strings.Join([]string{`&CSIDriver{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "CSIDriverSpec", "CSIDriverSpec", 1), `&`, ``, 1) + `,`, + `}`, + }, "") + return s +} +func (this *CSIDriverList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]CSIDriver{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "CSIDriver", "CSIDriver", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&CSIDriverList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} +func (this *CSIDriverSpec) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&CSIDriverSpec{`, + `AttachRequired:` + valueToStringGenerated(this.AttachRequired) + `,`, + `PodInfoOnMount:` + valueToStringGenerated(this.PodInfoOnMount) + `,`, + `VolumeLifecycleModes:` + fmt.Sprintf("%v", this.VolumeLifecycleModes) + `,`, + `StorageCapacity:` + valueToStringGenerated(this.StorageCapacity) + `,`, + `FSGroupPolicy:` + valueToStringGenerated(this.FSGroupPolicy) + `,`, + `}`, + }, "") + return s +} +func (this *CSINode) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&CSINode{`, `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, `Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "CSINodeSpec", "CSINodeSpec", 1), `&`, ``, 1) + `,`, `}`, @@ -1607,44 +1964,464 @@ func (this *VolumeAttachmentStatus) String() string { for _, k := range keysForAttachmentMetadata { mapStringForAttachmentMetadata += fmt.Sprintf("%v: %v,", k, this.AttachmentMetadata[k]) } - mapStringForAttachmentMetadata += "}" - s := strings.Join([]string{`&VolumeAttachmentStatus{`, - `Attached:` + fmt.Sprintf("%v", this.Attached) + `,`, - `AttachmentMetadata:` + mapStringForAttachmentMetadata + `,`, - `AttachError:` + strings.Replace(this.AttachError.String(), "VolumeError", "VolumeError", 1) + `,`, - `DetachError:` + strings.Replace(this.DetachError.String(), "VolumeError", "VolumeError", 1) + `,`, - `}`, - }, "") - return s -} -func (this *VolumeError) String() string { - if this == nil { - return "nil" + mapStringForAttachmentMetadata += "}" + s := strings.Join([]string{`&VolumeAttachmentStatus{`, + `Attached:` + fmt.Sprintf("%v", this.Attached) + `,`, + `AttachmentMetadata:` + mapStringForAttachmentMetadata + `,`, + `AttachError:` + strings.Replace(this.AttachError.String(), "VolumeError", "VolumeError", 1) + `,`, + `DetachError:` + strings.Replace(this.DetachError.String(), "VolumeError", "VolumeError", 1) + `,`, + `}`, + }, "") + return s +} +func (this *VolumeError) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&VolumeError{`, + `Time:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.Time), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, + `Message:` + fmt.Sprintf("%v", this.Message) + `,`, + `}`, + }, "") + return s +} +func (this *VolumeNodeResources) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&VolumeNodeResources{`, + `Count:` + valueToStringGenerated(this.Count) + `,`, + `}`, + }, "") + return s +} +func valueToStringGenerated(v interface{}) string { + rv := reflect.ValueOf(v) + if rv.IsNil() { + return "nil" + } + pv := reflect.Indirect(rv).Interface() + return fmt.Sprintf("*%v", pv) +} +func (m *CSIDriver) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CSIDriver: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CSIDriver: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *CSIDriverList) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CSIDriverList: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CSIDriverList: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Items = append(m.Items, CSIDriver{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF } - s := strings.Join([]string{`&VolumeError{`, - `Time:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.Time), "Time", "v1.Time", 1), `&`, ``, 1) + `,`, - `Message:` + fmt.Sprintf("%v", this.Message) + `,`, - `}`, - }, "") - return s + return nil } -func (this *VolumeNodeResources) String() string { - if this == nil { - return "nil" +func (m *CSIDriverSpec) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: CSIDriverSpec: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: CSIDriverSpec: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field AttachRequired", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + b := bool(v != 0) + m.AttachRequired = &b + case 2: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field PodInfoOnMount", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + b := bool(v != 0) + m.PodInfoOnMount = &b + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field VolumeLifecycleModes", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.VolumeLifecycleModes = append(m.VolumeLifecycleModes, VolumeLifecycleMode(dAtA[iNdEx:postIndex])) + iNdEx = postIndex + case 4: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field StorageCapacity", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + b := bool(v != 0) + m.StorageCapacity = &b + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field FSGroupPolicy", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := FSGroupPolicy(dAtA[iNdEx:postIndex]) + m.FSGroupPolicy = &s + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } } - s := strings.Join([]string{`&VolumeNodeResources{`, - `Count:` + valueToStringGenerated(this.Count) + `,`, - `}`, - }, "") - return s -} -func valueToStringGenerated(v interface{}) string { - rv := reflect.ValueOf(v) - if rv.IsNil() { - return "nil" + + if iNdEx > l { + return io.ErrUnexpectedEOF } - pv := reflect.Indirect(rv).Interface() - return fmt.Sprintf("*%v", pv) + return nil } func (m *CSINode) Unmarshal(dAtA []byte) error { l := len(dAtA) @@ -3685,6 +4462,7 @@ func (m *VolumeNodeResources) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -3716,10 +4494,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -3740,55 +4516,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/storage/v1/generated.proto b/vendor/k8s.io/api/storage/v1/generated.proto index e5004c84..a3526ca4 100644 --- a/vendor/k8s.io/api/storage/v1/generated.proto +++ b/vendor/k8s.io/api/storage/v1/generated.proto @@ -29,6 +29,125 @@ import "k8s.io/apimachinery/pkg/runtime/schema/generated.proto"; // Package-wide variables from generator "generated". option go_package = "v1"; +// CSIDriver captures information about a Container Storage Interface (CSI) +// volume driver deployed on the cluster. +// Kubernetes attach detach controller uses this object to determine whether attach is required. +// Kubelet uses this object to determine whether pod information needs to be passed on mount. +// CSIDriver objects are non-namespaced. +message CSIDriver { + // Standard object metadata. + // metadata.Name indicates the name of the CSI driver that this object + // refers to; it MUST be the same name returned by the CSI GetPluginName() + // call for that driver. + // The driver name must be 63 characters or less, beginning and ending with + // an alphanumeric character ([a-z0-9A-Z]) with dashes (-), dots (.), and + // alphanumerics between. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // Specification of the CSI Driver. + optional CSIDriverSpec spec = 2; +} + +// CSIDriverList is a collection of CSIDriver objects. +message CSIDriverList { + // Standard list metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; + + // items is the list of CSIDriver + repeated CSIDriver items = 2; +} + +// CSIDriverSpec is the specification of a CSIDriver. +message CSIDriverSpec { + // attachRequired indicates this CSI volume driver requires an attach + // operation (because it implements the CSI ControllerPublishVolume() + // method), and that the Kubernetes attach detach controller should call + // the attach volume interface which checks the volumeattachment status + // and waits until the volume is attached before proceeding to mounting. + // The CSI external-attacher coordinates with CSI volume driver and updates + // the volumeattachment status when the attach operation is complete. + // If the CSIDriverRegistry feature gate is enabled and the value is + // specified to false, the attach operation will be skipped. + // Otherwise the attach operation will be called. + // +optional + optional bool attachRequired = 1; + + // If set to true, podInfoOnMount indicates this CSI volume driver + // requires additional pod information (like podName, podUID, etc.) during + // mount operations. + // If set to false, pod information will not be passed on mount. + // Default is false. + // The CSI driver specifies podInfoOnMount as part of driver deployment. + // If true, Kubelet will pass pod information as VolumeContext in the CSI + // NodePublishVolume() calls. + // The CSI driver is responsible for parsing and validating the information + // passed in as VolumeContext. + // The following VolumeConext will be passed if podInfoOnMount is set to true. + // This list might grow, but the prefix will be used. + // "csi.storage.k8s.io/pod.name": pod.Name + // "csi.storage.k8s.io/pod.namespace": pod.Namespace + // "csi.storage.k8s.io/pod.uid": string(pod.UID) + // "csi.storage.k8s.io/ephemeral": "true" iff the volume is an ephemeral inline volume + // defined by a CSIVolumeSource, otherwise "false" + // + // "csi.storage.k8s.io/ephemeral" is a new feature in Kubernetes 1.16. It is only + // required for drivers which support both the "Persistent" and "Ephemeral" VolumeLifecycleMode. + // Other drivers can leave pod info disabled and/or ignore this field. + // As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when + // deployed on such a cluster and the deployment determines which mode that is, for example + // via a command line parameter of the driver. + // +optional + optional bool podInfoOnMount = 2; + + // volumeLifecycleModes defines what kind of volumes this CSI volume driver supports. + // The default if the list is empty is "Persistent", which is the usage + // defined by the CSI specification and implemented in Kubernetes via the usual + // PV/PVC mechanism. + // The other mode is "Ephemeral". In this mode, volumes are defined inline + // inside the pod spec with CSIVolumeSource and their lifecycle is tied to + // the lifecycle of that pod. A driver has to be aware of this + // because it is only going to get a NodePublishVolume call for such a volume. + // For more information about implementing this mode, see + // https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html + // A driver can support one or more of these modes and + // more modes may be added in the future. + // This field is beta. + // +optional + // +listType=set + repeated string volumeLifecycleModes = 3; + + // If set to true, storageCapacity indicates that the CSI + // volume driver wants pod scheduling to consider the storage + // capacity that the driver deployment will report by creating + // CSIStorageCapacity objects with capacity information. + // + // The check can be enabled immediately when deploying a driver. + // In that case, provisioning new volumes with late binding + // will pause until the driver deployment has published + // some suitable CSIStorageCapacity object. + // + // Alternatively, the driver can be deployed with the field + // unset or false and it can be flipped later when storage + // capacity information has been published. + // + // This is an alpha field and only available when the CSIStorageCapacity + // feature is enabled. The default is false. + // + // +optional + optional bool storageCapacity = 4; + + // Defines if the underlying volume supports changing ownership and + // permission of the volume before being mounted. + // Refer to the specific FSGroupPolicy values for additional details. + // This field is alpha-level, and is only honored by servers + // that enable the CSIVolumeFSGroupPolicy feature gate. + // +optional + optional string fsGroupPolicy = 5; +} + // CSINode holds information about all CSI drivers installed on a node. // CSI drivers do not need to create the CSINode object directly. As long as // they use the node-driver-registrar sidecar container, the kubelet will diff --git a/vendor/k8s.io/api/storage/v1/register.go b/vendor/k8s.io/api/storage/v1/register.go index 67493fd0..1a2f83d1 100644 --- a/vendor/k8s.io/api/storage/v1/register.go +++ b/vendor/k8s.io/api/storage/v1/register.go @@ -52,6 +52,9 @@ func addKnownTypes(scheme *runtime.Scheme) error { &CSINode{}, &CSINodeList{}, + + &CSIDriver{}, + &CSIDriverList{}, ) metav1.AddToGroupVersion(scheme, SchemeGroupVersion) diff --git a/vendor/k8s.io/api/storage/v1/types.go b/vendor/k8s.io/api/storage/v1/types.go index 86cb78b6..27e06deb 100644 --- a/vendor/k8s.io/api/storage/v1/types.go +++ b/vendor/k8s.io/api/storage/v1/types.go @@ -221,6 +221,191 @@ type VolumeError struct { // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// CSIDriver captures information about a Container Storage Interface (CSI) +// volume driver deployed on the cluster. +// Kubernetes attach detach controller uses this object to determine whether attach is required. +// Kubelet uses this object to determine whether pod information needs to be passed on mount. +// CSIDriver objects are non-namespaced. +type CSIDriver struct { + metav1.TypeMeta `json:",inline"` + + // Standard object metadata. + // metadata.Name indicates the name of the CSI driver that this object + // refers to; it MUST be the same name returned by the CSI GetPluginName() + // call for that driver. + // The driver name must be 63 characters or less, beginning and ending with + // an alphanumeric character ([a-z0-9A-Z]) with dashes (-), dots (.), and + // alphanumerics between. + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Specification of the CSI Driver. + Spec CSIDriverSpec `json:"spec" protobuf:"bytes,2,opt,name=spec"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// CSIDriverList is a collection of CSIDriver objects. +type CSIDriverList struct { + metav1.TypeMeta `json:",inline"` + + // Standard list metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // items is the list of CSIDriver + Items []CSIDriver `json:"items" protobuf:"bytes,2,rep,name=items"` +} + +// CSIDriverSpec is the specification of a CSIDriver. +type CSIDriverSpec struct { + // attachRequired indicates this CSI volume driver requires an attach + // operation (because it implements the CSI ControllerPublishVolume() + // method), and that the Kubernetes attach detach controller should call + // the attach volume interface which checks the volumeattachment status + // and waits until the volume is attached before proceeding to mounting. + // The CSI external-attacher coordinates with CSI volume driver and updates + // the volumeattachment status when the attach operation is complete. + // If the CSIDriverRegistry feature gate is enabled and the value is + // specified to false, the attach operation will be skipped. + // Otherwise the attach operation will be called. + // +optional + AttachRequired *bool `json:"attachRequired,omitempty" protobuf:"varint,1,opt,name=attachRequired"` + + // If set to true, podInfoOnMount indicates this CSI volume driver + // requires additional pod information (like podName, podUID, etc.) during + // mount operations. + // If set to false, pod information will not be passed on mount. + // Default is false. + // The CSI driver specifies podInfoOnMount as part of driver deployment. + // If true, Kubelet will pass pod information as VolumeContext in the CSI + // NodePublishVolume() calls. + // The CSI driver is responsible for parsing and validating the information + // passed in as VolumeContext. + // The following VolumeConext will be passed if podInfoOnMount is set to true. + // This list might grow, but the prefix will be used. + // "csi.storage.k8s.io/pod.name": pod.Name + // "csi.storage.k8s.io/pod.namespace": pod.Namespace + // "csi.storage.k8s.io/pod.uid": string(pod.UID) + // "csi.storage.k8s.io/ephemeral": "true" iff the volume is an ephemeral inline volume + // defined by a CSIVolumeSource, otherwise "false" + // + // "csi.storage.k8s.io/ephemeral" is a new feature in Kubernetes 1.16. It is only + // required for drivers which support both the "Persistent" and "Ephemeral" VolumeLifecycleMode. + // Other drivers can leave pod info disabled and/or ignore this field. + // As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when + // deployed on such a cluster and the deployment determines which mode that is, for example + // via a command line parameter of the driver. + // +optional + PodInfoOnMount *bool `json:"podInfoOnMount,omitempty" protobuf:"bytes,2,opt,name=podInfoOnMount"` + + // volumeLifecycleModes defines what kind of volumes this CSI volume driver supports. + // The default if the list is empty is "Persistent", which is the usage + // defined by the CSI specification and implemented in Kubernetes via the usual + // PV/PVC mechanism. + // The other mode is "Ephemeral". In this mode, volumes are defined inline + // inside the pod spec with CSIVolumeSource and their lifecycle is tied to + // the lifecycle of that pod. A driver has to be aware of this + // because it is only going to get a NodePublishVolume call for such a volume. + // For more information about implementing this mode, see + // https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html + // A driver can support one or more of these modes and + // more modes may be added in the future. + // This field is beta. + // +optional + // +listType=set + VolumeLifecycleModes []VolumeLifecycleMode `json:"volumeLifecycleModes,omitempty" protobuf:"bytes,3,opt,name=volumeLifecycleModes"` + + // If set to true, storageCapacity indicates that the CSI + // volume driver wants pod scheduling to consider the storage + // capacity that the driver deployment will report by creating + // CSIStorageCapacity objects with capacity information. + // + // The check can be enabled immediately when deploying a driver. + // In that case, provisioning new volumes with late binding + // will pause until the driver deployment has published + // some suitable CSIStorageCapacity object. + // + // Alternatively, the driver can be deployed with the field + // unset or false and it can be flipped later when storage + // capacity information has been published. + // + // This is an alpha field and only available when the CSIStorageCapacity + // feature is enabled. The default is false. + // + // +optional + StorageCapacity *bool `json:"storageCapacity,omitempty" protobuf:"bytes,4,opt,name=storageCapacity"` + + // Defines if the underlying volume supports changing ownership and + // permission of the volume before being mounted. + // Refer to the specific FSGroupPolicy values for additional details. + // This field is alpha-level, and is only honored by servers + // that enable the CSIVolumeFSGroupPolicy feature gate. + // +optional + FSGroupPolicy *FSGroupPolicy `json:"fsGroupPolicy,omitempty" protobuf:"bytes,5,opt,name=fsGroupPolicy"` +} + +// FSGroupPolicy specifies if a CSI Driver supports modifying +// volume ownership and permissions of the volume to be mounted. +// More modes may be added in the future. +type FSGroupPolicy string + +const ( + // ReadWriteOnceWithFSTypeFSGroupPolicy indicates that each volume will be examined + // to determine if the volume ownership and permissions + // should be modified. If a fstype is defined and the volume's access mode + // contains ReadWriteOnce, then the defined fsGroup will be applied. + // This mode should be defined if it's expected that the + // fsGroup may need to be modified depending on the pod's SecurityPolicy. + // This is the default behavior if no other FSGroupPolicy is defined. + ReadWriteOnceWithFSTypeFSGroupPolicy FSGroupPolicy = "ReadWriteOnceWithFSType" + + // FileFSGroupPolicy indicates that CSI driver supports volume ownership + // and permission change via fsGroup, and Kubernetes may use fsGroup + // to change permissions and ownership of the volume to match user requested fsGroup in + // the pod's SecurityPolicy regardless of fstype or access mode. + // This mode should be defined if the fsGroup is expected to always change on mount + FileFSGroupPolicy FSGroupPolicy = "File" + + // NoneFSGroupPolicy indicates that volumes will be mounted without performing + // any ownership or permission modifications, as the CSIDriver does not support + // these operations. + // This mode should be selected if the CSIDriver does not support fsGroup modifications, + // for example when Kubernetes cannot change ownership and permissions on a volume due + // to root-squash settings on a NFS volume. + NoneFSGroupPolicy FSGroupPolicy = "None" +) + +// VolumeLifecycleMode is an enumeration of possible usage modes for a volume +// provided by a CSI driver. More modes may be added in the future. +type VolumeLifecycleMode string + +const ( + // VolumeLifecyclePersistent explicitly confirms that the driver implements + // the full CSI spec. It is the default when CSIDriverSpec.VolumeLifecycleModes is not + // set. Such volumes are managed in Kubernetes via the persistent volume + // claim mechanism and have a lifecycle that is independent of the pods which + // use them. + VolumeLifecyclePersistent VolumeLifecycleMode = "Persistent" + + // VolumeLifecycleEphemeral indicates that the driver can be used for + // ephemeral inline volumes. Such volumes are specified inside the pod + // spec with a CSIVolumeSource and, as far as Kubernetes is concerned, have + // a lifecycle that is tied to the lifecycle of the pod. For example, such + // a volume might contain data that gets created specifically for that pod, + // like secrets. + // But how the volume actually gets created and managed is entirely up to + // the driver. It might also use reference counting to share the same volume + // instance among different pods if the CSIVolumeSource of those pods is + // identical. + VolumeLifecycleEphemeral VolumeLifecycleMode = "Ephemeral" +) + +// +genclient +// +genclient:nonNamespaced +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + // CSINode holds information about all CSI drivers installed on a node. // CSI drivers do not need to create the CSINode object directly. As long as // they use the node-driver-registrar sidecar container, the kubelet will diff --git a/vendor/k8s.io/api/storage/v1/types_swagger_doc_generated.go b/vendor/k8s.io/api/storage/v1/types_swagger_doc_generated.go index d6e3a162..606cda4d 100644 --- a/vendor/k8s.io/api/storage/v1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/storage/v1/types_swagger_doc_generated.go @@ -27,6 +27,39 @@ package v1 // Those methods can be generated by using hack/update-generated-swagger-docs.sh // AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. +var map_CSIDriver = map[string]string{ + "": "CSIDriver captures information about a Container Storage Interface (CSI) volume driver deployed on the cluster. Kubernetes attach detach controller uses this object to determine whether attach is required. Kubelet uses this object to determine whether pod information needs to be passed on mount. CSIDriver objects are non-namespaced.", + "metadata": "Standard object metadata. metadata.Name indicates the name of the CSI driver that this object refers to; it MUST be the same name returned by the CSI GetPluginName() call for that driver. The driver name must be 63 characters or less, beginning and ending with an alphanumeric character ([a-z0-9A-Z]) with dashes (-), dots (.), and alphanumerics between. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "spec": "Specification of the CSI Driver.", +} + +func (CSIDriver) SwaggerDoc() map[string]string { + return map_CSIDriver +} + +var map_CSIDriverList = map[string]string{ + "": "CSIDriverList is a collection of CSIDriver objects.", + "metadata": "Standard list metadata More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "items": "items is the list of CSIDriver", +} + +func (CSIDriverList) SwaggerDoc() map[string]string { + return map_CSIDriverList +} + +var map_CSIDriverSpec = map[string]string{ + "": "CSIDriverSpec is the specification of a CSIDriver.", + "attachRequired": "attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called.", + "podInfoOnMount": "If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. \"csi.storage.k8s.io/pod.name\": pod.Name \"csi.storage.k8s.io/pod.namespace\": pod.Namespace \"csi.storage.k8s.io/pod.uid\": string(pod.UID) \"csi.storage.k8s.io/ephemeral\": \"true\" iff the volume is an ephemeral inline volume\n defined by a CSIVolumeSource, otherwise \"false\"\n\n\"csi.storage.k8s.io/ephemeral\" is a new feature in Kubernetes 1.16. It is only required for drivers which support both the \"Persistent\" and \"Ephemeral\" VolumeLifecycleMode. Other drivers can leave pod info disabled and/or ignore this field. As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when deployed on such a cluster and the deployment determines which mode that is, for example via a command line parameter of the driver.", + "volumeLifecycleModes": "volumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is \"Persistent\", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is \"Ephemeral\". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future. This field is beta.", + "storageCapacity": "If set to true, storageCapacity indicates that the CSI volume driver wants pod scheduling to consider the storage capacity that the driver deployment will report by creating CSIStorageCapacity objects with capacity information.\n\nThe check can be enabled immediately when deploying a driver. In that case, provisioning new volumes with late binding will pause until the driver deployment has published some suitable CSIStorageCapacity object.\n\nAlternatively, the driver can be deployed with the field unset or false and it can be flipped later when storage capacity information has been published.\n\nThis is an alpha field and only available when the CSIStorageCapacity feature is enabled. The default is false.", + "fsGroupPolicy": "Defines if the underlying volume supports changing ownership and permission of the volume before being mounted. Refer to the specific FSGroupPolicy values for additional details. This field is alpha-level, and is only honored by servers that enable the CSIVolumeFSGroupPolicy feature gate.", +} + +func (CSIDriverSpec) SwaggerDoc() map[string]string { + return map_CSIDriverSpec +} + var map_CSINode = map[string]string{ "": "CSINode holds information about all CSI drivers installed on a node. CSI drivers do not need to create the CSINode object directly. As long as they use the node-driver-registrar sidecar container, the kubelet will automatically populate the CSINode object for the CSI driver as part of kubelet plugin registration. CSINode has the same name as a node. If the object is missing, it means either there are no CSI Drivers available on the node, or the Kubelet version is low enough that it doesn't create this object. CSINode has an OwnerReference that points to the corresponding node object.", "metadata": "metadata.name must be the Kubernetes node name.", diff --git a/vendor/k8s.io/api/storage/v1/zz_generated.deepcopy.go b/vendor/k8s.io/api/storage/v1/zz_generated.deepcopy.go index 76255a0a..5eb0225a 100644 --- a/vendor/k8s.io/api/storage/v1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/storage/v1/zz_generated.deepcopy.go @@ -25,6 +25,107 @@ import ( runtime "k8s.io/apimachinery/pkg/runtime" ) +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CSIDriver) DeepCopyInto(out *CSIDriver) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + in.Spec.DeepCopyInto(&out.Spec) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CSIDriver. +func (in *CSIDriver) DeepCopy() *CSIDriver { + if in == nil { + return nil + } + out := new(CSIDriver) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *CSIDriver) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CSIDriverList) DeepCopyInto(out *CSIDriverList) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ListMeta.DeepCopyInto(&out.ListMeta) + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]CSIDriver, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CSIDriverList. +func (in *CSIDriverList) DeepCopy() *CSIDriverList { + if in == nil { + return nil + } + out := new(CSIDriverList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *CSIDriverList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CSIDriverSpec) DeepCopyInto(out *CSIDriverSpec) { + *out = *in + if in.AttachRequired != nil { + in, out := &in.AttachRequired, &out.AttachRequired + *out = new(bool) + **out = **in + } + if in.PodInfoOnMount != nil { + in, out := &in.PodInfoOnMount, &out.PodInfoOnMount + *out = new(bool) + **out = **in + } + if in.VolumeLifecycleModes != nil { + in, out := &in.VolumeLifecycleModes, &out.VolumeLifecycleModes + *out = make([]VolumeLifecycleMode, len(*in)) + copy(*out, *in) + } + if in.StorageCapacity != nil { + in, out := &in.StorageCapacity, &out.StorageCapacity + *out = new(bool) + **out = **in + } + if in.FSGroupPolicy != nil { + in, out := &in.FSGroupPolicy, &out.FSGroupPolicy + *out = new(FSGroupPolicy) + **out = **in + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CSIDriverSpec. +func (in *CSIDriverSpec) DeepCopy() *CSIDriverSpec { + if in == nil { + return nil + } + out := new(CSIDriverSpec) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *CSINode) DeepCopyInto(out *CSINode) { *out = *in diff --git a/vendor/k8s.io/api/storage/v1alpha1/generated.pb.go b/vendor/k8s.io/api/storage/v1alpha1/generated.pb.go index 42324352..1b7767fd 100644 --- a/vendor/k8s.io/api/storage/v1alpha1/generated.pb.go +++ b/vendor/k8s.io/api/storage/v1alpha1/generated.pb.go @@ -27,6 +27,8 @@ import ( proto "github.com/gogo/protobuf/proto" github_com_gogo_protobuf_sortkeys "github.com/gogo/protobuf/sortkeys" v11 "k8s.io/api/core/v1" + resource "k8s.io/apimachinery/pkg/api/resource" + v1 "k8s.io/apimachinery/pkg/apis/meta/v1" math "math" math_bits "math/bits" @@ -43,12 +45,68 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package + +func (m *CSIStorageCapacity) Reset() { *m = CSIStorageCapacity{} } +func (*CSIStorageCapacity) ProtoMessage() {} +func (*CSIStorageCapacity) Descriptor() ([]byte, []int) { + return fileDescriptor_10f856db1e670dc4, []int{0} +} +func (m *CSIStorageCapacity) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CSIStorageCapacity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CSIStorageCapacity) XXX_Merge(src proto.Message) { + xxx_messageInfo_CSIStorageCapacity.Merge(m, src) +} +func (m *CSIStorageCapacity) XXX_Size() int { + return m.Size() +} +func (m *CSIStorageCapacity) XXX_DiscardUnknown() { + xxx_messageInfo_CSIStorageCapacity.DiscardUnknown(m) +} + +var xxx_messageInfo_CSIStorageCapacity proto.InternalMessageInfo + +func (m *CSIStorageCapacityList) Reset() { *m = CSIStorageCapacityList{} } +func (*CSIStorageCapacityList) ProtoMessage() {} +func (*CSIStorageCapacityList) Descriptor() ([]byte, []int) { + return fileDescriptor_10f856db1e670dc4, []int{1} +} +func (m *CSIStorageCapacityList) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *CSIStorageCapacityList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *CSIStorageCapacityList) XXX_Merge(src proto.Message) { + xxx_messageInfo_CSIStorageCapacityList.Merge(m, src) +} +func (m *CSIStorageCapacityList) XXX_Size() int { + return m.Size() +} +func (m *CSIStorageCapacityList) XXX_DiscardUnknown() { + xxx_messageInfo_CSIStorageCapacityList.DiscardUnknown(m) +} + +var xxx_messageInfo_CSIStorageCapacityList proto.InternalMessageInfo func (m *VolumeAttachment) Reset() { *m = VolumeAttachment{} } func (*VolumeAttachment) ProtoMessage() {} func (*VolumeAttachment) Descriptor() ([]byte, []int) { - return fileDescriptor_10f856db1e670dc4, []int{0} + return fileDescriptor_10f856db1e670dc4, []int{2} } func (m *VolumeAttachment) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -76,7 +134,7 @@ var xxx_messageInfo_VolumeAttachment proto.InternalMessageInfo func (m *VolumeAttachmentList) Reset() { *m = VolumeAttachmentList{} } func (*VolumeAttachmentList) ProtoMessage() {} func (*VolumeAttachmentList) Descriptor() ([]byte, []int) { - return fileDescriptor_10f856db1e670dc4, []int{1} + return fileDescriptor_10f856db1e670dc4, []int{3} } func (m *VolumeAttachmentList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -104,7 +162,7 @@ var xxx_messageInfo_VolumeAttachmentList proto.InternalMessageInfo func (m *VolumeAttachmentSource) Reset() { *m = VolumeAttachmentSource{} } func (*VolumeAttachmentSource) ProtoMessage() {} func (*VolumeAttachmentSource) Descriptor() ([]byte, []int) { - return fileDescriptor_10f856db1e670dc4, []int{2} + return fileDescriptor_10f856db1e670dc4, []int{4} } func (m *VolumeAttachmentSource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -132,7 +190,7 @@ var xxx_messageInfo_VolumeAttachmentSource proto.InternalMessageInfo func (m *VolumeAttachmentSpec) Reset() { *m = VolumeAttachmentSpec{} } func (*VolumeAttachmentSpec) ProtoMessage() {} func (*VolumeAttachmentSpec) Descriptor() ([]byte, []int) { - return fileDescriptor_10f856db1e670dc4, []int{3} + return fileDescriptor_10f856db1e670dc4, []int{5} } func (m *VolumeAttachmentSpec) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -160,7 +218,7 @@ var xxx_messageInfo_VolumeAttachmentSpec proto.InternalMessageInfo func (m *VolumeAttachmentStatus) Reset() { *m = VolumeAttachmentStatus{} } func (*VolumeAttachmentStatus) ProtoMessage() {} func (*VolumeAttachmentStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_10f856db1e670dc4, []int{4} + return fileDescriptor_10f856db1e670dc4, []int{6} } func (m *VolumeAttachmentStatus) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -188,7 +246,7 @@ var xxx_messageInfo_VolumeAttachmentStatus proto.InternalMessageInfo func (m *VolumeError) Reset() { *m = VolumeError{} } func (*VolumeError) ProtoMessage() {} func (*VolumeError) Descriptor() ([]byte, []int) { - return fileDescriptor_10f856db1e670dc4, []int{5} + return fileDescriptor_10f856db1e670dc4, []int{7} } func (m *VolumeError) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -214,6 +272,8 @@ func (m *VolumeError) XXX_DiscardUnknown() { var xxx_messageInfo_VolumeError proto.InternalMessageInfo func init() { + proto.RegisterType((*CSIStorageCapacity)(nil), "k8s.io.api.storage.v1alpha1.CSIStorageCapacity") + proto.RegisterType((*CSIStorageCapacityList)(nil), "k8s.io.api.storage.v1alpha1.CSIStorageCapacityList") proto.RegisterType((*VolumeAttachment)(nil), "k8s.io.api.storage.v1alpha1.VolumeAttachment") proto.RegisterType((*VolumeAttachmentList)(nil), "k8s.io.api.storage.v1alpha1.VolumeAttachmentList") proto.RegisterType((*VolumeAttachmentSource)(nil), "k8s.io.api.storage.v1alpha1.VolumeAttachmentSource") @@ -228,54 +288,172 @@ func init() { } var fileDescriptor_10f856db1e670dc4 = []byte{ - // 745 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x94, 0xcd, 0x6e, 0xd3, 0x40, - 0x10, 0xc7, 0xe3, 0x24, 0x6d, 0xd3, 0x0d, 0x1f, 0xd1, 0x2a, 0x82, 0x28, 0x48, 0x4e, 0x95, 0x53, - 0x40, 0x74, 0x4d, 0x0a, 0x42, 0x15, 0xb7, 0x58, 0xed, 0xa1, 0xa2, 0x2d, 0x68, 0x8b, 0x38, 0x00, - 0x07, 0x36, 0xf6, 0xe2, 0xb8, 0x89, 0x3f, 0xe4, 0x5d, 0x47, 0xea, 0x8d, 0x13, 0x67, 0x6e, 0xbc, - 0x01, 0xcf, 0x92, 0x1b, 0x15, 0xa7, 0x9e, 0x22, 0x6a, 0xde, 0x82, 0x0b, 0x68, 0xd7, 0x9b, 0xc4, - 0x24, 0x29, 0xb4, 0xbd, 0x79, 0x66, 0x67, 0x7e, 0x33, 0xf3, 0xdf, 0xf1, 0x82, 0x9d, 0xfe, 0x36, - 0x43, 0x6e, 0x60, 0xf4, 0xe3, 0x2e, 0x8d, 0x7c, 0xca, 0x29, 0x33, 0x86, 0xd4, 0xb7, 0x83, 0xc8, - 0x50, 0x07, 0x24, 0x74, 0x0d, 0xc6, 0x83, 0x88, 0x38, 0xd4, 0x18, 0xb6, 0xc9, 0x20, 0xec, 0x91, - 0xb6, 0xe1, 0x50, 0x9f, 0x46, 0x84, 0x53, 0x1b, 0x85, 0x51, 0xc0, 0x03, 0x78, 0x2f, 0x0d, 0x46, - 0x24, 0x74, 0x91, 0x0a, 0x46, 0x93, 0xe0, 0xfa, 0xa6, 0xe3, 0xf2, 0x5e, 0xdc, 0x45, 0x56, 0xe0, - 0x19, 0x4e, 0xe0, 0x04, 0x86, 0xcc, 0xe9, 0xc6, 0x1f, 0xa4, 0x25, 0x0d, 0xf9, 0x95, 0xb2, 0xea, - 0xcd, 0x4c, 0x61, 0x2b, 0x88, 0x44, 0xd5, 0xf9, 0x7a, 0xf5, 0x27, 0xb3, 0x18, 0x8f, 0x58, 0x3d, - 0xd7, 0xa7, 0xd1, 0x89, 0x11, 0xf6, 0x1d, 0xe1, 0x60, 0x86, 0x47, 0x39, 0x59, 0x96, 0x65, 0x5c, - 0x94, 0x15, 0xc5, 0x3e, 0x77, 0x3d, 0xba, 0x90, 0xf0, 0xf4, 0x7f, 0x09, 0xcc, 0xea, 0x51, 0x8f, - 0xcc, 0xe7, 0x35, 0xbf, 0xe6, 0x41, 0xe5, 0x75, 0x30, 0x88, 0x3d, 0xda, 0xe1, 0x9c, 0x58, 0x3d, - 0x8f, 0xfa, 0x1c, 0xbe, 0x07, 0x25, 0xd1, 0x98, 0x4d, 0x38, 0xa9, 0x69, 0x1b, 0x5a, 0xab, 0xbc, - 0xf5, 0x08, 0xcd, 0x64, 0x9b, 0xf2, 0x51, 0xd8, 0x77, 0x84, 0x83, 0x21, 0x11, 0x8d, 0x86, 0x6d, - 0xf4, 0xa2, 0x7b, 0x4c, 0x2d, 0x7e, 0x40, 0x39, 0x31, 0xe1, 0x68, 0xdc, 0xc8, 0x25, 0xe3, 0x06, - 0x98, 0xf9, 0xf0, 0x94, 0x0a, 0x8f, 0x40, 0x91, 0x85, 0xd4, 0xaa, 0xe5, 0x25, 0xbd, 0x8d, 0xfe, - 0x71, 0x29, 0x68, 0xbe, 0xbd, 0xa3, 0x90, 0x5a, 0xe6, 0x0d, 0x85, 0x2f, 0x0a, 0x0b, 0x4b, 0x18, - 0x7c, 0x0b, 0x56, 0x19, 0x27, 0x3c, 0x66, 0xb5, 0x82, 0xc4, 0x3e, 0xbe, 0x1a, 0x56, 0xa6, 0x9a, - 0xb7, 0x14, 0x78, 0x35, 0xb5, 0xb1, 0x42, 0x36, 0x47, 0x1a, 0xa8, 0xce, 0xa7, 0xec, 0xbb, 0x8c, - 0xc3, 0x77, 0x0b, 0x62, 0xa1, 0xcb, 0x89, 0x25, 0xb2, 0xa5, 0x54, 0x15, 0x55, 0xb2, 0x34, 0xf1, - 0x64, 0x84, 0xc2, 0x60, 0xc5, 0xe5, 0xd4, 0x63, 0xb5, 0xfc, 0x46, 0xa1, 0x55, 0xde, 0xda, 0xbc, - 0xd2, 0x48, 0xe6, 0x4d, 0x45, 0x5e, 0xd9, 0x13, 0x0c, 0x9c, 0xa2, 0x9a, 0xdf, 0x35, 0x70, 0x67, - 0x61, 0xfa, 0x20, 0x8e, 0x2c, 0x0a, 0xf7, 0x41, 0x35, 0xa4, 0x11, 0x73, 0x19, 0xa7, 0x3e, 0x4f, - 0x63, 0x0e, 0x89, 0x47, 0xe5, 0x60, 0xeb, 0x66, 0x2d, 0x19, 0x37, 0xaa, 0x2f, 0x97, 0x9c, 0xe3, - 0xa5, 0x59, 0xf0, 0x18, 0x54, 0x5c, 0x7f, 0xe0, 0xfa, 0x34, 0xf5, 0x1d, 0xcd, 0x6e, 0xbc, 0x95, - 0x9d, 0x43, 0xfc, 0x3a, 0x42, 0x90, 0x79, 0xb2, 0xbc, 0xe8, 0x6a, 0x32, 0x6e, 0x54, 0xf6, 0xe6, - 0x28, 0x78, 0x81, 0xdb, 0xfc, 0xb6, 0xe4, 0x7e, 0xc4, 0x01, 0x7c, 0x08, 0x4a, 0x44, 0x7a, 0x68, - 0xa4, 0xc6, 0x98, 0xea, 0xdd, 0x51, 0x7e, 0x3c, 0x8d, 0x90, 0x3b, 0x24, 0xa5, 0x50, 0x8d, 0x5e, - 0x71, 0x87, 0x64, 0x6a, 0x66, 0x87, 0xa4, 0x8d, 0x15, 0x52, 0xb4, 0xe2, 0x07, 0x76, 0xaa, 0x68, - 0xe1, 0xef, 0x56, 0x0e, 0x95, 0x1f, 0x4f, 0x23, 0x9a, 0xbf, 0x0b, 0x4b, 0xae, 0x49, 0x2e, 0x63, - 0x66, 0x26, 0x5b, 0xce, 0x54, 0x5a, 0x98, 0xc9, 0x9e, 0xce, 0x64, 0xc3, 0x2f, 0x1a, 0x80, 0x64, - 0x8a, 0x38, 0x98, 0x2c, 0x6b, 0xba, 0x51, 0xcf, 0xaf, 0xf1, 0x93, 0xa0, 0xce, 0x02, 0x6d, 0xd7, - 0xe7, 0xd1, 0x89, 0x59, 0x57, 0x5d, 0xc0, 0xc5, 0x00, 0xbc, 0xa4, 0x05, 0x78, 0x0c, 0xca, 0xa9, - 0x77, 0x37, 0x8a, 0x82, 0x48, 0xfd, 0xb6, 0xad, 0x4b, 0x74, 0x24, 0xe3, 0x4d, 0x3d, 0x19, 0x37, - 0xca, 0x9d, 0x19, 0xe0, 0xd7, 0xb8, 0x51, 0xce, 0x9c, 0xe3, 0x2c, 0x5c, 0xd4, 0xb2, 0xe9, 0xac, - 0x56, 0xf1, 0x3a, 0xb5, 0x76, 0xe8, 0xc5, 0xb5, 0x32, 0xf0, 0xfa, 0x2e, 0xb8, 0x7b, 0x81, 0x44, - 0xb0, 0x02, 0x0a, 0x7d, 0x7a, 0x92, 0x6e, 0x22, 0x16, 0x9f, 0xb0, 0x0a, 0x56, 0x86, 0x64, 0x10, - 0xa7, 0x1b, 0xb7, 0x8e, 0x53, 0xe3, 0x59, 0x7e, 0x5b, 0x6b, 0x7e, 0xd2, 0x40, 0xb6, 0x06, 0xdc, - 0x07, 0x45, 0xf1, 0x96, 0xab, 0x67, 0xe6, 0xc1, 0xe5, 0x9e, 0x99, 0x57, 0xae, 0x47, 0x67, 0xcf, - 0xa5, 0xb0, 0xb0, 0xa4, 0xc0, 0xfb, 0x60, 0xcd, 0xa3, 0x8c, 0x11, 0x47, 0x55, 0x36, 0x6f, 0xab, - 0xa0, 0xb5, 0x83, 0xd4, 0x8d, 0x27, 0xe7, 0x26, 0x1a, 0x9d, 0xeb, 0xb9, 0xd3, 0x73, 0x3d, 0x77, - 0x76, 0xae, 0xe7, 0x3e, 0x26, 0xba, 0x36, 0x4a, 0x74, 0xed, 0x34, 0xd1, 0xb5, 0xb3, 0x44, 0xd7, - 0x7e, 0x24, 0xba, 0xf6, 0xf9, 0xa7, 0x9e, 0x7b, 0x53, 0x9a, 0x08, 0xf7, 0x27, 0x00, 0x00, 0xff, - 0xff, 0xe8, 0x45, 0xe3, 0xba, 0xab, 0x07, 0x00, 0x00, + // 895 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x56, 0x4f, 0x6f, 0xe3, 0x44, + 0x14, 0x8f, 0x9b, 0x74, 0x37, 0x3b, 0x29, 0x10, 0x8d, 0xa2, 0x25, 0x0a, 0x92, 0x53, 0xe5, 0x14, + 0x10, 0x3b, 0xa6, 0x0b, 0x42, 0x2b, 0x6e, 0x75, 0xdb, 0x43, 0x45, 0x5b, 0x60, 0x52, 0x21, 0x04, + 0x1c, 0x98, 0x38, 0x0f, 0xc7, 0x4d, 0xfc, 0x47, 0x33, 0xe3, 0x4a, 0xb9, 0xc1, 0x85, 0x33, 0x37, + 0xbe, 0x01, 0x9f, 0xa5, 0x07, 0x24, 0x56, 0x9c, 0xf6, 0x14, 0x51, 0xf3, 0x2d, 0xb8, 0x80, 0x3c, + 0x9e, 0x38, 0x6e, 0x9c, 0x74, 0xb3, 0x7b, 0xd8, 0x9b, 0xe7, 0xcd, 0x7b, 0xbf, 0xdf, 0xfb, 0xf3, + 0x9b, 0x27, 0xa3, 0xe3, 0xc9, 0x33, 0x41, 0xbc, 0xd0, 0x9a, 0xc4, 0x43, 0xe0, 0x01, 0x48, 0x10, + 0xd6, 0x35, 0x04, 0xa3, 0x90, 0x5b, 0xfa, 0x82, 0x45, 0x9e, 0x25, 0x64, 0xc8, 0x99, 0x0b, 0xd6, + 0xf5, 0x01, 0x9b, 0x46, 0x63, 0x76, 0x60, 0xb9, 0x10, 0x00, 0x67, 0x12, 0x46, 0x24, 0xe2, 0xa1, + 0x0c, 0xf1, 0x7b, 0x99, 0x33, 0x61, 0x91, 0x47, 0xb4, 0x33, 0x59, 0x38, 0x77, 0x9e, 0xb8, 0x9e, + 0x1c, 0xc7, 0x43, 0xe2, 0x84, 0xbe, 0xe5, 0x86, 0x6e, 0x68, 0xa9, 0x98, 0x61, 0xfc, 0xa3, 0x3a, + 0xa9, 0x83, 0xfa, 0xca, 0xb0, 0x3a, 0xbd, 0x02, 0xb1, 0x13, 0xf2, 0x94, 0x75, 0x95, 0xaf, 0xf3, + 0xc9, 0xd2, 0xc7, 0x67, 0xce, 0xd8, 0x0b, 0x80, 0xcf, 0xac, 0x68, 0xe2, 0xaa, 0x20, 0x0e, 0x22, + 0x8c, 0xb9, 0x03, 0xaf, 0x14, 0x25, 0x2c, 0x1f, 0x24, 0x5b, 0xc7, 0x65, 0x6d, 0x8a, 0xe2, 0x71, + 0x20, 0x3d, 0xbf, 0x4c, 0xf3, 0xe9, 0xcb, 0x02, 0x84, 0x33, 0x06, 0x9f, 0xad, 0xc6, 0xf5, 0x7e, + 0xae, 0x22, 0x7c, 0x34, 0x38, 0x1d, 0x64, 0xfd, 0x3b, 0x62, 0x11, 0x73, 0x3c, 0x39, 0xc3, 0x3f, + 0xa0, 0x7a, 0x9a, 0xda, 0x88, 0x49, 0xd6, 0x36, 0xf6, 0x8d, 0x7e, 0xe3, 0xe9, 0x47, 0x64, 0xd9, + 0xee, 0x9c, 0x81, 0x44, 0x13, 0x37, 0x35, 0x08, 0x92, 0x7a, 0x93, 0xeb, 0x03, 0xf2, 0xc5, 0xf0, + 0x0a, 0x1c, 0x79, 0x0e, 0x92, 0xd9, 0xf8, 0x66, 0xde, 0xad, 0x24, 0xf3, 0x2e, 0x5a, 0xda, 0x68, + 0x8e, 0x8a, 0x3d, 0xb4, 0x17, 0x84, 0x23, 0xb8, 0x0c, 0xa3, 0x70, 0x1a, 0xba, 0xb3, 0xf6, 0x8e, + 0x62, 0xf9, 0x78, 0x3b, 0x96, 0x33, 0x36, 0x84, 0xe9, 0x00, 0xa6, 0xe0, 0xc8, 0x90, 0xdb, 0xcd, + 0x64, 0xde, 0xdd, 0xbb, 0x28, 0x80, 0xd1, 0x3b, 0xd0, 0xf8, 0x18, 0x35, 0xb5, 0x3e, 0x8e, 0xa6, + 0x4c, 0x88, 0x0b, 0xe6, 0x43, 0xbb, 0xba, 0x6f, 0xf4, 0x1f, 0xd9, 0x6d, 0x9d, 0x62, 0x73, 0xb0, + 0x72, 0x4f, 0x4b, 0x11, 0xf8, 0x1b, 0x54, 0x77, 0x74, 0x7b, 0xda, 0x35, 0x95, 0x2c, 0xb9, 0x2f, + 0x59, 0xb2, 0x50, 0x04, 0xf9, 0x2a, 0x66, 0x81, 0xf4, 0xe4, 0xcc, 0xde, 0x4b, 0xe6, 0xdd, 0xfa, + 0xa2, 0xc5, 0x34, 0x47, 0xeb, 0xfd, 0x61, 0xa0, 0xc7, 0xe5, 0x19, 0x9c, 0x79, 0x42, 0xe2, 0xef, + 0x4b, 0x73, 0x20, 0x5b, 0x76, 0xc8, 0x13, 0xd9, 0x14, 0x9a, 0xba, 0xc4, 0xfa, 0xc2, 0x52, 0x98, + 0xc1, 0x25, 0xda, 0xf5, 0x24, 0xf8, 0xa2, 0xbd, 0xb3, 0x5f, 0xed, 0x37, 0x9e, 0x5a, 0xe4, 0x9e, + 0x17, 0x45, 0xca, 0x19, 0xda, 0x6f, 0x69, 0xec, 0xdd, 0xd3, 0x14, 0x85, 0x66, 0x60, 0xbd, 0xdf, + 0x77, 0x50, 0xf3, 0xeb, 0x70, 0x1a, 0xfb, 0x70, 0x28, 0x25, 0x73, 0xc6, 0x3e, 0x04, 0xf2, 0x0d, + 0x08, 0x6a, 0x80, 0x6a, 0x22, 0x02, 0x47, 0x0b, 0xe9, 0xe0, 0xde, 0x5a, 0x56, 0xd3, 0x1b, 0x44, + 0xe0, 0xd8, 0x7b, 0x1a, 0xbe, 0x96, 0x9e, 0xa8, 0x02, 0xc3, 0xdf, 0xa1, 0x07, 0x42, 0x32, 0x19, + 0x0b, 0x25, 0x98, 0xbb, 0xfa, 0xdc, 0x02, 0x56, 0x85, 0xda, 0x6f, 0x6b, 0xe0, 0x07, 0xd9, 0x99, + 0x6a, 0xc8, 0xde, 0x8d, 0x81, 0x5a, 0xab, 0x21, 0x6f, 0x60, 0xea, 0xf4, 0xee, 0xd4, 0x9f, 0xbc, + 0x52, 0x49, 0x1b, 0x66, 0xfe, 0x97, 0x81, 0x1e, 0x97, 0xaa, 0x57, 0xf2, 0xc7, 0x67, 0xa8, 0x15, + 0x01, 0x17, 0x9e, 0x90, 0x10, 0xc8, 0xcc, 0x47, 0xbd, 0x40, 0x23, 0x7b, 0x81, 0xc9, 0xbc, 0xdb, + 0xfa, 0x72, 0xcd, 0x3d, 0x5d, 0x1b, 0x85, 0xaf, 0x50, 0xd3, 0x0b, 0xa6, 0x5e, 0x00, 0x99, 0x6d, + 0xb0, 0x9c, 0x78, 0xbf, 0x58, 0x47, 0xba, 0xc3, 0xd3, 0x86, 0xac, 0x22, 0xab, 0x41, 0xb7, 0xd2, + 0x17, 0x7f, 0xba, 0x82, 0x42, 0x4b, 0xb8, 0xbd, 0x3f, 0xd7, 0xcc, 0x27, 0xbd, 0xc0, 0x1f, 0xa2, + 0x3a, 0x53, 0x16, 0xe0, 0xba, 0x8c, 0xbc, 0xdf, 0x87, 0xda, 0x4e, 0x73, 0x0f, 0xa5, 0x21, 0xd5, + 0x8a, 0x35, 0x3b, 0x6e, 0x0b, 0x0d, 0xa9, 0xd0, 0x82, 0x86, 0xd4, 0x99, 0x6a, 0xc8, 0x34, 0x95, + 0x74, 0xd7, 0x15, 0x76, 0x5a, 0x9e, 0xca, 0x85, 0xb6, 0xd3, 0xdc, 0xa3, 0xf7, 0x5f, 0x75, 0xcd, + 0x98, 0x94, 0x18, 0x0b, 0x35, 0x8d, 0x54, 0x4d, 0xf5, 0x52, 0x4d, 0xa3, 0xbc, 0xa6, 0x11, 0xfe, + 0xcd, 0x40, 0x98, 0xe5, 0x10, 0xe7, 0x0b, 0xb1, 0x66, 0x8a, 0xfa, 0xfc, 0x35, 0x1e, 0x09, 0x39, + 0x2c, 0xa1, 0x9d, 0x04, 0x92, 0xcf, 0xec, 0x8e, 0xce, 0x02, 0x97, 0x1d, 0xe8, 0x9a, 0x14, 0xf0, + 0x15, 0x6a, 0x64, 0xd6, 0x13, 0xce, 0x43, 0xae, 0x9f, 0x6d, 0x7f, 0x8b, 0x8c, 0x94, 0xbf, 0x6d, + 0x26, 0xf3, 0x6e, 0xe3, 0x70, 0x09, 0xf0, 0xef, 0xbc, 0xdb, 0x28, 0xdc, 0xd3, 0x22, 0x78, 0xca, + 0x35, 0x82, 0x25, 0x57, 0xed, 0x75, 0xb8, 0x8e, 0x61, 0x33, 0x57, 0x01, 0xbc, 0x73, 0x82, 0xde, + 0xdd, 0xd0, 0x22, 0xdc, 0x44, 0xd5, 0x09, 0xcc, 0x32, 0x25, 0xd2, 0xf4, 0x13, 0xb7, 0xd0, 0xee, + 0x35, 0x9b, 0xc6, 0x99, 0xe2, 0x1e, 0xd1, 0xec, 0xf0, 0xd9, 0xce, 0x33, 0xa3, 0xf7, 0x8b, 0x81, + 0x8a, 0x1c, 0xf8, 0x0c, 0xd5, 0xd2, 0xdf, 0x03, 0xbd, 0x66, 0x3e, 0xd8, 0x6e, 0xcd, 0x5c, 0x7a, + 0x3e, 0x2c, 0xd7, 0x65, 0x7a, 0xa2, 0x0a, 0x05, 0xbf, 0x8f, 0x1e, 0xfa, 0x20, 0x04, 0x73, 0x35, + 0xb3, 0xfd, 0x8e, 0x76, 0x7a, 0x78, 0x9e, 0x99, 0xe9, 0xe2, 0xde, 0x26, 0x37, 0xb7, 0x66, 0xe5, + 0xf9, 0xad, 0x59, 0x79, 0x71, 0x6b, 0x56, 0x7e, 0x4a, 0x4c, 0xe3, 0x26, 0x31, 0x8d, 0xe7, 0x89, + 0x69, 0xbc, 0x48, 0x4c, 0xe3, 0xef, 0xc4, 0x34, 0x7e, 0xfd, 0xc7, 0xac, 0x7c, 0x5b, 0x5f, 0x34, + 0xee, 0xff, 0x00, 0x00, 0x00, 0xff, 0xff, 0x6e, 0xb9, 0x9d, 0xb3, 0x34, 0x0a, 0x00, 0x00, +} + +func (m *CSIStorageCapacity) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CSIStorageCapacity) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CSIStorageCapacity) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if m.Capacity != nil { + { + size, err := m.Capacity.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x22 + } + i -= len(m.StorageClassName) + copy(dAtA[i:], m.StorageClassName) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.StorageClassName))) + i-- + dAtA[i] = 0x1a + if m.NodeTopology != nil { + { + size, err := m.NodeTopology.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + { + size, err := m.ObjectMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + +func (m *CSIStorageCapacityList) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *CSIStorageCapacityList) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *CSIStorageCapacityList) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + if len(m.Items) > 0 { + for iNdEx := len(m.Items) - 1; iNdEx >= 0; iNdEx-- { + { + size, err := m.Items[iNdEx].MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x12 + } + } + { + size, err := m.ListMeta.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil } func (m *VolumeAttachment) Marshal() (dAtA []byte, err error) { @@ -591,6 +769,44 @@ func encodeVarintGenerated(dAtA []byte, offset int, v uint64) int { dAtA[offset] = uint8(v) return base } +func (m *CSIStorageCapacity) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ObjectMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if m.NodeTopology != nil { + l = m.NodeTopology.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + l = len(m.StorageClassName) + n += 1 + l + sovGenerated(uint64(l)) + if m.Capacity != nil { + l = m.Capacity.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + return n +} + +func (m *CSIStorageCapacityList) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = m.ListMeta.Size() + n += 1 + l + sovGenerated(uint64(l)) + if len(m.Items) > 0 { + for _, e := range m.Items { + l = e.Size() + n += 1 + l + sovGenerated(uint64(l)) + } + } + return n +} + func (m *VolumeAttachment) Size() (n int) { if m == nil { return 0 @@ -700,6 +916,35 @@ func sovGenerated(x uint64) (n int) { func sozGenerated(x uint64) (n int) { return sovGenerated(uint64((x << 1) ^ uint64((int64(x) >> 63)))) } +func (this *CSIStorageCapacity) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&CSIStorageCapacity{`, + `ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v1.ObjectMeta", 1), `&`, ``, 1) + `,`, + `NodeTopology:` + strings.Replace(fmt.Sprintf("%v", this.NodeTopology), "LabelSelector", "v1.LabelSelector", 1) + `,`, + `StorageClassName:` + fmt.Sprintf("%v", this.StorageClassName) + `,`, + `Capacity:` + strings.Replace(fmt.Sprintf("%v", this.Capacity), "Quantity", "resource.Quantity", 1) + `,`, + `}`, + }, "") + return s +} +func (this *CSIStorageCapacityList) String() string { + if this == nil { + return "nil" + } + repeatedStringForItems := "[]CSIStorageCapacity{" + for _, f := range this.Items { + repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "CSIStorageCapacity", "CSIStorageCapacity", 1), `&`, ``, 1) + "," + } + repeatedStringForItems += "}" + s := strings.Join([]string{`&CSIStorageCapacityList{`, + `ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v1.ListMeta", 1), `&`, ``, 1) + `,`, + `Items:` + repeatedStringForItems + `,`, + `}`, + }, "") + return s +} func (this *VolumeAttachment) String() string { if this == nil { return "nil" @@ -793,7 +1038,7 @@ func valueToStringGenerated(v interface{}) string { pv := reflect.Indirect(rv).Interface() return fmt.Sprintf("*%v", pv) } -func (m *VolumeAttachment) Unmarshal(dAtA []byte) error { +func (m *CSIStorageCapacity) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -816,10 +1061,10 @@ func (m *VolumeAttachment) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: VolumeAttachment: wiretype end group for non-group") + return fmt.Errorf("proto: CSIStorageCapacity: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: VolumeAttachment: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: CSIStorageCapacity: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: @@ -857,7 +1102,7 @@ func (m *VolumeAttachment) Unmarshal(dAtA []byte) error { iNdEx = postIndex case 2: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field NodeTopology", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -884,13 +1129,48 @@ func (m *VolumeAttachment) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if m.NodeTopology == nil { + m.NodeTopology = &v1.LabelSelector{} + } + if err := m.NodeTopology.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex case 3: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field StorageClassName", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.StorageClassName = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Capacity", wireType) } var msglen int for shift := uint(0); ; shift += 7 { @@ -917,7 +1197,10 @@ func (m *VolumeAttachment) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + if m.Capacity == nil { + m.Capacity = &resource.Quantity{} + } + if err := m.Capacity.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } iNdEx = postIndex @@ -945,7 +1228,7 @@ func (m *VolumeAttachment) Unmarshal(dAtA []byte) error { } return nil } -func (m *VolumeAttachmentList) Unmarshal(dAtA []byte) error { +func (m *CSIStorageCapacityList) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -968,10 +1251,10 @@ func (m *VolumeAttachmentList) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: VolumeAttachmentList: wiretype end group for non-group") + return fmt.Errorf("proto: CSIStorageCapacityList: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: VolumeAttachmentList: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: CSIStorageCapacityList: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: @@ -1036,7 +1319,7 @@ func (m *VolumeAttachmentList) Unmarshal(dAtA []byte) error { if postIndex > l { return io.ErrUnexpectedEOF } - m.Items = append(m.Items, VolumeAttachment{}) + m.Items = append(m.Items, CSIStorageCapacity{}) if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { return err } @@ -1065,7 +1348,7 @@ func (m *VolumeAttachmentList) Unmarshal(dAtA []byte) error { } return nil } -func (m *VolumeAttachmentSource) Unmarshal(dAtA []byte) error { +func (m *VolumeAttachment) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 for iNdEx < l { @@ -1088,17 +1371,17 @@ func (m *VolumeAttachmentSource) Unmarshal(dAtA []byte) error { fieldNum := int32(wire >> 3) wireType := int(wire & 0x7) if wireType == 4 { - return fmt.Errorf("proto: VolumeAttachmentSource: wiretype end group for non-group") + return fmt.Errorf("proto: VolumeAttachment: wiretype end group for non-group") } if fieldNum <= 0 { - return fmt.Errorf("proto: VolumeAttachmentSource: illegal tag %d (wire type %d)", fieldNum, wire) + return fmt.Errorf("proto: VolumeAttachment: illegal tag %d (wire type %d)", fieldNum, wire) } switch fieldNum { case 1: if wireType != 2 { - return fmt.Errorf("proto: wrong wireType = %d for field PersistentVolumeName", wireType) + return fmt.Errorf("proto: wrong wireType = %d for field ObjectMeta", wireType) } - var stringLen uint64 + var msglen int for shift := uint(0); ; shift += 7 { if shift >= 64 { return ErrIntOverflowGenerated @@ -1108,24 +1391,296 @@ func (m *VolumeAttachmentSource) Unmarshal(dAtA []byte) error { } b := dAtA[iNdEx] iNdEx++ - stringLen |= uint64(b&0x7F) << shift + msglen |= int(b&0x7F) << shift if b < 0x80 { break } } - intStringLen := int(stringLen) - if intStringLen < 0 { + if msglen < 0 { return ErrInvalidLengthGenerated } - postIndex := iNdEx + intStringLen + postIndex := iNdEx + msglen if postIndex < 0 { return ErrInvalidLengthGenerated } if postIndex > l { return io.ErrUnexpectedEOF } - s := string(dAtA[iNdEx:postIndex]) - m.PersistentVolumeName = &s + if err := m.ObjectMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Spec", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Spec.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 3: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.Status.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *VolumeAttachmentList) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: VolumeAttachmentList: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: VolumeAttachmentList: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ListMeta", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.ListMeta.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Items", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Items = append(m.Items, VolumeAttachment{}) + if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} +func (m *VolumeAttachmentSource) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: VolumeAttachmentSource: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: VolumeAttachmentSource: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field PersistentVolumeName", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.PersistentVolumeName = &s iNdEx = postIndex case 2: if wireType != 2 { @@ -1730,6 +2285,7 @@ func (m *VolumeError) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -1761,10 +2317,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -1785,55 +2339,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/storage/v1alpha1/generated.proto b/vendor/k8s.io/api/storage/v1alpha1/generated.proto index 76019639..40a76405 100644 --- a/vendor/k8s.io/api/storage/v1alpha1/generated.proto +++ b/vendor/k8s.io/api/storage/v1alpha1/generated.proto @@ -22,6 +22,7 @@ syntax = 'proto2'; package k8s.io.api.storage.v1alpha1; import "k8s.io/api/core/v1/generated.proto"; +import "k8s.io/apimachinery/pkg/api/resource/generated.proto"; import "k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto"; import "k8s.io/apimachinery/pkg/runtime/generated.proto"; import "k8s.io/apimachinery/pkg/runtime/schema/generated.proto"; @@ -29,6 +30,80 @@ import "k8s.io/apimachinery/pkg/runtime/schema/generated.proto"; // Package-wide variables from generator "generated". option go_package = "v1alpha1"; +// CSIStorageCapacity stores the result of one CSI GetCapacity call. +// For a given StorageClass, this describes the available capacity in a +// particular topology segment. This can be used when considering where to +// instantiate new PersistentVolumes. +// +// For example this can express things like: +// - StorageClass "standard" has "1234 GiB" available in "topology.kubernetes.io/zone=us-east1" +// - StorageClass "localssd" has "10 GiB" available in "kubernetes.io/hostname=knode-abc123" +// +// The following three cases all imply that no capacity is available for +// a certain combination: +// - no object exists with suitable topology and storage class name +// - such an object exists, but the capacity is unset +// - such an object exists, but the capacity is zero +// +// The producer of these objects can decide which approach is more suitable. +// +// This is an alpha feature and only available when the CSIStorageCapacity feature is enabled. +message CSIStorageCapacity { + // Standard object's metadata. The name has no particular meaning. It must be + // be a DNS subdomain (dots allowed, 253 characters). To ensure that + // there are no conflicts with other CSI drivers on the cluster, the recommendation + // is to use csisc-, a generated name, or a reverse-domain name which ends + // with the unique CSI driver name. + // + // Objects are namespaced. + // + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; + + // NodeTopology defines which nodes have access to the storage + // for which capacity was reported. If not set, the storage is + // not accessible from any node in the cluster. If empty, the + // storage is accessible from all nodes. This field is + // immutable. + // + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.LabelSelector nodeTopology = 2; + + // The name of the StorageClass that the reported capacity applies to. + // It must meet the same requirements as the name of a StorageClass + // object (non-empty, DNS subdomain). If that object no longer exists, + // the CSIStorageCapacity object is obsolete and should be removed by its + // creator. + // This field is immutable. + optional string storageClassName = 3; + + // Capacity is the value reported by the CSI driver in its GetCapacityResponse + // for a GetCapacityRequest with topology and parameters that match the + // previous fields. + // + // The semantic is currently (CSI spec 1.2) defined as: + // The available capacity, in bytes, of the storage that can be used + // to provision volumes. If not set, that information is currently + // unavailable and treated like zero capacity. + // + // +optional + optional k8s.io.apimachinery.pkg.api.resource.Quantity capacity = 4; +} + +// CSIStorageCapacityList is a collection of CSIStorageCapacity objects. +message CSIStorageCapacityList { + // Standard list metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1; + + // Items is the list of CSIStorageCapacity objects. + // +listType=map + // +listMapKey=name + repeated CSIStorageCapacity items = 2; +} + // VolumeAttachment captures the intent to attach or detach the specified volume // to/from the specified node. // diff --git a/vendor/k8s.io/api/storage/v1alpha1/register.go b/vendor/k8s.io/api/storage/v1alpha1/register.go index 7b81ee49..779c8580 100644 --- a/vendor/k8s.io/api/storage/v1alpha1/register.go +++ b/vendor/k8s.io/api/storage/v1alpha1/register.go @@ -43,6 +43,8 @@ func addKnownTypes(scheme *runtime.Scheme) error { scheme.AddKnownTypes(SchemeGroupVersion, &VolumeAttachment{}, &VolumeAttachmentList{}, + &CSIStorageCapacity{}, + &CSIStorageCapacityList{}, ) metav1.AddToGroupVersion(scheme, SchemeGroupVersion) diff --git a/vendor/k8s.io/api/storage/v1alpha1/types.go b/vendor/k8s.io/api/storage/v1alpha1/types.go index 39408857..5e65bceb 100644 --- a/vendor/k8s.io/api/storage/v1alpha1/types.go +++ b/vendor/k8s.io/api/storage/v1alpha1/types.go @@ -18,6 +18,7 @@ package v1alpha1 import ( "k8s.io/api/core/v1" + "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" ) @@ -134,3 +135,84 @@ type VolumeError struct { // +optional Message string `json:"message,omitempty" protobuf:"bytes,2,opt,name=message"` } + +// +genclient +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// CSIStorageCapacity stores the result of one CSI GetCapacity call. +// For a given StorageClass, this describes the available capacity in a +// particular topology segment. This can be used when considering where to +// instantiate new PersistentVolumes. +// +// For example this can express things like: +// - StorageClass "standard" has "1234 GiB" available in "topology.kubernetes.io/zone=us-east1" +// - StorageClass "localssd" has "10 GiB" available in "kubernetes.io/hostname=knode-abc123" +// +// The following three cases all imply that no capacity is available for +// a certain combination: +// - no object exists with suitable topology and storage class name +// - such an object exists, but the capacity is unset +// - such an object exists, but the capacity is zero +// +// The producer of these objects can decide which approach is more suitable. +// +// This is an alpha feature and only available when the CSIStorageCapacity feature is enabled. +type CSIStorageCapacity struct { + metav1.TypeMeta `json:",inline"` + // Standard object's metadata. The name has no particular meaning. It must be + // be a DNS subdomain (dots allowed, 253 characters). To ensure that + // there are no conflicts with other CSI drivers on the cluster, the recommendation + // is to use csisc-, a generated name, or a reverse-domain name which ends + // with the unique CSI driver name. + // + // Objects are namespaced. + // + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // NodeTopology defines which nodes have access to the storage + // for which capacity was reported. If not set, the storage is + // not accessible from any node in the cluster. If empty, the + // storage is accessible from all nodes. This field is + // immutable. + // + // +optional + NodeTopology *metav1.LabelSelector `json:"nodeTopology,omitempty" protobuf:"bytes,2,opt,name=nodeTopology"` + + // The name of the StorageClass that the reported capacity applies to. + // It must meet the same requirements as the name of a StorageClass + // object (non-empty, DNS subdomain). If that object no longer exists, + // the CSIStorageCapacity object is obsolete and should be removed by its + // creator. + // This field is immutable. + StorageClassName string `json:"storageClassName" protobuf:"bytes,3,name=storageClassName"` + + // Capacity is the value reported by the CSI driver in its GetCapacityResponse + // for a GetCapacityRequest with topology and parameters that match the + // previous fields. + // + // The semantic is currently (CSI spec 1.2) defined as: + // The available capacity, in bytes, of the storage that can be used + // to provision volumes. If not set, that information is currently + // unavailable and treated like zero capacity. + // + // +optional + Capacity *resource.Quantity `json:"capacity,omitempty" protobuf:"bytes,4,opt,name=capacity"` +} + +// +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object + +// CSIStorageCapacityList is a collection of CSIStorageCapacity objects. +type CSIStorageCapacityList struct { + metav1.TypeMeta `json:",inline"` + // Standard list metadata + // More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata + // +optional + metav1.ListMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` + + // Items is the list of CSIStorageCapacity objects. + // +listType=map + // +listMapKey=name + Items []CSIStorageCapacity `json:"items" protobuf:"bytes,2,rep,name=items"` +} diff --git a/vendor/k8s.io/api/storage/v1alpha1/types_swagger_doc_generated.go b/vendor/k8s.io/api/storage/v1alpha1/types_swagger_doc_generated.go index 2e821616..51778d18 100644 --- a/vendor/k8s.io/api/storage/v1alpha1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/storage/v1alpha1/types_swagger_doc_generated.go @@ -27,6 +27,28 @@ package v1alpha1 // Those methods can be generated by using hack/update-generated-swagger-docs.sh // AUTO-GENERATED FUNCTIONS START HERE. DO NOT EDIT. +var map_CSIStorageCapacity = map[string]string{ + "": "CSIStorageCapacity stores the result of one CSI GetCapacity call. For a given StorageClass, this describes the available capacity in a particular topology segment. This can be used when considering where to instantiate new PersistentVolumes.\n\nFor example this can express things like: - StorageClass \"standard\" has \"1234 GiB\" available in \"topology.kubernetes.io/zone=us-east1\" - StorageClass \"localssd\" has \"10 GiB\" available in \"kubernetes.io/hostname=knode-abc123\"\n\nThe following three cases all imply that no capacity is available for a certain combination: - no object exists with suitable topology and storage class name - such an object exists, but the capacity is unset - such an object exists, but the capacity is zero\n\nThe producer of these objects can decide which approach is more suitable.\n\nThis is an alpha feature and only available when the CSIStorageCapacity feature is enabled.", + "metadata": "Standard object's metadata. The name has no particular meaning. It must be be a DNS subdomain (dots allowed, 253 characters). To ensure that there are no conflicts with other CSI drivers on the cluster, the recommendation is to use csisc-, a generated name, or a reverse-domain name which ends with the unique CSI driver name.\n\nObjects are namespaced.\n\nMore info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "nodeTopology": "NodeTopology defines which nodes have access to the storage for which capacity was reported. If not set, the storage is not accessible from any node in the cluster. If empty, the storage is accessible from all nodes. This field is immutable.", + "storageClassName": "The name of the StorageClass that the reported capacity applies to. It must meet the same requirements as the name of a StorageClass object (non-empty, DNS subdomain). If that object no longer exists, the CSIStorageCapacity object is obsolete and should be removed by its creator. This field is immutable.", + "capacity": "Capacity is the value reported by the CSI driver in its GetCapacityResponse for a GetCapacityRequest with topology and parameters that match the previous fields.\n\nThe semantic is currently (CSI spec 1.2) defined as: The available capacity, in bytes, of the storage that can be used to provision volumes. If not set, that information is currently unavailable and treated like zero capacity.", +} + +func (CSIStorageCapacity) SwaggerDoc() map[string]string { + return map_CSIStorageCapacity +} + +var map_CSIStorageCapacityList = map[string]string{ + "": "CSIStorageCapacityList is a collection of CSIStorageCapacity objects.", + "metadata": "Standard list metadata More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", + "items": "Items is the list of CSIStorageCapacity objects.", +} + +func (CSIStorageCapacityList) SwaggerDoc() map[string]string { + return map_CSIStorageCapacityList +} + var map_VolumeAttachment = map[string]string{ "": "VolumeAttachment captures the intent to attach or detach the specified volume to/from the specified node.\n\nVolumeAttachment objects are non-namespaced.", "metadata": "Standard object metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata", diff --git a/vendor/k8s.io/api/storage/v1alpha1/zz_generated.deepcopy.go b/vendor/k8s.io/api/storage/v1alpha1/zz_generated.deepcopy.go index 3debf9df..7f3b357b 100644 --- a/vendor/k8s.io/api/storage/v1alpha1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/storage/v1alpha1/zz_generated.deepcopy.go @@ -21,10 +21,80 @@ limitations under the License. package v1alpha1 import ( - v1 "k8s.io/api/core/v1" + corev1 "k8s.io/api/core/v1" + v1 "k8s.io/apimachinery/pkg/apis/meta/v1" runtime "k8s.io/apimachinery/pkg/runtime" ) +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CSIStorageCapacity) DeepCopyInto(out *CSIStorageCapacity) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ObjectMeta.DeepCopyInto(&out.ObjectMeta) + if in.NodeTopology != nil { + in, out := &in.NodeTopology, &out.NodeTopology + *out = new(v1.LabelSelector) + (*in).DeepCopyInto(*out) + } + if in.Capacity != nil { + in, out := &in.Capacity, &out.Capacity + x := (*in).DeepCopy() + *out = &x + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CSIStorageCapacity. +func (in *CSIStorageCapacity) DeepCopy() *CSIStorageCapacity { + if in == nil { + return nil + } + out := new(CSIStorageCapacity) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *CSIStorageCapacity) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *CSIStorageCapacityList) DeepCopyInto(out *CSIStorageCapacityList) { + *out = *in + out.TypeMeta = in.TypeMeta + in.ListMeta.DeepCopyInto(&out.ListMeta) + if in.Items != nil { + in, out := &in.Items, &out.Items + *out = make([]CSIStorageCapacity, len(*in)) + for i := range *in { + (*in)[i].DeepCopyInto(&(*out)[i]) + } + } + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new CSIStorageCapacityList. +func (in *CSIStorageCapacityList) DeepCopy() *CSIStorageCapacityList { + if in == nil { + return nil + } + out := new(CSIStorageCapacityList) + in.DeepCopyInto(out) + return out +} + +// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object. +func (in *CSIStorageCapacityList) DeepCopyObject() runtime.Object { + if c := in.DeepCopy(); c != nil { + return c + } + return nil +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *VolumeAttachment) DeepCopyInto(out *VolumeAttachment) { *out = *in @@ -96,7 +166,7 @@ func (in *VolumeAttachmentSource) DeepCopyInto(out *VolumeAttachmentSource) { } if in.InlineVolumeSpec != nil { in, out := &in.InlineVolumeSpec, &out.InlineVolumeSpec - *out = new(v1.PersistentVolumeSpec) + *out = new(corev1.PersistentVolumeSpec) (*in).DeepCopyInto(*out) } return diff --git a/vendor/k8s.io/api/storage/v1beta1/doc.go b/vendor/k8s.io/api/storage/v1beta1/doc.go index e3e3626e..75142a62 100644 --- a/vendor/k8s.io/api/storage/v1beta1/doc.go +++ b/vendor/k8s.io/api/storage/v1beta1/doc.go @@ -18,5 +18,6 @@ limitations under the License. // +k8s:protobuf-gen=package // +groupName=storage.k8s.io // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true package v1beta1 // import "k8s.io/api/storage/v1beta1" diff --git a/vendor/k8s.io/api/storage/v1beta1/generated.pb.go b/vendor/k8s.io/api/storage/v1beta1/generated.pb.go index cd35af34..cec77515 100644 --- a/vendor/k8s.io/api/storage/v1beta1/generated.pb.go +++ b/vendor/k8s.io/api/storage/v1beta1/generated.pb.go @@ -44,7 +44,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *CSIDriver) Reset() { *m = CSIDriver{} } func (*CSIDriver) ProtoMessage() {} @@ -520,91 +520,95 @@ func init() { } var fileDescriptor_7d2980599fd0de80 = []byte{ - // 1344 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x57, 0xbd, 0x6f, 0xdb, 0x46, - 0x1b, 0x37, 0x2d, 0x7f, 0x9e, 0xec, 0x44, 0xbe, 0x18, 0xef, 0xab, 0x57, 0x83, 0x64, 0xe8, 0x45, - 0x1b, 0x27, 0x48, 0xc8, 0x24, 0x48, 0x83, 0x20, 0x40, 0x07, 0xd3, 0x31, 0x50, 0x25, 0x96, 0xe3, - 0x9e, 0x8d, 0xa0, 0x08, 0x3a, 0xf4, 0x44, 0x3e, 0x91, 0x19, 0x93, 0x3c, 0x86, 0x3c, 0xa9, 0xd5, - 0xd6, 0xa9, 0x73, 0xd1, 0xa1, 0x7f, 0x41, 0xff, 0x85, 0x16, 0x68, 0x97, 0x8e, 0xcd, 0x54, 0x04, - 0x9d, 0x32, 0x09, 0x0d, 0xbb, 0x76, 0xeb, 0x66, 0x74, 0x28, 0xee, 0x78, 0x12, 0x29, 0x89, 0x8a, - 0xed, 0x0e, 0xde, 0x78, 0xcf, 0xc7, 0xef, 0xf9, 0x7e, 0xee, 0x88, 0xb6, 0x8f, 0xef, 0x47, 0xba, - 0xc3, 0x8c, 0xe3, 0x4e, 0x0b, 0x42, 0x1f, 0x38, 0x44, 0x46, 0x17, 0x7c, 0x9b, 0x85, 0x86, 0x62, - 0xd0, 0xc0, 0x31, 0x22, 0xce, 0x42, 0xda, 0x06, 0xa3, 0x7b, 0xbb, 0x05, 0x9c, 0xde, 0x36, 0xda, - 0xe0, 0x43, 0x48, 0x39, 0xd8, 0x7a, 0x10, 0x32, 0xce, 0x70, 0x25, 0x91, 0xd5, 0x69, 0xe0, 0xe8, - 0x4a, 0x56, 0x57, 0xb2, 0x95, 0x9b, 0x6d, 0x87, 0x1f, 0x75, 0x5a, 0xba, 0xc5, 0x3c, 0xa3, 0xcd, - 0xda, 0xcc, 0x90, 0x2a, 0xad, 0xce, 0x73, 0x79, 0x92, 0x07, 0xf9, 0x95, 0x40, 0x55, 0xea, 0x19, - 0xb3, 0x16, 0x0b, 0x85, 0xcd, 0x71, 0x73, 0x95, 0xbb, 0xa9, 0x8c, 0x47, 0xad, 0x23, 0xc7, 0x87, - 0xb0, 0x67, 0x04, 0xc7, 0x6d, 0x41, 0x88, 0x0c, 0x0f, 0x38, 0xcd, 0xd3, 0x32, 0xa6, 0x69, 0x85, - 0x1d, 0x9f, 0x3b, 0x1e, 0x4c, 0x28, 0xdc, 0x3b, 0x4d, 0x21, 0xb2, 0x8e, 0xc0, 0xa3, 0xe3, 0x7a, - 0xf5, 0x9f, 0x34, 0xb4, 0xbc, 0x7d, 0xd0, 0x78, 0x18, 0x3a, 0x5d, 0x08, 0xf1, 0x67, 0x68, 0x49, - 0x78, 0x64, 0x53, 0x4e, 0xcb, 0xda, 0x86, 0xb6, 0x59, 0xbc, 0x73, 0x4b, 0x4f, 0xd3, 0x35, 0x04, - 0xd6, 0x83, 0xe3, 0xb6, 0x20, 0x44, 0xba, 0x90, 0xd6, 0xbb, 0xb7, 0xf5, 0x27, 0xad, 0x17, 0x60, - 0xf1, 0x26, 0x70, 0x6a, 0xe2, 0x57, 0xfd, 0xda, 0x4c, 0xdc, 0xaf, 0xa1, 0x94, 0x46, 0x86, 0xa8, - 0xf8, 0x31, 0x9a, 0x8b, 0x02, 0xb0, 0xca, 0xb3, 0x12, 0xfd, 0x9a, 0x3e, 0xbd, 0x18, 0xfa, 0xd0, - 0xad, 0x83, 0x00, 0x2c, 0x73, 0x45, 0xc1, 0xce, 0x89, 0x13, 0x91, 0x20, 0xf5, 0x1f, 0x35, 0xb4, - 0x3a, 0x94, 0xda, 0x75, 0x22, 0x8e, 0x3f, 0x9d, 0x08, 0x40, 0x3f, 0x5b, 0x00, 0x42, 0x5b, 0xba, - 0x5f, 0x52, 0x76, 0x96, 0x06, 0x94, 0x8c, 0xf3, 0x8f, 0xd0, 0xbc, 0xc3, 0xc1, 0x8b, 0xca, 0xb3, - 0x1b, 0x85, 0xcd, 0xe2, 0x9d, 0xf7, 0xce, 0xe4, 0xbd, 0xb9, 0xaa, 0x10, 0xe7, 0x1b, 0x42, 0x97, - 0x24, 0x10, 0xf5, 0x3f, 0xb3, 0xbe, 0x8b, 0x98, 0xf0, 0x03, 0x74, 0x89, 0x72, 0x4e, 0xad, 0x23, - 0x02, 0x2f, 0x3b, 0x4e, 0x08, 0xb6, 0x8c, 0x60, 0xc9, 0xc4, 0x71, 0xbf, 0x76, 0x69, 0x6b, 0x84, - 0x43, 0xc6, 0x24, 0x85, 0x6e, 0xc0, 0xec, 0x86, 0xff, 0x9c, 0x3d, 0xf1, 0x9b, 0xac, 0xe3, 0x73, - 0x99, 0x60, 0xa5, 0xbb, 0x3f, 0xc2, 0x21, 0x63, 0x92, 0xd8, 0x42, 0xeb, 0x5d, 0xe6, 0x76, 0x3c, - 0xd8, 0x75, 0x9e, 0x83, 0xd5, 0xb3, 0x5c, 0x68, 0x32, 0x1b, 0xa2, 0x72, 0x61, 0xa3, 0xb0, 0xb9, - 0x6c, 0x1a, 0x71, 0xbf, 0xb6, 0xfe, 0x34, 0x87, 0x7f, 0xd2, 0xaf, 0x5d, 0xc9, 0xa1, 0x93, 0x5c, - 0xb0, 0xfa, 0x0f, 0x1a, 0x5a, 0xdc, 0x3e, 0x68, 0xec, 0x31, 0x1b, 0x2e, 0xa0, 0xcb, 0x1a, 0x23, - 0x5d, 0x76, 0xf5, 0x94, 0x3a, 0x09, 0xa7, 0xa6, 0xf6, 0xd8, 0x5f, 0x49, 0x9d, 0x84, 0x8c, 0x1a, - 0x92, 0x0d, 0x34, 0xe7, 0x53, 0x0f, 0xa4, 0xeb, 0xcb, 0xa9, 0xce, 0x1e, 0xf5, 0x80, 0x48, 0x0e, - 0x7e, 0x1f, 0x2d, 0xf8, 0xcc, 0x86, 0xc6, 0x43, 0xe9, 0xc0, 0xb2, 0x79, 0x49, 0xc9, 0x2c, 0xec, - 0x49, 0x2a, 0x51, 0x5c, 0x7c, 0x17, 0xad, 0x70, 0x16, 0x30, 0x97, 0xb5, 0x7b, 0x8f, 0xa1, 0x37, - 0xc8, 0x78, 0x29, 0xee, 0xd7, 0x56, 0x0e, 0x33, 0x74, 0x32, 0x22, 0x85, 0x5b, 0xa8, 0x48, 0x5d, - 0x97, 0x59, 0x94, 0xd3, 0x96, 0x0b, 0xe5, 0x39, 0x19, 0xa3, 0xf1, 0xae, 0x18, 0x93, 0x32, 0x09, - 0xe3, 0x04, 0x22, 0xd6, 0x09, 0x2d, 0x88, 0xcc, 0xcb, 0x71, 0xbf, 0x56, 0xdc, 0x4a, 0x71, 0x48, - 0x16, 0xb4, 0xfe, 0xbd, 0x86, 0x8a, 0x2a, 0xea, 0x0b, 0x98, 0xab, 0x8f, 0x46, 0xe7, 0xea, 0xff, - 0x67, 0xa8, 0xd7, 0x94, 0xa9, 0xb2, 0x86, 0x6e, 0xcb, 0x91, 0x3a, 0x44, 0x8b, 0xb6, 0x2c, 0x5a, - 0x54, 0xd6, 0x24, 0xf4, 0xb5, 0x33, 0x40, 0xab, 0xb1, 0xbd, 0xac, 0x0c, 0x2c, 0x26, 0xe7, 0x88, - 0x0c, 0xa0, 0xea, 0xdf, 0x2c, 0xa0, 0x95, 0x83, 0x44, 0x77, 0xdb, 0xa5, 0x51, 0x74, 0x01, 0x0d, - 0xfd, 0x01, 0x2a, 0x06, 0x21, 0xeb, 0x3a, 0x91, 0xc3, 0x7c, 0x08, 0x55, 0x5b, 0x5d, 0x51, 0x2a, - 0xc5, 0xfd, 0x94, 0x45, 0xb2, 0x72, 0xd8, 0x45, 0x28, 0xa0, 0x21, 0xf5, 0x80, 0x8b, 0x14, 0x14, - 0x64, 0x0a, 0xee, 0xbf, 0x2b, 0x05, 0xd9, 0xb0, 0xf4, 0xfd, 0xa1, 0xea, 0x8e, 0xcf, 0xc3, 0x5e, - 0xea, 0x62, 0xca, 0x20, 0x19, 0x7c, 0x7c, 0x8c, 0x56, 0x43, 0xb0, 0x5c, 0xea, 0x78, 0xfb, 0xcc, - 0x75, 0xac, 0x9e, 0x6c, 0xcd, 0x65, 0x73, 0x27, 0xee, 0xd7, 0x56, 0x49, 0x96, 0x71, 0xd2, 0xaf, - 0xdd, 0x9a, 0xbc, 0x3a, 0xf5, 0x7d, 0x08, 0x23, 0x27, 0xe2, 0xe0, 0xf3, 0xa4, 0x61, 0x47, 0x74, - 0xc8, 0x28, 0xb6, 0x98, 0x1d, 0x4f, 0xac, 0xaf, 0x27, 0x01, 0x77, 0x98, 0x1f, 0x95, 0xe7, 0xd3, - 0xd9, 0x69, 0x66, 0xe8, 0x64, 0x44, 0x0a, 0xef, 0xa2, 0x75, 0xd1, 0xe6, 0x9f, 0x27, 0x06, 0x76, - 0xbe, 0x08, 0xa8, 0x2f, 0x52, 0x55, 0x5e, 0x90, 0xdb, 0xb2, 0x2c, 0x76, 0xdd, 0x56, 0x0e, 0x9f, - 0xe4, 0x6a, 0xe1, 0x4f, 0xd0, 0x5a, 0xb2, 0xec, 0x4c, 0xc7, 0xb7, 0x1d, 0xbf, 0x2d, 0x56, 0x5d, - 0x79, 0x51, 0x06, 0x7d, 0x3d, 0xee, 0xd7, 0xd6, 0x9e, 0x8e, 0x33, 0x4f, 0xf2, 0x88, 0x64, 0x12, - 0x04, 0xbf, 0x44, 0x6b, 0xd2, 0x22, 0xd8, 0x6a, 0x11, 0x38, 0x10, 0x95, 0x97, 0x64, 0xfd, 0x36, - 0xb3, 0xf5, 0x13, 0xa9, 0x13, 0x8d, 0x34, 0x58, 0x17, 0x07, 0xe0, 0x82, 0xc5, 0x59, 0x78, 0x08, - 0xa1, 0x67, 0xfe, 0x4f, 0xd5, 0x6b, 0x6d, 0x6b, 0x1c, 0x8a, 0x4c, 0xa2, 0x57, 0x3e, 0x44, 0x97, - 0xc7, 0x0a, 0x8e, 0x4b, 0xa8, 0x70, 0x0c, 0xbd, 0x64, 0xd1, 0x11, 0xf1, 0x89, 0xd7, 0xd1, 0x7c, - 0x97, 0xba, 0x1d, 0x48, 0x3a, 0x90, 0x24, 0x87, 0x07, 0xb3, 0xf7, 0xb5, 0xfa, 0xcf, 0x1a, 0x2a, - 0x65, 0xbb, 0xe7, 0x02, 0xd6, 0x46, 0x73, 0x74, 0x6d, 0x6c, 0x9e, 0xb5, 0xb1, 0xa7, 0xec, 0x8e, - 0xef, 0x66, 0x51, 0x29, 0x29, 0x4e, 0x72, 0xd9, 0x7a, 0xe0, 0xf3, 0x0b, 0x18, 0x6d, 0x32, 0x72, - 0x57, 0xdd, 0x3a, 0x7d, 0x8f, 0xa7, 0xde, 0x4d, 0xbb, 0xb4, 0xf0, 0x33, 0xb4, 0x10, 0x71, 0xca, - 0x3b, 0x62, 0xe6, 0x05, 0xea, 0x9d, 0x73, 0xa1, 0x4a, 0xcd, 0xf4, 0xd2, 0x4a, 0xce, 0x44, 0x21, - 0xd6, 0x7f, 0xd1, 0xd0, 0xfa, 0xb8, 0xca, 0x05, 0x14, 0xfb, 0xe3, 0xd1, 0x62, 0xdf, 0x38, 0x4f, - 0x44, 0x53, 0x0a, 0xfe, 0x9b, 0x86, 0xfe, 0x33, 0x11, 0xbc, 0xbc, 0x1e, 0xc5, 0x9e, 0x08, 0xc6, - 0xb6, 0xd1, 0x5e, 0x7a, 0xe7, 0xcb, 0x3d, 0xb1, 0x9f, 0xc3, 0x27, 0xb9, 0x5a, 0xf8, 0x05, 0x2a, - 0x39, 0xbe, 0xeb, 0xf8, 0x90, 0xd0, 0x0e, 0xd2, 0x72, 0xe7, 0x0e, 0xf3, 0x38, 0xb2, 0x2c, 0xf3, - 0x7a, 0xdc, 0xaf, 0x95, 0x1a, 0x63, 0x28, 0x64, 0x02, 0xb7, 0xfe, 0x6b, 0x4e, 0x79, 0xe4, 0x5d, - 0x78, 0x03, 0x2d, 0x25, 0x8f, 0x46, 0x08, 0x55, 0x18, 0xc3, 0x74, 0x6f, 0x29, 0x3a, 0x19, 0x4a, - 0xc8, 0x0e, 0x92, 0xa9, 0x50, 0x8e, 0x9e, 0xaf, 0x83, 0xa4, 0x66, 0xa6, 0x83, 0xe4, 0x99, 0x28, - 0x44, 0xe1, 0x89, 0x78, 0x00, 0xc9, 0x84, 0x16, 0x46, 0x3d, 0xd9, 0x53, 0x74, 0x32, 0x94, 0xa8, - 0xff, 0x5d, 0xc8, 0xa9, 0x92, 0x6c, 0xc5, 0x4c, 0x48, 0x83, 0xb7, 0xf2, 0x78, 0x48, 0xf6, 0x30, - 0x24, 0x1b, 0x7f, 0xab, 0x21, 0x4c, 0x87, 0x10, 0xcd, 0x41, 0xab, 0x26, 0xfd, 0xf4, 0xe8, 0xfc, - 0x13, 0xa2, 0x6f, 0x4d, 0x80, 0x25, 0xf7, 0x64, 0x45, 0x39, 0x81, 0x27, 0x05, 0x48, 0x8e, 0x07, - 0xd8, 0x41, 0xc5, 0x84, 0xba, 0x13, 0x86, 0x2c, 0x54, 0x23, 0x7b, 0xf5, 0x74, 0x87, 0xa4, 0xb8, - 0x59, 0x95, 0x0f, 0xb9, 0x54, 0xff, 0xa4, 0x5f, 0x2b, 0x66, 0xf8, 0x24, 0x8b, 0x2d, 0x4c, 0xd9, - 0x90, 0x9a, 0x9a, 0xfb, 0x17, 0xa6, 0x1e, 0xc2, 0x74, 0x53, 0x19, 0xec, 0xca, 0x0e, 0xfa, 0xef, - 0x94, 0x04, 0x9d, 0xeb, 0x5e, 0xf9, 0x4a, 0x43, 0x59, 0x1b, 0x78, 0x17, 0xcd, 0x89, 0xff, 0x59, - 0xb5, 0x61, 0xae, 0x9f, 0x6d, 0xc3, 0x1c, 0x3a, 0x1e, 0xa4, 0x8b, 0x52, 0x9c, 0x88, 0x44, 0xc1, - 0xd7, 0xd0, 0xa2, 0x07, 0x51, 0x44, 0xdb, 0xca, 0x72, 0xfa, 0xea, 0x6b, 0x26, 0x64, 0x32, 0xe0, - 0xd7, 0xef, 0xa1, 0x2b, 0x39, 0xef, 0x68, 0x5c, 0x43, 0xf3, 0x96, 0xfc, 0xe1, 0x12, 0x0e, 0xcd, - 0x9b, 0xcb, 0x62, 0xcb, 0x6c, 0xcb, 0xff, 0xac, 0x84, 0x6e, 0xde, 0x7c, 0xf5, 0xb6, 0x3a, 0xf3, - 0xfa, 0x6d, 0x75, 0xe6, 0xcd, 0xdb, 0xea, 0xcc, 0x97, 0x71, 0x55, 0x7b, 0x15, 0x57, 0xb5, 0xd7, - 0x71, 0x55, 0x7b, 0x13, 0x57, 0xb5, 0xdf, 0xe3, 0xaa, 0xf6, 0xf5, 0x1f, 0xd5, 0x99, 0x67, 0x8b, - 0x2a, 0xdf, 0xff, 0x04, 0x00, 0x00, 0xff, 0xff, 0x72, 0xff, 0xde, 0x2e, 0xe4, 0x10, 0x00, 0x00, + // 1400 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x57, 0x3d, 0x6f, 0xdb, 0x46, + 0x1f, 0x37, 0x2d, 0xc9, 0x2f, 0x27, 0x3b, 0x96, 0xcf, 0xc6, 0xf3, 0xe8, 0xd1, 0x20, 0x1a, 0x7a, + 0xd0, 0xc6, 0x09, 0x12, 0x2a, 0x31, 0xd2, 0x20, 0x08, 0x90, 0xc1, 0x72, 0xdc, 0x46, 0x89, 0xe5, + 0xb8, 0x27, 0x23, 0x28, 0x82, 0x0e, 0x3d, 0x91, 0x67, 0x99, 0xb1, 0xc8, 0x63, 0xc8, 0x93, 0x5a, + 0x6d, 0x9d, 0x3a, 0x17, 0x1d, 0xfa, 0x09, 0xfa, 0x15, 0x5a, 0xa0, 0x5d, 0x3a, 0x36, 0x53, 0x11, + 0x74, 0xca, 0x44, 0x34, 0xec, 0x47, 0x28, 0xba, 0x18, 0x1d, 0x8a, 0x3b, 0x9e, 0xc4, 0x17, 0x51, + 0xb1, 0xdd, 0xc1, 0x1b, 0xef, 0xff, 0xf2, 0xfb, 0xbf, 0xff, 0xef, 0x08, 0x76, 0x4e, 0xee, 0x79, + 0x9a, 0x49, 0xeb, 0x27, 0xfd, 0x0e, 0x71, 0x6d, 0xc2, 0x88, 0x57, 0x1f, 0x10, 0xdb, 0xa0, 0x6e, + 0x5d, 0x32, 0xb0, 0x63, 0xd6, 0x3d, 0x46, 0x5d, 0xdc, 0x25, 0xf5, 0xc1, 0xed, 0x0e, 0x61, 0xf8, + 0x76, 0xbd, 0x4b, 0x6c, 0xe2, 0x62, 0x46, 0x0c, 0xcd, 0x71, 0x29, 0xa3, 0xb0, 0x12, 0xca, 0x6a, + 0xd8, 0x31, 0x35, 0x29, 0xab, 0x49, 0xd9, 0xca, 0xcd, 0xae, 0xc9, 0x8e, 0xfb, 0x1d, 0x4d, 0xa7, + 0x56, 0xbd, 0x4b, 0xbb, 0xb4, 0x2e, 0x54, 0x3a, 0xfd, 0x23, 0x71, 0x12, 0x07, 0xf1, 0x15, 0x42, + 0x55, 0x6a, 0x31, 0xb3, 0x3a, 0x75, 0xb9, 0xcd, 0xb4, 0xb9, 0xca, 0x9d, 0x48, 0xc6, 0xc2, 0xfa, + 0xb1, 0x69, 0x13, 0x77, 0x58, 0x77, 0x4e, 0xba, 0x9c, 0xe0, 0xd5, 0x2d, 0xc2, 0x70, 0x96, 0x56, + 0x7d, 0x9a, 0x96, 0xdb, 0xb7, 0x99, 0x69, 0x91, 0x09, 0x85, 0xbb, 0x67, 0x29, 0x78, 0xfa, 0x31, + 0xb1, 0x70, 0x5a, 0xaf, 0xf6, 0x93, 0x02, 0x16, 0x77, 0xda, 0xcd, 0x87, 0xae, 0x39, 0x20, 0x2e, + 0xfc, 0x0c, 0x2c, 0x70, 0x8f, 0x0c, 0xcc, 0x70, 0x59, 0xd9, 0x50, 0x36, 0x8b, 0x5b, 0xb7, 0xb4, + 0x28, 0x5d, 0x63, 0x60, 0xcd, 0x39, 0xe9, 0x72, 0x82, 0xa7, 0x71, 0x69, 0x6d, 0x70, 0x5b, 0x7b, + 0xda, 0x79, 0x41, 0x74, 0xd6, 0x22, 0x0c, 0x37, 0xe0, 0x2b, 0x5f, 0x9d, 0x09, 0x7c, 0x15, 0x44, + 0x34, 0x34, 0x46, 0x85, 0x4f, 0x40, 0xde, 0x73, 0x88, 0x5e, 0x9e, 0x15, 0xe8, 0xd7, 0xb4, 0xe9, + 0xc5, 0xd0, 0xc6, 0x6e, 0xb5, 0x1d, 0xa2, 0x37, 0x96, 0x24, 0x6c, 0x9e, 0x9f, 0x90, 0x00, 0xa9, + 0xfd, 0xa8, 0x80, 0xe5, 0xb1, 0xd4, 0x9e, 0xe9, 0x31, 0xf8, 0xe9, 0x44, 0x00, 0xda, 0xf9, 0x02, + 0xe0, 0xda, 0xc2, 0xfd, 0x92, 0xb4, 0xb3, 0x30, 0xa2, 0xc4, 0x9c, 0x7f, 0x0c, 0x0a, 0x26, 0x23, + 0x96, 0x57, 0x9e, 0xdd, 0xc8, 0x6d, 0x16, 0xb7, 0xde, 0x3b, 0x97, 0xf7, 0x8d, 0x65, 0x89, 0x58, + 0x68, 0x72, 0x5d, 0x14, 0x42, 0xd4, 0xfe, 0x9a, 0x8d, 0xf9, 0xce, 0x63, 0x82, 0xf7, 0xc1, 0x15, + 0xcc, 0x18, 0xd6, 0x8f, 0x11, 0x79, 0xd9, 0x37, 0x5d, 0x62, 0x88, 0x08, 0x16, 0x1a, 0x30, 0xf0, + 0xd5, 0x2b, 0xdb, 0x09, 0x0e, 0x4a, 0x49, 0x72, 0x5d, 0x87, 0x1a, 0x4d, 0xfb, 0x88, 0x3e, 0xb5, + 0x5b, 0xb4, 0x6f, 0x33, 0x91, 0x60, 0xa9, 0x7b, 0x90, 0xe0, 0xa0, 0x94, 0x24, 0xd4, 0xc1, 0xfa, + 0x80, 0xf6, 0xfa, 0x16, 0xd9, 0x33, 0x8f, 0x88, 0x3e, 0xd4, 0x7b, 0xa4, 0x45, 0x0d, 0xe2, 0x95, + 0x73, 0x1b, 0xb9, 0xcd, 0xc5, 0x46, 0x3d, 0xf0, 0xd5, 0xf5, 0x67, 0x19, 0xfc, 0x53, 0x5f, 0x5d, + 0xcb, 0xa0, 0xa3, 0x4c, 0x30, 0xf8, 0x00, 0xac, 0xc8, 0x0c, 0xed, 0x60, 0x07, 0xeb, 0x26, 0x1b, + 0x96, 0xf3, 0xc2, 0xc3, 0xb5, 0xc0, 0x57, 0x57, 0xda, 0x49, 0x16, 0x4a, 0xcb, 0xc2, 0x47, 0x60, + 0xf9, 0xc8, 0xfb, 0xc8, 0xa5, 0x7d, 0xe7, 0x80, 0xf6, 0x4c, 0x7d, 0x58, 0x2e, 0x6c, 0x28, 0x9b, + 0x8b, 0x8d, 0x5a, 0xe0, 0xab, 0xcb, 0x1f, 0xb6, 0x63, 0x8c, 0xd3, 0x34, 0x01, 0x25, 0x15, 0x6b, + 0x3f, 0x28, 0x60, 0x7e, 0xa7, 0xdd, 0xdc, 0xa7, 0x06, 0xb9, 0x84, 0x76, 0x6f, 0x26, 0xda, 0xfd, + 0xea, 0x19, 0x0d, 0xc3, 0x9d, 0x9a, 0xda, 0xec, 0x7f, 0x86, 0xcd, 0xce, 0x65, 0xe4, 0xb4, 0x6e, + 0x80, 0xbc, 0x8d, 0x2d, 0x22, 0x5c, 0x5f, 0x8c, 0x74, 0xf6, 0xb1, 0x45, 0x90, 0xe0, 0xc0, 0xf7, + 0xc1, 0x9c, 0x4d, 0x0d, 0xd2, 0x7c, 0x28, 0x1c, 0x58, 0x6c, 0x5c, 0x91, 0x32, 0x73, 0xfb, 0x82, + 0x8a, 0x24, 0x17, 0xde, 0x01, 0x4b, 0x8c, 0x3a, 0xb4, 0x47, 0xbb, 0xc3, 0x27, 0x64, 0x38, 0x2a, + 0x7d, 0x29, 0xf0, 0xd5, 0xa5, 0xc3, 0x18, 0x1d, 0x25, 0xa4, 0x60, 0x07, 0x14, 0x71, 0xaf, 0x47, + 0x75, 0xcc, 0x70, 0xa7, 0x47, 0x44, 0x3d, 0x8b, 0x5b, 0xf5, 0x77, 0xc5, 0x18, 0xf6, 0x0b, 0x37, + 0x8e, 0x88, 0x47, 0xfb, 0xae, 0x4e, 0xbc, 0xc6, 0x4a, 0xe0, 0xab, 0xc5, 0xed, 0x08, 0x07, 0xc5, + 0x41, 0x6b, 0xdf, 0x2b, 0xa0, 0x28, 0xa3, 0xbe, 0x84, 0x01, 0x7f, 0x94, 0x1c, 0xf0, 0xff, 0x9f, + 0xa3, 0x5e, 0x53, 0xc6, 0x5b, 0x1f, 0xbb, 0x2d, 0x66, 0xfb, 0x10, 0xcc, 0x1b, 0xa2, 0x68, 0x5e, + 0x59, 0x11, 0xd0, 0xd7, 0xce, 0x01, 0x2d, 0xf7, 0xc7, 0x8a, 0x34, 0x30, 0x1f, 0x9e, 0x3d, 0x34, + 0x82, 0xaa, 0x7d, 0x33, 0x07, 0x96, 0x46, 0xa3, 0xd3, 0xc3, 0x9e, 0x77, 0x09, 0x0d, 0xfd, 0x01, + 0x28, 0x3a, 0x2e, 0x1d, 0x98, 0x9e, 0x49, 0x6d, 0xe2, 0xca, 0xb6, 0x5a, 0x93, 0x2a, 0xc5, 0x83, + 0x88, 0x85, 0xe2, 0x72, 0xb0, 0x07, 0x80, 0x83, 0x5d, 0x6c, 0x11, 0xc6, 0x53, 0x90, 0x13, 0x29, + 0xb8, 0xf7, 0xae, 0x14, 0xc4, 0xc3, 0xd2, 0x0e, 0xc6, 0xaa, 0xbb, 0x36, 0x73, 0x87, 0x91, 0x8b, + 0x11, 0x03, 0xc5, 0xf0, 0xe1, 0x09, 0x58, 0x76, 0x89, 0xde, 0xc3, 0xa6, 0x25, 0xb7, 0x45, 0x5e, + 0xb8, 0xb9, 0xcb, 0xb7, 0x05, 0x8a, 0x33, 0x4e, 0x7d, 0xf5, 0xd6, 0xe4, 0x1d, 0xae, 0x1d, 0x10, + 0xd7, 0x33, 0x3d, 0x46, 0x6c, 0x16, 0x36, 0x6c, 0x42, 0x07, 0x25, 0xb1, 0xf9, 0xec, 0x58, 0x7c, + 0x8f, 0x3e, 0x75, 0x98, 0x49, 0x6d, 0xaf, 0x5c, 0x88, 0x66, 0xa7, 0x15, 0xa3, 0xa3, 0x84, 0x14, + 0xdc, 0x03, 0xeb, 0xbc, 0xcd, 0x3f, 0x0f, 0x0d, 0xec, 0x7e, 0xe1, 0x60, 0x9b, 0xa7, 0xaa, 0x3c, + 0x27, 0x96, 0x62, 0x99, 0x2f, 0xdd, 0xed, 0x0c, 0x3e, 0xca, 0xd4, 0x82, 0x9f, 0x80, 0xd5, 0x70, + 0xeb, 0x36, 0x4c, 0xdb, 0x30, 0xed, 0x2e, 0xdf, 0xb9, 0xe5, 0x79, 0x11, 0xf4, 0xf5, 0xc0, 0x57, + 0x57, 0x9f, 0xa5, 0x99, 0xa7, 0x59, 0x44, 0x34, 0x09, 0x02, 0x5f, 0x82, 0x55, 0x61, 0x91, 0x18, + 0x72, 0x11, 0x98, 0xc4, 0x2b, 0x2f, 0x88, 0xfa, 0x6d, 0xc6, 0xeb, 0xc7, 0x53, 0xc7, 0x1b, 0x69, + 0xb4, 0x2e, 0xda, 0xa4, 0x47, 0x74, 0x46, 0xdd, 0x43, 0xe2, 0x5a, 0x8d, 0xff, 0xc9, 0x7a, 0xad, + 0x6e, 0xa7, 0xa1, 0xd0, 0x24, 0x7a, 0xe5, 0x01, 0x58, 0x49, 0x15, 0x1c, 0x96, 0x40, 0xee, 0x84, + 0x0c, 0xc3, 0x45, 0x87, 0xf8, 0x27, 0x5c, 0x07, 0x85, 0x01, 0xee, 0xf5, 0x49, 0xd8, 0x81, 0x28, + 0x3c, 0xdc, 0x9f, 0xbd, 0xa7, 0xd4, 0x7e, 0x56, 0x40, 0x29, 0xde, 0x3d, 0x97, 0xb0, 0x36, 0x5a, + 0xc9, 0xb5, 0xb1, 0x79, 0xde, 0xc6, 0x9e, 0xb2, 0x3b, 0xbe, 0x9b, 0x05, 0xa5, 0xb0, 0x38, 0xe1, + 0xad, 0x6f, 0x11, 0x9b, 0x5d, 0xc2, 0x68, 0xa3, 0xc4, 0x5d, 0x75, 0xeb, 0xec, 0x3d, 0x1e, 0x79, + 0x37, 0xed, 0xd2, 0x82, 0xcf, 0xc1, 0x9c, 0xc7, 0x30, 0xeb, 0xf3, 0x99, 0xe7, 0xa8, 0x5b, 0x17, + 0x42, 0x15, 0x9a, 0xd1, 0xa5, 0x15, 0x9e, 0x91, 0x44, 0xac, 0xfd, 0xa2, 0x80, 0xf5, 0xb4, 0xca, + 0x25, 0x14, 0xfb, 0xe3, 0x64, 0xb1, 0x6f, 0x5c, 0x24, 0xa2, 0x29, 0x05, 0xff, 0x4d, 0x01, 0xff, + 0x99, 0x08, 0x5e, 0x5c, 0x8f, 0x7c, 0x4f, 0x38, 0xa9, 0x6d, 0xb4, 0x1f, 0xdd, 0xf9, 0x62, 0x4f, + 0x1c, 0x64, 0xf0, 0x51, 0xa6, 0x16, 0x7c, 0x01, 0x4a, 0xa6, 0xdd, 0x33, 0x6d, 0x12, 0xd2, 0xda, + 0x51, 0xb9, 0x33, 0x87, 0x39, 0x8d, 0x2c, 0xca, 0xbc, 0x1e, 0xf8, 0x6a, 0xa9, 0x99, 0x42, 0x41, + 0x13, 0xb8, 0xb5, 0x5f, 0x33, 0xca, 0x23, 0xee, 0xc2, 0x1b, 0x60, 0x21, 0x7c, 0xbd, 0x12, 0x57, + 0x86, 0x31, 0x4e, 0xf7, 0xb6, 0xa4, 0xa3, 0xb1, 0x84, 0xe8, 0x20, 0x91, 0x0a, 0xe9, 0xe8, 0xc5, + 0x3a, 0x48, 0x68, 0xc6, 0x3a, 0x48, 0x9c, 0x91, 0x44, 0xe4, 0x9e, 0xf0, 0x07, 0x90, 0x48, 0x68, + 0x2e, 0xe9, 0xc9, 0xbe, 0xa4, 0xa3, 0xb1, 0x44, 0xed, 0xef, 0x5c, 0x46, 0x95, 0x44, 0x2b, 0xc6, + 0x42, 0x1a, 0x3d, 0xda, 0xd3, 0x21, 0x19, 0xe3, 0x90, 0x0c, 0xf8, 0xad, 0x02, 0x20, 0x1e, 0x43, + 0xb4, 0x46, 0xad, 0x1a, 0xf6, 0xd3, 0xe3, 0x8b, 0x4f, 0x88, 0xb6, 0x3d, 0x01, 0x16, 0xde, 0x93, + 0x15, 0xe9, 0x04, 0x9c, 0x14, 0x40, 0x19, 0x1e, 0x40, 0x13, 0x14, 0x43, 0xea, 0xae, 0xeb, 0x52, + 0x57, 0x8e, 0xec, 0xd5, 0xb3, 0x1d, 0x12, 0xe2, 0x8d, 0xaa, 0x78, 0xc8, 0x45, 0xfa, 0xa7, 0xbe, + 0x5a, 0x8c, 0xf1, 0x51, 0x1c, 0x9b, 0x9b, 0x32, 0x48, 0x64, 0x2a, 0xff, 0x2f, 0x4c, 0x3d, 0x24, + 0xd3, 0x4d, 0xc5, 0xb0, 0x2b, 0xbb, 0xe0, 0xbf, 0x53, 0x12, 0x74, 0xa1, 0x7b, 0xe5, 0x2b, 0x05, + 0xc4, 0x6d, 0xc0, 0x3d, 0x90, 0xe7, 0x3f, 0xd6, 0x72, 0xc3, 0x5c, 0x3f, 0xdf, 0x86, 0x39, 0x34, + 0x2d, 0x12, 0x2d, 0x4a, 0x7e, 0x42, 0x02, 0x05, 0x5e, 0x03, 0xf3, 0x16, 0xf1, 0x3c, 0xdc, 0x95, + 0x96, 0xa3, 0x57, 0x5f, 0x2b, 0x24, 0xa3, 0x11, 0xbf, 0x76, 0x17, 0xac, 0x65, 0xbc, 0xa3, 0xa1, + 0x0a, 0x0a, 0xba, 0xf8, 0xf3, 0xe3, 0x0e, 0x15, 0x1a, 0x8b, 0x7c, 0xcb, 0xec, 0x88, 0x1f, 0xbe, + 0x90, 0xde, 0xb8, 0xf9, 0xea, 0x6d, 0x75, 0xe6, 0xf5, 0xdb, 0xea, 0xcc, 0x9b, 0xb7, 0xd5, 0x99, + 0x2f, 0x83, 0xaa, 0xf2, 0x2a, 0xa8, 0x2a, 0xaf, 0x83, 0xaa, 0xf2, 0x26, 0xa8, 0x2a, 0xbf, 0x07, + 0x55, 0xe5, 0xeb, 0x3f, 0xaa, 0x33, 0xcf, 0xe7, 0x65, 0xbe, 0xff, 0x09, 0x00, 0x00, 0xff, 0xff, + 0x4b, 0x3f, 0x49, 0x6e, 0x6d, 0x11, 0x00, 0x00, } func (m *CSIDriver) Marshal() (dAtA []byte, err error) { @@ -717,6 +721,23 @@ func (m *CSIDriverSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + if m.FSGroupPolicy != nil { + i -= len(*m.FSGroupPolicy) + copy(dAtA[i:], *m.FSGroupPolicy) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.FSGroupPolicy))) + i-- + dAtA[i] = 0x2a + } + if m.StorageCapacity != nil { + i-- + if *m.StorageCapacity { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x20 + } if len(m.VolumeLifecycleModes) > 0 { for iNdEx := len(m.VolumeLifecycleModes) - 1; iNdEx >= 0; iNdEx-- { i -= len(m.VolumeLifecycleModes[iNdEx]) @@ -1475,6 +1496,13 @@ func (m *CSIDriverSpec) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) } } + if m.StorageCapacity != nil { + n += 2 + } + if m.FSGroupPolicy != nil { + l = len(*m.FSGroupPolicy) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -1763,6 +1791,8 @@ func (this *CSIDriverSpec) String() string { `AttachRequired:` + valueToStringGenerated(this.AttachRequired) + `,`, `PodInfoOnMount:` + valueToStringGenerated(this.PodInfoOnMount) + `,`, `VolumeLifecycleModes:` + fmt.Sprintf("%v", this.VolumeLifecycleModes) + `,`, + `StorageCapacity:` + valueToStringGenerated(this.StorageCapacity) + `,`, + `FSGroupPolicy:` + valueToStringGenerated(this.FSGroupPolicy) + `,`, `}`, }, "") return s @@ -2315,6 +2345,60 @@ func (m *CSIDriverSpec) Unmarshal(dAtA []byte) error { } m.VolumeLifecycleModes = append(m.VolumeLifecycleModes, VolumeLifecycleMode(dAtA[iNdEx:postIndex])) iNdEx = postIndex + case 4: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field StorageCapacity", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + b := bool(v != 0) + m.StorageCapacity = &b + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field FSGroupPolicy", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := FSGroupPolicy(dAtA[iNdEx:postIndex]) + m.FSGroupPolicy = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -4378,6 +4462,7 @@ func (m *VolumeNodeResources) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -4409,10 +4494,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -4433,55 +4516,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/api/storage/v1beta1/generated.proto b/vendor/k8s.io/api/storage/v1beta1/generated.proto index 373a154b..e61876ed 100644 --- a/vendor/k8s.io/api/storage/v1beta1/generated.proto +++ b/vendor/k8s.io/api/storage/v1beta1/generated.proto @@ -119,6 +119,34 @@ message CSIDriverSpec { // more modes may be added in the future. // +optional repeated string volumeLifecycleModes = 3; + + // If set to true, storageCapacity indicates that the CSI + // volume driver wants pod scheduling to consider the storage + // capacity that the driver deployment will report by creating + // CSIStorageCapacity objects with capacity information. + // + // The check can be enabled immediately when deploying a driver. + // In that case, provisioning new volumes with late binding + // will pause until the driver deployment has published + // some suitable CSIStorageCapacity object. + // + // Alternatively, the driver can be deployed with the field + // unset or false and it can be flipped later when storage + // capacity information has been published. + // + // This is an alpha field and only available when the CSIStorageCapacity + // feature is enabled. The default is false. + // + // +optional + optional bool storageCapacity = 4; + + // Defines if the underlying volume supports changing ownership and + // permission of the volume before being mounted. + // Refer to the specific FSGroupPolicy values for additional details. + // This field is alpha-level, and is only honored by servers + // that enable the CSIVolumeFSGroupPolicy feature gate. + // +optional + optional string fsGroupPolicy = 5; } // DEPRECATED - This group version of CSINode is deprecated by storage/v1/CSINode. diff --git a/vendor/k8s.io/api/storage/v1beta1/types.go b/vendor/k8s.io/api/storage/v1beta1/types.go index a8faeb9d..7946663a 100644 --- a/vendor/k8s.io/api/storage/v1beta1/types.go +++ b/vendor/k8s.io/api/storage/v1beta1/types.go @@ -24,6 +24,9 @@ import ( // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.4 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,StorageClass // StorageClass describes the parameters for a class of storage for // which PersistentVolumes can be dynamically provisioned. @@ -75,6 +78,9 @@ type StorageClass struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.4 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,StorageClassList // StorageClassList is a collection of storage classes. type StorageClassList struct { @@ -106,6 +112,9 @@ const ( // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.10 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,VolumeAttachment // VolumeAttachment captures the intent to attach or detach the specified volume // to/from the specified node. @@ -131,6 +140,9 @@ type VolumeAttachment struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.10 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,VolumeAttachmentList // VolumeAttachmentList is a collection of VolumeAttachment objects. type VolumeAttachmentList struct { @@ -220,6 +232,9 @@ type VolumeError struct { // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.14 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,CSIDriver // CSIDriver captures information about a Container Storage Interface (CSI) // volume driver deployed on the cluster. @@ -247,6 +262,9 @@ type CSIDriver struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.14 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,CSIDriverList // CSIDriverList is a collection of CSIDriver objects. type CSIDriverList struct { @@ -317,8 +335,62 @@ type CSIDriverSpec struct { // more modes may be added in the future. // +optional VolumeLifecycleModes []VolumeLifecycleMode `json:"volumeLifecycleModes,omitempty" protobuf:"bytes,3,opt,name=volumeLifecycleModes"` + + // If set to true, storageCapacity indicates that the CSI + // volume driver wants pod scheduling to consider the storage + // capacity that the driver deployment will report by creating + // CSIStorageCapacity objects with capacity information. + // + // + // The check can be enabled immediately when deploying a driver. + // In that case, provisioning new volumes with late binding + // will pause until the driver deployment has published + // some suitable CSIStorageCapacity object. + // + // Alternatively, the driver can be deployed with the field + // unset or false and it can be flipped later when storage + // capacity information has been published. + // + // This is an alpha field and only available when the CSIStorageCapacity + // feature is enabled. The default is false. + // + // +optional + StorageCapacity *bool `json:"storageCapacity,omitempty" protobuf:"bytes,4,opt,name=storageCapacity"` + + // Defines if the underlying volume supports changing ownership and + // permission of the volume before being mounted. + // Refer to the specific FSGroupPolicy values for additional details. + // This field is alpha-level, and is only honored by servers + // that enable the CSIVolumeFSGroupPolicy feature gate. + // +optional + FSGroupPolicy *FSGroupPolicy `json:"fsGroupPolicy,omitempty" protobuf:"bytes,5,opt,name=fsGroupPolicy"` } +// FSGroupPolicy specifies if a CSI Driver supports modifying +// volume ownership and permissions of the volume to be mounted. +// More modes may be added in the future. +type FSGroupPolicy string + +const ( + // ReadWriteOnceWithFSTypeFSGroupPolicy indicates that each volume will be examined + // to determine if the volume ownership and permissions + // should be modified. If a fstype is defined and the volume's access mode + // contains ReadWriteOnce, then the defined fsGroup will be applied. + // This is the default behavior if no other FSGroupPolicy is defined. + ReadWriteOnceWithFSTypeFSGroupPolicy FSGroupPolicy = "ReadWriteOnceWithFSType" + + // FileFSGroupPolicy indicates that CSI driver supports volume ownership + // and permission change via fsGroup, and Kubernetes may use fsGroup + // to change permissions and ownership of the volume to match user requested fsGroup in + // the pod's SecurityPolicy regardless of fstype or access mode. + FileFSGroupPolicy FSGroupPolicy = "File" + + // None indicates that volumes will be mounted without performing + // any ownership or permission modifications, as the CSIDriver does not support + // these operations. + NoneFSGroupPolicy FSGroupPolicy = "None" +) + // VolumeLifecycleMode is an enumeration of possible usage modes for a volume // provided by a CSI driver. More modes may be added in the future. type VolumeLifecycleMode string @@ -347,6 +419,10 @@ const ( // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.14 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,CSINode // DEPRECATED - This group version of CSINode is deprecated by storage/v1/CSINode. // See the release notes for more information. @@ -425,6 +501,10 @@ type VolumeNodeResources struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.14 +// +k8s:prerelease-lifecycle-gen:deprecated=1.17 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=storage.k8s.io,v1,CSINode // CSINodeList is a collection of CSINode objects. type CSINodeList struct { diff --git a/vendor/k8s.io/api/storage/v1beta1/types_swagger_doc_generated.go b/vendor/k8s.io/api/storage/v1beta1/types_swagger_doc_generated.go index 53fa666b..60cc4c6a 100644 --- a/vendor/k8s.io/api/storage/v1beta1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/api/storage/v1beta1/types_swagger_doc_generated.go @@ -52,6 +52,8 @@ var map_CSIDriverSpec = map[string]string{ "attachRequired": "attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called.", "podInfoOnMount": "If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. \"csi.storage.k8s.io/pod.name\": pod.Name \"csi.storage.k8s.io/pod.namespace\": pod.Namespace \"csi.storage.k8s.io/pod.uid\": string(pod.UID) \"csi.storage.k8s.io/ephemeral\": \"true\" iff the volume is an ephemeral inline volume\n defined by a CSIVolumeSource, otherwise \"false\"\n\n\"csi.storage.k8s.io/ephemeral\" is a new feature in Kubernetes 1.16. It is only required for drivers which support both the \"Persistent\" and \"Ephemeral\" VolumeLifecycleMode. Other drivers can leave pod info disabled and/or ignore this field. As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when deployed on such a cluster and the deployment determines which mode that is, for example via a command line parameter of the driver.", "volumeLifecycleModes": "VolumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is \"Persistent\", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is \"Ephemeral\". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future.", + "storageCapacity": "If set to true, storageCapacity indicates that the CSI volume driver wants pod scheduling to consider the storage capacity that the driver deployment will report by creating CSIStorageCapacity objects with capacity information.\n\nThe check can be enabled immediately when deploying a driver. In that case, provisioning new volumes with late binding will pause until the driver deployment has published some suitable CSIStorageCapacity object.\n\nAlternatively, the driver can be deployed with the field unset or false and it can be flipped later when storage capacity information has been published.\n\nThis is an alpha field and only available when the CSIStorageCapacity feature is enabled. The default is false.", + "fsGroupPolicy": "Defines if the underlying volume supports changing ownership and permission of the volume before being mounted. Refer to the specific FSGroupPolicy values for additional details. This field is alpha-level, and is only honored by servers that enable the CSIVolumeFSGroupPolicy feature gate.", } func (CSIDriverSpec) SwaggerDoc() map[string]string { diff --git a/vendor/k8s.io/api/storage/v1beta1/zz_generated.deepcopy.go b/vendor/k8s.io/api/storage/v1beta1/zz_generated.deepcopy.go index 52433fcd..a1538c13 100644 --- a/vendor/k8s.io/api/storage/v1beta1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/api/storage/v1beta1/zz_generated.deepcopy.go @@ -103,6 +103,16 @@ func (in *CSIDriverSpec) DeepCopyInto(out *CSIDriverSpec) { *out = make([]VolumeLifecycleMode, len(*in)) copy(*out, *in) } + if in.StorageCapacity != nil { + in, out := &in.StorageCapacity, &out.StorageCapacity + *out = new(bool) + **out = **in + } + if in.FSGroupPolicy != nil { + in, out := &in.FSGroupPolicy, &out.FSGroupPolicy + *out = new(FSGroupPolicy) + **out = **in + } return } diff --git a/vendor/k8s.io/api/storage/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/api/storage/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..8bb3de20 --- /dev/null +++ b/vendor/k8s.io/api/storage/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,217 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CSIDriver) APILifecycleIntroduced() (major, minor int) { + return 1, 14 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CSIDriver) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CSIDriver) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "CSIDriver"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CSIDriver) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CSIDriverList) APILifecycleIntroduced() (major, minor int) { + return 1, 14 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CSIDriverList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CSIDriverList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "CSIDriverList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CSIDriverList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CSINode) APILifecycleIntroduced() (major, minor int) { + return 1, 14 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CSINode) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CSINode) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "CSINode"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CSINode) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CSINodeList) APILifecycleIntroduced() (major, minor int) { + return 1, 14 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CSINodeList) APILifecycleDeprecated() (major, minor int) { + return 1, 17 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CSINodeList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "CSINode"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CSINodeList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *StorageClass) APILifecycleIntroduced() (major, minor int) { + return 1, 4 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *StorageClass) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *StorageClass) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "StorageClass"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *StorageClass) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *StorageClassList) APILifecycleIntroduced() (major, minor int) { + return 1, 4 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *StorageClassList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *StorageClassList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "StorageClassList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *StorageClassList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *VolumeAttachment) APILifecycleIntroduced() (major, minor int) { + return 1, 10 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *VolumeAttachment) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *VolumeAttachment) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "VolumeAttachment"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *VolumeAttachment) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *VolumeAttachmentList) APILifecycleIntroduced() (major, minor int) { + return 1, 10 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *VolumeAttachmentList) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *VolumeAttachmentList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "storage.k8s.io", Version: "v1", Kind: "VolumeAttachmentList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *VolumeAttachmentList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/types.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/types.go index 8f502e8b..ee77a422 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/types.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/types.go @@ -177,6 +177,15 @@ type CustomResourceDefinitionVersion struct { // Storage flags the version as storage version. There must be exactly one flagged // as storage version. Storage bool + // deprecated indicates this version of the custom resource API is deprecated. + // When set to true, API requests to this version receive a warning header in the server response. + // Defaults to false. + Deprecated bool + // deprecationWarning overrides the default warning returned to API clients. + // May only be set when `deprecated` is true. + // The default warning indicates this version is deprecated and recommends use + // of the newest served version of equal or greater stability, if one exists. + DeprecationWarning *string // Schema describes the schema for CustomResource used in validation, pruning, and defaulting. // Top-level and per-version schemas are mutually exclusive. // Per-version schemas must not all be set to identical values (top-level validation schema should be used instead) diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/.import-restrictions b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/.import-restrictions new file mode 100644 index 00000000..7408dd12 --- /dev/null +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/.import-restrictions @@ -0,0 +1,5 @@ +inverseRules: + # Allow use of this package in all k8s.io packages. + - selectorRegexp: k8s[.]io + allowedPrefixes: + - '' diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/doc.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/doc.go index 3f631dd4..7a92cb8b 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/doc.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/doc.go @@ -19,6 +19,7 @@ limitations under the License. // +k8s:conversion-gen=k8s.io/apiextensions-apiserver/pkg/apis/apiextensions // +k8s:defaulter-gen=TypeMeta // +k8s:openapi-gen=true +// +k8s:prerelease-lifecycle-gen=true // +groupName=apiextensions.k8s.io // Package v1beta1 is the v1beta1 version of the API. diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.pb.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.pb.go index 6e11dcc9..0382b044 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.pb.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.pb.go @@ -46,7 +46,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *ConversionRequest) Reset() { *m = ConversionRequest{} } func (*ConversionRequest) ProtoMessage() {} @@ -756,193 +756,196 @@ func init() { } var fileDescriptor_98a4cc6918394e53 = []byte{ - // 2976 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x5a, 0xcf, 0x73, 0x23, 0x47, - 0xf5, 0xdf, 0x91, 0x2c, 0x5b, 0x6e, 0xdb, 0x6b, 0xbb, 0x77, 0xed, 0xcc, 0x3a, 0x1b, 0xcb, 0xab, - 0x7c, 0xb3, 0x5f, 0x27, 0xd9, 0x95, 0x93, 0x25, 0x21, 0x21, 0x05, 0x45, 0x59, 0xb6, 0x13, 0x9c, - 0xac, 0x2d, 0xd3, 0xda, 0x4d, 0x0c, 0xf9, 0xd9, 0xd6, 0xb4, 0xe4, 0x59, 0xcf, 0xaf, 0x9d, 0x9e, - 0x91, 0xed, 0x0a, 0x50, 0xfc, 0xa8, 0x14, 0x14, 0x05, 0x84, 0x22, 0xb9, 0x50, 0x05, 0x87, 0x40, - 0x71, 0xe1, 0x00, 0x07, 0x28, 0x2e, 0xf0, 0x07, 0xe4, 0x98, 0xe2, 0x94, 0x03, 0xa5, 0x22, 0xca, - 0x95, 0x23, 0x55, 0x54, 0xf9, 0x44, 0xf5, 0x8f, 0xe9, 0x19, 0x8d, 0xa4, 0x5d, 0x57, 0x56, 0xca, - 0x72, 0xb3, 0xde, 0xaf, 0xcf, 0xeb, 0xd7, 0xaf, 0x5f, 0xbf, 0x7e, 0x63, 0x50, 0x3f, 0x78, 0x96, - 0x96, 0x4c, 0x77, 0xe5, 0x20, 0xdc, 0x23, 0xbe, 0x43, 0x02, 0x42, 0x57, 0x9a, 0xc4, 0x31, 0x5c, - 0x7f, 0x45, 0x32, 0xb0, 0x67, 0x92, 0xa3, 0x80, 0x38, 0xd4, 0x74, 0x1d, 0x7a, 0x15, 0x7b, 0x26, - 0x25, 0x7e, 0x93, 0xf8, 0x2b, 0xde, 0x41, 0x83, 0xf1, 0x68, 0xa7, 0xc0, 0x4a, 0xf3, 0xc9, 0x3d, - 0x12, 0xe0, 0x27, 0x57, 0x1a, 0xc4, 0x21, 0x3e, 0x0e, 0x88, 0x51, 0xf2, 0x7c, 0x37, 0x70, 0xe1, - 0x57, 0x84, 0xb9, 0x52, 0x87, 0xf4, 0x9b, 0xca, 0x5c, 0xc9, 0x3b, 0x68, 0x30, 0x1e, 0xed, 0x14, - 0x28, 0x49, 0x73, 0x0b, 0x57, 0x1b, 0x66, 0xb0, 0x1f, 0xee, 0x95, 0x6a, 0xae, 0xbd, 0xd2, 0x70, - 0x1b, 0xee, 0x0a, 0xb7, 0xba, 0x17, 0xd6, 0xf9, 0x2f, 0xfe, 0x83, 0xff, 0x25, 0xd0, 0x16, 0x9e, - 0x8a, 0x9d, 0xb7, 0x71, 0x6d, 0xdf, 0x74, 0x88, 0x7f, 0x1c, 0x7b, 0x6c, 0x93, 0x00, 0xaf, 0x34, - 0xbb, 0x7c, 0x5c, 0x58, 0xe9, 0xa7, 0xe5, 0x87, 0x4e, 0x60, 0xda, 0xa4, 0x4b, 0xe1, 0x8b, 0x77, - 0x53, 0xa0, 0xb5, 0x7d, 0x62, 0xe3, 0xb4, 0x5e, 0xf1, 0x44, 0x03, 0xb3, 0x6b, 0xae, 0xd3, 0x24, - 0x3e, 0x5b, 0x25, 0x22, 0xb7, 0x43, 0x42, 0x03, 0x58, 0x06, 0xd9, 0xd0, 0x34, 0x74, 0x6d, 0x49, - 0x5b, 0x1e, 0x2f, 0x3f, 0xf1, 0x61, 0xab, 0x70, 0xa6, 0xdd, 0x2a, 0x64, 0x6f, 0x6e, 0xae, 0x9f, - 0xb4, 0x0a, 0x97, 0xfa, 0x21, 0x05, 0xc7, 0x1e, 0xa1, 0xa5, 0x9b, 0x9b, 0xeb, 0x88, 0x29, 0xc3, - 0x17, 0xc0, 0xac, 0x41, 0xa8, 0xe9, 0x13, 0x63, 0x75, 0x67, 0xf3, 0x65, 0x61, 0x5f, 0xcf, 0x70, - 0x8b, 0x17, 0xa4, 0xc5, 0xd9, 0xf5, 0xb4, 0x00, 0xea, 0xd6, 0x81, 0xbb, 0x60, 0xcc, 0xdd, 0xbb, - 0x45, 0x6a, 0x01, 0xd5, 0xb3, 0x4b, 0xd9, 0xe5, 0x89, 0x6b, 0x57, 0x4b, 0xf1, 0x0e, 0x2a, 0x17, - 0xf8, 0xb6, 0xc9, 0xc5, 0x96, 0x10, 0x3e, 0xdc, 0x88, 0x76, 0xae, 0x3c, 0x2d, 0xd1, 0xc6, 0x2a, - 0xc2, 0x0a, 0x8a, 0xcc, 0x15, 0x7f, 0x9b, 0x01, 0x30, 0xb9, 0x78, 0xea, 0xb9, 0x0e, 0x25, 0x03, - 0x59, 0x3d, 0x05, 0x33, 0x35, 0x6e, 0x39, 0x20, 0x86, 0xc4, 0xd5, 0x33, 0x9f, 0xc5, 0x7b, 0x5d, - 0xe2, 0xcf, 0xac, 0xa5, 0xcc, 0xa1, 0x2e, 0x00, 0x78, 0x03, 0x8c, 0xfa, 0x84, 0x86, 0x56, 0xa0, - 0x67, 0x97, 0xb4, 0xe5, 0x89, 0x6b, 0x57, 0xfa, 0x42, 0xf1, 0xfc, 0x66, 0xc9, 0x57, 0x6a, 0x3e, - 0x59, 0xaa, 0x06, 0x38, 0x08, 0x69, 0xf9, 0xac, 0x44, 0x1a, 0x45, 0xdc, 0x06, 0x92, 0xb6, 0x8a, - 0x3f, 0xca, 0x80, 0x99, 0x64, 0x94, 0x9a, 0x26, 0x39, 0x84, 0x87, 0x60, 0xcc, 0x17, 0xc9, 0xc2, - 0xe3, 0x34, 0x71, 0x6d, 0xa7, 0x74, 0x4f, 0xc7, 0xaa, 0xd4, 0x95, 0x84, 0xe5, 0x09, 0xb6, 0x67, - 0xf2, 0x07, 0x8a, 0xd0, 0xe0, 0xdb, 0x20, 0xef, 0xcb, 0x8d, 0xe2, 0xd9, 0x34, 0x71, 0xed, 0xeb, - 0x03, 0x44, 0x16, 0x86, 0xcb, 0x93, 0xed, 0x56, 0x21, 0x1f, 0xfd, 0x42, 0x0a, 0xb0, 0xf8, 0x5e, - 0x06, 0x2c, 0xae, 0x85, 0x34, 0x70, 0x6d, 0x44, 0xa8, 0x1b, 0xfa, 0x35, 0xb2, 0xe6, 0x5a, 0xa1, - 0xed, 0xac, 0x93, 0xba, 0xe9, 0x98, 0x01, 0xcb, 0xd6, 0x25, 0x30, 0xe2, 0x60, 0x9b, 0xc8, 0xec, - 0x99, 0x94, 0x31, 0x1d, 0xd9, 0xc6, 0x36, 0x41, 0x9c, 0xc3, 0x24, 0x58, 0xb2, 0xc8, 0xb3, 0xa0, - 0x24, 0x6e, 0x1c, 0x7b, 0x04, 0x71, 0x0e, 0xbc, 0x0c, 0x46, 0xeb, 0xae, 0x6f, 0x63, 0xb1, 0x8f, - 0xe3, 0xf1, 0xce, 0x3c, 0xcf, 0xa9, 0x48, 0x72, 0xe1, 0xd3, 0x60, 0xc2, 0x20, 0xb4, 0xe6, 0x9b, - 0x1e, 0x83, 0xd6, 0x47, 0xb8, 0xf0, 0x39, 0x29, 0x3c, 0xb1, 0x1e, 0xb3, 0x50, 0x52, 0x0e, 0x5e, - 0x01, 0x79, 0xcf, 0x37, 0x5d, 0xdf, 0x0c, 0x8e, 0xf5, 0xdc, 0x92, 0xb6, 0x9c, 0x2b, 0xcf, 0x48, - 0x9d, 0xfc, 0x8e, 0xa4, 0x23, 0x25, 0x01, 0x97, 0x40, 0xfe, 0xc5, 0x6a, 0x65, 0x7b, 0x07, 0x07, - 0xfb, 0xfa, 0x28, 0x47, 0x18, 0x61, 0xd2, 0x28, 0x7f, 0x4b, 0x52, 0x8b, 0xff, 0xc8, 0x00, 0x3d, - 0x1d, 0x95, 0x28, 0xa4, 0xf0, 0x79, 0x90, 0xa7, 0x01, 0xab, 0x38, 0x8d, 0x63, 0x19, 0x93, 0xc7, - 0x22, 0xb0, 0xaa, 0xa4, 0x9f, 0xb4, 0x0a, 0xf3, 0xb1, 0x46, 0x44, 0xe5, 0xf1, 0x50, 0xba, 0xf0, - 0xd7, 0x1a, 0x38, 0x77, 0x48, 0xf6, 0xf6, 0x5d, 0xf7, 0x60, 0xcd, 0x32, 0x89, 0x13, 0xac, 0xb9, - 0x4e, 0xdd, 0x6c, 0xc8, 0x1c, 0x40, 0xf7, 0x98, 0x03, 0xaf, 0x74, 0x5b, 0x2e, 0x3f, 0xd0, 0x6e, - 0x15, 0xce, 0xf5, 0x60, 0xa0, 0x5e, 0x7e, 0xc0, 0x5d, 0xa0, 0xd7, 0x52, 0x87, 0x44, 0x16, 0x30, - 0x51, 0xb6, 0xc6, 0xcb, 0x17, 0xdb, 0xad, 0x82, 0xbe, 0xd6, 0x47, 0x06, 0xf5, 0xd5, 0x2e, 0xfe, - 0x20, 0x9b, 0x0e, 0x6f, 0x22, 0xdd, 0xde, 0x02, 0x79, 0x76, 0x8c, 0x0d, 0x1c, 0x60, 0x79, 0x10, - 0x9f, 0x38, 0xdd, 0xa1, 0x17, 0x35, 0x63, 0x8b, 0x04, 0xb8, 0x0c, 0xe5, 0x86, 0x80, 0x98, 0x86, - 0x94, 0x55, 0xf8, 0x6d, 0x30, 0x42, 0x3d, 0x52, 0x93, 0x81, 0x7e, 0xf5, 0x5e, 0x0f, 0x5b, 0x9f, - 0x85, 0x54, 0x3d, 0x52, 0x8b, 0xcf, 0x02, 0xfb, 0x85, 0x38, 0x2c, 0x7c, 0x47, 0x03, 0xa3, 0x94, - 0x17, 0x28, 0x59, 0xd4, 0x5e, 0x1f, 0x96, 0x07, 0xa9, 0x2a, 0x28, 0x7e, 0x23, 0x09, 0x5e, 0xfc, - 0x77, 0x06, 0x5c, 0xea, 0xa7, 0xba, 0xe6, 0x3a, 0x86, 0xd8, 0x8e, 0x4d, 0x79, 0xb6, 0x45, 0xa6, - 0x3f, 0x9d, 0x3c, 0xdb, 0x27, 0xad, 0xc2, 0x23, 0x77, 0x35, 0x90, 0x28, 0x02, 0x5f, 0x52, 0xeb, - 0x16, 0x85, 0xe2, 0x52, 0xa7, 0x63, 0x27, 0xad, 0xc2, 0xb4, 0x52, 0xeb, 0xf4, 0x15, 0x36, 0x01, - 0xb4, 0x30, 0x0d, 0x6e, 0xf8, 0xd8, 0xa1, 0xc2, 0xac, 0x69, 0x13, 0x19, 0xbe, 0xc7, 0x4e, 0x97, - 0x1e, 0x4c, 0xa3, 0xbc, 0x20, 0x21, 0xe1, 0xf5, 0x2e, 0x6b, 0xa8, 0x07, 0x02, 0xab, 0x5b, 0x3e, - 0xc1, 0x54, 0x95, 0xa2, 0xc4, 0x8d, 0xc2, 0xa8, 0x48, 0x72, 0xe1, 0xa3, 0x60, 0xcc, 0x26, 0x94, - 0xe2, 0x06, 0xe1, 0xf5, 0x67, 0x3c, 0xbe, 0xa2, 0xb7, 0x04, 0x19, 0x45, 0x7c, 0xd6, 0x9f, 0x5c, - 0xec, 0x17, 0xb5, 0xeb, 0x26, 0x0d, 0xe0, 0x6b, 0x5d, 0x07, 0xa0, 0x74, 0xba, 0x15, 0x32, 0x6d, - 0x9e, 0xfe, 0xaa, 0xf8, 0x45, 0x94, 0x44, 0xf2, 0x7f, 0x0b, 0xe4, 0xcc, 0x80, 0xd8, 0xd1, 0xdd, - 0xfd, 0xca, 0x90, 0x72, 0xaf, 0x3c, 0x25, 0x7d, 0xc8, 0x6d, 0x32, 0x34, 0x24, 0x40, 0x8b, 0xbf, - 0xcb, 0x80, 0x87, 0xfa, 0xa9, 0xb0, 0x0b, 0x85, 0xb2, 0x88, 0x7b, 0x56, 0xe8, 0x63, 0x4b, 0x66, - 0x9c, 0x8a, 0xf8, 0x0e, 0xa7, 0x22, 0xc9, 0x65, 0x25, 0x9f, 0x9a, 0x4e, 0x23, 0xb4, 0xb0, 0x2f, - 0xd3, 0x49, 0xad, 0xba, 0x2a, 0xe9, 0x48, 0x49, 0xc0, 0x12, 0x00, 0x74, 0xdf, 0xf5, 0x03, 0x8e, - 0x21, 0xab, 0xd7, 0x59, 0x56, 0x20, 0xaa, 0x8a, 0x8a, 0x12, 0x12, 0xec, 0x46, 0x3b, 0x30, 0x1d, - 0x43, 0xee, 0xba, 0x3a, 0xc5, 0x2f, 0x99, 0x8e, 0x81, 0x38, 0x87, 0xe1, 0x5b, 0x26, 0x0d, 0x18, - 0x45, 0x6e, 0x79, 0x47, 0xd4, 0xb9, 0xa4, 0x92, 0x60, 0xf8, 0x35, 0x56, 0xf5, 0x5d, 0xdf, 0x24, - 0x54, 0x1f, 0x8d, 0xf1, 0xd7, 0x14, 0x15, 0x25, 0x24, 0x8a, 0xff, 0xca, 0xf7, 0x4f, 0x12, 0x56, - 0x4a, 0xe0, 0xc3, 0x20, 0xd7, 0xf0, 0xdd, 0xd0, 0x93, 0x51, 0x52, 0xd1, 0x7e, 0x81, 0x11, 0x91, - 0xe0, 0xb1, 0xac, 0x6c, 0x76, 0xb4, 0xa9, 0x2a, 0x2b, 0xa3, 0xe6, 0x34, 0xe2, 0xc3, 0xef, 0x69, - 0x20, 0xe7, 0xc8, 0xe0, 0xb0, 0x94, 0x7b, 0x6d, 0x48, 0x79, 0xc1, 0xc3, 0x1b, 0xbb, 0x2b, 0x22, - 0x2f, 0x90, 0xe1, 0x53, 0x20, 0x47, 0x6b, 0xae, 0x47, 0x64, 0xd4, 0x17, 0x23, 0xa1, 0x2a, 0x23, - 0x9e, 0xb4, 0x0a, 0x53, 0x91, 0x39, 0x4e, 0x40, 0x42, 0x18, 0xfe, 0x50, 0x03, 0xa0, 0x89, 0x2d, - 0xd3, 0xc0, 0xbc, 0x65, 0xc8, 0x71, 0xf7, 0x07, 0x9b, 0xd6, 0x2f, 0x2b, 0xf3, 0x62, 0xd3, 0xe2, - 0xdf, 0x28, 0x01, 0x0d, 0xdf, 0xd5, 0xc0, 0x24, 0x0d, 0xf7, 0x7c, 0xa9, 0x45, 0x79, 0x73, 0x31, - 0x71, 0xed, 0x1b, 0x03, 0xf5, 0xa5, 0x9a, 0x00, 0x28, 0xcf, 0xb4, 0x5b, 0x85, 0xc9, 0x24, 0x05, - 0x75, 0x38, 0x00, 0x7f, 0xa2, 0x81, 0x7c, 0x33, 0xba, 0xb3, 0xc7, 0xf8, 0x81, 0x7f, 0x63, 0x48, - 0x1b, 0x2b, 0x33, 0x2a, 0x3e, 0x05, 0xaa, 0x0f, 0x50, 0x1e, 0xc0, 0xbf, 0x6a, 0x40, 0xc7, 0x86, - 0x28, 0xf0, 0xd8, 0xda, 0xf1, 0x4d, 0x27, 0x20, 0xbe, 0xe8, 0x37, 0xa9, 0x9e, 0xe7, 0xee, 0x0d, - 0xf6, 0x2e, 0x4c, 0xf7, 0xb2, 0xe5, 0x25, 0xe9, 0x9d, 0xbe, 0xda, 0xc7, 0x0d, 0xd4, 0xd7, 0x41, - 0x9e, 0x68, 0x71, 0x4b, 0xa3, 0x8f, 0x0f, 0x21, 0xd1, 0xe2, 0x5e, 0x4a, 0x56, 0x87, 0xb8, 0x83, - 0x4a, 0x40, 0xc3, 0x0a, 0x98, 0xf3, 0x7c, 0xc2, 0x01, 0x6e, 0x3a, 0x07, 0x8e, 0x7b, 0xe8, 0x3c, - 0x6f, 0x12, 0xcb, 0xa0, 0x3a, 0x58, 0xd2, 0x96, 0xf3, 0xe5, 0x0b, 0xed, 0x56, 0x61, 0x6e, 0xa7, - 0x97, 0x00, 0xea, 0xad, 0x57, 0x7c, 0x37, 0x9b, 0x7e, 0x05, 0xa4, 0xbb, 0x08, 0xf8, 0xbe, 0x58, - 0xbd, 0x88, 0x0d, 0xd5, 0x35, 0xbe, 0x5b, 0x6f, 0x0d, 0x29, 0x99, 0x54, 0x1b, 0x10, 0x77, 0x72, - 0x8a, 0x44, 0x51, 0xc2, 0x0f, 0xf8, 0x4b, 0x0d, 0x4c, 0xe1, 0x5a, 0x8d, 0x78, 0x01, 0x31, 0x44, - 0x71, 0xcf, 0x7c, 0x0e, 0xf5, 0x6b, 0x4e, 0x7a, 0x35, 0xb5, 0x9a, 0x84, 0x46, 0x9d, 0x9e, 0xc0, - 0xe7, 0xc0, 0x59, 0x1a, 0xb8, 0x3e, 0x31, 0x52, 0x6d, 0x33, 0x6c, 0xb7, 0x0a, 0x67, 0xab, 0x1d, - 0x1c, 0x94, 0x92, 0x2c, 0x7e, 0x3a, 0x02, 0x0a, 0x77, 0x39, 0x6a, 0xa7, 0x78, 0x98, 0x5d, 0x06, - 0xa3, 0x7c, 0xb9, 0x06, 0x8f, 0x4a, 0x3e, 0xd1, 0x0a, 0x72, 0x2a, 0x92, 0x5c, 0x76, 0x51, 0x30, - 0x7c, 0xd6, 0xbe, 0x64, 0xb9, 0xa0, 0xba, 0x28, 0xaa, 0x82, 0x8c, 0x22, 0x3e, 0x7c, 0x1b, 0x8c, - 0x8a, 0xc1, 0x0b, 0xaf, 0xd2, 0x43, 0xac, 0xb4, 0x80, 0xfb, 0xc9, 0xa1, 0x90, 0x84, 0xec, 0xae, - 0xb0, 0xb9, 0xfb, 0x5d, 0x61, 0xef, 0x58, 0xd2, 0x46, 0xff, 0xc7, 0x4b, 0x5a, 0xf1, 0x3f, 0x5a, - 0xfa, 0xdc, 0x27, 0x96, 0x5a, 0xad, 0x61, 0x8b, 0xc0, 0x75, 0x30, 0xc3, 0x5e, 0x2d, 0x88, 0x78, - 0x96, 0x59, 0xc3, 0x94, 0x3f, 0x9a, 0x45, 0xc2, 0xa9, 0x39, 0x4e, 0x35, 0xc5, 0x47, 0x5d, 0x1a, - 0xf0, 0x45, 0x00, 0x45, 0x27, 0xdf, 0x61, 0x47, 0x34, 0x25, 0xaa, 0x27, 0xaf, 0x76, 0x49, 0xa0, - 0x1e, 0x5a, 0x70, 0x0d, 0xcc, 0x5a, 0x78, 0x8f, 0x58, 0x55, 0x62, 0x91, 0x5a, 0xe0, 0xfa, 0xdc, - 0x94, 0x18, 0x2b, 0xcc, 0xb5, 0x5b, 0x85, 0xd9, 0xeb, 0x69, 0x26, 0xea, 0x96, 0x2f, 0x5e, 0x4a, - 0x1f, 0xaf, 0xe4, 0xc2, 0xc5, 0xfb, 0xe8, 0x83, 0x0c, 0x58, 0xe8, 0x9f, 0x19, 0xf0, 0xfb, 0xf1, - 0x33, 0x4e, 0x74, 0xe9, 0x6f, 0x0c, 0x2b, 0x0b, 0xe5, 0x3b, 0x0e, 0x74, 0xbf, 0xe1, 0xe0, 0x77, - 0x58, 0xcb, 0x84, 0xad, 0x68, 0x70, 0xf4, 0xfa, 0xd0, 0x5c, 0x60, 0x20, 0xe5, 0x71, 0xd1, 0x8d, - 0x61, 0x8b, 0x37, 0x5f, 0xd8, 0x22, 0xc5, 0xdf, 0x6b, 0xe9, 0x97, 0x7c, 0x7c, 0x82, 0xe1, 0x4f, - 0x35, 0x30, 0xed, 0x7a, 0xc4, 0x59, 0xdd, 0xd9, 0x7c, 0xf9, 0x0b, 0xe2, 0x24, 0xcb, 0x50, 0x6d, - 0xdf, 0xa3, 0x9f, 0x2f, 0x56, 0x2b, 0xdb, 0xc2, 0xe0, 0x8e, 0xef, 0x7a, 0xb4, 0x7c, 0xae, 0xdd, - 0x2a, 0x4c, 0x57, 0x3a, 0xa1, 0x50, 0x1a, 0xbb, 0x68, 0x83, 0xb9, 0x8d, 0xa3, 0x80, 0xf8, 0x0e, - 0xb6, 0xd6, 0xdd, 0x5a, 0x68, 0x13, 0x27, 0x10, 0x8e, 0xa6, 0xa6, 0x4e, 0xda, 0x29, 0xa7, 0x4e, - 0x0f, 0x81, 0x6c, 0xe8, 0x5b, 0x32, 0x8b, 0x27, 0xd4, 0x54, 0x15, 0x5d, 0x47, 0x8c, 0x5e, 0xbc, - 0x04, 0x46, 0x98, 0x9f, 0xf0, 0x02, 0xc8, 0xfa, 0xf8, 0x90, 0x5b, 0x9d, 0x2c, 0x8f, 0x31, 0x11, - 0x84, 0x0f, 0x11, 0xa3, 0x15, 0xff, 0xb2, 0x04, 0xa6, 0x53, 0x6b, 0x81, 0x0b, 0x20, 0xa3, 0x46, - 0xb5, 0x40, 0x1a, 0xcd, 0x6c, 0xae, 0xa3, 0x8c, 0x69, 0xc0, 0x67, 0x54, 0xf1, 0x15, 0xa0, 0x05, - 0x55, 0xcf, 0x39, 0x95, 0xf5, 0xc8, 0xb1, 0x39, 0xe6, 0x48, 0x54, 0x38, 0x99, 0x0f, 0xa4, 0x2e, - 0x4f, 0x89, 0xf0, 0x81, 0xd4, 0x11, 0xa3, 0x7d, 0xd6, 0x91, 0x5b, 0x34, 0xf3, 0xcb, 0x9d, 0x62, - 0xe6, 0x37, 0x7a, 0xc7, 0x99, 0xdf, 0xc3, 0x20, 0x17, 0x98, 0x81, 0x45, 0xf4, 0xb1, 0xce, 0xa7, - 0xcc, 0x0d, 0x46, 0x44, 0x82, 0x07, 0x6f, 0x81, 0x31, 0x83, 0xd4, 0x71, 0x68, 0x05, 0x7a, 0x9e, - 0xa7, 0xd0, 0xda, 0x00, 0x52, 0x48, 0x0c, 0x64, 0xd7, 0x85, 0x5d, 0x14, 0x01, 0xc0, 0x47, 0xc0, - 0x98, 0x8d, 0x8f, 0x4c, 0x3b, 0xb4, 0x79, 0x93, 0xa7, 0x09, 0xb1, 0x2d, 0x41, 0x42, 0x11, 0x8f, - 0x55, 0x46, 0x72, 0x54, 0xb3, 0x42, 0x6a, 0x36, 0x89, 0x64, 0xca, 0x06, 0x4c, 0x55, 0xc6, 0x8d, - 0x14, 0x1f, 0x75, 0x69, 0x70, 0x30, 0xd3, 0xe1, 0xca, 0x13, 0x09, 0x30, 0x41, 0x42, 0x11, 0xaf, - 0x13, 0x4c, 0xca, 0x4f, 0xf6, 0x03, 0x93, 0xca, 0x5d, 0x1a, 0xf0, 0x71, 0x30, 0x6e, 0xe3, 0xa3, - 0xeb, 0xc4, 0x69, 0x04, 0xfb, 0xfa, 0xd4, 0x92, 0xb6, 0x9c, 0x2d, 0x4f, 0xb5, 0x5b, 0x85, 0xf1, - 0xad, 0x88, 0x88, 0x62, 0x3e, 0x17, 0x36, 0x1d, 0x29, 0x7c, 0x36, 0x21, 0x1c, 0x11, 0x51, 0xcc, - 0x67, 0x1d, 0x84, 0x87, 0x03, 0x76, 0xb8, 0xf4, 0xe9, 0xce, 0xa7, 0xe6, 0x8e, 0x20, 0xa3, 0x88, - 0x0f, 0x97, 0x41, 0xde, 0xc6, 0x47, 0x7c, 0x2c, 0xa0, 0xcf, 0x70, 0xb3, 0x7c, 0x38, 0xbd, 0x25, - 0x69, 0x48, 0x71, 0xb9, 0xa4, 0xe9, 0x08, 0xc9, 0xd9, 0x84, 0xa4, 0xa4, 0x21, 0xc5, 0x65, 0x49, - 0x1c, 0x3a, 0xe6, 0xed, 0x90, 0x08, 0x61, 0xc8, 0x23, 0xa3, 0x92, 0xf8, 0x66, 0xcc, 0x42, 0x49, - 0x39, 0xf6, 0x2c, 0xb7, 0x43, 0x2b, 0x30, 0x3d, 0x8b, 0x54, 0xea, 0xfa, 0x39, 0x1e, 0x7f, 0xde, - 0x78, 0x6f, 0x29, 0x2a, 0x4a, 0x48, 0x40, 0x02, 0x46, 0x88, 0x13, 0xda, 0xfa, 0x79, 0x7e, 0xb1, - 0x0f, 0x24, 0x05, 0xd5, 0xc9, 0xd9, 0x70, 0x42, 0x1b, 0x71, 0xf3, 0xf0, 0x19, 0x30, 0x65, 0xe3, - 0x23, 0x56, 0x0e, 0x88, 0x1f, 0x98, 0x84, 0xea, 0x73, 0x7c, 0xf1, 0xb3, 0xac, 0xe3, 0xdc, 0x4a, - 0x32, 0x50, 0xa7, 0x1c, 0x57, 0x34, 0x9d, 0x84, 0xe2, 0x7c, 0x42, 0x31, 0xc9, 0x40, 0x9d, 0x72, - 0x2c, 0xd2, 0x3e, 0xb9, 0x1d, 0x9a, 0x3e, 0x31, 0xf4, 0x07, 0x78, 0x93, 0x2a, 0x3f, 0x18, 0x08, - 0x1a, 0x52, 0x5c, 0xd8, 0x8c, 0xe6, 0x47, 0x3a, 0x3f, 0x86, 0x37, 0x07, 0x5b, 0xc9, 0x2b, 0xfe, - 0xaa, 0xef, 0xe3, 0x63, 0x71, 0xd3, 0x24, 0x27, 0x47, 0x90, 0x82, 0x1c, 0xb6, 0xac, 0x4a, 0x5d, - 0xbf, 0xc0, 0x63, 0x3f, 0xe8, 0x1b, 0x44, 0x55, 0x9d, 0x55, 0x06, 0x82, 0x04, 0x16, 0x03, 0x75, - 0x1d, 0x96, 0x1a, 0x0b, 0xc3, 0x05, 0xad, 0x30, 0x10, 0x24, 0xb0, 0xf8, 0x4a, 0x9d, 0xe3, 0x4a, - 0x5d, 0x7f, 0x70, 0xc8, 0x2b, 0x65, 0x20, 0x48, 0x60, 0x41, 0x13, 0x64, 0x1d, 0x37, 0xd0, 0x2f, - 0x0e, 0xe5, 0x7a, 0xe6, 0x17, 0xce, 0xb6, 0x1b, 0x20, 0x86, 0x01, 0x7f, 0xa1, 0x01, 0xe0, 0xc5, - 0x29, 0xfa, 0xd0, 0x40, 0xc6, 0x12, 0x29, 0xc8, 0x52, 0x9c, 0xdb, 0x1b, 0x4e, 0xe0, 0x1f, 0xc7, - 0xef, 0xc8, 0xc4, 0x19, 0x48, 0x78, 0x01, 0x7f, 0xa3, 0x81, 0xf3, 0xc9, 0x36, 0x59, 0xb9, 0xb7, - 0xc8, 0x23, 0x72, 0x63, 0xd0, 0x69, 0x5e, 0x76, 0x5d, 0xab, 0xac, 0xb7, 0x5b, 0x85, 0xf3, 0xab, - 0x3d, 0x50, 0x51, 0x4f, 0x5f, 0xe0, 0x1f, 0x34, 0x30, 0x2b, 0xab, 0x68, 0xc2, 0xc3, 0x02, 0x0f, - 0x20, 0x19, 0x74, 0x00, 0xd3, 0x38, 0x22, 0x8e, 0xea, 0x43, 0x77, 0x17, 0x1f, 0x75, 0xbb, 0x06, - 0xff, 0xac, 0x81, 0x49, 0x83, 0x78, 0xc4, 0x31, 0x88, 0x53, 0x63, 0xbe, 0x2e, 0x0d, 0x64, 0x6c, - 0x90, 0xf6, 0x75, 0x3d, 0x01, 0x21, 0xdc, 0x2c, 0x49, 0x37, 0x27, 0x93, 0xac, 0x93, 0x56, 0x61, - 0x3e, 0x56, 0x4d, 0x72, 0x50, 0x87, 0x97, 0xf0, 0x3d, 0x0d, 0x4c, 0xc7, 0x1b, 0x20, 0xae, 0x94, - 0x4b, 0x43, 0xcc, 0x03, 0xde, 0xbe, 0xae, 0x76, 0x02, 0xa2, 0xb4, 0x07, 0xf0, 0x8f, 0x1a, 0xeb, - 0xd4, 0xa2, 0x77, 0x1f, 0xd5, 0x8b, 0x3c, 0x96, 0x6f, 0x0e, 0x3c, 0x96, 0x0a, 0x41, 0x84, 0xf2, - 0x4a, 0xdc, 0x0a, 0x2a, 0xce, 0x49, 0xab, 0x30, 0x97, 0x8c, 0xa4, 0x62, 0xa0, 0xa4, 0x87, 0xf0, - 0xc7, 0x1a, 0x98, 0x24, 0x71, 0xc7, 0x4d, 0xf5, 0x87, 0x07, 0x12, 0xc4, 0x9e, 0x4d, 0xbc, 0x78, - 0xa9, 0x27, 0x58, 0x14, 0x75, 0x60, 0xb3, 0x0e, 0x92, 0x1c, 0x61, 0xdb, 0xb3, 0x88, 0xfe, 0x7f, - 0x03, 0xee, 0x20, 0x37, 0x84, 0x5d, 0x14, 0x01, 0xc0, 0x2b, 0x20, 0xef, 0x84, 0x96, 0x85, 0xf7, - 0x2c, 0xa2, 0x3f, 0xc2, 0x7b, 0x11, 0x35, 0x16, 0xdd, 0x96, 0x74, 0xa4, 0x24, 0x60, 0x1d, 0x2c, - 0x1d, 0xbd, 0xa4, 0xfe, 0x45, 0xa8, 0xe7, 0xe0, 0x4e, 0xbf, 0xcc, 0xad, 0x2c, 0xb4, 0x5b, 0x85, - 0xf9, 0xdd, 0xde, 0xa3, 0xbd, 0xbb, 0xda, 0x80, 0xaf, 0x82, 0x07, 0x13, 0x32, 0x1b, 0xf6, 0x1e, - 0x31, 0x0c, 0x62, 0x44, 0x0f, 0x37, 0xfd, 0xff, 0xc5, 0xf0, 0x30, 0x3a, 0xe0, 0xbb, 0x69, 0x01, - 0x74, 0x27, 0x6d, 0x78, 0x1d, 0xcc, 0x27, 0xd8, 0x9b, 0x4e, 0x50, 0xf1, 0xab, 0x81, 0x6f, 0x3a, - 0x0d, 0x7d, 0x99, 0xdb, 0x3d, 0x1f, 0x9d, 0xc8, 0xdd, 0x04, 0x0f, 0xf5, 0xd1, 0x81, 0x5f, 0xeb, - 0xb0, 0xc6, 0x3f, 0x63, 0x61, 0xef, 0x25, 0x72, 0x4c, 0xf5, 0x47, 0x79, 0x77, 0xc2, 0x37, 0x7b, - 0x37, 0x41, 0x47, 0x7d, 0xe4, 0xe1, 0x57, 0xc1, 0xb9, 0x14, 0x87, 0x3d, 0x51, 0xf4, 0xc7, 0xc4, - 0x5b, 0x83, 0xf5, 0xb3, 0xbb, 0x11, 0x11, 0xf5, 0x92, 0x84, 0x5f, 0x06, 0x30, 0x41, 0xde, 0xc2, - 0x1e, 0xd7, 0x7f, 0x5c, 0x3c, 0x7b, 0xd8, 0x8e, 0xee, 0x4a, 0x1a, 0xea, 0x21, 0xb7, 0xc0, 0xde, - 0xc0, 0xa9, 0x1a, 0x0a, 0x67, 0x40, 0xf6, 0x80, 0xc8, 0xff, 0x1d, 0x40, 0xec, 0x4f, 0x68, 0x80, - 0x5c, 0x13, 0x5b, 0x61, 0xf4, 0x8c, 0x1f, 0xf0, 0xfd, 0x8b, 0x84, 0xf1, 0xe7, 0x32, 0xcf, 0x6a, - 0x0b, 0xef, 0x6b, 0x60, 0xbe, 0x77, 0x69, 0xbf, 0xaf, 0x6e, 0xfd, 0x4a, 0x03, 0xb3, 0x5d, 0x55, - 0xbc, 0x87, 0x47, 0xb7, 0x3b, 0x3d, 0x7a, 0x75, 0xd0, 0xe5, 0x58, 0xa4, 0x1f, 0xef, 0x41, 0x93, - 0xee, 0xfd, 0x4c, 0x03, 0x33, 0xe9, 0xc2, 0x78, 0x3f, 0xe3, 0x55, 0x7c, 0x3f, 0x03, 0xe6, 0x7b, - 0xb7, 0xce, 0xd0, 0x57, 0x33, 0x82, 0xe1, 0xcc, 0x5a, 0x7a, 0xcd, 0x65, 0xdf, 0xd1, 0xc0, 0xc4, - 0x2d, 0x25, 0x17, 0x7d, 0x5b, 0x1e, 0xf8, 0x94, 0x27, 0xba, 0x89, 0x62, 0x06, 0x45, 0x49, 0xdc, - 0xe2, 0x9f, 0x34, 0x30, 0xd7, 0xf3, 0x8a, 0x85, 0x97, 0xc1, 0x28, 0xb6, 0x2c, 0xf7, 0x50, 0x0c, - 0xeb, 0x12, 0x93, 0xf0, 0x55, 0x4e, 0x45, 0x92, 0x9b, 0x88, 0x5e, 0xe6, 0xf3, 0x8a, 0x5e, 0xf1, - 0x6f, 0x1a, 0xb8, 0x78, 0xa7, 0x4c, 0xbc, 0x2f, 0x5b, 0xba, 0x0c, 0xf2, 0xb2, 0x3d, 0x3e, 0xe6, - 0xdb, 0x29, 0x8b, 0x9d, 0x2c, 0x1a, 0xfc, 0xdf, 0xa9, 0xc4, 0x5f, 0xc5, 0x0f, 0x34, 0x30, 0x53, - 0x25, 0x7e, 0xd3, 0xac, 0x11, 0x44, 0xea, 0xc4, 0x27, 0x4e, 0x8d, 0xc0, 0x15, 0x30, 0xce, 0x3f, - 0xea, 0x7a, 0xb8, 0x16, 0x7d, 0xa0, 0x98, 0x95, 0x21, 0x1f, 0xdf, 0x8e, 0x18, 0x28, 0x96, 0x51, - 0x1f, 0x33, 0x32, 0x7d, 0x3f, 0x66, 0x5c, 0x04, 0x23, 0x5e, 0x3c, 0xea, 0xcd, 0x33, 0x2e, 0x9f, - 0xee, 0x72, 0x2a, 0xe7, 0xba, 0x7e, 0xc0, 0xe7, 0x57, 0x39, 0xc9, 0x75, 0xfd, 0x00, 0x71, 0x6a, - 0xf1, 0xef, 0x1a, 0xe8, 0xf5, 0x8f, 0x4f, 0xf0, 0x82, 0x18, 0xe1, 0x25, 0xe6, 0x62, 0xd1, 0xf8, - 0x0e, 0x36, 0xc1, 0x18, 0x15, 0xab, 0x92, 0x51, 0xaf, 0xdc, 0x63, 0xd4, 0xd3, 0x31, 0x12, 0xbd, - 0x43, 0x44, 0x8d, 0xc0, 0x58, 0xe0, 0x6b, 0xb8, 0x1c, 0x3a, 0x86, 0x9c, 0xea, 0x4e, 0x8a, 0xc0, - 0xaf, 0xad, 0x0a, 0x1a, 0x52, 0xdc, 0xf2, 0xd5, 0x0f, 0x3f, 0x59, 0x3c, 0xf3, 0xd1, 0x27, 0x8b, - 0x67, 0x3e, 0xfe, 0x64, 0xf1, 0xcc, 0x77, 0xdb, 0x8b, 0xda, 0x87, 0xed, 0x45, 0xed, 0xa3, 0xf6, - 0xa2, 0xf6, 0x71, 0x7b, 0x51, 0xfb, 0x67, 0x7b, 0x51, 0xfb, 0xf9, 0xa7, 0x8b, 0x67, 0xbe, 0x39, - 0x26, 0xf1, 0xff, 0x1b, 0x00, 0x00, 0xff, 0xff, 0x47, 0x22, 0xb1, 0x79, 0x8e, 0x2c, 0x00, 0x00, + // 3024 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x5a, 0xcd, 0x73, 0x23, 0x47, + 0xd9, 0xdf, 0x91, 0x2c, 0x5b, 0x6e, 0xdb, 0x6b, 0xbb, 0x77, 0xed, 0xcc, 0x3a, 0x1b, 0xcb, 0xab, + 0xbc, 0xd9, 0xd7, 0x49, 0x76, 0xe5, 0xc4, 0x24, 0x24, 0xa4, 0xa0, 0x28, 0xcb, 0xf6, 0x06, 0x27, + 0xeb, 0x0f, 0x5a, 0xbb, 0x89, 0x21, 0x9f, 0x6d, 0x4d, 0x4b, 0x9e, 0xf5, 0x7c, 0xed, 0xf4, 0x8c, + 0x6c, 0x57, 0x80, 0xe2, 0xa3, 0x52, 0x50, 0x14, 0x10, 0x8a, 0xe4, 0x42, 0x15, 0x1c, 0x02, 0xc5, + 0x85, 0x03, 0x1c, 0xa0, 0xb8, 0xc0, 0x1f, 0x90, 0x63, 0x8a, 0x53, 0x0e, 0x94, 0x60, 0xc5, 0x95, + 0x23, 0x55, 0x54, 0xf9, 0x44, 0xf5, 0xc7, 0xf4, 0x8c, 0x46, 0xd2, 0xae, 0x2b, 0x2b, 0x65, 0xb9, + 0x59, 0xcf, 0xd7, 0xef, 0xe9, 0xa7, 0x9f, 0x7e, 0xfa, 0xe9, 0x67, 0x0c, 0x6a, 0x07, 0xcf, 0xd3, + 0x92, 0xe9, 0x2e, 0x1d, 0x84, 0x7b, 0xc4, 0x77, 0x48, 0x40, 0xe8, 0x52, 0x83, 0x38, 0x86, 0xeb, + 0x2f, 0x49, 0x06, 0xf6, 0x4c, 0x72, 0x14, 0x10, 0x87, 0x9a, 0xae, 0x43, 0xaf, 0x62, 0xcf, 0xa4, + 0xc4, 0x6f, 0x10, 0x7f, 0xc9, 0x3b, 0xa8, 0x33, 0x1e, 0x6d, 0x17, 0x58, 0x6a, 0x3c, 0xbd, 0x47, + 0x02, 0xfc, 0xf4, 0x52, 0x9d, 0x38, 0xc4, 0xc7, 0x01, 0x31, 0x4a, 0x9e, 0xef, 0x06, 0x2e, 0xfc, + 0x92, 0x30, 0x57, 0x6a, 0x93, 0x7e, 0x4b, 0x99, 0x2b, 0x79, 0x07, 0x75, 0xc6, 0xa3, 0xed, 0x02, + 0x25, 0x69, 0x6e, 0xee, 0x6a, 0xdd, 0x0c, 0xf6, 0xc3, 0xbd, 0x52, 0xd5, 0xb5, 0x97, 0xea, 0x6e, + 0xdd, 0x5d, 0xe2, 0x56, 0xf7, 0xc2, 0x1a, 0xff, 0xc5, 0x7f, 0xf0, 0xbf, 0x04, 0xda, 0xdc, 0x33, + 0xb1, 0xf3, 0x36, 0xae, 0xee, 0x9b, 0x0e, 0xf1, 0x8f, 0x63, 0x8f, 0x6d, 0x12, 0xe0, 0xa5, 0x46, + 0x87, 0x8f, 0x73, 0x4b, 0xbd, 0xb4, 0xfc, 0xd0, 0x09, 0x4c, 0x9b, 0x74, 0x28, 0x7c, 0xfe, 0x5e, + 0x0a, 0xb4, 0xba, 0x4f, 0x6c, 0x9c, 0xd6, 0x2b, 0x9e, 0x68, 0x60, 0x7a, 0xd5, 0x75, 0x1a, 0xc4, + 0x67, 0xab, 0x44, 0xe4, 0x76, 0x48, 0x68, 0x00, 0xcb, 0x20, 0x1b, 0x9a, 0x86, 0xae, 0x2d, 0x68, + 0x8b, 0xa3, 0xe5, 0xa7, 0x3e, 0x6a, 0x16, 0xce, 0xb4, 0x9a, 0x85, 0xec, 0xcd, 0x8d, 0xb5, 0x93, + 0x66, 0xe1, 0x52, 0x2f, 0xa4, 0xe0, 0xd8, 0x23, 0xb4, 0x74, 0x73, 0x63, 0x0d, 0x31, 0x65, 0xf8, + 0x22, 0x98, 0x36, 0x08, 0x35, 0x7d, 0x62, 0xac, 0xec, 0x6c, 0xbc, 0x22, 0xec, 0xeb, 0x19, 0x6e, + 0xf1, 0x82, 0xb4, 0x38, 0xbd, 0x96, 0x16, 0x40, 0x9d, 0x3a, 0x70, 0x17, 0x8c, 0xb8, 0x7b, 0xb7, + 0x48, 0x35, 0xa0, 0x7a, 0x76, 0x21, 0xbb, 0x38, 0xb6, 0x7c, 0xb5, 0x14, 0xef, 0xa0, 0x72, 0x81, + 0x6f, 0x9b, 0x5c, 0x6c, 0x09, 0xe1, 0xc3, 0xf5, 0x68, 0xe7, 0xca, 0x93, 0x12, 0x6d, 0x64, 0x5b, + 0x58, 0x41, 0x91, 0xb9, 0xe2, 0xaf, 0x33, 0x00, 0x26, 0x17, 0x4f, 0x3d, 0xd7, 0xa1, 0xa4, 0x2f, + 0xab, 0xa7, 0x60, 0xaa, 0xca, 0x2d, 0x07, 0xc4, 0x90, 0xb8, 0x7a, 0xe6, 0xd3, 0x78, 0xaf, 0x4b, + 0xfc, 0xa9, 0xd5, 0x94, 0x39, 0xd4, 0x01, 0x00, 0x6f, 0x80, 0x61, 0x9f, 0xd0, 0xd0, 0x0a, 0xf4, + 0xec, 0x82, 0xb6, 0x38, 0xb6, 0x7c, 0xa5, 0x27, 0x14, 0xcf, 0x6f, 0x96, 0x7c, 0xa5, 0xc6, 0xd3, + 0xa5, 0x4a, 0x80, 0x83, 0x90, 0x96, 0xcf, 0x4a, 0xa4, 0x61, 0xc4, 0x6d, 0x20, 0x69, 0xab, 0xf8, + 0x83, 0x0c, 0x98, 0x4a, 0x46, 0xa9, 0x61, 0x92, 0x43, 0x78, 0x08, 0x46, 0x7c, 0x91, 0x2c, 0x3c, + 0x4e, 0x63, 0xcb, 0x3b, 0xa5, 0xfb, 0x3a, 0x56, 0xa5, 0x8e, 0x24, 0x2c, 0x8f, 0xb1, 0x3d, 0x93, + 0x3f, 0x50, 0x84, 0x06, 0xdf, 0x01, 0x79, 0x5f, 0x6e, 0x14, 0xcf, 0xa6, 0xb1, 0xe5, 0xaf, 0xf6, + 0x11, 0x59, 0x18, 0x2e, 0x8f, 0xb7, 0x9a, 0x85, 0x7c, 0xf4, 0x0b, 0x29, 0xc0, 0xe2, 0xfb, 0x19, + 0x30, 0xbf, 0x1a, 0xd2, 0xc0, 0xb5, 0x11, 0xa1, 0x6e, 0xe8, 0x57, 0xc9, 0xaa, 0x6b, 0x85, 0xb6, + 0xb3, 0x46, 0x6a, 0xa6, 0x63, 0x06, 0x2c, 0x5b, 0x17, 0xc0, 0x90, 0x83, 0x6d, 0x22, 0xb3, 0x67, + 0x5c, 0xc6, 0x74, 0x68, 0x0b, 0xdb, 0x04, 0x71, 0x0e, 0x93, 0x60, 0xc9, 0x22, 0xcf, 0x82, 0x92, + 0xb8, 0x71, 0xec, 0x11, 0xc4, 0x39, 0xf0, 0x32, 0x18, 0xae, 0xb9, 0xbe, 0x8d, 0xc5, 0x3e, 0x8e, + 0xc6, 0x3b, 0x73, 0x8d, 0x53, 0x91, 0xe4, 0xc2, 0x67, 0xc1, 0x98, 0x41, 0x68, 0xd5, 0x37, 0x3d, + 0x06, 0xad, 0x0f, 0x71, 0xe1, 0x73, 0x52, 0x78, 0x6c, 0x2d, 0x66, 0xa1, 0xa4, 0x1c, 0xbc, 0x02, + 0xf2, 0x9e, 0x6f, 0xba, 0xbe, 0x19, 0x1c, 0xeb, 0xb9, 0x05, 0x6d, 0x31, 0x57, 0x9e, 0x92, 0x3a, + 0xf9, 0x1d, 0x49, 0x47, 0x4a, 0x02, 0x2e, 0x80, 0xfc, 0x4b, 0x95, 0xed, 0xad, 0x1d, 0x1c, 0xec, + 0xeb, 0xc3, 0x1c, 0x61, 0x88, 0x49, 0xa3, 0xfc, 0x2d, 0x49, 0x2d, 0xfe, 0x2d, 0x03, 0xf4, 0x74, + 0x54, 0xa2, 0x90, 0xc2, 0x6b, 0x20, 0x4f, 0x03, 0x56, 0x71, 0xea, 0xc7, 0x32, 0x26, 0x4f, 0x44, + 0x60, 0x15, 0x49, 0x3f, 0x69, 0x16, 0x66, 0x63, 0x8d, 0x88, 0xca, 0xe3, 0xa1, 0x74, 0xe1, 0x2f, + 0x35, 0x70, 0xee, 0x90, 0xec, 0xed, 0xbb, 0xee, 0xc1, 0xaa, 0x65, 0x12, 0x27, 0x58, 0x75, 0x9d, + 0x9a, 0x59, 0x97, 0x39, 0x80, 0xee, 0x33, 0x07, 0x5e, 0xed, 0xb4, 0x5c, 0x7e, 0xa8, 0xd5, 0x2c, + 0x9c, 0xeb, 0xc2, 0x40, 0xdd, 0xfc, 0x80, 0xbb, 0x40, 0xaf, 0xa6, 0x0e, 0x89, 0x2c, 0x60, 0xa2, + 0x6c, 0x8d, 0x96, 0x2f, 0xb6, 0x9a, 0x05, 0x7d, 0xb5, 0x87, 0x0c, 0xea, 0xa9, 0x5d, 0xfc, 0x5e, + 0x36, 0x1d, 0xde, 0x44, 0xba, 0xbd, 0x0d, 0xf2, 0xec, 0x18, 0x1b, 0x38, 0xc0, 0xf2, 0x20, 0x3e, + 0x75, 0xba, 0x43, 0x2f, 0x6a, 0xc6, 0x26, 0x09, 0x70, 0x19, 0xca, 0x0d, 0x01, 0x31, 0x0d, 0x29, + 0xab, 0xf0, 0x9b, 0x60, 0x88, 0x7a, 0xa4, 0x2a, 0x03, 0xfd, 0xda, 0xfd, 0x1e, 0xb6, 0x1e, 0x0b, + 0xa9, 0x78, 0xa4, 0x1a, 0x9f, 0x05, 0xf6, 0x0b, 0x71, 0x58, 0xf8, 0xae, 0x06, 0x86, 0x29, 0x2f, + 0x50, 0xb2, 0xa8, 0xbd, 0x31, 0x28, 0x0f, 0x52, 0x55, 0x50, 0xfc, 0x46, 0x12, 0xbc, 0xf8, 0xef, + 0x0c, 0xb8, 0xd4, 0x4b, 0x75, 0xd5, 0x75, 0x0c, 0xb1, 0x1d, 0x1b, 0xf2, 0x6c, 0x8b, 0x4c, 0x7f, + 0x36, 0x79, 0xb6, 0x4f, 0x9a, 0x85, 0xc7, 0xee, 0x69, 0x20, 0x51, 0x04, 0xbe, 0xa0, 0xd6, 0x2d, + 0x0a, 0xc5, 0xa5, 0x76, 0xc7, 0x4e, 0x9a, 0x85, 0x49, 0xa5, 0xd6, 0xee, 0x2b, 0x6c, 0x00, 0x68, + 0x61, 0x1a, 0xdc, 0xf0, 0xb1, 0x43, 0x85, 0x59, 0xd3, 0x26, 0x32, 0x7c, 0x4f, 0x9c, 0x2e, 0x3d, + 0x98, 0x46, 0x79, 0x4e, 0x42, 0xc2, 0xeb, 0x1d, 0xd6, 0x50, 0x17, 0x04, 0x56, 0xb7, 0x7c, 0x82, + 0xa9, 0x2a, 0x45, 0x89, 0x1b, 0x85, 0x51, 0x91, 0xe4, 0xc2, 0xc7, 0xc1, 0x88, 0x4d, 0x28, 0xc5, + 0x75, 0xc2, 0xeb, 0xcf, 0x68, 0x7c, 0x45, 0x6f, 0x0a, 0x32, 0x8a, 0xf8, 0xac, 0x3f, 0xb9, 0xd8, + 0x2b, 0x6a, 0xd7, 0x4d, 0x1a, 0xc0, 0xd7, 0x3b, 0x0e, 0x40, 0xe9, 0x74, 0x2b, 0x64, 0xda, 0x3c, + 0xfd, 0x55, 0xf1, 0x8b, 0x28, 0x89, 0xe4, 0xff, 0x06, 0xc8, 0x99, 0x01, 0xb1, 0xa3, 0xbb, 0xfb, + 0xd5, 0x01, 0xe5, 0x5e, 0x79, 0x42, 0xfa, 0x90, 0xdb, 0x60, 0x68, 0x48, 0x80, 0x16, 0x7f, 0x93, + 0x01, 0x8f, 0xf4, 0x52, 0x61, 0x17, 0x0a, 0x65, 0x11, 0xf7, 0xac, 0xd0, 0xc7, 0x96, 0xcc, 0x38, + 0x15, 0xf1, 0x1d, 0x4e, 0x45, 0x92, 0xcb, 0x4a, 0x3e, 0x35, 0x9d, 0x7a, 0x68, 0x61, 0x5f, 0xa6, + 0x93, 0x5a, 0x75, 0x45, 0xd2, 0x91, 0x92, 0x80, 0x25, 0x00, 0xe8, 0xbe, 0xeb, 0x07, 0x1c, 0x43, + 0x56, 0xaf, 0xb3, 0xac, 0x40, 0x54, 0x14, 0x15, 0x25, 0x24, 0xd8, 0x8d, 0x76, 0x60, 0x3a, 0x86, + 0xdc, 0x75, 0x75, 0x8a, 0x5f, 0x36, 0x1d, 0x03, 0x71, 0x0e, 0xc3, 0xb7, 0x4c, 0x1a, 0x30, 0x8a, + 0xdc, 0xf2, 0xb6, 0xa8, 0x73, 0x49, 0x25, 0xc1, 0xf0, 0xab, 0xac, 0xea, 0xbb, 0xbe, 0x49, 0xa8, + 0x3e, 0x1c, 0xe3, 0xaf, 0x2a, 0x2a, 0x4a, 0x48, 0x14, 0xff, 0x95, 0xef, 0x9d, 0x24, 0xac, 0x94, + 0xc0, 0x47, 0x41, 0xae, 0xee, 0xbb, 0xa1, 0x27, 0xa3, 0xa4, 0xa2, 0xfd, 0x22, 0x23, 0x22, 0xc1, + 0x63, 0x59, 0xd9, 0x68, 0x6b, 0x53, 0x55, 0x56, 0x46, 0xcd, 0x69, 0xc4, 0x87, 0xdf, 0xd1, 0x40, + 0xce, 0x91, 0xc1, 0x61, 0x29, 0xf7, 0xfa, 0x80, 0xf2, 0x82, 0x87, 0x37, 0x76, 0x57, 0x44, 0x5e, + 0x20, 0xc3, 0x67, 0x40, 0x8e, 0x56, 0x5d, 0x8f, 0xc8, 0xa8, 0xcf, 0x47, 0x42, 0x15, 0x46, 0x3c, + 0x69, 0x16, 0x26, 0x22, 0x73, 0x9c, 0x80, 0x84, 0x30, 0xfc, 0xbe, 0x06, 0x40, 0x03, 0x5b, 0xa6, + 0x81, 0x79, 0xcb, 0x90, 0xe3, 0xee, 0xf7, 0x37, 0xad, 0x5f, 0x51, 0xe6, 0xc5, 0xa6, 0xc5, 0xbf, + 0x51, 0x02, 0x1a, 0xbe, 0xa7, 0x81, 0x71, 0x1a, 0xee, 0xf9, 0x52, 0x8b, 0xf2, 0xe6, 0x62, 0x6c, + 0xf9, 0x6b, 0x7d, 0xf5, 0xa5, 0x92, 0x00, 0x28, 0x4f, 0xb5, 0x9a, 0x85, 0xf1, 0x24, 0x05, 0xb5, + 0x39, 0x00, 0x7f, 0xa4, 0x81, 0x7c, 0x23, 0xba, 0xb3, 0x47, 0xf8, 0x81, 0x7f, 0x73, 0x40, 0x1b, + 0x2b, 0x33, 0x2a, 0x3e, 0x05, 0xaa, 0x0f, 0x50, 0x1e, 0xc0, 0x3f, 0x6b, 0x40, 0xc7, 0x86, 0x28, + 0xf0, 0xd8, 0xda, 0xf1, 0x4d, 0x27, 0x20, 0xbe, 0xe8, 0x37, 0xa9, 0x9e, 0xe7, 0xee, 0xf5, 0xf7, + 0x2e, 0x4c, 0xf7, 0xb2, 0xe5, 0x05, 0xe9, 0x9d, 0xbe, 0xd2, 0xc3, 0x0d, 0xd4, 0xd3, 0x41, 0x9e, + 0x68, 0x71, 0x4b, 0xa3, 0x8f, 0x0e, 0x20, 0xd1, 0xe2, 0x5e, 0x4a, 0x56, 0x87, 0xb8, 0x83, 0x4a, + 0x40, 0xc3, 0x6d, 0x30, 0xe3, 0xf9, 0x84, 0x03, 0xdc, 0x74, 0x0e, 0x1c, 0xf7, 0xd0, 0xb9, 0x66, + 0x12, 0xcb, 0xa0, 0x3a, 0x58, 0xd0, 0x16, 0xf3, 0xe5, 0x0b, 0xad, 0x66, 0x61, 0x66, 0xa7, 0x9b, + 0x00, 0xea, 0xae, 0x57, 0x7c, 0x2f, 0x9b, 0x7e, 0x05, 0xa4, 0xbb, 0x08, 0xf8, 0x81, 0x58, 0xbd, + 0x88, 0x0d, 0xd5, 0x35, 0xbe, 0x5b, 0x6f, 0x0f, 0x28, 0x99, 0x54, 0x1b, 0x10, 0x77, 0x72, 0x8a, + 0x44, 0x51, 0xc2, 0x0f, 0xf8, 0x73, 0x0d, 0x4c, 0xe0, 0x6a, 0x95, 0x78, 0x01, 0x31, 0x44, 0x71, + 0xcf, 0x7c, 0x06, 0xf5, 0x6b, 0x46, 0x7a, 0x35, 0xb1, 0x92, 0x84, 0x46, 0xed, 0x9e, 0xc0, 0x17, + 0xc0, 0x59, 0x1a, 0xb8, 0x3e, 0x31, 0x52, 0x6d, 0x33, 0x6c, 0x35, 0x0b, 0x67, 0x2b, 0x6d, 0x1c, + 0x94, 0x92, 0x2c, 0xfe, 0x3d, 0x07, 0x0a, 0xf7, 0x38, 0x6a, 0xa7, 0x78, 0x98, 0x5d, 0x06, 0xc3, + 0x7c, 0xb9, 0x06, 0x8f, 0x4a, 0x3e, 0xd1, 0x0a, 0x72, 0x2a, 0x92, 0x5c, 0x76, 0x51, 0x30, 0x7c, + 0xd6, 0xbe, 0x64, 0xb9, 0xa0, 0xba, 0x28, 0x2a, 0x82, 0x8c, 0x22, 0x3e, 0x5c, 0x06, 0xc0, 0x20, + 0x9e, 0x4f, 0xd8, 0x65, 0x65, 0xe8, 0x23, 0x5c, 0x5a, 0x6d, 0xd2, 0x9a, 0xe2, 0xa0, 0x84, 0x14, + 0xbc, 0x06, 0x60, 0xf4, 0xcb, 0x74, 0x9d, 0x57, 0xb1, 0xef, 0x98, 0x4e, 0x5d, 0xcf, 0x73, 0xb7, + 0x67, 0x59, 0x37, 0xb6, 0xd6, 0xc1, 0x45, 0x5d, 0x34, 0xe0, 0x3b, 0x60, 0x58, 0x0c, 0x7d, 0xf8, + 0x0d, 0x31, 0xc0, 0x2a, 0x0f, 0x78, 0x8c, 0x38, 0x14, 0x92, 0x90, 0x9d, 0xd5, 0x3d, 0xf7, 0xa0, + 0xab, 0xfb, 0x5d, 0xcb, 0xe9, 0xf0, 0xff, 0x78, 0x39, 0x2d, 0xfe, 0x47, 0x4b, 0xd7, 0x9c, 0xc4, + 0x52, 0x2b, 0x55, 0x6c, 0x11, 0xb8, 0x06, 0xa6, 0xd8, 0x8b, 0x09, 0x11, 0xcf, 0x32, 0xab, 0x98, + 0xf2, 0x07, 0xbb, 0x48, 0x76, 0x35, 0x43, 0xaa, 0xa4, 0xf8, 0xa8, 0x43, 0x03, 0xbe, 0x04, 0xa0, + 0x78, 0x45, 0xb4, 0xd9, 0x11, 0x0d, 0x91, 0x7a, 0x0f, 0x54, 0x3a, 0x24, 0x50, 0x17, 0x2d, 0xb8, + 0x0a, 0xa6, 0x2d, 0xbc, 0x47, 0xac, 0x0a, 0xb1, 0x48, 0x35, 0x70, 0x7d, 0x6e, 0x4a, 0x8c, 0x34, + 0x66, 0x5a, 0xcd, 0xc2, 0xf4, 0xf5, 0x34, 0x13, 0x75, 0xca, 0x17, 0x2f, 0xa5, 0x8f, 0x76, 0x72, + 0xe1, 0xe2, 0x6d, 0xf6, 0x61, 0x06, 0xcc, 0xf5, 0xce, 0x0c, 0xf8, 0xdd, 0xf8, 0x09, 0x29, 0x5e, + 0x08, 0x6f, 0x0e, 0x2a, 0x0b, 0xe5, 0x1b, 0x12, 0x74, 0xbe, 0x1f, 0xe1, 0xb7, 0x58, 0xbb, 0x86, + 0xad, 0x68, 0x68, 0xf5, 0xc6, 0xc0, 0x5c, 0x60, 0x20, 0xe5, 0x51, 0xd1, 0x09, 0x62, 0x8b, 0x37, + 0x7e, 0xd8, 0x22, 0xc5, 0xdf, 0x6a, 0xe9, 0x29, 0x42, 0x7c, 0x82, 0xe1, 0x8f, 0x35, 0x30, 0xe9, + 0x7a, 0xc4, 0x59, 0xd9, 0xd9, 0x78, 0xe5, 0x73, 0xe2, 0x24, 0xcb, 0x50, 0x6d, 0xdd, 0xa7, 0x9f, + 0x2f, 0x55, 0xb6, 0xb7, 0x84, 0xc1, 0x1d, 0xdf, 0xf5, 0x68, 0xf9, 0x5c, 0xab, 0x59, 0x98, 0xdc, + 0x6e, 0x87, 0x42, 0x69, 0xec, 0xa2, 0x0d, 0x66, 0xd6, 0x8f, 0x02, 0xe2, 0x3b, 0xd8, 0x5a, 0x73, + 0xab, 0xa1, 0x4d, 0x9c, 0x40, 0x38, 0x9a, 0x9a, 0x78, 0x69, 0xa7, 0x9c, 0x78, 0x3d, 0x02, 0xb2, + 0xa1, 0x6f, 0xc9, 0x2c, 0x1e, 0x53, 0x13, 0x5d, 0x74, 0x1d, 0x31, 0x7a, 0xf1, 0x12, 0x18, 0x62, + 0x7e, 0xc2, 0x0b, 0x20, 0xeb, 0xe3, 0x43, 0x6e, 0x75, 0xbc, 0x3c, 0xc2, 0x44, 0x10, 0x3e, 0x44, + 0x8c, 0x56, 0xfc, 0xd3, 0x02, 0x98, 0x4c, 0xad, 0x05, 0xce, 0x81, 0x8c, 0x1a, 0x13, 0x03, 0x69, + 0x34, 0xb3, 0xb1, 0x86, 0x32, 0xa6, 0x01, 0x9f, 0x53, 0xc5, 0x57, 0x80, 0x16, 0xd4, 0x5d, 0xc2, + 0xa9, 0xac, 0x3f, 0x8f, 0xcd, 0x31, 0x47, 0xa2, 0xc2, 0xc9, 0x7c, 0x20, 0x35, 0x79, 0x4a, 0x84, + 0x0f, 0xa4, 0x86, 0x18, 0xed, 0xd3, 0x8e, 0xfb, 0xa2, 0x79, 0x63, 0xee, 0x14, 0xf3, 0xc6, 0xe1, + 0xbb, 0xce, 0x1b, 0x1f, 0x05, 0xb9, 0xc0, 0x0c, 0x2c, 0xc2, 0x2f, 0xb2, 0xc4, 0x33, 0xea, 0x06, + 0x23, 0x22, 0xc1, 0x83, 0xb7, 0xc0, 0x88, 0x41, 0x6a, 0x38, 0xb4, 0x02, 0x7e, 0x67, 0x8d, 0x2d, + 0xaf, 0xf6, 0x21, 0x85, 0xc4, 0x30, 0x78, 0x4d, 0xd8, 0x45, 0x11, 0x00, 0x7c, 0x0c, 0x8c, 0xd8, + 0xf8, 0xc8, 0xb4, 0x43, 0x9b, 0x37, 0x98, 0x9a, 0x10, 0xdb, 0x14, 0x24, 0x14, 0xf1, 0x58, 0x65, + 0x24, 0x47, 0x55, 0x2b, 0xa4, 0x66, 0x83, 0x48, 0xa6, 0x6c, 0xfe, 0x54, 0x65, 0x5c, 0x4f, 0xf1, + 0x51, 0x87, 0x06, 0x07, 0x33, 0x1d, 0xae, 0x3c, 0x96, 0x00, 0x13, 0x24, 0x14, 0xf1, 0xda, 0xc1, + 0xa4, 0xfc, 0x78, 0x2f, 0x30, 0xa9, 0xdc, 0xa1, 0x01, 0x9f, 0x04, 0xa3, 0x36, 0x3e, 0xba, 0x4e, + 0x9c, 0x7a, 0xb0, 0xaf, 0x4f, 0x2c, 0x68, 0x8b, 0xd9, 0xf2, 0x44, 0xab, 0x59, 0x18, 0xdd, 0x8c, + 0x88, 0x28, 0xe6, 0x73, 0x61, 0xd3, 0x91, 0xc2, 0x67, 0x13, 0xc2, 0x11, 0x11, 0xc5, 0x7c, 0xd6, + 0xbd, 0x78, 0x38, 0x60, 0x87, 0x4b, 0x9f, 0x6c, 0x7f, 0xe6, 0xee, 0x08, 0x32, 0x8a, 0xf8, 0x70, + 0x11, 0xe4, 0x6d, 0x7c, 0xc4, 0x47, 0x12, 0xfa, 0x14, 0x37, 0xcb, 0x07, 0xe3, 0x9b, 0x92, 0x86, + 0x14, 0x97, 0x4b, 0x9a, 0x8e, 0x90, 0x9c, 0x4e, 0x48, 0x4a, 0x1a, 0x52, 0x5c, 0x96, 0xc4, 0xa1, + 0x63, 0xde, 0x0e, 0x89, 0x10, 0x86, 0x3c, 0x32, 0x2a, 0x89, 0x6f, 0xc6, 0x2c, 0x94, 0x94, 0x83, + 0x25, 0x00, 0xec, 0xd0, 0x0a, 0x4c, 0xcf, 0x22, 0xdb, 0x35, 0xfd, 0x1c, 0x8f, 0x3f, 0x6f, 0xfa, + 0x37, 0x15, 0x15, 0x25, 0x24, 0x20, 0x01, 0x43, 0xc4, 0x09, 0x6d, 0xfd, 0x3c, 0xbf, 0xd8, 0xfb, + 0x92, 0x82, 0xea, 0xe4, 0xac, 0x3b, 0xa1, 0x8d, 0xb8, 0x79, 0xf8, 0x1c, 0x98, 0xb0, 0xf1, 0x11, + 0x2b, 0x07, 0xc4, 0x0f, 0x4c, 0x42, 0xf5, 0x19, 0xbe, 0xf8, 0x69, 0xd6, 0xed, 0x6e, 0x26, 0x19, + 0xa8, 0x5d, 0x8e, 0x2b, 0x9a, 0x4e, 0x42, 0x71, 0x36, 0xa1, 0x98, 0x64, 0xa0, 0x76, 0x39, 0x16, + 0x69, 0x9f, 0xdc, 0x0e, 0x4d, 0x9f, 0x18, 0xfa, 0x43, 0xbc, 0x41, 0x96, 0x1f, 0x2b, 0x04, 0x0d, + 0x29, 0x2e, 0x6c, 0x44, 0xb3, 0x2b, 0x9d, 0x1f, 0xc3, 0x9b, 0xfd, 0xad, 0xe4, 0xdb, 0xfe, 0x8a, + 0xef, 0xe3, 0x63, 0x71, 0xd3, 0x24, 0xa7, 0x56, 0x90, 0x82, 0x1c, 0xb6, 0xac, 0xed, 0x9a, 0x7e, + 0x81, 0xc7, 0xbe, 0xdf, 0x37, 0x88, 0xaa, 0x3a, 0x2b, 0x0c, 0x04, 0x09, 0x2c, 0x06, 0xea, 0x3a, + 0x2c, 0x35, 0xe6, 0x06, 0x0b, 0xba, 0xcd, 0x40, 0x90, 0xc0, 0xe2, 0x2b, 0x75, 0x8e, 0xb7, 0x6b, + 0xfa, 0xc3, 0x03, 0x5e, 0x29, 0x03, 0x41, 0x02, 0x0b, 0x9a, 0x20, 0xeb, 0xb8, 0x81, 0x7e, 0x71, + 0x20, 0xd7, 0x33, 0xbf, 0x70, 0xb6, 0xdc, 0x00, 0x31, 0x0c, 0xf8, 0x33, 0x0d, 0x00, 0x2f, 0x4e, + 0xd1, 0x47, 0xfa, 0x32, 0x12, 0x49, 0x41, 0x96, 0xe2, 0xdc, 0x5e, 0x77, 0x02, 0xff, 0x38, 0x7e, + 0x1e, 0x25, 0xce, 0x40, 0xc2, 0x0b, 0xf8, 0x2b, 0x0d, 0x9c, 0x4f, 0xb6, 0xc9, 0xca, 0xbd, 0x79, + 0x1e, 0x91, 0x1b, 0xfd, 0x4e, 0xf3, 0xb2, 0xeb, 0x5a, 0x65, 0xbd, 0xd5, 0x2c, 0x9c, 0x5f, 0xe9, + 0x82, 0x8a, 0xba, 0xfa, 0x02, 0x7f, 0xa7, 0x81, 0x69, 0x59, 0x45, 0x13, 0x1e, 0x16, 0x78, 0x00, + 0x49, 0xbf, 0x03, 0x98, 0xc6, 0x11, 0x71, 0x54, 0x1f, 0xd9, 0x3b, 0xf8, 0xa8, 0xd3, 0x35, 0xf8, + 0x47, 0x0d, 0x8c, 0x1b, 0xc4, 0x23, 0x8e, 0x41, 0x9c, 0x2a, 0xf3, 0x75, 0xa1, 0x2f, 0x23, 0x8b, + 0xb4, 0xaf, 0x6b, 0x09, 0x08, 0xe1, 0x66, 0x49, 0xba, 0x39, 0x9e, 0x64, 0x9d, 0x34, 0x0b, 0xb3, + 0xb1, 0x6a, 0x92, 0x83, 0xda, 0xbc, 0x84, 0xef, 0x6b, 0x60, 0x32, 0xde, 0x00, 0x71, 0xa5, 0x5c, + 0x1a, 0x60, 0x1e, 0xf0, 0xf6, 0x75, 0xa5, 0x1d, 0x10, 0xa5, 0x3d, 0x80, 0xbf, 0xd7, 0x58, 0xa7, + 0x16, 0xbd, 0xfb, 0xa8, 0x5e, 0xe4, 0xb1, 0x7c, 0xab, 0xef, 0xb1, 0x54, 0x08, 0x22, 0x94, 0x57, + 0xe2, 0x56, 0x50, 0x71, 0x4e, 0x9a, 0x85, 0x99, 0x64, 0x24, 0x15, 0x03, 0x25, 0x3d, 0x84, 0x3f, + 0xd4, 0xc0, 0x38, 0x89, 0x3b, 0x6e, 0xaa, 0x3f, 0xda, 0x97, 0x20, 0x76, 0x6d, 0xe2, 0xc5, 0x4b, + 0x3d, 0xc1, 0xa2, 0xa8, 0x0d, 0x9b, 0x75, 0x90, 0xe4, 0x08, 0xdb, 0x9e, 0x45, 0xf4, 0xff, 0xeb, + 0x73, 0x07, 0xb9, 0x2e, 0xec, 0xa2, 0x08, 0x00, 0x5e, 0x01, 0x79, 0x27, 0xb4, 0x2c, 0xbc, 0x67, + 0x11, 0xfd, 0x31, 0xde, 0x8b, 0xa8, 0x91, 0xec, 0x96, 0xa4, 0x23, 0x25, 0x01, 0x6b, 0x60, 0xe1, + 0xe8, 0x65, 0xf5, 0xef, 0x49, 0x5d, 0x87, 0x86, 0xfa, 0x65, 0x6e, 0x65, 0xae, 0xd5, 0x2c, 0xcc, + 0xee, 0x76, 0x1f, 0x2b, 0xde, 0xd3, 0x06, 0x7c, 0x0d, 0x3c, 0x9c, 0x90, 0x59, 0xb7, 0xf7, 0x88, + 0x61, 0x10, 0x23, 0x7a, 0xb8, 0xe9, 0xff, 0x2f, 0x06, 0x97, 0xd1, 0x01, 0xdf, 0x4d, 0x0b, 0xa0, + 0xbb, 0x69, 0xc3, 0xeb, 0x60, 0x36, 0xc1, 0xde, 0x70, 0x82, 0x6d, 0xbf, 0x12, 0xf8, 0xa6, 0x53, + 0xd7, 0x17, 0xb9, 0xdd, 0xf3, 0xd1, 0x89, 0xdc, 0x4d, 0xf0, 0x50, 0x0f, 0x1d, 0xf8, 0x95, 0x36, + 0x6b, 0xfc, 0x13, 0x1a, 0xf6, 0x5e, 0x26, 0xc7, 0x54, 0x7f, 0x9c, 0x77, 0x27, 0x7c, 0xb3, 0x77, + 0x13, 0x74, 0xd4, 0x43, 0x1e, 0x7e, 0x19, 0x9c, 0x4b, 0x71, 0xd8, 0x13, 0x45, 0x7f, 0x42, 0xbc, + 0x35, 0x58, 0x3f, 0xbb, 0x1b, 0x11, 0x51, 0x37, 0x49, 0xf8, 0x45, 0x00, 0x13, 0xe4, 0x4d, 0xec, + 0x71, 0xfd, 0x27, 0xc5, 0xb3, 0x87, 0xed, 0xe8, 0xae, 0xa4, 0xa1, 0x2e, 0x72, 0x73, 0xec, 0x0d, + 0x9c, 0xaa, 0xa1, 0x70, 0x0a, 0x64, 0x0f, 0x88, 0xfc, 0xbf, 0x05, 0xc4, 0xfe, 0x84, 0x06, 0xc8, + 0x35, 0xb0, 0x15, 0x46, 0xcf, 0xf8, 0x3e, 0xdf, 0xbf, 0x48, 0x18, 0x7f, 0x21, 0xf3, 0xbc, 0x36, + 0xf7, 0x81, 0x06, 0x66, 0xbb, 0x97, 0xf6, 0x07, 0xea, 0xd6, 0x2f, 0x34, 0x30, 0xdd, 0x51, 0xc5, + 0xbb, 0x78, 0x74, 0xbb, 0xdd, 0xa3, 0xd7, 0xfa, 0x5d, 0x8e, 0x45, 0xfa, 0xf1, 0x1e, 0x34, 0xe9, + 0xde, 0x4f, 0x34, 0x30, 0x95, 0x2e, 0x8c, 0x0f, 0x32, 0x5e, 0xc5, 0x0f, 0x32, 0x60, 0xb6, 0x7b, + 0xeb, 0x0c, 0x7d, 0x35, 0x23, 0x18, 0xcc, 0xac, 0xa5, 0xdb, 0x5c, 0xf6, 0x5d, 0x0d, 0x8c, 0xdd, + 0x52, 0x72, 0xd1, 0x77, 0xed, 0xbe, 0x4f, 0x79, 0xa2, 0x9b, 0x28, 0x66, 0x50, 0x94, 0xc4, 0x2d, + 0xfe, 0x41, 0x03, 0x33, 0x5d, 0xaf, 0x58, 0x78, 0x19, 0x0c, 0x63, 0xcb, 0x72, 0x0f, 0xc5, 0xb0, + 0x2e, 0x31, 0x85, 0x5f, 0xe1, 0x54, 0x24, 0xb9, 0x89, 0xe8, 0x65, 0x3e, 0xab, 0xe8, 0x15, 0xff, + 0xa2, 0x81, 0x8b, 0x77, 0xcb, 0xc4, 0x07, 0xb2, 0xa5, 0x8b, 0x20, 0x2f, 0xdb, 0xe3, 0x63, 0xbe, + 0x9d, 0xb2, 0xd8, 0xc9, 0xa2, 0xc1, 0xff, 0x95, 0x4b, 0xfc, 0x55, 0xfc, 0x50, 0x03, 0x53, 0x15, + 0xe2, 0x37, 0xcc, 0x2a, 0x41, 0xa4, 0x46, 0x7c, 0xe2, 0x54, 0x09, 0x5c, 0x02, 0xa3, 0xfc, 0x83, + 0xb2, 0x87, 0xab, 0xd1, 0xc7, 0x91, 0x69, 0x19, 0xf2, 0xd1, 0xad, 0x88, 0x81, 0x62, 0x19, 0xf5, + 0x21, 0x25, 0xd3, 0xf3, 0x43, 0xca, 0x45, 0x30, 0xe4, 0xc5, 0xa3, 0xde, 0x3c, 0xe3, 0xf2, 0xe9, + 0x2e, 0xa7, 0x72, 0xae, 0xeb, 0x07, 0x7c, 0x7e, 0x95, 0x93, 0x5c, 0xd7, 0x0f, 0x10, 0xa7, 0x16, + 0xff, 0xaa, 0x81, 0x6e, 0xff, 0x74, 0x05, 0x2f, 0x88, 0x11, 0x5e, 0x62, 0x2e, 0x16, 0x8d, 0xef, + 0x60, 0x03, 0x8c, 0x50, 0xb1, 0x2a, 0x19, 0xf5, 0xed, 0xfb, 0x8c, 0x7a, 0x3a, 0x46, 0xa2, 0x77, + 0x88, 0xa8, 0x11, 0x18, 0x0b, 0x7c, 0x15, 0x97, 0x43, 0xc7, 0x90, 0x53, 0xdd, 0x71, 0x11, 0xf8, + 0xd5, 0x15, 0x41, 0x43, 0x8a, 0x5b, 0xbe, 0xfa, 0xd1, 0x9d, 0xf9, 0x33, 0x1f, 0xdf, 0x99, 0x3f, + 0xf3, 0xc9, 0x9d, 0xf9, 0x33, 0xdf, 0x6e, 0xcd, 0x6b, 0x1f, 0xb5, 0xe6, 0xb5, 0x8f, 0x5b, 0xf3, + 0xda, 0x27, 0xad, 0x79, 0xed, 0x1f, 0xad, 0x79, 0xed, 0xa7, 0xff, 0x9c, 0x3f, 0xf3, 0xf5, 0x11, + 0x89, 0xff, 0xdf, 0x00, 0x00, 0x00, 0xff, 0xff, 0xc6, 0xe5, 0x77, 0xcb, 0x0a, 0x2d, 0x00, 0x00, } func (m *ConversionRequest) Marshal() (dAtA []byte, err error) { @@ -1603,6 +1606,21 @@ func (m *CustomResourceDefinitionVersion) MarshalToSizedBuffer(dAtA []byte) (int _ = i var l int _ = l + if m.DeprecationWarning != nil { + i -= len(*m.DeprecationWarning) + copy(dAtA[i:], *m.DeprecationWarning) + i = encodeVarintGenerated(dAtA, i, uint64(len(*m.DeprecationWarning))) + i-- + dAtA[i] = 0x42 + } + i-- + if m.Deprecated { + dAtA[i] = 1 + } else { + dAtA[i] = 0 + } + i-- + dAtA[i] = 0x38 if len(m.AdditionalPrinterColumns) > 0 { for iNdEx := len(m.AdditionalPrinterColumns) - 1; iNdEx >= 0; iNdEx-- { { @@ -2896,6 +2914,11 @@ func (m *CustomResourceDefinitionVersion) Size() (n int) { n += 1 + l + sovGenerated(uint64(l)) } } + n += 2 + if m.DeprecationWarning != nil { + l = len(*m.DeprecationWarning) + n += 1 + l + sovGenerated(uint64(l)) + } return n } @@ -3442,6 +3465,8 @@ func (this *CustomResourceDefinitionVersion) String() string { `Schema:` + strings.Replace(this.Schema.String(), "CustomResourceValidation", "CustomResourceValidation", 1) + `,`, `Subresources:` + strings.Replace(this.Subresources.String(), "CustomResourceSubresources", "CustomResourceSubresources", 1) + `,`, `AdditionalPrinterColumns:` + repeatedStringForAdditionalPrinterColumns + `,`, + `Deprecated:` + fmt.Sprintf("%v", this.Deprecated) + `,`, + `DeprecationWarning:` + valueToStringGenerated(this.DeprecationWarning) + `,`, `}`, }, "") return s @@ -5974,6 +5999,59 @@ func (m *CustomResourceDefinitionVersion) Unmarshal(dAtA []byte) error { return err } iNdEx = postIndex + case 7: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field Deprecated", wireType) + } + var v int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + v |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + m.Deprecated = bool(v != 0) + case 8: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field DeprecationWarning", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + s := string(dAtA[iNdEx:postIndex]) + m.DeprecationWarning = &s + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -8942,6 +9020,7 @@ func (m *WebhookClientConfig) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -8973,10 +9052,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -8997,55 +9074,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.proto b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.proto index 705ca079..8a1f7b95 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.proto +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/generated.proto @@ -130,7 +130,7 @@ message CustomResourceConversion { // CustomResourceDefinition represents a resource that should be exposed on the API server. Its name MUST be in the format // <.spec.name>.<.spec.group>. -// Deprecated in v1.16, planned for removal in v1.19. Use apiextensions.k8s.io/v1 CustomResourceDefinition instead. +// Deprecated in v1.16, planned for removal in v1.22. Use apiextensions.k8s.io/v1 CustomResourceDefinition instead. message CustomResourceDefinition { optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1; @@ -284,6 +284,7 @@ message CustomResourceDefinitionStatus { // acceptedNames are the names that are actually being used to serve discovery. // They may be different than the names in spec. + // +optional optional CustomResourceDefinitionNames acceptedNames = 2; // storedVersions lists all versions of CustomResources that were ever persisted. Tracking these @@ -292,6 +293,7 @@ message CustomResourceDefinitionStatus { // no old objects are left in storage), and then remove the rest of the // versions from this list. // Versions may not be removed from `spec.versions` while they exist in this list. + // +optional repeated string storedVersions = 3; } @@ -308,6 +310,19 @@ message CustomResourceDefinitionVersion { // There must be exactly one version with storage=true. optional bool storage = 3; + // deprecated indicates this version of the custom resource API is deprecated. + // When set to true, API requests to this version receive a warning header in the server response. + // Defaults to false. + // +optional + optional bool deprecated = 7; + + // deprecationWarning overrides the default warning returned to API clients. + // May only be set when `deprecated` is true. + // The default warning indicates this version is deprecated and recommends use + // of the newest served version of equal or greater stability, if one exists. + // +optional + optional string deprecationWarning = 8; + // schema describes the schema used for validation and pruning of this version of the custom resource. // Top-level and per-version schemas are mutually exclusive. // Per-version schemas must not all be set to identical values (top-level validation schema should be used instead). @@ -543,6 +558,9 @@ message JSONSchemaProps { // extension set to "map". Also, the values specified for this attribute must // be a scalar typed field of the child structure (no nesting is supported). // + // The properties specified must either be required or have a default value, + // to ensure those properties are present for all list items. + // // +optional repeated string xKubernetesListMapKeys = 41; diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types.go index f6c260b6..806c68aa 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types.go @@ -200,6 +200,17 @@ type CustomResourceDefinitionVersion struct { // storage indicates this version should be used when persisting custom resources to storage. // There must be exactly one version with storage=true. Storage bool `json:"storage" protobuf:"varint,3,opt,name=storage"` + // deprecated indicates this version of the custom resource API is deprecated. + // When set to true, API requests to this version receive a warning header in the server response. + // Defaults to false. + // +optional + Deprecated bool `json:"deprecated,omitempty" protobuf:"varint,7,opt,name=deprecated"` + // deprecationWarning overrides the default warning returned to API clients. + // May only be set when `deprecated` is true. + // The default warning indicates this version is deprecated and recommends use + // of the newest served version of equal or greater stability, if one exists. + // +optional + DeprecationWarning *string `json:"deprecationWarning,omitempty" protobuf:"bytes,8,opt,name=deprecationWarning"` // schema describes the schema used for validation and pruning of this version of the custom resource. // Top-level and per-version schemas are mutually exclusive. // Per-version schemas must not all be set to identical values (top-level validation schema should be used instead). @@ -354,6 +365,7 @@ type CustomResourceDefinitionStatus struct { // acceptedNames are the names that are actually being used to serve discovery. // They may be different than the names in spec. + // +optional AcceptedNames CustomResourceDefinitionNames `json:"acceptedNames" protobuf:"bytes,2,opt,name=acceptedNames"` // storedVersions lists all versions of CustomResources that were ever persisted. Tracking these @@ -362,6 +374,7 @@ type CustomResourceDefinitionStatus struct { // no old objects are left in storage), and then remove the rest of the // versions from this list. // Versions may not be removed from `spec.versions` while they exist in this list. + // +optional StoredVersions []string `json:"storedVersions" protobuf:"bytes,3,rep,name=storedVersions"` } @@ -372,10 +385,14 @@ const CustomResourceCleanupFinalizer = "customresourcecleanup.apiextensions.k8s. // +genclient // +genclient:nonNamespaced // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.7 +// +k8s:prerelease-lifecycle-gen:deprecated=1.16 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=apiextensions.k8s.io,v1,CustomResourceDefinition // CustomResourceDefinition represents a resource that should be exposed on the API server. Its name MUST be in the format // <.spec.name>.<.spec.group>. -// Deprecated in v1.16, planned for removal in v1.19. Use apiextensions.k8s.io/v1 CustomResourceDefinition instead. +// Deprecated in v1.16, planned for removal in v1.22. Use apiextensions.k8s.io/v1 CustomResourceDefinition instead. type CustomResourceDefinition struct { metav1.TypeMeta `json:",inline"` metav1.ObjectMeta `json:"metadata,omitempty" protobuf:"bytes,1,opt,name=metadata"` @@ -388,6 +405,10 @@ type CustomResourceDefinition struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.7 +// +k8s:prerelease-lifecycle-gen:deprecated=1.16 +// +k8s:prerelease-lifecycle-gen:removed=1.22 +// +k8s:prerelease-lifecycle-gen:replacement=apiextensions.k8s.io,v1,CustomResourceDefinitionList // CustomResourceDefinitionList is a list of CustomResourceDefinition objects. type CustomResourceDefinitionList struct { @@ -452,6 +473,11 @@ type CustomResourceSubresourceScale struct { } // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object +// +k8s:prerelease-lifecycle-gen:introduced=1.13 +// +k8s:prerelease-lifecycle-gen:deprecated=1.19 +// This API is never served. It is used for outbound requests from apiservers. This will ensure it never gets served accidentally +// and having the generator against this group will protect future APIs which may be served. +// +k8s:prerelease-lifecycle-gen:replacement=apiextensions.k8s.io,v1,ConversionReview // ConversionReview describes a conversion request/response. type ConversionReview struct { diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types_jsonschema.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types_jsonschema.go index b51a3249..1837723a 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types_jsonschema.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/types_jsonschema.go @@ -126,6 +126,9 @@ type JSONSchemaProps struct { // extension set to "map". Also, the values specified for this attribute must // be a scalar typed field of the child structure (no nesting is supported). // + // The properties specified must either be required or have a default value, + // to ensure those properties are present for all list items. + // // +optional XListMapKeys []string `json:"x-kubernetes-list-map-keys,omitempty" protobuf:"bytes,41,rep,name=xKubernetesListMapKeys"` diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.conversion.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.conversion.go index 95d430c5..322b4d28 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.conversion.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.conversion.go @@ -577,6 +577,8 @@ func autoConvert_v1beta1_CustomResourceDefinitionVersion_To_apiextensions_Custom out.Name = in.Name out.Served = in.Served out.Storage = in.Storage + out.Deprecated = in.Deprecated + out.DeprecationWarning = (*string)(unsafe.Pointer(in.DeprecationWarning)) if in.Schema != nil { in, out := &in.Schema, &out.Schema *out = new(apiextensions.CustomResourceValidation) @@ -600,6 +602,8 @@ func autoConvert_apiextensions_CustomResourceDefinitionVersion_To_v1beta1_Custom out.Name = in.Name out.Served = in.Served out.Storage = in.Storage + out.Deprecated = in.Deprecated + out.DeprecationWarning = (*string)(unsafe.Pointer(in.DeprecationWarning)) if in.Schema != nil { in, out := &in.Schema, &out.Schema *out = new(CustomResourceValidation) diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.deepcopy.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.deepcopy.go index 82bbb2be..b39a5900 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.deepcopy.go @@ -333,6 +333,11 @@ func (in *CustomResourceDefinitionStatus) DeepCopy() *CustomResourceDefinitionSt // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *CustomResourceDefinitionVersion) DeepCopyInto(out *CustomResourceDefinitionVersion) { *out = *in + if in.DeprecationWarning != nil { + in, out := &in.DeprecationWarning, &out.DeprecationWarning + *out = new(string) + **out = **in + } if in.Schema != nil { in, out := &in.Schema, &out.Schema *out = new(CustomResourceValidation) diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.prerelease-lifecycle.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.prerelease-lifecycle.go new file mode 100644 index 00000000..7b73b9cf --- /dev/null +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1/zz_generated.prerelease-lifecycle.go @@ -0,0 +1,97 @@ +// +build !ignore_autogenerated + +/* +Copyright The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Code generated by prerelease-lifecycle-gen. DO NOT EDIT. + +package v1beta1 + +import ( + schema "k8s.io/apimachinery/pkg/runtime/schema" +) + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *ConversionReview) APILifecycleIntroduced() (major, minor int) { + return 1, 13 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *ConversionReview) APILifecycleDeprecated() (major, minor int) { + return 1, 19 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *ConversionReview) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apiextensions.k8s.io", Version: "v1", Kind: "ConversionReview"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *ConversionReview) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CustomResourceDefinition) APILifecycleIntroduced() (major, minor int) { + return 1, 7 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CustomResourceDefinition) APILifecycleDeprecated() (major, minor int) { + return 1, 16 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CustomResourceDefinition) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apiextensions.k8s.io", Version: "v1", Kind: "CustomResourceDefinition"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CustomResourceDefinition) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} + +// APILifecycleIntroduced is an autogenerated function, returning the release in which the API struct was introduced as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:introduced" tags in types.go. +func (in *CustomResourceDefinitionList) APILifecycleIntroduced() (major, minor int) { + return 1, 7 +} + +// APILifecycleDeprecated is an autogenerated function, returning the release in which the API struct was or will be deprecated as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:deprecated" tags in types.go or "k8s:prerelease-lifecycle-gen:introduced" plus three minor. +func (in *CustomResourceDefinitionList) APILifecycleDeprecated() (major, minor int) { + return 1, 16 +} + +// APILifecycleReplacement is an autogenerated function, returning the group, version, and kind that should be used instead of this deprecated type. +// It is controlled by "k8s:prerelease-lifecycle-gen:replacement=,," tags in types.go. +func (in *CustomResourceDefinitionList) APILifecycleReplacement() schema.GroupVersionKind { + return schema.GroupVersionKind{Group: "apiextensions.k8s.io", Version: "v1", Kind: "CustomResourceDefinitionList"} +} + +// APILifecycleRemoved is an autogenerated function, returning the release in which the API is no longer served as int versions of major and minor for comparison. +// It is controlled by "k8s:prerelease-lifecycle-gen:removed" tags in types.go or "k8s:prerelease-lifecycle-gen:deprecated" plus three minor. +func (in *CustomResourceDefinitionList) APILifecycleRemoved() (major, minor int) { + return 1, 22 +} diff --git a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/zz_generated.deepcopy.go b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/zz_generated.deepcopy.go index 682e6fd4..ba9e31e9 100644 --- a/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/zz_generated.deepcopy.go +++ b/vendor/k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/zz_generated.deepcopy.go @@ -251,6 +251,11 @@ func (in *CustomResourceDefinitionStatus) DeepCopy() *CustomResourceDefinitionSt // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *CustomResourceDefinitionVersion) DeepCopyInto(out *CustomResourceDefinitionVersion) { *out = *in + if in.DeprecationWarning != nil { + in, out := &in.DeprecationWarning, &out.DeprecationWarning + *out = new(string) + **out = **in + } if in.Schema != nil { in, out := &in.Schema, &out.Schema *out = new(CustomResourceValidation) diff --git a/vendor/k8s.io/apimachinery/pkg/api/errors/OWNERS b/vendor/k8s.io/apimachinery/pkg/api/errors/OWNERS index 435297a8..d18a1788 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/errors/OWNERS +++ b/vendor/k8s.io/apimachinery/pkg/api/errors/OWNERS @@ -17,9 +17,7 @@ reviewers: - saad-ali - janetkuo - tallclair -- eparis - dims - hongchaodeng - krousey - cjcullen -- david-mcmahon diff --git a/vendor/k8s.io/apimachinery/pkg/api/errors/errors.go b/vendor/k8s.io/apimachinery/pkg/api/errors/errors.go index e53c3e61..d3927d81 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/errors/errors.go +++ b/vendor/k8s.io/apimachinery/pkg/api/errors/errors.go @@ -18,6 +18,7 @@ package errors import ( "encoding/json" + "errors" "fmt" "net/http" "reflect" @@ -29,14 +30,6 @@ import ( "k8s.io/apimachinery/pkg/util/validation/field" ) -const ( - // StatusTooManyRequests means the server experienced too many requests within a - // given window and that the client must wait to perform the action again. - // DEPRECATED: please use http.StatusTooManyRequests, this will be removed in - // the future version. - StatusTooManyRequests = http.StatusTooManyRequests -) - // StatusError is an error intended for consumption by a REST API server; it can also be // reconstructed by clients from a REST response. Public to allow easy type switches. type StatusError struct { @@ -483,127 +476,141 @@ func NewGenericServerResponse(code int, verb string, qualifiedResource schema.Gr } // IsNotFound returns true if the specified error was created by NewNotFound. +// It supports wrapped errors. func IsNotFound(err error) bool { return ReasonForError(err) == metav1.StatusReasonNotFound } // IsAlreadyExists determines if the err is an error which indicates that a specified resource already exists. +// It supports wrapped errors. func IsAlreadyExists(err error) bool { return ReasonForError(err) == metav1.StatusReasonAlreadyExists } // IsConflict determines if the err is an error which indicates the provided update conflicts. +// It supports wrapped errors. func IsConflict(err error) bool { return ReasonForError(err) == metav1.StatusReasonConflict } // IsInvalid determines if the err is an error which indicates the provided resource is not valid. +// It supports wrapped errors. func IsInvalid(err error) bool { return ReasonForError(err) == metav1.StatusReasonInvalid } // IsGone is true if the error indicates the requested resource is no longer available. +// It supports wrapped errors. func IsGone(err error) bool { return ReasonForError(err) == metav1.StatusReasonGone } // IsResourceExpired is true if the error indicates the resource has expired and the current action is // no longer possible. +// It supports wrapped errors. func IsResourceExpired(err error) bool { return ReasonForError(err) == metav1.StatusReasonExpired } // IsNotAcceptable determines if err is an error which indicates that the request failed due to an invalid Accept header +// It supports wrapped errors. func IsNotAcceptable(err error) bool { return ReasonForError(err) == metav1.StatusReasonNotAcceptable } // IsUnsupportedMediaType determines if err is an error which indicates that the request failed due to an invalid Content-Type header +// It supports wrapped errors. func IsUnsupportedMediaType(err error) bool { return ReasonForError(err) == metav1.StatusReasonUnsupportedMediaType } // IsMethodNotSupported determines if the err is an error which indicates the provided action could not // be performed because it is not supported by the server. +// It supports wrapped errors. func IsMethodNotSupported(err error) bool { return ReasonForError(err) == metav1.StatusReasonMethodNotAllowed } // IsServiceUnavailable is true if the error indicates the underlying service is no longer available. +// It supports wrapped errors. func IsServiceUnavailable(err error) bool { return ReasonForError(err) == metav1.StatusReasonServiceUnavailable } // IsBadRequest determines if err is an error which indicates that the request is invalid. +// It supports wrapped errors. func IsBadRequest(err error) bool { return ReasonForError(err) == metav1.StatusReasonBadRequest } // IsUnauthorized determines if err is an error which indicates that the request is unauthorized and // requires authentication by the user. +// It supports wrapped errors. func IsUnauthorized(err error) bool { return ReasonForError(err) == metav1.StatusReasonUnauthorized } // IsForbidden determines if err is an error which indicates that the request is forbidden and cannot // be completed as requested. +// It supports wrapped errors. func IsForbidden(err error) bool { return ReasonForError(err) == metav1.StatusReasonForbidden } // IsTimeout determines if err is an error which indicates that request times out due to long // processing. +// It supports wrapped errors. func IsTimeout(err error) bool { return ReasonForError(err) == metav1.StatusReasonTimeout } // IsServerTimeout determines if err is an error which indicates that the request needs to be retried // by the client. +// It supports wrapped errors. func IsServerTimeout(err error) bool { return ReasonForError(err) == metav1.StatusReasonServerTimeout } // IsInternalError determines if err is an error which indicates an internal server error. +// It supports wrapped errors. func IsInternalError(err error) bool { return ReasonForError(err) == metav1.StatusReasonInternalError } // IsTooManyRequests determines if err is an error which indicates that there are too many requests // that the server cannot handle. +// It supports wrapped errors. func IsTooManyRequests(err error) bool { if ReasonForError(err) == metav1.StatusReasonTooManyRequests { return true } - switch t := err.(type) { - case APIStatus: - return t.Status().Code == http.StatusTooManyRequests + if status := APIStatus(nil); errors.As(err, &status) { + return status.Status().Code == http.StatusTooManyRequests } return false } // IsRequestEntityTooLargeError determines if err is an error which indicates // the request entity is too large. +// It supports wrapped errors. func IsRequestEntityTooLargeError(err error) bool { if ReasonForError(err) == metav1.StatusReasonRequestEntityTooLarge { return true } - switch t := err.(type) { - case APIStatus: - return t.Status().Code == http.StatusRequestEntityTooLarge + if status := APIStatus(nil); errors.As(err, &status) { + return status.Status().Code == http.StatusRequestEntityTooLarge } return false } // IsUnexpectedServerError returns true if the server response was not in the expected API format, // and may be the result of another HTTP actor. +// It supports wrapped errors. func IsUnexpectedServerError(err error) bool { - switch t := err.(type) { - case APIStatus: - if d := t.Status().Details; d != nil { - for _, cause := range d.Causes { - if cause.Type == metav1.CauseTypeUnexpectedServerResponse { - return true - } + if status := APIStatus(nil); errors.As(err, &status) && status.Status().Details != nil { + for _, cause := range status.Status().Details.Causes { + if cause.Type == metav1.CauseTypeUnexpectedServerResponse { + return true } } } @@ -611,38 +618,37 @@ func IsUnexpectedServerError(err error) bool { } // IsUnexpectedObjectError determines if err is due to an unexpected object from the master. +// It supports wrapped errors. func IsUnexpectedObjectError(err error) bool { - _, ok := err.(*UnexpectedObjectError) - return err != nil && ok + uoe := &UnexpectedObjectError{} + return err != nil && errors.As(err, &uoe) } // SuggestsClientDelay returns true if this error suggests a client delay as well as the // suggested seconds to wait, or false if the error does not imply a wait. It does not // address whether the error *should* be retried, since some errors (like a 3xx) may // request delay without retry. +// It supports wrapped errors. func SuggestsClientDelay(err error) (int, bool) { - switch t := err.(type) { - case APIStatus: - if t.Status().Details != nil { - switch t.Status().Reason { - // this StatusReason explicitly requests the caller to delay the action - case metav1.StatusReasonServerTimeout: - return int(t.Status().Details.RetryAfterSeconds), true - } - // If the client requests that we retry after a certain number of seconds - if t.Status().Details.RetryAfterSeconds > 0 { - return int(t.Status().Details.RetryAfterSeconds), true - } + if t := APIStatus(nil); errors.As(err, &t) && t.Status().Details != nil { + switch t.Status().Reason { + // this StatusReason explicitly requests the caller to delay the action + case metav1.StatusReasonServerTimeout: + return int(t.Status().Details.RetryAfterSeconds), true + } + // If the client requests that we retry after a certain number of seconds + if t.Status().Details.RetryAfterSeconds > 0 { + return int(t.Status().Details.RetryAfterSeconds), true } } return 0, false } // ReasonForError returns the HTTP status for a particular error. +// It supports wrapped errors. func ReasonForError(err error) metav1.StatusReason { - switch t := err.(type) { - case APIStatus: - return t.Status().Reason + if status := APIStatus(nil); errors.As(err, &status) { + return status.Status().Reason } return metav1.StatusReasonUnknown } diff --git a/vendor/k8s.io/apimachinery/pkg/api/meta/OWNERS b/vendor/k8s.io/apimachinery/pkg/api/meta/OWNERS index 96bccff1..68b8d353 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/meta/OWNERS +++ b/vendor/k8s.io/apimachinery/pkg/api/meta/OWNERS @@ -14,10 +14,8 @@ reviewers: - gmarek - janetkuo - ncdc -- eparis - dims - krousey - resouer -- david-mcmahon - mfojtik - jianhuiz diff --git a/vendor/k8s.io/apimachinery/pkg/api/meta/conditions.go b/vendor/k8s.io/apimachinery/pkg/api/meta/conditions.go new file mode 100644 index 00000000..934790dc --- /dev/null +++ b/vendor/k8s.io/apimachinery/pkg/api/meta/conditions.go @@ -0,0 +1,101 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package meta + +import ( + "time" + + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" +) + +// SetStatusCondition sets the corresponding condition in conditions to newCondition. +// conditions must be non-nil. +// 1. if the condition of the specified type already exists (all fields of the existing condition are updated to +// newCondition, LastTransitionTime is set to now if the new status differs from the old status) +// 2. if a condition of the specified type does not exist (LastTransitionTime is set to now() if unset, and newCondition is appended) +func SetStatusCondition(conditions *[]metav1.Condition, newCondition metav1.Condition) { + if conditions == nil { + return + } + existingCondition := FindStatusCondition(*conditions, newCondition.Type) + if existingCondition == nil { + if newCondition.LastTransitionTime.IsZero() { + newCondition.LastTransitionTime = metav1.NewTime(time.Now()) + } + *conditions = append(*conditions, newCondition) + return + } + + if existingCondition.Status != newCondition.Status { + existingCondition.Status = newCondition.Status + if !newCondition.LastTransitionTime.IsZero() { + existingCondition.LastTransitionTime = newCondition.LastTransitionTime + } else { + existingCondition.LastTransitionTime = metav1.NewTime(time.Now()) + } + } + + existingCondition.Reason = newCondition.Reason + existingCondition.Message = newCondition.Message +} + +// RemoveStatusCondition removes the corresponding conditionType from conditions. +// conditions must be non-nil. +func RemoveStatusCondition(conditions *[]metav1.Condition, conditionType string) { + if conditions == nil { + return + } + newConditions := make([]metav1.Condition, 0, len(*conditions)-1) + for _, condition := range *conditions { + if condition.Type != conditionType { + newConditions = append(newConditions, condition) + } + } + + *conditions = newConditions +} + +// FindStatusCondition finds the conditionType in conditions. +func FindStatusCondition(conditions []metav1.Condition, conditionType string) *metav1.Condition { + for i := range conditions { + if conditions[i].Type == conditionType { + return &conditions[i] + } + } + + return nil +} + +// IsStatusConditionTrue returns true when the conditionType is present and set to `metav1.ConditionTrue` +func IsStatusConditionTrue(conditions []metav1.Condition, conditionType string) bool { + return IsStatusConditionPresentAndEqual(conditions, conditionType, metav1.ConditionTrue) +} + +// IsStatusConditionFalse returns true when the conditionType is present and set to `metav1.ConditionFalse` +func IsStatusConditionFalse(conditions []metav1.Condition, conditionType string) bool { + return IsStatusConditionPresentAndEqual(conditions, conditionType, metav1.ConditionFalse) +} + +// IsStatusConditionPresentAndEqual returns true when conditionType is present and equal to status. +func IsStatusConditionPresentAndEqual(conditions []metav1.Condition, conditionType string, status metav1.ConditionStatus) bool { + for _, condition := range conditions { + if condition.Type == conditionType { + return condition.Status == status + } + } + return false +} diff --git a/vendor/k8s.io/apimachinery/pkg/api/meta/meta.go b/vendor/k8s.io/apimachinery/pkg/api/meta/meta.go index fa4b7673..9ca34c9f 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/meta/meta.go +++ b/vendor/k8s.io/apimachinery/pkg/api/meta/meta.go @@ -25,7 +25,7 @@ import ( "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apimachinery/pkg/types" - "k8s.io/klog" + "k8s.io/klog/v2" ) // errNotList is returned when an object implements the Object style interfaces but not the List style diff --git a/vendor/k8s.io/apimachinery/pkg/api/resource/OWNERS b/vendor/k8s.io/apimachinery/pkg/api/resource/OWNERS index dc774019..7ac0fe11 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/resource/OWNERS +++ b/vendor/k8s.io/apimachinery/pkg/api/resource/OWNERS @@ -9,8 +9,5 @@ reviewers: - mikedanese - saad-ali - janetkuo -- tallclair -- eparis - xiang90 - mbohlool -- david-mcmahon diff --git a/vendor/k8s.io/apimachinery/pkg/api/resource/generated.pb.go b/vendor/k8s.io/apimachinery/pkg/api/resource/generated.pb.go index 9fca2e16..2e09f4fa 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/resource/generated.pb.go +++ b/vendor/k8s.io/apimachinery/pkg/api/resource/generated.pb.go @@ -36,7 +36,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *Quantity) Reset() { *m = Quantity{} } func (*Quantity) ProtoMessage() {} diff --git a/vendor/k8s.io/apimachinery/pkg/api/resource/math.go b/vendor/k8s.io/apimachinery/pkg/api/resource/math.go index 7f63175d..8ffcb9f0 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/resource/math.go +++ b/vendor/k8s.io/apimachinery/pkg/api/resource/math.go @@ -37,12 +37,8 @@ var ( big1024 = big.NewInt(1024) // Commonly needed inf.Dec values-- treat as read only! - decZero = inf.NewDec(0, 0) - decOne = inf.NewDec(1, 0) - decMinusOne = inf.NewDec(-1, 0) - decThousand = inf.NewDec(1000, 0) - dec1024 = inf.NewDec(1024, 0) - decMinus1024 = inf.NewDec(-1024, 0) + decZero = inf.NewDec(0, 0) + decOne = inf.NewDec(1, 0) // Largest (in magnitude) number allowed. maxAllowed = infDecAmount{inf.NewDec((1<<63)-1, 0)} // == max int64 diff --git a/vendor/k8s.io/apimachinery/pkg/api/resource/quantity.go b/vendor/k8s.io/apimachinery/pkg/api/resource/quantity.go index 516d041d..d95e03aa 100644 --- a/vendor/k8s.io/apimachinery/pkg/api/resource/quantity.go +++ b/vendor/k8s.io/apimachinery/pkg/api/resource/quantity.go @@ -634,6 +634,11 @@ func (q Quantity) MarshalJSON() ([]byte, error) { return result, nil } +// ToUnstructured implements the value.UnstructuredConverter interface. +func (q Quantity) ToUnstructured() interface{} { + return q.String() +} + // UnmarshalJSON implements the json.Unmarshaller interface. // TODO: Remove support for leading/trailing whitespace func (q *Quantity) UnmarshalJSON(value []byte) error { diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/OWNERS b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/OWNERS index 77cfb0c1..15b4c875 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/OWNERS +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/OWNERS @@ -25,7 +25,6 @@ reviewers: - krousey - mml - mbohlool -- david-mcmahon - therc - mqliang - kevin-wangzefeng diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/conversion.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/conversion.go index 285a41a4..8eaebb80 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/conversion.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/conversion.go @@ -26,69 +26,10 @@ import ( "k8s.io/apimachinery/pkg/conversion" "k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/labels" - "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/types" "k8s.io/apimachinery/pkg/util/intstr" ) -func AddConversionFuncs(scheme *runtime.Scheme) error { - return scheme.AddConversionFuncs( - Convert_v1_TypeMeta_To_v1_TypeMeta, - - Convert_v1_ListMeta_To_v1_ListMeta, - - Convert_v1_DeleteOptions_To_v1_DeleteOptions, - - Convert_intstr_IntOrString_To_intstr_IntOrString, - Convert_Pointer_intstr_IntOrString_To_intstr_IntOrString, - Convert_intstr_IntOrString_To_Pointer_intstr_IntOrString, - - Convert_Pointer_v1_Duration_To_v1_Duration, - Convert_v1_Duration_To_Pointer_v1_Duration, - - Convert_Slice_string_To_v1_Time, - Convert_Slice_string_To_Pointer_v1_Time, - - Convert_v1_Time_To_v1_Time, - Convert_v1_MicroTime_To_v1_MicroTime, - - Convert_resource_Quantity_To_resource_Quantity, - - Convert_string_To_labels_Selector, - Convert_labels_Selector_To_string, - - Convert_string_To_fields_Selector, - Convert_fields_Selector_To_string, - - Convert_Pointer_bool_To_bool, - Convert_bool_To_Pointer_bool, - - Convert_Pointer_string_To_string, - Convert_string_To_Pointer_string, - - Convert_Pointer_int64_To_int, - Convert_int_To_Pointer_int64, - - Convert_Pointer_int32_To_int32, - Convert_int32_To_Pointer_int32, - - Convert_Pointer_int64_To_int64, - Convert_int64_To_Pointer_int64, - - Convert_Pointer_float64_To_float64, - Convert_float64_To_Pointer_float64, - - Convert_Map_string_To_string_To_v1_LabelSelector, - Convert_v1_LabelSelector_To_Map_string_To_string, - - Convert_Slice_string_To_Slice_int32, - - Convert_Slice_string_To_Pointer_v1_DeletionPropagation, - - Convert_Slice_string_To_v1_IncludeObjectPolicy, - ) -} - func Convert_Pointer_float64_To_float64(in **float64, out *float64, s conversion.Scope) error { if *in == nil { *out = 0 @@ -404,3 +345,11 @@ func Convert_url_Values_To_v1_DeleteOptions(in *url.Values, out *DeleteOptions, } return nil } + +// Convert_Slice_string_To_v1_ResourceVersionMatch allows converting a URL query parameter to ResourceVersionMatch +func Convert_Slice_string_To_v1_ResourceVersionMatch(in *[]string, out *ResourceVersionMatch, s conversion.Scope) error { + if len(*in) > 0 { + *out = ResourceVersionMatch((*in)[0]) + } + return nil +} diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/duration.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/duration.go index babe8a8b..a22b0787 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/duration.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/duration.go @@ -49,6 +49,11 @@ func (d Duration) MarshalJSON() ([]byte, error) { return json.Marshal(d.Duration.String()) } +// ToUnstructured implements the value.UnstructuredConverter interface. +func (d Duration) ToUnstructured() interface{} { + return d.Duration.String() +} + // OpenAPISchemaType is used by the kube-openapi generator when constructing // the OpenAPI spec of this type. // diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.pb.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.pb.go index 31b1d955..e74a5109 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.pb.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.pb.go @@ -47,7 +47,7 @@ var _ = time.Kitchen // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *APIGroup) Reset() { *m = APIGroup{} } func (*APIGroup) ProtoMessage() {} @@ -189,10 +189,38 @@ func (m *APIVersions) XXX_DiscardUnknown() { var xxx_messageInfo_APIVersions proto.InternalMessageInfo +func (m *Condition) Reset() { *m = Condition{} } +func (*Condition) ProtoMessage() {} +func (*Condition) Descriptor() ([]byte, []int) { + return fileDescriptor_cf52fa777ced5367, []int{5} +} +func (m *Condition) XXX_Unmarshal(b []byte) error { + return m.Unmarshal(b) +} +func (m *Condition) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + b = b[:cap(b)] + n, err := m.MarshalToSizedBuffer(b) + if err != nil { + return nil, err + } + return b[:n], nil +} +func (m *Condition) XXX_Merge(src proto.Message) { + xxx_messageInfo_Condition.Merge(m, src) +} +func (m *Condition) XXX_Size() int { + return m.Size() +} +func (m *Condition) XXX_DiscardUnknown() { + xxx_messageInfo_Condition.DiscardUnknown(m) +} + +var xxx_messageInfo_Condition proto.InternalMessageInfo + func (m *CreateOptions) Reset() { *m = CreateOptions{} } func (*CreateOptions) ProtoMessage() {} func (*CreateOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{5} + return fileDescriptor_cf52fa777ced5367, []int{6} } func (m *CreateOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -220,7 +248,7 @@ var xxx_messageInfo_CreateOptions proto.InternalMessageInfo func (m *DeleteOptions) Reset() { *m = DeleteOptions{} } func (*DeleteOptions) ProtoMessage() {} func (*DeleteOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{6} + return fileDescriptor_cf52fa777ced5367, []int{7} } func (m *DeleteOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -248,7 +276,7 @@ var xxx_messageInfo_DeleteOptions proto.InternalMessageInfo func (m *Duration) Reset() { *m = Duration{} } func (*Duration) ProtoMessage() {} func (*Duration) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{7} + return fileDescriptor_cf52fa777ced5367, []int{8} } func (m *Duration) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -276,7 +304,7 @@ var xxx_messageInfo_Duration proto.InternalMessageInfo func (m *ExportOptions) Reset() { *m = ExportOptions{} } func (*ExportOptions) ProtoMessage() {} func (*ExportOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{8} + return fileDescriptor_cf52fa777ced5367, []int{9} } func (m *ExportOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -304,7 +332,7 @@ var xxx_messageInfo_ExportOptions proto.InternalMessageInfo func (m *FieldsV1) Reset() { *m = FieldsV1{} } func (*FieldsV1) ProtoMessage() {} func (*FieldsV1) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{9} + return fileDescriptor_cf52fa777ced5367, []int{10} } func (m *FieldsV1) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -332,7 +360,7 @@ var xxx_messageInfo_FieldsV1 proto.InternalMessageInfo func (m *GetOptions) Reset() { *m = GetOptions{} } func (*GetOptions) ProtoMessage() {} func (*GetOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{10} + return fileDescriptor_cf52fa777ced5367, []int{11} } func (m *GetOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -360,7 +388,7 @@ var xxx_messageInfo_GetOptions proto.InternalMessageInfo func (m *GroupKind) Reset() { *m = GroupKind{} } func (*GroupKind) ProtoMessage() {} func (*GroupKind) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{11} + return fileDescriptor_cf52fa777ced5367, []int{12} } func (m *GroupKind) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -388,7 +416,7 @@ var xxx_messageInfo_GroupKind proto.InternalMessageInfo func (m *GroupResource) Reset() { *m = GroupResource{} } func (*GroupResource) ProtoMessage() {} func (*GroupResource) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{12} + return fileDescriptor_cf52fa777ced5367, []int{13} } func (m *GroupResource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -416,7 +444,7 @@ var xxx_messageInfo_GroupResource proto.InternalMessageInfo func (m *GroupVersion) Reset() { *m = GroupVersion{} } func (*GroupVersion) ProtoMessage() {} func (*GroupVersion) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{13} + return fileDescriptor_cf52fa777ced5367, []int{14} } func (m *GroupVersion) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -444,7 +472,7 @@ var xxx_messageInfo_GroupVersion proto.InternalMessageInfo func (m *GroupVersionForDiscovery) Reset() { *m = GroupVersionForDiscovery{} } func (*GroupVersionForDiscovery) ProtoMessage() {} func (*GroupVersionForDiscovery) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{14} + return fileDescriptor_cf52fa777ced5367, []int{15} } func (m *GroupVersionForDiscovery) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -472,7 +500,7 @@ var xxx_messageInfo_GroupVersionForDiscovery proto.InternalMessageInfo func (m *GroupVersionKind) Reset() { *m = GroupVersionKind{} } func (*GroupVersionKind) ProtoMessage() {} func (*GroupVersionKind) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{15} + return fileDescriptor_cf52fa777ced5367, []int{16} } func (m *GroupVersionKind) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -500,7 +528,7 @@ var xxx_messageInfo_GroupVersionKind proto.InternalMessageInfo func (m *GroupVersionResource) Reset() { *m = GroupVersionResource{} } func (*GroupVersionResource) ProtoMessage() {} func (*GroupVersionResource) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{16} + return fileDescriptor_cf52fa777ced5367, []int{17} } func (m *GroupVersionResource) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -528,7 +556,7 @@ var xxx_messageInfo_GroupVersionResource proto.InternalMessageInfo func (m *LabelSelector) Reset() { *m = LabelSelector{} } func (*LabelSelector) ProtoMessage() {} func (*LabelSelector) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{17} + return fileDescriptor_cf52fa777ced5367, []int{18} } func (m *LabelSelector) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -556,7 +584,7 @@ var xxx_messageInfo_LabelSelector proto.InternalMessageInfo func (m *LabelSelectorRequirement) Reset() { *m = LabelSelectorRequirement{} } func (*LabelSelectorRequirement) ProtoMessage() {} func (*LabelSelectorRequirement) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{18} + return fileDescriptor_cf52fa777ced5367, []int{19} } func (m *LabelSelectorRequirement) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -584,7 +612,7 @@ var xxx_messageInfo_LabelSelectorRequirement proto.InternalMessageInfo func (m *List) Reset() { *m = List{} } func (*List) ProtoMessage() {} func (*List) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{19} + return fileDescriptor_cf52fa777ced5367, []int{20} } func (m *List) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -612,7 +640,7 @@ var xxx_messageInfo_List proto.InternalMessageInfo func (m *ListMeta) Reset() { *m = ListMeta{} } func (*ListMeta) ProtoMessage() {} func (*ListMeta) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{20} + return fileDescriptor_cf52fa777ced5367, []int{21} } func (m *ListMeta) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -640,7 +668,7 @@ var xxx_messageInfo_ListMeta proto.InternalMessageInfo func (m *ListOptions) Reset() { *m = ListOptions{} } func (*ListOptions) ProtoMessage() {} func (*ListOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{21} + return fileDescriptor_cf52fa777ced5367, []int{22} } func (m *ListOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -668,7 +696,7 @@ var xxx_messageInfo_ListOptions proto.InternalMessageInfo func (m *ManagedFieldsEntry) Reset() { *m = ManagedFieldsEntry{} } func (*ManagedFieldsEntry) ProtoMessage() {} func (*ManagedFieldsEntry) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{22} + return fileDescriptor_cf52fa777ced5367, []int{23} } func (m *ManagedFieldsEntry) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -696,7 +724,7 @@ var xxx_messageInfo_ManagedFieldsEntry proto.InternalMessageInfo func (m *MicroTime) Reset() { *m = MicroTime{} } func (*MicroTime) ProtoMessage() {} func (*MicroTime) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{23} + return fileDescriptor_cf52fa777ced5367, []int{24} } func (m *MicroTime) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MicroTime.Unmarshal(m, b) @@ -719,7 +747,7 @@ var xxx_messageInfo_MicroTime proto.InternalMessageInfo func (m *ObjectMeta) Reset() { *m = ObjectMeta{} } func (*ObjectMeta) ProtoMessage() {} func (*ObjectMeta) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{24} + return fileDescriptor_cf52fa777ced5367, []int{25} } func (m *ObjectMeta) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -747,7 +775,7 @@ var xxx_messageInfo_ObjectMeta proto.InternalMessageInfo func (m *OwnerReference) Reset() { *m = OwnerReference{} } func (*OwnerReference) ProtoMessage() {} func (*OwnerReference) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{25} + return fileDescriptor_cf52fa777ced5367, []int{26} } func (m *OwnerReference) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -775,7 +803,7 @@ var xxx_messageInfo_OwnerReference proto.InternalMessageInfo func (m *PartialObjectMetadata) Reset() { *m = PartialObjectMetadata{} } func (*PartialObjectMetadata) ProtoMessage() {} func (*PartialObjectMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{26} + return fileDescriptor_cf52fa777ced5367, []int{27} } func (m *PartialObjectMetadata) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -803,7 +831,7 @@ var xxx_messageInfo_PartialObjectMetadata proto.InternalMessageInfo func (m *PartialObjectMetadataList) Reset() { *m = PartialObjectMetadataList{} } func (*PartialObjectMetadataList) ProtoMessage() {} func (*PartialObjectMetadataList) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{27} + return fileDescriptor_cf52fa777ced5367, []int{28} } func (m *PartialObjectMetadataList) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -831,7 +859,7 @@ var xxx_messageInfo_PartialObjectMetadataList proto.InternalMessageInfo func (m *Patch) Reset() { *m = Patch{} } func (*Patch) ProtoMessage() {} func (*Patch) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{28} + return fileDescriptor_cf52fa777ced5367, []int{29} } func (m *Patch) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -859,7 +887,7 @@ var xxx_messageInfo_Patch proto.InternalMessageInfo func (m *PatchOptions) Reset() { *m = PatchOptions{} } func (*PatchOptions) ProtoMessage() {} func (*PatchOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{29} + return fileDescriptor_cf52fa777ced5367, []int{30} } func (m *PatchOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -887,7 +915,7 @@ var xxx_messageInfo_PatchOptions proto.InternalMessageInfo func (m *Preconditions) Reset() { *m = Preconditions{} } func (*Preconditions) ProtoMessage() {} func (*Preconditions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{30} + return fileDescriptor_cf52fa777ced5367, []int{31} } func (m *Preconditions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -915,7 +943,7 @@ var xxx_messageInfo_Preconditions proto.InternalMessageInfo func (m *RootPaths) Reset() { *m = RootPaths{} } func (*RootPaths) ProtoMessage() {} func (*RootPaths) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{31} + return fileDescriptor_cf52fa777ced5367, []int{32} } func (m *RootPaths) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -943,7 +971,7 @@ var xxx_messageInfo_RootPaths proto.InternalMessageInfo func (m *ServerAddressByClientCIDR) Reset() { *m = ServerAddressByClientCIDR{} } func (*ServerAddressByClientCIDR) ProtoMessage() {} func (*ServerAddressByClientCIDR) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{32} + return fileDescriptor_cf52fa777ced5367, []int{33} } func (m *ServerAddressByClientCIDR) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -971,7 +999,7 @@ var xxx_messageInfo_ServerAddressByClientCIDR proto.InternalMessageInfo func (m *Status) Reset() { *m = Status{} } func (*Status) ProtoMessage() {} func (*Status) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{33} + return fileDescriptor_cf52fa777ced5367, []int{34} } func (m *Status) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -999,7 +1027,7 @@ var xxx_messageInfo_Status proto.InternalMessageInfo func (m *StatusCause) Reset() { *m = StatusCause{} } func (*StatusCause) ProtoMessage() {} func (*StatusCause) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{34} + return fileDescriptor_cf52fa777ced5367, []int{35} } func (m *StatusCause) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1027,7 +1055,7 @@ var xxx_messageInfo_StatusCause proto.InternalMessageInfo func (m *StatusDetails) Reset() { *m = StatusDetails{} } func (*StatusDetails) ProtoMessage() {} func (*StatusDetails) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{35} + return fileDescriptor_cf52fa777ced5367, []int{36} } func (m *StatusDetails) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1055,7 +1083,7 @@ var xxx_messageInfo_StatusDetails proto.InternalMessageInfo func (m *TableOptions) Reset() { *m = TableOptions{} } func (*TableOptions) ProtoMessage() {} func (*TableOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{36} + return fileDescriptor_cf52fa777ced5367, []int{37} } func (m *TableOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1083,7 +1111,7 @@ var xxx_messageInfo_TableOptions proto.InternalMessageInfo func (m *Time) Reset() { *m = Time{} } func (*Time) ProtoMessage() {} func (*Time) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{37} + return fileDescriptor_cf52fa777ced5367, []int{38} } func (m *Time) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Time.Unmarshal(m, b) @@ -1106,7 +1134,7 @@ var xxx_messageInfo_Time proto.InternalMessageInfo func (m *Timestamp) Reset() { *m = Timestamp{} } func (*Timestamp) ProtoMessage() {} func (*Timestamp) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{38} + return fileDescriptor_cf52fa777ced5367, []int{39} } func (m *Timestamp) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1134,7 +1162,7 @@ var xxx_messageInfo_Timestamp proto.InternalMessageInfo func (m *TypeMeta) Reset() { *m = TypeMeta{} } func (*TypeMeta) ProtoMessage() {} func (*TypeMeta) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{39} + return fileDescriptor_cf52fa777ced5367, []int{40} } func (m *TypeMeta) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1162,7 +1190,7 @@ var xxx_messageInfo_TypeMeta proto.InternalMessageInfo func (m *UpdateOptions) Reset() { *m = UpdateOptions{} } func (*UpdateOptions) ProtoMessage() {} func (*UpdateOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{40} + return fileDescriptor_cf52fa777ced5367, []int{41} } func (m *UpdateOptions) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1190,7 +1218,7 @@ var xxx_messageInfo_UpdateOptions proto.InternalMessageInfo func (m *Verbs) Reset() { *m = Verbs{} } func (*Verbs) ProtoMessage() {} func (*Verbs) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{41} + return fileDescriptor_cf52fa777ced5367, []int{42} } func (m *Verbs) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1218,7 +1246,7 @@ var xxx_messageInfo_Verbs proto.InternalMessageInfo func (m *WatchEvent) Reset() { *m = WatchEvent{} } func (*WatchEvent) ProtoMessage() {} func (*WatchEvent) Descriptor() ([]byte, []int) { - return fileDescriptor_cf52fa777ced5367, []int{42} + return fileDescriptor_cf52fa777ced5367, []int{43} } func (m *WatchEvent) XXX_Unmarshal(b []byte) error { return m.Unmarshal(b) @@ -1249,6 +1277,7 @@ func init() { proto.RegisterType((*APIResource)(nil), "k8s.io.apimachinery.pkg.apis.meta.v1.APIResource") proto.RegisterType((*APIResourceList)(nil), "k8s.io.apimachinery.pkg.apis.meta.v1.APIResourceList") proto.RegisterType((*APIVersions)(nil), "k8s.io.apimachinery.pkg.apis.meta.v1.APIVersions") + proto.RegisterType((*Condition)(nil), "k8s.io.apimachinery.pkg.apis.meta.v1.Condition") proto.RegisterType((*CreateOptions)(nil), "k8s.io.apimachinery.pkg.apis.meta.v1.CreateOptions") proto.RegisterType((*DeleteOptions)(nil), "k8s.io.apimachinery.pkg.apis.meta.v1.DeleteOptions") proto.RegisterType((*Duration)(nil), "k8s.io.apimachinery.pkg.apis.meta.v1.Duration") @@ -1297,177 +1326,184 @@ func init() { } var fileDescriptor_cf52fa777ced5367 = []byte{ - // 2713 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x19, 0xcd, 0x6f, 0x1b, 0x59, - 0x3d, 0x63, 0xc7, 0x8e, 0xfd, 0x73, 0x9c, 0x8f, 0x97, 0x16, 0xdc, 0x00, 0x71, 0x76, 0x16, 0xad, - 0x52, 0xe8, 0x3a, 0x9b, 0x02, 0xab, 0xd2, 0x65, 0x0b, 0x71, 0x9c, 0x74, 0xc3, 0x36, 0x4d, 0xf4, - 0xd2, 0x16, 0x28, 0x15, 0xea, 0x64, 0xe6, 0xc5, 0x19, 0x32, 0x9e, 0xf1, 0xbe, 0x19, 0x27, 0x35, - 0x1c, 0xd8, 0x03, 0x08, 0x90, 0x60, 0xd5, 0x23, 0xe2, 0x80, 0xb6, 0x82, 0xbf, 0x80, 0x13, 0x7f, - 0x00, 0x12, 0xbd, 0x20, 0xad, 0xc4, 0x65, 0x25, 0x90, 0xb5, 0x0d, 0x07, 0x8e, 0x88, 0x6b, 0x4e, - 0xe8, 0x7d, 0xcd, 0x87, 0x1d, 0x37, 0x63, 0xba, 0xac, 0xf6, 0xe6, 0xf9, 0x7d, 0xff, 0xde, 0xfb, - 0xbd, 0xdf, 0x97, 0x61, 0xeb, 0xf0, 0x9a, 0x5f, 0xb3, 0xbd, 0xe5, 0xc3, 0xce, 0x1e, 0xa1, 0x2e, - 0x09, 0x88, 0xbf, 0x7c, 0x44, 0x5c, 0xcb, 0xa3, 0xcb, 0x12, 0x61, 0xb4, 0xed, 0x96, 0x61, 0x1e, - 0xd8, 0x2e, 0xa1, 0xdd, 0xe5, 0xf6, 0x61, 0x93, 0x01, 0xfc, 0xe5, 0x16, 0x09, 0x8c, 0xe5, 0xa3, - 0x95, 0xe5, 0x26, 0x71, 0x09, 0x35, 0x02, 0x62, 0xd5, 0xda, 0xd4, 0x0b, 0x3c, 0xf4, 0x45, 0xc1, - 0x55, 0x8b, 0x73, 0xd5, 0xda, 0x87, 0x4d, 0x06, 0xf0, 0x6b, 0x8c, 0xab, 0x76, 0xb4, 0x32, 0xff, - 0x6a, 0xd3, 0x0e, 0x0e, 0x3a, 0x7b, 0x35, 0xd3, 0x6b, 0x2d, 0x37, 0xbd, 0xa6, 0xb7, 0xcc, 0x99, - 0xf7, 0x3a, 0xfb, 0xfc, 0x8b, 0x7f, 0xf0, 0x5f, 0x42, 0xe8, 0xfc, 0x50, 0x53, 0x68, 0xc7, 0x0d, - 0xec, 0x16, 0xe9, 0xb7, 0x62, 0xfe, 0xf5, 0xf3, 0x18, 0x7c, 0xf3, 0x80, 0xb4, 0x8c, 0x7e, 0x3e, - 0xfd, 0x2f, 0x59, 0x28, 0xac, 0xee, 0x6c, 0xde, 0xa4, 0x5e, 0xa7, 0x8d, 0x16, 0x61, 0xdc, 0x35, - 0x5a, 0xa4, 0xa2, 0x2d, 0x6a, 0x4b, 0xc5, 0xfa, 0xe4, 0xd3, 0x5e, 0x75, 0xec, 0xa4, 0x57, 0x1d, - 0xbf, 0x6d, 0xb4, 0x08, 0xe6, 0x18, 0xe4, 0x40, 0xe1, 0x88, 0x50, 0xdf, 0xf6, 0x5c, 0xbf, 0x92, - 0x59, 0xcc, 0x2e, 0x95, 0xae, 0xde, 0xa8, 0xa5, 0xf1, 0xbf, 0xc6, 0x15, 0xdc, 0x13, 0xac, 0x1b, - 0x1e, 0x6d, 0xd8, 0xbe, 0xe9, 0x1d, 0x11, 0xda, 0xad, 0xcf, 0x48, 0x2d, 0x05, 0x89, 0xf4, 0x71, - 0xa8, 0x01, 0xfd, 0x54, 0x83, 0x99, 0x36, 0x25, 0xfb, 0x84, 0x52, 0x62, 0x49, 0x7c, 0x25, 0xbb, - 0xa8, 0x7d, 0x0c, 0x6a, 0x2b, 0x52, 0xed, 0xcc, 0x4e, 0x9f, 0x7c, 0x3c, 0xa0, 0x11, 0xfd, 0x5e, - 0x83, 0x79, 0x9f, 0xd0, 0x23, 0x42, 0x57, 0x2d, 0x8b, 0x12, 0xdf, 0xaf, 0x77, 0xd7, 0x1c, 0x9b, - 0xb8, 0xc1, 0xda, 0x66, 0x03, 0xfb, 0x95, 0x71, 0x7e, 0x0e, 0xdf, 0x4c, 0x67, 0xd0, 0xee, 0x30, - 0x39, 0x75, 0x5d, 0x5a, 0x34, 0x3f, 0x94, 0xc4, 0xc7, 0xcf, 0x31, 0x43, 0xdf, 0x87, 0x49, 0x75, - 0x91, 0xb7, 0x6c, 0x3f, 0x40, 0xf7, 0x20, 0xdf, 0x64, 0x1f, 0x7e, 0x45, 0xe3, 0x06, 0xd6, 0xd2, - 0x19, 0xa8, 0x64, 0xd4, 0xa7, 0xa4, 0x3d, 0x79, 0xfe, 0xe9, 0x63, 0x29, 0x4d, 0xff, 0xe5, 0x38, - 0x94, 0x56, 0x77, 0x36, 0x31, 0xf1, 0xbd, 0x0e, 0x35, 0x49, 0x8a, 0xa0, 0xb9, 0x06, 0x93, 0xbe, - 0xed, 0x36, 0x3b, 0x8e, 0x41, 0x19, 0xb4, 0x92, 0xe7, 0x94, 0x17, 0x24, 0xe5, 0xe4, 0x6e, 0x0c, - 0x87, 0x13, 0x94, 0xe8, 0x2a, 0x00, 0x93, 0xe0, 0xb7, 0x0d, 0x93, 0x58, 0x95, 0xcc, 0xa2, 0xb6, - 0x54, 0xa8, 0x23, 0xc9, 0x07, 0xb7, 0x43, 0x0c, 0x8e, 0x51, 0xa1, 0x97, 0x21, 0xc7, 0x2d, 0xad, - 0x14, 0xb8, 0x9a, 0xb2, 0x24, 0xcf, 0x71, 0x37, 0xb0, 0xc0, 0xa1, 0xcb, 0x30, 0x21, 0xa3, 0xac, - 0x52, 0xe4, 0x64, 0xd3, 0x92, 0x6c, 0x42, 0x85, 0x81, 0xc2, 0x33, 0xff, 0x0e, 0x6d, 0xd7, 0xe2, - 0x71, 0x17, 0xf3, 0xef, 0x6d, 0xdb, 0xb5, 0x30, 0xc7, 0xa0, 0x5b, 0x90, 0x3b, 0x22, 0x74, 0x8f, - 0x45, 0x02, 0x0b, 0xcd, 0x2f, 0xa7, 0x3b, 0xe8, 0x7b, 0x8c, 0xa5, 0x5e, 0x64, 0xa6, 0xf1, 0x9f, - 0x58, 0x08, 0x41, 0x35, 0x00, 0xff, 0xc0, 0xa3, 0x01, 0x77, 0xaf, 0x92, 0x5b, 0xcc, 0x2e, 0x15, - 0xeb, 0x53, 0xcc, 0xdf, 0xdd, 0x10, 0x8a, 0x63, 0x14, 0x8c, 0xde, 0x34, 0x02, 0xd2, 0xf4, 0xa8, - 0x4d, 0xfc, 0xca, 0x44, 0x44, 0xbf, 0x16, 0x42, 0x71, 0x8c, 0x02, 0x7d, 0x1b, 0x90, 0x1f, 0x78, - 0xd4, 0x68, 0x12, 0xe9, 0xea, 0x5b, 0x86, 0x7f, 0x50, 0x01, 0xee, 0xdd, 0xbc, 0xf4, 0x0e, 0xed, - 0x0e, 0x50, 0xe0, 0x33, 0xb8, 0xf4, 0x3f, 0x6a, 0x30, 0x1d, 0x8b, 0x05, 0x1e, 0x77, 0xd7, 0x60, - 0xb2, 0x19, 0x7b, 0x75, 0x32, 0x2e, 0xc2, 0xdb, 0x8e, 0xbf, 0x48, 0x9c, 0xa0, 0x44, 0x04, 0x8a, - 0x54, 0x4a, 0x52, 0xd9, 0x65, 0x25, 0x75, 0xd0, 0x2a, 0x1b, 0x22, 0x4d, 0x31, 0xa0, 0x8f, 0x23, - 0xc9, 0xfa, 0xbf, 0x34, 0x1e, 0xc0, 0x2a, 0xdf, 0xa0, 0xa5, 0x58, 0x4e, 0xd3, 0xf8, 0xf1, 0x4d, - 0x0e, 0xc9, 0x47, 0xe7, 0x24, 0x82, 0xcc, 0xa7, 0x22, 0x11, 0x5c, 0x2f, 0xfc, 0xe6, 0xfd, 0xea, - 0xd8, 0xbb, 0xff, 0x58, 0x1c, 0xd3, 0x5b, 0x50, 0x5e, 0xa3, 0xc4, 0x08, 0xc8, 0x76, 0x3b, 0xe0, - 0x0e, 0xe8, 0x90, 0xb7, 0x68, 0x17, 0x77, 0x5c, 0xe9, 0x28, 0xb0, 0xf7, 0xdd, 0xe0, 0x10, 0x2c, - 0x31, 0xec, 0xfe, 0xf6, 0x6d, 0xe2, 0x58, 0x5b, 0x86, 0x6b, 0x34, 0x09, 0x95, 0x71, 0x1f, 0x9e, - 0xea, 0x46, 0x0c, 0x87, 0x13, 0x94, 0xfa, 0xcf, 0xb3, 0x50, 0x6e, 0x10, 0x87, 0x44, 0xfa, 0x36, - 0x00, 0x35, 0xa9, 0x61, 0x92, 0x1d, 0x42, 0x6d, 0xcf, 0xda, 0x25, 0xa6, 0xe7, 0x5a, 0x3e, 0x8f, - 0x88, 0x6c, 0xfd, 0x33, 0x2c, 0xce, 0x6e, 0x0e, 0x60, 0xf1, 0x19, 0x1c, 0xc8, 0x81, 0x72, 0x9b, - 0xf2, 0xdf, 0x76, 0x20, 0x6b, 0x0f, 0x7b, 0x69, 0x5f, 0x49, 0x77, 0xd4, 0x3b, 0x71, 0xd6, 0xfa, - 0xec, 0x49, 0xaf, 0x5a, 0x4e, 0x80, 0x70, 0x52, 0x38, 0xfa, 0x16, 0xcc, 0x78, 0xb4, 0x7d, 0x60, - 0xb8, 0x0d, 0xd2, 0x26, 0xae, 0x45, 0xdc, 0xc0, 0xe7, 0xa7, 0x50, 0xa8, 0x5f, 0x60, 0x15, 0x63, - 0xbb, 0x0f, 0x87, 0x07, 0xa8, 0xd1, 0x7d, 0x98, 0x6d, 0x53, 0xaf, 0x6d, 0x34, 0x0d, 0x26, 0x71, - 0xc7, 0x73, 0x6c, 0xb3, 0xcb, 0xb3, 0x43, 0xb1, 0x7e, 0xe5, 0xa4, 0x57, 0x9d, 0xdd, 0xe9, 0x47, - 0x9e, 0xf6, 0xaa, 0x73, 0xfc, 0xe8, 0x18, 0x24, 0x42, 0xe2, 0x41, 0x31, 0xb1, 0x3b, 0xcc, 0x0d, - 0xbb, 0x43, 0x7d, 0x13, 0x0a, 0x8d, 0x0e, 0xe5, 0x5c, 0xe8, 0x4d, 0x28, 0x58, 0xf2, 0xb7, 0x3c, - 0xf9, 0x97, 0x54, 0xc9, 0x55, 0x34, 0xa7, 0xbd, 0x6a, 0x99, 0x35, 0x09, 0x35, 0x05, 0xc0, 0x21, - 0x8b, 0xfe, 0x00, 0xca, 0xeb, 0x8f, 0xda, 0x1e, 0x0d, 0xd4, 0x9d, 0xbe, 0x02, 0x79, 0xc2, 0x01, - 0x5c, 0x5a, 0x21, 0xaa, 0x13, 0x82, 0x0c, 0x4b, 0x2c, 0xcb, 0xc3, 0xe4, 0x91, 0x61, 0x06, 0x32, - 0x6d, 0x87, 0x79, 0x78, 0x9d, 0x01, 0xb1, 0xc0, 0xe9, 0x9f, 0x87, 0x02, 0x0f, 0x28, 0xff, 0xde, - 0x0a, 0x9a, 0x81, 0x2c, 0x36, 0x8e, 0xb9, 0xd4, 0x49, 0x9c, 0xa5, 0xc6, 0xb1, 0xbe, 0x0d, 0x70, - 0x93, 0x84, 0x8a, 0x57, 0x61, 0x5a, 0x3d, 0xe2, 0x64, 0x6e, 0xf9, 0xac, 0x14, 0x3d, 0x8d, 0x93, - 0x68, 0xdc, 0x4f, 0xaf, 0x3f, 0x80, 0x22, 0xcf, 0x3f, 0x2c, 0x79, 0x47, 0x85, 0x42, 0x7b, 0x4e, - 0xa1, 0x50, 0xd9, 0x3f, 0x33, 0x2c, 0xfb, 0xc7, 0x9e, 0x9b, 0x03, 0x65, 0xc1, 0xab, 0x4a, 0x63, - 0x2a, 0x0d, 0x57, 0xa0, 0xa0, 0xcc, 0x94, 0x5a, 0xc2, 0x96, 0x48, 0x09, 0xc2, 0x21, 0x45, 0x4c, - 0xdb, 0x01, 0x24, 0x72, 0x69, 0x3a, 0x65, 0xb1, 0xba, 0x97, 0x79, 0x7e, 0xdd, 0x8b, 0x69, 0xfa, - 0x09, 0x54, 0x86, 0xf5, 0x51, 0x2f, 0x90, 0xed, 0xd3, 0x9b, 0xa2, 0xbf, 0xa7, 0xc1, 0x4c, 0x5c, - 0x52, 0xfa, 0xeb, 0x4b, 0xaf, 0xe4, 0xfc, 0x3a, 0x1f, 0x3b, 0x91, 0xdf, 0x69, 0x70, 0x21, 0xe1, - 0xda, 0x48, 0x37, 0x3e, 0x82, 0x51, 0xf1, 0xe0, 0xc8, 0x8e, 0x10, 0x1c, 0x7f, 0xcb, 0x40, 0xf9, - 0x96, 0xb1, 0x47, 0x9c, 0x5d, 0xe2, 0x10, 0x33, 0xf0, 0x28, 0xfa, 0x31, 0x94, 0x5a, 0x46, 0x60, - 0x1e, 0x70, 0xa8, 0xea, 0x09, 0x1b, 0xe9, 0x12, 0x68, 0x42, 0x52, 0x6d, 0x2b, 0x12, 0xb3, 0xee, - 0x06, 0xb4, 0x5b, 0x9f, 0x93, 0x26, 0x95, 0x62, 0x18, 0x1c, 0xd7, 0xc6, 0x1b, 0x79, 0xfe, 0xbd, - 0xfe, 0xa8, 0xcd, 0x0a, 0xd6, 0xe8, 0xf3, 0x43, 0xc2, 0x04, 0x4c, 0xde, 0xe9, 0xd8, 0x94, 0xb4, - 0x88, 0x1b, 0x44, 0x8d, 0xfc, 0x56, 0x9f, 0x7c, 0x3c, 0xa0, 0x71, 0xfe, 0x06, 0xcc, 0xf4, 0x1b, - 0xcf, 0xb2, 0xce, 0x21, 0xe9, 0x8a, 0xfb, 0xc2, 0xec, 0x27, 0xba, 0x00, 0xb9, 0x23, 0xc3, 0xe9, - 0xc8, 0xd7, 0x88, 0xc5, 0xc7, 0xf5, 0xcc, 0x35, 0x4d, 0xff, 0x83, 0x06, 0x95, 0x61, 0x86, 0xa0, - 0x2f, 0xc4, 0x04, 0xd5, 0x4b, 0xd2, 0xaa, 0xec, 0xdb, 0xa4, 0x2b, 0xa4, 0xae, 0x43, 0xc1, 0x6b, - 0xb3, 0xd1, 0xcb, 0xa3, 0xf2, 0xd6, 0x2f, 0xab, 0x9b, 0xdc, 0x96, 0xf0, 0xd3, 0x5e, 0xf5, 0x62, - 0x42, 0xbc, 0x42, 0xe0, 0x90, 0x95, 0x65, 0x7f, 0x6e, 0x0f, 0xab, 0x48, 0x61, 0xf6, 0xbf, 0xc7, - 0x21, 0x58, 0x62, 0xf4, 0x3f, 0x69, 0x30, 0xce, 0x5b, 0xb1, 0x07, 0x50, 0x60, 0xe7, 0x67, 0x19, - 0x81, 0xc1, 0xed, 0x4a, 0x3d, 0x04, 0x30, 0xee, 0x2d, 0x12, 0x18, 0x51, 0xb4, 0x29, 0x08, 0x0e, - 0x25, 0x22, 0x0c, 0x39, 0x3b, 0x20, 0x2d, 0x75, 0x91, 0xaf, 0x0e, 0x15, 0x2d, 0x47, 0xd0, 0x1a, - 0x36, 0x8e, 0xd7, 0x1f, 0x05, 0xc4, 0x65, 0x97, 0x11, 0x3d, 0x8d, 0x4d, 0x26, 0x03, 0x0b, 0x51, - 0xfa, 0x7f, 0x34, 0x08, 0x55, 0xb1, 0xe0, 0xf7, 0x89, 0xb3, 0x7f, 0xcb, 0x76, 0x0f, 0xe5, 0xb1, - 0x86, 0xe6, 0xec, 0x4a, 0x38, 0x0e, 0x29, 0xce, 0x2a, 0x0f, 0x99, 0xd1, 0xca, 0x03, 0x53, 0x68, - 0x7a, 0x6e, 0x60, 0xbb, 0x9d, 0x81, 0xd7, 0xb6, 0x26, 0xe1, 0x38, 0xa4, 0x60, 0xcd, 0x0d, 0x25, - 0x2d, 0xc3, 0x76, 0x6d, 0xb7, 0xc9, 0x9c, 0x58, 0xf3, 0x3a, 0x6e, 0xc0, 0xab, 0xbc, 0x6c, 0x6e, - 0xf0, 0x00, 0x16, 0x9f, 0xc1, 0xa1, 0xff, 0x35, 0x0b, 0x25, 0xe6, 0xb3, 0xaa, 0x73, 0x6f, 0x40, - 0xd9, 0x89, 0x47, 0x81, 0xf4, 0xfd, 0xa2, 0x34, 0x25, 0xf9, 0xae, 0x71, 0x92, 0x96, 0x31, 0xf3, - 0x9e, 0x2c, 0x64, 0xce, 0x24, 0x99, 0x37, 0xe2, 0x48, 0x9c, 0xa4, 0x65, 0xd9, 0xeb, 0x98, 0xbd, - 0x0f, 0xd9, 0xed, 0x84, 0x57, 0xf4, 0x1d, 0x06, 0xc4, 0x02, 0x87, 0xb6, 0x60, 0xce, 0x70, 0x1c, - 0xef, 0x98, 0x03, 0xeb, 0x9e, 0x77, 0xd8, 0x32, 0xe8, 0xa1, 0xcf, 0xc7, 0xa8, 0x42, 0xfd, 0x73, - 0x92, 0x65, 0x6e, 0x75, 0x90, 0x04, 0x9f, 0xc5, 0x77, 0xd6, 0xb5, 0x8d, 0x8f, 0x78, 0x6d, 0xd7, - 0x61, 0x8a, 0xc5, 0x97, 0xd7, 0x09, 0x54, 0x87, 0x99, 0xe3, 0x97, 0x80, 0x4e, 0x7a, 0xd5, 0xa9, - 0x3b, 0x09, 0x0c, 0xee, 0xa3, 0x64, 0x2e, 0x3b, 0x76, 0xcb, 0x0e, 0x2a, 0x13, 0x9c, 0x25, 0x74, - 0xf9, 0x16, 0x03, 0x62, 0x81, 0x4b, 0xc4, 0x45, 0xe1, 0xbc, 0xb8, 0xd0, 0x7f, 0x9b, 0x05, 0x24, - 0x5a, 0x62, 0x4b, 0xf4, 0x36, 0x22, 0xd1, 0x5c, 0x86, 0x89, 0x96, 0x6c, 0xa9, 0xb5, 0x64, 0xd6, - 0x57, 0xdd, 0xb4, 0xc2, 0xa3, 0x2d, 0x28, 0x8a, 0x07, 0x1f, 0x05, 0xf1, 0xb2, 0x24, 0x2e, 0x6e, - 0x2b, 0xc4, 0x69, 0xaf, 0x3a, 0x9f, 0x50, 0x13, 0x62, 0xee, 0x74, 0xdb, 0x04, 0x47, 0x12, 0xd8, - 0x14, 0x6d, 0xb4, 0xed, 0xf8, 0xfe, 0xa4, 0x18, 0x4d, 0xd1, 0xd1, 0x24, 0x84, 0x63, 0x54, 0xe8, - 0x2d, 0x18, 0x67, 0x27, 0x25, 0x47, 0xda, 0x2f, 0xa5, 0x4b, 0x1b, 0xec, 0xac, 0xeb, 0x05, 0x56, - 0x35, 0xd9, 0x2f, 0xcc, 0x25, 0x30, 0xed, 0x3c, 0xca, 0x7c, 0x66, 0x96, 0x9c, 0xfd, 0x43, 0xed, - 0x1b, 0x21, 0x06, 0xc7, 0xa8, 0xd0, 0x77, 0xa1, 0xb0, 0x2f, 0xdb, 0x42, 0x7e, 0x31, 0xa9, 0x13, - 0x97, 0x6a, 0x26, 0xc5, 0x08, 0xa7, 0xbe, 0x70, 0x28, 0x4d, 0x7f, 0x07, 0x8a, 0x5b, 0xb6, 0x49, - 0x3d, 0x66, 0x20, 0xbb, 0x12, 0x3f, 0x31, 0x93, 0x84, 0x57, 0xa2, 0xc2, 0x45, 0xe1, 0x59, 0x9c, - 0xb8, 0x86, 0xeb, 0x89, 0xc9, 0x23, 0x17, 0xc5, 0xc9, 0x6d, 0x06, 0xc4, 0x02, 0x77, 0xfd, 0x02, - 0xab, 0xbf, 0xbf, 0x78, 0x52, 0x1d, 0x7b, 0xfc, 0xa4, 0x3a, 0xf6, 0xfe, 0x13, 0x59, 0x8b, 0x4f, - 0x01, 0x60, 0x7b, 0xef, 0x87, 0xc4, 0x14, 0x59, 0x2d, 0xd5, 0xbe, 0x44, 0xad, 0xe9, 0xf8, 0xbe, - 0x24, 0xd3, 0xd7, 0x53, 0xc5, 0x70, 0x38, 0x41, 0x89, 0x96, 0xa1, 0x18, 0x6e, 0x42, 0xe4, 0x45, - 0xcf, 0xaa, 0xc0, 0x09, 0xd7, 0x25, 0x38, 0xa2, 0x49, 0xa4, 0xd8, 0xf1, 0x73, 0x53, 0x6c, 0x1d, - 0xb2, 0x1d, 0xdb, 0xe2, 0xaf, 0xab, 0x58, 0x7f, 0x4d, 0x95, 0xb8, 0xbb, 0x9b, 0x8d, 0xd3, 0x5e, - 0xf5, 0xa5, 0x61, 0x0b, 0xc8, 0xa0, 0xdb, 0x26, 0x7e, 0xed, 0xee, 0x66, 0x03, 0x33, 0xe6, 0xb3, - 0xde, 0x7b, 0x7e, 0xc4, 0xf7, 0x7e, 0x15, 0x40, 0x7a, 0xcd, 0xb8, 0xc5, 0xc3, 0x0d, 0x23, 0xea, - 0x66, 0x88, 0xc1, 0x31, 0x2a, 0xe4, 0xc3, 0xac, 0xc9, 0x46, 0x61, 0xf6, 0x3c, 0xec, 0x16, 0xf1, - 0x03, 0xa3, 0x25, 0x36, 0x44, 0xa3, 0x05, 0xf7, 0x25, 0xa9, 0x66, 0x76, 0xad, 0x5f, 0x18, 0x1e, - 0x94, 0x8f, 0x3c, 0x98, 0xb5, 0xe4, 0x50, 0x17, 0x29, 0x2d, 0x8e, 0xac, 0xf4, 0x22, 0x53, 0xd8, - 0xe8, 0x17, 0x84, 0x07, 0x65, 0xa3, 0x1f, 0xc0, 0xbc, 0x02, 0x0e, 0x4e, 0xd6, 0x7c, 0xc7, 0x93, - 0xad, 0x2f, 0x9c, 0xf4, 0xaa, 0xf3, 0x8d, 0xa1, 0x54, 0xf8, 0x39, 0x12, 0x90, 0x05, 0x79, 0x47, - 0xf4, 0x8f, 0x25, 0x5e, 0xf3, 0xbf, 0x91, 0xce, 0x8b, 0x28, 0xfa, 0x6b, 0xf1, 0xbe, 0x31, 0x9c, - 0x1c, 0x65, 0xcb, 0x28, 0x65, 0xa3, 0x47, 0x50, 0x32, 0x5c, 0xd7, 0x0b, 0x0c, 0x31, 0xeb, 0x4f, - 0x72, 0x55, 0xab, 0x23, 0xab, 0x5a, 0x8d, 0x64, 0xf4, 0xf5, 0xa9, 0x31, 0x0c, 0x8e, 0xab, 0x42, - 0xc7, 0x30, 0xed, 0x1d, 0xbb, 0x84, 0x62, 0xb2, 0x4f, 0x28, 0x71, 0x4d, 0xe2, 0x57, 0xca, 0x5c, - 0xfb, 0x57, 0x53, 0x6a, 0x4f, 0x30, 0x47, 0x21, 0x9d, 0x84, 0xfb, 0xb8, 0x5f, 0x0b, 0xaa, 0xb1, - 0x24, 0xe9, 0x1a, 0x8e, 0xfd, 0x23, 0x42, 0xfd, 0xca, 0x54, 0xb4, 0xc4, 0xdb, 0x08, 0xa1, 0x38, - 0x46, 0x81, 0xbe, 0x06, 0x25, 0xd3, 0xe9, 0xf8, 0x01, 0x11, 0x1b, 0xd5, 0x69, 0xfe, 0x82, 0x42, - 0xff, 0xd6, 0x22, 0x14, 0x8e, 0xd3, 0xa1, 0x0e, 0x94, 0x5b, 0xf1, 0x92, 0x51, 0x99, 0xe5, 0xde, - 0x5d, 0x4b, 0xe7, 0xdd, 0x60, 0x51, 0x8b, 0xfa, 0x8a, 0x04, 0x0e, 0x27, 0xb5, 0xcc, 0x7f, 0x1d, - 0x4a, 0xff, 0x63, 0xcb, 0xcd, 0x5a, 0xf6, 0xfe, 0x7b, 0x1c, 0xa9, 0x65, 0xff, 0x73, 0x06, 0xa6, - 0x92, 0xa7, 0xdf, 0x57, 0x0e, 0x73, 0xa9, 0xca, 0xa1, 0x1a, 0x0e, 0xb5, 0xa1, 0x4b, 0x60, 0x95, - 0xd6, 0xb3, 0x43, 0xd3, 0xba, 0xcc, 0x9e, 0xe3, 0x2f, 0x92, 0x3d, 0x6b, 0x00, 0xac, 0xcf, 0xa0, - 0x9e, 0xe3, 0x10, 0xca, 0x13, 0x67, 0x41, 0x2e, 0x7b, 0x43, 0x28, 0x8e, 0x51, 0xb0, 0x1e, 0x75, - 0xcf, 0xf1, 0xcc, 0x43, 0x7e, 0x04, 0xea, 0xd1, 0xf3, 0x94, 0x59, 0x10, 0x3d, 0x6a, 0x7d, 0x00, - 0x8b, 0xcf, 0xe0, 0xd0, 0xbb, 0x70, 0x71, 0xc7, 0xa0, 0x81, 0x6d, 0x38, 0xd1, 0x03, 0xe3, 0x43, - 0xc0, 0xc3, 0x81, 0x11, 0xe3, 0xb5, 0x51, 0x1f, 0x6a, 0x74, 0xf8, 0x11, 0x2c, 0x1a, 0x33, 0xf4, - 0xbf, 0x6b, 0x70, 0xe9, 0x4c, 0xdd, 0x9f, 0xc0, 0x88, 0xf3, 0x30, 0x39, 0xe2, 0xbc, 0x91, 0x72, - 0xdf, 0x78, 0x96, 0xb5, 0x43, 0x06, 0x9e, 0x09, 0xc8, 0xed, 0xb0, 0x86, 0x58, 0xff, 0xb5, 0x06, - 0x93, 0xfc, 0xd7, 0x28, 0xbb, 0xda, 0x2a, 0xe4, 0xf6, 0x3d, 0xb5, 0x38, 0x2a, 0x88, 0x3f, 0x13, - 0x36, 0x18, 0x00, 0x0b, 0xf8, 0x0b, 0x2c, 0x73, 0xdf, 0xd3, 0x20, 0xb9, 0x25, 0x45, 0x37, 0x44, - 0xfc, 0x6a, 0xe1, 0x1a, 0x73, 0xc4, 0xd8, 0x7d, 0x73, 0xd8, 0x80, 0x36, 0x97, 0x6a, 0x77, 0x77, - 0x05, 0x8a, 0xd8, 0xf3, 0x82, 0x1d, 0x23, 0x38, 0xf0, 0x99, 0xe3, 0x6d, 0xf6, 0x43, 0x9e, 0x0d, - 0x77, 0x9c, 0x63, 0xb0, 0x80, 0xeb, 0xbf, 0xd2, 0xe0, 0xd2, 0xd0, 0xfd, 0x39, 0x4b, 0x01, 0x66, - 0xf8, 0x25, 0x3d, 0x0a, 0xa3, 0x30, 0xa2, 0xc3, 0x31, 0x2a, 0x36, 0x59, 0x25, 0x96, 0xee, 0xfd, - 0x93, 0x55, 0x42, 0x1b, 0x4e, 0xd2, 0xea, 0xff, 0xce, 0x40, 0x7e, 0x37, 0x30, 0x82, 0x8e, 0xff, - 0x7f, 0x8e, 0xd8, 0x57, 0x20, 0xef, 0x73, 0x3d, 0xd2, 0xbc, 0xb0, 0xc6, 0x0a, 0xed, 0x58, 0x62, - 0xf9, 0x34, 0x42, 0x7c, 0xdf, 0x68, 0xaa, 0x8c, 0x15, 0x4d, 0x23, 0x02, 0x8c, 0x15, 0x1e, 0xbd, - 0x0e, 0x79, 0x4a, 0x0c, 0x3f, 0x1c, 0xcc, 0x16, 0x94, 0x48, 0xcc, 0xa1, 0xa7, 0xbd, 0xea, 0xa4, - 0x14, 0xce, 0xbf, 0xb1, 0xa4, 0x46, 0xf7, 0x61, 0xc2, 0x22, 0x81, 0x61, 0x3b, 0x62, 0x1e, 0x4b, - 0xbd, 0xae, 0x17, 0xc2, 0x1a, 0x82, 0xb5, 0x5e, 0x62, 0x36, 0xc9, 0x0f, 0xac, 0x04, 0xb2, 0x6c, - 0x6b, 0x7a, 0x96, 0x18, 0x27, 0x72, 0x51, 0xb6, 0x5d, 0xf3, 0x2c, 0x82, 0x39, 0x46, 0x7f, 0xac, - 0x41, 0x49, 0x48, 0x5a, 0x33, 0x3a, 0x3e, 0x41, 0x2b, 0xa1, 0x17, 0xe2, 0xba, 0x55, 0x27, 0x37, - 0xce, 0x06, 0x8e, 0xd3, 0x5e, 0xb5, 0xc8, 0xc9, 0xf8, 0x24, 0xa2, 0x1c, 0x88, 0x9d, 0x51, 0xe6, - 0x9c, 0x33, 0x7a, 0x19, 0x72, 0xfc, 0xf5, 0xc8, 0xc3, 0x0c, 0xdf, 0x3a, 0x7f, 0x60, 0x58, 0xe0, - 0xf4, 0x8f, 0x32, 0x50, 0x4e, 0x38, 0x97, 0x62, 0x16, 0x08, 0x17, 0x8a, 0x99, 0x14, 0x4b, 0xea, - 0xe1, 0x7f, 0x51, 0xca, 0xda, 0x93, 0x7f, 0x91, 0xda, 0xf3, 0x3d, 0xc8, 0x9b, 0xec, 0x8c, 0xd4, - 0x3f, 0xde, 0x2b, 0xa3, 0x5c, 0x27, 0x3f, 0xdd, 0x28, 0x1a, 0xf9, 0xa7, 0x8f, 0xa5, 0x40, 0x74, - 0x13, 0x66, 0x29, 0x09, 0x68, 0x77, 0x75, 0x3f, 0x20, 0x34, 0x3e, 0xc4, 0xe7, 0xa2, 0x8e, 0x1b, - 0xf7, 0x13, 0xe0, 0x41, 0x1e, 0x7d, 0x0f, 0x26, 0xef, 0x18, 0x7b, 0x4e, 0xf8, 0x07, 0x14, 0x86, - 0xb2, 0xed, 0x9a, 0x4e, 0xc7, 0x22, 0x22, 0x1b, 0xab, 0xec, 0xa5, 0x1e, 0xed, 0x66, 0x1c, 0x79, - 0xda, 0xab, 0xce, 0x25, 0x00, 0xe2, 0x1f, 0x17, 0x9c, 0x14, 0xa1, 0x3b, 0x30, 0xfe, 0x09, 0x4e, - 0x8f, 0xdf, 0x87, 0x62, 0xd4, 0xdf, 0x7f, 0xcc, 0x2a, 0xf5, 0x87, 0x50, 0x60, 0x11, 0xaf, 0xe6, - 0xd2, 0x73, 0x5a, 0x9c, 0x64, 0xe3, 0x94, 0x49, 0xd3, 0x38, 0xe9, 0x2d, 0x28, 0xdf, 0x6d, 0x5b, - 0x2f, 0xf8, 0x17, 0x64, 0x26, 0x75, 0xd5, 0xba, 0x0a, 0xe2, 0xcf, 0x74, 0x56, 0x20, 0x44, 0xe5, - 0x8e, 0x15, 0x88, 0x78, 0xe1, 0x8d, 0xed, 0xca, 0x7f, 0xa6, 0x01, 0xf0, 0xa5, 0xd4, 0xfa, 0x11, - 0x71, 0x03, 0x76, 0x0e, 0x2c, 0xf0, 0xfb, 0xcf, 0x81, 0x67, 0x06, 0x8e, 0x41, 0x77, 0x21, 0xef, - 0x89, 0x68, 0x12, 0x7f, 0x43, 0x8e, 0xb8, 0xf9, 0x0c, 0x1f, 0x81, 0x88, 0x27, 0x2c, 0x85, 0xd5, - 0x97, 0x9e, 0x3e, 0x5b, 0x18, 0xfb, 0xe0, 0xd9, 0xc2, 0xd8, 0x87, 0xcf, 0x16, 0xc6, 0xde, 0x3d, - 0x59, 0xd0, 0x9e, 0x9e, 0x2c, 0x68, 0x1f, 0x9c, 0x2c, 0x68, 0x1f, 0x9e, 0x2c, 0x68, 0x1f, 0x9d, - 0x2c, 0x68, 0x8f, 0xff, 0xb9, 0x30, 0x76, 0x3f, 0x73, 0xb4, 0xf2, 0xdf, 0x00, 0x00, 0x00, 0xff, - 0xff, 0x61, 0xb7, 0xc5, 0x7c, 0xc2, 0x24, 0x00, 0x00, + // 2832 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x1a, 0xcd, 0x6f, 0x23, 0x57, + 0x3d, 0x63, 0xc7, 0x89, 0xfd, 0x73, 0x9c, 0x8f, 0xb7, 0x59, 0xf0, 0x06, 0x11, 0xa7, 0x53, 0xb4, + 0xda, 0x42, 0xeb, 0x34, 0x4b, 0xa9, 0xb6, 0x5b, 0x5a, 0x88, 0xe3, 0x64, 0x1b, 0x9a, 0x34, 0xd1, + 0xcb, 0xee, 0x02, 0xa5, 0x42, 0x9d, 0x78, 0x5e, 0x9c, 0x21, 0xe3, 0x19, 0xf7, 0xbd, 0x71, 0xb2, + 0x86, 0x03, 0x3d, 0x80, 0x00, 0x09, 0xaa, 0x1e, 0x11, 0x07, 0xd4, 0x0a, 0xfe, 0x02, 0x2e, 0xf0, + 0x07, 0x20, 0xd1, 0x63, 0x25, 0x2e, 0x95, 0x40, 0x56, 0x37, 0x1c, 0x38, 0x22, 0xae, 0xb9, 0x80, + 0xde, 0xc7, 0xcc, 0xbc, 0xf1, 0xc7, 0x66, 0xdc, 0x2d, 0x15, 0x37, 0xcf, 0xef, 0xfb, 0xbd, 0xf7, + 0x7b, 0xbf, 0xaf, 0x67, 0xd8, 0x3d, 0xb9, 0xc5, 0xaa, 0x8e, 0xbf, 0x7a, 0xd2, 0x39, 0x24, 0xd4, + 0x23, 0x01, 0x61, 0xab, 0xa7, 0xc4, 0xb3, 0x7d, 0xba, 0xaa, 0x10, 0x56, 0xdb, 0x69, 0x59, 0x8d, + 0x63, 0xc7, 0x23, 0xb4, 0xbb, 0xda, 0x3e, 0x69, 0x72, 0x00, 0x5b, 0x6d, 0x91, 0xc0, 0x5a, 0x3d, + 0x5d, 0x5b, 0x6d, 0x12, 0x8f, 0x50, 0x2b, 0x20, 0x76, 0xb5, 0x4d, 0xfd, 0xc0, 0x47, 0x5f, 0x92, + 0x5c, 0x55, 0x9d, 0xab, 0xda, 0x3e, 0x69, 0x72, 0x00, 0xab, 0x72, 0xae, 0xea, 0xe9, 0xda, 0xd2, + 0x33, 0x4d, 0x27, 0x38, 0xee, 0x1c, 0x56, 0x1b, 0x7e, 0x6b, 0xb5, 0xe9, 0x37, 0xfd, 0x55, 0xc1, + 0x7c, 0xd8, 0x39, 0x12, 0x5f, 0xe2, 0x43, 0xfc, 0x92, 0x42, 0x97, 0x46, 0x9a, 0x42, 0x3b, 0x5e, + 0xe0, 0xb4, 0x48, 0xbf, 0x15, 0x4b, 0xcf, 0x5f, 0xc6, 0xc0, 0x1a, 0xc7, 0xa4, 0x65, 0xf5, 0xf3, + 0x99, 0x7f, 0xc9, 0x42, 0x7e, 0x7d, 0x7f, 0xfb, 0x0e, 0xf5, 0x3b, 0x6d, 0xb4, 0x02, 0x93, 0x9e, + 0xd5, 0x22, 0x65, 0x63, 0xc5, 0xb8, 0x51, 0xa8, 0xcd, 0x7c, 0xd0, 0xab, 0x4c, 0x9c, 0xf7, 0x2a, + 0x93, 0xaf, 0x59, 0x2d, 0x82, 0x05, 0x06, 0xb9, 0x90, 0x3f, 0x25, 0x94, 0x39, 0xbe, 0xc7, 0xca, + 0x99, 0x95, 0xec, 0x8d, 0xe2, 0xcd, 0x97, 0xab, 0x69, 0xd6, 0x5f, 0x15, 0x0a, 0xee, 0x4b, 0xd6, + 0x2d, 0x9f, 0xd6, 0x1d, 0xd6, 0xf0, 0x4f, 0x09, 0xed, 0xd6, 0xe6, 0x95, 0x96, 0xbc, 0x42, 0x32, + 0x1c, 0x69, 0x40, 0x3f, 0x31, 0x60, 0xbe, 0x4d, 0xc9, 0x11, 0xa1, 0x94, 0xd8, 0x0a, 0x5f, 0xce, + 0xae, 0x18, 0x9f, 0x82, 0xda, 0xb2, 0x52, 0x3b, 0xbf, 0xdf, 0x27, 0x1f, 0x0f, 0x68, 0x44, 0xbf, + 0x33, 0x60, 0x89, 0x11, 0x7a, 0x4a, 0xe8, 0xba, 0x6d, 0x53, 0xc2, 0x58, 0xad, 0xbb, 0xe1, 0x3a, + 0xc4, 0x0b, 0x36, 0xb6, 0xeb, 0x98, 0x95, 0x27, 0xc5, 0x3e, 0x7c, 0x23, 0x9d, 0x41, 0x07, 0xa3, + 0xe4, 0xd4, 0x4c, 0x65, 0xd1, 0xd2, 0x48, 0x12, 0x86, 0x1f, 0x61, 0x86, 0x79, 0x04, 0x33, 0xe1, + 0x41, 0xee, 0x38, 0x2c, 0x40, 0xf7, 0x61, 0xaa, 0xc9, 0x3f, 0x58, 0xd9, 0x10, 0x06, 0x56, 0xd3, + 0x19, 0x18, 0xca, 0xa8, 0xcd, 0x2a, 0x7b, 0xa6, 0xc4, 0x27, 0xc3, 0x4a, 0x9a, 0xf9, 0x8b, 0x49, + 0x28, 0xae, 0xef, 0x6f, 0x63, 0xc2, 0xfc, 0x0e, 0x6d, 0x90, 0x14, 0x4e, 0x73, 0x0b, 0x66, 0x98, + 0xe3, 0x35, 0x3b, 0xae, 0x45, 0x39, 0xb4, 0x3c, 0x25, 0x28, 0x17, 0x15, 0xe5, 0xcc, 0x81, 0x86, + 0xc3, 0x09, 0x4a, 0x74, 0x13, 0x80, 0x4b, 0x60, 0x6d, 0xab, 0x41, 0xec, 0x72, 0x66, 0xc5, 0xb8, + 0x91, 0xaf, 0x21, 0xc5, 0x07, 0xaf, 0x45, 0x18, 0xac, 0x51, 0xa1, 0x27, 0x21, 0x27, 0x2c, 0x2d, + 0xe7, 0x85, 0x9a, 0x92, 0x22, 0xcf, 0x89, 0x65, 0x60, 0x89, 0x43, 0x4f, 0xc1, 0xb4, 0xf2, 0xb2, + 0x72, 0x41, 0x90, 0xcd, 0x29, 0xb2, 0xe9, 0xd0, 0x0d, 0x42, 0x3c, 0x5f, 0xdf, 0x89, 0xe3, 0xd9, + 0xc2, 0xef, 0xb4, 0xf5, 0xbd, 0xea, 0x78, 0x36, 0x16, 0x18, 0xb4, 0x03, 0xb9, 0x53, 0x42, 0x0f, + 0xb9, 0x27, 0x70, 0xd7, 0xfc, 0x4a, 0xba, 0x8d, 0xbe, 0xcf, 0x59, 0x6a, 0x05, 0x6e, 0x9a, 0xf8, + 0x89, 0xa5, 0x10, 0x54, 0x05, 0x60, 0xc7, 0x3e, 0x0d, 0xc4, 0xf2, 0xca, 0xb9, 0x95, 0xec, 0x8d, + 0x42, 0x6d, 0x96, 0xaf, 0xf7, 0x20, 0x82, 0x62, 0x8d, 0x82, 0xd3, 0x37, 0xac, 0x80, 0x34, 0x7d, + 0xea, 0x10, 0x56, 0x9e, 0x8e, 0xe9, 0x37, 0x22, 0x28, 0xd6, 0x28, 0xd0, 0xb7, 0x00, 0xb1, 0xc0, + 0xa7, 0x56, 0x93, 0xa8, 0xa5, 0xbe, 0x62, 0xb1, 0xe3, 0x32, 0x88, 0xd5, 0x2d, 0xa9, 0xd5, 0xa1, + 0x83, 0x01, 0x0a, 0x3c, 0x84, 0xcb, 0xfc, 0x83, 0x01, 0x73, 0x9a, 0x2f, 0x08, 0xbf, 0xbb, 0x05, + 0x33, 0x4d, 0xed, 0xd6, 0x29, 0xbf, 0x88, 0x4e, 0x5b, 0xbf, 0x91, 0x38, 0x41, 0x89, 0x08, 0x14, + 0xa8, 0x92, 0x14, 0x46, 0x97, 0xb5, 0xd4, 0x4e, 0x1b, 0xda, 0x10, 0x6b, 0xd2, 0x80, 0x0c, 0xc7, + 0x92, 0xcd, 0x7f, 0x1a, 0xc2, 0x81, 0xc3, 0x78, 0x83, 0x6e, 0x68, 0x31, 0xcd, 0x10, 0xdb, 0x37, + 0x33, 0x22, 0x1e, 0x5d, 0x12, 0x08, 0x32, 0xff, 0x17, 0x81, 0xe0, 0x76, 0xfe, 0xd7, 0xef, 0x55, + 0x26, 0xde, 0xfe, 0xfb, 0xca, 0x84, 0xf9, 0x9f, 0x0c, 0x14, 0x36, 0x7c, 0xcf, 0x76, 0x02, 0xe5, + 0xc8, 0x41, 0xb7, 0x3d, 0x70, 0x51, 0xef, 0x76, 0xdb, 0x04, 0x0b, 0x0c, 0x7a, 0x01, 0xa6, 0x58, + 0x60, 0x05, 0x1d, 0x26, 0xae, 0x5a, 0xa1, 0xf6, 0x44, 0x18, 0x02, 0x0e, 0x04, 0xf4, 0xa2, 0x57, + 0x99, 0x8b, 0xc4, 0x49, 0x10, 0x56, 0x0c, 0xdc, 0xab, 0xfc, 0x43, 0x61, 0x94, 0x7d, 0x47, 0xa6, + 0x98, 0x30, 0x56, 0x67, 0x63, 0xaf, 0xda, 0x1b, 0xa0, 0xc0, 0x43, 0xb8, 0xd0, 0x29, 0x20, 0xd7, + 0x62, 0xc1, 0x5d, 0x6a, 0x79, 0x4c, 0xe8, 0xba, 0xeb, 0xb4, 0x88, 0xba, 0x5c, 0x5f, 0x4e, 0xb7, + 0xbb, 0x9c, 0x23, 0xd6, 0xbb, 0x33, 0x20, 0x0d, 0x0f, 0xd1, 0x80, 0xae, 0xc3, 0x14, 0x25, 0x16, + 0xf3, 0xbd, 0x72, 0x4e, 0x2c, 0x3f, 0x8a, 0x80, 0x58, 0x40, 0xb1, 0xc2, 0xf2, 0xe0, 0xd1, 0x22, + 0x8c, 0x59, 0xcd, 0x30, 0x94, 0x45, 0xc1, 0x63, 0x57, 0x82, 0x71, 0x88, 0x37, 0x5b, 0x50, 0xda, + 0xa0, 0xc4, 0x0a, 0xc8, 0x5e, 0x3b, 0x10, 0x2e, 0x64, 0xc2, 0x94, 0x4d, 0xbb, 0xb8, 0xe3, 0x29, + 0x57, 0x03, 0x2e, 0xbf, 0x2e, 0x20, 0x58, 0x61, 0xf8, 0x0d, 0x3a, 0x72, 0x88, 0x6b, 0xef, 0x5a, + 0x9e, 0xd5, 0x24, 0x54, 0x45, 0x9e, 0xc8, 0xaf, 0xb7, 0x34, 0x1c, 0x4e, 0x50, 0x9a, 0x3f, 0xcb, + 0x42, 0xa9, 0x4e, 0x5c, 0x12, 0xeb, 0xdb, 0x02, 0xd4, 0xa4, 0x56, 0x83, 0xec, 0x13, 0xea, 0xf8, + 0xf6, 0x01, 0x69, 0xf8, 0x9e, 0xcd, 0x84, 0x0b, 0x64, 0x6b, 0x9f, 0xe3, 0x7b, 0x73, 0x67, 0x00, + 0x8b, 0x87, 0x70, 0x20, 0x17, 0x4a, 0x6d, 0x2a, 0x7e, 0x8b, 0xfd, 0x92, 0x1e, 0x52, 0xbc, 0xf9, + 0xd5, 0x74, 0xc7, 0xb1, 0xaf, 0xb3, 0xd6, 0x16, 0xce, 0x7b, 0x95, 0x52, 0x02, 0x84, 0x93, 0xc2, + 0xd1, 0x37, 0x61, 0xde, 0xa7, 0xed, 0x63, 0xcb, 0xab, 0x93, 0x36, 0xf1, 0x6c, 0xe2, 0x05, 0x4c, + 0xec, 0x42, 0xbe, 0xb6, 0xc8, 0x73, 0xf6, 0x5e, 0x1f, 0x0e, 0x0f, 0x50, 0xa3, 0xd7, 0x61, 0xa1, + 0x4d, 0xfd, 0xb6, 0xd5, 0x14, 0x2e, 0xb5, 0xef, 0xbb, 0x4e, 0xa3, 0x2b, 0x5c, 0xa8, 0x50, 0x7b, + 0xfa, 0xbc, 0x57, 0x59, 0xd8, 0xef, 0x47, 0x5e, 0xf4, 0x2a, 0x57, 0xc4, 0xd6, 0x71, 0x48, 0x8c, + 0xc4, 0x83, 0x62, 0xb4, 0x33, 0xcc, 0x8d, 0x3a, 0x43, 0x73, 0x1b, 0xf2, 0xf5, 0x8e, 0xf2, 0xe7, + 0x97, 0x20, 0x6f, 0xab, 0xdf, 0x6a, 0xe7, 0xc3, 0x8b, 0x15, 0xd1, 0x5c, 0xf4, 0x2a, 0x25, 0x5e, + 0xa6, 0x55, 0x43, 0x00, 0x8e, 0x58, 0xcc, 0x37, 0xa0, 0xb4, 0xf9, 0xa0, 0xed, 0xd3, 0x20, 0x3c, + 0xd3, 0xeb, 0x30, 0x45, 0x04, 0x40, 0x48, 0xcb, 0xc7, 0x7e, 0x2a, 0xc9, 0xb0, 0xc2, 0xf2, 0x4c, + 0x48, 0x1e, 0x58, 0x8d, 0x40, 0x25, 0xce, 0x28, 0x13, 0x6e, 0x72, 0x20, 0x96, 0x38, 0xf3, 0x3a, + 0xe4, 0x85, 0x43, 0xb1, 0xfb, 0x6b, 0x68, 0x1e, 0xb2, 0xd8, 0x3a, 0x13, 0x52, 0x67, 0x70, 0x96, + 0x5a, 0x67, 0x5a, 0x2c, 0xd9, 0x03, 0xb8, 0x43, 0x22, 0x13, 0xd6, 0x61, 0x2e, 0x0c, 0xa8, 0xc9, + 0x38, 0xff, 0x79, 0xa5, 0x64, 0x0e, 0x27, 0xd1, 0xb8, 0x9f, 0xde, 0x7c, 0x03, 0x0a, 0x22, 0x17, + 0xf0, 0x44, 0x1a, 0x27, 0x6d, 0xe3, 0x11, 0x49, 0x3b, 0xcc, 0xc4, 0x99, 0x51, 0x99, 0x58, 0x33, + 0xd7, 0x85, 0x92, 0xe4, 0x0d, 0xcb, 0x94, 0x54, 0x1a, 0x9e, 0x86, 0x7c, 0x68, 0xa6, 0xd2, 0x12, + 0x95, 0xa7, 0xa1, 0x20, 0x1c, 0x51, 0x68, 0xda, 0x8e, 0x21, 0x91, 0xd7, 0xd2, 0x29, 0xd3, 0x6a, + 0x90, 0xcc, 0xa3, 0x6b, 0x10, 0x4d, 0xd3, 0x8f, 0xa1, 0x3c, 0xaa, 0xa6, 0x7d, 0x8c, 0xcc, 0x9b, + 0xde, 0x14, 0xf3, 0x1d, 0x03, 0xe6, 0x75, 0x49, 0xe9, 0x8f, 0x2f, 0xbd, 0x92, 0xcb, 0x6b, 0x2e, + 0x6d, 0x47, 0x7e, 0x6b, 0xc0, 0x62, 0x62, 0x69, 0x63, 0x9d, 0xf8, 0x18, 0x46, 0xe9, 0xce, 0x91, + 0x1d, 0xc3, 0x39, 0xfe, 0x9a, 0x81, 0xd2, 0x8e, 0x75, 0x48, 0xdc, 0x03, 0xe2, 0x92, 0x46, 0xe0, + 0x53, 0xf4, 0x23, 0x28, 0xb6, 0xac, 0xa0, 0x71, 0x2c, 0xa0, 0x61, 0x7d, 0x5e, 0x4f, 0x17, 0x4a, + 0x13, 0x92, 0xaa, 0xbb, 0xb1, 0x98, 0x4d, 0x2f, 0xa0, 0xdd, 0xda, 0x15, 0x65, 0x52, 0x51, 0xc3, + 0x60, 0x5d, 0x9b, 0x68, 0xaa, 0xc4, 0xf7, 0xe6, 0x83, 0x36, 0x2f, 0x1e, 0xc6, 0xef, 0xe5, 0x12, + 0x26, 0x60, 0xf2, 0x56, 0xc7, 0xa1, 0xa4, 0x45, 0xbc, 0x20, 0x6e, 0xaa, 0x76, 0xfb, 0xe4, 0xe3, + 0x01, 0x8d, 0x4b, 0x2f, 0xc3, 0x7c, 0xbf, 0xf1, 0x3c, 0xfe, 0x9c, 0x90, 0xae, 0x3c, 0x2f, 0xcc, + 0x7f, 0xa2, 0x45, 0xc8, 0x9d, 0x5a, 0x6e, 0x47, 0xdd, 0x46, 0x2c, 0x3f, 0x6e, 0x67, 0x6e, 0x19, + 0xe6, 0xef, 0x0d, 0x28, 0x8f, 0x32, 0x04, 0x7d, 0x51, 0x13, 0x54, 0x2b, 0x2a, 0xab, 0xb2, 0xaf, + 0x92, 0xae, 0x94, 0xba, 0x09, 0x79, 0xbf, 0xcd, 0xab, 0x0d, 0x9f, 0xaa, 0x53, 0x7f, 0x2a, 0x3c, + 0xc9, 0x3d, 0x05, 0xbf, 0xe8, 0x55, 0xae, 0x26, 0xc4, 0x87, 0x08, 0x1c, 0xb1, 0xf2, 0x3c, 0x20, + 0xec, 0xe1, 0xb9, 0x29, 0xca, 0x03, 0xf7, 0x05, 0x04, 0x2b, 0x8c, 0xf9, 0x27, 0x03, 0x26, 0x45, + 0x59, 0xfc, 0x06, 0xe4, 0xf9, 0xfe, 0xd9, 0x56, 0x60, 0x09, 0xbb, 0x52, 0x37, 0x64, 0x9c, 0x7b, + 0x97, 0x04, 0x56, 0xec, 0x6d, 0x21, 0x04, 0x47, 0x12, 0x11, 0x86, 0x9c, 0x13, 0x90, 0x56, 0x78, + 0x90, 0xcf, 0x8c, 0x14, 0xad, 0xc6, 0x01, 0x55, 0x6c, 0x9d, 0x6d, 0x3e, 0x08, 0x88, 0xc7, 0x0f, + 0x23, 0xbe, 0x1a, 0xdb, 0x5c, 0x06, 0x96, 0xa2, 0xcc, 0x7f, 0x1b, 0x10, 0xa9, 0xe2, 0xce, 0xcf, + 0x88, 0x7b, 0xb4, 0xe3, 0x78, 0x27, 0x6a, 0x5b, 0x23, 0x73, 0x0e, 0x14, 0x1c, 0x47, 0x14, 0xc3, + 0xd2, 0x43, 0x66, 0xbc, 0xf4, 0xc0, 0x15, 0x36, 0x7c, 0x2f, 0x70, 0xbc, 0xce, 0xc0, 0x6d, 0xdb, + 0x50, 0x70, 0x1c, 0x51, 0xf0, 0x32, 0x87, 0x92, 0x96, 0xe5, 0x78, 0x8e, 0xd7, 0xe4, 0x8b, 0xd8, + 0xf0, 0x3b, 0x5e, 0x20, 0xf2, 0xbd, 0x2a, 0x73, 0xf0, 0x00, 0x16, 0x0f, 0xe1, 0x30, 0xff, 0x38, + 0x09, 0x45, 0xbe, 0xe6, 0x30, 0xcf, 0xbd, 0x08, 0x25, 0x57, 0xf7, 0x02, 0xb5, 0xf6, 0xab, 0xca, + 0x94, 0xe4, 0xbd, 0xc6, 0x49, 0x5a, 0xce, 0x2c, 0xaa, 0xb3, 0x88, 0x39, 0x93, 0x64, 0xde, 0xd2, + 0x91, 0x38, 0x49, 0xcb, 0xa3, 0xd7, 0x19, 0xbf, 0x1f, 0xaa, 0xee, 0x89, 0x8e, 0xe8, 0xdb, 0x1c, + 0x88, 0x25, 0x0e, 0xed, 0xc2, 0x15, 0xcb, 0x75, 0xfd, 0x33, 0x01, 0xac, 0xf9, 0xfe, 0x49, 0xcb, + 0xa2, 0x27, 0x4c, 0xb4, 0xb4, 0xf9, 0xda, 0x17, 0x14, 0xcb, 0x95, 0xf5, 0x41, 0x12, 0x3c, 0x8c, + 0x6f, 0xd8, 0xb1, 0x4d, 0x8e, 0x79, 0x6c, 0xc7, 0xb0, 0xd8, 0x07, 0x12, 0xb7, 0x5c, 0xf5, 0x97, + 0xcf, 0x29, 0x39, 0x8b, 0x78, 0x08, 0xcd, 0xc5, 0x08, 0x38, 0x1e, 0x2a, 0x11, 0xdd, 0x86, 0x59, + 0xee, 0xc9, 0x7e, 0x27, 0x08, 0xab, 0xda, 0x9c, 0x38, 0x6e, 0x74, 0xde, 0xab, 0xcc, 0xde, 0x4d, + 0x60, 0x70, 0x1f, 0x25, 0xdf, 0x5c, 0xd7, 0x69, 0x39, 0x41, 0x79, 0x5a, 0xb0, 0x44, 0x9b, 0xbb, + 0xc3, 0x81, 0x58, 0xe2, 0x12, 0x1e, 0x98, 0xbf, 0xcc, 0x03, 0xcd, 0xdf, 0x64, 0x01, 0xc9, 0x32, + 0xdc, 0x96, 0xf5, 0x94, 0x0c, 0x69, 0xbc, 0x57, 0x50, 0x65, 0xbc, 0xd1, 0xd7, 0x2b, 0xa8, 0x0a, + 0x3e, 0xc4, 0xa3, 0x5d, 0x28, 0xc8, 0xd0, 0x12, 0x5f, 0x97, 0x55, 0x45, 0x5c, 0xd8, 0x0b, 0x11, + 0x17, 0xbd, 0xca, 0x52, 0x42, 0x4d, 0x84, 0x11, 0x7d, 0x5c, 0x2c, 0x01, 0xdd, 0x04, 0xb0, 0xda, + 0x8e, 0x3e, 0x35, 0x2b, 0xc4, 0xb3, 0x93, 0xb8, 0xff, 0xc5, 0x1a, 0x15, 0x7a, 0x05, 0x26, 0x83, + 0x4f, 0xd6, 0x6b, 0xe5, 0x45, 0x2b, 0xc9, 0x3b, 0x2b, 0x21, 0x81, 0x6b, 0x17, 0xfe, 0xcc, 0xb8, + 0x59, 0xaa, 0x4d, 0x8a, 0xb4, 0x6f, 0x45, 0x18, 0xac, 0x51, 0xa1, 0xef, 0x40, 0xfe, 0x48, 0x95, + 0xa2, 0xe2, 0x60, 0x52, 0x87, 0xc8, 0xb0, 0x80, 0x95, 0x8d, 0x7b, 0xf8, 0x85, 0x23, 0x69, 0xe6, + 0x5b, 0x50, 0xd8, 0x75, 0x1a, 0xd4, 0x17, 0x6d, 0xde, 0x53, 0x30, 0xcd, 0x12, 0x7d, 0x50, 0x74, + 0x24, 0xa1, 0xbb, 0x84, 0x78, 0xee, 0x27, 0x9e, 0xe5, 0xf9, 0xb2, 0xdb, 0xc9, 0xc5, 0x7e, 0xf2, + 0x1a, 0x07, 0x62, 0x89, 0xbb, 0xbd, 0xc8, 0x33, 0xfd, 0xcf, 0xdf, 0xaf, 0x4c, 0xbc, 0xfb, 0x7e, + 0x65, 0xe2, 0xbd, 0xf7, 0x55, 0xd6, 0xbf, 0x00, 0x80, 0xbd, 0xc3, 0x1f, 0x90, 0x86, 0x8c, 0x9f, + 0xa9, 0xa6, 0x64, 0xe1, 0x70, 0x56, 0x4c, 0xc9, 0x32, 0x7d, 0xd5, 0x9b, 0x86, 0xc3, 0x09, 0x4a, + 0xb4, 0x0a, 0x85, 0x68, 0xfe, 0xa5, 0x0e, 0x7a, 0x21, 0x74, 0x9c, 0x68, 0x48, 0x86, 0x63, 0x9a, + 0x44, 0x30, 0x9f, 0xbc, 0x34, 0x98, 0xd7, 0x20, 0xdb, 0x71, 0x6c, 0xd5, 0x13, 0x3f, 0x1b, 0x26, + 0xd3, 0x7b, 0xdb, 0xf5, 0x8b, 0x5e, 0xe5, 0x89, 0x51, 0x63, 0xe7, 0xa0, 0xdb, 0x26, 0xac, 0x7a, + 0x6f, 0xbb, 0x8e, 0x39, 0xf3, 0xb0, 0xc8, 0x32, 0x35, 0x66, 0x64, 0xb9, 0x09, 0xd0, 0x8c, 0x27, + 0x0b, 0xf2, 0xe2, 0x46, 0x1e, 0xa5, 0x4d, 0x14, 0x34, 0x2a, 0xc4, 0x60, 0xa1, 0xc1, 0xdb, 0x6f, + 0xd5, 0xe1, 0xb3, 0xc0, 0x6a, 0xc9, 0xb9, 0xe0, 0x78, 0xce, 0x7d, 0x4d, 0xa9, 0x59, 0xd8, 0xe8, + 0x17, 0x86, 0x07, 0xe5, 0x23, 0x1f, 0x16, 0x6c, 0xd5, 0x48, 0xc6, 0x4a, 0x0b, 0x63, 0x2b, 0xbd, + 0xca, 0x15, 0xd6, 0xfb, 0x05, 0xe1, 0x41, 0xd9, 0xe8, 0xfb, 0xb0, 0x14, 0x02, 0x07, 0xbb, 0x79, + 0x11, 0x79, 0xb3, 0xb5, 0xe5, 0xf3, 0x5e, 0x65, 0xa9, 0x3e, 0x92, 0x0a, 0x3f, 0x42, 0x02, 0xb2, + 0x61, 0xca, 0x95, 0x95, 0x6a, 0x51, 0x54, 0x17, 0x5f, 0x4f, 0xb7, 0x8a, 0xd8, 0xfb, 0xab, 0x7a, + 0x85, 0x1a, 0x75, 0xab, 0xaa, 0x38, 0x55, 0xb2, 0xd1, 0x03, 0x28, 0x5a, 0x9e, 0xe7, 0x07, 0x96, + 0x9c, 0x2f, 0xcc, 0x08, 0x55, 0xeb, 0x63, 0xab, 0x5a, 0x8f, 0x65, 0xf4, 0x55, 0xc4, 0x1a, 0x06, + 0xeb, 0xaa, 0xd0, 0x19, 0xcc, 0xf9, 0x67, 0x1e, 0xa1, 0x98, 0x1c, 0x11, 0x4a, 0xbc, 0x06, 0x61, + 0xe5, 0x92, 0xd0, 0xfe, 0x5c, 0x4a, 0xed, 0x09, 0xe6, 0xd8, 0xa5, 0x93, 0x70, 0x86, 0xfb, 0xb5, + 0xa0, 0x2a, 0x0f, 0x92, 0x9e, 0xe5, 0x3a, 0x3f, 0x24, 0x94, 0x95, 0x67, 0xe3, 0xd1, 0xed, 0x56, + 0x04, 0xc5, 0x1a, 0x05, 0xfa, 0x1a, 0x14, 0x1b, 0x6e, 0x87, 0x05, 0x44, 0xce, 0xd1, 0xe7, 0xc4, + 0x0d, 0x8a, 0xd6, 0xb7, 0x11, 0xa3, 0xb0, 0x4e, 0x87, 0x3a, 0x50, 0x6a, 0xe9, 0x29, 0xa3, 0xbc, + 0x20, 0x56, 0x77, 0x2b, 0xdd, 0xea, 0x06, 0x93, 0x5a, 0x5c, 0xc1, 0x24, 0x70, 0x38, 0xa9, 0x65, + 0xe9, 0x05, 0x28, 0x7e, 0xc2, 0xe2, 0x9e, 0x37, 0x07, 0xfd, 0xe7, 0x38, 0x56, 0x73, 0xf0, 0xe7, + 0x0c, 0xcc, 0x26, 0x77, 0xbf, 0x2f, 0x1d, 0xe6, 0x52, 0xa5, 0xc3, 0xb0, 0x0d, 0x35, 0x46, 0x8e, + 0xfe, 0xc3, 0xb0, 0x9e, 0x1d, 0x19, 0xd6, 0x55, 0xf4, 0x9c, 0x7c, 0x9c, 0xe8, 0x59, 0x05, 0xe0, + 0x75, 0x06, 0xf5, 0x5d, 0x97, 0x50, 0x11, 0x38, 0xf3, 0x6a, 0xc4, 0x1f, 0x41, 0xb1, 0x46, 0xc1, + 0xab, 0xe1, 0x43, 0xd7, 0x6f, 0x9c, 0x88, 0x2d, 0x08, 0x2f, 0xbd, 0x08, 0x99, 0x79, 0x59, 0x0d, + 0xd7, 0x06, 0xb0, 0x78, 0x08, 0x87, 0xd9, 0x85, 0xab, 0xfb, 0x16, 0x0d, 0x1c, 0xcb, 0x8d, 0x2f, + 0x98, 0x68, 0x37, 0xde, 0x1c, 0x68, 0x66, 0x9e, 0x1d, 0xf7, 0xa2, 0xc6, 0x9b, 0x1f, 0xc3, 0xe2, + 0x86, 0xc6, 0xfc, 0x9b, 0x01, 0xd7, 0x86, 0xea, 0xfe, 0x0c, 0x9a, 0xa9, 0x37, 0x93, 0xcd, 0xd4, + 0x8b, 0x29, 0x67, 0x9c, 0xc3, 0xac, 0x1d, 0xd1, 0x5a, 0x4d, 0x43, 0x6e, 0x9f, 0x17, 0xb1, 0xe6, + 0xaf, 0x0c, 0x98, 0x11, 0xbf, 0xc6, 0x99, 0x0f, 0x57, 0x20, 0x77, 0xe4, 0x87, 0x23, 0xaa, 0xbc, + 0x7c, 0x42, 0xda, 0xe2, 0x00, 0x2c, 0xe1, 0x8f, 0x31, 0x40, 0x7e, 0xc7, 0x80, 0xe4, 0x64, 0x16, + 0xbd, 0x2c, 0xfd, 0xd7, 0x88, 0x46, 0xa7, 0x63, 0xfa, 0xee, 0x4b, 0xa3, 0x5a, 0xc1, 0x2b, 0xa9, + 0xa6, 0x84, 0x4f, 0x43, 0x01, 0xfb, 0x7e, 0xb0, 0x6f, 0x05, 0xc7, 0x8c, 0x2f, 0xbc, 0xcd, 0x7f, + 0xa8, 0xbd, 0x11, 0x0b, 0x17, 0x18, 0x2c, 0xe1, 0xe6, 0x2f, 0x0d, 0xb8, 0x36, 0xf2, 0xd5, 0x84, + 0x87, 0x80, 0x46, 0xf4, 0xa5, 0x56, 0x14, 0x79, 0x61, 0x4c, 0x87, 0x35, 0x2a, 0xde, 0xc3, 0x25, + 0x9e, 0x5a, 0xfa, 0x7b, 0xb8, 0x84, 0x36, 0x9c, 0xa4, 0x35, 0xff, 0x95, 0x01, 0xf5, 0x74, 0xf2, + 0x3f, 0xf6, 0xd8, 0xeb, 0x7d, 0x0f, 0x37, 0xb3, 0xc9, 0x87, 0x9b, 0xe8, 0x95, 0x46, 0x7b, 0xb9, + 0xc8, 0x3e, 0xfa, 0xe5, 0x02, 0x3d, 0x1f, 0x3d, 0x86, 0xc8, 0xd0, 0xb5, 0x9c, 0x7c, 0x0c, 0xb9, + 0xe8, 0x55, 0x66, 0x94, 0xf0, 0xe4, 0xe3, 0xc8, 0xeb, 0x30, 0x6d, 0x93, 0xc0, 0x72, 0x5c, 0xd9, + 0x8f, 0xa5, 0x7e, 0x22, 0x90, 0xc2, 0xea, 0x92, 0xb5, 0x56, 0xe4, 0x36, 0xa9, 0x0f, 0x1c, 0x0a, + 0xe4, 0xd1, 0xb6, 0xe1, 0xdb, 0xb2, 0x9d, 0xc8, 0xc5, 0xd1, 0x76, 0xc3, 0xb7, 0x09, 0x16, 0x18, + 0xf3, 0x5d, 0x03, 0x8a, 0x52, 0xd2, 0x86, 0xd5, 0x61, 0x04, 0xad, 0x45, 0xab, 0x90, 0xc7, 0x7d, + 0x4d, 0x7f, 0xf5, 0xba, 0xe8, 0x55, 0x0a, 0x82, 0x4c, 0x74, 0x22, 0x43, 0x5e, 0x77, 0x32, 0x97, + 0xec, 0xd1, 0x93, 0x90, 0x13, 0xb7, 0x47, 0x6d, 0x66, 0x74, 0xd7, 0xc5, 0x05, 0xc3, 0x12, 0x67, + 0x7e, 0x9c, 0x81, 0x52, 0x62, 0x71, 0x29, 0x7a, 0x81, 0x68, 0x74, 0x99, 0x49, 0x31, 0x0e, 0x1f, + 0xfd, 0x30, 0xad, 0x72, 0xcf, 0xd4, 0xe3, 0xe4, 0x9e, 0xef, 0xc2, 0x54, 0x83, 0xef, 0x51, 0xf8, + 0x3f, 0x87, 0xb5, 0x71, 0x8e, 0x53, 0xec, 0x6e, 0xec, 0x8d, 0xe2, 0x93, 0x61, 0x25, 0x10, 0xdd, + 0x81, 0x05, 0x4a, 0x02, 0xda, 0x5d, 0x3f, 0x0a, 0x08, 0xd5, 0x9b, 0xf8, 0x5c, 0x5c, 0x71, 0xe3, + 0x7e, 0x02, 0x3c, 0xc8, 0x63, 0x1e, 0xc2, 0xcc, 0x5d, 0xeb, 0xd0, 0x8d, 0x1e, 0xbd, 0x30, 0x94, + 0x1c, 0xaf, 0xe1, 0x76, 0x6c, 0x22, 0xa3, 0x71, 0x18, 0xbd, 0xc2, 0x4b, 0xbb, 0xad, 0x23, 0x2f, + 0x7a, 0x95, 0x2b, 0x09, 0x80, 0x7c, 0xe5, 0xc1, 0x49, 0x11, 0xa6, 0x0b, 0x93, 0x9f, 0x61, 0xf7, + 0xf8, 0x3d, 0x28, 0xc4, 0xf5, 0xfd, 0xa7, 0xac, 0xd2, 0x7c, 0x13, 0xf2, 0xdc, 0xe3, 0xc3, 0xbe, + 0xf4, 0x92, 0x12, 0x27, 0x59, 0x38, 0x65, 0xd2, 0x14, 0x4e, 0x66, 0x0b, 0x4a, 0xf7, 0xda, 0xf6, + 0x63, 0x3e, 0x7b, 0x66, 0x52, 0x67, 0xad, 0x9b, 0x20, 0xff, 0x42, 0xc1, 0x13, 0x84, 0xcc, 0xdc, + 0x5a, 0x82, 0xd0, 0x13, 0xaf, 0x36, 0x95, 0xff, 0xa9, 0x01, 0x20, 0xc6, 0x5f, 0x9b, 0xa7, 0xc4, + 0x0b, 0x52, 0x3c, 0x8e, 0xdf, 0x83, 0x29, 0x5f, 0x7a, 0x93, 0x7c, 0xfa, 0x1c, 0x73, 0xc6, 0x1a, + 0x5d, 0x02, 0xe9, 0x4f, 0x58, 0x09, 0xab, 0xdd, 0xf8, 0xe0, 0xe1, 0xf2, 0xc4, 0x87, 0x0f, 0x97, + 0x27, 0x3e, 0x7a, 0xb8, 0x3c, 0xf1, 0xf6, 0xf9, 0xb2, 0xf1, 0xc1, 0xf9, 0xb2, 0xf1, 0xe1, 0xf9, + 0xb2, 0xf1, 0xd1, 0xf9, 0xb2, 0xf1, 0xf1, 0xf9, 0xb2, 0xf1, 0xee, 0x3f, 0x96, 0x27, 0x5e, 0xcf, + 0x9c, 0xae, 0xfd, 0x37, 0x00, 0x00, 0xff, 0xff, 0x82, 0x62, 0x88, 0xff, 0xb8, 0x26, 0x00, 0x00, } func (m *APIGroup) Marshal() (dAtA []byte, err error) { @@ -1752,6 +1788,62 @@ func (m *APIVersions) MarshalToSizedBuffer(dAtA []byte) (int, error) { return len(dAtA) - i, nil } +func (m *Condition) Marshal() (dAtA []byte, err error) { + size := m.Size() + dAtA = make([]byte, size) + n, err := m.MarshalToSizedBuffer(dAtA[:size]) + if err != nil { + return nil, err + } + return dAtA[:n], nil +} + +func (m *Condition) MarshalTo(dAtA []byte) (int, error) { + size := m.Size() + return m.MarshalToSizedBuffer(dAtA[:size]) +} + +func (m *Condition) MarshalToSizedBuffer(dAtA []byte) (int, error) { + i := len(dAtA) + _ = i + var l int + _ = l + i -= len(m.Message) + copy(dAtA[i:], m.Message) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Message))) + i-- + dAtA[i] = 0x32 + i -= len(m.Reason) + copy(dAtA[i:], m.Reason) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Reason))) + i-- + dAtA[i] = 0x2a + { + size, err := m.LastTransitionTime.MarshalToSizedBuffer(dAtA[:i]) + if err != nil { + return 0, err + } + i -= size + i = encodeVarintGenerated(dAtA, i, uint64(size)) + } + i-- + dAtA[i] = 0x22 + i = encodeVarintGenerated(dAtA, i, uint64(m.ObservedGeneration)) + i-- + dAtA[i] = 0x18 + i -= len(m.Status) + copy(dAtA[i:], m.Status) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Status))) + i-- + dAtA[i] = 0x12 + i -= len(m.Type) + copy(dAtA[i:], m.Type) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.Type))) + i-- + dAtA[i] = 0xa + return len(dAtA) - i, nil +} + func (m *CreateOptions) Marshal() (dAtA []byte, err error) { size := m.Size() dAtA = make([]byte, size) @@ -2399,6 +2491,11 @@ func (m *ListOptions) MarshalToSizedBuffer(dAtA []byte) (int, error) { _ = i var l int _ = l + i -= len(m.ResourceVersionMatch) + copy(dAtA[i:], m.ResourceVersionMatch) + i = encodeVarintGenerated(dAtA, i, uint64(len(m.ResourceVersionMatch))) + i-- + dAtA[i] = 0x52 i-- if m.AllowWatchBookmarks { dAtA[i] = 1 @@ -3488,6 +3585,26 @@ func (m *APIVersions) Size() (n int) { return n } +func (m *Condition) Size() (n int) { + if m == nil { + return 0 + } + var l int + _ = l + l = len(m.Type) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Status) + n += 1 + l + sovGenerated(uint64(l)) + n += 1 + sovGenerated(uint64(m.ObservedGeneration)) + l = m.LastTransitionTime.Size() + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Reason) + n += 1 + l + sovGenerated(uint64(l)) + l = len(m.Message) + n += 1 + l + sovGenerated(uint64(l)) + return n +} + func (m *CreateOptions) Size() (n int) { if m == nil { return 0 @@ -3758,6 +3875,8 @@ func (m *ListOptions) Size() (n int) { l = len(m.Continue) n += 1 + l + sovGenerated(uint64(l)) n += 2 + l = len(m.ResourceVersionMatch) + n += 1 + l + sovGenerated(uint64(l)) return n } @@ -4198,6 +4317,21 @@ func (this *APIResourceList) String() string { }, "") return s } +func (this *Condition) String() string { + if this == nil { + return "nil" + } + s := strings.Join([]string{`&Condition{`, + `Type:` + fmt.Sprintf("%v", this.Type) + `,`, + `Status:` + fmt.Sprintf("%v", this.Status) + `,`, + `ObservedGeneration:` + fmt.Sprintf("%v", this.ObservedGeneration) + `,`, + `LastTransitionTime:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.LastTransitionTime), "Time", "Time", 1), `&`, ``, 1) + `,`, + `Reason:` + fmt.Sprintf("%v", this.Reason) + `,`, + `Message:` + fmt.Sprintf("%v", this.Message) + `,`, + `}`, + }, "") + return s +} func (this *CreateOptions) String() string { if this == nil { return "nil" @@ -4244,16 +4378,6 @@ func (this *ExportOptions) String() string { }, "") return s } -func (this *FieldsV1) String() string { - if this == nil { - return "nil" - } - s := strings.Join([]string{`&FieldsV1{`, - `Raw:` + valueToStringGenerated(this.Raw) + `,`, - `}`, - }, "") - return s -} func (this *GetOptions) String() string { if this == nil { return "nil" @@ -4355,6 +4479,7 @@ func (this *ListOptions) String() string { `Limit:` + fmt.Sprintf("%v", this.Limit) + `,`, `Continue:` + fmt.Sprintf("%v", this.Continue) + `,`, `AllowWatchBookmarks:` + fmt.Sprintf("%v", this.AllowWatchBookmarks) + `,`, + `ResourceVersionMatch:` + fmt.Sprintf("%v", this.ResourceVersionMatch) + `,`, `}`, }, "") return s @@ -4369,7 +4494,7 @@ func (this *ManagedFieldsEntry) String() string { `APIVersion:` + fmt.Sprintf("%v", this.APIVersion) + `,`, `Time:` + strings.Replace(fmt.Sprintf("%v", this.Time), "Time", "Time", 1) + `,`, `FieldsType:` + fmt.Sprintf("%v", this.FieldsType) + `,`, - `FieldsV1:` + strings.Replace(this.FieldsV1.String(), "FieldsV1", "FieldsV1", 1) + `,`, + `FieldsV1:` + strings.Replace(fmt.Sprintf("%v", this.FieldsV1), "FieldsV1", "FieldsV1", 1) + `,`, `}`, }, "") return s @@ -5508,6 +5633,239 @@ func (m *APIVersions) Unmarshal(dAtA []byte) error { } return nil } +func (m *Condition) Unmarshal(dAtA []byte) error { + l := len(dAtA) + iNdEx := 0 + for iNdEx < l { + preIndex := iNdEx + var wire uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + wire |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + fieldNum := int32(wire >> 3) + wireType := int(wire & 0x7) + if wireType == 4 { + return fmt.Errorf("proto: Condition: wiretype end group for non-group") + } + if fieldNum <= 0 { + return fmt.Errorf("proto: Condition: illegal tag %d (wire type %d)", fieldNum, wire) + } + switch fieldNum { + case 1: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Type", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Type = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 2: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Status", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Status = ConditionStatus(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 3: + if wireType != 0 { + return fmt.Errorf("proto: wrong wireType = %d for field ObservedGeneration", wireType) + } + m.ObservedGeneration = 0 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + m.ObservedGeneration |= int64(b&0x7F) << shift + if b < 0x80 { + break + } + } + case 4: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field LastTransitionTime", wireType) + } + var msglen int + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + msglen |= int(b&0x7F) << shift + if b < 0x80 { + break + } + } + if msglen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + msglen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + if err := m.LastTransitionTime.Unmarshal(dAtA[iNdEx:postIndex]); err != nil { + return err + } + iNdEx = postIndex + case 5: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Reason", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Reason = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + case 6: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field Message", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.Message = string(dAtA[iNdEx:postIndex]) + iNdEx = postIndex + default: + iNdEx = preIndex + skippy, err := skipGenerated(dAtA[iNdEx:]) + if err != nil { + return err + } + if skippy < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) < 0 { + return ErrInvalidLengthGenerated + } + if (iNdEx + skippy) > l { + return io.ErrUnexpectedEOF + } + iNdEx += skippy + } + } + + if iNdEx > l { + return io.ErrUnexpectedEOF + } + return nil +} func (m *CreateOptions) Unmarshal(dAtA []byte) error { l := len(dAtA) iNdEx := 0 @@ -7811,6 +8169,38 @@ func (m *ListOptions) Unmarshal(dAtA []byte) error { } } m.AllowWatchBookmarks = bool(v != 0) + case 10: + if wireType != 2 { + return fmt.Errorf("proto: wrong wireType = %d for field ResourceVersionMatch", wireType) + } + var stringLen uint64 + for shift := uint(0); ; shift += 7 { + if shift >= 64 { + return ErrIntOverflowGenerated + } + if iNdEx >= l { + return io.ErrUnexpectedEOF + } + b := dAtA[iNdEx] + iNdEx++ + stringLen |= uint64(b&0x7F) << shift + if b < 0x80 { + break + } + } + intStringLen := int(stringLen) + if intStringLen < 0 { + return ErrInvalidLengthGenerated + } + postIndex := iNdEx + intStringLen + if postIndex < 0 { + return ErrInvalidLengthGenerated + } + if postIndex > l { + return io.ErrUnexpectedEOF + } + m.ResourceVersionMatch = ResourceVersionMatch(dAtA[iNdEx:postIndex]) + iNdEx = postIndex default: iNdEx = preIndex skippy, err := skipGenerated(dAtA[iNdEx:]) @@ -11004,6 +11394,7 @@ func (m *WatchEvent) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -11035,10 +11426,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -11059,55 +11448,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto index ba1194dc..b72d43ff 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/generated.proto @@ -134,6 +134,73 @@ message APIVersions { repeated ServerAddressByClientCIDR serverAddressByClientCIDRs = 2; } +// Condition contains details for one aspect of the current state of this API Resource. +// --- +// This struct is intended for direct use as an array at the field path .status.conditions. For example, +// type FooStatus struct{ +// // Represents the observations of a foo's current state. +// // Known .status.conditions.type are: "Available", "Progressing", and "Degraded" +// // +patchMergeKey=type +// // +patchStrategy=merge +// // +listType=map +// // +listMapKey=type +// Conditions []metav1.Condition `json:"conditions,omitempty" patchStrategy:"merge" patchMergeKey:"type" protobuf:"bytes,1,rep,name=conditions"` +// +// // other fields +// } +message Condition { + // type of condition in CamelCase or in foo.example.com/CamelCase. + // --- + // Many .condition.type values are consistent across resources like Available, but because arbitrary conditions can be + // useful (see .node.status.conditions), the ability to deconflict is important. + // The regex it matches is (dns1123SubdomainFmt/)?(qualifiedNameFmt) + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:Pattern=`^([a-z0-9]([-a-z0-9]*[a-z0-9])?(\.[a-z0-9]([-a-z0-9]*[a-z0-9])?)*/)?(([A-Za-z0-9][-A-Za-z0-9_.]*)?[A-Za-z0-9])$` + // +kubebuilder:validation:MaxLength=316 + optional string type = 1; + + // status of the condition, one of True, False, Unknown. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:Enum=True;False;Unknown + optional string status = 2; + + // observedGeneration represents the .metadata.generation that the condition was set based upon. + // For instance, if .metadata.generation is currently 12, but the .status.conditions[x].observedGeneration is 9, the condition is out of date + // with respect to the current state of the instance. + // +optional + // +kubebuilder:validation:Minimum=0 + optional int64 observedGeneration = 3; + + // lastTransitionTime is the last time the condition transitioned from one status to another. + // This should be when the underlying condition changed. If that is not known, then using the time when the API field changed is acceptable. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:Type=string + // +kubebuilder:validation:Format=date-time + optional Time lastTransitionTime = 4; + + // reason contains a programmatic identifier indicating the reason for the condition's last transition. + // Producers of specific condition types may define expected values and meanings for this field, + // and whether the values are considered a guaranteed API. + // The value should be a CamelCase string. + // This field may not be empty. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:MaxLength=1024 + // +kubebuilder:validation:MinLength=1 + // +kubebuilder:validation:Pattern=`^[A-Za-z]([A-Za-z0-9_,:]*[A-Za-z0-9_])?$` + optional string reason = 5; + + // message is a human readable message indicating details about the transition. + // This may be an empty string. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:MaxLength=32768 + optional string message = 6; +} + // CreateOptions may be provided when creating an API object. message CreateOptions { // When present, indicates that modifications should not be @@ -224,6 +291,7 @@ message ExportOptions { // If a key maps to an empty Fields value, the field that key represents is part of the set. // // The exact format is defined in sigs.k8s.io/structured-merge-diff +// +protobuf.options.(gogoproto.goproto_stringer)=false message FieldsV1 { // Raw is the underlying serialization of this object. optional bytes Raw = 1; @@ -231,10 +299,12 @@ message FieldsV1 { // GetOptions is the standard query options to the standard REST get call. message GetOptions { - // When specified: - // - if unset, then the result is returned from remote storage based on quorum-read flag; - // - if it's 0, then we simply return what we currently have in cache, no guarantee; - // - if set to non zero, then the result is at least as fresh as given rv. + // resourceVersion sets a constraint on what resource versions a request may be served from. + // See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for + // details. + // + // Defaults to unset + // +optional optional string resourceVersion = 1; } @@ -420,15 +490,24 @@ message ListOptions { // +optional optional bool allowWatchBookmarks = 9; - // When specified with a watch call, shows changes that occur after that particular version of a resource. - // Defaults to changes from the beginning of history. - // When specified for list: - // - if unset, then the result is returned from remote storage based on quorum-read flag; - // - if it's 0, then we simply return what we currently have in cache, no guarantee; - // - if set to non zero, then the result is at least as fresh as given rv. + // resourceVersion sets a constraint on what resource versions a request may be served from. + // See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for + // details. + // + // Defaults to unset // +optional optional string resourceVersion = 4; + // resourceVersionMatch determines how resourceVersion is applied to list calls. + // It is highly recommended that resourceVersionMatch be set for list calls where + // resourceVersion is set + // See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for + // details. + // + // Defaults to unset + // +optional + optional string resourceVersionMatch = 10; + // Timeout for the list/watch call. // This limits the duration of the call, regardless of any activity or inactivity. // +optional @@ -546,7 +625,7 @@ message ObjectMeta { // +optional optional string generateName = 2; - // Namespace defines the space within each name must be unique. An empty namespace is + // Namespace defines the space within which each name must be unique. An empty namespace is // equivalent to the "default" namespace, but "default" is the canonical representation. // Not all objects are required to be scoped to a namespace - the value of this field for // those objects will be empty. diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/helpers.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/helpers.go index ec016fd3..ad989ad7 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/helpers.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/helpers.go @@ -252,7 +252,9 @@ func ResetObjectMetaForStatus(meta, existingMeta Object) { meta.SetAnnotations(existingMeta.GetAnnotations()) meta.SetFinalizers(existingMeta.GetFinalizers()) meta.SetOwnerReferences(existingMeta.GetOwnerReferences()) - meta.SetManagedFields(existingMeta.GetManagedFields()) + // managedFields must be preserved since it's been modified to + // track changed fields in the status update. + //meta.SetManagedFields(existingMeta.GetManagedFields()) } // MarshalJSON implements json.Marshaler diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/meta.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/meta.go index 912cf203..2002f91b 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/meta.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/meta.go @@ -156,7 +156,9 @@ func (meta *ObjectMeta) GetDeletionTimestamp() *Time { return meta.DeletionTimes func (meta *ObjectMeta) SetDeletionTimestamp(deletionTimestamp *Time) { meta.DeletionTimestamp = deletionTimestamp } -func (meta *ObjectMeta) GetDeletionGracePeriodSeconds() *int64 { return meta.DeletionGracePeriodSeconds } +func (meta *ObjectMeta) GetDeletionGracePeriodSeconds() *int64 { + return meta.DeletionGracePeriodSeconds +} func (meta *ObjectMeta) SetDeletionGracePeriodSeconds(deletionGracePeriodSeconds *int64) { meta.DeletionGracePeriodSeconds = deletionGracePeriodSeconds } diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/register.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/register.go index a7b8aa34..c1a07717 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/register.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/register.go @@ -53,15 +53,6 @@ var scheme = runtime.NewScheme() // ParameterCodec knows about query parameters used with the meta v1 API spec. var ParameterCodec = runtime.NewParameterCodec(scheme) -func addEventConversionFuncs(scheme *runtime.Scheme) error { - return scheme.AddConversionFuncs( - Convert_v1_WatchEvent_To_watch_Event, - Convert_v1_InternalEvent_To_v1_WatchEvent, - Convert_watch_Event_To_v1_WatchEvent, - Convert_v1_WatchEvent_To_v1_InternalEvent, - ) -} - var optionsTypes = []runtime.Object{ &ListOptions{}, &ExportOptions{}, @@ -90,10 +81,8 @@ func AddToGroupVersion(scheme *runtime.Scheme, groupVersion schema.GroupVersion) &APIResourceList{}, ) - utilruntime.Must(addEventConversionFuncs(scheme)) - // register manually. This usually goes through the SchemeBuilder, which we cannot use here. - utilruntime.Must(AddConversionFuncs(scheme)) + utilruntime.Must(RegisterConversions(scheme)) utilruntime.Must(RegisterDefaults(scheme)) } @@ -106,9 +95,7 @@ func AddMetaToScheme(scheme *runtime.Scheme) error { &PartialObjectMetadataList{}, ) - return scheme.AddConversionFuncs( - Convert_Slice_string_To_v1_IncludeObjectPolicy, - ) + return nil } func init() { diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/time.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/time.go index fe510ed9..4a1d89cf 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/time.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/time.go @@ -153,6 +153,16 @@ func (t Time) MarshalJSON() ([]byte, error) { return buf, nil } +// ToUnstructured implements the value.UnstructuredConverter interface. +func (t Time) ToUnstructured() interface{} { + if t.IsZero() { + return nil + } + buf := make([]byte, 0, len(time.RFC3339)) + buf = t.UTC().AppendFormat(buf, time.RFC3339) + return string(buf) +} + // OpenAPISchemaType is used by the kube-openapi generator when constructing // the OpenAPI spec of this type. // diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types.go index bf125b62..bb57f2cc 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types.go @@ -135,7 +135,7 @@ type ObjectMeta struct { // +optional GenerateName string `json:"generateName,omitempty" protobuf:"bytes,2,opt,name=generateName"` - // Namespace defines the space within each name must be unique. An empty namespace is + // Namespace defines the space within which each name must be unique. An empty namespace is // equivalent to the "default" namespace, but "default" is the canonical representation. // Not all objects are required to be scoped to a namespace - the value of this field for // those objects will be empty. @@ -355,14 +355,23 @@ type ListOptions struct { // +optional AllowWatchBookmarks bool `json:"allowWatchBookmarks,omitempty" protobuf:"varint,9,opt,name=allowWatchBookmarks"` - // When specified with a watch call, shows changes that occur after that particular version of a resource. - // Defaults to changes from the beginning of history. - // When specified for list: - // - if unset, then the result is returned from remote storage based on quorum-read flag; - // - if it's 0, then we simply return what we currently have in cache, no guarantee; - // - if set to non zero, then the result is at least as fresh as given rv. + // resourceVersion sets a constraint on what resource versions a request may be served from. + // See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for + // details. + // + // Defaults to unset // +optional ResourceVersion string `json:"resourceVersion,omitempty" protobuf:"bytes,4,opt,name=resourceVersion"` + + // resourceVersionMatch determines how resourceVersion is applied to list calls. + // It is highly recommended that resourceVersionMatch be set for list calls where + // resourceVersion is set + // See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for + // details. + // + // Defaults to unset + // +optional + ResourceVersionMatch ResourceVersionMatch `json:"resourceVersionMatch,omitempty" protobuf:"bytes,10,opt,name=resourceVersionMatch,casttype=ResourceVersionMatch"` // Timeout for the list/watch call. // This limits the duration of the call, regardless of any activity or inactivity. // +optional @@ -402,6 +411,25 @@ type ListOptions struct { Continue string `json:"continue,omitempty" protobuf:"bytes,8,opt,name=continue"` } +// resourceVersionMatch specifies how the resourceVersion parameter is applied. resourceVersionMatch +// may only be set if resourceVersion is also set. +// +// "NotOlderThan" matches data at least as new as the provided resourceVersion. +// "Exact" matches data at the exact resourceVersion provided. +// +// See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for +// details. +type ResourceVersionMatch string + +const ( + // ResourceVersionMatchNotOlderThan matches data at least as new as the provided + // resourceVersion. + ResourceVersionMatchNotOlderThan ResourceVersionMatch = "NotOlderThan" + // ResourceVersionMatchExact matches data at the exact resourceVersion + // provided. + ResourceVersionMatchExact ResourceVersionMatch = "Exact" +) + // +k8s:conversion-gen:explicit-from=net/url.Values // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object @@ -423,10 +451,12 @@ type ExportOptions struct { // GetOptions is the standard query options to the standard REST get call. type GetOptions struct { TypeMeta `json:",inline"` - // When specified: - // - if unset, then the result is returned from remote storage based on quorum-read flag; - // - if it's 0, then we simply return what we currently have in cache, no guarantee; - // - if set to non zero, then the result is at least as fresh as given rv. + // resourceVersion sets a constraint on what resource versions a request may be served from. + // See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for + // details. + // + // Defaults to unset + // +optional ResourceVersion string `json:"resourceVersion,omitempty" protobuf:"bytes,1,opt,name=resourceVersion"` // +k8s:deprecated=includeUninitialized,protobuf=2 } @@ -873,6 +903,9 @@ const ( // FieldManagerConflict is used to report when another client claims to manage this field, // It should only be returned for a request using server-side apply. CauseTypeFieldManagerConflict CauseType = "FieldManagerConflict" + // CauseTypeResourceVersionTooLarge is used to report that the requested resource version + // is newer than the data observed by the API server, so the request cannot be served. + CauseTypeResourceVersionTooLarge CauseType = "ResourceVersionTooLarge" ) // +k8s:deepcopy-gen:interfaces=k8s.io/apimachinery/pkg/runtime.Object @@ -1145,11 +1178,16 @@ const ( // If a key maps to an empty Fields value, the field that key represents is part of the set. // // The exact format is defined in sigs.k8s.io/structured-merge-diff +// +protobuf.options.(gogoproto.goproto_stringer)=false type FieldsV1 struct { // Raw is the underlying serialization of this object. Raw []byte `json:"-" protobuf:"bytes,1,opt,name=Raw"` } +func (f FieldsV1) String() string { + return string(f.Raw) +} + // TODO: Table does not generate to protobuf because of the interface{} - fix protobuf // generation to support a meta type that can accept any valid JSON. This can be introduced // in a v1 because clients a) receive an error if they try to access proto today, and b) @@ -1311,3 +1349,65 @@ type PartialObjectMetadataList struct { // items contains each of the included items. Items []PartialObjectMetadata `json:"items" protobuf:"bytes,2,rep,name=items"` } + +// Condition contains details for one aspect of the current state of this API Resource. +// --- +// This struct is intended for direct use as an array at the field path .status.conditions. For example, +// type FooStatus struct{ +// // Represents the observations of a foo's current state. +// // Known .status.conditions.type are: "Available", "Progressing", and "Degraded" +// // +patchMergeKey=type +// // +patchStrategy=merge +// // +listType=map +// // +listMapKey=type +// Conditions []metav1.Condition `json:"conditions,omitempty" patchStrategy:"merge" patchMergeKey:"type" protobuf:"bytes,1,rep,name=conditions"` +// +// // other fields +// } +type Condition struct { + // type of condition in CamelCase or in foo.example.com/CamelCase. + // --- + // Many .condition.type values are consistent across resources like Available, but because arbitrary conditions can be + // useful (see .node.status.conditions), the ability to deconflict is important. + // The regex it matches is (dns1123SubdomainFmt/)?(qualifiedNameFmt) + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:Pattern=`^([a-z0-9]([-a-z0-9]*[a-z0-9])?(\.[a-z0-9]([-a-z0-9]*[a-z0-9])?)*/)?(([A-Za-z0-9][-A-Za-z0-9_.]*)?[A-Za-z0-9])$` + // +kubebuilder:validation:MaxLength=316 + Type string `json:"type" protobuf:"bytes,1,opt,name=type"` + // status of the condition, one of True, False, Unknown. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:Enum=True;False;Unknown + Status ConditionStatus `json:"status" protobuf:"bytes,2,opt,name=status"` + // observedGeneration represents the .metadata.generation that the condition was set based upon. + // For instance, if .metadata.generation is currently 12, but the .status.conditions[x].observedGeneration is 9, the condition is out of date + // with respect to the current state of the instance. + // +optional + // +kubebuilder:validation:Minimum=0 + ObservedGeneration int64 `json:"observedGeneration,omitempty" protobuf:"varint,3,opt,name=observedGeneration"` + // lastTransitionTime is the last time the condition transitioned from one status to another. + // This should be when the underlying condition changed. If that is not known, then using the time when the API field changed is acceptable. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:Type=string + // +kubebuilder:validation:Format=date-time + LastTransitionTime Time `json:"lastTransitionTime" protobuf:"bytes,4,opt,name=lastTransitionTime"` + // reason contains a programmatic identifier indicating the reason for the condition's last transition. + // Producers of specific condition types may define expected values and meanings for this field, + // and whether the values are considered a guaranteed API. + // The value should be a CamelCase string. + // This field may not be empty. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:MaxLength=1024 + // +kubebuilder:validation:MinLength=1 + // +kubebuilder:validation:Pattern=`^[A-Za-z]([A-Za-z0-9_,:]*[A-Za-z0-9_])?$` + Reason string `json:"reason" protobuf:"bytes,5,opt,name=reason"` + // message is a human readable message indicating details about the transition. + // This may be an empty string. + // +required + // +kubebuilder:validation:Required + // +kubebuilder:validation:MaxLength=32768 + Message string `json:"message" protobuf:"bytes,6,opt,name=message"` +} diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types_swagger_doc_generated.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types_swagger_doc_generated.go index b62e591e..ace0abfb 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types_swagger_doc_generated.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/types_swagger_doc_generated.go @@ -86,6 +86,20 @@ func (APIVersions) SwaggerDoc() map[string]string { return map_APIVersions } +var map_Condition = map[string]string{ + "": "Condition contains details for one aspect of the current state of this API Resource.", + "type": "type of condition in CamelCase or in foo.example.com/CamelCase.", + "status": "status of the condition, one of True, False, Unknown.", + "observedGeneration": "observedGeneration represents the .metadata.generation that the condition was set based upon. For instance, if .metadata.generation is currently 12, but the .status.conditions[x].observedGeneration is 9, the condition is out of date with respect to the current state of the instance.", + "lastTransitionTime": "lastTransitionTime is the last time the condition transitioned from one status to another. This should be when the underlying condition changed. If that is not known, then using the time when the API field changed is acceptable.", + "reason": "reason contains a programmatic identifier indicating the reason for the condition's last transition. Producers of specific condition types may define expected values and meanings for this field, and whether the values are considered a guaranteed API. The value should be a CamelCase string. This field may not be empty.", + "message": "message is a human readable message indicating details about the transition. This may be an empty string.", +} + +func (Condition) SwaggerDoc() map[string]string { + return map_Condition +} + var map_CreateOptions = map[string]string{ "": "CreateOptions may be provided when creating an API object.", "dryRun": "When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed", @@ -129,7 +143,7 @@ func (FieldsV1) SwaggerDoc() map[string]string { var map_GetOptions = map[string]string{ "": "GetOptions is the standard query options to the standard REST get call.", - "resourceVersion": "When specified: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv.", + "resourceVersion": "resourceVersion sets a constraint on what resource versions a request may be served from. See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for details.\n\nDefaults to unset", } func (GetOptions) SwaggerDoc() map[string]string { @@ -190,15 +204,16 @@ func (ListMeta) SwaggerDoc() map[string]string { } var map_ListOptions = map[string]string{ - "": "ListOptions is the query options to a standard REST list call.", - "labelSelector": "A selector to restrict the list of returned objects by their labels. Defaults to everything.", - "fieldSelector": "A selector to restrict the list of returned objects by their fields. Defaults to everything.", - "watch": "Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion.", - "allowWatchBookmarks": "allowWatchBookmarks requests watch events with type \"BOOKMARK\". Servers that do not implement bookmarks may ignore this flag and bookmarks are sent at the server's discretion. Clients should not assume bookmarks are returned at any specific interval, nor may they assume the server will send any BOOKMARK event during a session. If this is not a watch, this field is ignored. If the feature gate WatchBookmarks is not enabled in apiserver, this field is ignored.", - "resourceVersion": "When specified with a watch call, shows changes that occur after that particular version of a resource. Defaults to changes from the beginning of history. When specified for list: - if unset, then the result is returned from remote storage based on quorum-read flag; - if it's 0, then we simply return what we currently have in cache, no guarantee; - if set to non zero, then the result is at least as fresh as given rv.", - "timeoutSeconds": "Timeout for the list/watch call. This limits the duration of the call, regardless of any activity or inactivity.", - "limit": "limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true.\n\nThe server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned.", - "continue": "The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server, the server will respond with a 410 ResourceExpired error together with a continue token. If the client needs a consistent list, it must restart their list without the continue field. Otherwise, the client may send another list request with the token received with the 410 error, the server will respond with a list starting from the next key, but from the latest snapshot, which is inconsistent from the previous list results - objects that are created, modified, or deleted after the first list request will be included in the response, as long as their keys are after the \"next key\".\n\nThis field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications.", + "": "ListOptions is the query options to a standard REST list call.", + "labelSelector": "A selector to restrict the list of returned objects by their labels. Defaults to everything.", + "fieldSelector": "A selector to restrict the list of returned objects by their fields. Defaults to everything.", + "watch": "Watch for changes to the described resources and return them as a stream of add, update, and remove notifications. Specify resourceVersion.", + "allowWatchBookmarks": "allowWatchBookmarks requests watch events with type \"BOOKMARK\". Servers that do not implement bookmarks may ignore this flag and bookmarks are sent at the server's discretion. Clients should not assume bookmarks are returned at any specific interval, nor may they assume the server will send any BOOKMARK event during a session. If this is not a watch, this field is ignored. If the feature gate WatchBookmarks is not enabled in apiserver, this field is ignored.", + "resourceVersion": "resourceVersion sets a constraint on what resource versions a request may be served from. See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for details.\n\nDefaults to unset", + "resourceVersionMatch": "resourceVersionMatch determines how resourceVersion is applied to list calls. It is highly recommended that resourceVersionMatch be set for list calls where resourceVersion is set See https://kubernetes.io/docs/reference/using-api/api-concepts/#resource-versions for details.\n\nDefaults to unset", + "timeoutSeconds": "Timeout for the list/watch call. This limits the duration of the call, regardless of any activity or inactivity.", + "limit": "limit is a maximum number of responses to return for a list call. If more items exist, the server will set the `continue` field on the list metadata to a value that can be used with the same initial query to retrieve the next set of results. Setting a limit may return fewer than the requested amount of items (up to zero items) in the event all requested objects are filtered out and clients should only use the presence of the continue field to determine whether more results are available. Servers may choose not to support the limit argument and will return all of the available results. If limit is specified and the continue field is empty, clients may assume that no more results are available. This field is not supported if watch is true.\n\nThe server guarantees that the objects returned when using continue will be identical to issuing a single list call without a limit - that is, no objects created, modified, or deleted after the first request is issued will be included in any subsequent continued requests. This is sometimes referred to as a consistent snapshot, and ensures that a client that is using limit to receive smaller chunks of a very large result can ensure they see all possible objects. If objects are updated during a chunked list the version of the object that was present at the time the first list result was calculated is returned.", + "continue": "The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server, the server will respond with a 410 ResourceExpired error together with a continue token. If the client needs a consistent list, it must restart their list without the continue field. Otherwise, the client may send another list request with the token received with the 410 error, the server will respond with a list starting from the next key, but from the latest snapshot, which is inconsistent from the previous list results - objects that are created, modified, or deleted after the first list request will be included in the response, as long as their keys are after the \"next key\".\n\nThis field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications.", } func (ListOptions) SwaggerDoc() map[string]string { @@ -223,7 +238,7 @@ var map_ObjectMeta = map[string]string{ "": "ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.", "name": "Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names", "generateName": "GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency", - "namespace": "Namespace defines the space within each name must be unique. An empty namespace is equivalent to the \"default\" namespace, but \"default\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces", + "namespace": "Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \"default\" namespace, but \"default\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces", "selfLink": "SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.", "uid": "UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids", "resourceVersion": "An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency", diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/unstructured/helpers.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/unstructured/helpers.go index 4244b8a6..54a231e4 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/unstructured/helpers.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/unstructured/helpers.go @@ -27,7 +27,7 @@ import ( "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apimachinery/pkg/types" "k8s.io/apimachinery/pkg/util/json" - "k8s.io/klog" + "k8s.io/klog/v2" ) // NestedFieldCopy returns a deep copy of the value of a nested field. diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.conversion.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.conversion.go index 2ade69dd..06afd9b5 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.conversion.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.conversion.go @@ -145,6 +145,11 @@ func RegisterConversions(s *runtime.Scheme) error { }); err != nil { return err } + if err := s.AddConversionFunc((*[]string)(nil), (*ResourceVersionMatch)(nil), func(a, b interface{}, scope conversion.Scope) error { + return Convert_Slice_string_To_v1_ResourceVersionMatch(a.(*[]string), b.(*ResourceVersionMatch), scope) + }); err != nil { + return err + } if err := s.AddConversionFunc((*[]string)(nil), (*Time)(nil), func(a, b interface{}, scope conversion.Scope) error { return Convert_Slice_string_To_v1_Time(a.(*[]string), b.(*Time), scope) }); err != nil { @@ -415,6 +420,13 @@ func autoConvert_url_Values_To_v1_ListOptions(in *url.Values, out *ListOptions, } else { out.ResourceVersion = "" } + if values, ok := map[string][]string(*in)["resourceVersionMatch"]; ok && len(values) > 0 { + if err := Convert_Slice_string_To_v1_ResourceVersionMatch(&values, &out.ResourceVersionMatch, s); err != nil { + return err + } + } else { + out.ResourceVersionMatch = "" + } if values, ok := map[string][]string(*in)["timeoutSeconds"]; ok && len(values) > 0 { if err := runtime.Convert_Slice_string_To_Pointer_int64(&values, &out.TimeoutSeconds, s); err != nil { return err diff --git a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.deepcopy.go b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.deepcopy.go index b82fdf20..1aa73bd2 100644 --- a/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.deepcopy.go +++ b/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/zz_generated.deepcopy.go @@ -191,6 +191,23 @@ func (in *APIVersions) DeepCopyObject() runtime.Object { return nil } +// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. +func (in *Condition) DeepCopyInto(out *Condition) { + *out = *in + in.LastTransitionTime.DeepCopyInto(&out.LastTransitionTime) + return +} + +// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new Condition. +func (in *Condition) DeepCopy() *Condition { + if in == nil { + return nil + } + out := new(Condition) + in.DeepCopyInto(out) + return out +} + // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *CreateOptions) DeepCopyInto(out *CreateOptions) { *out = *in diff --git a/vendor/k8s.io/apimachinery/pkg/conversion/converter.go b/vendor/k8s.io/apimachinery/pkg/conversion/converter.go index bc615dc3..838d5b0a 100644 --- a/vendor/k8s.io/apimachinery/pkg/conversion/converter.go +++ b/vendor/k8s.io/apimachinery/pkg/conversion/converter.go @@ -54,7 +54,8 @@ type Converter struct { generatedConversionFuncs ConversionFuncs // Set of conversions that should be treated as a no-op - ignoredConversions map[typePair]struct{} + ignoredConversions map[typePair]struct{} + ignoredUntypedConversions map[typePair]struct{} // This is a map from a source field type and name, to a list of destination // field type and name. @@ -83,17 +84,23 @@ type Converter struct { // NewConverter creates a new Converter object. func NewConverter(nameFn NameFunc) *Converter { c := &Converter{ - conversionFuncs: NewConversionFuncs(), - generatedConversionFuncs: NewConversionFuncs(), - ignoredConversions: make(map[typePair]struct{}), - nameFunc: nameFn, - structFieldDests: make(map[typeNamePair][]typeNamePair), - structFieldSources: make(map[typeNamePair][]typeNamePair), + conversionFuncs: NewConversionFuncs(), + generatedConversionFuncs: NewConversionFuncs(), + ignoredConversions: make(map[typePair]struct{}), + ignoredUntypedConversions: make(map[typePair]struct{}), + nameFunc: nameFn, + structFieldDests: make(map[typeNamePair][]typeNamePair), + structFieldSources: make(map[typeNamePair][]typeNamePair), inputFieldMappingFuncs: make(map[reflect.Type]FieldMappingFunc), inputDefaultFlags: make(map[reflect.Type]FieldMatchingFlags), } - c.RegisterConversionFunc(Convert_Slice_byte_To_Slice_byte) + c.RegisterUntypedConversionFunc( + (*[]byte)(nil), (*[]byte)(nil), + func(a, b interface{}, s Scope) error { + return Convert_Slice_byte_To_Slice_byte(a.(*[]byte), b.(*[]byte), s) + }, + ) return c } @@ -131,10 +138,6 @@ type Scope interface { // parameters, you'll run out of stack space before anything useful happens. Convert(src, dest interface{}, flags FieldMatchingFlags) error - // DefaultConvert performs the default conversion, without calling a conversion func - // on the current stack frame. This makes it safe to call from a conversion func. - DefaultConvert(src, dest interface{}, flags FieldMatchingFlags) error - // SrcTags and DestTags contain the struct tags that src and dest had, respectively. // If the enclosing object was not a struct, then these will contain no tags, of course. SrcTag() reflect.StructTag @@ -153,31 +156,14 @@ type FieldMappingFunc func(key string, sourceTag, destTag reflect.StructTag) (so func NewConversionFuncs() ConversionFuncs { return ConversionFuncs{ - fns: make(map[typePair]reflect.Value), untyped: make(map[typePair]ConversionFunc), } } type ConversionFuncs struct { - fns map[typePair]reflect.Value untyped map[typePair]ConversionFunc } -// Add adds the provided conversion functions to the lookup table - they must have the signature -// `func(type1, type2, Scope) error`. Functions are added in the order passed and will override -// previously registered pairs. -func (c ConversionFuncs) Add(fns ...interface{}) error { - for _, fn := range fns { - fv := reflect.ValueOf(fn) - ft := fv.Type() - if err := verifyConversionFunctionSignature(ft); err != nil { - return err - } - c.fns[typePair{ft.In(0).Elem(), ft.In(1).Elem()}] = fv - } - return nil -} - // AddUntyped adds the provided conversion function to the lookup table for the types that are // supplied as a and b. a and b must be pointers or an error is returned. This method overwrites // previously defined functions. @@ -197,12 +183,6 @@ func (c ConversionFuncs) AddUntyped(a, b interface{}, fn ConversionFunc) error { // both other and c, with other conversions taking precedence. func (c ConversionFuncs) Merge(other ConversionFuncs) ConversionFuncs { merged := NewConversionFuncs() - for k, v := range c.fns { - merged.fns[k] = v - } - for k, v := range other.fns { - merged.fns[k] = v - } for k, v := range c.untyped { merged.untyped[k] = v } @@ -290,12 +270,6 @@ func (s *scope) Convert(src, dest interface{}, flags FieldMatchingFlags) error { return s.converter.Convert(src, dest, flags, s.meta) } -// DefaultConvert continues a conversion, performing a default conversion (no conversion func) -// for the current stack frame. -func (s *scope) DefaultConvert(src, dest interface{}, flags FieldMatchingFlags) error { - return s.converter.DefaultConvert(src, dest, flags, s.meta) -} - // SrcTag returns the tag of the struct containing the current source item, if any. func (s *scope) SrcTag() reflect.StructTag { return s.srcStack.top().tag @@ -360,29 +334,6 @@ func verifyConversionFunctionSignature(ft reflect.Type) error { return nil } -// RegisterConversionFunc registers a conversion func with the -// Converter. conversionFunc must take three parameters: a pointer to the input -// type, a pointer to the output type, and a conversion.Scope (which should be -// used if recursive conversion calls are desired). It must return an error. -// -// Example: -// c.RegisterConversionFunc( -// func(in *Pod, out *v1.Pod, s Scope) error { -// // conversion logic... -// return nil -// }) -// DEPRECATED: Will be removed in favor of RegisterUntypedConversionFunc -func (c *Converter) RegisterConversionFunc(conversionFunc interface{}) error { - return c.conversionFuncs.Add(conversionFunc) -} - -// Similar to RegisterConversionFunc, but registers conversion function that were -// automatically generated. -// DEPRECATED: Will be removed in favor of RegisterGeneratedUntypedConversionFunc -func (c *Converter) RegisterGeneratedConversionFunc(conversionFunc interface{}) error { - return c.generatedConversionFuncs.Add(conversionFunc) -} - // RegisterUntypedConversionFunc registers a function that converts between a and b by passing objects of those // types to the provided function. The function *must* accept objects of a and b - this machinery will not enforce // any other guarantee. @@ -409,6 +360,7 @@ func (c *Converter) RegisterIgnoredConversion(from, to interface{}) error { return fmt.Errorf("expected pointer arg for 'to' param 1, got: %v", typeTo) } c.ignoredConversions[typePair{typeFrom.Elem(), typeTo.Elem()}] = struct{}{} + c.ignoredUntypedConversions[typePair{typeFrom, typeTo}] = struct{}{} return nil } @@ -470,18 +422,6 @@ func (c *Converter) Convert(src, dest interface{}, flags FieldMatchingFlags, met return c.doConversion(src, dest, flags, meta, c.convert) } -// DefaultConvert will translate src to dest if it knows how. Both must be pointers. -// No conversion func is used. If the default copying mechanism -// doesn't work on this type pair, an error will be returned. -// Read the comments on the various FieldMatchingFlags constants to understand -// what the 'flags' parameter does. -// 'meta' is given to allow you to pass information to conversion functions, -// it is not used by DefaultConvert() other than storing it in the scope. -// Not safe for objects with cyclic references! -func (c *Converter) DefaultConvert(src, dest interface{}, flags FieldMatchingFlags, meta *Meta) error { - return c.doConversion(src, dest, flags, meta, c.defaultConvert) -} - type conversionFunc func(sv, dv reflect.Value, scope *scope) error func (c *Converter) doConversion(src, dest interface{}, flags FieldMatchingFlags, meta *Meta, f conversionFunc) error { @@ -491,59 +431,51 @@ func (c *Converter) doConversion(src, dest interface{}, flags FieldMatchingFlags flags: flags, meta: meta, } + + // ignore conversions of this type + if _, ok := c.ignoredUntypedConversions[pair]; ok { + return nil + } if fn, ok := c.conversionFuncs.untyped[pair]; ok { return fn(src, dest, scope) } if fn, ok := c.generatedConversionFuncs.untyped[pair]; ok { return fn(src, dest, scope) } - // TODO: consider everything past this point deprecated - we want to support only point to point top level - // conversions dv, err := EnforcePtr(dest) if err != nil { return err } - if !dv.CanAddr() && !dv.CanSet() { - return fmt.Errorf("can't write to dest") - } sv, err := EnforcePtr(src) if err != nil { return err } + return fmt.Errorf("converting (%s) to (%s): unknown conversion", sv.Type(), dv.Type()) + + // TODO: Everything past this point is deprecated. + // Remove in 1.20 once we're sure it didn't break anything. + // Leave something on the stack, so that calls to struct tag getters never fail. scope.srcStack.push(scopeStackElem{}) scope.destStack.push(scopeStackElem{}) return f(sv, dv, scope) } -// callCustom calls 'custom' with sv & dv. custom must be a conversion function. -func (c *Converter) callCustom(sv, dv, custom reflect.Value, scope *scope) error { - if !sv.CanAddr() { - sv2 := reflect.New(sv.Type()) - sv2.Elem().Set(sv) - sv = sv2 - } else { - sv = sv.Addr() - } +// callUntyped calls predefined conversion func. +func (c *Converter) callUntyped(sv, dv reflect.Value, f ConversionFunc, scope *scope) error { if !dv.CanAddr() { - if !dv.CanSet() { - return scope.errorf("can't addr or set dest.") - } - dvOrig := dv - dv := reflect.New(dvOrig.Type()) - defer func() { dvOrig.Set(dv) }() - } else { - dv = dv.Addr() + return scope.errorf("cant addr dest") } - args := []reflect.Value{sv, dv, reflect.ValueOf(scope)} - ret := custom.Call(args)[0].Interface() - // This convolution is necessary because nil interfaces won't convert - // to errors. - if ret == nil { - return nil + var svPointer reflect.Value + if sv.CanAddr() { + svPointer = sv.Addr() + } else { + svPointer = reflect.New(sv.Type()) + svPointer.Elem().Set(sv) } - return ret.(error) + dvPointer := dv.Addr() + return f(svPointer.Interface(), dvPointer.Interface(), scope) } // convert recursively copies sv into dv, calling an appropriate conversion function if @@ -561,27 +493,14 @@ func (c *Converter) convert(sv, dv reflect.Value, scope *scope) error { } // Convert sv to dv. - if fv, ok := c.conversionFuncs.fns[pair]; ok { - if c.Debug != nil { - c.Debug.Logf("Calling custom conversion of '%v' to '%v'", st, dt) - } - return c.callCustom(sv, dv, fv, scope) + pair = typePair{reflect.PtrTo(sv.Type()), reflect.PtrTo(dv.Type())} + if f, ok := c.conversionFuncs.untyped[pair]; ok { + return c.callUntyped(sv, dv, f, scope) } - if fv, ok := c.generatedConversionFuncs.fns[pair]; ok { - if c.Debug != nil { - c.Debug.Logf("Calling generated conversion of '%v' to '%v'", st, dt) - } - return c.callCustom(sv, dv, fv, scope) + if f, ok := c.generatedConversionFuncs.untyped[pair]; ok { + return c.callUntyped(sv, dv, f, scope) } - return c.defaultConvert(sv, dv, scope) -} - -// defaultConvert recursively copies sv into dv. no conversion function is called -// for the current stack frame (but conversion functions may be called for nested objects) -func (c *Converter) defaultConvert(sv, dv reflect.Value, scope *scope) error { - dt, st := dv.Type(), sv.Type() - if !dv.CanSet() { return scope.errorf("Cannot set dest. (Tried to deep copy something with unexported fields?)") } diff --git a/vendor/k8s.io/apimachinery/pkg/fields/selector.go b/vendor/k8s.io/apimachinery/pkg/fields/selector.go index e3e4453b..a9e20497 100644 --- a/vendor/k8s.io/apimachinery/pkg/fields/selector.go +++ b/vendor/k8s.io/apimachinery/pkg/fields/selector.go @@ -57,13 +57,15 @@ type Selector interface { type nothingSelector struct{} -func (n nothingSelector) Matches(_ Fields) bool { return false } -func (n nothingSelector) Empty() bool { return false } -func (n nothingSelector) String() string { return "" } -func (n nothingSelector) Requirements() Requirements { return nil } -func (n nothingSelector) DeepCopySelector() Selector { return n } -func (n nothingSelector) RequiresExactMatch(field string) (value string, found bool) { return "", false } -func (n nothingSelector) Transform(fn TransformFunc) (Selector, error) { return n, nil } +func (n nothingSelector) Matches(_ Fields) bool { return false } +func (n nothingSelector) Empty() bool { return false } +func (n nothingSelector) String() string { return "" } +func (n nothingSelector) Requirements() Requirements { return nil } +func (n nothingSelector) DeepCopySelector() Selector { return n } +func (n nothingSelector) RequiresExactMatch(field string) (value string, found bool) { + return "", false +} +func (n nothingSelector) Transform(fn TransformFunc) (Selector, error) { return n, nil } // Nothing returns a selector that matches no fields func Nothing() Selector { diff --git a/vendor/k8s.io/apimachinery/pkg/labels/labels.go b/vendor/k8s.io/apimachinery/pkg/labels/labels.go index abf3ace6..d9eeb4f9 100644 --- a/vendor/k8s.io/apimachinery/pkg/labels/labels.go +++ b/vendor/k8s.io/apimachinery/pkg/labels/labels.go @@ -57,14 +57,22 @@ func (ls Set) Get(label string) string { return ls[label] } -// AsSelector converts labels into a selectors. +// AsSelector converts labels into a selectors. It does not +// perform any validation, which means the server will reject +// the request if the Set contains invalid values. func (ls Set) AsSelector() Selector { return SelectorFromSet(ls) } +// AsValidatedSelector converts labels into a selectors. +// The Set is validated client-side, which allows to catch errors early. +func (ls Set) AsValidatedSelector() (Selector, error) { + return ValidatedSelectorFromSet(ls) +} + // AsSelectorPreValidated converts labels into a selector, but -// assumes that labels are already validated and thus don't -// preform any validation. +// assumes that labels are already validated and thus doesn't +// perform any validation. // According to our measurements this is significantly faster // in codepaths that matter at high scale. func (ls Set) AsSelectorPreValidated() Selector { diff --git a/vendor/k8s.io/apimachinery/pkg/labels/selector.go b/vendor/k8s.io/apimachinery/pkg/labels/selector.go index 2f8e1e2b..bf62f98a 100644 --- a/vendor/k8s.io/apimachinery/pkg/labels/selector.go +++ b/vendor/k8s.io/apimachinery/pkg/labels/selector.go @@ -26,7 +26,7 @@ import ( "k8s.io/apimachinery/pkg/selection" "k8s.io/apimachinery/pkg/util/sets" "k8s.io/apimachinery/pkg/util/validation" - "k8s.io/klog" + "k8s.io/klog/v2" ) // Requirements is AND of all requirements. @@ -68,13 +68,15 @@ func Everything() Selector { type nothingSelector struct{} -func (n nothingSelector) Matches(_ Labels) bool { return false } -func (n nothingSelector) Empty() bool { return false } -func (n nothingSelector) String() string { return "" } -func (n nothingSelector) Add(_ ...Requirement) Selector { return n } -func (n nothingSelector) Requirements() (Requirements, bool) { return nil, false } -func (n nothingSelector) DeepCopySelector() Selector { return n } -func (n nothingSelector) RequiresExactMatch(label string) (value string, found bool) { return "", false } +func (n nothingSelector) Matches(_ Labels) bool { return false } +func (n nothingSelector) Empty() bool { return false } +func (n nothingSelector) String() string { return "" } +func (n nothingSelector) Add(_ ...Requirement) Selector { return n } +func (n nothingSelector) Requirements() (Requirements, bool) { return nil, false } +func (n nothingSelector) DeepCopySelector() Selector { return n } +func (n nothingSelector) RequiresExactMatch(label string) (value string, found bool) { + return "", false +} // Nothing returns a selector that matches no labels func Nothing() Selector { @@ -221,7 +223,7 @@ func (r *Requirement) Matches(ls Labels) bool { return false } - // There should be only one strValue in r.strValues, and can be converted to a integer. + // There should be only one strValue in r.strValues, and can be converted to an integer. if len(r.strValues) != 1 { klog.V(10).Infof("Invalid values count %+v of requirement %#v, for 'Gt', 'Lt' operators, exactly one value is required", len(r.strValues), r) return false @@ -869,23 +871,30 @@ func validateLabelValue(k, v string) error { // SelectorFromSet returns a Selector which will match exactly the given Set. A // nil and empty Sets are considered equivalent to Everything(). +// It does not perform any validation, which means the server will reject +// the request if the Set contains invalid values. func SelectorFromSet(ls Set) Selector { + return SelectorFromValidatedSet(ls) +} + +// ValidatedSelectorFromSet returns a Selector which will match exactly the given Set. A +// nil and empty Sets are considered equivalent to Everything(). +// The Set is validated client-side, which allows to catch errors early. +func ValidatedSelectorFromSet(ls Set) (Selector, error) { if ls == nil || len(ls) == 0 { - return internalSelector{} + return internalSelector{}, nil } requirements := make([]Requirement, 0, len(ls)) for label, value := range ls { r, err := NewRequirement(label, selection.Equals, []string{value}) - if err == nil { - requirements = append(requirements, *r) - } else { - //TODO: double check errors when input comes from serialization? - return internalSelector{} + if err != nil { + return nil, err } + requirements = append(requirements, *r) } // sort to have deterministic string representation sort.Sort(ByKey(requirements)) - return internalSelector(requirements) + return internalSelector(requirements), nil } // SelectorFromValidatedSet returns a Selector which will match exactly the given Set. diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/codec.go b/vendor/k8s.io/apimachinery/pkg/runtime/codec.go index 0bccf9dd..a9286313 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/codec.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/codec.go @@ -29,7 +29,7 @@ import ( "k8s.io/apimachinery/pkg/conversion/queryparams" "k8s.io/apimachinery/pkg/runtime/schema" - "k8s.io/klog" + "k8s.io/klog/v2" ) // codec binds an encoder and decoder. diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/codec_check.go b/vendor/k8s.io/apimachinery/pkg/runtime/codec_check.go index 510444a4..00022806 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/codec_check.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/codec_check.go @@ -21,6 +21,7 @@ import ( "reflect" "k8s.io/apimachinery/pkg/runtime/schema" + "k8s.io/apimachinery/pkg/util/json" ) // CheckCodec makes sure that the codec can encode objects like internalType, @@ -32,7 +33,14 @@ func CheckCodec(c Codec, internalType Object, externalTypes ...schema.GroupVersi return fmt.Errorf("Internal type not encodable: %v", err) } for _, et := range externalTypes { - exBytes := []byte(fmt.Sprintf(`{"kind":"%v","apiVersion":"%v"}`, et.Kind, et.GroupVersion().String())) + typeMeta := TypeMeta{ + Kind: et.Kind, + APIVersion: et.GroupVersion().String(), + } + exBytes, err := json.Marshal(&typeMeta) + if err != nil { + return err + } obj, err := Decode(c, exBytes) if err != nil { return fmt.Errorf("external type %s not interpretable: %v", et, err) diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/conversion.go b/vendor/k8s.io/apimachinery/pkg/runtime/conversion.go index 0947dce7..d04d701f 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/conversion.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/conversion.go @@ -53,14 +53,6 @@ func JSONKeyMapper(key string, sourceTag, destTag reflect.StructTag) (string, st return key, key } -// DefaultStringConversions are helpers for converting []string and string to real values. -var DefaultStringConversions = []interface{}{ - Convert_Slice_string_To_string, - Convert_Slice_string_To_int, - Convert_Slice_string_To_bool, - Convert_Slice_string_To_int64, -} - func Convert_Slice_string_To_string(in *[]string, out *string, s conversion.Scope) error { if len(*in) == 0 { *out = "" @@ -178,3 +170,27 @@ func Convert_Slice_string_To_Pointer_int64(in *[]string, out **int64, s conversi *out = &i return nil } + +func RegisterStringConversions(s *Scheme) error { + if err := s.AddConversionFunc((*[]string)(nil), (*string)(nil), func(a, b interface{}, scope conversion.Scope) error { + return Convert_Slice_string_To_string(a.(*[]string), b.(*string), scope) + }); err != nil { + return err + } + if err := s.AddConversionFunc((*[]string)(nil), (*int)(nil), func(a, b interface{}, scope conversion.Scope) error { + return Convert_Slice_string_To_int(a.(*[]string), b.(*int), scope) + }); err != nil { + return err + } + if err := s.AddConversionFunc((*[]string)(nil), (*bool)(nil), func(a, b interface{}, scope conversion.Scope) error { + return Convert_Slice_string_To_bool(a.(*[]string), b.(*bool), scope) + }); err != nil { + return err + } + if err := s.AddConversionFunc((*[]string)(nil), (*int64)(nil), func(a, b interface{}, scope conversion.Scope) error { + return Convert_Slice_string_To_int64(a.(*[]string), b.(*int64), scope) + }); err != nil { + return err + } + return nil +} diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/converter.go b/vendor/k8s.io/apimachinery/pkg/runtime/converter.go index b3e8a53b..871e4c8c 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/converter.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/converter.go @@ -17,7 +17,6 @@ limitations under the License. package runtime import ( - "bytes" encodingjson "encoding/json" "fmt" "math" @@ -32,8 +31,9 @@ import ( "k8s.io/apimachinery/pkg/conversion" "k8s.io/apimachinery/pkg/util/json" utilruntime "k8s.io/apimachinery/pkg/util/runtime" + "sigs.k8s.io/structured-merge-diff/v4/value" - "k8s.io/klog" + "k8s.io/klog/v2" ) // UnstructuredConverter is an interface for converting between interface{} @@ -68,13 +68,8 @@ func newFieldsCache() *fieldsCache { } var ( - marshalerType = reflect.TypeOf(new(encodingjson.Marshaler)).Elem() - unmarshalerType = reflect.TypeOf(new(encodingjson.Unmarshaler)).Elem() mapStringInterfaceType = reflect.TypeOf(map[string]interface{}{}) stringType = reflect.TypeOf(string("")) - int64Type = reflect.TypeOf(int64(0)) - float64Type = reflect.TypeOf(float64(0)) - boolType = reflect.TypeOf(bool(false)) fieldCache = newFieldsCache() // DefaultUnstructuredConverter performs unstructured to Go typed object conversions. @@ -208,13 +203,9 @@ func fromUnstructured(sv, dv reflect.Value) error { } // Check if the object has a custom JSON marshaller/unmarshaller. - if reflect.PtrTo(dt).Implements(unmarshalerType) { - data, err := json.Marshal(sv.Interface()) - if err != nil { - return fmt.Errorf("error encoding %s to json: %v", st.String(), err) - } - unmarshaler := dv.Addr().Interface().(encodingjson.Unmarshaler) - return unmarshaler.UnmarshalJSON(data) + entry := value.TypeReflectEntryOf(dv.Type()) + if entry.CanConvertFromUnstructured() { + return entry.FromUnstructured(sv, dv) } switch dt.Kind() { @@ -256,6 +247,7 @@ func fieldInfoFromField(structType reflect.Type, field int) *fieldInfo { for i := range items { if items[i] == "omitempty" { info.omitempty = true + break } } } @@ -483,112 +475,28 @@ func toUnstructuredViaJSON(obj interface{}, u *map[string]interface{}) error { return json.Unmarshal(data, u) } -var ( - nullBytes = []byte("null") - trueBytes = []byte("true") - falseBytes = []byte("false") -) - -func getMarshaler(v reflect.Value) (encodingjson.Marshaler, bool) { - // Check value receivers if v is not a pointer and pointer receivers if v is a pointer - if v.Type().Implements(marshalerType) { - return v.Interface().(encodingjson.Marshaler), true - } - // Check pointer receivers if v is not a pointer - if v.Kind() != reflect.Ptr && v.CanAddr() { - v = v.Addr() - if v.Type().Implements(marshalerType) { - return v.Interface().(encodingjson.Marshaler), true - } - } - return nil, false -} - func toUnstructured(sv, dv reflect.Value) error { - // Check if the object has a custom JSON marshaller/unmarshaller. - if marshaler, ok := getMarshaler(sv); ok { - if sv.Kind() == reflect.Ptr && sv.IsNil() { - // We're done - we don't need to store anything. - return nil - } - - data, err := marshaler.MarshalJSON() + // Check if the object has a custom string converter. + entry := value.TypeReflectEntryOf(sv.Type()) + if entry.CanConvertToUnstructured() { + v, err := entry.ToUnstructured(sv) if err != nil { return err } - switch { - case len(data) == 0: - return fmt.Errorf("error decoding from json: empty value") - - case bytes.Equal(data, nullBytes): - // We're done - we don't need to store anything. - - case bytes.Equal(data, trueBytes): - dv.Set(reflect.ValueOf(true)) - - case bytes.Equal(data, falseBytes): - dv.Set(reflect.ValueOf(false)) - - case data[0] == '"': - var result string - err := json.Unmarshal(data, &result) - if err != nil { - return fmt.Errorf("error decoding string from json: %v", err) - } - dv.Set(reflect.ValueOf(result)) - - case data[0] == '{': - result := make(map[string]interface{}) - err := json.Unmarshal(data, &result) - if err != nil { - return fmt.Errorf("error decoding object from json: %v", err) - } - dv.Set(reflect.ValueOf(result)) - - case data[0] == '[': - result := make([]interface{}, 0) - err := json.Unmarshal(data, &result) - if err != nil { - return fmt.Errorf("error decoding array from json: %v", err) - } - dv.Set(reflect.ValueOf(result)) - - default: - var ( - resultInt int64 - resultFloat float64 - err error - ) - if err = json.Unmarshal(data, &resultInt); err == nil { - dv.Set(reflect.ValueOf(resultInt)) - } else if err = json.Unmarshal(data, &resultFloat); err == nil { - dv.Set(reflect.ValueOf(resultFloat)) - } else { - return fmt.Errorf("error decoding number from json: %v", err) - } + if v != nil { + dv.Set(reflect.ValueOf(v)) } - return nil } - - st, dt := sv.Type(), dv.Type() + st := sv.Type() switch st.Kind() { case reflect.String: - if dt.Kind() == reflect.Interface && dv.NumMethod() == 0 { - dv.Set(reflect.New(stringType)) - } dv.Set(reflect.ValueOf(sv.String())) return nil case reflect.Bool: - if dt.Kind() == reflect.Interface && dv.NumMethod() == 0 { - dv.Set(reflect.New(boolType)) - } dv.Set(reflect.ValueOf(sv.Bool())) return nil case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: - if dt.Kind() == reflect.Interface && dv.NumMethod() == 0 { - dv.Set(reflect.New(int64Type)) - } dv.Set(reflect.ValueOf(sv.Int())) return nil case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64: @@ -596,15 +504,9 @@ func toUnstructured(sv, dv reflect.Value) error { if uVal > math.MaxInt64 { return fmt.Errorf("unsigned value %d does not fit into int64 (overflow)", uVal) } - if dt.Kind() == reflect.Interface && dv.NumMethod() == 0 { - dv.Set(reflect.New(int64Type)) - } dv.Set(reflect.ValueOf(int64(uVal))) return nil case reflect.Float32, reflect.Float64: - if dt.Kind() == reflect.Interface && dv.NumMethod() == 0 { - dv.Set(reflect.New(float64Type)) - } dv.Set(reflect.ValueOf(sv.Float())) return nil case reflect.Map: diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/embedded.go b/vendor/k8s.io/apimachinery/pkg/runtime/embedded.go index db11eb8b..7251e65f 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/embedded.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/embedded.go @@ -134,9 +134,16 @@ func Convert_runtime_RawExtension_To_runtime_Object(in *RawExtension, out *Objec return nil } -func DefaultEmbeddedConversions() []interface{} { - return []interface{}{ - Convert_runtime_Object_To_runtime_RawExtension, - Convert_runtime_RawExtension_To_runtime_Object, +func RegisterEmbeddedConversions(s *Scheme) error { + if err := s.AddConversionFunc((*Object)(nil), (*RawExtension)(nil), func(a, b interface{}, scope conversion.Scope) error { + return Convert_runtime_Object_To_runtime_RawExtension(a.(*Object), b.(*RawExtension), scope) + }); err != nil { + return err } + if err := s.AddConversionFunc((*RawExtension)(nil), (*Object)(nil), func(a, b interface{}, scope conversion.Scope) error { + return Convert_runtime_RawExtension_To_runtime_Object(a.(*RawExtension), b.(*Object), scope) + }); err != nil { + return err + } + return nil } diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/generated.pb.go b/vendor/k8s.io/apimachinery/pkg/runtime/generated.pb.go index af2f076b..07197181 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/generated.pb.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/generated.pb.go @@ -40,7 +40,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *RawExtension) Reset() { *m = RawExtension{} } func (*RawExtension) ProtoMessage() {} @@ -772,6 +772,7 @@ func (m *Unknown) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -803,10 +804,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -827,55 +826,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/schema/generated.pb.go b/vendor/k8s.io/apimachinery/pkg/runtime/schema/generated.pb.go index a7276649..29d3ac45 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/schema/generated.pb.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/schema/generated.pb.go @@ -36,7 +36,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func init() { proto.RegisterFile("k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/runtime/schema/generated.proto", fileDescriptor_0462724132518e0d) diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/schema/group_version.go b/vendor/k8s.io/apimachinery/pkg/runtime/schema/group_version.go index 63610331..994a3e3f 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/schema/group_version.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/schema/group_version.go @@ -176,15 +176,6 @@ func (gv GroupVersion) Empty() bool { // String puts "group" and "version" into a single "group/version" string. For the legacy v1 // it returns "v1". func (gv GroupVersion) String() string { - // special case the internal apiVersion for the legacy kube types - if gv.Empty() { - return "" - } - - // special case of "v1" for backward compatibility - if len(gv.Group) == 0 && gv.Version == "v1" { - return gv.Version - } if len(gv.Group) > 0 { return gv.Group + "/" + gv.Version } @@ -252,10 +243,10 @@ func (gv GroupVersion) WithResource(resource string) GroupVersionResource { type GroupVersions []GroupVersion // Identifier implements runtime.GroupVersioner interface. -func (gv GroupVersions) Identifier() string { - groupVersions := make([]string, 0, len(gv)) - for i := range gv { - groupVersions = append(groupVersions, gv[i].String()) +func (gvs GroupVersions) Identifier() string { + groupVersions := make([]string, 0, len(gvs)) + for i := range gvs { + groupVersions = append(groupVersions, gvs[i].String()) } return fmt.Sprintf("[%s]", strings.Join(groupVersions, ",")) } diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/scheme.go b/vendor/k8s.io/apimachinery/pkg/runtime/scheme.go index fd37e293..3b254961 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/scheme.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/scheme.go @@ -102,10 +102,10 @@ func NewScheme() *Scheme { } s.converter = conversion.NewConverter(s.nameFunc) - utilruntime.Must(s.AddConversionFuncs(DefaultEmbeddedConversions()...)) + // Enable couple default conversions by default. + utilruntime.Must(RegisterEmbeddedConversions(s)) + utilruntime.Must(RegisterStringConversions(s)) - // Enable map[string][]string conversions by default - utilruntime.Must(s.AddConversionFuncs(DefaultStringConversions...)) utilruntime.Must(s.RegisterInputDefaults(&map[string][]string{}, JSONKeyMapper, conversion.AllowDifferentFieldTypeNames|conversion.IgnoreMissingFields)) utilruntime.Must(s.RegisterInputDefaults(&url.Values{}, JSONKeyMapper, conversion.AllowDifferentFieldTypeNames|conversion.IgnoreMissingFields)) return s @@ -211,6 +211,19 @@ func (s *Scheme) AddKnownTypeWithName(gvk schema.GroupVersionKind, obj Object) { } } s.typeToGVK[t] = append(s.typeToGVK[t], gvk) + + // if the type implements DeepCopyInto(), register a self-conversion + if m := reflect.ValueOf(obj).MethodByName("DeepCopyInto"); m.IsValid() && m.Type().NumIn() == 1 && m.Type().NumOut() == 0 && m.Type().In(0) == reflect.TypeOf(obj) { + if err := s.AddGeneratedConversionFunc(obj, obj, func(a, b interface{}, scope conversion.Scope) error { + // copy a to b + reflect.ValueOf(a).MethodByName("DeepCopyInto").Call([]reflect.Value{reflect.ValueOf(b)}) + // clear TypeMeta to match legacy reflective conversion + b.(Object).GetObjectKind().SetGroupVersionKind(schema.GroupVersionKind{}) + return nil + }); err != nil { + panic(err) + } + } } // KnownTypes returns the types known for the given version. @@ -308,45 +321,6 @@ func (s *Scheme) AddIgnoredConversionType(from, to interface{}) error { return s.converter.RegisterIgnoredConversion(from, to) } -// AddConversionFuncs adds functions to the list of conversion functions. The given -// functions should know how to convert between two of your API objects, or their -// sub-objects. We deduce how to call these functions from the types of their two -// parameters; see the comment for Converter.Register. -// -// Note that, if you need to copy sub-objects that didn't change, you can use the -// conversion.Scope object that will be passed to your conversion function. -// Additionally, all conversions started by Scheme will set the SrcVersion and -// DestVersion fields on the Meta object. Example: -// -// s.AddConversionFuncs( -// func(in *InternalObject, out *ExternalObject, scope conversion.Scope) error { -// // You can depend on Meta() being non-nil, and this being set to -// // the source version, e.g., "" -// s.Meta().SrcVersion -// // You can depend on this being set to the destination version, -// // e.g., "v1". -// s.Meta().DestVersion -// // Call scope.Convert to copy sub-fields. -// s.Convert(&in.SubFieldThatMoved, &out.NewLocation.NewName, 0) -// return nil -// }, -// ) -// -// (For more detail about conversion functions, see Converter.Register's comment.) -// -// Also note that the default behavior, if you don't add a conversion function, is to -// sanely copy fields that have the same names and same type names. It's OK if the -// destination type has extra fields, but it must not remove any. So you only need to -// add conversion functions for things with changed/removed fields. -func (s *Scheme) AddConversionFuncs(conversionFuncs ...interface{}) error { - for _, f := range conversionFuncs { - if err := s.converter.RegisterConversionFunc(f); err != nil { - return err - } - } - return nil -} - // AddConversionFunc registers a function that converts between a and b by passing objects of those // types to the provided function. The function *must* accept objects of a and b - this machinery will not enforce // any other guarantee. diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/serializer/json/json.go b/vendor/k8s.io/apimachinery/pkg/runtime/serializer/json/json.go index 9d17f09e..e081d7ff 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/serializer/json/json.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/serializer/json/json.go @@ -31,7 +31,7 @@ import ( "k8s.io/apimachinery/pkg/runtime/serializer/recognizer" "k8s.io/apimachinery/pkg/util/framer" utilyaml "k8s.io/apimachinery/pkg/util/yaml" - "k8s.io/klog" + "k8s.io/klog/v2" ) // NewSerializer creates a JSON serializer that handles encoding versioned objects into the proper JSON form. If typer diff --git a/vendor/k8s.io/apimachinery/pkg/runtime/serializer/versioning/versioning.go b/vendor/k8s.io/apimachinery/pkg/runtime/serializer/versioning/versioning.go index ced184c9..718c5dfb 100644 --- a/vendor/k8s.io/apimachinery/pkg/runtime/serializer/versioning/versioning.go +++ b/vendor/k8s.io/apimachinery/pkg/runtime/serializer/versioning/versioning.go @@ -25,7 +25,7 @@ import ( "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/runtime/schema" - "k8s.io/klog" + "k8s.io/klog/v2" ) // NewDefaultingCodecForScheme is a convenience method for callers that are using a scheme. diff --git a/vendor/k8s.io/apimachinery/pkg/util/clock/clock.go b/vendor/k8s.io/apimachinery/pkg/util/clock/clock.go index 1689e62e..6cf13d83 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/clock/clock.go +++ b/vendor/k8s.io/apimachinery/pkg/util/clock/clock.go @@ -52,23 +52,26 @@ func (RealClock) Since(ts time.Time) time.Duration { return time.Since(ts) } -// Same as time.After(d). +// After is the same as time.After(d). func (RealClock) After(d time.Duration) <-chan time.Time { return time.After(d) } +// NewTimer returns a new Timer. func (RealClock) NewTimer(d time.Duration) Timer { return &realTimer{ timer: time.NewTimer(d), } } +// NewTicker returns a new Ticker. func (RealClock) NewTicker(d time.Duration) Ticker { return &realTicker{ ticker: time.NewTicker(d), } } +// Sleep pauses the RealClock for duration d. func (RealClock) Sleep(d time.Duration) { time.Sleep(d) } @@ -94,12 +97,14 @@ type fakeClockWaiter struct { destChan chan time.Time } +// NewFakePassiveClock returns a new FakePassiveClock. func NewFakePassiveClock(t time.Time) *FakePassiveClock { return &FakePassiveClock{ time: t, } } +// NewFakeClock returns a new FakeClock func NewFakeClock(t time.Time) *FakeClock { return &FakeClock{ FakePassiveClock: *NewFakePassiveClock(t), @@ -120,14 +125,14 @@ func (f *FakePassiveClock) Since(ts time.Time) time.Duration { return f.time.Sub(ts) } -// Sets the time. +// SetTime sets the time on the FakePassiveClock. func (f *FakePassiveClock) SetTime(t time.Time) { f.lock.Lock() defer f.lock.Unlock() f.time = t } -// Fake version of time.After(d). +// After is the Fake version of time.After(d). func (f *FakeClock) After(d time.Duration) <-chan time.Time { f.lock.Lock() defer f.lock.Unlock() @@ -140,7 +145,7 @@ func (f *FakeClock) After(d time.Duration) <-chan time.Time { return ch } -// Fake version of time.NewTimer(d). +// NewTimer is the Fake version of time.NewTimer(d). func (f *FakeClock) NewTimer(d time.Duration) Timer { f.lock.Lock() defer f.lock.Unlock() @@ -157,6 +162,7 @@ func (f *FakeClock) NewTimer(d time.Duration) Timer { return timer } +// NewTicker returns a new Ticker. func (f *FakeClock) NewTicker(d time.Duration) Ticker { f.lock.Lock() defer f.lock.Unlock() @@ -174,14 +180,14 @@ func (f *FakeClock) NewTicker(d time.Duration) Ticker { } } -// Move clock by Duration, notify anyone that's called After, Tick, or NewTimer +// Step moves clock by Duration, notifies anyone that's called After, Tick, or NewTimer func (f *FakeClock) Step(d time.Duration) { f.lock.Lock() defer f.lock.Unlock() f.setTimeLocked(f.time.Add(d)) } -// Sets the time. +// SetTime sets the time on a FakeClock. func (f *FakeClock) SetTime(t time.Time) { f.lock.Lock() defer f.lock.Unlock() @@ -219,7 +225,7 @@ func (f *FakeClock) setTimeLocked(t time.Time) { f.waiters = newWaiters } -// Returns true if After has been called on f but not yet satisfied (so you can +// HasWaiters returns true if After has been called on f but not yet satisfied (so you can // write race-free tests). func (f *FakeClock) HasWaiters() bool { f.lock.RLock() @@ -227,6 +233,7 @@ func (f *FakeClock) HasWaiters() bool { return len(f.waiters) > 0 } +// Sleep pauses the FakeClock for duration d. func (f *FakeClock) Sleep(d time.Duration) { f.Step(d) } @@ -248,24 +255,25 @@ func (i *IntervalClock) Since(ts time.Time) time.Duration { return i.Time.Sub(ts) } -// Unimplemented, will panic. +// After is currently unimplemented, will panic. // TODO: make interval clock use FakeClock so this can be implemented. func (*IntervalClock) After(d time.Duration) <-chan time.Time { panic("IntervalClock doesn't implement After") } -// Unimplemented, will panic. +// NewTimer is currently unimplemented, will panic. // TODO: make interval clock use FakeClock so this can be implemented. func (*IntervalClock) NewTimer(d time.Duration) Timer { panic("IntervalClock doesn't implement NewTimer") } -// Unimplemented, will panic. +// NewTicker is currently unimplemented, will panic. // TODO: make interval clock use FakeClock so this can be implemented. func (*IntervalClock) NewTicker(d time.Duration) Ticker { panic("IntervalClock doesn't implement NewTicker") } +// Sleep is currently unimplemented; will panic. func (*IntervalClock) Sleep(d time.Duration) { panic("IntervalClock doesn't implement Sleep") } @@ -355,6 +363,7 @@ func (f *fakeTimer) Reset(d time.Duration) bool { return false } +// Ticker defines the Ticker interface type Ticker interface { C() <-chan time.Time Stop() diff --git a/vendor/k8s.io/apimachinery/pkg/util/errors/errors.go b/vendor/k8s.io/apimachinery/pkg/util/errors/errors.go index 62a73f34..5bafc218 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/errors/errors.go +++ b/vendor/k8s.io/apimachinery/pkg/util/errors/errors.go @@ -28,9 +28,14 @@ type MessageCountMap map[string]int // Aggregate represents an object that contains multiple errors, but does not // necessarily have singular semantic meaning. +// The aggregate can be used with `errors.Is()` to check for the occurrence of +// a specific error type. +// Errors.As() is not supported, because the caller presumably cares about a +// specific error of potentially multiple that match the given type. type Aggregate interface { error Errors() []error + Is(error) bool } // NewAggregate converts a slice of errors into an Aggregate interface, which @@ -71,16 +76,17 @@ func (agg aggregate) Error() string { } seenerrs := sets.NewString() result := "" - agg.visit(func(err error) { + agg.visit(func(err error) bool { msg := err.Error() if seenerrs.Has(msg) { - return + return false } seenerrs.Insert(msg) if len(seenerrs) > 1 { result += ", " } result += msg + return false }) if len(seenerrs) == 1 { return result @@ -88,19 +94,33 @@ func (agg aggregate) Error() string { return "[" + result + "]" } -func (agg aggregate) visit(f func(err error)) { +func (agg aggregate) Is(target error) bool { + return agg.visit(func(err error) bool { + return errors.Is(err, target) + }) +} + +func (agg aggregate) visit(f func(err error) bool) bool { for _, err := range agg { switch err := err.(type) { case aggregate: - err.visit(f) + if match := err.visit(f); match { + return match + } case Aggregate: for _, nestedErr := range err.Errors() { - f(nestedErr) + if match := f(nestedErr); match { + return match + } } default: - f(err) + if match := f(err); match { + return match + } } } + + return false } // Errors is part of the Aggregate interface. diff --git a/vendor/k8s.io/apimachinery/pkg/util/intstr/generated.pb.go b/vendor/k8s.io/apimachinery/pkg/util/intstr/generated.pb.go index 64cbc770..ec1cb70f 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/intstr/generated.pb.go +++ b/vendor/k8s.io/apimachinery/pkg/util/intstr/generated.pb.go @@ -38,7 +38,7 @@ var _ = math.Inf // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. -const _ = proto.GoGoProtoPackageIsVersion2 // please upgrade the proto package +const _ = proto.GoGoProtoPackageIsVersion3 // please upgrade the proto package func (m *IntOrString) Reset() { *m = IntOrString{} } func (*IntOrString) ProtoMessage() {} @@ -289,6 +289,7 @@ func (m *IntOrString) Unmarshal(dAtA []byte) error { func skipGenerated(dAtA []byte) (n int, err error) { l := len(dAtA) iNdEx := 0 + depth := 0 for iNdEx < l { var wire uint64 for shift := uint(0); ; shift += 7 { @@ -320,10 +321,8 @@ func skipGenerated(dAtA []byte) (n int, err error) { break } } - return iNdEx, nil case 1: iNdEx += 8 - return iNdEx, nil case 2: var length int for shift := uint(0); ; shift += 7 { @@ -344,55 +343,30 @@ func skipGenerated(dAtA []byte) (n int, err error) { return 0, ErrInvalidLengthGenerated } iNdEx += length - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - return iNdEx, nil case 3: - for { - var innerWire uint64 - var start int = iNdEx - for shift := uint(0); ; shift += 7 { - if shift >= 64 { - return 0, ErrIntOverflowGenerated - } - if iNdEx >= l { - return 0, io.ErrUnexpectedEOF - } - b := dAtA[iNdEx] - iNdEx++ - innerWire |= (uint64(b) & 0x7F) << shift - if b < 0x80 { - break - } - } - innerWireType := int(innerWire & 0x7) - if innerWireType == 4 { - break - } - next, err := skipGenerated(dAtA[start:]) - if err != nil { - return 0, err - } - iNdEx = start + next - if iNdEx < 0 { - return 0, ErrInvalidLengthGenerated - } - } - return iNdEx, nil + depth++ case 4: - return iNdEx, nil + if depth == 0 { + return 0, ErrUnexpectedEndOfGroupGenerated + } + depth-- case 5: iNdEx += 4 - return iNdEx, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } + if iNdEx < 0 { + return 0, ErrInvalidLengthGenerated + } + if depth == 0 { + return iNdEx, nil + } } - panic("unreachable") + return 0, io.ErrUnexpectedEOF } var ( - ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") - ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrInvalidLengthGenerated = fmt.Errorf("proto: negative length found during unmarshaling") + ErrIntOverflowGenerated = fmt.Errorf("proto: integer overflow") + ErrUnexpectedEndOfGroupGenerated = fmt.Errorf("proto: unexpected end of group") ) diff --git a/vendor/k8s.io/apimachinery/pkg/util/intstr/intstr.go b/vendor/k8s.io/apimachinery/pkg/util/intstr/intstr.go index 2df62955..6576def8 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/intstr/intstr.go +++ b/vendor/k8s.io/apimachinery/pkg/util/intstr/intstr.go @@ -26,7 +26,7 @@ import ( "strings" "github.com/google/gofuzz" - "k8s.io/klog" + "k8s.io/klog/v2" ) // IntOrString is a type that can hold an int32 or a string. When used in @@ -97,7 +97,8 @@ func (intstr *IntOrString) String() string { } // IntValue returns the IntVal if type Int, or if -// it is a String, will attempt a conversion to int. +// it is a String, will attempt a conversion to int, +// returning 0 if a parsing error occurs. func (intstr *IntOrString) IntValue() int { if intstr.Type == String { i, _ := strconv.Atoi(intstr.StrVal) diff --git a/vendor/k8s.io/apimachinery/pkg/util/json/json.go b/vendor/k8s.io/apimachinery/pkg/util/json/json.go index 0e2e3017..20483488 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/json/json.go +++ b/vendor/k8s.io/apimachinery/pkg/util/json/json.go @@ -66,11 +66,36 @@ func Unmarshal(data []byte, v interface{}) error { // If the decode succeeds, post-process the map to convert json.Number objects to int64 or float64 return convertSliceNumbers(*v, 0) + case *interface{}: + // Build a decoder from the given data + decoder := json.NewDecoder(bytes.NewBuffer(data)) + // Preserve numbers, rather than casting to float64 automatically + decoder.UseNumber() + // Run the decode + if err := decoder.Decode(v); err != nil { + return err + } + // If the decode succeeds, post-process the map to convert json.Number objects to int64 or float64 + return convertInterfaceNumbers(v, 0) + default: return json.Unmarshal(data, v) } } +func convertInterfaceNumbers(v *interface{}, depth int) error { + var err error + switch v2 := (*v).(type) { + case json.Number: + *v, err = convertNumber(v2) + case map[string]interface{}: + err = convertMapNumbers(v2, depth+1) + case []interface{}: + err = convertSliceNumbers(v2, depth+1) + } + return err +} + // convertMapNumbers traverses the map, converting any json.Number values to int64 or float64. // values which are map[string]interface{} or []interface{} are recursively visited func convertMapNumbers(m map[string]interface{}, depth int) error { diff --git a/vendor/k8s.io/apimachinery/pkg/util/net/http.go b/vendor/k8s.io/apimachinery/pkg/util/net/http.go index f9540c63..945886c4 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/net/http.go +++ b/vendor/k8s.io/apimachinery/pkg/util/net/http.go @@ -21,18 +21,23 @@ import ( "bytes" "context" "crypto/tls" + "errors" "fmt" "io" + "mime" "net" "net/http" "net/url" "os" "path" + "regexp" "strconv" "strings" + "unicode" + "unicode/utf8" "golang.org/x/net/http2" - "k8s.io/klog" + "k8s.io/klog/v2" ) // JoinPreservingTrailingSlash does a path.Join of the specified elements, @@ -55,6 +60,15 @@ func JoinPreservingTrailingSlash(elem ...string) string { return result } +// IsTimeout returns true if the given error is a network timeout error +func IsTimeout(err error) bool { + var neterr net.Error + if errors.As(err, &neterr) { + return neterr != nil && neterr.Timeout() + } + return false +} + // IsProbableEOF returns true if the given error resembles a connection termination // scenario that would justify assuming that the watch is empty. // These errors are what the Go http stack returns back to us which are general @@ -65,13 +79,16 @@ func IsProbableEOF(err error) bool { if err == nil { return false } - if uerr, ok := err.(*url.Error); ok { + var uerr *url.Error + if errors.As(err, &uerr) { err = uerr.Err } msg := err.Error() switch { case err == io.EOF: return true + case err == io.ErrUnexpectedEOF: + return true case msg == "http: can't write HTTP request on broken connection": return true case strings.Contains(msg, "http2: server sent GOAWAY and closed the connection"): @@ -206,13 +223,17 @@ func GetHTTPClient(req *http.Request) string { return "unknown" } -// SourceIPs splits the comma separated X-Forwarded-For header or returns the X-Real-Ip header or req.RemoteAddr, -// in that order, ignoring invalid IPs. It returns nil if all of these are empty or invalid. +// SourceIPs splits the comma separated X-Forwarded-For header and joins it with +// the X-Real-Ip header and/or req.RemoteAddr, ignoring invalid IPs. +// The X-Real-Ip is omitted if it's already present in the X-Forwarded-For chain. +// The req.RemoteAddr is always the last IP in the returned list. +// It returns nil if all of these are empty or invalid. func SourceIPs(req *http.Request) []net.IP { + var srcIPs []net.IP + hdr := req.Header // First check the X-Forwarded-For header for requests via proxy. hdrForwardedFor := hdr.Get("X-Forwarded-For") - forwardedForIPs := []net.IP{} if hdrForwardedFor != "" { // X-Forwarded-For can be a csv of IPs in case of multiple proxies. // Use the first valid one. @@ -220,38 +241,49 @@ func SourceIPs(req *http.Request) []net.IP { for _, part := range parts { ip := net.ParseIP(strings.TrimSpace(part)) if ip != nil { - forwardedForIPs = append(forwardedForIPs, ip) + srcIPs = append(srcIPs, ip) } } } - if len(forwardedForIPs) > 0 { - return forwardedForIPs - } // Try the X-Real-Ip header. hdrRealIp := hdr.Get("X-Real-Ip") if hdrRealIp != "" { ip := net.ParseIP(hdrRealIp) - if ip != nil { - return []net.IP{ip} + // Only append the X-Real-Ip if it's not already contained in the X-Forwarded-For chain. + if ip != nil && !containsIP(srcIPs, ip) { + srcIPs = append(srcIPs, ip) } } - // Fallback to Remote Address in request, which will give the correct client IP when there is no proxy. + // Always include the request Remote Address as it cannot be easily spoofed. + var remoteIP net.IP // Remote Address in Go's HTTP server is in the form host:port so we need to split that first. host, _, err := net.SplitHostPort(req.RemoteAddr) if err == nil { - if remoteIP := net.ParseIP(host); remoteIP != nil { - return []net.IP{remoteIP} - } + remoteIP = net.ParseIP(host) } - // Fallback if Remote Address was just IP. - if remoteIP := net.ParseIP(req.RemoteAddr); remoteIP != nil { - return []net.IP{remoteIP} + if remoteIP == nil { + remoteIP = net.ParseIP(req.RemoteAddr) + } + + // Don't duplicate remote IP if it's already the last address in the chain. + if remoteIP != nil && (len(srcIPs) == 0 || !remoteIP.Equal(srcIPs[len(srcIPs)-1])) { + srcIPs = append(srcIPs, remoteIP) } - return nil + return srcIPs +} + +// Checks whether the given IP address is contained in the list of IPs. +func containsIP(ips []net.IP, ip net.IP) bool { + for _, v := range ips { + if v.Equal(ip) { + return true + } + } + return false } // Extracts and returns the clients IP from the given request. @@ -425,7 +457,7 @@ redirectLoop: // Only follow redirects to the same host. Otherwise, propagate the redirect response back. if requireSameHostRedirects && location.Hostname() != originalLocation.Hostname() { - break redirectLoop + return nil, nil, fmt.Errorf("hostname mismatch: expected %s, found %s", originalLocation.Hostname(), location.Hostname()) } // Reset the connection. @@ -461,3 +493,232 @@ func CloneHeader(in http.Header) http.Header { } return out } + +// WarningHeader contains a single RFC2616 14.46 warnings header +type WarningHeader struct { + // Codeindicates the type of warning. 299 is a miscellaneous persistent warning + Code int + // Agent contains the name or pseudonym of the server adding the Warning header. + // A single "-" is recommended when agent is unknown. + Agent string + // Warning text + Text string +} + +// ParseWarningHeaders extract RFC2616 14.46 warnings headers from the specified set of header values. +// Multiple comma-separated warnings per header are supported. +// If errors are encountered on a header, the remainder of that header are skipped and subsequent headers are parsed. +// Returns successfully parsed warnings and any errors encountered. +func ParseWarningHeaders(headers []string) ([]WarningHeader, []error) { + var ( + results []WarningHeader + errs []error + ) + for _, header := range headers { + for len(header) > 0 { + result, remainder, err := ParseWarningHeader(header) + if err != nil { + errs = append(errs, err) + break + } + results = append(results, result) + header = remainder + } + } + return results, errs +} + +var ( + codeMatcher = regexp.MustCompile(`^[0-9]{3}$`) + wordDecoder = &mime.WordDecoder{} +) + +// ParseWarningHeader extracts one RFC2616 14.46 warning from the specified header, +// returning an error if the header does not contain a correctly formatted warning. +// Any remaining content in the header is returned. +func ParseWarningHeader(header string) (result WarningHeader, remainder string, err error) { + // https://tools.ietf.org/html/rfc2616#section-14.46 + // updated by + // https://tools.ietf.org/html/rfc7234#section-5.5 + // https://tools.ietf.org/html/rfc7234#appendix-A + // Some requirements regarding production and processing of the Warning + // header fields have been relaxed, as it is not widely implemented. + // Furthermore, the Warning header field no longer uses RFC 2047 + // encoding, nor does it allow multiple languages, as these aspects were + // not implemented. + // + // Format is one of: + // warn-code warn-agent "warn-text" + // warn-code warn-agent "warn-text" "warn-date" + // + // warn-code is a three digit number + // warn-agent is unquoted and contains no spaces + // warn-text is quoted with backslash escaping (RFC2047-encoded according to RFC2616, not encoded according to RFC7234) + // warn-date is optional, quoted, and in HTTP-date format (no embedded or escaped quotes) + // + // additional warnings can optionally be included in the same header by comma-separating them: + // warn-code warn-agent "warn-text" "warn-date"[, warn-code warn-agent "warn-text" "warn-date", ...] + + // tolerate leading whitespace + header = strings.TrimSpace(header) + + parts := strings.SplitN(header, " ", 3) + if len(parts) != 3 { + return WarningHeader{}, "", errors.New("invalid warning header: fewer than 3 segments") + } + code, agent, textDateRemainder := parts[0], parts[1], parts[2] + + // verify code format + if !codeMatcher.Match([]byte(code)) { + return WarningHeader{}, "", errors.New("invalid warning header: code segment is not 3 digits between 100-299") + } + codeInt, _ := strconv.ParseInt(code, 10, 64) + + // verify agent presence + if len(agent) == 0 { + return WarningHeader{}, "", errors.New("invalid warning header: empty agent segment") + } + if !utf8.ValidString(agent) || hasAnyRunes(agent, unicode.IsControl) { + return WarningHeader{}, "", errors.New("invalid warning header: invalid agent") + } + + // verify textDateRemainder presence + if len(textDateRemainder) == 0 { + return WarningHeader{}, "", errors.New("invalid warning header: empty text segment") + } + + // extract text + text, dateAndRemainder, err := parseQuotedString(textDateRemainder) + if err != nil { + return WarningHeader{}, "", fmt.Errorf("invalid warning header: %v", err) + } + // tolerate RFC2047-encoded text from warnings produced according to RFC2616 + if decodedText, err := wordDecoder.DecodeHeader(text); err == nil { + text = decodedText + } + if !utf8.ValidString(text) || hasAnyRunes(text, unicode.IsControl) { + return WarningHeader{}, "", errors.New("invalid warning header: invalid text") + } + result = WarningHeader{Code: int(codeInt), Agent: agent, Text: text} + + if len(dateAndRemainder) > 0 { + if dateAndRemainder[0] == '"' { + // consume date + foundEndQuote := false + for i := 1; i < len(dateAndRemainder); i++ { + if dateAndRemainder[i] == '"' { + foundEndQuote = true + remainder = strings.TrimSpace(dateAndRemainder[i+1:]) + break + } + } + if !foundEndQuote { + return WarningHeader{}, "", errors.New("invalid warning header: unterminated date segment") + } + } else { + remainder = dateAndRemainder + } + } + if len(remainder) > 0 { + if remainder[0] == ',' { + // consume comma if present + remainder = strings.TrimSpace(remainder[1:]) + } else { + return WarningHeader{}, "", errors.New("invalid warning header: unexpected token after warn-date") + } + } + + return result, remainder, nil +} + +func parseQuotedString(quotedString string) (string, string, error) { + if len(quotedString) == 0 { + return "", "", errors.New("invalid quoted string: 0-length") + } + + if quotedString[0] != '"' { + return "", "", errors.New("invalid quoted string: missing initial quote") + } + + quotedString = quotedString[1:] + var remainder string + escaping := false + closedQuote := false + result := &bytes.Buffer{} +loop: + for i := 0; i < len(quotedString); i++ { + b := quotedString[i] + switch b { + case '"': + if escaping { + result.WriteByte(b) + escaping = false + } else { + closedQuote = true + remainder = strings.TrimSpace(quotedString[i+1:]) + break loop + } + case '\\': + if escaping { + result.WriteByte(b) + escaping = false + } else { + escaping = true + } + default: + result.WriteByte(b) + escaping = false + } + } + + if !closedQuote { + return "", "", errors.New("invalid quoted string: missing closing quote") + } + return result.String(), remainder, nil +} + +func NewWarningHeader(code int, agent, text string) (string, error) { + if code < 0 || code > 999 { + return "", errors.New("code must be between 0 and 999") + } + if len(agent) == 0 { + agent = "-" + } else if !utf8.ValidString(agent) || strings.ContainsAny(agent, `\"`) || hasAnyRunes(agent, unicode.IsSpace, unicode.IsControl) { + return "", errors.New("agent must be valid UTF-8 and must not contain spaces, quotes, backslashes, or control characters") + } + if !utf8.ValidString(text) || hasAnyRunes(text, unicode.IsControl) { + return "", errors.New("text must be valid UTF-8 and must not contain control characters") + } + return fmt.Sprintf("%03d %s %s", code, agent, makeQuotedString(text)), nil +} + +func hasAnyRunes(s string, runeCheckers ...func(rune) bool) bool { + for _, r := range s { + for _, checker := range runeCheckers { + if checker(r) { + return true + } + } + } + return false +} + +func makeQuotedString(s string) string { + result := &bytes.Buffer{} + // opening quote + result.WriteRune('"') + for _, c := range s { + switch c { + case '"', '\\': + // escape " and \ + result.WriteRune('\\') + result.WriteRune(c) + default: + // write everything else as-is + result.WriteRune(c) + } + } + // closing quote + result.WriteRune('"') + return result.String() +} diff --git a/vendor/k8s.io/apimachinery/pkg/util/net/interface.go b/vendor/k8s.io/apimachinery/pkg/util/net/interface.go index 836494d5..204e223c 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/net/interface.go +++ b/vendor/k8s.io/apimachinery/pkg/util/net/interface.go @@ -26,7 +26,7 @@ import ( "strings" - "k8s.io/klog" + "k8s.io/klog/v2" ) type AddressFamily uint diff --git a/vendor/k8s.io/apimachinery/pkg/util/net/util.go b/vendor/k8s.io/apimachinery/pkg/util/net/util.go index 2e7cb949..5950087e 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/net/util.go +++ b/vendor/k8s.io/apimachinery/pkg/util/net/util.go @@ -17,9 +17,8 @@ limitations under the License. package net import ( + "errors" "net" - "net/url" - "os" "reflect" "syscall" ) @@ -40,34 +39,18 @@ func IPNetEqual(ipnet1, ipnet2 *net.IPNet) bool { // Returns if the given err is "connection reset by peer" error. func IsConnectionReset(err error) bool { - if urlErr, ok := err.(*url.Error); ok { - err = urlErr.Err - } - if opErr, ok := err.(*net.OpError); ok { - err = opErr.Err - } - if osErr, ok := err.(*os.SyscallError); ok { - err = osErr.Err - } - if errno, ok := err.(syscall.Errno); ok && errno == syscall.ECONNRESET { - return true + var errno syscall.Errno + if errors.As(err, &errno) { + return errno == syscall.ECONNRESET } return false } // Returns if the given err is "connection refused" error func IsConnectionRefused(err error) bool { - if urlErr, ok := err.(*url.Error); ok { - err = urlErr.Err - } - if opErr, ok := err.(*net.OpError); ok { - err = opErr.Err - } - if osErr, ok := err.(*os.SyscallError); ok { - err = osErr.Err - } - if errno, ok := err.(syscall.Errno); ok && errno == syscall.ECONNREFUSED { - return true + var errno syscall.Errno + if errors.As(err, &errno) { + return errno == syscall.ECONNREFUSED } return false } diff --git a/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go b/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go index 1428443f..e8a9f609 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go +++ b/vendor/k8s.io/apimachinery/pkg/util/runtime/runtime.go @@ -23,7 +23,7 @@ import ( "sync" "time" - "k8s.io/klog" + "k8s.io/klog/v2" ) var ( diff --git a/vendor/k8s.io/apimachinery/pkg/util/strategicpatch/patch.go b/vendor/k8s.io/apimachinery/pkg/util/strategicpatch/patch.go index c55894e5..600f3bef 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/strategicpatch/patch.go +++ b/vendor/k8s.io/apimachinery/pkg/util/strategicpatch/patch.go @@ -1321,9 +1321,7 @@ func mergeMap(original, patch map[string]interface{}, schema LookupPatchMeta, me // Preserving the null value is useful when we want to send an explicit // delete to the API server. if patchV == nil { - if _, ok := original[k]; ok { - delete(original, k) - } + delete(original, k) if mergeOptions.IgnoreUnmatchedNulls { continue } diff --git a/vendor/k8s.io/apimachinery/pkg/util/validation/validation.go b/vendor/k8s.io/apimachinery/pkg/util/validation/validation.go index 8e1907c2..4752b29a 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/validation/validation.go +++ b/vendor/k8s.io/apimachinery/pkg/util/validation/validation.go @@ -106,9 +106,52 @@ func IsFullyQualifiedDomainName(fldPath *field.Path, name string) field.ErrorLis if len(strings.Split(name, ".")) < 2 { return append(allErrors, field.Invalid(fldPath, name, "should be a domain with at least two segments separated by dots")) } + for _, label := range strings.Split(name, ".") { + if errs := IsDNS1123Label(label); len(errs) > 0 { + return append(allErrors, field.Invalid(fldPath, label, strings.Join(errs, ","))) + } + } return allErrors } +// Allowed characters in an HTTP Path as defined by RFC 3986. A HTTP path may +// contain: +// * unreserved characters (alphanumeric, '-', '.', '_', '~') +// * percent-encoded octets +// * sub-delims ("!", "$", "&", "'", "(", ")", "*", "+", ",", ";", "=") +// * a colon character (":") +const httpPathFmt string = `[A-Za-z0-9/\-._~%!$&'()*+,;=:]+` + +var httpPathRegexp = regexp.MustCompile("^" + httpPathFmt + "$") + +// IsDomainPrefixedPath checks if the given string is a domain-prefixed path +// (e.g. acme.io/foo). All characters before the first "/" must be a valid +// subdomain as defined by RFC 1123. All characters trailing the first "/" must +// be valid HTTP Path characters as defined by RFC 3986. +func IsDomainPrefixedPath(fldPath *field.Path, dpPath string) field.ErrorList { + var allErrs field.ErrorList + if len(dpPath) == 0 { + return append(allErrs, field.Required(fldPath, "")) + } + + segments := strings.SplitN(dpPath, "/", 2) + if len(segments) != 2 || len(segments[0]) == 0 || len(segments[1]) == 0 { + return append(allErrs, field.Invalid(fldPath, dpPath, "must be a domain-prefixed path (such as \"acme.io/foo\")")) + } + + host := segments[0] + for _, err := range IsDNS1123Subdomain(host) { + allErrs = append(allErrs, field.Invalid(fldPath, host, err)) + } + + path := segments[1] + if !httpPathRegexp.MatchString(path) { + return append(allErrs, field.Invalid(fldPath, path, RegexError("Invalid path", httpPathFmt))) + } + + return allErrs +} + const labelValueFmt string = "(" + qualifiedNameFmt + ")?" const labelValueErrMsg string = "a valid label must be an empty string or consist of alphanumeric characters, '-', '_' or '.', and must start and end with an alphanumeric character" diff --git a/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go b/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go index 386c3e7e..d759d912 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go +++ b/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go @@ -19,10 +19,12 @@ package wait import ( "context" "errors" + "math" "math/rand" "sync" "time" + "k8s.io/apimachinery/pkg/util/clock" "k8s.io/apimachinery/pkg/util/runtime" ) @@ -128,9 +130,15 @@ func NonSlidingUntilWithContext(ctx context.Context, f func(context.Context), pe // Close stopCh to stop. f may not be invoked if stop channel is already // closed. Pass NeverStop to if you don't want it stop. func JitterUntil(f func(), period time.Duration, jitterFactor float64, sliding bool, stopCh <-chan struct{}) { - var t *time.Timer - var sawTimeout bool + BackoffUntil(f, NewJitteredBackoffManager(period, jitterFactor, &clock.RealClock{}), sliding, stopCh) +} +// BackoffUntil loops until stop channel is closed, run f every duration given by BackoffManager. +// +// If sliding is true, the period is computed after f runs. If it is false then +// period includes the runtime for f. +func BackoffUntil(f func(), backoff BackoffManager, sliding bool, stopCh <-chan struct{}) { + var t clock.Timer for { select { case <-stopCh: @@ -138,13 +146,8 @@ func JitterUntil(f func(), period time.Duration, jitterFactor float64, sliding b default: } - jitteredPeriod := period - if jitterFactor > 0.0 { - jitteredPeriod = Jitter(period, jitterFactor) - } - if !sliding { - t = resetOrReuseTimer(t, jitteredPeriod, sawTimeout) + t = backoff.Backoff() } func() { @@ -153,7 +156,7 @@ func JitterUntil(f func(), period time.Duration, jitterFactor float64, sliding b }() if sliding { - t = resetOrReuseTimer(t, jitteredPeriod, sawTimeout) + t = backoff.Backoff() } // NOTE: b/c there is no priority selection in golang @@ -164,8 +167,7 @@ func JitterUntil(f func(), period time.Duration, jitterFactor float64, sliding b select { case <-stopCh: return - case <-t.C: - sawTimeout = true + case <-t.C(): } } } @@ -203,6 +205,12 @@ var ErrWaitTimeout = errors.New("timed out waiting for the condition") // if the loop should be aborted. type ConditionFunc func() (done bool, err error) +// runConditionWithCrashProtection runs a ConditionFunc with crash protection +func runConditionWithCrashProtection(condition ConditionFunc) (bool, error) { + defer runtime.HandleCrash() + return condition() +} + // Backoff holds parameters applied to a Backoff function. type Backoff struct { // The initial duration. @@ -277,6 +285,105 @@ func contextForChannel(parentCh <-chan struct{}) (context.Context, context.Cance return ctx, cancel } +// BackoffManager manages backoff with a particular scheme based on its underlying implementation. It provides +// an interface to return a timer for backoff, and caller shall backoff until Timer.C() drains. If the second Backoff() +// is called before the timer from the first Backoff() call finishes, the first timer will NOT be drained and result in +// undetermined behavior. +// The BackoffManager is supposed to be called in a single-threaded environment. +type BackoffManager interface { + Backoff() clock.Timer +} + +type exponentialBackoffManagerImpl struct { + backoff *Backoff + backoffTimer clock.Timer + lastBackoffStart time.Time + initialBackoff time.Duration + backoffResetDuration time.Duration + clock clock.Clock +} + +// NewExponentialBackoffManager returns a manager for managing exponential backoff. Each backoff is jittered and +// backoff will not exceed the given max. If the backoff is not called within resetDuration, the backoff is reset. +// This backoff manager is used to reduce load during upstream unhealthiness. +func NewExponentialBackoffManager(initBackoff, maxBackoff, resetDuration time.Duration, backoffFactor, jitter float64, c clock.Clock) BackoffManager { + return &exponentialBackoffManagerImpl{ + backoff: &Backoff{ + Duration: initBackoff, + Factor: backoffFactor, + Jitter: jitter, + + // the current impl of wait.Backoff returns Backoff.Duration once steps are used up, which is not + // what we ideally need here, we set it to max int and assume we will never use up the steps + Steps: math.MaxInt32, + Cap: maxBackoff, + }, + backoffTimer: nil, + initialBackoff: initBackoff, + lastBackoffStart: c.Now(), + backoffResetDuration: resetDuration, + clock: c, + } +} + +func (b *exponentialBackoffManagerImpl) getNextBackoff() time.Duration { + if b.clock.Now().Sub(b.lastBackoffStart) > b.backoffResetDuration { + b.backoff.Steps = math.MaxInt32 + b.backoff.Duration = b.initialBackoff + } + b.lastBackoffStart = b.clock.Now() + return b.backoff.Step() +} + +// Backoff implements BackoffManager.Backoff, it returns a timer so caller can block on the timer for exponential backoff. +// The returned timer must be drained before calling Backoff() the second time +func (b *exponentialBackoffManagerImpl) Backoff() clock.Timer { + if b.backoffTimer == nil { + b.backoffTimer = b.clock.NewTimer(b.getNextBackoff()) + } else { + b.backoffTimer.Reset(b.getNextBackoff()) + } + return b.backoffTimer +} + +type jitteredBackoffManagerImpl struct { + clock clock.Clock + duration time.Duration + jitter float64 + backoffTimer clock.Timer +} + +// NewJitteredBackoffManager returns a BackoffManager that backoffs with given duration plus given jitter. If the jitter +// is negative, backoff will not be jittered. +func NewJitteredBackoffManager(duration time.Duration, jitter float64, c clock.Clock) BackoffManager { + return &jitteredBackoffManagerImpl{ + clock: c, + duration: duration, + jitter: jitter, + backoffTimer: nil, + } +} + +func (j *jitteredBackoffManagerImpl) getNextBackoff() time.Duration { + jitteredPeriod := j.duration + if j.jitter > 0.0 { + jitteredPeriod = Jitter(j.duration, j.jitter) + } + return jitteredPeriod +} + +// Backoff implements BackoffManager.Backoff, it returns a timer so caller can block on the timer for jittered backoff. +// The returned timer must be drained before calling Backoff() the second time +func (j *jitteredBackoffManagerImpl) Backoff() clock.Timer { + backoff := j.getNextBackoff() + if j.backoffTimer == nil { + j.backoffTimer = j.clock.NewTimer(backoff) + } else { + j.backoffTimer.Reset(backoff) + } + return j.backoffTimer +} + // ExponentialBackoff repeats a condition check with exponential backoff. // // It repeatedly checks the condition and then sleeps, using `backoff.Step()` @@ -289,7 +396,7 @@ func contextForChannel(parentCh <-chan struct{}) (context.Context, context.Cance // In all other cases, ErrWaitTimeout is returned. func ExponentialBackoff(backoff Backoff, condition ConditionFunc) error { for backoff.Steps > 0 { - if ok, err := condition(); err != nil || ok { + if ok, err := runConditionWithCrashProtection(condition); err != nil || ok { return err } if backoff.Steps == 1 { @@ -335,7 +442,7 @@ func PollImmediate(interval, timeout time.Duration, condition ConditionFunc) err } func pollImmediateInternal(wait WaitFunc, condition ConditionFunc) error { - done, err := condition() + done, err := runConditionWithCrashProtection(condition) if err != nil { return err } @@ -364,7 +471,7 @@ func PollInfinite(interval time.Duration, condition ConditionFunc) error { // Some intervals may be missed if the condition takes too long or the time // window is too short. func PollImmediateInfinite(interval time.Duration, condition ConditionFunc) error { - done, err := condition() + done, err := runConditionWithCrashProtection(condition) if err != nil { return err } @@ -431,7 +538,7 @@ func WaitFor(wait WaitFunc, fn ConditionFunc, done <-chan struct{}) error { for { select { case _, open := <-c: - ok, err := fn() + ok, err := runConditionWithCrashProtection(fn) if err != nil { return err } @@ -497,16 +604,3 @@ func poller(interval, timeout time.Duration) WaitFunc { return ch }) } - -// resetOrReuseTimer avoids allocating a new timer if one is already in use. -// Not safe for multiple threads. -func resetOrReuseTimer(t *time.Timer, d time.Duration, sawTimeout bool) *time.Timer { - if t == nil { - return time.NewTimer(d) - } - if !t.Stop() && !sawTimeout { - <-t.C - } - t.Reset(d) - return t -} diff --git a/vendor/k8s.io/apimachinery/pkg/util/yaml/decoder.go b/vendor/k8s.io/apimachinery/pkg/util/yaml/decoder.go index a9a3853a..492171fa 100644 --- a/vendor/k8s.io/apimachinery/pkg/util/yaml/decoder.go +++ b/vendor/k8s.io/apimachinery/pkg/util/yaml/decoder.go @@ -26,7 +26,7 @@ import ( "strings" "unicode" - "k8s.io/klog" + "k8s.io/klog/v2" "sigs.k8s.io/yaml" ) @@ -92,6 +92,10 @@ type YAMLDecoder struct { // the caller in framing the chunk. func NewDocumentDecoder(r io.ReadCloser) io.ReadCloser { scanner := bufio.NewScanner(r) + // the size of initial allocation for buffer 4k + buf := make([]byte, 4*1024) + // the maximum size used to buffer a token 5M + scanner.Buffer(buf, 5*1024*1024) scanner.Split(splitYAMLDocument) return &YAMLDecoder{ r: r, diff --git a/vendor/k8s.io/apimachinery/pkg/watch/streamwatcher.go b/vendor/k8s.io/apimachinery/pkg/watch/streamwatcher.go index 8af256eb..8271e9b7 100644 --- a/vendor/k8s.io/apimachinery/pkg/watch/streamwatcher.go +++ b/vendor/k8s.io/apimachinery/pkg/watch/streamwatcher.go @@ -21,7 +21,7 @@ import ( "io" "sync" - "k8s.io/klog" + "k8s.io/klog/v2" "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/util/net" @@ -113,7 +113,7 @@ func (sw *StreamWatcher) receive() { case io.ErrUnexpectedEOF: klog.V(1).Infof("Unexpected EOF during watch stream event decoding: %v", err) default: - if net.IsProbableEOF(err) { + if net.IsProbableEOF(err) || net.IsTimeout(err) { klog.V(5).Infof("Unable to decode an event from the watch stream: %v", err) } else { sw.result <- Event{ diff --git a/vendor/k8s.io/apimachinery/pkg/watch/watch.go b/vendor/k8s.io/apimachinery/pkg/watch/watch.go index 3945be3a..1f4911a3 100644 --- a/vendor/k8s.io/apimachinery/pkg/watch/watch.go +++ b/vendor/k8s.io/apimachinery/pkg/watch/watch.go @@ -20,7 +20,7 @@ import ( "fmt" "sync" - "k8s.io/klog" + "k8s.io/klog/v2" "k8s.io/apimachinery/pkg/runtime" ) @@ -32,8 +32,8 @@ type Interface interface { Stop() // Returns a chan which will receive all the events. If an error occurs - // or Stop() is called, this channel will be closed, in which case the - // watch should be completely cleaned up. + // or Stop() is called, the implementation will close this channel and + // release any resources used by the watch. ResultChan() <-chan Event } @@ -46,7 +46,9 @@ const ( Deleted EventType = "DELETED" Bookmark EventType = "BOOKMARK" Error EventType = "ERROR" +) +var ( DefaultChanSize int32 = 100 ) @@ -90,7 +92,7 @@ func (w emptyWatch) ResultChan() <-chan Event { // FakeWatcher lets you test anything that consumes a watch.Interface; threadsafe. type FakeWatcher struct { result chan Event - Stopped bool + stopped bool sync.Mutex } @@ -110,24 +112,24 @@ func NewFakeWithChanSize(size int, blocking bool) *FakeWatcher { func (f *FakeWatcher) Stop() { f.Lock() defer f.Unlock() - if !f.Stopped { + if !f.stopped { klog.V(4).Infof("Stopping fake watcher.") close(f.result) - f.Stopped = true + f.stopped = true } } func (f *FakeWatcher) IsStopped() bool { f.Lock() defer f.Unlock() - return f.Stopped + return f.stopped } // Reset prepares the watcher to be reused. func (f *FakeWatcher) Reset() { f.Lock() defer f.Unlock() - f.Stopped = false + f.stopped = false f.result = make(chan Event) } diff --git a/vendor/k8s.io/apimachinery/third_party/forked/golang/reflect/deep_equal.go b/vendor/k8s.io/apimachinery/third_party/forked/golang/reflect/deep_equal.go index 7ed1d1cf..6be80349 100644 --- a/vendor/k8s.io/apimachinery/third_party/forked/golang/reflect/deep_equal.go +++ b/vendor/k8s.io/apimachinery/third_party/forked/golang/reflect/deep_equal.go @@ -16,7 +16,7 @@ import ( // that type. type Equalities map[reflect.Type]reflect.Value -// For convenience, panics on errrors +// For convenience, panics on errors func EqualitiesOrDie(funcs ...interface{}) Equalities { e := Equalities{} if err := e.AddFuncs(funcs...); err != nil { @@ -229,7 +229,7 @@ func (e Equalities) deepValueEqual(v1, v2 reflect.Value, visited map[visit]bool, // // An empty slice *is* equal to a nil slice for our purposes; same for maps. // -// Unexported field members cannot be compared and will cause an imformative panic; you must add an Equality +// Unexported field members cannot be compared and will cause an informative panic; you must add an Equality // function for these types. func (e Equalities) DeepEqual(a1, a2 interface{}) bool { if a1 == nil || a2 == nil { diff --git a/vendor/k8s.io/client-go/discovery/discovery_client.go b/vendor/k8s.io/client-go/discovery/discovery_client.go index 5d89457c..6c8e87e2 100644 --- a/vendor/k8s.io/client-go/discovery/discovery_client.go +++ b/vendor/k8s.io/client-go/discovery/discovery_client.go @@ -17,6 +17,7 @@ limitations under the License. package discovery import ( + "context" "encoding/json" "fmt" "net/url" @@ -26,7 +27,7 @@ import ( "time" "github.com/golang/protobuf/proto" - openapi_v2 "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" "k8s.io/apimachinery/pkg/api/errors" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" @@ -155,7 +156,7 @@ func apiVersionsToAPIGroup(apiVersions *metav1.APIVersions) (apiGroup metav1.API func (d *DiscoveryClient) ServerGroups() (apiGroupList *metav1.APIGroupList, err error) { // Get the groupVersions exposed at /api v := &metav1.APIVersions{} - err = d.restClient.Get().AbsPath(d.LegacyPrefix).Do().Into(v) + err = d.restClient.Get().AbsPath(d.LegacyPrefix).Do(context.TODO()).Into(v) apiGroup := metav1.APIGroup{} if err == nil && len(v.Versions) != 0 { apiGroup = apiVersionsToAPIGroup(v) @@ -166,7 +167,7 @@ func (d *DiscoveryClient) ServerGroups() (apiGroupList *metav1.APIGroupList, err // Get the groupVersions exposed at /apis apiGroupList = &metav1.APIGroupList{} - err = d.restClient.Get().AbsPath("/apis").Do().Into(apiGroupList) + err = d.restClient.Get().AbsPath("/apis").Do(context.TODO()).Into(apiGroupList) if err != nil && !errors.IsNotFound(err) && !errors.IsForbidden(err) { return nil, err } @@ -196,7 +197,7 @@ func (d *DiscoveryClient) ServerResourcesForGroupVersion(groupVersion string) (r resources = &metav1.APIResourceList{ GroupVersion: groupVersion, } - err = d.restClient.Get().AbsPath(url.String()).Do().Into(resources) + err = d.restClient.Get().AbsPath(url.String()).Do(context.TODO()).Into(resources) if err != nil { // ignore 403 or 404 error to be compatible with an v1.0 server. if groupVersion == "v1" && (errors.IsNotFound(err) || errors.IsForbidden(err)) { @@ -405,7 +406,7 @@ func ServerPreferredNamespacedResources(d DiscoveryInterface) ([]*metav1.APIReso // ServerVersion retrieves and parses the server's version (git version). func (d *DiscoveryClient) ServerVersion() (*version.Info, error) { - body, err := d.restClient.Get().AbsPath("/version").Do().Raw() + body, err := d.restClient.Get().AbsPath("/version").Do(context.TODO()).Raw() if err != nil { return nil, err } @@ -419,12 +420,12 @@ func (d *DiscoveryClient) ServerVersion() (*version.Info, error) { // OpenAPISchema fetches the open api schema using a rest client and parses the proto. func (d *DiscoveryClient) OpenAPISchema() (*openapi_v2.Document, error) { - data, err := d.restClient.Get().AbsPath("/openapi/v2").SetHeader("Accept", mimePb).Do().Raw() + data, err := d.restClient.Get().AbsPath("/openapi/v2").SetHeader("Accept", mimePb).Do(context.TODO()).Raw() if err != nil { if errors.IsForbidden(err) || errors.IsNotFound(err) || errors.IsNotAcceptable(err) { // single endpoint not found/registered in old server, try to fetch old endpoint // TODO: remove this when kubectl/client-go don't work with 1.9 server - data, err = d.restClient.Get().AbsPath("/swagger-2.0.0.pb-v1").Do().Raw() + data, err = d.restClient.Get().AbsPath("/swagger-2.0.0.pb-v1").Do(context.TODO()).Raw() if err != nil { return nil, err } diff --git a/vendor/k8s.io/client-go/discovery/fake/discovery.go b/vendor/k8s.io/client-go/discovery/fake/discovery.go index 335473dd..f0cc2dbf 100644 --- a/vendor/k8s.io/client-go/discovery/fake/discovery.go +++ b/vendor/k8s.io/client-go/discovery/fake/discovery.go @@ -19,7 +19,7 @@ package fake import ( "fmt" - "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime/schema" diff --git a/vendor/k8s.io/client-go/dynamic/interface.go b/vendor/k8s.io/client-go/dynamic/interface.go index 70756a4f..b08067c3 100644 --- a/vendor/k8s.io/client-go/dynamic/interface.go +++ b/vendor/k8s.io/client-go/dynamic/interface.go @@ -17,6 +17,8 @@ limitations under the License. package dynamic import ( + "context" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apimachinery/pkg/runtime/schema" @@ -29,15 +31,15 @@ type Interface interface { } type ResourceInterface interface { - Create(obj *unstructured.Unstructured, options metav1.CreateOptions, subresources ...string) (*unstructured.Unstructured, error) - Update(obj *unstructured.Unstructured, options metav1.UpdateOptions, subresources ...string) (*unstructured.Unstructured, error) - UpdateStatus(obj *unstructured.Unstructured, options metav1.UpdateOptions) (*unstructured.Unstructured, error) - Delete(name string, options *metav1.DeleteOptions, subresources ...string) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions, subresources ...string) (*unstructured.Unstructured, error) - List(opts metav1.ListOptions) (*unstructured.UnstructuredList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, options metav1.PatchOptions, subresources ...string) (*unstructured.Unstructured, error) + Create(ctx context.Context, obj *unstructured.Unstructured, options metav1.CreateOptions, subresources ...string) (*unstructured.Unstructured, error) + Update(ctx context.Context, obj *unstructured.Unstructured, options metav1.UpdateOptions, subresources ...string) (*unstructured.Unstructured, error) + UpdateStatus(ctx context.Context, obj *unstructured.Unstructured, options metav1.UpdateOptions) (*unstructured.Unstructured, error) + Delete(ctx context.Context, name string, options metav1.DeleteOptions, subresources ...string) error + DeleteCollection(ctx context.Context, options metav1.DeleteOptions, listOptions metav1.ListOptions) error + Get(ctx context.Context, name string, options metav1.GetOptions, subresources ...string) (*unstructured.Unstructured, error) + List(ctx context.Context, opts metav1.ListOptions) (*unstructured.UnstructuredList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, options metav1.PatchOptions, subresources ...string) (*unstructured.Unstructured, error) } type NamespaceableResourceInterface interface { diff --git a/vendor/k8s.io/client-go/dynamic/simple.go b/vendor/k8s.io/client-go/dynamic/simple.go index 8a026788..9ae320d3 100644 --- a/vendor/k8s.io/client-go/dynamic/simple.go +++ b/vendor/k8s.io/client-go/dynamic/simple.go @@ -17,6 +17,7 @@ limitations under the License. package dynamic import ( + "context" "fmt" "k8s.io/apimachinery/pkg/api/meta" @@ -89,7 +90,7 @@ func (c *dynamicResourceClient) Namespace(ns string) ResourceInterface { return &ret } -func (c *dynamicResourceClient) Create(obj *unstructured.Unstructured, opts metav1.CreateOptions, subresources ...string) (*unstructured.Unstructured, error) { +func (c *dynamicResourceClient) Create(ctx context.Context, obj *unstructured.Unstructured, opts metav1.CreateOptions, subresources ...string) (*unstructured.Unstructured, error) { outBytes, err := runtime.Encode(unstructured.UnstructuredJSONScheme, obj) if err != nil { return nil, err @@ -111,7 +112,7 @@ func (c *dynamicResourceClient) Create(obj *unstructured.Unstructured, opts meta AbsPath(append(c.makeURLSegments(name), subresources...)...). Body(outBytes). SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1). - Do() + Do(ctx) if err := result.Error(); err != nil { return nil, err } @@ -127,7 +128,7 @@ func (c *dynamicResourceClient) Create(obj *unstructured.Unstructured, opts meta return uncastObj.(*unstructured.Unstructured), nil } -func (c *dynamicResourceClient) Update(obj *unstructured.Unstructured, opts metav1.UpdateOptions, subresources ...string) (*unstructured.Unstructured, error) { +func (c *dynamicResourceClient) Update(ctx context.Context, obj *unstructured.Unstructured, opts metav1.UpdateOptions, subresources ...string) (*unstructured.Unstructured, error) { accessor, err := meta.Accessor(obj) if err != nil { return nil, err @@ -146,7 +147,7 @@ func (c *dynamicResourceClient) Update(obj *unstructured.Unstructured, opts meta AbsPath(append(c.makeURLSegments(name), subresources...)...). Body(outBytes). SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1). - Do() + Do(ctx) if err := result.Error(); err != nil { return nil, err } @@ -162,7 +163,7 @@ func (c *dynamicResourceClient) Update(obj *unstructured.Unstructured, opts meta return uncastObj.(*unstructured.Unstructured), nil } -func (c *dynamicResourceClient) UpdateStatus(obj *unstructured.Unstructured, opts metav1.UpdateOptions) (*unstructured.Unstructured, error) { +func (c *dynamicResourceClient) UpdateStatus(ctx context.Context, obj *unstructured.Unstructured, opts metav1.UpdateOptions) (*unstructured.Unstructured, error) { accessor, err := meta.Accessor(obj) if err != nil { return nil, err @@ -182,7 +183,7 @@ func (c *dynamicResourceClient) UpdateStatus(obj *unstructured.Unstructured, opt AbsPath(append(c.makeURLSegments(name), "status")...). Body(outBytes). SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1). - Do() + Do(ctx) if err := result.Error(); err != nil { return nil, err } @@ -198,14 +199,11 @@ func (c *dynamicResourceClient) UpdateStatus(obj *unstructured.Unstructured, opt return uncastObj.(*unstructured.Unstructured), nil } -func (c *dynamicResourceClient) Delete(name string, opts *metav1.DeleteOptions, subresources ...string) error { +func (c *dynamicResourceClient) Delete(ctx context.Context, name string, opts metav1.DeleteOptions, subresources ...string) error { if len(name) == 0 { return fmt.Errorf("name is required") } - if opts == nil { - opts = &metav1.DeleteOptions{} - } - deleteOptionsByte, err := runtime.Encode(deleteOptionsCodec.LegacyCodec(schema.GroupVersion{Version: "v1"}), opts) + deleteOptionsByte, err := runtime.Encode(deleteOptionsCodec.LegacyCodec(schema.GroupVersion{Version: "v1"}), &opts) if err != nil { return err } @@ -214,15 +212,12 @@ func (c *dynamicResourceClient) Delete(name string, opts *metav1.DeleteOptions, Delete(). AbsPath(append(c.makeURLSegments(name), subresources...)...). Body(deleteOptionsByte). - Do() + Do(ctx) return result.Error() } -func (c *dynamicResourceClient) DeleteCollection(opts *metav1.DeleteOptions, listOptions metav1.ListOptions) error { - if opts == nil { - opts = &metav1.DeleteOptions{} - } - deleteOptionsByte, err := runtime.Encode(deleteOptionsCodec.LegacyCodec(schema.GroupVersion{Version: "v1"}), opts) +func (c *dynamicResourceClient) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOptions metav1.ListOptions) error { + deleteOptionsByte, err := runtime.Encode(deleteOptionsCodec.LegacyCodec(schema.GroupVersion{Version: "v1"}), &opts) if err != nil { return err } @@ -232,15 +227,15 @@ func (c *dynamicResourceClient) DeleteCollection(opts *metav1.DeleteOptions, lis AbsPath(c.makeURLSegments("")...). Body(deleteOptionsByte). SpecificallyVersionedParams(&listOptions, dynamicParameterCodec, versionV1). - Do() + Do(ctx) return result.Error() } -func (c *dynamicResourceClient) Get(name string, opts metav1.GetOptions, subresources ...string) (*unstructured.Unstructured, error) { +func (c *dynamicResourceClient) Get(ctx context.Context, name string, opts metav1.GetOptions, subresources ...string) (*unstructured.Unstructured, error) { if len(name) == 0 { return nil, fmt.Errorf("name is required") } - result := c.client.client.Get().AbsPath(append(c.makeURLSegments(name), subresources...)...).SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1).Do() + result := c.client.client.Get().AbsPath(append(c.makeURLSegments(name), subresources...)...).SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1).Do(ctx) if err := result.Error(); err != nil { return nil, err } @@ -255,8 +250,8 @@ func (c *dynamicResourceClient) Get(name string, opts metav1.GetOptions, subreso return uncastObj.(*unstructured.Unstructured), nil } -func (c *dynamicResourceClient) List(opts metav1.ListOptions) (*unstructured.UnstructuredList, error) { - result := c.client.client.Get().AbsPath(c.makeURLSegments("")...).SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1).Do() +func (c *dynamicResourceClient) List(ctx context.Context, opts metav1.ListOptions) (*unstructured.UnstructuredList, error) { + result := c.client.client.Get().AbsPath(c.makeURLSegments("")...).SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1).Do(ctx) if err := result.Error(); err != nil { return nil, err } @@ -279,14 +274,14 @@ func (c *dynamicResourceClient) List(opts metav1.ListOptions) (*unstructured.Uns return list, nil } -func (c *dynamicResourceClient) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *dynamicResourceClient) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { opts.Watch = true return c.client.client.Get().AbsPath(c.makeURLSegments("")...). SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1). - Watch() + Watch(ctx) } -func (c *dynamicResourceClient) Patch(name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (*unstructured.Unstructured, error) { +func (c *dynamicResourceClient) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (*unstructured.Unstructured, error) { if len(name) == 0 { return nil, fmt.Errorf("name is required") } @@ -295,7 +290,7 @@ func (c *dynamicResourceClient) Patch(name string, pt types.PatchType, data []by AbsPath(append(c.makeURLSegments(name), subresources...)...). Body(data). SpecificallyVersionedParams(&opts, dynamicParameterCodec, versionV1). - Do() + Do(ctx) if err := result.Error(); err != nil { return nil, err } diff --git a/vendor/k8s.io/client-go/kubernetes/scheme/register.go b/vendor/k8s.io/client-go/kubernetes/scheme/register.go index 4d8e8b7f..2710bf2d 100644 --- a/vendor/k8s.io/client-go/kubernetes/scheme/register.go +++ b/vendor/k8s.io/client-go/kubernetes/scheme/register.go @@ -24,7 +24,6 @@ import ( appsv1 "k8s.io/api/apps/v1" appsv1beta1 "k8s.io/api/apps/v1beta1" appsv1beta2 "k8s.io/api/apps/v1beta2" - auditregistrationv1alpha1 "k8s.io/api/auditregistration/v1alpha1" authenticationv1 "k8s.io/api/authentication/v1" authenticationv1beta1 "k8s.io/api/authentication/v1beta1" authorizationv1 "k8s.io/api/authorization/v1" @@ -35,12 +34,14 @@ import ( batchv1 "k8s.io/api/batch/v1" batchv1beta1 "k8s.io/api/batch/v1beta1" batchv2alpha1 "k8s.io/api/batch/v2alpha1" + certificatesv1 "k8s.io/api/certificates/v1" certificatesv1beta1 "k8s.io/api/certificates/v1beta1" coordinationv1 "k8s.io/api/coordination/v1" coordinationv1beta1 "k8s.io/api/coordination/v1beta1" corev1 "k8s.io/api/core/v1" discoveryv1alpha1 "k8s.io/api/discovery/v1alpha1" discoveryv1beta1 "k8s.io/api/discovery/v1beta1" + eventsv1 "k8s.io/api/events/v1" eventsv1beta1 "k8s.io/api/events/v1beta1" extensionsv1beta1 "k8s.io/api/extensions/v1beta1" flowcontrolv1alpha1 "k8s.io/api/flowcontrol/v1alpha1" @@ -75,7 +76,6 @@ var localSchemeBuilder = runtime.SchemeBuilder{ appsv1.AddToScheme, appsv1beta1.AddToScheme, appsv1beta2.AddToScheme, - auditregistrationv1alpha1.AddToScheme, authenticationv1.AddToScheme, authenticationv1beta1.AddToScheme, authorizationv1.AddToScheme, @@ -86,12 +86,14 @@ var localSchemeBuilder = runtime.SchemeBuilder{ batchv1.AddToScheme, batchv1beta1.AddToScheme, batchv2alpha1.AddToScheme, + certificatesv1.AddToScheme, certificatesv1beta1.AddToScheme, coordinationv1beta1.AddToScheme, coordinationv1.AddToScheme, corev1.AddToScheme, discoveryv1alpha1.AddToScheme, discoveryv1beta1.AddToScheme, + eventsv1.AddToScheme, eventsv1beta1.AddToScheme, extensionsv1beta1.AddToScheme, flowcontrolv1alpha1.AddToScheme, diff --git a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/controllerrevision.go b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/controllerrevision.go index e28e4d2a..dba06207 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/controllerrevision.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/controllerrevision.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/apps/v1" @@ -37,14 +38,14 @@ type ControllerRevisionsGetter interface { // ControllerRevisionInterface has methods to work with ControllerRevision resources. type ControllerRevisionInterface interface { - Create(*v1.ControllerRevision) (*v1.ControllerRevision, error) - Update(*v1.ControllerRevision) (*v1.ControllerRevision, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.ControllerRevision, error) - List(opts metav1.ListOptions) (*v1.ControllerRevisionList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ControllerRevision, err error) + Create(ctx context.Context, controllerRevision *v1.ControllerRevision, opts metav1.CreateOptions) (*v1.ControllerRevision, error) + Update(ctx context.Context, controllerRevision *v1.ControllerRevision, opts metav1.UpdateOptions) (*v1.ControllerRevision, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.ControllerRevision, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ControllerRevisionList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ControllerRevision, err error) ControllerRevisionExpansion } @@ -63,20 +64,20 @@ func newControllerRevisions(c *AppsV1Client, namespace string) *controllerRevisi } // Get takes name of the controllerRevision, and returns the corresponding controllerRevision object, and an error if there is any. -func (c *controllerRevisions) Get(name string, options metav1.GetOptions) (result *v1.ControllerRevision, err error) { +func (c *controllerRevisions) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.ControllerRevision, err error) { result = &v1.ControllerRevision{} err = c.client.Get(). Namespace(c.ns). Resource("controllerrevisions"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of ControllerRevisions that match those selectors. -func (c *controllerRevisions) List(opts metav1.ListOptions) (result *v1.ControllerRevisionList, err error) { +func (c *controllerRevisions) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ControllerRevisionList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *controllerRevisions) List(opts metav1.ListOptions) (result *v1.Controll Resource("controllerrevisions"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested controllerRevisions. -func (c *controllerRevisions) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *controllerRevisions) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *controllerRevisions) Watch(opts metav1.ListOptions) (watch.Interface, e Resource("controllerrevisions"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a controllerRevision and creates it. Returns the server's representation of the controllerRevision, and an error, if there is any. -func (c *controllerRevisions) Create(controllerRevision *v1.ControllerRevision) (result *v1.ControllerRevision, err error) { +func (c *controllerRevisions) Create(ctx context.Context, controllerRevision *v1.ControllerRevision, opts metav1.CreateOptions) (result *v1.ControllerRevision, err error) { result = &v1.ControllerRevision{} err = c.client.Post(). Namespace(c.ns). Resource("controllerrevisions"). + VersionedParams(&opts, scheme.ParameterCodec). Body(controllerRevision). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a controllerRevision and updates it. Returns the server's representation of the controllerRevision, and an error, if there is any. -func (c *controllerRevisions) Update(controllerRevision *v1.ControllerRevision) (result *v1.ControllerRevision, err error) { +func (c *controllerRevisions) Update(ctx context.Context, controllerRevision *v1.ControllerRevision, opts metav1.UpdateOptions) (result *v1.ControllerRevision, err error) { result = &v1.ControllerRevision{} err = c.client.Put(). Namespace(c.ns). Resource("controllerrevisions"). Name(controllerRevision.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(controllerRevision). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the controllerRevision and deletes it. Returns an error if one occurs. -func (c *controllerRevisions) Delete(name string, options *metav1.DeleteOptions) error { +func (c *controllerRevisions) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("controllerrevisions"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *controllerRevisions) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *controllerRevisions) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("controllerrevisions"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched controllerRevision. -func (c *controllerRevisions) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ControllerRevision, err error) { +func (c *controllerRevisions) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ControllerRevision, err error) { result = &v1.ControllerRevision{} err = c.client.Patch(pt). Namespace(c.ns). Resource("controllerrevisions"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/daemonset.go b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/daemonset.go index a535cdab..0bb397af 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/daemonset.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/daemonset.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/apps/v1" @@ -37,15 +38,15 @@ type DaemonSetsGetter interface { // DaemonSetInterface has methods to work with DaemonSet resources. type DaemonSetInterface interface { - Create(*v1.DaemonSet) (*v1.DaemonSet, error) - Update(*v1.DaemonSet) (*v1.DaemonSet, error) - UpdateStatus(*v1.DaemonSet) (*v1.DaemonSet, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.DaemonSet, error) - List(opts metav1.ListOptions) (*v1.DaemonSetList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.DaemonSet, err error) + Create(ctx context.Context, daemonSet *v1.DaemonSet, opts metav1.CreateOptions) (*v1.DaemonSet, error) + Update(ctx context.Context, daemonSet *v1.DaemonSet, opts metav1.UpdateOptions) (*v1.DaemonSet, error) + UpdateStatus(ctx context.Context, daemonSet *v1.DaemonSet, opts metav1.UpdateOptions) (*v1.DaemonSet, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.DaemonSet, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.DaemonSetList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.DaemonSet, err error) DaemonSetExpansion } @@ -64,20 +65,20 @@ func newDaemonSets(c *AppsV1Client, namespace string) *daemonSets { } // Get takes name of the daemonSet, and returns the corresponding daemonSet object, and an error if there is any. -func (c *daemonSets) Get(name string, options metav1.GetOptions) (result *v1.DaemonSet, err error) { +func (c *daemonSets) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.DaemonSet, err error) { result = &v1.DaemonSet{} err = c.client.Get(). Namespace(c.ns). Resource("daemonsets"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of DaemonSets that match those selectors. -func (c *daemonSets) List(opts metav1.ListOptions) (result *v1.DaemonSetList, err error) { +func (c *daemonSets) List(ctx context.Context, opts metav1.ListOptions) (result *v1.DaemonSetList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -88,13 +89,13 @@ func (c *daemonSets) List(opts metav1.ListOptions) (result *v1.DaemonSetList, er Resource("daemonsets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested daemonSets. -func (c *daemonSets) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *daemonSets) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -105,87 +106,90 @@ func (c *daemonSets) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("daemonsets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a daemonSet and creates it. Returns the server's representation of the daemonSet, and an error, if there is any. -func (c *daemonSets) Create(daemonSet *v1.DaemonSet) (result *v1.DaemonSet, err error) { +func (c *daemonSets) Create(ctx context.Context, daemonSet *v1.DaemonSet, opts metav1.CreateOptions) (result *v1.DaemonSet, err error) { result = &v1.DaemonSet{} err = c.client.Post(). Namespace(c.ns). Resource("daemonsets"). + VersionedParams(&opts, scheme.ParameterCodec). Body(daemonSet). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a daemonSet and updates it. Returns the server's representation of the daemonSet, and an error, if there is any. -func (c *daemonSets) Update(daemonSet *v1.DaemonSet) (result *v1.DaemonSet, err error) { +func (c *daemonSets) Update(ctx context.Context, daemonSet *v1.DaemonSet, opts metav1.UpdateOptions) (result *v1.DaemonSet, err error) { result = &v1.DaemonSet{} err = c.client.Put(). Namespace(c.ns). Resource("daemonsets"). Name(daemonSet.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(daemonSet). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *daemonSets) UpdateStatus(daemonSet *v1.DaemonSet) (result *v1.DaemonSet, err error) { +func (c *daemonSets) UpdateStatus(ctx context.Context, daemonSet *v1.DaemonSet, opts metav1.UpdateOptions) (result *v1.DaemonSet, err error) { result = &v1.DaemonSet{} err = c.client.Put(). Namespace(c.ns). Resource("daemonsets"). Name(daemonSet.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(daemonSet). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the daemonSet and deletes it. Returns an error if one occurs. -func (c *daemonSets) Delete(name string, options *metav1.DeleteOptions) error { +func (c *daemonSets) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("daemonsets"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *daemonSets) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *daemonSets) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("daemonsets"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched daemonSet. -func (c *daemonSets) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.DaemonSet, err error) { +func (c *daemonSets) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.DaemonSet, err error) { result = &v1.DaemonSet{} err = c.client.Patch(pt). Namespace(c.ns). Resource("daemonsets"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/deployment.go b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/deployment.go index f9799a45..69d1b86d 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/deployment.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/deployment.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/apps/v1" @@ -38,17 +39,17 @@ type DeploymentsGetter interface { // DeploymentInterface has methods to work with Deployment resources. type DeploymentInterface interface { - Create(*v1.Deployment) (*v1.Deployment, error) - Update(*v1.Deployment) (*v1.Deployment, error) - UpdateStatus(*v1.Deployment) (*v1.Deployment, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.Deployment, error) - List(opts metav1.ListOptions) (*v1.DeploymentList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Deployment, err error) - GetScale(deploymentName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) - UpdateScale(deploymentName string, scale *autoscalingv1.Scale) (*autoscalingv1.Scale, error) + Create(ctx context.Context, deployment *v1.Deployment, opts metav1.CreateOptions) (*v1.Deployment, error) + Update(ctx context.Context, deployment *v1.Deployment, opts metav1.UpdateOptions) (*v1.Deployment, error) + UpdateStatus(ctx context.Context, deployment *v1.Deployment, opts metav1.UpdateOptions) (*v1.Deployment, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Deployment, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.DeploymentList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Deployment, err error) + GetScale(ctx context.Context, deploymentName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) + UpdateScale(ctx context.Context, deploymentName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (*autoscalingv1.Scale, error) DeploymentExpansion } @@ -68,20 +69,20 @@ func newDeployments(c *AppsV1Client, namespace string) *deployments { } // Get takes name of the deployment, and returns the corresponding deployment object, and an error if there is any. -func (c *deployments) Get(name string, options metav1.GetOptions) (result *v1.Deployment, err error) { +func (c *deployments) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Deployment, err error) { result = &v1.Deployment{} err = c.client.Get(). Namespace(c.ns). Resource("deployments"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Deployments that match those selectors. -func (c *deployments) List(opts metav1.ListOptions) (result *v1.DeploymentList, err error) { +func (c *deployments) List(ctx context.Context, opts metav1.ListOptions) (result *v1.DeploymentList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -92,13 +93,13 @@ func (c *deployments) List(opts metav1.ListOptions) (result *v1.DeploymentList, Resource("deployments"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested deployments. -func (c *deployments) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *deployments) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -109,93 +110,96 @@ func (c *deployments) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("deployments"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a deployment and creates it. Returns the server's representation of the deployment, and an error, if there is any. -func (c *deployments) Create(deployment *v1.Deployment) (result *v1.Deployment, err error) { +func (c *deployments) Create(ctx context.Context, deployment *v1.Deployment, opts metav1.CreateOptions) (result *v1.Deployment, err error) { result = &v1.Deployment{} err = c.client.Post(). Namespace(c.ns). Resource("deployments"). + VersionedParams(&opts, scheme.ParameterCodec). Body(deployment). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a deployment and updates it. Returns the server's representation of the deployment, and an error, if there is any. -func (c *deployments) Update(deployment *v1.Deployment) (result *v1.Deployment, err error) { +func (c *deployments) Update(ctx context.Context, deployment *v1.Deployment, opts metav1.UpdateOptions) (result *v1.Deployment, err error) { result = &v1.Deployment{} err = c.client.Put(). Namespace(c.ns). Resource("deployments"). Name(deployment.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(deployment). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *deployments) UpdateStatus(deployment *v1.Deployment) (result *v1.Deployment, err error) { +func (c *deployments) UpdateStatus(ctx context.Context, deployment *v1.Deployment, opts metav1.UpdateOptions) (result *v1.Deployment, err error) { result = &v1.Deployment{} err = c.client.Put(). Namespace(c.ns). Resource("deployments"). Name(deployment.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(deployment). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the deployment and deletes it. Returns an error if one occurs. -func (c *deployments) Delete(name string, options *metav1.DeleteOptions) error { +func (c *deployments) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("deployments"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *deployments) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *deployments) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("deployments"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched deployment. -func (c *deployments) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Deployment, err error) { +func (c *deployments) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Deployment, err error) { result = &v1.Deployment{} err = c.client.Patch(pt). Namespace(c.ns). Resource("deployments"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } // GetScale takes name of the deployment, and returns the corresponding autoscalingv1.Scale object, and an error if there is any. -func (c *deployments) GetScale(deploymentName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { +func (c *deployments) GetScale(ctx context.Context, deploymentName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Get(). Namespace(c.ns). @@ -203,21 +207,22 @@ func (c *deployments) GetScale(deploymentName string, options metav1.GetOptions) Name(deploymentName). SubResource("scale"). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // UpdateScale takes the top resource name and the representation of a scale and updates it. Returns the server's representation of the scale, and an error, if there is any. -func (c *deployments) UpdateScale(deploymentName string, scale *autoscalingv1.Scale) (result *autoscalingv1.Scale, err error) { +func (c *deployments) UpdateScale(ctx context.Context, deploymentName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Put(). Namespace(c.ns). Resource("deployments"). Name(deploymentName). SubResource("scale"). + VersionedParams(&opts, scheme.ParameterCodec). Body(scale). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/replicaset.go b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/replicaset.go index ff3504e7..377b9ca3 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/replicaset.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/replicaset.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/apps/v1" @@ -38,17 +39,17 @@ type ReplicaSetsGetter interface { // ReplicaSetInterface has methods to work with ReplicaSet resources. type ReplicaSetInterface interface { - Create(*v1.ReplicaSet) (*v1.ReplicaSet, error) - Update(*v1.ReplicaSet) (*v1.ReplicaSet, error) - UpdateStatus(*v1.ReplicaSet) (*v1.ReplicaSet, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.ReplicaSet, error) - List(opts metav1.ListOptions) (*v1.ReplicaSetList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ReplicaSet, err error) - GetScale(replicaSetName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) - UpdateScale(replicaSetName string, scale *autoscalingv1.Scale) (*autoscalingv1.Scale, error) + Create(ctx context.Context, replicaSet *v1.ReplicaSet, opts metav1.CreateOptions) (*v1.ReplicaSet, error) + Update(ctx context.Context, replicaSet *v1.ReplicaSet, opts metav1.UpdateOptions) (*v1.ReplicaSet, error) + UpdateStatus(ctx context.Context, replicaSet *v1.ReplicaSet, opts metav1.UpdateOptions) (*v1.ReplicaSet, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.ReplicaSet, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ReplicaSetList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ReplicaSet, err error) + GetScale(ctx context.Context, replicaSetName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) + UpdateScale(ctx context.Context, replicaSetName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (*autoscalingv1.Scale, error) ReplicaSetExpansion } @@ -68,20 +69,20 @@ func newReplicaSets(c *AppsV1Client, namespace string) *replicaSets { } // Get takes name of the replicaSet, and returns the corresponding replicaSet object, and an error if there is any. -func (c *replicaSets) Get(name string, options metav1.GetOptions) (result *v1.ReplicaSet, err error) { +func (c *replicaSets) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.ReplicaSet, err error) { result = &v1.ReplicaSet{} err = c.client.Get(). Namespace(c.ns). Resource("replicasets"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of ReplicaSets that match those selectors. -func (c *replicaSets) List(opts metav1.ListOptions) (result *v1.ReplicaSetList, err error) { +func (c *replicaSets) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ReplicaSetList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -92,13 +93,13 @@ func (c *replicaSets) List(opts metav1.ListOptions) (result *v1.ReplicaSetList, Resource("replicasets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested replicaSets. -func (c *replicaSets) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *replicaSets) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -109,93 +110,96 @@ func (c *replicaSets) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("replicasets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a replicaSet and creates it. Returns the server's representation of the replicaSet, and an error, if there is any. -func (c *replicaSets) Create(replicaSet *v1.ReplicaSet) (result *v1.ReplicaSet, err error) { +func (c *replicaSets) Create(ctx context.Context, replicaSet *v1.ReplicaSet, opts metav1.CreateOptions) (result *v1.ReplicaSet, err error) { result = &v1.ReplicaSet{} err = c.client.Post(). Namespace(c.ns). Resource("replicasets"). + VersionedParams(&opts, scheme.ParameterCodec). Body(replicaSet). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a replicaSet and updates it. Returns the server's representation of the replicaSet, and an error, if there is any. -func (c *replicaSets) Update(replicaSet *v1.ReplicaSet) (result *v1.ReplicaSet, err error) { +func (c *replicaSets) Update(ctx context.Context, replicaSet *v1.ReplicaSet, opts metav1.UpdateOptions) (result *v1.ReplicaSet, err error) { result = &v1.ReplicaSet{} err = c.client.Put(). Namespace(c.ns). Resource("replicasets"). Name(replicaSet.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(replicaSet). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *replicaSets) UpdateStatus(replicaSet *v1.ReplicaSet) (result *v1.ReplicaSet, err error) { +func (c *replicaSets) UpdateStatus(ctx context.Context, replicaSet *v1.ReplicaSet, opts metav1.UpdateOptions) (result *v1.ReplicaSet, err error) { result = &v1.ReplicaSet{} err = c.client.Put(). Namespace(c.ns). Resource("replicasets"). Name(replicaSet.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(replicaSet). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the replicaSet and deletes it. Returns an error if one occurs. -func (c *replicaSets) Delete(name string, options *metav1.DeleteOptions) error { +func (c *replicaSets) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("replicasets"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *replicaSets) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *replicaSets) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("replicasets"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched replicaSet. -func (c *replicaSets) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ReplicaSet, err error) { +func (c *replicaSets) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ReplicaSet, err error) { result = &v1.ReplicaSet{} err = c.client.Patch(pt). Namespace(c.ns). Resource("replicasets"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } // GetScale takes name of the replicaSet, and returns the corresponding autoscalingv1.Scale object, and an error if there is any. -func (c *replicaSets) GetScale(replicaSetName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { +func (c *replicaSets) GetScale(ctx context.Context, replicaSetName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Get(). Namespace(c.ns). @@ -203,21 +207,22 @@ func (c *replicaSets) GetScale(replicaSetName string, options metav1.GetOptions) Name(replicaSetName). SubResource("scale"). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // UpdateScale takes the top resource name and the representation of a scale and updates it. Returns the server's representation of the scale, and an error, if there is any. -func (c *replicaSets) UpdateScale(replicaSetName string, scale *autoscalingv1.Scale) (result *autoscalingv1.Scale, err error) { +func (c *replicaSets) UpdateScale(ctx context.Context, replicaSetName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Put(). Namespace(c.ns). Resource("replicasets"). Name(replicaSetName). SubResource("scale"). + VersionedParams(&opts, scheme.ParameterCodec). Body(scale). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/statefulset.go b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/statefulset.go index c12c470b..33a9f535 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/statefulset.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/apps/v1/statefulset.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/apps/v1" @@ -38,17 +39,17 @@ type StatefulSetsGetter interface { // StatefulSetInterface has methods to work with StatefulSet resources. type StatefulSetInterface interface { - Create(*v1.StatefulSet) (*v1.StatefulSet, error) - Update(*v1.StatefulSet) (*v1.StatefulSet, error) - UpdateStatus(*v1.StatefulSet) (*v1.StatefulSet, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.StatefulSet, error) - List(opts metav1.ListOptions) (*v1.StatefulSetList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.StatefulSet, err error) - GetScale(statefulSetName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) - UpdateScale(statefulSetName string, scale *autoscalingv1.Scale) (*autoscalingv1.Scale, error) + Create(ctx context.Context, statefulSet *v1.StatefulSet, opts metav1.CreateOptions) (*v1.StatefulSet, error) + Update(ctx context.Context, statefulSet *v1.StatefulSet, opts metav1.UpdateOptions) (*v1.StatefulSet, error) + UpdateStatus(ctx context.Context, statefulSet *v1.StatefulSet, opts metav1.UpdateOptions) (*v1.StatefulSet, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.StatefulSet, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.StatefulSetList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.StatefulSet, err error) + GetScale(ctx context.Context, statefulSetName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) + UpdateScale(ctx context.Context, statefulSetName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (*autoscalingv1.Scale, error) StatefulSetExpansion } @@ -68,20 +69,20 @@ func newStatefulSets(c *AppsV1Client, namespace string) *statefulSets { } // Get takes name of the statefulSet, and returns the corresponding statefulSet object, and an error if there is any. -func (c *statefulSets) Get(name string, options metav1.GetOptions) (result *v1.StatefulSet, err error) { +func (c *statefulSets) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.StatefulSet, err error) { result = &v1.StatefulSet{} err = c.client.Get(). Namespace(c.ns). Resource("statefulsets"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of StatefulSets that match those selectors. -func (c *statefulSets) List(opts metav1.ListOptions) (result *v1.StatefulSetList, err error) { +func (c *statefulSets) List(ctx context.Context, opts metav1.ListOptions) (result *v1.StatefulSetList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -92,13 +93,13 @@ func (c *statefulSets) List(opts metav1.ListOptions) (result *v1.StatefulSetList Resource("statefulsets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested statefulSets. -func (c *statefulSets) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *statefulSets) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -109,93 +110,96 @@ func (c *statefulSets) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("statefulsets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a statefulSet and creates it. Returns the server's representation of the statefulSet, and an error, if there is any. -func (c *statefulSets) Create(statefulSet *v1.StatefulSet) (result *v1.StatefulSet, err error) { +func (c *statefulSets) Create(ctx context.Context, statefulSet *v1.StatefulSet, opts metav1.CreateOptions) (result *v1.StatefulSet, err error) { result = &v1.StatefulSet{} err = c.client.Post(). Namespace(c.ns). Resource("statefulsets"). + VersionedParams(&opts, scheme.ParameterCodec). Body(statefulSet). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a statefulSet and updates it. Returns the server's representation of the statefulSet, and an error, if there is any. -func (c *statefulSets) Update(statefulSet *v1.StatefulSet) (result *v1.StatefulSet, err error) { +func (c *statefulSets) Update(ctx context.Context, statefulSet *v1.StatefulSet, opts metav1.UpdateOptions) (result *v1.StatefulSet, err error) { result = &v1.StatefulSet{} err = c.client.Put(). Namespace(c.ns). Resource("statefulsets"). Name(statefulSet.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(statefulSet). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *statefulSets) UpdateStatus(statefulSet *v1.StatefulSet) (result *v1.StatefulSet, err error) { +func (c *statefulSets) UpdateStatus(ctx context.Context, statefulSet *v1.StatefulSet, opts metav1.UpdateOptions) (result *v1.StatefulSet, err error) { result = &v1.StatefulSet{} err = c.client.Put(). Namespace(c.ns). Resource("statefulsets"). Name(statefulSet.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(statefulSet). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the statefulSet and deletes it. Returns an error if one occurs. -func (c *statefulSets) Delete(name string, options *metav1.DeleteOptions) error { +func (c *statefulSets) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("statefulsets"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *statefulSets) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *statefulSets) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("statefulsets"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched statefulSet. -func (c *statefulSets) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.StatefulSet, err error) { +func (c *statefulSets) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.StatefulSet, err error) { result = &v1.StatefulSet{} err = c.client.Patch(pt). Namespace(c.ns). Resource("statefulsets"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } // GetScale takes name of the statefulSet, and returns the corresponding autoscalingv1.Scale object, and an error if there is any. -func (c *statefulSets) GetScale(statefulSetName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { +func (c *statefulSets) GetScale(ctx context.Context, statefulSetName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Get(). Namespace(c.ns). @@ -203,21 +207,22 @@ func (c *statefulSets) GetScale(statefulSetName string, options metav1.GetOption Name(statefulSetName). SubResource("scale"). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // UpdateScale takes the top resource name and the representation of a scale and updates it. Returns the server's representation of the scale, and an error, if there is any. -func (c *statefulSets) UpdateScale(statefulSetName string, scale *autoscalingv1.Scale) (result *autoscalingv1.Scale, err error) { +func (c *statefulSets) UpdateScale(ctx context.Context, statefulSetName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Put(). Namespace(c.ns). Resource("statefulsets"). Name(statefulSetName). SubResource("scale"). + VersionedParams(&opts, scheme.ParameterCodec). Body(scale). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/componentstatus.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/componentstatus.go index 302b2fdc..faf5d19c 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/componentstatus.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/componentstatus.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type ComponentStatusesGetter interface { // ComponentStatusInterface has methods to work with ComponentStatus resources. type ComponentStatusInterface interface { - Create(*v1.ComponentStatus) (*v1.ComponentStatus, error) - Update(*v1.ComponentStatus) (*v1.ComponentStatus, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.ComponentStatus, error) - List(opts metav1.ListOptions) (*v1.ComponentStatusList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ComponentStatus, err error) + Create(ctx context.Context, componentStatus *v1.ComponentStatus, opts metav1.CreateOptions) (*v1.ComponentStatus, error) + Update(ctx context.Context, componentStatus *v1.ComponentStatus, opts metav1.UpdateOptions) (*v1.ComponentStatus, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.ComponentStatus, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ComponentStatusList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ComponentStatus, err error) ComponentStatusExpansion } @@ -61,19 +62,19 @@ func newComponentStatuses(c *CoreV1Client) *componentStatuses { } // Get takes name of the componentStatus, and returns the corresponding componentStatus object, and an error if there is any. -func (c *componentStatuses) Get(name string, options metav1.GetOptions) (result *v1.ComponentStatus, err error) { +func (c *componentStatuses) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.ComponentStatus, err error) { result = &v1.ComponentStatus{} err = c.client.Get(). Resource("componentstatuses"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of ComponentStatuses that match those selectors. -func (c *componentStatuses) List(opts metav1.ListOptions) (result *v1.ComponentStatusList, err error) { +func (c *componentStatuses) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ComponentStatusList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -83,13 +84,13 @@ func (c *componentStatuses) List(opts metav1.ListOptions) (result *v1.ComponentS Resource("componentstatuses"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested componentStatuses. -func (c *componentStatuses) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *componentStatuses) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -99,66 +100,69 @@ func (c *componentStatuses) Watch(opts metav1.ListOptions) (watch.Interface, err Resource("componentstatuses"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a componentStatus and creates it. Returns the server's representation of the componentStatus, and an error, if there is any. -func (c *componentStatuses) Create(componentStatus *v1.ComponentStatus) (result *v1.ComponentStatus, err error) { +func (c *componentStatuses) Create(ctx context.Context, componentStatus *v1.ComponentStatus, opts metav1.CreateOptions) (result *v1.ComponentStatus, err error) { result = &v1.ComponentStatus{} err = c.client.Post(). Resource("componentstatuses"). + VersionedParams(&opts, scheme.ParameterCodec). Body(componentStatus). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a componentStatus and updates it. Returns the server's representation of the componentStatus, and an error, if there is any. -func (c *componentStatuses) Update(componentStatus *v1.ComponentStatus) (result *v1.ComponentStatus, err error) { +func (c *componentStatuses) Update(ctx context.Context, componentStatus *v1.ComponentStatus, opts metav1.UpdateOptions) (result *v1.ComponentStatus, err error) { result = &v1.ComponentStatus{} err = c.client.Put(). Resource("componentstatuses"). Name(componentStatus.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(componentStatus). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the componentStatus and deletes it. Returns an error if one occurs. -func (c *componentStatuses) Delete(name string, options *metav1.DeleteOptions) error { +func (c *componentStatuses) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Resource("componentstatuses"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *componentStatuses) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *componentStatuses) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Resource("componentstatuses"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched componentStatus. -func (c *componentStatuses) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ComponentStatus, err error) { +func (c *componentStatuses) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ComponentStatus, err error) { result = &v1.ComponentStatus{} err = c.client.Patch(pt). Resource("componentstatuses"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/configmap.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/configmap.go index 18ce954a..407d25a4 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/configmap.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/configmap.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type ConfigMapsGetter interface { // ConfigMapInterface has methods to work with ConfigMap resources. type ConfigMapInterface interface { - Create(*v1.ConfigMap) (*v1.ConfigMap, error) - Update(*v1.ConfigMap) (*v1.ConfigMap, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.ConfigMap, error) - List(opts metav1.ListOptions) (*v1.ConfigMapList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ConfigMap, err error) + Create(ctx context.Context, configMap *v1.ConfigMap, opts metav1.CreateOptions) (*v1.ConfigMap, error) + Update(ctx context.Context, configMap *v1.ConfigMap, opts metav1.UpdateOptions) (*v1.ConfigMap, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.ConfigMap, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ConfigMapList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ConfigMap, err error) ConfigMapExpansion } @@ -63,20 +64,20 @@ func newConfigMaps(c *CoreV1Client, namespace string) *configMaps { } // Get takes name of the configMap, and returns the corresponding configMap object, and an error if there is any. -func (c *configMaps) Get(name string, options metav1.GetOptions) (result *v1.ConfigMap, err error) { +func (c *configMaps) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.ConfigMap, err error) { result = &v1.ConfigMap{} err = c.client.Get(). Namespace(c.ns). Resource("configmaps"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of ConfigMaps that match those selectors. -func (c *configMaps) List(opts metav1.ListOptions) (result *v1.ConfigMapList, err error) { +func (c *configMaps) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ConfigMapList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *configMaps) List(opts metav1.ListOptions) (result *v1.ConfigMapList, er Resource("configmaps"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested configMaps. -func (c *configMaps) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *configMaps) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *configMaps) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("configmaps"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a configMap and creates it. Returns the server's representation of the configMap, and an error, if there is any. -func (c *configMaps) Create(configMap *v1.ConfigMap) (result *v1.ConfigMap, err error) { +func (c *configMaps) Create(ctx context.Context, configMap *v1.ConfigMap, opts metav1.CreateOptions) (result *v1.ConfigMap, err error) { result = &v1.ConfigMap{} err = c.client.Post(). Namespace(c.ns). Resource("configmaps"). + VersionedParams(&opts, scheme.ParameterCodec). Body(configMap). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a configMap and updates it. Returns the server's representation of the configMap, and an error, if there is any. -func (c *configMaps) Update(configMap *v1.ConfigMap) (result *v1.ConfigMap, err error) { +func (c *configMaps) Update(ctx context.Context, configMap *v1.ConfigMap, opts metav1.UpdateOptions) (result *v1.ConfigMap, err error) { result = &v1.ConfigMap{} err = c.client.Put(). Namespace(c.ns). Resource("configmaps"). Name(configMap.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(configMap). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the configMap and deletes it. Returns an error if one occurs. -func (c *configMaps) Delete(name string, options *metav1.DeleteOptions) error { +func (c *configMaps) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("configmaps"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *configMaps) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *configMaps) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("configmaps"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched configMap. -func (c *configMaps) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ConfigMap, err error) { +func (c *configMaps) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ConfigMap, err error) { result = &v1.ConfigMap{} err = c.client.Patch(pt). Namespace(c.ns). Resource("configmaps"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/endpoints.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/endpoints.go index 978a2a19..c36eaaa4 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/endpoints.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/endpoints.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type EndpointsGetter interface { // EndpointsInterface has methods to work with Endpoints resources. type EndpointsInterface interface { - Create(*v1.Endpoints) (*v1.Endpoints, error) - Update(*v1.Endpoints) (*v1.Endpoints, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.Endpoints, error) - List(opts metav1.ListOptions) (*v1.EndpointsList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Endpoints, err error) + Create(ctx context.Context, endpoints *v1.Endpoints, opts metav1.CreateOptions) (*v1.Endpoints, error) + Update(ctx context.Context, endpoints *v1.Endpoints, opts metav1.UpdateOptions) (*v1.Endpoints, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Endpoints, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.EndpointsList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Endpoints, err error) EndpointsExpansion } @@ -63,20 +64,20 @@ func newEndpoints(c *CoreV1Client, namespace string) *endpoints { } // Get takes name of the endpoints, and returns the corresponding endpoints object, and an error if there is any. -func (c *endpoints) Get(name string, options metav1.GetOptions) (result *v1.Endpoints, err error) { +func (c *endpoints) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Endpoints, err error) { result = &v1.Endpoints{} err = c.client.Get(). Namespace(c.ns). Resource("endpoints"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Endpoints that match those selectors. -func (c *endpoints) List(opts metav1.ListOptions) (result *v1.EndpointsList, err error) { +func (c *endpoints) List(ctx context.Context, opts metav1.ListOptions) (result *v1.EndpointsList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *endpoints) List(opts metav1.ListOptions) (result *v1.EndpointsList, err Resource("endpoints"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested endpoints. -func (c *endpoints) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *endpoints) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *endpoints) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("endpoints"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a endpoints and creates it. Returns the server's representation of the endpoints, and an error, if there is any. -func (c *endpoints) Create(endpoints *v1.Endpoints) (result *v1.Endpoints, err error) { +func (c *endpoints) Create(ctx context.Context, endpoints *v1.Endpoints, opts metav1.CreateOptions) (result *v1.Endpoints, err error) { result = &v1.Endpoints{} err = c.client.Post(). Namespace(c.ns). Resource("endpoints"). + VersionedParams(&opts, scheme.ParameterCodec). Body(endpoints). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a endpoints and updates it. Returns the server's representation of the endpoints, and an error, if there is any. -func (c *endpoints) Update(endpoints *v1.Endpoints) (result *v1.Endpoints, err error) { +func (c *endpoints) Update(ctx context.Context, endpoints *v1.Endpoints, opts metav1.UpdateOptions) (result *v1.Endpoints, err error) { result = &v1.Endpoints{} err = c.client.Put(). Namespace(c.ns). Resource("endpoints"). Name(endpoints.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(endpoints). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the endpoints and deletes it. Returns an error if one occurs. -func (c *endpoints) Delete(name string, options *metav1.DeleteOptions) error { +func (c *endpoints) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("endpoints"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *endpoints) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *endpoints) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("endpoints"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched endpoints. -func (c *endpoints) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Endpoints, err error) { +func (c *endpoints) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Endpoints, err error) { result = &v1.Endpoints{} err = c.client.Patch(pt). Namespace(c.ns). Resource("endpoints"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event.go index 55cfa090..9b669920 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type EventsGetter interface { // EventInterface has methods to work with Event resources. type EventInterface interface { - Create(*v1.Event) (*v1.Event, error) - Update(*v1.Event) (*v1.Event, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.Event, error) - List(opts metav1.ListOptions) (*v1.EventList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Event, err error) + Create(ctx context.Context, event *v1.Event, opts metav1.CreateOptions) (*v1.Event, error) + Update(ctx context.Context, event *v1.Event, opts metav1.UpdateOptions) (*v1.Event, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Event, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.EventList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Event, err error) EventExpansion } @@ -63,20 +64,20 @@ func newEvents(c *CoreV1Client, namespace string) *events { } // Get takes name of the event, and returns the corresponding event object, and an error if there is any. -func (c *events) Get(name string, options metav1.GetOptions) (result *v1.Event, err error) { +func (c *events) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Event, err error) { result = &v1.Event{} err = c.client.Get(). Namespace(c.ns). Resource("events"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Events that match those selectors. -func (c *events) List(opts metav1.ListOptions) (result *v1.EventList, err error) { +func (c *events) List(ctx context.Context, opts metav1.ListOptions) (result *v1.EventList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *events) List(opts metav1.ListOptions) (result *v1.EventList, err error) Resource("events"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested events. -func (c *events) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *events) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *events) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("events"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a event and creates it. Returns the server's representation of the event, and an error, if there is any. -func (c *events) Create(event *v1.Event) (result *v1.Event, err error) { +func (c *events) Create(ctx context.Context, event *v1.Event, opts metav1.CreateOptions) (result *v1.Event, err error) { result = &v1.Event{} err = c.client.Post(). Namespace(c.ns). Resource("events"). + VersionedParams(&opts, scheme.ParameterCodec). Body(event). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a event and updates it. Returns the server's representation of the event, and an error, if there is any. -func (c *events) Update(event *v1.Event) (result *v1.Event, err error) { +func (c *events) Update(ctx context.Context, event *v1.Event, opts metav1.UpdateOptions) (result *v1.Event, err error) { result = &v1.Event{} err = c.client.Put(). Namespace(c.ns). Resource("events"). Name(event.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(event). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the event and deletes it. Returns an error if one occurs. -func (c *events) Delete(name string, options *metav1.DeleteOptions) error { +func (c *events) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("events"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *events) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *events) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("events"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched event. -func (c *events) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Event, err error) { +func (c *events) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Event, err error) { result = &v1.Event{} err = c.client.Patch(pt). Namespace(c.ns). Resource("events"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event_expansion.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event_expansion.go index 5a82afa4..211cf060 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event_expansion.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/event_expansion.go @@ -17,9 +17,10 @@ limitations under the License. package v1 import ( + "context" "fmt" - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/fields" "k8s.io/apimachinery/pkg/runtime" @@ -54,7 +55,7 @@ func (e *events) CreateWithEventNamespace(event *v1.Event) (*v1.Event, error) { NamespaceIfScoped(event.Namespace, len(event.Namespace) > 0). Resource("events"). Body(event). - Do(). + Do(context.TODO()). Into(result) return result, err } @@ -71,7 +72,7 @@ func (e *events) UpdateWithEventNamespace(event *v1.Event) (*v1.Event, error) { Resource("events"). Name(event.Name). Body(event). - Do(). + Do(context.TODO()). Into(result) return result, err } @@ -91,7 +92,7 @@ func (e *events) PatchWithEventNamespace(incompleteEvent *v1.Event, data []byte) Resource("events"). Name(incompleteEvent.Name). Body(data). - Do(). + Do(context.TODO()). Into(result) return result, err } @@ -118,7 +119,7 @@ func (e *events) Search(scheme *runtime.Scheme, objOrRef runtime.Object) (*v1.Ev refUID = &stringRefUID } fieldSelector := e.GetFieldSelector(&ref.Name, &ref.Namespace, refKind, refUID) - return e.List(metav1.ListOptions{FieldSelector: fieldSelector.String()}) + return e.List(context.TODO(), metav1.ListOptions{FieldSelector: fieldSelector.String()}) } // Returns the appropriate field selector based on the API version being used to communicate with the server. diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/generated_expansion.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/generated_expansion.go index 6e8591b1..2cb81aa4 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/generated_expansion.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/generated_expansion.go @@ -37,3 +37,5 @@ type ReplicationControllerExpansion interface{} type ResourceQuotaExpansion interface{} type SecretExpansion interface{} + +type ServiceAccountExpansion interface{} diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/limitrange.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/limitrange.go index 2eeae11a..7031cd77 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/limitrange.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/limitrange.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type LimitRangesGetter interface { // LimitRangeInterface has methods to work with LimitRange resources. type LimitRangeInterface interface { - Create(*v1.LimitRange) (*v1.LimitRange, error) - Update(*v1.LimitRange) (*v1.LimitRange, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.LimitRange, error) - List(opts metav1.ListOptions) (*v1.LimitRangeList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.LimitRange, err error) + Create(ctx context.Context, limitRange *v1.LimitRange, opts metav1.CreateOptions) (*v1.LimitRange, error) + Update(ctx context.Context, limitRange *v1.LimitRange, opts metav1.UpdateOptions) (*v1.LimitRange, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.LimitRange, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.LimitRangeList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.LimitRange, err error) LimitRangeExpansion } @@ -63,20 +64,20 @@ func newLimitRanges(c *CoreV1Client, namespace string) *limitRanges { } // Get takes name of the limitRange, and returns the corresponding limitRange object, and an error if there is any. -func (c *limitRanges) Get(name string, options metav1.GetOptions) (result *v1.LimitRange, err error) { +func (c *limitRanges) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.LimitRange, err error) { result = &v1.LimitRange{} err = c.client.Get(). Namespace(c.ns). Resource("limitranges"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of LimitRanges that match those selectors. -func (c *limitRanges) List(opts metav1.ListOptions) (result *v1.LimitRangeList, err error) { +func (c *limitRanges) List(ctx context.Context, opts metav1.ListOptions) (result *v1.LimitRangeList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *limitRanges) List(opts metav1.ListOptions) (result *v1.LimitRangeList, Resource("limitranges"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested limitRanges. -func (c *limitRanges) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *limitRanges) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *limitRanges) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("limitranges"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a limitRange and creates it. Returns the server's representation of the limitRange, and an error, if there is any. -func (c *limitRanges) Create(limitRange *v1.LimitRange) (result *v1.LimitRange, err error) { +func (c *limitRanges) Create(ctx context.Context, limitRange *v1.LimitRange, opts metav1.CreateOptions) (result *v1.LimitRange, err error) { result = &v1.LimitRange{} err = c.client.Post(). Namespace(c.ns). Resource("limitranges"). + VersionedParams(&opts, scheme.ParameterCodec). Body(limitRange). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a limitRange and updates it. Returns the server's representation of the limitRange, and an error, if there is any. -func (c *limitRanges) Update(limitRange *v1.LimitRange) (result *v1.LimitRange, err error) { +func (c *limitRanges) Update(ctx context.Context, limitRange *v1.LimitRange, opts metav1.UpdateOptions) (result *v1.LimitRange, err error) { result = &v1.LimitRange{} err = c.client.Put(). Namespace(c.ns). Resource("limitranges"). Name(limitRange.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(limitRange). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the limitRange and deletes it. Returns an error if one occurs. -func (c *limitRanges) Delete(name string, options *metav1.DeleteOptions) error { +func (c *limitRanges) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("limitranges"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *limitRanges) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *limitRanges) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("limitranges"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched limitRange. -func (c *limitRanges) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.LimitRange, err error) { +func (c *limitRanges) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.LimitRange, err error) { result = &v1.LimitRange{} err = c.client.Patch(pt). Namespace(c.ns). Resource("limitranges"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace.go index 8a81fe85..55b03d65 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type NamespacesGetter interface { // NamespaceInterface has methods to work with Namespace resources. type NamespaceInterface interface { - Create(*v1.Namespace) (*v1.Namespace, error) - Update(*v1.Namespace) (*v1.Namespace, error) - UpdateStatus(*v1.Namespace) (*v1.Namespace, error) - Delete(name string, options *metav1.DeleteOptions) error - Get(name string, options metav1.GetOptions) (*v1.Namespace, error) - List(opts metav1.ListOptions) (*v1.NamespaceList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Namespace, err error) + Create(ctx context.Context, namespace *v1.Namespace, opts metav1.CreateOptions) (*v1.Namespace, error) + Update(ctx context.Context, namespace *v1.Namespace, opts metav1.UpdateOptions) (*v1.Namespace, error) + UpdateStatus(ctx context.Context, namespace *v1.Namespace, opts metav1.UpdateOptions) (*v1.Namespace, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Namespace, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.NamespaceList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Namespace, err error) NamespaceExpansion } @@ -61,19 +62,19 @@ func newNamespaces(c *CoreV1Client) *namespaces { } // Get takes name of the namespace, and returns the corresponding namespace object, and an error if there is any. -func (c *namespaces) Get(name string, options metav1.GetOptions) (result *v1.Namespace, err error) { +func (c *namespaces) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Namespace, err error) { result = &v1.Namespace{} err = c.client.Get(). Resource("namespaces"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Namespaces that match those selectors. -func (c *namespaces) List(opts metav1.ListOptions) (result *v1.NamespaceList, err error) { +func (c *namespaces) List(ctx context.Context, opts metav1.ListOptions) (result *v1.NamespaceList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -83,13 +84,13 @@ func (c *namespaces) List(opts metav1.ListOptions) (result *v1.NamespaceList, er Resource("namespaces"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested namespaces. -func (c *namespaces) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *namespaces) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -99,66 +100,69 @@ func (c *namespaces) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("namespaces"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a namespace and creates it. Returns the server's representation of the namespace, and an error, if there is any. -func (c *namespaces) Create(namespace *v1.Namespace) (result *v1.Namespace, err error) { +func (c *namespaces) Create(ctx context.Context, namespace *v1.Namespace, opts metav1.CreateOptions) (result *v1.Namespace, err error) { result = &v1.Namespace{} err = c.client.Post(). Resource("namespaces"). + VersionedParams(&opts, scheme.ParameterCodec). Body(namespace). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a namespace and updates it. Returns the server's representation of the namespace, and an error, if there is any. -func (c *namespaces) Update(namespace *v1.Namespace) (result *v1.Namespace, err error) { +func (c *namespaces) Update(ctx context.Context, namespace *v1.Namespace, opts metav1.UpdateOptions) (result *v1.Namespace, err error) { result = &v1.Namespace{} err = c.client.Put(). Resource("namespaces"). Name(namespace.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(namespace). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *namespaces) UpdateStatus(namespace *v1.Namespace) (result *v1.Namespace, err error) { +func (c *namespaces) UpdateStatus(ctx context.Context, namespace *v1.Namespace, opts metav1.UpdateOptions) (result *v1.Namespace, err error) { result = &v1.Namespace{} err = c.client.Put(). Resource("namespaces"). Name(namespace.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(namespace). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the namespace and deletes it. Returns an error if one occurs. -func (c *namespaces) Delete(name string, options *metav1.DeleteOptions) error { +func (c *namespaces) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Resource("namespaces"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched namespace. -func (c *namespaces) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Namespace, err error) { +func (c *namespaces) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Namespace, err error) { result = &v1.Namespace{} err = c.client.Patch(pt). Resource("namespaces"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace_expansion.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace_expansion.go index 17effe29..be1116db 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace_expansion.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/namespace_expansion.go @@ -16,16 +16,22 @@ limitations under the License. package v1 -import "k8s.io/api/core/v1" +import ( + "context" + + v1 "k8s.io/api/core/v1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" + scheme "k8s.io/client-go/kubernetes/scheme" +) // The NamespaceExpansion interface allows manually adding extra methods to the NamespaceInterface. type NamespaceExpansion interface { - Finalize(item *v1.Namespace) (*v1.Namespace, error) + Finalize(ctx context.Context, item *v1.Namespace, opts metav1.UpdateOptions) (*v1.Namespace, error) } // Finalize takes the representation of a namespace to update. Returns the server's representation of the namespace, and an error, if it occurs. -func (c *namespaces) Finalize(namespace *v1.Namespace) (result *v1.Namespace, err error) { +func (c *namespaces) Finalize(ctx context.Context, namespace *v1.Namespace, opts metav1.UpdateOptions) (result *v1.Namespace, err error) { result = &v1.Namespace{} - err = c.client.Put().Resource("namespaces").Name(namespace.Name).SubResource("finalize").Body(namespace).Do().Into(result) + err = c.client.Put().Resource("namespaces").Name(namespace.Name).VersionedParams(&opts, scheme.ParameterCodec).SubResource("finalize").Body(namespace).Do(ctx).Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node.go index d19fab89..6176808f 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,15 +38,15 @@ type NodesGetter interface { // NodeInterface has methods to work with Node resources. type NodeInterface interface { - Create(*v1.Node) (*v1.Node, error) - Update(*v1.Node) (*v1.Node, error) - UpdateStatus(*v1.Node) (*v1.Node, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.Node, error) - List(opts metav1.ListOptions) (*v1.NodeList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Node, err error) + Create(ctx context.Context, node *v1.Node, opts metav1.CreateOptions) (*v1.Node, error) + Update(ctx context.Context, node *v1.Node, opts metav1.UpdateOptions) (*v1.Node, error) + UpdateStatus(ctx context.Context, node *v1.Node, opts metav1.UpdateOptions) (*v1.Node, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Node, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.NodeList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Node, err error) NodeExpansion } @@ -62,19 +63,19 @@ func newNodes(c *CoreV1Client) *nodes { } // Get takes name of the node, and returns the corresponding node object, and an error if there is any. -func (c *nodes) Get(name string, options metav1.GetOptions) (result *v1.Node, err error) { +func (c *nodes) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Node, err error) { result = &v1.Node{} err = c.client.Get(). Resource("nodes"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Nodes that match those selectors. -func (c *nodes) List(opts metav1.ListOptions) (result *v1.NodeList, err error) { +func (c *nodes) List(ctx context.Context, opts metav1.ListOptions) (result *v1.NodeList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -84,13 +85,13 @@ func (c *nodes) List(opts metav1.ListOptions) (result *v1.NodeList, err error) { Resource("nodes"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested nodes. -func (c *nodes) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *nodes) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -100,81 +101,84 @@ func (c *nodes) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("nodes"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a node and creates it. Returns the server's representation of the node, and an error, if there is any. -func (c *nodes) Create(node *v1.Node) (result *v1.Node, err error) { +func (c *nodes) Create(ctx context.Context, node *v1.Node, opts metav1.CreateOptions) (result *v1.Node, err error) { result = &v1.Node{} err = c.client.Post(). Resource("nodes"). + VersionedParams(&opts, scheme.ParameterCodec). Body(node). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a node and updates it. Returns the server's representation of the node, and an error, if there is any. -func (c *nodes) Update(node *v1.Node) (result *v1.Node, err error) { +func (c *nodes) Update(ctx context.Context, node *v1.Node, opts metav1.UpdateOptions) (result *v1.Node, err error) { result = &v1.Node{} err = c.client.Put(). Resource("nodes"). Name(node.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(node). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *nodes) UpdateStatus(node *v1.Node) (result *v1.Node, err error) { +func (c *nodes) UpdateStatus(ctx context.Context, node *v1.Node, opts metav1.UpdateOptions) (result *v1.Node, err error) { result = &v1.Node{} err = c.client.Put(). Resource("nodes"). Name(node.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(node). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the node and deletes it. Returns an error if one occurs. -func (c *nodes) Delete(name string, options *metav1.DeleteOptions) error { +func (c *nodes) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Resource("nodes"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *nodes) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *nodes) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Resource("nodes"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched node. -func (c *nodes) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Node, err error) { +func (c *nodes) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Node, err error) { result = &v1.Node{} err = c.client.Patch(pt). Resource("nodes"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node_expansion.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node_expansion.go index 5db29c3f..bdf7bfed 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node_expansion.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/node_expansion.go @@ -17,7 +17,9 @@ limitations under the License. package v1 import ( - "k8s.io/api/core/v1" + "context" + + v1 "k8s.io/api/core/v1" "k8s.io/apimachinery/pkg/types" ) @@ -25,19 +27,19 @@ import ( type NodeExpansion interface { // PatchStatus modifies the status of an existing node. It returns the copy // of the node that the server returns, or an error. - PatchStatus(nodeName string, data []byte) (*v1.Node, error) + PatchStatus(ctx context.Context, nodeName string, data []byte) (*v1.Node, error) } // PatchStatus modifies the status of an existing node. It returns the copy of // the node that the server returns, or an error. -func (c *nodes) PatchStatus(nodeName string, data []byte) (*v1.Node, error) { +func (c *nodes) PatchStatus(ctx context.Context, nodeName string, data []byte) (*v1.Node, error) { result := &v1.Node{} err := c.client.Patch(types.StrategicMergePatchType). Resource("nodes"). Name(nodeName). SubResource("status"). Body(data). - Do(). + Do(ctx). Into(result) return result, err } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolume.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolume.go index 74514825..1eb9db63 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolume.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolume.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,15 +38,15 @@ type PersistentVolumesGetter interface { // PersistentVolumeInterface has methods to work with PersistentVolume resources. type PersistentVolumeInterface interface { - Create(*v1.PersistentVolume) (*v1.PersistentVolume, error) - Update(*v1.PersistentVolume) (*v1.PersistentVolume, error) - UpdateStatus(*v1.PersistentVolume) (*v1.PersistentVolume, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.PersistentVolume, error) - List(opts metav1.ListOptions) (*v1.PersistentVolumeList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.PersistentVolume, err error) + Create(ctx context.Context, persistentVolume *v1.PersistentVolume, opts metav1.CreateOptions) (*v1.PersistentVolume, error) + Update(ctx context.Context, persistentVolume *v1.PersistentVolume, opts metav1.UpdateOptions) (*v1.PersistentVolume, error) + UpdateStatus(ctx context.Context, persistentVolume *v1.PersistentVolume, opts metav1.UpdateOptions) (*v1.PersistentVolume, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.PersistentVolume, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.PersistentVolumeList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.PersistentVolume, err error) PersistentVolumeExpansion } @@ -62,19 +63,19 @@ func newPersistentVolumes(c *CoreV1Client) *persistentVolumes { } // Get takes name of the persistentVolume, and returns the corresponding persistentVolume object, and an error if there is any. -func (c *persistentVolumes) Get(name string, options metav1.GetOptions) (result *v1.PersistentVolume, err error) { +func (c *persistentVolumes) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.PersistentVolume, err error) { result = &v1.PersistentVolume{} err = c.client.Get(). Resource("persistentvolumes"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of PersistentVolumes that match those selectors. -func (c *persistentVolumes) List(opts metav1.ListOptions) (result *v1.PersistentVolumeList, err error) { +func (c *persistentVolumes) List(ctx context.Context, opts metav1.ListOptions) (result *v1.PersistentVolumeList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -84,13 +85,13 @@ func (c *persistentVolumes) List(opts metav1.ListOptions) (result *v1.Persistent Resource("persistentvolumes"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested persistentVolumes. -func (c *persistentVolumes) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *persistentVolumes) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -100,81 +101,84 @@ func (c *persistentVolumes) Watch(opts metav1.ListOptions) (watch.Interface, err Resource("persistentvolumes"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a persistentVolume and creates it. Returns the server's representation of the persistentVolume, and an error, if there is any. -func (c *persistentVolumes) Create(persistentVolume *v1.PersistentVolume) (result *v1.PersistentVolume, err error) { +func (c *persistentVolumes) Create(ctx context.Context, persistentVolume *v1.PersistentVolume, opts metav1.CreateOptions) (result *v1.PersistentVolume, err error) { result = &v1.PersistentVolume{} err = c.client.Post(). Resource("persistentvolumes"). + VersionedParams(&opts, scheme.ParameterCodec). Body(persistentVolume). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a persistentVolume and updates it. Returns the server's representation of the persistentVolume, and an error, if there is any. -func (c *persistentVolumes) Update(persistentVolume *v1.PersistentVolume) (result *v1.PersistentVolume, err error) { +func (c *persistentVolumes) Update(ctx context.Context, persistentVolume *v1.PersistentVolume, opts metav1.UpdateOptions) (result *v1.PersistentVolume, err error) { result = &v1.PersistentVolume{} err = c.client.Put(). Resource("persistentvolumes"). Name(persistentVolume.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(persistentVolume). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *persistentVolumes) UpdateStatus(persistentVolume *v1.PersistentVolume) (result *v1.PersistentVolume, err error) { +func (c *persistentVolumes) UpdateStatus(ctx context.Context, persistentVolume *v1.PersistentVolume, opts metav1.UpdateOptions) (result *v1.PersistentVolume, err error) { result = &v1.PersistentVolume{} err = c.client.Put(). Resource("persistentvolumes"). Name(persistentVolume.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(persistentVolume). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the persistentVolume and deletes it. Returns an error if one occurs. -func (c *persistentVolumes) Delete(name string, options *metav1.DeleteOptions) error { +func (c *persistentVolumes) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Resource("persistentvolumes"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *persistentVolumes) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *persistentVolumes) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Resource("persistentvolumes"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched persistentVolume. -func (c *persistentVolumes) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.PersistentVolume, err error) { +func (c *persistentVolumes) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.PersistentVolume, err error) { result = &v1.PersistentVolume{} err = c.client.Patch(pt). Resource("persistentvolumes"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolumeclaim.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolumeclaim.go index 410ab37d..f4e205f4 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolumeclaim.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/persistentvolumeclaim.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,15 +38,15 @@ type PersistentVolumeClaimsGetter interface { // PersistentVolumeClaimInterface has methods to work with PersistentVolumeClaim resources. type PersistentVolumeClaimInterface interface { - Create(*v1.PersistentVolumeClaim) (*v1.PersistentVolumeClaim, error) - Update(*v1.PersistentVolumeClaim) (*v1.PersistentVolumeClaim, error) - UpdateStatus(*v1.PersistentVolumeClaim) (*v1.PersistentVolumeClaim, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.PersistentVolumeClaim, error) - List(opts metav1.ListOptions) (*v1.PersistentVolumeClaimList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.PersistentVolumeClaim, err error) + Create(ctx context.Context, persistentVolumeClaim *v1.PersistentVolumeClaim, opts metav1.CreateOptions) (*v1.PersistentVolumeClaim, error) + Update(ctx context.Context, persistentVolumeClaim *v1.PersistentVolumeClaim, opts metav1.UpdateOptions) (*v1.PersistentVolumeClaim, error) + UpdateStatus(ctx context.Context, persistentVolumeClaim *v1.PersistentVolumeClaim, opts metav1.UpdateOptions) (*v1.PersistentVolumeClaim, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.PersistentVolumeClaim, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.PersistentVolumeClaimList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.PersistentVolumeClaim, err error) PersistentVolumeClaimExpansion } @@ -64,20 +65,20 @@ func newPersistentVolumeClaims(c *CoreV1Client, namespace string) *persistentVol } // Get takes name of the persistentVolumeClaim, and returns the corresponding persistentVolumeClaim object, and an error if there is any. -func (c *persistentVolumeClaims) Get(name string, options metav1.GetOptions) (result *v1.PersistentVolumeClaim, err error) { +func (c *persistentVolumeClaims) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.PersistentVolumeClaim, err error) { result = &v1.PersistentVolumeClaim{} err = c.client.Get(). Namespace(c.ns). Resource("persistentvolumeclaims"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of PersistentVolumeClaims that match those selectors. -func (c *persistentVolumeClaims) List(opts metav1.ListOptions) (result *v1.PersistentVolumeClaimList, err error) { +func (c *persistentVolumeClaims) List(ctx context.Context, opts metav1.ListOptions) (result *v1.PersistentVolumeClaimList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -88,13 +89,13 @@ func (c *persistentVolumeClaims) List(opts metav1.ListOptions) (result *v1.Persi Resource("persistentvolumeclaims"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested persistentVolumeClaims. -func (c *persistentVolumeClaims) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *persistentVolumeClaims) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -105,87 +106,90 @@ func (c *persistentVolumeClaims) Watch(opts metav1.ListOptions) (watch.Interface Resource("persistentvolumeclaims"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a persistentVolumeClaim and creates it. Returns the server's representation of the persistentVolumeClaim, and an error, if there is any. -func (c *persistentVolumeClaims) Create(persistentVolumeClaim *v1.PersistentVolumeClaim) (result *v1.PersistentVolumeClaim, err error) { +func (c *persistentVolumeClaims) Create(ctx context.Context, persistentVolumeClaim *v1.PersistentVolumeClaim, opts metav1.CreateOptions) (result *v1.PersistentVolumeClaim, err error) { result = &v1.PersistentVolumeClaim{} err = c.client.Post(). Namespace(c.ns). Resource("persistentvolumeclaims"). + VersionedParams(&opts, scheme.ParameterCodec). Body(persistentVolumeClaim). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a persistentVolumeClaim and updates it. Returns the server's representation of the persistentVolumeClaim, and an error, if there is any. -func (c *persistentVolumeClaims) Update(persistentVolumeClaim *v1.PersistentVolumeClaim) (result *v1.PersistentVolumeClaim, err error) { +func (c *persistentVolumeClaims) Update(ctx context.Context, persistentVolumeClaim *v1.PersistentVolumeClaim, opts metav1.UpdateOptions) (result *v1.PersistentVolumeClaim, err error) { result = &v1.PersistentVolumeClaim{} err = c.client.Put(). Namespace(c.ns). Resource("persistentvolumeclaims"). Name(persistentVolumeClaim.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(persistentVolumeClaim). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *persistentVolumeClaims) UpdateStatus(persistentVolumeClaim *v1.PersistentVolumeClaim) (result *v1.PersistentVolumeClaim, err error) { +func (c *persistentVolumeClaims) UpdateStatus(ctx context.Context, persistentVolumeClaim *v1.PersistentVolumeClaim, opts metav1.UpdateOptions) (result *v1.PersistentVolumeClaim, err error) { result = &v1.PersistentVolumeClaim{} err = c.client.Put(). Namespace(c.ns). Resource("persistentvolumeclaims"). Name(persistentVolumeClaim.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(persistentVolumeClaim). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the persistentVolumeClaim and deletes it. Returns an error if one occurs. -func (c *persistentVolumeClaims) Delete(name string, options *metav1.DeleteOptions) error { +func (c *persistentVolumeClaims) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("persistentvolumeclaims"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *persistentVolumeClaims) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *persistentVolumeClaims) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("persistentvolumeclaims"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched persistentVolumeClaim. -func (c *persistentVolumeClaims) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.PersistentVolumeClaim, err error) { +func (c *persistentVolumeClaims) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.PersistentVolumeClaim, err error) { result = &v1.PersistentVolumeClaim{} err = c.client.Patch(pt). Namespace(c.ns). Resource("persistentvolumeclaims"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod.go index feacd307..36092ab6 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,17 +38,17 @@ type PodsGetter interface { // PodInterface has methods to work with Pod resources. type PodInterface interface { - Create(*v1.Pod) (*v1.Pod, error) - Update(*v1.Pod) (*v1.Pod, error) - UpdateStatus(*v1.Pod) (*v1.Pod, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.Pod, error) - List(opts metav1.ListOptions) (*v1.PodList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Pod, err error) - GetEphemeralContainers(podName string, options metav1.GetOptions) (*v1.EphemeralContainers, error) - UpdateEphemeralContainers(podName string, ephemeralContainers *v1.EphemeralContainers) (*v1.EphemeralContainers, error) + Create(ctx context.Context, pod *v1.Pod, opts metav1.CreateOptions) (*v1.Pod, error) + Update(ctx context.Context, pod *v1.Pod, opts metav1.UpdateOptions) (*v1.Pod, error) + UpdateStatus(ctx context.Context, pod *v1.Pod, opts metav1.UpdateOptions) (*v1.Pod, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Pod, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.PodList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Pod, err error) + GetEphemeralContainers(ctx context.Context, podName string, options metav1.GetOptions) (*v1.EphemeralContainers, error) + UpdateEphemeralContainers(ctx context.Context, podName string, ephemeralContainers *v1.EphemeralContainers, opts metav1.UpdateOptions) (*v1.EphemeralContainers, error) PodExpansion } @@ -67,20 +68,20 @@ func newPods(c *CoreV1Client, namespace string) *pods { } // Get takes name of the pod, and returns the corresponding pod object, and an error if there is any. -func (c *pods) Get(name string, options metav1.GetOptions) (result *v1.Pod, err error) { +func (c *pods) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Pod, err error) { result = &v1.Pod{} err = c.client.Get(). Namespace(c.ns). Resource("pods"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Pods that match those selectors. -func (c *pods) List(opts metav1.ListOptions) (result *v1.PodList, err error) { +func (c *pods) List(ctx context.Context, opts metav1.ListOptions) (result *v1.PodList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -91,13 +92,13 @@ func (c *pods) List(opts metav1.ListOptions) (result *v1.PodList, err error) { Resource("pods"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested pods. -func (c *pods) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *pods) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -108,93 +109,96 @@ func (c *pods) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("pods"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a pod and creates it. Returns the server's representation of the pod, and an error, if there is any. -func (c *pods) Create(pod *v1.Pod) (result *v1.Pod, err error) { +func (c *pods) Create(ctx context.Context, pod *v1.Pod, opts metav1.CreateOptions) (result *v1.Pod, err error) { result = &v1.Pod{} err = c.client.Post(). Namespace(c.ns). Resource("pods"). + VersionedParams(&opts, scheme.ParameterCodec). Body(pod). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a pod and updates it. Returns the server's representation of the pod, and an error, if there is any. -func (c *pods) Update(pod *v1.Pod) (result *v1.Pod, err error) { +func (c *pods) Update(ctx context.Context, pod *v1.Pod, opts metav1.UpdateOptions) (result *v1.Pod, err error) { result = &v1.Pod{} err = c.client.Put(). Namespace(c.ns). Resource("pods"). Name(pod.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(pod). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *pods) UpdateStatus(pod *v1.Pod) (result *v1.Pod, err error) { +func (c *pods) UpdateStatus(ctx context.Context, pod *v1.Pod, opts metav1.UpdateOptions) (result *v1.Pod, err error) { result = &v1.Pod{} err = c.client.Put(). Namespace(c.ns). Resource("pods"). Name(pod.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(pod). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the pod and deletes it. Returns an error if one occurs. -func (c *pods) Delete(name string, options *metav1.DeleteOptions) error { +func (c *pods) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("pods"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *pods) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *pods) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("pods"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched pod. -func (c *pods) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Pod, err error) { +func (c *pods) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Pod, err error) { result = &v1.Pod{} err = c.client.Patch(pt). Namespace(c.ns). Resource("pods"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } // GetEphemeralContainers takes name of the pod, and returns the corresponding v1.EphemeralContainers object, and an error if there is any. -func (c *pods) GetEphemeralContainers(podName string, options metav1.GetOptions) (result *v1.EphemeralContainers, err error) { +func (c *pods) GetEphemeralContainers(ctx context.Context, podName string, options metav1.GetOptions) (result *v1.EphemeralContainers, err error) { result = &v1.EphemeralContainers{} err = c.client.Get(). Namespace(c.ns). @@ -202,21 +206,22 @@ func (c *pods) GetEphemeralContainers(podName string, options metav1.GetOptions) Name(podName). SubResource("ephemeralcontainers"). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // UpdateEphemeralContainers takes the top resource name and the representation of a ephemeralContainers and updates it. Returns the server's representation of the ephemeralContainers, and an error, if there is any. -func (c *pods) UpdateEphemeralContainers(podName string, ephemeralContainers *v1.EphemeralContainers) (result *v1.EphemeralContainers, err error) { +func (c *pods) UpdateEphemeralContainers(ctx context.Context, podName string, ephemeralContainers *v1.EphemeralContainers, opts metav1.UpdateOptions) (result *v1.EphemeralContainers, err error) { result = &v1.EphemeralContainers{} err = c.client.Put(). Namespace(c.ns). Resource("pods"). Name(podName). SubResource("ephemeralcontainers"). + VersionedParams(&opts, scheme.ParameterCodec). Body(ephemeralContainers). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod_expansion.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod_expansion.go index ed876be8..759fe0ff 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod_expansion.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/pod_expansion.go @@ -17,29 +17,48 @@ limitations under the License. package v1 import ( - "k8s.io/api/core/v1" + "context" + + v1 "k8s.io/api/core/v1" policy "k8s.io/api/policy/v1beta1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" + "k8s.io/apimachinery/pkg/util/net" "k8s.io/client-go/kubernetes/scheme" restclient "k8s.io/client-go/rest" ) // The PodExpansion interface allows manually adding extra methods to the PodInterface. type PodExpansion interface { - Bind(binding *v1.Binding) error - Evict(eviction *policy.Eviction) error + Bind(ctx context.Context, binding *v1.Binding, opts metav1.CreateOptions) error + Evict(ctx context.Context, eviction *policy.Eviction) error GetLogs(name string, opts *v1.PodLogOptions) *restclient.Request + ProxyGet(scheme, name, port, path string, params map[string]string) restclient.ResponseWrapper } // Bind applies the provided binding to the named pod in the current namespace (binding.Namespace is ignored). -func (c *pods) Bind(binding *v1.Binding) error { - return c.client.Post().Namespace(c.ns).Resource("pods").Name(binding.Name).SubResource("binding").Body(binding).Do().Error() +func (c *pods) Bind(ctx context.Context, binding *v1.Binding, opts metav1.CreateOptions) error { + return c.client.Post().Namespace(c.ns).Resource("pods").Name(binding.Name).VersionedParams(&opts, scheme.ParameterCodec).SubResource("binding").Body(binding).Do(ctx).Error() } -func (c *pods) Evict(eviction *policy.Eviction) error { - return c.client.Post().Namespace(c.ns).Resource("pods").Name(eviction.Name).SubResource("eviction").Body(eviction).Do().Error() +func (c *pods) Evict(ctx context.Context, eviction *policy.Eviction) error { + return c.client.Post().Namespace(c.ns).Resource("pods").Name(eviction.Name).SubResource("eviction").Body(eviction).Do(ctx).Error() } // Get constructs a request for getting the logs for a pod func (c *pods) GetLogs(name string, opts *v1.PodLogOptions) *restclient.Request { return c.client.Get().Namespace(c.ns).Name(name).Resource("pods").SubResource("log").VersionedParams(opts, scheme.ParameterCodec) } + +// ProxyGet returns a response of the pod by calling it through the proxy. +func (c *pods) ProxyGet(scheme, name, port, path string, params map[string]string) restclient.ResponseWrapper { + request := c.client.Get(). + Namespace(c.ns). + Resource("pods"). + SubResource("proxy"). + Name(net.JoinSchemeNamePort(scheme, name, port)). + Suffix(path) + for k, v := range params { + request = request.Param(k, v) + } + return request +} diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/podtemplate.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/podtemplate.go index 84d7c980..012d3b52 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/podtemplate.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/podtemplate.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type PodTemplatesGetter interface { // PodTemplateInterface has methods to work with PodTemplate resources. type PodTemplateInterface interface { - Create(*v1.PodTemplate) (*v1.PodTemplate, error) - Update(*v1.PodTemplate) (*v1.PodTemplate, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.PodTemplate, error) - List(opts metav1.ListOptions) (*v1.PodTemplateList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.PodTemplate, err error) + Create(ctx context.Context, podTemplate *v1.PodTemplate, opts metav1.CreateOptions) (*v1.PodTemplate, error) + Update(ctx context.Context, podTemplate *v1.PodTemplate, opts metav1.UpdateOptions) (*v1.PodTemplate, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.PodTemplate, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.PodTemplateList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.PodTemplate, err error) PodTemplateExpansion } @@ -63,20 +64,20 @@ func newPodTemplates(c *CoreV1Client, namespace string) *podTemplates { } // Get takes name of the podTemplate, and returns the corresponding podTemplate object, and an error if there is any. -func (c *podTemplates) Get(name string, options metav1.GetOptions) (result *v1.PodTemplate, err error) { +func (c *podTemplates) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.PodTemplate, err error) { result = &v1.PodTemplate{} err = c.client.Get(). Namespace(c.ns). Resource("podtemplates"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of PodTemplates that match those selectors. -func (c *podTemplates) List(opts metav1.ListOptions) (result *v1.PodTemplateList, err error) { +func (c *podTemplates) List(ctx context.Context, opts metav1.ListOptions) (result *v1.PodTemplateList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *podTemplates) List(opts metav1.ListOptions) (result *v1.PodTemplateList Resource("podtemplates"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested podTemplates. -func (c *podTemplates) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *podTemplates) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *podTemplates) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("podtemplates"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a podTemplate and creates it. Returns the server's representation of the podTemplate, and an error, if there is any. -func (c *podTemplates) Create(podTemplate *v1.PodTemplate) (result *v1.PodTemplate, err error) { +func (c *podTemplates) Create(ctx context.Context, podTemplate *v1.PodTemplate, opts metav1.CreateOptions) (result *v1.PodTemplate, err error) { result = &v1.PodTemplate{} err = c.client.Post(). Namespace(c.ns). Resource("podtemplates"). + VersionedParams(&opts, scheme.ParameterCodec). Body(podTemplate). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a podTemplate and updates it. Returns the server's representation of the podTemplate, and an error, if there is any. -func (c *podTemplates) Update(podTemplate *v1.PodTemplate) (result *v1.PodTemplate, err error) { +func (c *podTemplates) Update(ctx context.Context, podTemplate *v1.PodTemplate, opts metav1.UpdateOptions) (result *v1.PodTemplate, err error) { result = &v1.PodTemplate{} err = c.client.Put(). Namespace(c.ns). Resource("podtemplates"). Name(podTemplate.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(podTemplate). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the podTemplate and deletes it. Returns an error if one occurs. -func (c *podTemplates) Delete(name string, options *metav1.DeleteOptions) error { +func (c *podTemplates) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("podtemplates"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *podTemplates) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *podTemplates) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("podtemplates"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched podTemplate. -func (c *podTemplates) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.PodTemplate, err error) { +func (c *podTemplates) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.PodTemplate, err error) { result = &v1.PodTemplate{} err = c.client.Patch(pt). Namespace(c.ns). Resource("podtemplates"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/replicationcontroller.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/replicationcontroller.go index dd3182db..8e9ccd59 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/replicationcontroller.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/replicationcontroller.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" autoscalingv1 "k8s.io/api/autoscaling/v1" @@ -38,17 +39,17 @@ type ReplicationControllersGetter interface { // ReplicationControllerInterface has methods to work with ReplicationController resources. type ReplicationControllerInterface interface { - Create(*v1.ReplicationController) (*v1.ReplicationController, error) - Update(*v1.ReplicationController) (*v1.ReplicationController, error) - UpdateStatus(*v1.ReplicationController) (*v1.ReplicationController, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.ReplicationController, error) - List(opts metav1.ListOptions) (*v1.ReplicationControllerList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ReplicationController, err error) - GetScale(replicationControllerName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) - UpdateScale(replicationControllerName string, scale *autoscalingv1.Scale) (*autoscalingv1.Scale, error) + Create(ctx context.Context, replicationController *v1.ReplicationController, opts metav1.CreateOptions) (*v1.ReplicationController, error) + Update(ctx context.Context, replicationController *v1.ReplicationController, opts metav1.UpdateOptions) (*v1.ReplicationController, error) + UpdateStatus(ctx context.Context, replicationController *v1.ReplicationController, opts metav1.UpdateOptions) (*v1.ReplicationController, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.ReplicationController, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ReplicationControllerList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ReplicationController, err error) + GetScale(ctx context.Context, replicationControllerName string, options metav1.GetOptions) (*autoscalingv1.Scale, error) + UpdateScale(ctx context.Context, replicationControllerName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (*autoscalingv1.Scale, error) ReplicationControllerExpansion } @@ -68,20 +69,20 @@ func newReplicationControllers(c *CoreV1Client, namespace string) *replicationCo } // Get takes name of the replicationController, and returns the corresponding replicationController object, and an error if there is any. -func (c *replicationControllers) Get(name string, options metav1.GetOptions) (result *v1.ReplicationController, err error) { +func (c *replicationControllers) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.ReplicationController, err error) { result = &v1.ReplicationController{} err = c.client.Get(). Namespace(c.ns). Resource("replicationcontrollers"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of ReplicationControllers that match those selectors. -func (c *replicationControllers) List(opts metav1.ListOptions) (result *v1.ReplicationControllerList, err error) { +func (c *replicationControllers) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ReplicationControllerList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -92,13 +93,13 @@ func (c *replicationControllers) List(opts metav1.ListOptions) (result *v1.Repli Resource("replicationcontrollers"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested replicationControllers. -func (c *replicationControllers) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *replicationControllers) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -109,93 +110,96 @@ func (c *replicationControllers) Watch(opts metav1.ListOptions) (watch.Interface Resource("replicationcontrollers"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a replicationController and creates it. Returns the server's representation of the replicationController, and an error, if there is any. -func (c *replicationControllers) Create(replicationController *v1.ReplicationController) (result *v1.ReplicationController, err error) { +func (c *replicationControllers) Create(ctx context.Context, replicationController *v1.ReplicationController, opts metav1.CreateOptions) (result *v1.ReplicationController, err error) { result = &v1.ReplicationController{} err = c.client.Post(). Namespace(c.ns). Resource("replicationcontrollers"). + VersionedParams(&opts, scheme.ParameterCodec). Body(replicationController). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a replicationController and updates it. Returns the server's representation of the replicationController, and an error, if there is any. -func (c *replicationControllers) Update(replicationController *v1.ReplicationController) (result *v1.ReplicationController, err error) { +func (c *replicationControllers) Update(ctx context.Context, replicationController *v1.ReplicationController, opts metav1.UpdateOptions) (result *v1.ReplicationController, err error) { result = &v1.ReplicationController{} err = c.client.Put(). Namespace(c.ns). Resource("replicationcontrollers"). Name(replicationController.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(replicationController). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *replicationControllers) UpdateStatus(replicationController *v1.ReplicationController) (result *v1.ReplicationController, err error) { +func (c *replicationControllers) UpdateStatus(ctx context.Context, replicationController *v1.ReplicationController, opts metav1.UpdateOptions) (result *v1.ReplicationController, err error) { result = &v1.ReplicationController{} err = c.client.Put(). Namespace(c.ns). Resource("replicationcontrollers"). Name(replicationController.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(replicationController). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the replicationController and deletes it. Returns an error if one occurs. -func (c *replicationControllers) Delete(name string, options *metav1.DeleteOptions) error { +func (c *replicationControllers) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("replicationcontrollers"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *replicationControllers) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *replicationControllers) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("replicationcontrollers"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched replicationController. -func (c *replicationControllers) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ReplicationController, err error) { +func (c *replicationControllers) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ReplicationController, err error) { result = &v1.ReplicationController{} err = c.client.Patch(pt). Namespace(c.ns). Resource("replicationcontrollers"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } // GetScale takes name of the replicationController, and returns the corresponding autoscalingv1.Scale object, and an error if there is any. -func (c *replicationControllers) GetScale(replicationControllerName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { +func (c *replicationControllers) GetScale(ctx context.Context, replicationControllerName string, options metav1.GetOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Get(). Namespace(c.ns). @@ -203,21 +207,22 @@ func (c *replicationControllers) GetScale(replicationControllerName string, opti Name(replicationControllerName). SubResource("scale"). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // UpdateScale takes the top resource name and the representation of a scale and updates it. Returns the server's representation of the scale, and an error, if there is any. -func (c *replicationControllers) UpdateScale(replicationControllerName string, scale *autoscalingv1.Scale) (result *autoscalingv1.Scale, err error) { +func (c *replicationControllers) UpdateScale(ctx context.Context, replicationControllerName string, scale *autoscalingv1.Scale, opts metav1.UpdateOptions) (result *autoscalingv1.Scale, err error) { result = &autoscalingv1.Scale{} err = c.client.Put(). Namespace(c.ns). Resource("replicationcontrollers"). Name(replicationControllerName). SubResource("scale"). + VersionedParams(&opts, scheme.ParameterCodec). Body(scale). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/resourcequota.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/resourcequota.go index 5a178990..6a41e35f 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/resourcequota.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/resourcequota.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,15 +38,15 @@ type ResourceQuotasGetter interface { // ResourceQuotaInterface has methods to work with ResourceQuota resources. type ResourceQuotaInterface interface { - Create(*v1.ResourceQuota) (*v1.ResourceQuota, error) - Update(*v1.ResourceQuota) (*v1.ResourceQuota, error) - UpdateStatus(*v1.ResourceQuota) (*v1.ResourceQuota, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.ResourceQuota, error) - List(opts metav1.ListOptions) (*v1.ResourceQuotaList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ResourceQuota, err error) + Create(ctx context.Context, resourceQuota *v1.ResourceQuota, opts metav1.CreateOptions) (*v1.ResourceQuota, error) + Update(ctx context.Context, resourceQuota *v1.ResourceQuota, opts metav1.UpdateOptions) (*v1.ResourceQuota, error) + UpdateStatus(ctx context.Context, resourceQuota *v1.ResourceQuota, opts metav1.UpdateOptions) (*v1.ResourceQuota, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.ResourceQuota, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ResourceQuotaList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ResourceQuota, err error) ResourceQuotaExpansion } @@ -64,20 +65,20 @@ func newResourceQuotas(c *CoreV1Client, namespace string) *resourceQuotas { } // Get takes name of the resourceQuota, and returns the corresponding resourceQuota object, and an error if there is any. -func (c *resourceQuotas) Get(name string, options metav1.GetOptions) (result *v1.ResourceQuota, err error) { +func (c *resourceQuotas) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.ResourceQuota, err error) { result = &v1.ResourceQuota{} err = c.client.Get(). Namespace(c.ns). Resource("resourcequotas"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of ResourceQuotas that match those selectors. -func (c *resourceQuotas) List(opts metav1.ListOptions) (result *v1.ResourceQuotaList, err error) { +func (c *resourceQuotas) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ResourceQuotaList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -88,13 +89,13 @@ func (c *resourceQuotas) List(opts metav1.ListOptions) (result *v1.ResourceQuota Resource("resourcequotas"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested resourceQuotas. -func (c *resourceQuotas) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *resourceQuotas) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -105,87 +106,90 @@ func (c *resourceQuotas) Watch(opts metav1.ListOptions) (watch.Interface, error) Resource("resourcequotas"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a resourceQuota and creates it. Returns the server's representation of the resourceQuota, and an error, if there is any. -func (c *resourceQuotas) Create(resourceQuota *v1.ResourceQuota) (result *v1.ResourceQuota, err error) { +func (c *resourceQuotas) Create(ctx context.Context, resourceQuota *v1.ResourceQuota, opts metav1.CreateOptions) (result *v1.ResourceQuota, err error) { result = &v1.ResourceQuota{} err = c.client.Post(). Namespace(c.ns). Resource("resourcequotas"). + VersionedParams(&opts, scheme.ParameterCodec). Body(resourceQuota). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a resourceQuota and updates it. Returns the server's representation of the resourceQuota, and an error, if there is any. -func (c *resourceQuotas) Update(resourceQuota *v1.ResourceQuota) (result *v1.ResourceQuota, err error) { +func (c *resourceQuotas) Update(ctx context.Context, resourceQuota *v1.ResourceQuota, opts metav1.UpdateOptions) (result *v1.ResourceQuota, err error) { result = &v1.ResourceQuota{} err = c.client.Put(). Namespace(c.ns). Resource("resourcequotas"). Name(resourceQuota.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(resourceQuota). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *resourceQuotas) UpdateStatus(resourceQuota *v1.ResourceQuota) (result *v1.ResourceQuota, err error) { +func (c *resourceQuotas) UpdateStatus(ctx context.Context, resourceQuota *v1.ResourceQuota, opts metav1.UpdateOptions) (result *v1.ResourceQuota, err error) { result = &v1.ResourceQuota{} err = c.client.Put(). Namespace(c.ns). Resource("resourcequotas"). Name(resourceQuota.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(resourceQuota). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the resourceQuota and deletes it. Returns an error if one occurs. -func (c *resourceQuotas) Delete(name string, options *metav1.DeleteOptions) error { +func (c *resourceQuotas) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("resourcequotas"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *resourceQuotas) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *resourceQuotas) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("resourcequotas"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched resourceQuota. -func (c *resourceQuotas) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ResourceQuota, err error) { +func (c *resourceQuotas) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ResourceQuota, err error) { result = &v1.ResourceQuota{} err = c.client.Patch(pt). Namespace(c.ns). Resource("resourcequotas"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/secret.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/secret.go index 85c143b1..b2bd80ba 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/secret.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/secret.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type SecretsGetter interface { // SecretInterface has methods to work with Secret resources. type SecretInterface interface { - Create(*v1.Secret) (*v1.Secret, error) - Update(*v1.Secret) (*v1.Secret, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.Secret, error) - List(opts metav1.ListOptions) (*v1.SecretList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Secret, err error) + Create(ctx context.Context, secret *v1.Secret, opts metav1.CreateOptions) (*v1.Secret, error) + Update(ctx context.Context, secret *v1.Secret, opts metav1.UpdateOptions) (*v1.Secret, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Secret, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.SecretList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Secret, err error) SecretExpansion } @@ -63,20 +64,20 @@ func newSecrets(c *CoreV1Client, namespace string) *secrets { } // Get takes name of the secret, and returns the corresponding secret object, and an error if there is any. -func (c *secrets) Get(name string, options metav1.GetOptions) (result *v1.Secret, err error) { +func (c *secrets) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Secret, err error) { result = &v1.Secret{} err = c.client.Get(). Namespace(c.ns). Resource("secrets"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Secrets that match those selectors. -func (c *secrets) List(opts metav1.ListOptions) (result *v1.SecretList, err error) { +func (c *secrets) List(ctx context.Context, opts metav1.ListOptions) (result *v1.SecretList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *secrets) List(opts metav1.ListOptions) (result *v1.SecretList, err erro Resource("secrets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested secrets. -func (c *secrets) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *secrets) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *secrets) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("secrets"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a secret and creates it. Returns the server's representation of the secret, and an error, if there is any. -func (c *secrets) Create(secret *v1.Secret) (result *v1.Secret, err error) { +func (c *secrets) Create(ctx context.Context, secret *v1.Secret, opts metav1.CreateOptions) (result *v1.Secret, err error) { result = &v1.Secret{} err = c.client.Post(). Namespace(c.ns). Resource("secrets"). + VersionedParams(&opts, scheme.ParameterCodec). Body(secret). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a secret and updates it. Returns the server's representation of the secret, and an error, if there is any. -func (c *secrets) Update(secret *v1.Secret) (result *v1.Secret, err error) { +func (c *secrets) Update(ctx context.Context, secret *v1.Secret, opts metav1.UpdateOptions) (result *v1.Secret, err error) { result = &v1.Secret{} err = c.client.Put(). Namespace(c.ns). Resource("secrets"). Name(secret.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(secret). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the secret and deletes it. Returns an error if one occurs. -func (c *secrets) Delete(name string, options *metav1.DeleteOptions) error { +func (c *secrets) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("secrets"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *secrets) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *secrets) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("secrets"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched secret. -func (c *secrets) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Secret, err error) { +func (c *secrets) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Secret, err error) { result = &v1.Secret{} err = c.client.Patch(pt). Namespace(c.ns). Resource("secrets"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go index b0e09413..ddde2ec6 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go @@ -19,6 +19,7 @@ limitations under the License. package v1 import ( + "context" "time" v1 "k8s.io/api/core/v1" @@ -37,14 +38,14 @@ type ServicesGetter interface { // ServiceInterface has methods to work with Service resources. type ServiceInterface interface { - Create(*v1.Service) (*v1.Service, error) - Update(*v1.Service) (*v1.Service, error) - UpdateStatus(*v1.Service) (*v1.Service, error) - Delete(name string, options *metav1.DeleteOptions) error - Get(name string, options metav1.GetOptions) (*v1.Service, error) - List(opts metav1.ListOptions) (*v1.ServiceList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Service, err error) + Create(ctx context.Context, service *v1.Service, opts metav1.CreateOptions) (*v1.Service, error) + Update(ctx context.Context, service *v1.Service, opts metav1.UpdateOptions) (*v1.Service, error) + UpdateStatus(ctx context.Context, service *v1.Service, opts metav1.UpdateOptions) (*v1.Service, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.Service, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ServiceList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Service, err error) ServiceExpansion } @@ -63,20 +64,20 @@ func newServices(c *CoreV1Client, namespace string) *services { } // Get takes name of the service, and returns the corresponding service object, and an error if there is any. -func (c *services) Get(name string, options metav1.GetOptions) (result *v1.Service, err error) { +func (c *services) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.Service, err error) { result = &v1.Service{} err = c.client.Get(). Namespace(c.ns). Resource("services"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of Services that match those selectors. -func (c *services) List(opts metav1.ListOptions) (result *v1.ServiceList, err error) { +func (c *services) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ServiceList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +88,13 @@ func (c *services) List(opts metav1.ListOptions) (result *v1.ServiceList, err er Resource("services"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested services. -func (c *services) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *services) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +105,74 @@ func (c *services) Watch(opts metav1.ListOptions) (watch.Interface, error) { Resource("services"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a service and creates it. Returns the server's representation of the service, and an error, if there is any. -func (c *services) Create(service *v1.Service) (result *v1.Service, err error) { +func (c *services) Create(ctx context.Context, service *v1.Service, opts metav1.CreateOptions) (result *v1.Service, err error) { result = &v1.Service{} err = c.client.Post(). Namespace(c.ns). Resource("services"). + VersionedParams(&opts, scheme.ParameterCodec). Body(service). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a service and updates it. Returns the server's representation of the service, and an error, if there is any. -func (c *services) Update(service *v1.Service) (result *v1.Service, err error) { +func (c *services) Update(ctx context.Context, service *v1.Service, opts metav1.UpdateOptions) (result *v1.Service, err error) { result = &v1.Service{} err = c.client.Put(). Namespace(c.ns). Resource("services"). Name(service.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(service). - Do(). + Do(ctx). Into(result) return } // UpdateStatus was generated because the type contains a Status member. // Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus(). - -func (c *services) UpdateStatus(service *v1.Service) (result *v1.Service, err error) { +func (c *services) UpdateStatus(ctx context.Context, service *v1.Service, opts metav1.UpdateOptions) (result *v1.Service, err error) { result = &v1.Service{} err = c.client.Put(). Namespace(c.ns). Resource("services"). Name(service.Name). SubResource("status"). + VersionedParams(&opts, scheme.ParameterCodec). Body(service). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the service and deletes it. Returns an error if one occurs. -func (c *services) Delete(name string, options *metav1.DeleteOptions) error { +func (c *services) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("services"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched service. -func (c *services) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.Service, err error) { +func (c *services) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.Service, err error) { result = &v1.Service{} err = c.client.Patch(pt). Namespace(c.ns). Resource("services"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount.go index 50af6a21..c2ddfbfd 100644 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount.go +++ b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount.go @@ -19,8 +19,10 @@ limitations under the License. package v1 import ( + "context" "time" + authenticationv1 "k8s.io/api/authentication/v1" v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" types "k8s.io/apimachinery/pkg/types" @@ -37,14 +39,16 @@ type ServiceAccountsGetter interface { // ServiceAccountInterface has methods to work with ServiceAccount resources. type ServiceAccountInterface interface { - Create(*v1.ServiceAccount) (*v1.ServiceAccount, error) - Update(*v1.ServiceAccount) (*v1.ServiceAccount, error) - Delete(name string, options *metav1.DeleteOptions) error - DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error - Get(name string, options metav1.GetOptions) (*v1.ServiceAccount, error) - List(opts metav1.ListOptions) (*v1.ServiceAccountList, error) - Watch(opts metav1.ListOptions) (watch.Interface, error) - Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ServiceAccount, err error) + Create(ctx context.Context, serviceAccount *v1.ServiceAccount, opts metav1.CreateOptions) (*v1.ServiceAccount, error) + Update(ctx context.Context, serviceAccount *v1.ServiceAccount, opts metav1.UpdateOptions) (*v1.ServiceAccount, error) + Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error + DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error + Get(ctx context.Context, name string, opts metav1.GetOptions) (*v1.ServiceAccount, error) + List(ctx context.Context, opts metav1.ListOptions) (*v1.ServiceAccountList, error) + Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) + Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ServiceAccount, err error) + CreateToken(ctx context.Context, serviceAccountName string, tokenRequest *authenticationv1.TokenRequest, opts metav1.CreateOptions) (*authenticationv1.TokenRequest, error) + ServiceAccountExpansion } @@ -63,20 +67,20 @@ func newServiceAccounts(c *CoreV1Client, namespace string) *serviceAccounts { } // Get takes name of the serviceAccount, and returns the corresponding serviceAccount object, and an error if there is any. -func (c *serviceAccounts) Get(name string, options metav1.GetOptions) (result *v1.ServiceAccount, err error) { +func (c *serviceAccounts) Get(ctx context.Context, name string, options metav1.GetOptions) (result *v1.ServiceAccount, err error) { result = &v1.ServiceAccount{} err = c.client.Get(). Namespace(c.ns). Resource("serviceaccounts"). Name(name). VersionedParams(&options, scheme.ParameterCodec). - Do(). + Do(ctx). Into(result) return } // List takes label and field selectors, and returns the list of ServiceAccounts that match those selectors. -func (c *serviceAccounts) List(opts metav1.ListOptions) (result *v1.ServiceAccountList, err error) { +func (c *serviceAccounts) List(ctx context.Context, opts metav1.ListOptions) (result *v1.ServiceAccountList, err error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -87,13 +91,13 @@ func (c *serviceAccounts) List(opts metav1.ListOptions) (result *v1.ServiceAccou Resource("serviceaccounts"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Do(). + Do(ctx). Into(result) return } // Watch returns a watch.Interface that watches the requested serviceAccounts. -func (c *serviceAccounts) Watch(opts metav1.ListOptions) (watch.Interface, error) { +func (c *serviceAccounts) Watch(ctx context.Context, opts metav1.ListOptions) (watch.Interface, error) { var timeout time.Duration if opts.TimeoutSeconds != nil { timeout = time.Duration(*opts.TimeoutSeconds) * time.Second @@ -104,71 +108,89 @@ func (c *serviceAccounts) Watch(opts metav1.ListOptions) (watch.Interface, error Resource("serviceaccounts"). VersionedParams(&opts, scheme.ParameterCodec). Timeout(timeout). - Watch() + Watch(ctx) } // Create takes the representation of a serviceAccount and creates it. Returns the server's representation of the serviceAccount, and an error, if there is any. -func (c *serviceAccounts) Create(serviceAccount *v1.ServiceAccount) (result *v1.ServiceAccount, err error) { +func (c *serviceAccounts) Create(ctx context.Context, serviceAccount *v1.ServiceAccount, opts metav1.CreateOptions) (result *v1.ServiceAccount, err error) { result = &v1.ServiceAccount{} err = c.client.Post(). Namespace(c.ns). Resource("serviceaccounts"). + VersionedParams(&opts, scheme.ParameterCodec). Body(serviceAccount). - Do(). + Do(ctx). Into(result) return } // Update takes the representation of a serviceAccount and updates it. Returns the server's representation of the serviceAccount, and an error, if there is any. -func (c *serviceAccounts) Update(serviceAccount *v1.ServiceAccount) (result *v1.ServiceAccount, err error) { +func (c *serviceAccounts) Update(ctx context.Context, serviceAccount *v1.ServiceAccount, opts metav1.UpdateOptions) (result *v1.ServiceAccount, err error) { result = &v1.ServiceAccount{} err = c.client.Put(). Namespace(c.ns). Resource("serviceaccounts"). Name(serviceAccount.Name). + VersionedParams(&opts, scheme.ParameterCodec). Body(serviceAccount). - Do(). + Do(ctx). Into(result) return } // Delete takes name of the serviceAccount and deletes it. Returns an error if one occurs. -func (c *serviceAccounts) Delete(name string, options *metav1.DeleteOptions) error { +func (c *serviceAccounts) Delete(ctx context.Context, name string, opts metav1.DeleteOptions) error { return c.client.Delete(). Namespace(c.ns). Resource("serviceaccounts"). Name(name). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // DeleteCollection deletes a collection of objects. -func (c *serviceAccounts) DeleteCollection(options *metav1.DeleteOptions, listOptions metav1.ListOptions) error { +func (c *serviceAccounts) DeleteCollection(ctx context.Context, opts metav1.DeleteOptions, listOpts metav1.ListOptions) error { var timeout time.Duration - if listOptions.TimeoutSeconds != nil { - timeout = time.Duration(*listOptions.TimeoutSeconds) * time.Second + if listOpts.TimeoutSeconds != nil { + timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second } return c.client.Delete(). Namespace(c.ns). Resource("serviceaccounts"). - VersionedParams(&listOptions, scheme.ParameterCodec). + VersionedParams(&listOpts, scheme.ParameterCodec). Timeout(timeout). - Body(options). - Do(). + Body(&opts). + Do(ctx). Error() } // Patch applies the patch and returns the patched serviceAccount. -func (c *serviceAccounts) Patch(name string, pt types.PatchType, data []byte, subresources ...string) (result *v1.ServiceAccount, err error) { +func (c *serviceAccounts) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts metav1.PatchOptions, subresources ...string) (result *v1.ServiceAccount, err error) { result = &v1.ServiceAccount{} err = c.client.Patch(pt). Namespace(c.ns). Resource("serviceaccounts"). - SubResource(subresources...). Name(name). + SubResource(subresources...). + VersionedParams(&opts, scheme.ParameterCodec). Body(data). - Do(). + Do(ctx). + Into(result) + return +} + +// CreateToken takes the representation of a tokenRequest and creates it. Returns the server's representation of the tokenRequest, and an error, if there is any. +func (c *serviceAccounts) CreateToken(ctx context.Context, serviceAccountName string, tokenRequest *authenticationv1.TokenRequest, opts metav1.CreateOptions) (result *authenticationv1.TokenRequest, err error) { + result = &authenticationv1.TokenRequest{} + err = c.client.Post(). + Namespace(c.ns). + Resource("serviceaccounts"). + Name(serviceAccountName). + SubResource("token"). + VersionedParams(&opts, scheme.ParameterCodec). + Body(tokenRequest). + Do(ctx). Into(result) return } diff --git a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount_expansion.go b/vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount_expansion.go deleted file mode 100644 index eaf643f1..00000000 --- a/vendor/k8s.io/client-go/kubernetes/typed/core/v1/serviceaccount_expansion.go +++ /dev/null @@ -1,41 +0,0 @@ -/* -Copyright 2018 The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -package v1 - -import ( - authenticationv1 "k8s.io/api/authentication/v1" -) - -// The ServiceAccountExpansion interface allows manually adding extra methods -// to the ServiceAccountInterface. -type ServiceAccountExpansion interface { - CreateToken(name string, tr *authenticationv1.TokenRequest) (*authenticationv1.TokenRequest, error) -} - -// CreateToken creates a new token for a serviceaccount. -func (c *serviceAccounts) CreateToken(name string, tr *authenticationv1.TokenRequest) (*authenticationv1.TokenRequest, error) { - result := &authenticationv1.TokenRequest{} - err := c.client.Post(). - Namespace(c.ns). - Resource("serviceaccounts"). - SubResource("token"). - Name(name). - Body(tr). - Do(). - Into(result) - return result, err -} diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/README.md b/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/README.md index 21a39ae8..096f99e1 100644 --- a/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/README.md +++ b/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/README.md @@ -38,7 +38,13 @@ This plugin provides an integration with Azure Active Directory device flow. If * Replace `APISERVER_APPLICATION_ID` with the application ID of your `apiserver` application ID * Be sure to also (create and) select a context that uses above user - 6. The access token is acquired when first `kubectl` command is executed +6. (Optionally) the AAD token has `aud` claim with `spn:` prefix. To omit that, add following auth configuration: + + ``` + --auth-provider-arg=config-mode="1" + ``` + + 7. The access token is acquired when first `kubectl` command is executed ``` kubectl get pods diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/azure.go b/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/azure.go index e583100c..a5218029 100644 --- a/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/azure.go +++ b/vendor/k8s.io/client-go/plugin/pkg/client/auth/azure/azure.go @@ -22,17 +22,20 @@ import ( "fmt" "net/http" "os" + "strconv" "sync" "github.com/Azure/go-autorest/autorest" "github.com/Azure/go-autorest/autorest/adal" "github.com/Azure/go-autorest/autorest/azure" - "k8s.io/klog" + "k8s.io/klog/v2" "k8s.io/apimachinery/pkg/util/net" restclient "k8s.io/client-go/rest" ) +type configMode int + const ( azureTokenKey = "azureTokenKey" tokenType = "Bearer" @@ -46,6 +49,10 @@ const ( cfgExpiresOn = "expires-on" cfgEnvironment = "environment" cfgApiserverID = "apiserver-id" + cfgConfigMode = "config-mode" + + configModeDefault configMode = 0 + configModeOmitSPNPrefix configMode = 1 ) func init() { @@ -78,17 +85,37 @@ func (c *azureTokenCache) setToken(tokenKey string, token *azureToken) { } func newAzureAuthProvider(_ string, cfg map[string]string, persister restclient.AuthProviderConfigPersister) (restclient.AuthProvider, error) { - var ts tokenSource - - environment, err := azure.EnvironmentFromName(cfg[cfgEnvironment]) + var ( + ts tokenSource + environment azure.Environment + err error + mode configMode + ) + + environment, err = azure.EnvironmentFromName(cfg[cfgEnvironment]) if err != nil { environment = azure.PublicCloud } - ts, err = newAzureTokenSourceDeviceCode(environment, cfg[cfgClientID], cfg[cfgTenantID], cfg[cfgApiserverID]) + + mode = configModeDefault + if cfg[cfgConfigMode] != "" { + configModeInt, err := strconv.Atoi(cfg[cfgConfigMode]) + if err != nil { + return nil, fmt.Errorf("failed to parse %s, error: %s", cfgConfigMode, err) + } + mode = configMode(configModeInt) + switch mode { + case configModeOmitSPNPrefix: + case configModeDefault: + default: + return nil, fmt.Errorf("%s:%s is not a valid mode", cfgConfigMode, cfg[cfgConfigMode]) + } + } + ts, err = newAzureTokenSourceDeviceCode(environment, cfg[cfgClientID], cfg[cfgTenantID], cfg[cfgApiserverID], mode) if err != nil { return nil, fmt.Errorf("creating a new azure token source for device code authentication: %v", err) } - cacheSource := newAzureTokenSource(ts, cache, cfg, persister) + cacheSource := newAzureTokenSource(ts, cache, cfg, mode, persister) return &azureAuthProvider{ tokenSource: cacheSource, @@ -153,22 +180,25 @@ type azureToken struct { type tokenSource interface { Token() (*azureToken, error) + Refresh(*azureToken) (*azureToken, error) } type azureTokenSource struct { - source tokenSource - cache *azureTokenCache - lock sync.Mutex - cfg map[string]string - persister restclient.AuthProviderConfigPersister + source tokenSource + cache *azureTokenCache + lock sync.Mutex + configMode configMode + cfg map[string]string + persister restclient.AuthProviderConfigPersister } -func newAzureTokenSource(source tokenSource, cache *azureTokenCache, cfg map[string]string, persister restclient.AuthProviderConfigPersister) tokenSource { +func newAzureTokenSource(source tokenSource, cache *azureTokenCache, cfg map[string]string, configMode configMode, persister restclient.AuthProviderConfigPersister) tokenSource { return &azureTokenSource{ - source: source, - cache: cache, - cfg: cfg, - persister: persister, + source: source, + cache: cache, + cfg: cfg, + persister: persister, + configMode: configMode, } } @@ -181,33 +211,66 @@ func (ts *azureTokenSource) Token() (*azureToken, error) { var err error token := ts.cache.getToken(azureTokenKey) + + if token != nil && !token.token.IsExpired() { + return token, nil + } + + // retrieve from config if no cache if token == nil { - token, err = ts.retrieveTokenFromCfg() - if err != nil { - token, err = ts.source.Token() - if err != nil { - return nil, fmt.Errorf("acquiring a new fresh token: %v", err) - } + tokenFromCfg, err := ts.retrieveTokenFromCfg() + + if err == nil { + token = tokenFromCfg } + } + + if token != nil { + // cache and return if the token is as good + // avoids frequent persistor calls if !token.token.IsExpired() { ts.cache.setToken(azureTokenKey, token) - err = ts.storeTokenInCfg(token) - if err != nil { - return nil, fmt.Errorf("storing the token in configuration: %v", err) - } + return token, nil } - } - if token.token.IsExpired() { - token, err = ts.refreshToken(token) - if err != nil { - return nil, fmt.Errorf("refreshing the expired token: %v", err) + + klog.V(4).Info("Refreshing token.") + tokenFromRefresh, err := ts.Refresh(token) + switch { + case err == nil: + token = tokenFromRefresh + case autorest.IsTokenRefreshError(err): + klog.V(4).Infof("Failed to refresh expired token, proceed to auth: %v", err) + // reset token to nil so that the token source will be used to acquire new + token = nil + default: + return nil, fmt.Errorf("unexpected error when refreshing token: %v", err) } - ts.cache.setToken(azureTokenKey, token) - err = ts.storeTokenInCfg(token) + } + + if token == nil { + tokenFromSource, err := ts.source.Token() if err != nil { - return nil, fmt.Errorf("storing the refreshed token in configuration: %v", err) + return nil, fmt.Errorf("failed acquiring new token: %v", err) } + token = tokenFromSource } + + // sanity check + if token == nil { + return nil, fmt.Errorf("unable to acquire token") + } + + // corner condition, newly got token is valid but expired + if token.token.IsExpired() { + return nil, fmt.Errorf("newly acquired token is expired") + } + + err = ts.storeTokenInCfg(token) + if err != nil { + return nil, fmt.Errorf("storing the refreshed token in configuration: %v", err) + } + ts.cache.setToken(azureTokenKey, token) + return token, nil } @@ -232,9 +295,9 @@ func (ts *azureTokenSource) retrieveTokenFromCfg() (*azureToken, error) { if tenantID == "" { return nil, fmt.Errorf("no tenant ID in cfg: %s", cfgTenantID) } - apiserverID := ts.cfg[cfgApiserverID] - if apiserverID == "" { - return nil, fmt.Errorf("no apiserver ID in cfg: %s", apiserverID) + resourceID := ts.cfg[cfgApiserverID] + if resourceID == "" { + return nil, fmt.Errorf("no apiserver ID in cfg: %s", cfgApiserverID) } expiresIn := ts.cfg[cfgExpiresIn] if expiresIn == "" { @@ -244,6 +307,10 @@ func (ts *azureTokenSource) retrieveTokenFromCfg() (*azureToken, error) { if expiresOn == "" { return nil, fmt.Errorf("no expiresOn in cfg: %s", cfgExpiresOn) } + tokenAudience := resourceID + if ts.configMode == configModeDefault { + tokenAudience = fmt.Sprintf("spn:%s", resourceID) + } return &azureToken{ token: adal.Token{ @@ -252,13 +319,13 @@ func (ts *azureTokenSource) retrieveTokenFromCfg() (*azureToken, error) { ExpiresIn: json.Number(expiresIn), ExpiresOn: json.Number(expiresOn), NotBefore: json.Number(expiresOn), - Resource: fmt.Sprintf("spn:%s", apiserverID), + Resource: tokenAudience, Type: tokenType, }, environment: environment, clientID: clientID, tenantID: tenantID, - apiserverID: apiserverID, + apiserverID: resourceID, }, nil } @@ -272,6 +339,7 @@ func (ts *azureTokenSource) storeTokenInCfg(token *azureToken) error { newCfg[cfgApiserverID] = token.apiserverID newCfg[cfgExpiresIn] = string(token.token.ExpiresIn) newCfg[cfgExpiresOn] = string(token.token.ExpiresOn) + newCfg[cfgConfigMode] = strconv.Itoa(int(ts.configMode)) err := ts.persister.Persist(newCfg) if err != nil { @@ -281,15 +349,29 @@ func (ts *azureTokenSource) storeTokenInCfg(token *azureToken) error { return nil } -func (ts *azureTokenSource) refreshToken(token *azureToken) (*azureToken, error) { +func (ts *azureTokenSource) Refresh(token *azureToken) (*azureToken, error) { + return ts.source.Refresh(token) +} + +// refresh outdated token with adal. +// adal.RefreshTokenError will be returned if error occur during refreshing. +func (ts *azureTokenSourceDeviceCode) Refresh(token *azureToken) (*azureToken, error) { env, err := azure.EnvironmentFromName(token.environment) if err != nil { return nil, err } - oauthConfig, err := adal.NewOAuthConfig(env.ActiveDirectoryEndpoint, token.tenantID) - if err != nil { - return nil, fmt.Errorf("building the OAuth configuration for token refresh: %v", err) + var oauthConfig *adal.OAuthConfig + if ts.configMode == configModeOmitSPNPrefix { + oauthConfig, err = adal.NewOAuthConfigWithAPIVersion(env.ActiveDirectoryEndpoint, token.tenantID, nil) + if err != nil { + return nil, fmt.Errorf("building the OAuth configuration without api-version for token refresh: %v", err) + } + } else { + oauthConfig, err = adal.NewOAuthConfig(env.ActiveDirectoryEndpoint, token.tenantID) + if err != nil { + return nil, fmt.Errorf("building the OAuth configuration for token refresh: %v", err) + } } callback := func(t adal.Token) error { @@ -323,9 +405,10 @@ type azureTokenSourceDeviceCode struct { clientID string tenantID string apiserverID string + configMode configMode } -func newAzureTokenSourceDeviceCode(environment azure.Environment, clientID string, tenantID string, apiserverID string) (tokenSource, error) { +func newAzureTokenSourceDeviceCode(environment azure.Environment, clientID string, tenantID string, apiserverID string, configMode configMode) (tokenSource, error) { if clientID == "" { return nil, errors.New("client-id is empty") } @@ -340,13 +423,25 @@ func newAzureTokenSourceDeviceCode(environment azure.Environment, clientID strin clientID: clientID, tenantID: tenantID, apiserverID: apiserverID, + configMode: configMode, }, nil } func (ts *azureTokenSourceDeviceCode) Token() (*azureToken, error) { - oauthConfig, err := adal.NewOAuthConfig(ts.environment.ActiveDirectoryEndpoint, ts.tenantID) - if err != nil { - return nil, fmt.Errorf("building the OAuth configuration for device code authentication: %v", err) + var ( + oauthConfig *adal.OAuthConfig + err error + ) + if ts.configMode == configModeOmitSPNPrefix { + oauthConfig, err = adal.NewOAuthConfigWithAPIVersion(ts.environment.ActiveDirectoryEndpoint, ts.tenantID, nil) + if err != nil { + return nil, fmt.Errorf("building the OAuth configuration without api-version for device code authentication: %v", err) + } + } else { + oauthConfig, err = adal.NewOAuthConfig(ts.environment.ActiveDirectoryEndpoint, ts.tenantID) + if err != nil { + return nil, fmt.Errorf("building the OAuth configuration for device code authentication: %v", err) + } } client := &autorest.Client{} deviceCode, err := adal.InitiateDeviceAuth(client, *oauthConfig, ts.clientID, ts.apiserverID) diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/exec.go b/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/exec.go index 741729bb..627bb2de 100644 --- a/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/exec.go +++ b/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/exec.go @@ -20,6 +20,7 @@ import ( "bytes" "context" "crypto/tls" + "crypto/x509" "errors" "fmt" "io" @@ -28,6 +29,7 @@ import ( "os" "os/exec" "reflect" + "strings" "sync" "time" @@ -37,18 +39,26 @@ import ( "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apimachinery/pkg/runtime/serializer" + "k8s.io/apimachinery/pkg/util/clock" utilruntime "k8s.io/apimachinery/pkg/util/runtime" "k8s.io/client-go/pkg/apis/clientauthentication" "k8s.io/client-go/pkg/apis/clientauthentication/v1alpha1" "k8s.io/client-go/pkg/apis/clientauthentication/v1beta1" "k8s.io/client-go/tools/clientcmd/api" + "k8s.io/client-go/tools/metrics" "k8s.io/client-go/transport" "k8s.io/client-go/util/connrotation" - "k8s.io/klog" + "k8s.io/klog/v2" ) const execInfoEnv = "KUBERNETES_EXEC_INFO" const onRotateListWarningLength = 1000 +const installHintVerboseHelp = ` + +It looks like you are trying to use a client-go credential plugin that is not installed. + +To learn more about this feature, consult the documentation available at: + https://kubernetes.io/docs/reference/access-authn-authz/authentication/#client-go-credential-plugins` var scheme = runtime.NewScheme() var codecs = serializer.NewCodecFactory(scheme) @@ -106,6 +116,44 @@ func (c *cache) put(s string, a *Authenticator) *Authenticator { return a } +// sometimes rate limits how often a function f() is called. Specifically, Do() +// will run the provided function f() up to threshold times every interval +// duration. +type sometimes struct { + threshold int + interval time.Duration + + clock clock.Clock + mu sync.Mutex + + count int // times we have called f() in this window + window time.Time // beginning of current window of length interval +} + +func (s *sometimes) Do(f func()) { + s.mu.Lock() + defer s.mu.Unlock() + + now := s.clock.Now() + if s.window.IsZero() { + s.window = now + } + + // If we are no longer in our saved time window, then we get to reset our run + // count back to 0 and start increasing towards the threshold again. + if inWindow := now.Sub(s.window) < s.interval; !inWindow { + s.window = now + s.count = 0 + } + + // If we have not run the function more than threshold times in this current + // time window, we get to run it now! + if underThreshold := s.count < s.threshold; underThreshold { + s.count++ + f() + } +} + // GetAuthenticator returns an exec-based plugin for providing client credentials. func GetAuthenticator(config *api.ExecConfig) (*Authenticator, error) { return newAuthenticator(globalCache, config) @@ -127,6 +175,13 @@ func newAuthenticator(c *cache, config *api.ExecConfig) (*Authenticator, error) args: config.Args, group: gv, + installHint: config.InstallHint, + sometimes: &sometimes{ + threshold: 10, + interval: time.Hour, + clock: clock.RealClock{}, + }, + stdin: os.Stdin, stderr: os.Stderr, interactive: terminal.IsTerminal(int(os.Stdout.Fd())), @@ -150,6 +205,12 @@ type Authenticator struct { group schema.GroupVersion env []string + // Used to avoid log spew by rate limiting install hint printing. We didn't do + // this by interval based rate limiting alone since that way may have prevented + // the install hint from showing up for kubectl users. + sometimes *sometimes + installHint string + // Stubbable for testing stdin io.Reader stderr io.Writer @@ -176,6 +237,15 @@ type credentials struct { // UpdateTransportConfig updates the transport.Config to use credentials // returned by the plugin. func (a *Authenticator) UpdateTransportConfig(c *transport.Config) error { + // If a bearer token is present in the request - avoid the GetCert callback when + // setting up the transport, as that triggers the exec action if the server is + // also configured to allow client certificates for authentication. For requests + // like "kubectl get --token (token) pods" we should assume the intention is to + // use the provided token for authentication. + if c.HasTokenAuth() { + return nil + } + c.Wrap(func(rt http.RoundTripper) http.RoundTripper { return &roundTripper{a, rt} }) @@ -260,6 +330,7 @@ func (a *Authenticator) cert() (*tls.Certificate, error) { func (a *Authenticator) getCreds() (*credentials, error) { a.mu.Lock() defer a.mu.Unlock() + if a.cachedCreds != nil && !a.credsExpired() { return a.cachedCreds, nil } @@ -267,6 +338,7 @@ func (a *Authenticator) getCreds() (*credentials, error) { if err := a.refreshCredsLocked(nil); err != nil { return nil, err } + return a.cachedCreds, nil } @@ -319,7 +391,7 @@ func (a *Authenticator) refreshCredsLocked(r *clientauthentication.Response) err } if err := cmd.Run(); err != nil { - return fmt.Errorf("exec: %v", err) + return a.wrapCmdRunErrorLocked(err) } _, gvk, err := codecs.UniversalDecoder(a.group).Decode(stdout.Bytes(), nil, cred) @@ -355,6 +427,17 @@ func (a *Authenticator) refreshCredsLocked(r *clientauthentication.Response) err if err != nil { return fmt.Errorf("failed parsing client key/certificate: %v", err) } + + // Leaf is initialized to be nil: + // https://golang.org/pkg/crypto/tls/#X509KeyPair + // Leaf certificate is the first certificate: + // https://golang.org/pkg/crypto/tls/#Certificate + // Populating leaf is useful for quickly accessing the underlying x509 + // certificate values. + cert.Leaf, err = x509.ParseCertificate(cert.Certificate[0]) + if err != nil { + return fmt.Errorf("failed parsing client leaf certificate: %v", err) + } newCreds.cert = &cert } @@ -362,10 +445,52 @@ func (a *Authenticator) refreshCredsLocked(r *clientauthentication.Response) err a.cachedCreds = newCreds // Only close all connections when TLS cert rotates. Token rotation doesn't // need the extra noise. - if len(a.onRotateList) > 0 && oldCreds != nil && !reflect.DeepEqual(oldCreds.cert, a.cachedCreds.cert) { + if oldCreds != nil && !reflect.DeepEqual(oldCreds.cert, a.cachedCreds.cert) { + // Can be nil if the exec auth plugin only returned token auth. + if oldCreds.cert != nil && oldCreds.cert.Leaf != nil { + metrics.ClientCertRotationAge.Observe(time.Now().Sub(oldCreds.cert.Leaf.NotBefore)) + } for _, onRotate := range a.onRotateList { onRotate() } } + + expiry := time.Time{} + if a.cachedCreds.cert != nil && a.cachedCreds.cert.Leaf != nil { + expiry = a.cachedCreds.cert.Leaf.NotAfter + } + expirationMetrics.set(a, expiry) return nil } + +// wrapCmdRunErrorLocked pulls out the code to construct a helpful error message +// for when the exec plugin's binary fails to Run(). +// +// It must be called while holding the Authenticator's mutex. +func (a *Authenticator) wrapCmdRunErrorLocked(err error) error { + switch err.(type) { + case *exec.Error: // Binary does not exist (see exec.Error). + builder := strings.Builder{} + fmt.Fprintf(&builder, "exec: executable %s not found", a.cmd) + + a.sometimes.Do(func() { + fmt.Fprint(&builder, installHintVerboseHelp) + if a.installHint != "" { + fmt.Fprintf(&builder, "\n\n%s", a.installHint) + } + }) + + return errors.New(builder.String()) + + case *exec.ExitError: // Binary execution failed (see exec.Cmd.Run()). + e := err.(*exec.ExitError) + return fmt.Errorf( + "exec: executable %s failed with exit code %d", + a.cmd, + e.ProcessState.ExitCode(), + ) + + default: + return fmt.Errorf("exec: %v", err) + } +} diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/metrics.go b/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/metrics.go new file mode 100644 index 00000000..caf0cca3 --- /dev/null +++ b/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec/metrics.go @@ -0,0 +1,60 @@ +/* +Copyright 2018 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package exec + +import ( + "sync" + "time" + + "k8s.io/client-go/tools/metrics" +) + +type certificateExpirationTracker struct { + mu sync.RWMutex + m map[*Authenticator]time.Time + metricSet func(*time.Time) +} + +var expirationMetrics = &certificateExpirationTracker{ + m: map[*Authenticator]time.Time{}, + metricSet: func(e *time.Time) { + metrics.ClientCertExpiry.Set(e) + }, +} + +// set stores the given expiration time and updates the updates the certificate +// expiry metric to the earliest expiration time. +func (c *certificateExpirationTracker) set(a *Authenticator, t time.Time) { + c.mu.Lock() + defer c.mu.Unlock() + c.m[a] = t + + earliest := time.Time{} + for _, t := range c.m { + if t.IsZero() { + continue + } + if earliest.IsZero() || earliest.After(t) { + earliest = t + } + } + if earliest.IsZero() { + c.metricSet(nil) + } else { + c.metricSet(&earliest) + } +} diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/gcp/gcp.go b/vendor/k8s.io/client-go/plugin/pkg/client/auth/gcp/gcp.go index e44c2ada..389dc6c7 100644 --- a/vendor/k8s.io/client-go/plugin/pkg/client/auth/gcp/gcp.go +++ b/vendor/k8s.io/client-go/plugin/pkg/client/auth/gcp/gcp.go @@ -33,7 +33,7 @@ import ( "k8s.io/apimachinery/pkg/util/yaml" restclient "k8s.io/client-go/rest" "k8s.io/client-go/util/jsonpath" - "k8s.io/klog" + "k8s.io/klog/v2" ) func init() { diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/oidc/oidc.go b/vendor/k8s.io/client-go/plugin/pkg/client/auth/oidc/oidc.go index 1383a97c..8fc37692 100644 --- a/vendor/k8s.io/client-go/plugin/pkg/client/auth/oidc/oidc.go +++ b/vendor/k8s.io/client-go/plugin/pkg/client/auth/oidc/oidc.go @@ -31,11 +31,11 @@ import ( "golang.org/x/oauth2" "k8s.io/apimachinery/pkg/util/net" restclient "k8s.io/client-go/rest" - "k8s.io/klog" + "k8s.io/klog/v2" ) const ( - cfgIssuerUrl = "idp-issuer-url" + cfgIssuerURL = "idp-issuer-url" cfgClientID = "client-id" cfgClientSecret = "client-secret" cfgCertificateAuthority = "idp-certificate-authority" @@ -76,24 +76,25 @@ func newClientCache() *clientCache { } type cacheKey struct { + clusterAddress string // Canonical issuer URL string of the provider. issuerURL string clientID string } -func (c *clientCache) getClient(issuer, clientID string) (*oidcAuthProvider, bool) { +func (c *clientCache) getClient(clusterAddress, issuer, clientID string) (*oidcAuthProvider, bool) { c.mu.RLock() defer c.mu.RUnlock() - client, ok := c.cache[cacheKey{issuer, clientID}] + client, ok := c.cache[cacheKey{clusterAddress: clusterAddress, issuerURL: issuer, clientID: clientID}] return client, ok } // setClient attempts to put the client in the cache but may return any clients // with the same keys set before. This is so there's only ever one client for a provider. -func (c *clientCache) setClient(issuer, clientID string, client *oidcAuthProvider) *oidcAuthProvider { +func (c *clientCache) setClient(clusterAddress, issuer, clientID string, client *oidcAuthProvider) *oidcAuthProvider { c.mu.Lock() defer c.mu.Unlock() - key := cacheKey{issuer, clientID} + key := cacheKey{clusterAddress: clusterAddress, issuerURL: issuer, clientID: clientID} // If another client has already initialized a client for the given provider we want // to use that client instead of the one we're trying to set. This is so all transports @@ -107,10 +108,10 @@ func (c *clientCache) setClient(issuer, clientID string, client *oidcAuthProvide return client } -func newOIDCAuthProvider(_ string, cfg map[string]string, persister restclient.AuthProviderConfigPersister) (restclient.AuthProvider, error) { - issuer := cfg[cfgIssuerUrl] +func newOIDCAuthProvider(clusterAddress string, cfg map[string]string, persister restclient.AuthProviderConfigPersister) (restclient.AuthProvider, error) { + issuer := cfg[cfgIssuerURL] if issuer == "" { - return nil, fmt.Errorf("Must provide %s", cfgIssuerUrl) + return nil, fmt.Errorf("Must provide %s", cfgIssuerURL) } clientID := cfg[cfgClientID] @@ -119,7 +120,7 @@ func newOIDCAuthProvider(_ string, cfg map[string]string, persister restclient.A } // Check cache for existing provider. - if provider, ok := cache.getClient(issuer, clientID); ok { + if provider, ok := cache.getClient(clusterAddress, issuer, clientID); ok { return provider, nil } @@ -157,7 +158,7 @@ func newOIDCAuthProvider(_ string, cfg map[string]string, persister restclient.A persister: persister, } - return cache.setClient(issuer, clientID, provider), nil + return cache.setClient(clusterAddress, issuer, clientID, provider), nil } type oidcAuthProvider struct { @@ -215,7 +216,7 @@ func (r *roundTripper) RoundTrip(req *http.Request) (*http.Response, error) { return r.wrapped.RoundTrip(r2) } -func (t *roundTripper) WrappedRoundTripper() http.RoundTripper { return t.wrapped } +func (r *roundTripper) WrappedRoundTripper() http.RoundTripper { return r.wrapped } func (p *oidcAuthProvider) idToken() (string, error) { p.mu.Lock() @@ -239,7 +240,7 @@ func (p *oidcAuthProvider) idToken() (string, error) { } // Determine provider's OAuth2 token endpoint. - tokenURL, err := tokenEndpoint(p.client, p.cfg[cfgIssuerUrl]) + tokenURL, err := tokenEndpoint(p.client, p.cfg[cfgIssuerURL]) if err != nil { return "", err } @@ -262,7 +263,7 @@ func (p *oidcAuthProvider) idToken() (string, error) { // providers (Okta) don't return this value. // // See https://github.com/kubernetes/kubernetes/issues/36847 - return "", fmt.Errorf("token response did not contain an id_token, either the scope \"openid\" wasn't requested upon login, or the provider doesn't support id_tokens as part of the refresh response.") + return "", fmt.Errorf("token response did not contain an id_token, either the scope \"openid\" wasn't requested upon login, or the provider doesn't support id_tokens as part of the refresh response") } // Create a new config to persist. diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack.go b/vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack.go deleted file mode 100644 index fab5104e..00000000 --- a/vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack.go +++ /dev/null @@ -1,193 +0,0 @@ -/* -Copyright 2017 The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -package openstack - -import ( - "fmt" - "net/http" - "sync" - "time" - - "github.com/gophercloud/gophercloud" - "github.com/gophercloud/gophercloud/openstack" - "k8s.io/klog" - - "k8s.io/apimachinery/pkg/util/net" - restclient "k8s.io/client-go/rest" -) - -func init() { - if err := restclient.RegisterAuthProviderPlugin("openstack", newOpenstackAuthProvider); err != nil { - klog.Fatalf("Failed to register openstack auth plugin: %s", err) - } -} - -// DefaultTTLDuration is the time before a token gets expired. -const DefaultTTLDuration = 10 * time.Minute - -// openstackAuthProvider is an authprovider for openstack. this provider reads -// the environment variables to determine the client identity, and generates a -// token which will be inserted into the request header later. -type openstackAuthProvider struct { - ttl time.Duration - tokenGetter TokenGetter -} - -// TokenGetter returns a bearer token that can be inserted into request. -type TokenGetter interface { - Token() (string, error) -} - -type tokenGetter struct { - authOpt *gophercloud.AuthOptions -} - -// Token creates a token by authenticate with keystone. -func (t *tokenGetter) Token() (string, error) { - var options gophercloud.AuthOptions - var err error - if t.authOpt == nil { - // reads the config from the environment - klog.V(4).Info("reading openstack config from the environment variables") - options, err = openstack.AuthOptionsFromEnv() - if err != nil { - return "", fmt.Errorf("failed to read openstack env vars: %s", err) - } - } else { - options = *t.authOpt - } - client, err := openstack.AuthenticatedClient(options) - if err != nil { - return "", fmt.Errorf("authentication failed: %s", err) - } - return client.TokenID, nil -} - -// cachedGetter caches a token until it gets expired, after the expiration, it will -// generate another token and cache it. -type cachedGetter struct { - mutex sync.Mutex - tokenGetter TokenGetter - - token string - born time.Time - ttl time.Duration -} - -// Token returns the current available token, create a new one if expired. -func (c *cachedGetter) Token() (string, error) { - c.mutex.Lock() - defer c.mutex.Unlock() - - var err error - // no token or exceeds the TTL - if c.token == "" || time.Since(c.born) > c.ttl { - c.token, err = c.tokenGetter.Token() - if err != nil { - return "", fmt.Errorf("failed to get token: %s", err) - } - c.born = time.Now() - } - return c.token, nil -} - -// tokenRoundTripper implements the RoundTripper interface: adding the bearer token -// into the request header. -type tokenRoundTripper struct { - http.RoundTripper - - tokenGetter TokenGetter -} - -var _ net.RoundTripperWrapper = &tokenRoundTripper{} - -// RoundTrip adds the bearer token into the request. -func (t *tokenRoundTripper) RoundTrip(req *http.Request) (*http.Response, error) { - // if the authorization header already present, use it. - if req.Header.Get("Authorization") != "" { - return t.RoundTripper.RoundTrip(req) - } - - token, err := t.tokenGetter.Token() - if err == nil { - req.Header.Set("Authorization", "Bearer "+token) - } else { - klog.V(4).Infof("failed to get token: %s", err) - } - - return t.RoundTripper.RoundTrip(req) -} - -func (t *tokenRoundTripper) WrappedRoundTripper() http.RoundTripper { return t.RoundTripper } - -// newOpenstackAuthProvider creates an auth provider which works with openstack -// environment. -func newOpenstackAuthProvider(_ string, config map[string]string, persister restclient.AuthProviderConfigPersister) (restclient.AuthProvider, error) { - var ttlDuration time.Duration - var err error - - klog.Warningf("WARNING: in-tree openstack auth plugin is now deprecated. please use the \"client-keystone-auth\" kubectl/client-go credential plugin instead") - ttl, found := config["ttl"] - if !found { - ttlDuration = DefaultTTLDuration - // persist to config - config["ttl"] = ttlDuration.String() - if err = persister.Persist(config); err != nil { - return nil, fmt.Errorf("failed to persist config: %s", err) - } - } else { - ttlDuration, err = time.ParseDuration(ttl) - if err != nil { - return nil, fmt.Errorf("failed to parse ttl config: %s", err) - } - } - - authOpt := gophercloud.AuthOptions{ - IdentityEndpoint: config["identityEndpoint"], - Username: config["username"], - Password: config["password"], - DomainName: config["name"], - TenantID: config["tenantId"], - TenantName: config["tenantName"], - } - - getter := tokenGetter{} - // not empty - if (authOpt != gophercloud.AuthOptions{}) { - if len(authOpt.IdentityEndpoint) == 0 { - return nil, fmt.Errorf("empty %q in the config for openstack auth provider", "identityEndpoint") - } - getter.authOpt = &authOpt - } - - return &openstackAuthProvider{ - ttl: ttlDuration, - tokenGetter: &getter, - }, nil -} - -func (oap *openstackAuthProvider) WrapTransport(rt http.RoundTripper) http.RoundTripper { - return &tokenRoundTripper{ - RoundTripper: rt, - tokenGetter: &cachedGetter{ - tokenGetter: oap.tokenGetter, - ttl: oap.ttl, - }, - } -} - -func (oap *openstackAuthProvider) Login() error { return nil } diff --git a/vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack_stub.go b/vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack_stub.go new file mode 100644 index 00000000..6e404bed --- /dev/null +++ b/vendor/k8s.io/client-go/plugin/pkg/client/auth/openstack/openstack_stub.go @@ -0,0 +1,36 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package openstack + +import ( + "errors" + + "k8s.io/client-go/rest" + "k8s.io/klog/v2" +) + +func init() { + if err := rest.RegisterAuthProviderPlugin("openstack", newOpenstackAuthProvider); err != nil { + klog.Fatalf("Failed to register openstack auth plugin: %s", err) + } +} + +func newOpenstackAuthProvider(_ string, _ map[string]string, _ rest.AuthProviderConfigPersister) (rest.AuthProvider, error) { + return nil, errors.New(`The openstack auth plugin has been removed. +Please use the "client-keystone-auth" kubectl/client-go credential plugin instead. +See https://github.com/kubernetes/cloud-provider-openstack/blob/master/docs/using-client-keystone-auth.md for further details`) +} diff --git a/vendor/k8s.io/client-go/rest/client.go b/vendor/k8s.io/client-go/rest/client.go index 53c6abd3..f35955d4 100644 --- a/vendor/k8s.io/client-go/rest/client.go +++ b/vendor/k8s.io/client-go/rest/client.go @@ -94,6 +94,10 @@ type RESTClient struct { // overridden. rateLimiter flowcontrol.RateLimiter + // warningHandler is shared among all requests created by this client. + // If not set, defaultWarningHandler is used. + warningHandler WarningHandler + // Set specific behavior of the client. If not set http.DefaultClient will be used. Client *http.Client } diff --git a/vendor/k8s.io/client-go/rest/config.go b/vendor/k8s.io/client-go/rest/config.go index f58f5183..6e50eef5 100644 --- a/vendor/k8s.io/client-go/rest/config.go +++ b/vendor/k8s.io/client-go/rest/config.go @@ -23,6 +23,7 @@ import ( "io/ioutil" "net" "net/http" + "net/url" "os" "path/filepath" gruntime "runtime" @@ -37,7 +38,7 @@ import ( "k8s.io/client-go/transport" certutil "k8s.io/client-go/util/cert" "k8s.io/client-go/util/flowcontrol" - "k8s.io/klog" + "k8s.io/klog/v2" ) const ( @@ -122,12 +123,23 @@ type Config struct { // Rate limiter for limiting connections to the master from this client. If present overwrites QPS/Burst RateLimiter flowcontrol.RateLimiter + // WarningHandler handles warnings in server responses. + // If not set, the default warning handler is used. + WarningHandler WarningHandler + // The maximum length of time to wait before giving up on a server request. A value of zero means no timeout. Timeout time.Duration // Dial specifies the dial function for creating unencrypted TCP connections. Dial func(ctx context.Context, network, address string) (net.Conn, error) + // Proxy is the the proxy func to be used for all requests made by this + // transport. If Proxy is nil, http.ProxyFromEnvironment is used. If Proxy + // returns a nil *URL, no proxy is used. + // + // socks5 proxying does not currently support spdy streaming endpoints. + Proxy func(*http.Request) (*url.URL, error) + // Version forces a specific version to be used (if registered) // Do we need this? // Version string @@ -331,7 +343,11 @@ func RESTClientFor(config *Config) (*RESTClient, error) { Negotiator: runtime.NewClientNegotiator(config.NegotiatedSerializer, gv), } - return NewRESTClient(baseURL, versionedAPIPath, clientContent, rateLimiter, httpClient) + restClient, err := NewRESTClient(baseURL, versionedAPIPath, clientContent, rateLimiter, httpClient) + if err == nil && config.WarningHandler != nil { + restClient.warningHandler = config.WarningHandler + } + return restClient, err } // UnversionedRESTClientFor is the same as RESTClientFor, except that it allows @@ -385,7 +401,11 @@ func UnversionedRESTClientFor(config *Config) (*RESTClient, error) { Negotiator: runtime.NewClientNegotiator(config.NegotiatedSerializer, gv), } - return NewRESTClient(baseURL, versionedAPIPath, clientContent, rateLimiter, httpClient) + restClient, err := NewRESTClient(baseURL, versionedAPIPath, clientContent, rateLimiter, httpClient) + if err == nil && config.WarningHandler != nil { + restClient.warningHandler = config.WarningHandler + } + return restClient, err } // SetKubernetesDefaults sets default values on the provided client config for accessing the @@ -554,12 +574,14 @@ func AnonymousClientConfig(config *Config) *Config { NextProtos: config.TLSClientConfig.NextProtos, }, RateLimiter: config.RateLimiter, + WarningHandler: config.WarningHandler, UserAgent: config.UserAgent, DisableCompression: config.DisableCompression, QPS: config.QPS, Burst: config.Burst, Timeout: config.Timeout, Dial: config.Dial, + Proxy: config.Proxy, } } @@ -599,7 +621,9 @@ func CopyConfig(config *Config) *Config { QPS: config.QPS, Burst: config.Burst, RateLimiter: config.RateLimiter, + WarningHandler: config.WarningHandler, Timeout: config.Timeout, Dial: config.Dial, + Proxy: config.Proxy, } } diff --git a/vendor/k8s.io/client-go/rest/plugin.go b/vendor/k8s.io/client-go/rest/plugin.go index 0bc2d03f..33d146cd 100644 --- a/vendor/k8s.io/client-go/rest/plugin.go +++ b/vendor/k8s.io/client-go/rest/plugin.go @@ -21,7 +21,7 @@ import ( "net/http" "sync" - "k8s.io/klog" + "k8s.io/klog/v2" clientcmdapi "k8s.io/client-go/tools/clientcmd/api" ) diff --git a/vendor/k8s.io/client-go/rest/request.go b/vendor/k8s.io/client-go/rest/request.go index 9e0c2611..0ed7def7 100644 --- a/vendor/k8s.io/client-go/rest/request.go +++ b/vendor/k8s.io/client-go/rest/request.go @@ -30,6 +30,7 @@ import ( "reflect" "strconv" "strings" + "sync" "time" "golang.org/x/net/http2" @@ -38,12 +39,13 @@ import ( "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/apimachinery/pkg/runtime/serializer/streaming" + utilclock "k8s.io/apimachinery/pkg/util/clock" "k8s.io/apimachinery/pkg/util/net" "k8s.io/apimachinery/pkg/watch" restclientwatch "k8s.io/client-go/rest/watch" "k8s.io/client-go/tools/metrics" "k8s.io/client-go/util/flowcontrol" - "k8s.io/klog" + "k8s.io/klog/v2" ) var ( @@ -51,6 +53,9 @@ var ( // throttled (via the provided rateLimiter) for more than longThrottleLatency will // be logged. longThrottleLatency = 50 * time.Millisecond + + // extraLongThrottleLatency defines the threshold for logging requests at log level 2. + extraLongThrottleLatency = 1 * time.Second ) // HTTPClient is an interface for testing a request object. @@ -61,8 +66,8 @@ type HTTPClient interface { // ResponseWrapper is an interface for getting a response. // The response may be either accessed as a raw data (the whole output is put into memory) or as a stream. type ResponseWrapper interface { - DoRaw() ([]byte, error) - Stream() (io.ReadCloser, error) + DoRaw(context.Context) ([]byte, error) + Stream(context.Context) (io.ReadCloser, error) } // RequestConstructionError is returned when there's an error assembling a request. @@ -83,9 +88,12 @@ var noBackoff = &NoBackoff{} type Request struct { c *RESTClient + warningHandler WarningHandler + rateLimiter flowcontrol.RateLimiter backoff BackoffManager timeout time.Duration + maxRetries int // generic components accessible via method setters verb string @@ -104,9 +112,6 @@ type Request struct { // output err error body io.Reader - - // This is only used for per-request timeouts, deadlines, and cancellations. - ctx context.Context } // NewRequest creates a new request helper object for accessing runtime.Objects on a server. @@ -132,11 +137,13 @@ func NewRequest(c *RESTClient) *Request { } r := &Request{ - c: c, - rateLimiter: c.rateLimiter, - backoff: backoff, - timeout: timeout, - pathPrefix: pathPrefix, + c: c, + rateLimiter: c.rateLimiter, + backoff: backoff, + timeout: timeout, + pathPrefix: pathPrefix, + maxRetries: 10, + warningHandler: c.warningHandler, } switch { @@ -214,6 +221,13 @@ func (r *Request) BackOff(manager BackoffManager) *Request { return r } +// WarningHandler sets the handler this client uses when warning headers are encountered. +// If set to nil, this client will use the default warning handler (see SetDefaultWarningHandler). +func (r *Request) WarningHandler(handler WarningHandler) *Request { + r.warningHandler = handler + return r +} + // Throttle receives a rate-limiter and sets or replaces an existing request limiter func (r *Request) Throttle(limiter flowcontrol.RateLimiter) *Request { r.rateLimiter = limiter @@ -389,6 +403,18 @@ func (r *Request) Timeout(d time.Duration) *Request { return r } +// MaxRetries makes the request use the given integer as a ceiling of retrying upon receiving +// "Retry-After" headers and 429 status-code in the response. The default is 10 unless this +// function is specifically called with a different value. +// A zero maxRetries prevent it from doing retires and return an error immediately. +func (r *Request) MaxRetries(maxRetries int) *Request { + if maxRetries < 0 { + maxRetries = 0 + } + r.maxRetries = maxRetries + return r +} + // Body makes the request use obj as the body. Optional. // If obj is a string, try to read a file of that name. // If obj is a []byte, send it directly. @@ -438,13 +464,6 @@ func (r *Request) Body(obj interface{}) *Request { return r } -// Context adds a context to the request. Contexts are only used for -// timeouts, deadlines, and cancellations. -func (r *Request) Context(ctx context.Context) *Request { - r.ctx = ctx - return r -} - // URL returns the current working URL. func (r *Request) URL() *url.URL { p := r.pathPrefix @@ -548,29 +567,88 @@ func (r Request) finalURLTemplate() url.URL { return *url } -func (r *Request) tryThrottle() error { +func (r *Request) tryThrottle(ctx context.Context) error { if r.rateLimiter == nil { return nil } now := time.Now() - var err error - if r.ctx != nil { - err = r.rateLimiter.Wait(r.ctx) - } else { - r.rateLimiter.Accept() - } - if latency := time.Since(now); latency > longThrottleLatency { - klog.V(4).Infof("Throttling request took %v, request: %s:%s", latency, r.verb, r.URL().String()) + err := r.rateLimiter.Wait(ctx) + + latency := time.Since(now) + if latency > longThrottleLatency { + klog.V(3).Infof("Throttling request took %v, request: %s:%s", latency, r.verb, r.URL().String()) } + if latency > extraLongThrottleLatency { + // If the rate limiter latency is very high, the log message should be printed at a higher log level, + // but we use a throttled logger to prevent spamming. + globalThrottledLogger.Infof("Throttling request took %v, request: %s:%s", latency, r.verb, r.URL().String()) + } + metrics.RateLimiterLatency.Observe(r.verb, r.finalURLTemplate(), latency) return err } +type throttleSettings struct { + logLevel klog.Level + minLogInterval time.Duration + + lastLogTime time.Time + lock sync.RWMutex +} + +type throttledLogger struct { + clock utilclock.PassiveClock + settings []*throttleSettings +} + +var globalThrottledLogger = &throttledLogger{ + clock: utilclock.RealClock{}, + settings: []*throttleSettings{ + { + logLevel: 2, + minLogInterval: 1 * time.Second, + }, { + logLevel: 0, + minLogInterval: 10 * time.Second, + }, + }, +} + +func (b *throttledLogger) attemptToLog() (klog.Level, bool) { + for _, setting := range b.settings { + if bool(klog.V(setting.logLevel).Enabled()) { + // Return early without write locking if possible. + if func() bool { + setting.lock.RLock() + defer setting.lock.RUnlock() + return b.clock.Since(setting.lastLogTime) >= setting.minLogInterval + }() { + setting.lock.Lock() + defer setting.lock.Unlock() + if b.clock.Since(setting.lastLogTime) >= setting.minLogInterval { + setting.lastLogTime = b.clock.Now() + return setting.logLevel, true + } + } + return -1, false + } + } + return -1, false +} + +// Infof will write a log message at each logLevel specified by the reciever's throttleSettings +// as long as it hasn't written a log message more recently than minLogInterval. +func (b *throttledLogger) Infof(message string, args ...interface{}) { + if logLevel, ok := b.attemptToLog(); ok { + klog.V(logLevel).Infof(message, args...) + } +} + // Watch attempts to begin watching the requested location. // Returns a watch.Interface, or an error. -func (r *Request) Watch() (watch.Interface, error) { +func (r *Request) Watch(ctx context.Context) (watch.Interface, error) { // We specifically don't want to rate limit watches, so we // don't use r.rateLimiter here. if r.err != nil { @@ -582,9 +660,7 @@ func (r *Request) Watch() (watch.Interface, error) { if err != nil { return nil, err } - if r.ctx != nil { - req = req.WithContext(r.ctx) - } + req = req.WithContext(ctx) req.Header = r.headers client := r.c.Client if client == nil { @@ -603,7 +679,7 @@ func (r *Request) Watch() (watch.Interface, error) { if err != nil { // The watch stream mechanism handles many common partial data errors, so closed // connections can be retried in many cases. - if net.IsProbableEOF(err) { + if net.IsProbableEOF(err) || net.IsTimeout(err) { return watch.NewEmptyWatch(), nil } return nil, err @@ -626,6 +702,8 @@ func (r *Request) Watch() (watch.Interface, error) { return nil, err } + handleWarnings(resp.Header, r.warningHandler) + frameReader := framer.NewFrameReader(resp.Body) watchEventDecoder := streaming.NewDecoder(frameReader, streamingSerializer) @@ -659,12 +737,12 @@ func updateURLMetrics(req *Request, resp *http.Response, err error) { // Returns io.ReadCloser which could be used for streaming of the response, or an error // Any non-2xx http status code causes an error. If we get a non-2xx code, we try to convert the body into an APIStatus object. // If we can, we return that as an error. Otherwise, we create an error that lists the http status and the content of the response. -func (r *Request) Stream() (io.ReadCloser, error) { +func (r *Request) Stream(ctx context.Context) (io.ReadCloser, error) { if r.err != nil { return nil, r.err } - if err := r.tryThrottle(); err != nil { + if err := r.tryThrottle(ctx); err != nil { return nil, err } @@ -676,9 +754,7 @@ func (r *Request) Stream() (io.ReadCloser, error) { if r.body != nil { req.Body = ioutil.NopCloser(r.body) } - if r.ctx != nil { - req = req.WithContext(r.ctx) - } + req = req.WithContext(ctx) req.Header = r.headers client := r.c.Client if client == nil { @@ -700,6 +776,7 @@ func (r *Request) Stream() (io.ReadCloser, error) { switch { case (resp.StatusCode >= 200) && (resp.StatusCode < 300): + handleWarnings(resp.Header, r.warningHandler) return resp.Body, nil default: @@ -746,7 +823,7 @@ func (r *Request) requestPreflightCheck() error { // received. It handles retry behavior and up front validation of requests. It will invoke // fn at most once. It will return an error if a problem occurred prior to connecting to the // server - the provided function is responsible for handling server errors. -func (r *Request) request(fn func(*http.Request, *http.Response)) error { +func (r *Request) request(ctx context.Context, fn func(*http.Request, *http.Response)) error { //Metrics for total request latency start := time.Now() defer func() { @@ -767,26 +844,29 @@ func (r *Request) request(fn func(*http.Request, *http.Response)) error { client = http.DefaultClient } + // Throttle the first try before setting up the timeout configured on the + // client. We don't want a throttled client to return timeouts to callers + // before it makes a single request. + if err := r.tryThrottle(ctx); err != nil { + return err + } + + if r.timeout > 0 { + var cancel context.CancelFunc + ctx, cancel = context.WithTimeout(ctx, r.timeout) + defer cancel() + } + // Right now we make about ten retry attempts if we get a Retry-After response. - maxRetries := 10 retries := 0 for { + url := r.URL().String() req, err := http.NewRequest(r.verb, url, r.body) if err != nil { return err } - if r.timeout > 0 { - if r.ctx == nil { - r.ctx = context.Background() - } - var cancelFn context.CancelFunc - r.ctx, cancelFn = context.WithTimeout(r.ctx, r.timeout) - defer cancelFn() - } - if r.ctx != nil { - req = req.WithContext(r.ctx) - } + req = req.WithContext(ctx) req.Header = r.headers r.backoff.Sleep(r.backoff.CalculateBackoff(r.URL())) @@ -794,7 +874,7 @@ func (r *Request) request(fn func(*http.Request, *http.Response)) error { // We are retrying the request that we already send to apiserver // at least once before. // This request should also be throttled with the client-internal rate limiter. - if err := r.tryThrottle(); err != nil { + if err := r.tryThrottle(ctx); err != nil { return err } } @@ -806,19 +886,24 @@ func (r *Request) request(fn func(*http.Request, *http.Response)) error { r.backoff.UpdateBackoff(r.URL(), err, resp.StatusCode) } if err != nil { - // "Connection reset by peer" is usually a transient error. + // "Connection reset by peer" or "apiserver is shutting down" are usually a transient errors. // Thus in case of "GET" operations, we simply retry it. // We are not automatically retrying "write" operations, as // they are not idempotent. - if !net.IsConnectionReset(err) || r.verb != "GET" { + if r.verb != "GET" { return err } - // For the purpose of retry, we set the artificial "retry-after" response. - // TODO: Should we clean the original response if it exists? - resp = &http.Response{ - StatusCode: http.StatusInternalServerError, - Header: http.Header{"Retry-After": []string{"1"}}, - Body: ioutil.NopCloser(bytes.NewReader([]byte{})), + // For connection errors and apiserver shutdown errors retry. + if net.IsConnectionReset(err) || net.IsProbableEOF(err) { + // For the purpose of retry, we set the artificial "retry-after" response. + // TODO: Should we clean the original response if it exists? + resp = &http.Response{ + StatusCode: http.StatusInternalServerError, + Header: http.Header{"Retry-After": []string{"1"}}, + Body: ioutil.NopCloser(bytes.NewReader([]byte{})), + } + } else { + return err } } @@ -835,7 +920,7 @@ func (r *Request) request(fn func(*http.Request, *http.Response)) error { }() retries++ - if seconds, wait := checkWait(resp); wait && retries < maxRetries { + if seconds, wait := checkWait(resp); wait && retries <= r.maxRetries { if seeker, ok := r.body.(io.Seeker); ok && r.body != nil { _, err := seeker.Seek(0, 0) if err != nil { @@ -864,13 +949,9 @@ func (r *Request) request(fn func(*http.Request, *http.Response)) error { // Error type: // * If the server responds with a status: *errors.StatusError or *errors.UnexpectedObjectError // * http.Client.Do errors are returned directly. -func (r *Request) Do() Result { - if err := r.tryThrottle(); err != nil { - return Result{err: err} - } - +func (r *Request) Do(ctx context.Context) Result { var result Result - err := r.request(func(req *http.Request, resp *http.Response) { + err := r.request(ctx, func(req *http.Request, resp *http.Response) { result = r.transformResponse(resp, req) }) if err != nil { @@ -880,13 +961,9 @@ func (r *Request) Do() Result { } // DoRaw executes the request but does not process the response body. -func (r *Request) DoRaw() ([]byte, error) { - if err := r.tryThrottle(); err != nil { - return nil, err - } - +func (r *Request) DoRaw(ctx context.Context) ([]byte, error) { var result Result - err := r.request(func(req *http.Request, resp *http.Response) { + err := r.request(ctx, func(req *http.Request, resp *http.Response) { result.body, result.err = ioutil.ReadAll(resp.Body) glogBody("Response Body", result.body) if resp.StatusCode < http.StatusOK || resp.StatusCode > http.StatusPartialContent { @@ -956,6 +1033,7 @@ func (r *Request) transformResponse(resp *http.Response, req *http.Request) Resu body: body, contentType: contentType, statusCode: resp.StatusCode, + warnings: handleWarnings(resp.Header, r.warningHandler), } } } @@ -974,6 +1052,7 @@ func (r *Request) transformResponse(resp *http.Response, req *http.Request) Resu statusCode: resp.StatusCode, decoder: decoder, err: err, + warnings: handleWarnings(resp.Header, r.warningHandler), } } @@ -982,6 +1061,7 @@ func (r *Request) transformResponse(resp *http.Response, req *http.Request) Resu contentType: contentType, statusCode: resp.StatusCode, decoder: decoder, + warnings: handleWarnings(resp.Header, r.warningHandler), } } @@ -989,11 +1069,11 @@ func (r *Request) transformResponse(resp *http.Response, req *http.Request) Resu func truncateBody(body string) string { max := 0 switch { - case bool(klog.V(10)): + case bool(klog.V(10).Enabled()): return body - case bool(klog.V(9)): + case bool(klog.V(9).Enabled()): max = 10240 - case bool(klog.V(8)): + case bool(klog.V(8).Enabled()): max = 1024 } @@ -1008,7 +1088,7 @@ func truncateBody(body string) string { // allocating a new string for the body output unless necessary. Uses a simple heuristic to determine // whether the body is printable. func glogBody(prefix string, body []byte) { - if klog.V(8) { + if klog.V(8).Enabled() { if bytes.IndexFunc(body, func(r rune) bool { return r < 0x0a }) != -1 { @@ -1117,6 +1197,7 @@ func retryAfterSeconds(resp *http.Response) (int, bool) { // Result contains the result of calling Request.Do(). type Result struct { body []byte + warnings []net.WarningHeader contentType string err error statusCode int @@ -1230,6 +1311,11 @@ func (r Result) Error() error { return r.err } +// Warnings returns any warning headers received in the response +func (r Result) Warnings() []net.WarningHeader { + return r.warnings +} + // NameMayNotBe specifies strings that cannot be used as names specified as path segments (like the REST API or etcd store) var NameMayNotBe = []string{".", ".."} diff --git a/vendor/k8s.io/client-go/rest/transport.go b/vendor/k8s.io/client-go/rest/transport.go index 0800e4ec..450edc6e 100644 --- a/vendor/k8s.io/client-go/rest/transport.go +++ b/vendor/k8s.io/client-go/rest/transport.go @@ -85,7 +85,8 @@ func (c *Config) TransportConfig() (*transport.Config, error) { Groups: c.Impersonate.Groups, Extra: c.Impersonate.Extra, }, - Dial: c.Dial, + Dial: c.Dial, + Proxy: c.Proxy, } if c.ExecProvider != nil && c.AuthProvider != nil { diff --git a/vendor/k8s.io/client-go/rest/urlbackoff.go b/vendor/k8s.io/client-go/rest/urlbackoff.go index d00e42f8..2f9962d7 100644 --- a/vendor/k8s.io/client-go/rest/urlbackoff.go +++ b/vendor/k8s.io/client-go/rest/urlbackoff.go @@ -22,7 +22,7 @@ import ( "k8s.io/apimachinery/pkg/util/sets" "k8s.io/client-go/util/flowcontrol" - "k8s.io/klog" + "k8s.io/klog/v2" ) // Set of resp. Codes that we backoff for. diff --git a/vendor/k8s.io/client-go/rest/warnings.go b/vendor/k8s.io/client-go/rest/warnings.go new file mode 100644 index 00000000..45c1c3b2 --- /dev/null +++ b/vendor/k8s.io/client-go/rest/warnings.go @@ -0,0 +1,144 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package rest + +import ( + "fmt" + "io" + "net/http" + "sync" + + "k8s.io/klog/v2" + + "k8s.io/apimachinery/pkg/util/net" +) + +// WarningHandler is an interface for handling warning headers +type WarningHandler interface { + // HandleWarningHeader is called with the warn code, agent, and text when a warning header is countered. + HandleWarningHeader(code int, agent string, text string) +} + +var ( + defaultWarningHandler WarningHandler = WarningLogger{} + defaultWarningHandlerLock sync.RWMutex +) + +// SetDefaultWarningHandler sets the default handler client uses when warning headers are encountered. +// By default, warnings are printed to stderr. +func SetDefaultWarningHandler(l WarningHandler) { + defaultWarningHandlerLock.Lock() + defer defaultWarningHandlerLock.Unlock() + defaultWarningHandler = l +} +func getDefaultWarningHandler() WarningHandler { + defaultWarningHandlerLock.RLock() + defer defaultWarningHandlerLock.RUnlock() + l := defaultWarningHandler + return l +} + +// NoWarnings is an implementation of WarningHandler that suppresses warnings. +type NoWarnings struct{} + +func (NoWarnings) HandleWarningHeader(code int, agent string, message string) {} + +// WarningLogger is an implementation of WarningHandler that logs code 299 warnings +type WarningLogger struct{} + +func (WarningLogger) HandleWarningHeader(code int, agent string, message string) { + if code != 299 || len(message) == 0 { + return + } + klog.Warning(message) +} + +type warningWriter struct { + // out is the writer to output warnings to + out io.Writer + // opts contains options controlling warning output + opts WarningWriterOptions + // writtenLock guards written and writtenCount + writtenLock sync.Mutex + writtenCount int + written map[string]struct{} +} + +// WarningWriterOptions controls the behavior of a WarningHandler constructed using NewWarningWriter() +type WarningWriterOptions struct { + // Deduplicate indicates a given warning message should only be written once. + // Setting this to true in a long-running process handling many warnings can result in increased memory use. + Deduplicate bool + // Color indicates that warning output can include ANSI color codes + Color bool +} + +// NewWarningWriter returns an implementation of WarningHandler that outputs code 299 warnings to the specified writer. +func NewWarningWriter(out io.Writer, opts WarningWriterOptions) *warningWriter { + h := &warningWriter{out: out, opts: opts} + if opts.Deduplicate { + h.written = map[string]struct{}{} + } + return h +} + +const ( + yellowColor = "\u001b[33;1m" + resetColor = "\u001b[0m" +) + +// HandleWarningHeader prints warnings with code=299 to the configured writer. +func (w *warningWriter) HandleWarningHeader(code int, agent string, message string) { + if code != 299 || len(message) == 0 { + return + } + + w.writtenLock.Lock() + defer w.writtenLock.Unlock() + + if w.opts.Deduplicate { + if _, alreadyWritten := w.written[message]; alreadyWritten { + return + } + w.written[message] = struct{}{} + } + w.writtenCount++ + + if w.opts.Color { + fmt.Fprintf(w.out, "%sWarning:%s %s\n", yellowColor, resetColor, message) + } else { + fmt.Fprintf(w.out, "Warning: %s\n", message) + } +} + +func (w *warningWriter) WarningCount() int { + w.writtenLock.Lock() + defer w.writtenLock.Unlock() + return w.writtenCount +} + +func handleWarnings(headers http.Header, handler WarningHandler) []net.WarningHeader { + if handler == nil { + handler = getDefaultWarningHandler() + } + + warnings, _ := net.ParseWarningHeaders(headers["Warning"]) + for _, warning := range warnings { + handler.HandleWarningHeader(warning.Code, warning.Agent, warning.Text) + } + return warnings +} diff --git a/vendor/k8s.io/client-go/restmapper/discovery.go b/vendor/k8s.io/client-go/restmapper/discovery.go index f8d7080d..19ae95e1 100644 --- a/vendor/k8s.io/client-go/restmapper/discovery.go +++ b/vendor/k8s.io/client-go/restmapper/discovery.go @@ -26,7 +26,7 @@ import ( "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/client-go/discovery" - "k8s.io/klog" + "k8s.io/klog/v2" ) // APIGroupResources is an API group with a mapping of versions to diff --git a/vendor/k8s.io/client-go/restmapper/shortcut.go b/vendor/k8s.io/client-go/restmapper/shortcut.go index 6f3c9d93..6903ec80 100644 --- a/vendor/k8s.io/client-go/restmapper/shortcut.go +++ b/vendor/k8s.io/client-go/restmapper/shortcut.go @@ -19,7 +19,7 @@ package restmapper import ( "strings" - "k8s.io/klog" + "k8s.io/klog/v2" "k8s.io/apimachinery/pkg/api/meta" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" diff --git a/vendor/k8s.io/client-go/testing/actions.go b/vendor/k8s.io/client-go/testing/actions.go index f56b34ee..b6b2c1f2 100644 --- a/vendor/k8s.io/client-go/testing/actions.go +++ b/vendor/k8s.io/client-go/testing/actions.go @@ -439,8 +439,18 @@ func (a ActionImpl) GetSubresource() string { return a.Subresource } func (a ActionImpl) Matches(verb, resource string) bool { + // Stay backwards compatible. + if !strings.Contains(resource, "/") { + return strings.EqualFold(verb, a.Verb) && + strings.EqualFold(resource, a.Resource.Resource) + } + + parts := strings.SplitN(resource, "/", 2) + topresource, subresource := parts[0], parts[1] + return strings.EqualFold(verb, a.Verb) && - strings.EqualFold(resource, a.Resource.Resource) + strings.EqualFold(topresource, a.Resource.Resource) && + strings.EqualFold(subresource, a.Subresource) } func (a ActionImpl) DeepCopy() Action { ret := a diff --git a/vendor/k8s.io/client-go/testing/fixture.go b/vendor/k8s.io/client-go/testing/fixture.go index 54f600ad..d3b93724 100644 --- a/vendor/k8s.io/client-go/testing/fixture.go +++ b/vendor/k8s.io/client-go/testing/fixture.go @@ -19,6 +19,7 @@ package testing import ( "fmt" "reflect" + "sort" "sync" jsonpatch "github.com/evanphx/json-patch" @@ -197,7 +198,7 @@ type tracker struct { scheme ObjectScheme decoder runtime.Decoder lock sync.RWMutex - objects map[schema.GroupVersionResource][]runtime.Object + objects map[schema.GroupVersionResource]map[types.NamespacedName]runtime.Object // The value type of watchers is a map of which the key is either a namespace or // all/non namespace aka "" and its value is list of fake watchers. // Manipulations on resources will broadcast the notification events into the @@ -214,7 +215,7 @@ func NewObjectTracker(scheme ObjectScheme, decoder runtime.Decoder) ObjectTracke return &tracker{ scheme: scheme, decoder: decoder, - objects: make(map[schema.GroupVersionResource][]runtime.Object), + objects: make(map[schema.GroupVersionResource]map[types.NamespacedName]runtime.Object), watchers: make(map[schema.GroupVersionResource]map[string][]*watch.RaceFreeFakeWatcher), } } @@ -282,31 +283,15 @@ func (t *tracker) Get(gvr schema.GroupVersionResource, ns, name string) (runtime return nil, errNotFound } - var matchingObjs []runtime.Object - for _, obj := range objs { - acc, err := meta.Accessor(obj) - if err != nil { - return nil, err - } - if acc.GetNamespace() != ns { - continue - } - if acc.GetName() != name { - continue - } - matchingObjs = append(matchingObjs, obj) - } - if len(matchingObjs) == 0 { + matchingObj, ok := objs[types.NamespacedName{Namespace: ns, Name: name}] + if !ok { return nil, errNotFound } - if len(matchingObjs) > 1 { - return nil, fmt.Errorf("more than one object matched gvr %s, ns: %q name: %q", gvr, ns, name) - } // Only one object should match in the tracker if it works // correctly, as Add/Update methods enforce kind/namespace/name // uniqueness. - obj := matchingObjs[0].DeepCopyObject() + obj := matchingObj.DeepCopyObject() if status, ok := obj.(*metav1.Status); ok { if status.Status != metav1.StatusSuccess { return nil, &errors.StatusError{ErrStatus: *status} @@ -405,21 +390,21 @@ func (t *tracker) add(gvr schema.GroupVersionResource, obj runtime.Object, ns st return errors.NewBadRequest(msg) } - for i, existingObj := range t.objects[gvr] { - oldMeta, err := meta.Accessor(existingObj) - if err != nil { - return err - } - if oldMeta.GetNamespace() == newMeta.GetNamespace() && oldMeta.GetName() == newMeta.GetName() { - if replaceExisting { - for _, w := range t.getWatches(gvr, ns) { - w.Modify(obj) - } - t.objects[gvr][i] = obj - return nil + _, ok := t.objects[gvr] + if !ok { + t.objects[gvr] = make(map[types.NamespacedName]runtime.Object) + } + + namespacedName := types.NamespacedName{Namespace: newMeta.GetNamespace(), Name: newMeta.GetName()} + if _, ok = t.objects[gvr][namespacedName]; ok { + if replaceExisting { + for _, w := range t.getWatches(gvr, ns) { + w.Modify(obj) } - return errors.NewAlreadyExists(gr, newMeta.GetName()) + t.objects[gvr][namespacedName] = obj + return nil } + return errors.NewAlreadyExists(gr, newMeta.GetName()) } if replaceExisting { @@ -427,7 +412,7 @@ func (t *tracker) add(gvr schema.GroupVersionResource, obj runtime.Object, ns st return errors.NewNotFound(gr, newMeta.GetName()) } - t.objects[gvr] = append(t.objects[gvr], obj) + t.objects[gvr][namespacedName] = obj for _, w := range t.getWatches(gvr, ns) { w.Add(obj) @@ -457,35 +442,28 @@ func (t *tracker) Delete(gvr schema.GroupVersionResource, ns, name string) error t.lock.Lock() defer t.lock.Unlock() - found := false - - for i, existingObj := range t.objects[gvr] { - objMeta, err := meta.Accessor(existingObj) - if err != nil { - return err - } - if objMeta.GetNamespace() == ns && objMeta.GetName() == name { - obj := t.objects[gvr][i] - t.objects[gvr] = append(t.objects[gvr][:i], t.objects[gvr][i+1:]...) - for _, w := range t.getWatches(gvr, ns) { - w.Delete(obj) - } - found = true - break - } + objs, ok := t.objects[gvr] + if !ok { + return errors.NewNotFound(gvr.GroupResource(), name) } - if found { - return nil + namespacedName := types.NamespacedName{Namespace: ns, Name: name} + obj, ok := objs[namespacedName] + if !ok { + return errors.NewNotFound(gvr.GroupResource(), name) } - return errors.NewNotFound(gvr.GroupResource(), name) + delete(objs, namespacedName) + for _, w := range t.getWatches(gvr, ns) { + w.Delete(obj) + } + return nil } // filterByNamespace returns all objects in the collection that // match provided namespace. Empty namespace matches // non-namespaced objects. -func filterByNamespace(objs []runtime.Object, ns string) ([]runtime.Object, error) { +func filterByNamespace(objs map[types.NamespacedName]runtime.Object, ns string) ([]runtime.Object, error) { var res []runtime.Object for _, obj := range objs { @@ -499,6 +477,15 @@ func filterByNamespace(objs []runtime.Object, ns string) ([]runtime.Object, erro res = append(res, obj) } + // Sort res to get deterministic order. + sort.Slice(res, func(i, j int) bool { + acc1, _ := meta.Accessor(res[i]) + acc2, _ := meta.Accessor(res[j]) + if acc1.GetNamespace() != acc2.GetNamespace() { + return acc1.GetNamespace() < acc2.GetNamespace() + } + return acc1.GetName() < acc2.GetName() + }) return res, nil } diff --git a/vendor/k8s.io/client-go/tools/clientcmd/api/helpers.go b/vendor/k8s.io/client-go/tools/clientcmd/api/helpers.go index 65a36936..d677d651 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/api/helpers.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/api/helpers.go @@ -98,6 +98,9 @@ func ShortenConfig(config *Config) { if len(authInfo.ClientCertificateData) > 0 { authInfo.ClientCertificateData = redactedBytes } + if len(authInfo.Token) > 0 { + authInfo.Token = "REDACTED" + } config.AuthInfos[key] = authInfo } for key, cluster := range config.Clusters { diff --git a/vendor/k8s.io/client-go/tools/clientcmd/api/types.go b/vendor/k8s.io/client-go/tools/clientcmd/api/types.go index 1f1209f8..829424dc 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/api/types.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/api/types.go @@ -70,6 +70,9 @@ type Cluster struct { LocationOfOrigin string // Server is the address of the kubernetes cluster (https://hostname:port). Server string `json:"server"` + // TLSServerName is used to check server certificate. If TLSServerName is empty, the hostname used to contact the server is used. + // +optional + TLSServerName string `json:"tls-server-name,omitempty"` // InsecureSkipTLSVerify skips the validity check for the server's certificate. This will make your HTTPS connections insecure. // +optional InsecureSkipTLSVerify bool `json:"insecure-skip-tls-verify,omitempty"` @@ -79,6 +82,17 @@ type Cluster struct { // CertificateAuthorityData contains PEM-encoded certificate authority certificates. Overrides CertificateAuthority // +optional CertificateAuthorityData []byte `json:"certificate-authority-data,omitempty"` + // ProxyURL is the URL to the proxy to be used for all requests made by this + // client. URLs with "http", "https", and "socks5" schemes are supported. If + // this configuration is not provided or the empty string, the client + // attempts to construct a proxy configuration from http_proxy and + // https_proxy environment variables. If these environment variables are not + // set, the client does not attempt to proxy requests. + // + // socks5 proxying does not currently support spdy streaming endpoints (exec, + // attach, port forward). + // +optional + ProxyURL string `json:"proxy-url,omitempty"` // Extensions holds additional information. This is useful for extenders so that reads and writes don't clobber unknown fields // +optional Extensions map[string]runtime.Object `json:"extensions,omitempty"` @@ -179,7 +193,7 @@ func (c AuthProviderConfig) String() string { // ExecConfig specifies a command to provide client credentials. The command is exec'd // and outputs structured stdout holding credentials. // -// See the client.authentiction.k8s.io API group for specifications of the exact input +// See the client.authentication.k8s.io API group for specifications of the exact input // and output format type ExecConfig struct { // Command to execute. @@ -196,6 +210,11 @@ type ExecConfig struct { // Preferred input version of the ExecInfo. The returned ExecCredentials MUST use // the same encoding version as the input. APIVersion string `json:"apiVersion,omitempty"` + + // This text is shown to the user when the executable doesn't seem to be + // present. For example, `brew install foo-cli` might be a good InstallHint for + // foo-cli on Mac OS systems. + InstallHint string `json:"installHint,omitempty"` } var _ fmt.Stringer = new(ExecConfig) diff --git a/vendor/k8s.io/client-go/tools/clientcmd/api/v1/types.go b/vendor/k8s.io/client-go/tools/clientcmd/api/v1/types.go index 2159ffc7..0395f860 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/api/v1/types.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/api/v1/types.go @@ -63,6 +63,9 @@ type Preferences struct { type Cluster struct { // Server is the address of the kubernetes cluster (https://hostname:port). Server string `json:"server"` + // TLSServerName is used to check server certificate. If TLSServerName is empty, the hostname used to contact the server is used. + // +optional + TLSServerName string `json:"tls-server-name,omitempty"` // InsecureSkipTLSVerify skips the validity check for the server's certificate. This will make your HTTPS connections insecure. // +optional InsecureSkipTLSVerify bool `json:"insecure-skip-tls-verify,omitempty"` @@ -72,6 +75,17 @@ type Cluster struct { // CertificateAuthorityData contains PEM-encoded certificate authority certificates. Overrides CertificateAuthority // +optional CertificateAuthorityData []byte `json:"certificate-authority-data,omitempty"` + // ProxyURL is the URL to the proxy to be used for all requests made by this + // client. URLs with "http", "https", and "socks5" schemes are supported. If + // this configuration is not provided or the empty string, the client + // attempts to construct a proxy configuration from http_proxy and + // https_proxy environment variables. If these environment variables are not + // set, the client does not attempt to proxy requests. + // + // socks5 proxying does not currently support spdy streaming endpoints (exec, + // attach, port forward). + // +optional + ProxyURL string `json:"proxy-url,omitempty"` // Extensions holds additional information. This is useful for extenders so that reads and writes don't clobber unknown fields // +optional Extensions []NamedExtension `json:"extensions,omitempty"` @@ -178,7 +192,7 @@ type AuthProviderConfig struct { // ExecConfig specifies a command to provide client credentials. The command is exec'd // and outputs structured stdout holding credentials. // -// See the client.authentiction.k8s.io API group for specifications of the exact input +// See the client.authentication.k8s.io API group for specifications of the exact input // and output format type ExecConfig struct { // Command to execute. @@ -195,6 +209,11 @@ type ExecConfig struct { // Preferred input version of the ExecInfo. The returned ExecCredentials MUST use // the same encoding version as the input. APIVersion string `json:"apiVersion,omitempty"` + + // This text is shown to the user when the executable doesn't seem to be + // present. For example, `brew install foo-cli` might be a good InstallHint for + // foo-cli on Mac OS systems. + InstallHint string `json:"installHint,omitempty"` } // ExecEnvVar is used for setting environment variables when executing an exec-based diff --git a/vendor/k8s.io/client-go/tools/clientcmd/api/v1/zz_generated.conversion.go b/vendor/k8s.io/client-go/tools/clientcmd/api/v1/zz_generated.conversion.go index 31e00ea6..bf9eaeca 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/api/v1/zz_generated.conversion.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/api/v1/zz_generated.conversion.go @@ -233,9 +233,11 @@ func Convert_api_AuthProviderConfig_To_v1_AuthProviderConfig(in *api.AuthProvide func autoConvert_v1_Cluster_To_api_Cluster(in *Cluster, out *api.Cluster, s conversion.Scope) error { out.Server = in.Server + out.TLSServerName = in.TLSServerName out.InsecureSkipTLSVerify = in.InsecureSkipTLSVerify out.CertificateAuthority = in.CertificateAuthority out.CertificateAuthorityData = *(*[]byte)(unsafe.Pointer(&in.CertificateAuthorityData)) + out.ProxyURL = in.ProxyURL if err := Convert_Slice_v1_NamedExtension_To_Map_string_To_runtime_Object(&in.Extensions, &out.Extensions, s); err != nil { return err } @@ -250,9 +252,11 @@ func Convert_v1_Cluster_To_api_Cluster(in *Cluster, out *api.Cluster, s conversi func autoConvert_api_Cluster_To_v1_Cluster(in *api.Cluster, out *Cluster, s conversion.Scope) error { // INFO: in.LocationOfOrigin opted out of conversion generation out.Server = in.Server + out.TLSServerName = in.TLSServerName out.InsecureSkipTLSVerify = in.InsecureSkipTLSVerify out.CertificateAuthority = in.CertificateAuthority out.CertificateAuthorityData = *(*[]byte)(unsafe.Pointer(&in.CertificateAuthorityData)) + out.ProxyURL = in.ProxyURL if err := Convert_Map_string_To_runtime_Object_To_Slice_v1_NamedExtension(&in.Extensions, &out.Extensions, s); err != nil { return err } @@ -354,6 +358,7 @@ func autoConvert_v1_ExecConfig_To_api_ExecConfig(in *ExecConfig, out *api.ExecCo out.Args = *(*[]string)(unsafe.Pointer(&in.Args)) out.Env = *(*[]api.ExecEnvVar)(unsafe.Pointer(&in.Env)) out.APIVersion = in.APIVersion + out.InstallHint = in.InstallHint return nil } @@ -367,6 +372,7 @@ func autoConvert_api_ExecConfig_To_v1_ExecConfig(in *api.ExecConfig, out *ExecCo out.Args = *(*[]string)(unsafe.Pointer(&in.Args)) out.Env = *(*[]ExecEnvVar)(unsafe.Pointer(&in.Env)) out.APIVersion = in.APIVersion + out.InstallHint = in.InstallHint return nil } diff --git a/vendor/k8s.io/client-go/tools/clientcmd/client_config.go b/vendor/k8s.io/client-go/tools/clientcmd/client_config.go index 44115130..690afce0 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/client_config.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/client_config.go @@ -20,16 +20,18 @@ import ( "fmt" "io" "io/ioutil" + "net/http" "net/url" "os" "strings" - - "github.com/imdario/mergo" - "k8s.io/klog" + "unicode" restclient "k8s.io/client-go/rest" clientauth "k8s.io/client-go/tools/auth" clientcmdapi "k8s.io/client-go/tools/clientcmd/api" + "k8s.io/klog/v2" + + "github.com/imdario/mergo" ) var ( @@ -150,8 +152,15 @@ func (config *DirectClientConfig) ClientConfig() (*restclient.Config, error) { clientConfig := &restclient.Config{} clientConfig.Host = configClusterInfo.Server + if configClusterInfo.ProxyURL != "" { + u, err := parseProxyURL(configClusterInfo.ProxyURL) + if err != nil { + return nil, err + } + clientConfig.Proxy = http.ProxyURL(u) + } - if len(config.overrides.Timeout) > 0 { + if config.overrides != nil && len(config.overrides.Timeout) > 0 { timeout, err := ParseTimeout(config.overrides.Timeout) if err != nil { return nil, err @@ -210,6 +219,7 @@ func getServerIdentificationPartialConfig(configAuthInfo clientcmdapi.AuthInfo, configClientConfig.CAFile = configClusterInfo.CertificateAuthority configClientConfig.CAData = configClusterInfo.CertificateAuthorityData configClientConfig.Insecure = configClusterInfo.InsecureSkipTLSVerify + configClientConfig.ServerName = configClusterInfo.TLSServerName mergo.MergeWithOverwrite(mergedConfig, configClientConfig) return mergedConfig, nil @@ -260,6 +270,7 @@ func (config *DirectClientConfig) getUserIdentificationPartialConfig(configAuthI } if configAuthInfo.Exec != nil { mergedConfig.ExecProvider = configAuthInfo.Exec + mergedConfig.ExecProvider.InstallHint = cleanANSIEscapeCodes(mergedConfig.ExecProvider.InstallHint) } // if there still isn't enough information to authenticate the user, try prompting @@ -305,6 +316,41 @@ func canIdentifyUser(config restclient.Config) bool { config.ExecProvider != nil } +// cleanANSIEscapeCodes takes an arbitrary string and ensures that there are no +// ANSI escape sequences that could put the terminal in a weird state (e.g., +// "\e[1m" bolds text) +func cleanANSIEscapeCodes(s string) string { + // spaceControlCharacters includes tab, new line, vertical tab, new page, and + // carriage return. These are in the unicode.Cc category, but that category also + // contains ESC (U+001B) which we don't want. + spaceControlCharacters := unicode.RangeTable{ + R16: []unicode.Range16{ + {Lo: 0x0009, Hi: 0x000D, Stride: 1}, + }, + } + + // Why not make this deny-only (instead of allow-only)? Because unicode.C + // contains newline and tab characters that we want. + allowedRanges := []*unicode.RangeTable{ + unicode.L, + unicode.M, + unicode.N, + unicode.P, + unicode.S, + unicode.Z, + &spaceControlCharacters, + } + builder := strings.Builder{} + for _, roon := range s { + if unicode.IsOneOf(allowedRanges, roon) { + builder.WriteRune(roon) // returns nil error, per go doc + } else { + fmt.Fprintf(&builder, "%U", roon) + } + } + return builder.String() +} + // Namespace implements ClientConfig func (config *DirectClientConfig) Namespace() (string, bool, error) { if config.overrides != nil && config.overrides.Context.Namespace != "" { @@ -372,7 +418,7 @@ func (config *DirectClientConfig) ConfirmUsable() error { // getContextName returns the default, or user-set context name, and a boolean that indicates // whether the default context name has been overwritten by a user-set flag, or left as its default value func (config *DirectClientConfig) getContextName() (string, bool) { - if len(config.overrides.CurrentContext) != 0 { + if config.overrides != nil && len(config.overrides.CurrentContext) != 0 { return config.overrides.CurrentContext, true } if len(config.contextName) != 0 { @@ -386,7 +432,7 @@ func (config *DirectClientConfig) getContextName() (string, bool) { // and a boolean indicating whether the default authInfo name is overwritten by a user-set flag, or // left as its default value func (config *DirectClientConfig) getAuthInfoName() (string, bool) { - if len(config.overrides.Context.AuthInfo) != 0 { + if config.overrides != nil && len(config.overrides.Context.AuthInfo) != 0 { return config.overrides.Context.AuthInfo, true } context, _ := config.getContext() @@ -397,7 +443,7 @@ func (config *DirectClientConfig) getAuthInfoName() (string, bool) { // indicating whether the default clusterName has been overwritten by a user-set flag, or left as // its default value func (config *DirectClientConfig) getClusterName() (string, bool) { - if len(config.overrides.Context.Cluster) != 0 { + if config.overrides != nil && len(config.overrides.Context.Cluster) != 0 { return config.overrides.Context.Cluster, true } context, _ := config.getContext() @@ -415,7 +461,9 @@ func (config *DirectClientConfig) getContext() (clientcmdapi.Context, error) { } else if required { return clientcmdapi.Context{}, fmt.Errorf("context %q does not exist", contextName) } - mergo.MergeWithOverwrite(mergedContext, config.overrides.Context) + if config.overrides != nil { + mergo.MergeWithOverwrite(mergedContext, config.overrides.Context) + } return *mergedContext, nil } @@ -431,7 +479,9 @@ func (config *DirectClientConfig) getAuthInfo() (clientcmdapi.AuthInfo, error) { } else if required { return clientcmdapi.AuthInfo{}, fmt.Errorf("auth info %q does not exist", authInfoName) } - mergo.MergeWithOverwrite(mergedAuthInfo, config.overrides.AuthInfo) + if config.overrides != nil { + mergo.MergeWithOverwrite(mergedAuthInfo, config.overrides.AuthInfo) + } return *mergedAuthInfo, nil } @@ -442,22 +492,37 @@ func (config *DirectClientConfig) getCluster() (clientcmdapi.Cluster, error) { clusterInfoName, required := config.getClusterName() mergedClusterInfo := clientcmdapi.NewCluster() - mergo.MergeWithOverwrite(mergedClusterInfo, config.overrides.ClusterDefaults) + if config.overrides != nil { + mergo.MergeWithOverwrite(mergedClusterInfo, config.overrides.ClusterDefaults) + } if configClusterInfo, exists := clusterInfos[clusterInfoName]; exists { mergo.MergeWithOverwrite(mergedClusterInfo, configClusterInfo) } else if required { return clientcmdapi.Cluster{}, fmt.Errorf("cluster %q does not exist", clusterInfoName) } - mergo.MergeWithOverwrite(mergedClusterInfo, config.overrides.ClusterInfo) + if config.overrides != nil { + mergo.MergeWithOverwrite(mergedClusterInfo, config.overrides.ClusterInfo) + } + // * An override of --insecure-skip-tls-verify=true and no accompanying CA/CA data should clear already-set CA/CA data // otherwise, a kubeconfig containing a CA reference would return an error that "CA and insecure-skip-tls-verify couldn't both be set". // * An override of --certificate-authority should also override TLS skip settings and CA data, otherwise existing CA data will take precedence. - caLen := len(config.overrides.ClusterInfo.CertificateAuthority) - caDataLen := len(config.overrides.ClusterInfo.CertificateAuthorityData) - if config.overrides.ClusterInfo.InsecureSkipTLSVerify || caLen > 0 || caDataLen > 0 { - mergedClusterInfo.InsecureSkipTLSVerify = config.overrides.ClusterInfo.InsecureSkipTLSVerify - mergedClusterInfo.CertificateAuthority = config.overrides.ClusterInfo.CertificateAuthority - mergedClusterInfo.CertificateAuthorityData = config.overrides.ClusterInfo.CertificateAuthorityData + if config.overrides != nil { + caLen := len(config.overrides.ClusterInfo.CertificateAuthority) + caDataLen := len(config.overrides.ClusterInfo.CertificateAuthorityData) + if config.overrides.ClusterInfo.InsecureSkipTLSVerify || caLen > 0 || caDataLen > 0 { + mergedClusterInfo.InsecureSkipTLSVerify = config.overrides.ClusterInfo.InsecureSkipTLSVerify + mergedClusterInfo.CertificateAuthority = config.overrides.ClusterInfo.CertificateAuthority + mergedClusterInfo.CertificateAuthorityData = config.overrides.ClusterInfo.CertificateAuthorityData + } + + // if the --tls-server-name has been set in overrides, use that value. + // if the --server has been set in overrides, then use the value of --tls-server-name specified on the CLI too. This gives the property + // that setting a --server will effectively clear the KUBECONFIG value of tls-server-name if it is specified on the command line which is + // usually correct. + if config.overrides.ClusterInfo.TLSServerName != "" || config.overrides.ClusterInfo.Server != "" { + mergedClusterInfo.TLSServerName = config.overrides.ClusterInfo.TLSServerName + } } return *mergedClusterInfo, nil diff --git a/vendor/k8s.io/client-go/tools/clientcmd/config.go b/vendor/k8s.io/client-go/tools/clientcmd/config.go index b8cc3968..5f1660bf 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/config.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/config.go @@ -24,7 +24,7 @@ import ( "reflect" "sort" - "k8s.io/klog" + "k8s.io/klog/v2" restclient "k8s.io/client-go/rest" clientcmdapi "k8s.io/client-go/tools/clientcmd/api" @@ -58,6 +58,15 @@ type PathOptions struct { LoadingRules *ClientConfigLoadingRules } +var ( + // UseModifyConfigLock ensures that access to kubeconfig file using ModifyConfig method + // is being guarded by a lock file. + // This variable is intentionaly made public so other consumers of this library + // can modify its default behavior, but be caution when disabling it since + // this will make your code not threadsafe. + UseModifyConfigLock = true +) + func (o *PathOptions) GetEnvVarFiles() []string { if len(o.EnvVar) == 0 { return []string{} @@ -156,15 +165,17 @@ func NewDefaultPathOptions() *PathOptions { // that means that this code will only write into a single file. If you want to relativizePaths, you must provide a fully qualified path in any // modified element. func ModifyConfig(configAccess ConfigAccess, newConfig clientcmdapi.Config, relativizePaths bool) error { - possibleSources := configAccess.GetLoadingPrecedence() - // sort the possible kubeconfig files so we always "lock" in the same order - // to avoid deadlock (note: this can fail w/ symlinks, but... come on). - sort.Strings(possibleSources) - for _, filename := range possibleSources { - if err := lockFile(filename); err != nil { - return err + if UseModifyConfigLock { + possibleSources := configAccess.GetLoadingPrecedence() + // sort the possible kubeconfig files so we always "lock" in the same order + // to avoid deadlock (note: this can fail w/ symlinks, but... come on). + sort.Strings(possibleSources) + for _, filename := range possibleSources { + if err := lockFile(filename); err != nil { + return err + } + defer unlockFile(filename) } - defer unlockFile(filename) } startingConfig, err := configAccess.GetStartingConfig() diff --git a/vendor/k8s.io/client-go/tools/clientcmd/helpers.go b/vendor/k8s.io/client-go/tools/clientcmd/helpers.go index b609d1a7..d7572232 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/helpers.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/helpers.go @@ -18,6 +18,7 @@ package clientcmd import ( "fmt" + "net/url" "strconv" "time" ) @@ -33,3 +34,17 @@ func ParseTimeout(duration string) (time.Duration, error) { } return 0, fmt.Errorf("Invalid timeout value. Timeout must be a single integer in seconds, or an integer followed by a corresponding time unit (e.g. 1s | 2m | 3h)") } + +func parseProxyURL(proxyURL string) (*url.URL, error) { + u, err := url.Parse(proxyURL) + if err != nil { + return nil, fmt.Errorf("could not parse: %v", proxyURL) + } + + switch u.Scheme { + case "http", "https", "socks5": + default: + return nil, fmt.Errorf("unsupported scheme %q, must be http, https, or socks5", u.Scheme) + } + return u, nil +} diff --git a/vendor/k8s.io/client-go/tools/clientcmd/loader.go b/vendor/k8s.io/client-go/tools/clientcmd/loader.go index 4e37e792..b0672291 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/loader.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/loader.go @@ -28,7 +28,7 @@ import ( "strings" "github.com/imdario/mergo" - "k8s.io/klog" + "k8s.io/klog/v2" "k8s.io/apimachinery/pkg/runtime" "k8s.io/apimachinery/pkg/runtime/schema" diff --git a/vendor/k8s.io/client-go/tools/clientcmd/merged_client_builder.go b/vendor/k8s.io/client-go/tools/clientcmd/merged_client_builder.go index 9cc112a2..10744156 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/merged_client_builder.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/merged_client_builder.go @@ -20,7 +20,7 @@ import ( "io" "sync" - "k8s.io/klog" + "k8s.io/klog/v2" restclient "k8s.io/client-go/rest" clientcmdapi "k8s.io/client-go/tools/clientcmd/api" @@ -60,27 +60,26 @@ func NewInteractiveDeferredLoadingClientConfig(loader ClientConfigLoader, overri } func (config *DeferredLoadingClientConfig) createClientConfig() (ClientConfig, error) { - if config.clientConfig == nil { - config.loadingLock.Lock() - defer config.loadingLock.Unlock() - - if config.clientConfig == nil { - mergedConfig, err := config.loader.Load() - if err != nil { - return nil, err - } - - var mergedClientConfig ClientConfig - if config.fallbackReader != nil { - mergedClientConfig = NewInteractiveClientConfig(*mergedConfig, config.overrides.CurrentContext, config.overrides, config.fallbackReader, config.loader) - } else { - mergedClientConfig = NewNonInteractiveClientConfig(*mergedConfig, config.overrides.CurrentContext, config.overrides, config.loader) - } + config.loadingLock.Lock() + defer config.loadingLock.Unlock() - config.clientConfig = mergedClientConfig - } + if config.clientConfig != nil { + return config.clientConfig, nil + } + mergedConfig, err := config.loader.Load() + if err != nil { + return nil, err } + var currentContext string + if config.overrides != nil { + currentContext = config.overrides.CurrentContext + } + if config.fallbackReader != nil { + config.clientConfig = NewInteractiveClientConfig(*mergedConfig, currentContext, config.overrides, config.fallbackReader, config.loader) + } else { + config.clientConfig = NewNonInteractiveClientConfig(*mergedConfig, currentContext, config.overrides, config.loader) + } return config.clientConfig, nil } diff --git a/vendor/k8s.io/client-go/tools/clientcmd/overrides.go b/vendor/k8s.io/client-go/tools/clientcmd/overrides.go index bfca0328..95cba0fa 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/overrides.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/overrides.go @@ -71,6 +71,7 @@ type ClusterOverrideFlags struct { APIVersion FlagInfo CertificateAuthority FlagInfo InsecureSkipTLSVerify FlagInfo + TLSServerName FlagInfo } // FlagInfo contains information about how to register a flag. This struct is useful if you want to provide a way for an extender to @@ -145,6 +146,7 @@ const ( FlagContext = "context" FlagNamespace = "namespace" FlagAPIServer = "server" + FlagTLSServerName = "tls-server-name" FlagInsecure = "insecure-skip-tls-verify" FlagCertFile = "client-certificate" FlagKeyFile = "client-key" @@ -189,6 +191,7 @@ func RecommendedClusterOverrideFlags(prefix string) ClusterOverrideFlags { APIServer: FlagInfo{prefix + FlagAPIServer, "", "", "The address and port of the Kubernetes API server"}, CertificateAuthority: FlagInfo{prefix + FlagCAFile, "", "", "Path to a cert file for the certificate authority"}, InsecureSkipTLSVerify: FlagInfo{prefix + FlagInsecure, "", "false", "If true, the server's certificate will not be checked for validity. This will make your HTTPS connections insecure"}, + TLSServerName: FlagInfo{prefix + FlagTLSServerName, "", "", "If provided, this name will be used to validate server certificate. If this is not provided, hostname used to contact the server is used."}, } } @@ -226,6 +229,7 @@ func BindClusterFlags(clusterInfo *clientcmdapi.Cluster, flags *pflag.FlagSet, f flagNames.APIServer.BindStringFlag(flags, &clusterInfo.Server) flagNames.CertificateAuthority.BindStringFlag(flags, &clusterInfo.CertificateAuthority) flagNames.InsecureSkipTLSVerify.BindBoolFlag(flags, &clusterInfo.InsecureSkipTLSVerify) + flagNames.TLSServerName.BindStringFlag(flags, &clusterInfo.TLSServerName) } // BindFlags is a convenience method to bind the specified flags to their associated variables diff --git a/vendor/k8s.io/client-go/tools/clientcmd/validation.go b/vendor/k8s.io/client-go/tools/clientcmd/validation.go index 2f927072..f77ef04f 100644 --- a/vendor/k8s.io/client-go/tools/clientcmd/validation.go +++ b/vendor/k8s.io/client-go/tools/clientcmd/validation.go @@ -30,11 +30,24 @@ import ( var ( ErrNoContext = errors.New("no context chosen") - ErrEmptyConfig = errors.New("no configuration has been provided") + ErrEmptyConfig = NewEmptyConfigError("no configuration has been provided, try setting KUBERNETES_MASTER environment variable") // message is for consistency with old behavior ErrEmptyCluster = errors.New("cluster has no server defined") ) +// NewEmptyConfigError returns an error wrapping the given message which IsEmptyConfig() will recognize as an empty config error +func NewEmptyConfigError(message string) error { + return &errEmptyConfig{message} +} + +type errEmptyConfig struct { + message string +} + +func (e *errEmptyConfig) Error() string { + return e.message +} + type errContextNotFound struct { ContextName string } @@ -60,9 +73,14 @@ func IsContextNotFound(err error) bool { func IsEmptyConfig(err error) bool { switch t := err.(type) { case errConfigurationInvalid: - return len(t) == 1 && t[0] == ErrEmptyConfig + if len(t) != 1 { + return false + } + _, ok := t[0].(*errEmptyConfig) + return ok } - return err == ErrEmptyConfig + _, ok := err.(*errEmptyConfig) + return ok } // errConfigurationInvalid is a set of errors indicating the configuration is invalid. @@ -86,11 +104,41 @@ func (e errConfigurationInvalid) Error() string { return fmt.Sprintf("invalid configuration: %v", utilerrors.NewAggregate(e).Error()) } -// Errors implements the AggregateError interface +// Errors implements the utilerrors.Aggregate interface func (e errConfigurationInvalid) Errors() []error { return e } +// Is implements the utilerrors.Aggregate interface +func (e errConfigurationInvalid) Is(target error) bool { + return e.visit(func(err error) bool { + return errors.Is(err, target) + }) +} + +func (e errConfigurationInvalid) visit(f func(err error) bool) bool { + for _, err := range e { + switch err := err.(type) { + case errConfigurationInvalid: + if match := err.visit(f); match { + return match + } + case utilerrors.Aggregate: + for _, nestedErr := range err.Errors() { + if match := f(nestedErr); match { + return match + } + } + default: + if match := f(err); match { + return match + } + } + } + + return false +} + // IsConfigurationInvalid returns true if the provided error indicates the configuration is invalid. func IsConfigurationInvalid(err error) bool { switch err.(type) { @@ -179,6 +227,11 @@ func validateClusterInfo(clusterName string, clusterInfo clientcmdapi.Cluster) [ validationErrors = append(validationErrors, fmt.Errorf("no server found for cluster %q", clusterName)) } } + if proxyURL := clusterInfo.ProxyURL; proxyURL != "" { + if _, err := parseProxyURL(proxyURL); err != nil { + validationErrors = append(validationErrors, fmt.Errorf("invalid 'proxy-url' %q for cluster %q: %v", proxyURL, clusterName, err)) + } + } // Make sure CA data and CA file aren't both specified if len(clusterInfo.CertificateAuthority) != 0 && len(clusterInfo.CertificateAuthorityData) != 0 { validationErrors = append(validationErrors, fmt.Errorf("certificate-authority-data and certificate-authority are both specified for %v. certificate-authority-data will override.", clusterName)) diff --git a/vendor/k8s.io/client-go/tools/metrics/OWNERS b/vendor/k8s.io/client-go/tools/metrics/OWNERS index ad52d0c5..77bcb509 100644 --- a/vendor/k8s.io/client-go/tools/metrics/OWNERS +++ b/vendor/k8s.io/client-go/tools/metrics/OWNERS @@ -2,6 +2,5 @@ reviewers: - wojtek-t -- eparis - krousey - jayunit100 diff --git a/vendor/k8s.io/client-go/tools/metrics/metrics.go b/vendor/k8s.io/client-go/tools/metrics/metrics.go index a01306c6..5194026b 100644 --- a/vendor/k8s.io/client-go/tools/metrics/metrics.go +++ b/vendor/k8s.io/client-go/tools/metrics/metrics.go @@ -26,6 +26,16 @@ import ( var registerMetrics sync.Once +// DurationMetric is a measurement of some amount of time. +type DurationMetric interface { + Observe(duration time.Duration) +} + +// ExpiryMetric sets some time of expiry. If nil, assume not relevant. +type ExpiryMetric interface { + Set(expiry *time.Time) +} + // LatencyMetric observes client latency partitioned by verb and url. type LatencyMetric interface { Observe(verb string, u url.URL, latency time.Duration) @@ -37,21 +47,57 @@ type ResultMetric interface { } var ( + // ClientCertExpiry is the expiry time of a client certificate + ClientCertExpiry ExpiryMetric = noopExpiry{} + // ClientCertRotationAge is the age of a certificate that has just been rotated. + ClientCertRotationAge DurationMetric = noopDuration{} // RequestLatency is the latency metric that rest clients will update. RequestLatency LatencyMetric = noopLatency{} + // RateLimiterLatency is the client side rate limiter latency metric. + RateLimiterLatency LatencyMetric = noopLatency{} // RequestResult is the result metric that rest clients will update. RequestResult ResultMetric = noopResult{} ) +// RegisterOpts contains all the metrics to register. Metrics may be nil. +type RegisterOpts struct { + ClientCertExpiry ExpiryMetric + ClientCertRotationAge DurationMetric + RequestLatency LatencyMetric + RateLimiterLatency LatencyMetric + RequestResult ResultMetric +} + // Register registers metrics for the rest client to use. This can // only be called once. -func Register(lm LatencyMetric, rm ResultMetric) { +func Register(opts RegisterOpts) { registerMetrics.Do(func() { - RequestLatency = lm - RequestResult = rm + if opts.ClientCertExpiry != nil { + ClientCertExpiry = opts.ClientCertExpiry + } + if opts.ClientCertRotationAge != nil { + ClientCertRotationAge = opts.ClientCertRotationAge + } + if opts.RequestLatency != nil { + RequestLatency = opts.RequestLatency + } + if opts.RateLimiterLatency != nil { + RateLimiterLatency = opts.RateLimiterLatency + } + if opts.RequestResult != nil { + RequestResult = opts.RequestResult + } }) } +type noopDuration struct{} + +func (noopDuration) Observe(time.Duration) {} + +type noopExpiry struct{} + +func (noopExpiry) Set(*time.Time) {} + type noopLatency struct{} func (noopLatency) Observe(string, url.URL, time.Duration) {} diff --git a/vendor/k8s.io/client-go/transport/cache.go b/vendor/k8s.io/client-go/transport/cache.go index 980d36ae..3ec4e193 100644 --- a/vendor/k8s.io/client-go/transport/cache.go +++ b/vendor/k8s.io/client-go/transport/cache.go @@ -25,6 +25,7 @@ import ( "time" utilnet "k8s.io/apimachinery/pkg/util/net" + "k8s.io/apimachinery/pkg/util/wait" ) // TlsTransportCache caches TLS http.RoundTrippers different configurations. The @@ -44,11 +45,14 @@ type tlsCacheKey struct { caData string certData string keyData string + certFile string + keyFile string getCert string serverName string nextProtos string dial string disableCompression bool + proxy string } func (t tlsCacheKey) String() string { @@ -56,7 +60,7 @@ func (t tlsCacheKey) String() string { if len(t.keyData) > 0 { keyText = "" } - return fmt.Sprintf("insecure:%v, caData:%#v, certData:%#v, keyData:%s, getCert: %s, serverName:%s, dial:%s disableCompression:%t", t.insecure, t.caData, t.certData, keyText, t.getCert, t.serverName, t.dial, t.disableCompression) + return fmt.Sprintf("insecure:%v, caData:%#v, certData:%#v, keyData:%s, getCert: %s, serverName:%s, dial:%s disableCompression:%t, proxy: %s", t.insecure, t.caData, t.certData, keyText, t.getCert, t.serverName, t.dial, t.disableCompression, t.proxy) } func (c *tlsTransportCache) get(config *Config) (http.RoundTripper, error) { @@ -80,7 +84,7 @@ func (c *tlsTransportCache) get(config *Config) (http.RoundTripper, error) { return nil, err } // The options didn't require a custom TLS config - if tlsConfig == nil && config.Dial == nil { + if tlsConfig == nil && config.Dial == nil && config.Proxy == nil { return http.DefaultTransport, nil } @@ -91,9 +95,24 @@ func (c *tlsTransportCache) get(config *Config) (http.RoundTripper, error) { KeepAlive: 30 * time.Second, }).DialContext } + + // If we use are reloading files, we need to handle certificate rotation properly + // TODO(jackkleeman): We can also add rotation here when config.HasCertCallback() is true + if config.TLS.ReloadTLSFiles { + dynamicCertDialer := certRotatingDialer(tlsConfig.GetClientCertificate, dial) + tlsConfig.GetClientCertificate = dynamicCertDialer.GetClientCertificate + dial = dynamicCertDialer.connDialer.DialContext + go dynamicCertDialer.Run(wait.NeverStop) + } + + proxy := http.ProxyFromEnvironment + if config.Proxy != nil { + proxy = config.Proxy + } + // Cache a single transport for these options c.transports[key] = utilnet.SetTransportDefaults(&http.Transport{ - Proxy: http.ProxyFromEnvironment, + Proxy: proxy, TLSHandshakeTimeout: 10 * time.Second, TLSClientConfig: tlsConfig, MaxIdleConnsPerHost: idleConnsPerHost, @@ -109,15 +128,24 @@ func tlsConfigKey(c *Config) (tlsCacheKey, error) { if err := loadTLSFiles(c); err != nil { return tlsCacheKey{}, err } - return tlsCacheKey{ + k := tlsCacheKey{ insecure: c.TLS.Insecure, caData: string(c.TLS.CAData), - certData: string(c.TLS.CertData), - keyData: string(c.TLS.KeyData), getCert: fmt.Sprintf("%p", c.TLS.GetCert), serverName: c.TLS.ServerName, nextProtos: strings.Join(c.TLS.NextProtos, ","), dial: fmt.Sprintf("%p", c.Dial), disableCompression: c.DisableCompression, - }, nil + proxy: fmt.Sprintf("%p", c.Proxy), + } + + if c.TLS.ReloadTLSFiles { + k.certFile = c.TLS.CertFile + k.keyFile = c.TLS.KeyFile + } else { + k.certData = string(c.TLS.CertData) + k.keyData = string(c.TLS.KeyData) + } + + return k, nil } diff --git a/vendor/k8s.io/client-go/transport/cert_rotation.go b/vendor/k8s.io/client-go/transport/cert_rotation.go new file mode 100644 index 00000000..dc22b6ec --- /dev/null +++ b/vendor/k8s.io/client-go/transport/cert_rotation.go @@ -0,0 +1,176 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package transport + +import ( + "bytes" + "crypto/tls" + "fmt" + "reflect" + "sync" + "time" + + utilnet "k8s.io/apimachinery/pkg/util/net" + utilruntime "k8s.io/apimachinery/pkg/util/runtime" + "k8s.io/apimachinery/pkg/util/wait" + "k8s.io/client-go/util/connrotation" + "k8s.io/client-go/util/workqueue" + "k8s.io/klog/v2" +) + +const workItemKey = "key" + +// CertCallbackRefreshDuration is exposed so that integration tests can crank up the reload speed. +var CertCallbackRefreshDuration = 5 * time.Minute + +type reloadFunc func(*tls.CertificateRequestInfo) (*tls.Certificate, error) + +type dynamicClientCert struct { + clientCert *tls.Certificate + certMtx sync.RWMutex + + reload reloadFunc + connDialer *connrotation.Dialer + + // queue only ever has one item, but it has nice error handling backoff/retry semantics + queue workqueue.RateLimitingInterface +} + +func certRotatingDialer(reload reloadFunc, dial utilnet.DialFunc) *dynamicClientCert { + d := &dynamicClientCert{ + reload: reload, + connDialer: connrotation.NewDialer(connrotation.DialFunc(dial)), + queue: workqueue.NewNamedRateLimitingQueue(workqueue.DefaultControllerRateLimiter(), "DynamicClientCertificate"), + } + + return d +} + +// loadClientCert calls the callback and rotates connections if needed +func (c *dynamicClientCert) loadClientCert() (*tls.Certificate, error) { + cert, err := c.reload(nil) + if err != nil { + return nil, err + } + + // check to see if we have a change. If the values are the same, do nothing. + c.certMtx.RLock() + haveCert := c.clientCert != nil + if certsEqual(c.clientCert, cert) { + c.certMtx.RUnlock() + return c.clientCert, nil + } + c.certMtx.RUnlock() + + c.certMtx.Lock() + c.clientCert = cert + c.certMtx.Unlock() + + // The first certificate requested is not a rotation that is worth closing connections for + if !haveCert { + return cert, nil + } + + klog.V(1).Infof("certificate rotation detected, shutting down client connections to start using new credentials") + c.connDialer.CloseAll() + + return cert, nil +} + +// certsEqual compares tls Certificates, ignoring the Leaf which may get filled in dynamically +func certsEqual(left, right *tls.Certificate) bool { + if left == nil || right == nil { + return left == right + } + + if !byteMatrixEqual(left.Certificate, right.Certificate) { + return false + } + + if !reflect.DeepEqual(left.PrivateKey, right.PrivateKey) { + return false + } + + if !byteMatrixEqual(left.SignedCertificateTimestamps, right.SignedCertificateTimestamps) { + return false + } + + if !bytes.Equal(left.OCSPStaple, right.OCSPStaple) { + return false + } + + return true +} + +func byteMatrixEqual(left, right [][]byte) bool { + if len(left) != len(right) { + return false + } + + for i := range left { + if !bytes.Equal(left[i], right[i]) { + return false + } + } + return true +} + +// run starts the controller and blocks until stopCh is closed. +func (c *dynamicClientCert) Run(stopCh <-chan struct{}) { + defer utilruntime.HandleCrash() + defer c.queue.ShutDown() + + klog.V(3).Infof("Starting client certificate rotation controller") + defer klog.V(3).Infof("Shutting down client certificate rotation controller") + + go wait.Until(c.runWorker, time.Second, stopCh) + + go wait.PollImmediateUntil(CertCallbackRefreshDuration, func() (bool, error) { + c.queue.Add(workItemKey) + return false, nil + }, stopCh) + + <-stopCh +} + +func (c *dynamicClientCert) runWorker() { + for c.processNextWorkItem() { + } +} + +func (c *dynamicClientCert) processNextWorkItem() bool { + dsKey, quit := c.queue.Get() + if quit { + return false + } + defer c.queue.Done(dsKey) + + _, err := c.loadClientCert() + if err == nil { + c.queue.Forget(dsKey) + return true + } + + utilruntime.HandleError(fmt.Errorf("%v failed with : %v", dsKey, err)) + c.queue.AddRateLimited(dsKey) + + return true +} + +func (c *dynamicClientCert) GetClientCertificate(*tls.CertificateRequestInfo) (*tls.Certificate, error) { + return c.loadClientCert() +} diff --git a/vendor/k8s.io/client-go/transport/config.go b/vendor/k8s.io/client-go/transport/config.go index 9e18d11d..45db2486 100644 --- a/vendor/k8s.io/client-go/transport/config.go +++ b/vendor/k8s.io/client-go/transport/config.go @@ -21,6 +21,7 @@ import ( "crypto/tls" "net" "net/http" + "net/url" ) // Config holds various options for establishing a transport. @@ -68,6 +69,13 @@ type Config struct { // Dial specifies the dial function for creating unencrypted TCP connections. Dial func(ctx context.Context, network, address string) (net.Conn, error) + + // Proxy is the the proxy func to be used for all requests made by this + // transport. If Proxy is nil, http.ProxyFromEnvironment is used. If Proxy + // returns a nil *URL, no proxy is used. + // + // socks5 proxying does not currently support spdy streaming endpoints. + Proxy func(*http.Request) (*url.URL, error) } // ImpersonationConfig has all the available impersonation options @@ -115,9 +123,10 @@ func (c *Config) Wrap(fn WrapperFunc) { // TLSConfig holds the information needed to set up a TLS transport. type TLSConfig struct { - CAFile string // Path of the PEM-encoded server trusted root certificates. - CertFile string // Path of the PEM-encoded client certificate. - KeyFile string // Path of the PEM-encoded client key. + CAFile string // Path of the PEM-encoded server trusted root certificates. + CertFile string // Path of the PEM-encoded client certificate. + KeyFile string // Path of the PEM-encoded client key. + ReloadTLSFiles bool // Set to indicate that the original config provided files, and that they should be reloaded Insecure bool // Server should be accessed without verifying the certificate. For testing only. ServerName string // Override for the server name passed to the server for SNI and used to verify certificates. diff --git a/vendor/k8s.io/client-go/transport/round_trippers.go b/vendor/k8s.io/client-go/transport/round_trippers.go index a272753a..a05208d9 100644 --- a/vendor/k8s.io/client-go/transport/round_trippers.go +++ b/vendor/k8s.io/client-go/transport/round_trippers.go @@ -23,7 +23,7 @@ import ( "time" "golang.org/x/oauth2" - "k8s.io/klog" + "k8s.io/klog/v2" utilnet "k8s.io/apimachinery/pkg/util/net" ) @@ -67,13 +67,13 @@ func HTTPWrappersForConfig(config *Config, rt http.RoundTripper) (http.RoundTrip // DebugWrappers wraps a round tripper and logs based on the current log level. func DebugWrappers(rt http.RoundTripper) http.RoundTripper { switch { - case bool(klog.V(9)): + case bool(klog.V(9).Enabled()): rt = newDebuggingRoundTripper(rt, debugCurlCommand, debugURLTiming, debugResponseHeaders) - case bool(klog.V(8)): + case bool(klog.V(8).Enabled()): rt = newDebuggingRoundTripper(rt, debugJustURL, debugRequestHeaders, debugResponseStatus, debugResponseHeaders) - case bool(klog.V(7)): + case bool(klog.V(7).Enabled()): rt = newDebuggingRoundTripper(rt, debugJustURL, debugRequestHeaders, debugResponseStatus) - case bool(klog.V(6)): + case bool(klog.V(6).Enabled()): rt = newDebuggingRoundTripper(rt, debugURLTiming) } diff --git a/vendor/k8s.io/client-go/transport/token_source.go b/vendor/k8s.io/client-go/transport/token_source.go index bb32c3b4..f730c397 100644 --- a/vendor/k8s.io/client-go/transport/token_source.go +++ b/vendor/k8s.io/client-go/transport/token_source.go @@ -26,7 +26,7 @@ import ( "golang.org/x/oauth2" - "k8s.io/klog" + "k8s.io/klog/v2" ) // TokenSourceWrapTransport returns a WrapTransport that injects bearer tokens diff --git a/vendor/k8s.io/client-go/transport/transport.go b/vendor/k8s.io/client-go/transport/transport.go index cd8de982..88d89494 100644 --- a/vendor/k8s.io/client-go/transport/transport.go +++ b/vendor/k8s.io/client-go/transport/transport.go @@ -23,9 +23,11 @@ import ( "fmt" "io/ioutil" "net/http" + "sync" + "time" utilnet "k8s.io/apimachinery/pkg/util/net" - "k8s.io/klog" + "k8s.io/klog/v2" ) // New returns an http.RoundTripper that will provide the authentication @@ -81,7 +83,8 @@ func TLSConfigFor(c *Config) (*tls.Config, error) { } var staticCert *tls.Certificate - if c.HasCertAuth() { + // Treat cert as static if either key or cert was data, not a file + if c.HasCertAuth() && !c.TLS.ReloadTLSFiles { // If key/cert were provided, verify them before setting up // tlsConfig.GetClientCertificate. cert, err := tls.X509KeyPair(c.TLS.CertData, c.TLS.KeyData) @@ -91,6 +94,11 @@ func TLSConfigFor(c *Config) (*tls.Config, error) { staticCert = &cert } + var dynamicCertLoader func() (*tls.Certificate, error) + if c.TLS.ReloadTLSFiles { + dynamicCertLoader = cachingCertificateLoader(c.TLS.CertFile, c.TLS.KeyFile) + } + if c.HasCertAuth() || c.HasCertCallback() { tlsConfig.GetClientCertificate = func(*tls.CertificateRequestInfo) (*tls.Certificate, error) { // Note: static key/cert data always take precedence over cert @@ -98,6 +106,10 @@ func TLSConfigFor(c *Config) (*tls.Config, error) { if staticCert != nil { return staticCert, nil } + // key/cert files lead to ReloadTLSFiles being set - takes precedence over cert callback + if dynamicCertLoader != nil { + return dynamicCertLoader() + } if c.HasCertCallback() { cert, err := c.TLS.GetCert() if err != nil { @@ -129,6 +141,11 @@ func loadTLSFiles(c *Config) error { return err } + // Check that we are purely loading from files + if len(c.TLS.CertFile) > 0 && len(c.TLS.CertData) == 0 && len(c.TLS.KeyFile) > 0 && len(c.TLS.KeyData) == 0 { + c.TLS.ReloadTLSFiles = true + } + c.TLS.CertData, err = dataFromSliceOrFile(c.TLS.CertData, c.TLS.CertFile) if err != nil { return err @@ -243,3 +260,44 @@ func tryCancelRequest(rt http.RoundTripper, req *http.Request) { klog.Warningf("Unable to cancel request for %T", rt) } } + +type certificateCacheEntry struct { + cert *tls.Certificate + err error + birth time.Time +} + +// isStale returns true when this cache entry is too old to be usable +func (c *certificateCacheEntry) isStale() bool { + return time.Now().Sub(c.birth) > time.Second +} + +func newCertificateCacheEntry(certFile, keyFile string) certificateCacheEntry { + cert, err := tls.LoadX509KeyPair(certFile, keyFile) + return certificateCacheEntry{cert: &cert, err: err, birth: time.Now()} +} + +// cachingCertificateLoader ensures that we don't hammer the filesystem when opening many connections +// the underlying cert files are read at most once every second +func cachingCertificateLoader(certFile, keyFile string) func() (*tls.Certificate, error) { + current := newCertificateCacheEntry(certFile, keyFile) + var currentMtx sync.RWMutex + + return func() (*tls.Certificate, error) { + currentMtx.RLock() + if current.isStale() { + currentMtx.RUnlock() + + currentMtx.Lock() + defer currentMtx.Unlock() + + if current.isStale() { + current = newCertificateCacheEntry(certFile, keyFile) + } + } else { + defer currentMtx.RUnlock() + } + + return current.cert, current.err + } +} diff --git a/vendor/k8s.io/client-go/util/cert/cert.go b/vendor/k8s.io/client-go/util/cert/cert.go index 9fd097af..3da14416 100644 --- a/vendor/k8s.io/client-go/util/cert/cert.go +++ b/vendor/k8s.io/client-go/util/cert/cert.go @@ -28,7 +28,7 @@ import ( "io/ioutil" "math/big" "net" - "path" + "path/filepath" "strings" "time" @@ -96,8 +96,8 @@ func GenerateSelfSignedCertKeyWithFixtures(host string, alternateIPs []net.IP, a maxAge := time.Hour * 24 * 365 // one year self-signed certs baseName := fmt.Sprintf("%s_%s_%s", host, strings.Join(ipsToStrings(alternateIPs), "-"), strings.Join(alternateDNS, "-")) - certFixturePath := path.Join(fixtureDirectory, baseName+".crt") - keyFixturePath := path.Join(fixtureDirectory, baseName+".key") + certFixturePath := filepath.Join(fixtureDirectory, baseName+".crt") + keyFixturePath := filepath.Join(fixtureDirectory, baseName+".key") if len(fixtureDirectory) > 0 { cert, err := ioutil.ReadFile(certFixturePath) if err == nil { diff --git a/vendor/k8s.io/client-go/util/jsonpath/jsonpath.go b/vendor/k8s.io/client-go/util/jsonpath/jsonpath.go index 78b6b678..9740fe69 100644 --- a/vendor/k8s.io/client-go/util/jsonpath/jsonpath.go +++ b/vendor/k8s.io/client-go/util/jsonpath/jsonpath.go @@ -18,6 +18,7 @@ package jsonpath import ( "bytes" + "encoding/json" "fmt" "io" "reflect" @@ -29,13 +30,14 @@ import ( type JSONPath struct { name string parser *Parser - stack [][]reflect.Value // push and pop values in different scopes - cur []reflect.Value // current scope values beginRange int inRange int endRange int + lastEndNode *Node + allowMissingKeys bool + outputJSON bool } // New creates a new JSONPath with the given name. @@ -81,12 +83,12 @@ func (j *JSONPath) FindResults(data interface{}) ([][]reflect.Value, error) { return nil, fmt.Errorf("%s is an incomplete jsonpath template", j.name) } - j.cur = []reflect.Value{reflect.ValueOf(data)} + cur := []reflect.Value{reflect.ValueOf(data)} nodes := j.parser.Root.Nodes fullResult := [][]reflect.Value{} for i := 0; i < len(nodes); i++ { node := nodes[i] - results, err := j.walk(j.cur, node) + results, err := j.walk(cur, node) if err != nil { return nil, err } @@ -94,34 +96,80 @@ func (j *JSONPath) FindResults(data interface{}) ([][]reflect.Value, error) { // encounter an end node, break the current block if j.endRange > 0 && j.endRange <= j.inRange { j.endRange-- + j.lastEndNode = &nodes[i] break } // encounter a range node, start a range loop if j.beginRange > 0 { j.beginRange-- j.inRange++ - for k, value := range results { + for _, value := range results { j.parser.Root.Nodes = nodes[i+1:] - if k == len(results)-1 { - j.inRange-- - } nextResults, err := j.FindResults(value.Interface()) if err != nil { return nil, err } fullResult = append(fullResult, nextResults...) } - break + j.inRange-- + + // Fast forward to resume processing after the most recent end node that was encountered + for k := i + 1; k < len(nodes); k++ { + if &nodes[k] == j.lastEndNode { + i = k + break + } + } + continue } fullResult = append(fullResult, results) } return fullResult, nil } +// EnableJSONOutput changes the PrintResults behavior to return a JSON array of results +func (j *JSONPath) EnableJSONOutput(v bool) { + j.outputJSON = v +} + // PrintResults writes the results into writer func (j *JSONPath) PrintResults(wr io.Writer, results []reflect.Value) error { + if j.outputJSON { + // convert the []reflect.Value to something that json + // will be able to marshal + r := make([]interface{}, 0, len(results)) + for i := range results { + r = append(r, results[i].Interface()) + } + results = []reflect.Value{reflect.ValueOf(r)} + } for i, r := range results { - text, err := j.evalToText(r) + var text []byte + var err error + outputJSON := true + kind := r.Kind() + if kind == reflect.Interface { + kind = r.Elem().Kind() + } + switch kind { + case reflect.Map: + case reflect.Array: + case reflect.Slice: + case reflect.Struct: + default: + outputJSON = false + } + switch { + case outputJSON || j.outputJSON: + if j.outputJSON { + text, err = json.MarshalIndent(r.Interface(), "", " ") + text = append(text, '\n') + } else { + text, err = json.Marshal(r.Interface()) + } + default: + text, err = j.evalToText(r) + } if err != nil { return err } @@ -132,7 +180,9 @@ func (j *JSONPath) PrintResults(wr io.Writer, results []reflect.Value) error { return err } } + return nil + } // walk visits tree rooted at the given node in DFS order @@ -212,17 +262,11 @@ func (j *JSONPath) evalIdentifier(input []reflect.Value, node *IdentifierNode) ( results := []reflect.Value{} switch node.Name { case "range": - j.stack = append(j.stack, j.cur) j.beginRange++ results = input case "end": - if j.endRange < j.inRange { // inside a loop, break the current block + if j.inRange > 0 { j.endRange++ - break - } - // the loop is about to end, pop value and continue the following execution - if len(j.stack) > 0 { - j.cur, j.stack = j.stack[len(j.stack)-1], j.stack[:len(j.stack)-1] } else { return results, fmt.Errorf("not in range, nothing to end") } diff --git a/vendor/k8s.io/client-go/util/jsonpath/parser.go b/vendor/k8s.io/client-go/util/jsonpath/parser.go index e1aab680..b84016a9 100644 --- a/vendor/k8s.io/client-go/util/jsonpath/parser.go +++ b/vendor/k8s.io/client-go/util/jsonpath/parser.go @@ -214,8 +214,11 @@ func (p *Parser) parseIdentifier(cur *ListNode) error { return p.parseInsideAction(cur) } -// parseRecursive scans the recursive desent operator .. +// parseRecursive scans the recursive descent operator .. func (p *Parser) parseRecursive(cur *ListNode) error { + if lastIndex := len(cur.Nodes) - 1; lastIndex >= 0 && cur.Nodes[lastIndex].Type() == NodeRecursive { + return fmt.Errorf("invalid multiple recursive descent") + } p.pos += len("..") p.consumeText() cur.append(newRecursive()) diff --git a/vendor/k8s.io/client-go/util/workqueue/default_rate_limiters.go b/vendor/k8s.io/client-go/util/workqueue/default_rate_limiters.go new file mode 100644 index 00000000..71bb6322 --- /dev/null +++ b/vendor/k8s.io/client-go/util/workqueue/default_rate_limiters.go @@ -0,0 +1,211 @@ +/* +Copyright 2016 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package workqueue + +import ( + "math" + "sync" + "time" + + "golang.org/x/time/rate" +) + +type RateLimiter interface { + // When gets an item and gets to decide how long that item should wait + When(item interface{}) time.Duration + // Forget indicates that an item is finished being retried. Doesn't matter whether its for perm failing + // or for success, we'll stop tracking it + Forget(item interface{}) + // NumRequeues returns back how many failures the item has had + NumRequeues(item interface{}) int +} + +// DefaultControllerRateLimiter is a no-arg constructor for a default rate limiter for a workqueue. It has +// both overall and per-item rate limiting. The overall is a token bucket and the per-item is exponential +func DefaultControllerRateLimiter() RateLimiter { + return NewMaxOfRateLimiter( + NewItemExponentialFailureRateLimiter(5*time.Millisecond, 1000*time.Second), + // 10 qps, 100 bucket size. This is only for retry speed and its only the overall factor (not per item) + &BucketRateLimiter{Limiter: rate.NewLimiter(rate.Limit(10), 100)}, + ) +} + +// BucketRateLimiter adapts a standard bucket to the workqueue ratelimiter API +type BucketRateLimiter struct { + *rate.Limiter +} + +var _ RateLimiter = &BucketRateLimiter{} + +func (r *BucketRateLimiter) When(item interface{}) time.Duration { + return r.Limiter.Reserve().Delay() +} + +func (r *BucketRateLimiter) NumRequeues(item interface{}) int { + return 0 +} + +func (r *BucketRateLimiter) Forget(item interface{}) { +} + +// ItemExponentialFailureRateLimiter does a simple baseDelay*2^ limit +// dealing with max failures and expiration are up to the caller +type ItemExponentialFailureRateLimiter struct { + failuresLock sync.Mutex + failures map[interface{}]int + + baseDelay time.Duration + maxDelay time.Duration +} + +var _ RateLimiter = &ItemExponentialFailureRateLimiter{} + +func NewItemExponentialFailureRateLimiter(baseDelay time.Duration, maxDelay time.Duration) RateLimiter { + return &ItemExponentialFailureRateLimiter{ + failures: map[interface{}]int{}, + baseDelay: baseDelay, + maxDelay: maxDelay, + } +} + +func DefaultItemBasedRateLimiter() RateLimiter { + return NewItemExponentialFailureRateLimiter(time.Millisecond, 1000*time.Second) +} + +func (r *ItemExponentialFailureRateLimiter) When(item interface{}) time.Duration { + r.failuresLock.Lock() + defer r.failuresLock.Unlock() + + exp := r.failures[item] + r.failures[item] = r.failures[item] + 1 + + // The backoff is capped such that 'calculated' value never overflows. + backoff := float64(r.baseDelay.Nanoseconds()) * math.Pow(2, float64(exp)) + if backoff > math.MaxInt64 { + return r.maxDelay + } + + calculated := time.Duration(backoff) + if calculated > r.maxDelay { + return r.maxDelay + } + + return calculated +} + +func (r *ItemExponentialFailureRateLimiter) NumRequeues(item interface{}) int { + r.failuresLock.Lock() + defer r.failuresLock.Unlock() + + return r.failures[item] +} + +func (r *ItemExponentialFailureRateLimiter) Forget(item interface{}) { + r.failuresLock.Lock() + defer r.failuresLock.Unlock() + + delete(r.failures, item) +} + +// ItemFastSlowRateLimiter does a quick retry for a certain number of attempts, then a slow retry after that +type ItemFastSlowRateLimiter struct { + failuresLock sync.Mutex + failures map[interface{}]int + + maxFastAttempts int + fastDelay time.Duration + slowDelay time.Duration +} + +var _ RateLimiter = &ItemFastSlowRateLimiter{} + +func NewItemFastSlowRateLimiter(fastDelay, slowDelay time.Duration, maxFastAttempts int) RateLimiter { + return &ItemFastSlowRateLimiter{ + failures: map[interface{}]int{}, + fastDelay: fastDelay, + slowDelay: slowDelay, + maxFastAttempts: maxFastAttempts, + } +} + +func (r *ItemFastSlowRateLimiter) When(item interface{}) time.Duration { + r.failuresLock.Lock() + defer r.failuresLock.Unlock() + + r.failures[item] = r.failures[item] + 1 + + if r.failures[item] <= r.maxFastAttempts { + return r.fastDelay + } + + return r.slowDelay +} + +func (r *ItemFastSlowRateLimiter) NumRequeues(item interface{}) int { + r.failuresLock.Lock() + defer r.failuresLock.Unlock() + + return r.failures[item] +} + +func (r *ItemFastSlowRateLimiter) Forget(item interface{}) { + r.failuresLock.Lock() + defer r.failuresLock.Unlock() + + delete(r.failures, item) +} + +// MaxOfRateLimiter calls every RateLimiter and returns the worst case response +// When used with a token bucket limiter, the burst could be apparently exceeded in cases where particular items +// were separately delayed a longer time. +type MaxOfRateLimiter struct { + limiters []RateLimiter +} + +func (r *MaxOfRateLimiter) When(item interface{}) time.Duration { + ret := time.Duration(0) + for _, limiter := range r.limiters { + curr := limiter.When(item) + if curr > ret { + ret = curr + } + } + + return ret +} + +func NewMaxOfRateLimiter(limiters ...RateLimiter) RateLimiter { + return &MaxOfRateLimiter{limiters: limiters} +} + +func (r *MaxOfRateLimiter) NumRequeues(item interface{}) int { + ret := 0 + for _, limiter := range r.limiters { + curr := limiter.NumRequeues(item) + if curr > ret { + ret = curr + } + } + + return ret +} + +func (r *MaxOfRateLimiter) Forget(item interface{}) { + for _, limiter := range r.limiters { + limiter.Forget(item) + } +} diff --git a/vendor/k8s.io/client-go/util/workqueue/delaying_queue.go b/vendor/k8s.io/client-go/util/workqueue/delaying_queue.go new file mode 100644 index 00000000..31d9182d --- /dev/null +++ b/vendor/k8s.io/client-go/util/workqueue/delaying_queue.go @@ -0,0 +1,280 @@ +/* +Copyright 2016 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package workqueue + +import ( + "container/heap" + "sync" + "time" + + "k8s.io/apimachinery/pkg/util/clock" + utilruntime "k8s.io/apimachinery/pkg/util/runtime" +) + +// DelayingInterface is an Interface that can Add an item at a later time. This makes it easier to +// requeue items after failures without ending up in a hot-loop. +type DelayingInterface interface { + Interface + // AddAfter adds an item to the workqueue after the indicated duration has passed + AddAfter(item interface{}, duration time.Duration) +} + +// NewDelayingQueue constructs a new workqueue with delayed queuing ability +func NewDelayingQueue() DelayingInterface { + return NewDelayingQueueWithCustomClock(clock.RealClock{}, "") +} + +// NewDelayingQueueWithCustomQueue constructs a new workqueue with ability to +// inject custom queue Interface instead of the default one +func NewDelayingQueueWithCustomQueue(q Interface, name string) DelayingInterface { + return newDelayingQueue(clock.RealClock{}, q, name) +} + +// NewNamedDelayingQueue constructs a new named workqueue with delayed queuing ability +func NewNamedDelayingQueue(name string) DelayingInterface { + return NewDelayingQueueWithCustomClock(clock.RealClock{}, name) +} + +// NewDelayingQueueWithCustomClock constructs a new named workqueue +// with ability to inject real or fake clock for testing purposes +func NewDelayingQueueWithCustomClock(clock clock.Clock, name string) DelayingInterface { + return newDelayingQueue(clock, NewNamed(name), name) +} + +func newDelayingQueue(clock clock.Clock, q Interface, name string) *delayingType { + ret := &delayingType{ + Interface: q, + clock: clock, + heartbeat: clock.NewTicker(maxWait), + stopCh: make(chan struct{}), + waitingForAddCh: make(chan *waitFor, 1000), + metrics: newRetryMetrics(name), + } + + go ret.waitingLoop() + return ret +} + +// delayingType wraps an Interface and provides delayed re-enquing +type delayingType struct { + Interface + + // clock tracks time for delayed firing + clock clock.Clock + + // stopCh lets us signal a shutdown to the waiting loop + stopCh chan struct{} + // stopOnce guarantees we only signal shutdown a single time + stopOnce sync.Once + + // heartbeat ensures we wait no more than maxWait before firing + heartbeat clock.Ticker + + // waitingForAddCh is a buffered channel that feeds waitingForAdd + waitingForAddCh chan *waitFor + + // metrics counts the number of retries + metrics retryMetrics +} + +// waitFor holds the data to add and the time it should be added +type waitFor struct { + data t + readyAt time.Time + // index in the priority queue (heap) + index int +} + +// waitForPriorityQueue implements a priority queue for waitFor items. +// +// waitForPriorityQueue implements heap.Interface. The item occurring next in +// time (i.e., the item with the smallest readyAt) is at the root (index 0). +// Peek returns this minimum item at index 0. Pop returns the minimum item after +// it has been removed from the queue and placed at index Len()-1 by +// container/heap. Push adds an item at index Len(), and container/heap +// percolates it into the correct location. +type waitForPriorityQueue []*waitFor + +func (pq waitForPriorityQueue) Len() int { + return len(pq) +} +func (pq waitForPriorityQueue) Less(i, j int) bool { + return pq[i].readyAt.Before(pq[j].readyAt) +} +func (pq waitForPriorityQueue) Swap(i, j int) { + pq[i], pq[j] = pq[j], pq[i] + pq[i].index = i + pq[j].index = j +} + +// Push adds an item to the queue. Push should not be called directly; instead, +// use `heap.Push`. +func (pq *waitForPriorityQueue) Push(x interface{}) { + n := len(*pq) + item := x.(*waitFor) + item.index = n + *pq = append(*pq, item) +} + +// Pop removes an item from the queue. Pop should not be called directly; +// instead, use `heap.Pop`. +func (pq *waitForPriorityQueue) Pop() interface{} { + n := len(*pq) + item := (*pq)[n-1] + item.index = -1 + *pq = (*pq)[0:(n - 1)] + return item +} + +// Peek returns the item at the beginning of the queue, without removing the +// item or otherwise mutating the queue. It is safe to call directly. +func (pq waitForPriorityQueue) Peek() interface{} { + return pq[0] +} + +// ShutDown stops the queue. After the queue drains, the returned shutdown bool +// on Get() will be true. This method may be invoked more than once. +func (q *delayingType) ShutDown() { + q.stopOnce.Do(func() { + q.Interface.ShutDown() + close(q.stopCh) + q.heartbeat.Stop() + }) +} + +// AddAfter adds the given item to the work queue after the given delay +func (q *delayingType) AddAfter(item interface{}, duration time.Duration) { + // don't add if we're already shutting down + if q.ShuttingDown() { + return + } + + q.metrics.retry() + + // immediately add things with no delay + if duration <= 0 { + q.Add(item) + return + } + + select { + case <-q.stopCh: + // unblock if ShutDown() is called + case q.waitingForAddCh <- &waitFor{data: item, readyAt: q.clock.Now().Add(duration)}: + } +} + +// maxWait keeps a max bound on the wait time. It's just insurance against weird things happening. +// Checking the queue every 10 seconds isn't expensive and we know that we'll never end up with an +// expired item sitting for more than 10 seconds. +const maxWait = 10 * time.Second + +// waitingLoop runs until the workqueue is shutdown and keeps a check on the list of items to be added. +func (q *delayingType) waitingLoop() { + defer utilruntime.HandleCrash() + + // Make a placeholder channel to use when there are no items in our list + never := make(<-chan time.Time) + + // Make a timer that expires when the item at the head of the waiting queue is ready + var nextReadyAtTimer clock.Timer + + waitingForQueue := &waitForPriorityQueue{} + heap.Init(waitingForQueue) + + waitingEntryByData := map[t]*waitFor{} + + for { + if q.Interface.ShuttingDown() { + return + } + + now := q.clock.Now() + + // Add ready entries + for waitingForQueue.Len() > 0 { + entry := waitingForQueue.Peek().(*waitFor) + if entry.readyAt.After(now) { + break + } + + entry = heap.Pop(waitingForQueue).(*waitFor) + q.Add(entry.data) + delete(waitingEntryByData, entry.data) + } + + // Set up a wait for the first item's readyAt (if one exists) + nextReadyAt := never + if waitingForQueue.Len() > 0 { + if nextReadyAtTimer != nil { + nextReadyAtTimer.Stop() + } + entry := waitingForQueue.Peek().(*waitFor) + nextReadyAtTimer = q.clock.NewTimer(entry.readyAt.Sub(now)) + nextReadyAt = nextReadyAtTimer.C() + } + + select { + case <-q.stopCh: + return + + case <-q.heartbeat.C(): + // continue the loop, which will add ready items + + case <-nextReadyAt: + // continue the loop, which will add ready items + + case waitEntry := <-q.waitingForAddCh: + if waitEntry.readyAt.After(q.clock.Now()) { + insert(waitingForQueue, waitingEntryByData, waitEntry) + } else { + q.Add(waitEntry.data) + } + + drained := false + for !drained { + select { + case waitEntry := <-q.waitingForAddCh: + if waitEntry.readyAt.After(q.clock.Now()) { + insert(waitingForQueue, waitingEntryByData, waitEntry) + } else { + q.Add(waitEntry.data) + } + default: + drained = true + } + } + } + } +} + +// insert adds the entry to the priority queue, or updates the readyAt if it already exists in the queue +func insert(q *waitForPriorityQueue, knownEntries map[t]*waitFor, entry *waitFor) { + // if the entry already exists, update the time only if it would cause the item to be queued sooner + existing, exists := knownEntries[entry.data] + if exists { + if existing.readyAt.After(entry.readyAt) { + existing.readyAt = entry.readyAt + heap.Fix(q, existing.index) + } + + return + } + + heap.Push(q, entry) + knownEntries[entry.data] = entry +} diff --git a/vendor/k8s.io/client-go/util/workqueue/doc.go b/vendor/k8s.io/client-go/util/workqueue/doc.go new file mode 100644 index 00000000..a5c976e0 --- /dev/null +++ b/vendor/k8s.io/client-go/util/workqueue/doc.go @@ -0,0 +1,26 @@ +/* +Copyright 2014 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Package workqueue provides a simple queue that supports the following +// features: +// * Fair: items processed in the order in which they are added. +// * Stingy: a single item will not be processed multiple times concurrently, +// and if an item is added multiple times before it can be processed, it +// will only be processed once. +// * Multiple consumers and producers. In particular, it is allowed for an +// item to be reenqueued while it is being processed. +// * Shutdown notifications. +package workqueue // import "k8s.io/client-go/util/workqueue" diff --git a/vendor/k8s.io/client-go/util/workqueue/metrics.go b/vendor/k8s.io/client-go/util/workqueue/metrics.go new file mode 100644 index 00000000..556e6432 --- /dev/null +++ b/vendor/k8s.io/client-go/util/workqueue/metrics.go @@ -0,0 +1,261 @@ +/* +Copyright 2016 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package workqueue + +import ( + "sync" + "time" + + "k8s.io/apimachinery/pkg/util/clock" +) + +// This file provides abstractions for setting the provider (e.g., prometheus) +// of metrics. + +type queueMetrics interface { + add(item t) + get(item t) + done(item t) + updateUnfinishedWork() +} + +// GaugeMetric represents a single numerical value that can arbitrarily go up +// and down. +type GaugeMetric interface { + Inc() + Dec() +} + +// SettableGaugeMetric represents a single numerical value that can arbitrarily go up +// and down. (Separate from GaugeMetric to preserve backwards compatibility.) +type SettableGaugeMetric interface { + Set(float64) +} + +// CounterMetric represents a single numerical value that only ever +// goes up. +type CounterMetric interface { + Inc() +} + +// SummaryMetric captures individual observations. +type SummaryMetric interface { + Observe(float64) +} + +// HistogramMetric counts individual observations. +type HistogramMetric interface { + Observe(float64) +} + +type noopMetric struct{} + +func (noopMetric) Inc() {} +func (noopMetric) Dec() {} +func (noopMetric) Set(float64) {} +func (noopMetric) Observe(float64) {} + +// defaultQueueMetrics expects the caller to lock before setting any metrics. +type defaultQueueMetrics struct { + clock clock.Clock + + // current depth of a workqueue + depth GaugeMetric + // total number of adds handled by a workqueue + adds CounterMetric + // how long an item stays in a workqueue + latency HistogramMetric + // how long processing an item from a workqueue takes + workDuration HistogramMetric + addTimes map[t]time.Time + processingStartTimes map[t]time.Time + + // how long have current threads been working? + unfinishedWorkSeconds SettableGaugeMetric + longestRunningProcessor SettableGaugeMetric +} + +func (m *defaultQueueMetrics) add(item t) { + if m == nil { + return + } + + m.adds.Inc() + m.depth.Inc() + if _, exists := m.addTimes[item]; !exists { + m.addTimes[item] = m.clock.Now() + } +} + +func (m *defaultQueueMetrics) get(item t) { + if m == nil { + return + } + + m.depth.Dec() + m.processingStartTimes[item] = m.clock.Now() + if startTime, exists := m.addTimes[item]; exists { + m.latency.Observe(m.sinceInSeconds(startTime)) + delete(m.addTimes, item) + } +} + +func (m *defaultQueueMetrics) done(item t) { + if m == nil { + return + } + + if startTime, exists := m.processingStartTimes[item]; exists { + m.workDuration.Observe(m.sinceInSeconds(startTime)) + delete(m.processingStartTimes, item) + } +} + +func (m *defaultQueueMetrics) updateUnfinishedWork() { + // Note that a summary metric would be better for this, but prometheus + // doesn't seem to have non-hacky ways to reset the summary metrics. + var total float64 + var oldest float64 + for _, t := range m.processingStartTimes { + age := m.sinceInSeconds(t) + total += age + if age > oldest { + oldest = age + } + } + m.unfinishedWorkSeconds.Set(total) + m.longestRunningProcessor.Set(oldest) +} + +type noMetrics struct{} + +func (noMetrics) add(item t) {} +func (noMetrics) get(item t) {} +func (noMetrics) done(item t) {} +func (noMetrics) updateUnfinishedWork() {} + +// Gets the time since the specified start in seconds. +func (m *defaultQueueMetrics) sinceInSeconds(start time.Time) float64 { + return m.clock.Since(start).Seconds() +} + +type retryMetrics interface { + retry() +} + +type defaultRetryMetrics struct { + retries CounterMetric +} + +func (m *defaultRetryMetrics) retry() { + if m == nil { + return + } + + m.retries.Inc() +} + +// MetricsProvider generates various metrics used by the queue. +type MetricsProvider interface { + NewDepthMetric(name string) GaugeMetric + NewAddsMetric(name string) CounterMetric + NewLatencyMetric(name string) HistogramMetric + NewWorkDurationMetric(name string) HistogramMetric + NewUnfinishedWorkSecondsMetric(name string) SettableGaugeMetric + NewLongestRunningProcessorSecondsMetric(name string) SettableGaugeMetric + NewRetriesMetric(name string) CounterMetric +} + +type noopMetricsProvider struct{} + +func (_ noopMetricsProvider) NewDepthMetric(name string) GaugeMetric { + return noopMetric{} +} + +func (_ noopMetricsProvider) NewAddsMetric(name string) CounterMetric { + return noopMetric{} +} + +func (_ noopMetricsProvider) NewLatencyMetric(name string) HistogramMetric { + return noopMetric{} +} + +func (_ noopMetricsProvider) NewWorkDurationMetric(name string) HistogramMetric { + return noopMetric{} +} + +func (_ noopMetricsProvider) NewUnfinishedWorkSecondsMetric(name string) SettableGaugeMetric { + return noopMetric{} +} + +func (_ noopMetricsProvider) NewLongestRunningProcessorSecondsMetric(name string) SettableGaugeMetric { + return noopMetric{} +} + +func (_ noopMetricsProvider) NewRetriesMetric(name string) CounterMetric { + return noopMetric{} +} + +var globalMetricsFactory = queueMetricsFactory{ + metricsProvider: noopMetricsProvider{}, +} + +type queueMetricsFactory struct { + metricsProvider MetricsProvider + + onlyOnce sync.Once +} + +func (f *queueMetricsFactory) setProvider(mp MetricsProvider) { + f.onlyOnce.Do(func() { + f.metricsProvider = mp + }) +} + +func (f *queueMetricsFactory) newQueueMetrics(name string, clock clock.Clock) queueMetrics { + mp := f.metricsProvider + if len(name) == 0 || mp == (noopMetricsProvider{}) { + return noMetrics{} + } + return &defaultQueueMetrics{ + clock: clock, + depth: mp.NewDepthMetric(name), + adds: mp.NewAddsMetric(name), + latency: mp.NewLatencyMetric(name), + workDuration: mp.NewWorkDurationMetric(name), + unfinishedWorkSeconds: mp.NewUnfinishedWorkSecondsMetric(name), + longestRunningProcessor: mp.NewLongestRunningProcessorSecondsMetric(name), + addTimes: map[t]time.Time{}, + processingStartTimes: map[t]time.Time{}, + } +} + +func newRetryMetrics(name string) retryMetrics { + var ret *defaultRetryMetrics + if len(name) == 0 { + return ret + } + return &defaultRetryMetrics{ + retries: globalMetricsFactory.metricsProvider.NewRetriesMetric(name), + } +} + +// SetProvider sets the metrics provider for all subsequently created work +// queues. Only the first call has an effect. +func SetProvider(metricsProvider MetricsProvider) { + globalMetricsFactory.setProvider(metricsProvider) +} diff --git a/vendor/k8s.io/client-go/util/workqueue/parallelizer.go b/vendor/k8s.io/client-go/util/workqueue/parallelizer.go new file mode 100644 index 00000000..366bf20a --- /dev/null +++ b/vendor/k8s.io/client-go/util/workqueue/parallelizer.go @@ -0,0 +1,101 @@ +/* +Copyright 2016 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package workqueue + +import ( + "context" + "sync" + + utilruntime "k8s.io/apimachinery/pkg/util/runtime" +) + +type DoWorkPieceFunc func(piece int) + +type options struct { + chunkSize int +} + +type Options func(*options) + +// WithChunkSize allows to set chunks of work items to the workers, rather than +// processing one by one. +// It is recommended to use this option if the number of pieces significantly +// higher than the number of workers and the work done for each item is small. +func WithChunkSize(c int) func(*options) { + return func(o *options) { + o.chunkSize = c + } +} + +// ParallelizeUntil is a framework that allows for parallelizing N +// independent pieces of work until done or the context is canceled. +func ParallelizeUntil(ctx context.Context, workers, pieces int, doWorkPiece DoWorkPieceFunc, opts ...Options) { + if pieces == 0 { + return + } + o := options{} + for _, opt := range opts { + opt(&o) + } + chunkSize := o.chunkSize + if chunkSize < 1 { + chunkSize = 1 + } + + chunks := ceilDiv(pieces, chunkSize) + toProcess := make(chan int, chunks) + for i := 0; i < chunks; i++ { + toProcess <- i + } + close(toProcess) + + var stop <-chan struct{} + if ctx != nil { + stop = ctx.Done() + } + if chunks < workers { + workers = chunks + } + wg := sync.WaitGroup{} + wg.Add(workers) + for i := 0; i < workers; i++ { + go func() { + defer utilruntime.HandleCrash() + defer wg.Done() + for chunk := range toProcess { + start := chunk * chunkSize + end := start + chunkSize + if end > pieces { + end = pieces + } + for p := start; p < end; p++ { + select { + case <-stop: + return + default: + doWorkPiece(p) + } + } + } + }() + } + wg.Wait() +} + +func ceilDiv(a, b int) int { + return (a + b - 1) / b +} diff --git a/vendor/k8s.io/client-go/util/workqueue/queue.go b/vendor/k8s.io/client-go/util/workqueue/queue.go new file mode 100644 index 00000000..39009b8e --- /dev/null +++ b/vendor/k8s.io/client-go/util/workqueue/queue.go @@ -0,0 +1,212 @@ +/* +Copyright 2015 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package workqueue + +import ( + "sync" + "time" + + "k8s.io/apimachinery/pkg/util/clock" +) + +type Interface interface { + Add(item interface{}) + Len() int + Get() (item interface{}, shutdown bool) + Done(item interface{}) + ShutDown() + ShuttingDown() bool +} + +// New constructs a new work queue (see the package comment). +func New() *Type { + return NewNamed("") +} + +func NewNamed(name string) *Type { + rc := clock.RealClock{} + return newQueue( + rc, + globalMetricsFactory.newQueueMetrics(name, rc), + defaultUnfinishedWorkUpdatePeriod, + ) +} + +func newQueue(c clock.Clock, metrics queueMetrics, updatePeriod time.Duration) *Type { + t := &Type{ + clock: c, + dirty: set{}, + processing: set{}, + cond: sync.NewCond(&sync.Mutex{}), + metrics: metrics, + unfinishedWorkUpdatePeriod: updatePeriod, + } + go t.updateUnfinishedWorkLoop() + return t +} + +const defaultUnfinishedWorkUpdatePeriod = 500 * time.Millisecond + +// Type is a work queue (see the package comment). +type Type struct { + // queue defines the order in which we will work on items. Every + // element of queue should be in the dirty set and not in the + // processing set. + queue []t + + // dirty defines all of the items that need to be processed. + dirty set + + // Things that are currently being processed are in the processing set. + // These things may be simultaneously in the dirty set. When we finish + // processing something and remove it from this set, we'll check if + // it's in the dirty set, and if so, add it to the queue. + processing set + + cond *sync.Cond + + shuttingDown bool + + metrics queueMetrics + + unfinishedWorkUpdatePeriod time.Duration + clock clock.Clock +} + +type empty struct{} +type t interface{} +type set map[t]empty + +func (s set) has(item t) bool { + _, exists := s[item] + return exists +} + +func (s set) insert(item t) { + s[item] = empty{} +} + +func (s set) delete(item t) { + delete(s, item) +} + +// Add marks item as needing processing. +func (q *Type) Add(item interface{}) { + q.cond.L.Lock() + defer q.cond.L.Unlock() + if q.shuttingDown { + return + } + if q.dirty.has(item) { + return + } + + q.metrics.add(item) + + q.dirty.insert(item) + if q.processing.has(item) { + return + } + + q.queue = append(q.queue, item) + q.cond.Signal() +} + +// Len returns the current queue length, for informational purposes only. You +// shouldn't e.g. gate a call to Add() or Get() on Len() being a particular +// value, that can't be synchronized properly. +func (q *Type) Len() int { + q.cond.L.Lock() + defer q.cond.L.Unlock() + return len(q.queue) +} + +// Get blocks until it can return an item to be processed. If shutdown = true, +// the caller should end their goroutine. You must call Done with item when you +// have finished processing it. +func (q *Type) Get() (item interface{}, shutdown bool) { + q.cond.L.Lock() + defer q.cond.L.Unlock() + for len(q.queue) == 0 && !q.shuttingDown { + q.cond.Wait() + } + if len(q.queue) == 0 { + // We must be shutting down. + return nil, true + } + + item, q.queue = q.queue[0], q.queue[1:] + + q.metrics.get(item) + + q.processing.insert(item) + q.dirty.delete(item) + + return item, false +} + +// Done marks item as done processing, and if it has been marked as dirty again +// while it was being processed, it will be re-added to the queue for +// re-processing. +func (q *Type) Done(item interface{}) { + q.cond.L.Lock() + defer q.cond.L.Unlock() + + q.metrics.done(item) + + q.processing.delete(item) + if q.dirty.has(item) { + q.queue = append(q.queue, item) + q.cond.Signal() + } +} + +// ShutDown will cause q to ignore all new items added to it. As soon as the +// worker goroutines have drained the existing items in the queue, they will be +// instructed to exit. +func (q *Type) ShutDown() { + q.cond.L.Lock() + defer q.cond.L.Unlock() + q.shuttingDown = true + q.cond.Broadcast() +} + +func (q *Type) ShuttingDown() bool { + q.cond.L.Lock() + defer q.cond.L.Unlock() + + return q.shuttingDown +} + +func (q *Type) updateUnfinishedWorkLoop() { + t := q.clock.NewTicker(q.unfinishedWorkUpdatePeriod) + defer t.Stop() + for range t.C() { + if !func() bool { + q.cond.L.Lock() + defer q.cond.L.Unlock() + if !q.shuttingDown { + q.metrics.updateUnfinishedWork() + return true + } + return false + + }() { + return + } + } +} diff --git a/vendor/k8s.io/client-go/util/workqueue/rate_limiting_queue.go b/vendor/k8s.io/client-go/util/workqueue/rate_limiting_queue.go new file mode 100644 index 00000000..8321876a --- /dev/null +++ b/vendor/k8s.io/client-go/util/workqueue/rate_limiting_queue.go @@ -0,0 +1,69 @@ +/* +Copyright 2016 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package workqueue + +// RateLimitingInterface is an interface that rate limits items being added to the queue. +type RateLimitingInterface interface { + DelayingInterface + + // AddRateLimited adds an item to the workqueue after the rate limiter says it's ok + AddRateLimited(item interface{}) + + // Forget indicates that an item is finished being retried. Doesn't matter whether it's for perm failing + // or for success, we'll stop the rate limiter from tracking it. This only clears the `rateLimiter`, you + // still have to call `Done` on the queue. + Forget(item interface{}) + + // NumRequeues returns back how many times the item was requeued + NumRequeues(item interface{}) int +} + +// NewRateLimitingQueue constructs a new workqueue with rateLimited queuing ability +// Remember to call Forget! If you don't, you may end up tracking failures forever. +func NewRateLimitingQueue(rateLimiter RateLimiter) RateLimitingInterface { + return &rateLimitingType{ + DelayingInterface: NewDelayingQueue(), + rateLimiter: rateLimiter, + } +} + +func NewNamedRateLimitingQueue(rateLimiter RateLimiter, name string) RateLimitingInterface { + return &rateLimitingType{ + DelayingInterface: NewNamedDelayingQueue(name), + rateLimiter: rateLimiter, + } +} + +// rateLimitingType wraps an Interface and provides rateLimited re-enquing +type rateLimitingType struct { + DelayingInterface + + rateLimiter RateLimiter +} + +// AddRateLimited AddAfter's the item based on the time when the rate limiter says it's ok +func (q *rateLimitingType) AddRateLimited(item interface{}) { + q.DelayingInterface.AddAfter(item, q.rateLimiter.When(item)) +} + +func (q *rateLimitingType) NumRequeues(item interface{}) int { + return q.rateLimiter.NumRequeues(item) +} + +func (q *rateLimitingType) Forget(item interface{}) { + q.rateLimiter.Forget(item) +} diff --git a/vendor/k8s.io/klog/v2/.gitignore b/vendor/k8s.io/klog/v2/.gitignore new file mode 100644 index 00000000..0aa20023 --- /dev/null +++ b/vendor/k8s.io/klog/v2/.gitignore @@ -0,0 +1,17 @@ +# OSX leaves these everywhere on SMB shares +._* + +# OSX trash +.DS_Store + +# Eclipse files +.classpath +.project +.settings/** + +# Files generated by JetBrains IDEs, e.g. IntelliJ IDEA +.idea/ +*.iml + +# Vscode files +.vscode diff --git a/vendor/k8s.io/klog/v2/CONTRIBUTING.md b/vendor/k8s.io/klog/v2/CONTRIBUTING.md new file mode 100644 index 00000000..2641b1f4 --- /dev/null +++ b/vendor/k8s.io/klog/v2/CONTRIBUTING.md @@ -0,0 +1,22 @@ +# Contributing Guidelines + +Welcome to Kubernetes. We are excited about the prospect of you joining our [community](https://github.com/kubernetes/community)! The Kubernetes community abides by the CNCF [code of conduct](code-of-conduct.md). Here is an excerpt: + +_As contributors and maintainers of this project, and in the interest of fostering an open and welcoming community, we pledge to respect all people who contribute through reporting issues, posting feature requests, updating documentation, submitting pull requests or patches, and other activities._ + +## Getting Started + +We have full documentation on how to get started contributing here: + +- [Contributor License Agreement](https://git.k8s.io/community/CLA.md) Kubernetes projects require that you sign a Contributor License Agreement (CLA) before we can accept your pull requests +- [Kubernetes Contributor Guide](http://git.k8s.io/community/contributors/guide) - Main contributor documentation, or you can just jump directly to the [contributing section](http://git.k8s.io/community/contributors/guide#contributing) +- [Contributor Cheat Sheet](https://git.k8s.io/community/contributors/guide/contributor-cheatsheet) - Common resources for existing developers + +## Mentorship + +- [Mentoring Initiatives](https://git.k8s.io/community/mentoring) - We have a diverse set of mentorship programs available that are always looking for volunteers! + +## Contact Information + +- [Slack](https://kubernetes.slack.com/messages/sig-architecture) +- [Mailing List](https://groups.google.com/forum/#!forum/kubernetes-sig-architecture) diff --git a/vendor/k8s.io/klog/v2/LICENSE b/vendor/k8s.io/klog/v2/LICENSE new file mode 100644 index 00000000..37ec93a1 --- /dev/null +++ b/vendor/k8s.io/klog/v2/LICENSE @@ -0,0 +1,191 @@ +Apache License +Version 2.0, January 2004 +http://www.apache.org/licenses/ + +TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + +1. Definitions. + +"License" shall mean the terms and conditions for use, reproduction, and +distribution as defined by Sections 1 through 9 of this document. + +"Licensor" shall mean the copyright owner or entity authorized by the copyright +owner that is granting the License. + +"Legal Entity" shall mean the union of the acting entity and all other entities +that control, are controlled by, or are under common control with that entity. +For the purposes of this definition, "control" means (i) the power, direct or +indirect, to cause the direction or management of such entity, whether by +contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the +outstanding shares, or (iii) beneficial ownership of such entity. + +"You" (or "Your") shall mean an individual or Legal Entity exercising +permissions granted by this License. + +"Source" form shall mean the preferred form for making modifications, including +but not limited to software source code, documentation source, and configuration +files. + +"Object" form shall mean any form resulting from mechanical transformation or +translation of a Source form, including but not limited to compiled object code, +generated documentation, and conversions to other media types. + +"Work" shall mean the work of authorship, whether in Source or Object form, made +available under the License, as indicated by a copyright notice that is included +in or attached to the work (an example is provided in the Appendix below). + +"Derivative Works" shall mean any work, whether in Source or Object form, that +is based on (or derived from) the Work and for which the editorial revisions, +annotations, elaborations, or other modifications represent, as a whole, an +original work of authorship. For the purposes of this License, Derivative Works +shall not include works that remain separable from, or merely link (or bind by +name) to the interfaces of, the Work and Derivative Works thereof. + +"Contribution" shall mean any work of authorship, including the original version +of the Work and any modifications or additions to that Work or Derivative Works +thereof, that is intentionally submitted to Licensor for inclusion in the Work +by the copyright owner or by an individual or Legal Entity authorized to submit +on behalf of the copyright owner. For the purposes of this definition, +"submitted" means any form of electronic, verbal, or written communication sent +to the Licensor or its representatives, including but not limited to +communication on electronic mailing lists, source code control systems, and +issue tracking systems that are managed by, or on behalf of, the Licensor for +the purpose of discussing and improving the Work, but excluding communication +that is conspicuously marked or otherwise designated in writing by the copyright +owner as "Not a Contribution." + +"Contributor" shall mean Licensor and any individual or Legal Entity on behalf +of whom a Contribution has been received by Licensor and subsequently +incorporated within the Work. + +2. Grant of Copyright License. + +Subject to the terms and conditions of this License, each Contributor hereby +grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, +irrevocable copyright license to reproduce, prepare Derivative Works of, +publicly display, publicly perform, sublicense, and distribute the Work and such +Derivative Works in Source or Object form. + +3. Grant of Patent License. + +Subject to the terms and conditions of this License, each Contributor hereby +grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, +irrevocable (except as stated in this section) patent license to make, have +made, use, offer to sell, sell, import, and otherwise transfer the Work, where +such license applies only to those patent claims licensable by such Contributor +that are necessarily infringed by their Contribution(s) alone or by combination +of their Contribution(s) with the Work to which such Contribution(s) was +submitted. If You institute patent litigation against any entity (including a +cross-claim or counterclaim in a lawsuit) alleging that the Work or a +Contribution incorporated within the Work constitutes direct or contributory +patent infringement, then any patent licenses granted to You under this License +for that Work shall terminate as of the date such litigation is filed. + +4. Redistribution. + +You may reproduce and distribute copies of the Work or Derivative Works thereof +in any medium, with or without modifications, and in Source or Object form, +provided that You meet the following conditions: + +You must give any other recipients of the Work or Derivative Works a copy of +this License; and +You must cause any modified files to carry prominent notices stating that You +changed the files; and +You must retain, in the Source form of any Derivative Works that You distribute, +all copyright, patent, trademark, and attribution notices from the Source form +of the Work, excluding those notices that do not pertain to any part of the +Derivative Works; and +If the Work includes a "NOTICE" text file as part of its distribution, then any +Derivative Works that You distribute must include a readable copy of the +attribution notices contained within such NOTICE file, excluding those notices +that do not pertain to any part of the Derivative Works, in at least one of the +following places: within a NOTICE text file distributed as part of the +Derivative Works; within the Source form or documentation, if provided along +with the Derivative Works; or, within a display generated by the Derivative +Works, if and wherever such third-party notices normally appear. The contents of +the NOTICE file are for informational purposes only and do not modify the +License. You may add Your own attribution notices within Derivative Works that +You distribute, alongside or as an addendum to the NOTICE text from the Work, +provided that such additional attribution notices cannot be construed as +modifying the License. +You may add Your own copyright statement to Your modifications and may provide +additional or different license terms and conditions for use, reproduction, or +distribution of Your modifications, or for any such Derivative Works as a whole, +provided Your use, reproduction, and distribution of the Work otherwise complies +with the conditions stated in this License. + +5. Submission of Contributions. + +Unless You explicitly state otherwise, any Contribution intentionally submitted +for inclusion in the Work by You to the Licensor shall be under the terms and +conditions of this License, without any additional terms or conditions. +Notwithstanding the above, nothing herein shall supersede or modify the terms of +any separate license agreement you may have executed with Licensor regarding +such Contributions. + +6. Trademarks. + +This License does not grant permission to use the trade names, trademarks, +service marks, or product names of the Licensor, except as required for +reasonable and customary use in describing the origin of the Work and +reproducing the content of the NOTICE file. + +7. Disclaimer of Warranty. + +Unless required by applicable law or agreed to in writing, Licensor provides the +Work (and each Contributor provides its Contributions) on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, +including, without limitation, any warranties or conditions of TITLE, +NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are +solely responsible for determining the appropriateness of using or +redistributing the Work and assume any risks associated with Your exercise of +permissions under this License. + +8. Limitation of Liability. + +In no event and under no legal theory, whether in tort (including negligence), +contract, or otherwise, unless required by applicable law (such as deliberate +and grossly negligent acts) or agreed to in writing, shall any Contributor be +liable to You for damages, including any direct, indirect, special, incidental, +or consequential damages of any character arising as a result of this License or +out of the use or inability to use the Work (including but not limited to +damages for loss of goodwill, work stoppage, computer failure or malfunction, or +any and all other commercial damages or losses), even if such Contributor has +been advised of the possibility of such damages. + +9. Accepting Warranty or Additional Liability. + +While redistributing the Work or Derivative Works thereof, You may choose to +offer, and charge a fee for, acceptance of support, warranty, indemnity, or +other liability obligations and/or rights consistent with this License. However, +in accepting such obligations, You may act only on Your own behalf and on Your +sole responsibility, not on behalf of any other Contributor, and only if You +agree to indemnify, defend, and hold each Contributor harmless for any liability +incurred by, or claims asserted against, such Contributor by reason of your +accepting any such warranty or additional liability. + +END OF TERMS AND CONDITIONS + +APPENDIX: How to apply the Apache License to your work + +To apply the Apache License to your work, attach the following boilerplate +notice, with the fields enclosed by brackets "[]" replaced with your own +identifying information. (Don't include the brackets!) The text should be +enclosed in the appropriate comment syntax for the file format. We also +recommend that a file or class name and description of purpose be included on +the same "printed page" as the copyright notice for easier identification within +third-party archives. + + Copyright [yyyy] [name of copyright owner] + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/k8s.io/klog/v2/OWNERS b/vendor/k8s.io/klog/v2/OWNERS new file mode 100644 index 00000000..380e514f --- /dev/null +++ b/vendor/k8s.io/klog/v2/OWNERS @@ -0,0 +1,19 @@ +# See the OWNERS docs at https://go.k8s.io/owners +reviewers: + - jayunit100 + - hoegaarden + - andyxning + - neolit123 + - pohly + - yagonobre + - vincepri + - detiber +approvers: + - dims + - thockin + - justinsb + - tallclair + - piosz + - brancz + - DirectXMan12 + - lavalamp diff --git a/vendor/k8s.io/klog/v2/README.md b/vendor/k8s.io/klog/v2/README.md new file mode 100644 index 00000000..2f9f6f0d --- /dev/null +++ b/vendor/k8s.io/klog/v2/README.md @@ -0,0 +1,99 @@ +klog +==== + +klog is a permanent fork of https://github.com/golang/glog. + +## Why was klog created? + +The decision to create klog was one that wasn't made lightly, but it was necessary due to some +drawbacks that are present in [glog](https://github.com/golang/glog). Ultimately, the fork was created due to glog not being under active development; this can be seen in the glog README: + +> The code in this repo [...] is not itself under development + +This makes us unable to solve many use cases without a fork. The factors that contributed to needing feature development are listed below: + + * `glog` [presents a lot "gotchas"](https://github.com/kubernetes/kubernetes/issues/61006) and introduces challenges in containerized environments, all of which aren't well documented. + * `glog` doesn't provide an easy way to test logs, which detracts from the stability of software using it + * A long term goal is to implement a logging interface that allows us to add context, change output format, etc. + +Historical context is available here: + + * https://github.com/kubernetes/kubernetes/issues/61006 + * https://github.com/kubernetes/kubernetes/issues/70264 + * https://groups.google.com/forum/#!msg/kubernetes-sig-architecture/wCWiWf3Juzs/hXRVBH90CgAJ + * https://groups.google.com/forum/#!msg/kubernetes-dev/7vnijOMhLS0/1oRiNtigBgAJ + +---- + +How to use klog +=============== +- Replace imports for `github.com/golang/glog` with `k8s.io/klog` +- Use `klog.InitFlags(nil)` explicitly for initializing global flags as we no longer use `init()` method to register the flags +- You can now use `log_file` instead of `log_dir` for logging to a single file (See `examples/log_file/usage_log_file.go`) +- If you want to redirect everything logged using klog somewhere else (say syslog!), you can use `klog.SetOutput()` method and supply a `io.Writer`. (See `examples/set_output/usage_set_output.go`) +- For more logging conventions (See [Logging Conventions](https://github.com/kubernetes/community/blob/master/contributors/devel/sig-instrumentation/logging.md)) + +**NOTE**: please use the newer go versions that support semantic import versioning in modules, ideally go 1.11.4 or greater. + +### Coexisting with glog +This package can be used side by side with glog. [This example](examples/coexist_glog/coexist_glog.go) shows how to initialize and synchronize flags from the global `flag.CommandLine` FlagSet. In addition, the example makes use of stderr as combined output by setting `alsologtostderr` (or `logtostderr`) to `true`. + +## Community, discussion, contribution, and support + +Learn how to engage with the Kubernetes community on the [community page](http://kubernetes.io/community/). + +You can reach the maintainers of this project at: + +- [Slack](https://kubernetes.slack.com/messages/klog) +- [Mailing List](https://groups.google.com/forum/#!forum/kubernetes-sig-architecture) + +### Code of conduct + +Participation in the Kubernetes community is governed by the [Kubernetes Code of Conduct](code-of-conduct.md). + +---- + +glog +==== + +Leveled execution logs for Go. + +This is an efficient pure Go implementation of leveled logs in the +manner of the open source C++ package + https://github.com/google/glog + +By binding methods to booleans it is possible to use the log package +without paying the expense of evaluating the arguments to the log. +Through the -vmodule flag, the package also provides fine-grained +control over logging at the file level. + +The comment from glog.go introduces the ideas: + + Package glog implements logging analogous to the Google-internal + C++ INFO/ERROR/V setup. It provides functions Info, Warning, + Error, Fatal, plus formatting variants such as Infof. It + also provides V-style logging controlled by the -v and + -vmodule=file=2 flags. + + Basic examples: + + glog.Info("Prepare to repel boarders") + + glog.Fatalf("Initialization failed: %s", err) + + See the documentation for the V function for an explanation + of these examples: + + if glog.V(2) { + glog.Info("Starting transaction...") + } + + glog.V(2).Infoln("Processed", nItems, "elements") + + +The repository contains an open source version of the log package +used inside Google. The master copy of the source lives inside +Google, not here. The code in this repo is for export only and is not itself +under development. Feature requests will be ignored. + +Send bug reports to golang-nuts@googlegroups.com. diff --git a/vendor/k8s.io/klog/v2/RELEASE.md b/vendor/k8s.io/klog/v2/RELEASE.md new file mode 100644 index 00000000..b53eb960 --- /dev/null +++ b/vendor/k8s.io/klog/v2/RELEASE.md @@ -0,0 +1,9 @@ +# Release Process + +The `klog` is released on an as-needed basis. The process is as follows: + +1. An issue is proposing a new release with a changelog since the last release +1. All [OWNERS](OWNERS) must LGTM this release +1. An OWNER runs `git tag -s $VERSION` and inserts the changelog and pushes the tag with `git push $VERSION` +1. The release issue is closed +1. An announcement email is sent to `kubernetes-dev@googlegroups.com` with the subject `[ANNOUNCE] kubernetes-template-project $VERSION is released` diff --git a/vendor/k8s.io/klog/v2/SECURITY_CONTACTS b/vendor/k8s.io/klog/v2/SECURITY_CONTACTS new file mode 100644 index 00000000..6128a586 --- /dev/null +++ b/vendor/k8s.io/klog/v2/SECURITY_CONTACTS @@ -0,0 +1,20 @@ +# Defined below are the security contacts for this repo. +# +# They are the contact point for the Product Security Committee to reach out +# to for triaging and handling of incoming issues. +# +# The below names agree to abide by the +# [Embargo Policy](https://git.k8s.io/security/private-distributors-list.md#embargo-policy) +# and will be removed and replaced if they violate that agreement. +# +# DO NOT REPORT SECURITY VULNERABILITIES DIRECTLY TO THESE NAMES, FOLLOW THE +# INSTRUCTIONS AT https://kubernetes.io/security/ + +dims +thockin +justinsb +tallclair +piosz +brancz +DirectXMan12 +lavalamp diff --git a/vendor/k8s.io/klog/v2/code-of-conduct.md b/vendor/k8s.io/klog/v2/code-of-conduct.md new file mode 100644 index 00000000..0d15c00c --- /dev/null +++ b/vendor/k8s.io/klog/v2/code-of-conduct.md @@ -0,0 +1,3 @@ +# Kubernetes Community Code of Conduct + +Please refer to our [Kubernetes Community Code of Conduct](https://git.k8s.io/community/code-of-conduct.md) diff --git a/vendor/k8s.io/klog/v2/go.mod b/vendor/k8s.io/klog/v2/go.mod new file mode 100644 index 00000000..e396e31c --- /dev/null +++ b/vendor/k8s.io/klog/v2/go.mod @@ -0,0 +1,5 @@ +module k8s.io/klog/v2 + +go 1.13 + +require github.com/go-logr/logr v0.2.0 diff --git a/vendor/k8s.io/klog/v2/go.sum b/vendor/k8s.io/klog/v2/go.sum new file mode 100644 index 00000000..8dfa7854 --- /dev/null +++ b/vendor/k8s.io/klog/v2/go.sum @@ -0,0 +1,2 @@ +github.com/go-logr/logr v0.2.0 h1:QvGt2nLcHH0WK9orKa+ppBPAxREcH364nPUedEpK0TY= +github.com/go-logr/logr v0.2.0/go.mod h1:z6/tIYblkpsD+a4lm/fGIIU9mZ+XfAiaFtq7xTgseGU= diff --git a/vendor/k8s.io/klog/v2/klog.go b/vendor/k8s.io/klog/v2/klog.go new file mode 100644 index 00000000..ae2b8613 --- /dev/null +++ b/vendor/k8s.io/klog/v2/klog.go @@ -0,0 +1,1525 @@ +// Go support for leveled logs, analogous to https://code.google.com/p/google-glog/ +// +// Copyright 2013 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Package klog implements logging analogous to the Google-internal C++ INFO/ERROR/V setup. +// It provides functions Info, Warning, Error, Fatal, plus formatting variants such as +// Infof. It also provides V-style logging controlled by the -v and -vmodule=file=2 flags. +// +// Basic examples: +// +// klog.Info("Prepare to repel boarders") +// +// klog.Fatalf("Initialization failed: %s", err) +// +// See the documentation for the V function for an explanation of these examples: +// +// if klog.V(2) { +// klog.Info("Starting transaction...") +// } +// +// klog.V(2).Infoln("Processed", nItems, "elements") +// +// Log output is buffered and written periodically using Flush. Programs +// should call Flush before exiting to guarantee all log output is written. +// +// By default, all log statements write to standard error. +// This package provides several flags that modify this behavior. +// As a result, flag.Parse must be called before any logging is done. +// +// -logtostderr=true +// Logs are written to standard error instead of to files. +// -alsologtostderr=false +// Logs are written to standard error as well as to files. +// -stderrthreshold=ERROR +// Log events at or above this severity are logged to standard +// error as well as to files. +// -log_dir="" +// Log files will be written to this directory instead of the +// default temporary directory. +// +// Other flags provide aids to debugging. +// +// -log_backtrace_at="" +// When set to a file and line number holding a logging statement, +// such as +// -log_backtrace_at=gopherflakes.go:234 +// a stack trace will be written to the Info log whenever execution +// hits that statement. (Unlike with -vmodule, the ".go" must be +// present.) +// -v=0 +// Enable V-leveled logging at the specified level. +// -vmodule="" +// The syntax of the argument is a comma-separated list of pattern=N, +// where pattern is a literal file name (minus the ".go" suffix) or +// "glob" pattern and N is a V level. For instance, +// -vmodule=gopher*=3 +// sets the V level to 3 in all Go files whose names begin "gopher". +// +package klog + +import ( + "bufio" + "bytes" + "errors" + "flag" + "fmt" + "io" + stdLog "log" + "math" + "os" + "path/filepath" + "runtime" + "strconv" + "strings" + "sync" + "sync/atomic" + "time" + + "github.com/go-logr/logr" +) + +// severity identifies the sort of log: info, warning etc. It also implements +// the flag.Value interface. The -stderrthreshold flag is of type severity and +// should be modified only through the flag.Value interface. The values match +// the corresponding constants in C++. +type severity int32 // sync/atomic int32 + +// These constants identify the log levels in order of increasing severity. +// A message written to a high-severity log file is also written to each +// lower-severity log file. +const ( + infoLog severity = iota + warningLog + errorLog + fatalLog + numSeverity = 4 +) + +const severityChar = "IWEF" + +var severityName = []string{ + infoLog: "INFO", + warningLog: "WARNING", + errorLog: "ERROR", + fatalLog: "FATAL", +} + +// get returns the value of the severity. +func (s *severity) get() severity { + return severity(atomic.LoadInt32((*int32)(s))) +} + +// set sets the value of the severity. +func (s *severity) set(val severity) { + atomic.StoreInt32((*int32)(s), int32(val)) +} + +// String is part of the flag.Value interface. +func (s *severity) String() string { + return strconv.FormatInt(int64(*s), 10) +} + +// Get is part of the flag.Getter interface. +func (s *severity) Get() interface{} { + return *s +} + +// Set is part of the flag.Value interface. +func (s *severity) Set(value string) error { + var threshold severity + // Is it a known name? + if v, ok := severityByName(value); ok { + threshold = v + } else { + v, err := strconv.ParseInt(value, 10, 32) + if err != nil { + return err + } + threshold = severity(v) + } + logging.stderrThreshold.set(threshold) + return nil +} + +func severityByName(s string) (severity, bool) { + s = strings.ToUpper(s) + for i, name := range severityName { + if name == s { + return severity(i), true + } + } + return 0, false +} + +// OutputStats tracks the number of output lines and bytes written. +type OutputStats struct { + lines int64 + bytes int64 +} + +// Lines returns the number of lines written. +func (s *OutputStats) Lines() int64 { + return atomic.LoadInt64(&s.lines) +} + +// Bytes returns the number of bytes written. +func (s *OutputStats) Bytes() int64 { + return atomic.LoadInt64(&s.bytes) +} + +// Stats tracks the number of lines of output and number of bytes +// per severity level. Values must be read with atomic.LoadInt64. +var Stats struct { + Info, Warning, Error OutputStats +} + +var severityStats = [numSeverity]*OutputStats{ + infoLog: &Stats.Info, + warningLog: &Stats.Warning, + errorLog: &Stats.Error, +} + +// Level is exported because it appears in the arguments to V and is +// the type of the v flag, which can be set programmatically. +// It's a distinct type because we want to discriminate it from logType. +// Variables of type level are only changed under logging.mu. +// The -v flag is read only with atomic ops, so the state of the logging +// module is consistent. + +// Level is treated as a sync/atomic int32. + +// Level specifies a level of verbosity for V logs. *Level implements +// flag.Value; the -v flag is of type Level and should be modified +// only through the flag.Value interface. +type Level int32 + +// get returns the value of the Level. +func (l *Level) get() Level { + return Level(atomic.LoadInt32((*int32)(l))) +} + +// set sets the value of the Level. +func (l *Level) set(val Level) { + atomic.StoreInt32((*int32)(l), int32(val)) +} + +// String is part of the flag.Value interface. +func (l *Level) String() string { + return strconv.FormatInt(int64(*l), 10) +} + +// Get is part of the flag.Getter interface. +func (l *Level) Get() interface{} { + return *l +} + +// Set is part of the flag.Value interface. +func (l *Level) Set(value string) error { + v, err := strconv.ParseInt(value, 10, 32) + if err != nil { + return err + } + logging.mu.Lock() + defer logging.mu.Unlock() + logging.setVState(Level(v), logging.vmodule.filter, false) + return nil +} + +// moduleSpec represents the setting of the -vmodule flag. +type moduleSpec struct { + filter []modulePat +} + +// modulePat contains a filter for the -vmodule flag. +// It holds a verbosity level and a file pattern to match. +type modulePat struct { + pattern string + literal bool // The pattern is a literal string + level Level +} + +// match reports whether the file matches the pattern. It uses a string +// comparison if the pattern contains no metacharacters. +func (m *modulePat) match(file string) bool { + if m.literal { + return file == m.pattern + } + match, _ := filepath.Match(m.pattern, file) + return match +} + +func (m *moduleSpec) String() string { + // Lock because the type is not atomic. TODO: clean this up. + logging.mu.Lock() + defer logging.mu.Unlock() + var b bytes.Buffer + for i, f := range m.filter { + if i > 0 { + b.WriteRune(',') + } + fmt.Fprintf(&b, "%s=%d", f.pattern, f.level) + } + return b.String() +} + +// Get is part of the (Go 1.2) flag.Getter interface. It always returns nil for this flag type since the +// struct is not exported. +func (m *moduleSpec) Get() interface{} { + return nil +} + +var errVmoduleSyntax = errors.New("syntax error: expect comma-separated list of filename=N") + +// Syntax: -vmodule=recordio=2,file=1,gfs*=3 +func (m *moduleSpec) Set(value string) error { + var filter []modulePat + for _, pat := range strings.Split(value, ",") { + if len(pat) == 0 { + // Empty strings such as from a trailing comma can be ignored. + continue + } + patLev := strings.Split(pat, "=") + if len(patLev) != 2 || len(patLev[0]) == 0 || len(patLev[1]) == 0 { + return errVmoduleSyntax + } + pattern := patLev[0] + v, err := strconv.ParseInt(patLev[1], 10, 32) + if err != nil { + return errors.New("syntax error: expect comma-separated list of filename=N") + } + if v < 0 { + return errors.New("negative value for vmodule level") + } + if v == 0 { + continue // Ignore. It's harmless but no point in paying the overhead. + } + // TODO: check syntax of filter? + filter = append(filter, modulePat{pattern, isLiteral(pattern), Level(v)}) + } + logging.mu.Lock() + defer logging.mu.Unlock() + logging.setVState(logging.verbosity, filter, true) + return nil +} + +// isLiteral reports whether the pattern is a literal string, that is, has no metacharacters +// that require filepath.Match to be called to match the pattern. +func isLiteral(pattern string) bool { + return !strings.ContainsAny(pattern, `\*?[]`) +} + +// traceLocation represents the setting of the -log_backtrace_at flag. +type traceLocation struct { + file string + line int +} + +// isSet reports whether the trace location has been specified. +// logging.mu is held. +func (t *traceLocation) isSet() bool { + return t.line > 0 +} + +// match reports whether the specified file and line matches the trace location. +// The argument file name is the full path, not the basename specified in the flag. +// logging.mu is held. +func (t *traceLocation) match(file string, line int) bool { + if t.line != line { + return false + } + if i := strings.LastIndex(file, "/"); i >= 0 { + file = file[i+1:] + } + return t.file == file +} + +func (t *traceLocation) String() string { + // Lock because the type is not atomic. TODO: clean this up. + logging.mu.Lock() + defer logging.mu.Unlock() + return fmt.Sprintf("%s:%d", t.file, t.line) +} + +// Get is part of the (Go 1.2) flag.Getter interface. It always returns nil for this flag type since the +// struct is not exported +func (t *traceLocation) Get() interface{} { + return nil +} + +var errTraceSyntax = errors.New("syntax error: expect file.go:234") + +// Syntax: -log_backtrace_at=gopherflakes.go:234 +// Note that unlike vmodule the file extension is included here. +func (t *traceLocation) Set(value string) error { + if value == "" { + // Unset. + logging.mu.Lock() + defer logging.mu.Unlock() + t.line = 0 + t.file = "" + return nil + } + fields := strings.Split(value, ":") + if len(fields) != 2 { + return errTraceSyntax + } + file, line := fields[0], fields[1] + if !strings.Contains(file, ".") { + return errTraceSyntax + } + v, err := strconv.Atoi(line) + if err != nil { + return errTraceSyntax + } + if v <= 0 { + return errors.New("negative or zero value for level") + } + logging.mu.Lock() + defer logging.mu.Unlock() + t.line = v + t.file = file + return nil +} + +// flushSyncWriter is the interface satisfied by logging destinations. +type flushSyncWriter interface { + Flush() error + Sync() error + io.Writer +} + +// init sets up the defaults and runs flushDaemon. +func init() { + logging.stderrThreshold = errorLog // Default stderrThreshold is ERROR. + logging.setVState(0, nil, false) + logging.logDir = "" + logging.logFile = "" + logging.logFileMaxSizeMB = 1800 + logging.toStderr = true + logging.alsoToStderr = false + logging.skipHeaders = false + logging.addDirHeader = false + logging.skipLogHeaders = false + go logging.flushDaemon() +} + +// InitFlags is for explicitly initializing the flags. +func InitFlags(flagset *flag.FlagSet) { + if flagset == nil { + flagset = flag.CommandLine + } + + flagset.StringVar(&logging.logDir, "log_dir", logging.logDir, "If non-empty, write log files in this directory") + flagset.StringVar(&logging.logFile, "log_file", logging.logFile, "If non-empty, use this log file") + flagset.Uint64Var(&logging.logFileMaxSizeMB, "log_file_max_size", logging.logFileMaxSizeMB, + "Defines the maximum size a log file can grow to. Unit is megabytes. "+ + "If the value is 0, the maximum file size is unlimited.") + flagset.BoolVar(&logging.toStderr, "logtostderr", logging.toStderr, "log to standard error instead of files") + flagset.BoolVar(&logging.alsoToStderr, "alsologtostderr", logging.alsoToStderr, "log to standard error as well as files") + flagset.Var(&logging.verbosity, "v", "number for the log level verbosity") + flagset.BoolVar(&logging.addDirHeader, "add_dir_header", logging.addDirHeader, "If true, adds the file directory to the header of the log messages") + flagset.BoolVar(&logging.skipHeaders, "skip_headers", logging.skipHeaders, "If true, avoid header prefixes in the log messages") + flagset.BoolVar(&logging.skipLogHeaders, "skip_log_headers", logging.skipLogHeaders, "If true, avoid headers when opening log files") + flagset.Var(&logging.stderrThreshold, "stderrthreshold", "logs at or above this threshold go to stderr") + flagset.Var(&logging.vmodule, "vmodule", "comma-separated list of pattern=N settings for file-filtered logging") + flagset.Var(&logging.traceLocation, "log_backtrace_at", "when logging hits line file:N, emit a stack trace") +} + +// Flush flushes all pending log I/O. +func Flush() { + logging.lockAndFlushAll() +} + +// loggingT collects all the global state of the logging setup. +type loggingT struct { + // Boolean flags. Not handled atomically because the flag.Value interface + // does not let us avoid the =true, and that shorthand is necessary for + // compatibility. TODO: does this matter enough to fix? Seems unlikely. + toStderr bool // The -logtostderr flag. + alsoToStderr bool // The -alsologtostderr flag. + + // Level flag. Handled atomically. + stderrThreshold severity // The -stderrthreshold flag. + + // freeList is a list of byte buffers, maintained under freeListMu. + freeList *buffer + // freeListMu maintains the free list. It is separate from the main mutex + // so buffers can be grabbed and printed to without holding the main lock, + // for better parallelization. + freeListMu sync.Mutex + + // mu protects the remaining elements of this structure and is + // used to synchronize logging. + mu sync.Mutex + // file holds writer for each of the log types. + file [numSeverity]flushSyncWriter + // pcs is used in V to avoid an allocation when computing the caller's PC. + pcs [1]uintptr + // vmap is a cache of the V Level for each V() call site, identified by PC. + // It is wiped whenever the vmodule flag changes state. + vmap map[uintptr]Level + // filterLength stores the length of the vmodule filter chain. If greater + // than zero, it means vmodule is enabled. It may be read safely + // using sync.LoadInt32, but is only modified under mu. + filterLength int32 + // traceLocation is the state of the -log_backtrace_at flag. + traceLocation traceLocation + // These flags are modified only under lock, although verbosity may be fetched + // safely using atomic.LoadInt32. + vmodule moduleSpec // The state of the -vmodule flag. + verbosity Level // V logging level, the value of the -v flag/ + + // If non-empty, overrides the choice of directory in which to write logs. + // See createLogDirs for the full list of possible destinations. + logDir string + + // If non-empty, specifies the path of the file to write logs. mutually exclusive + // with the log_dir option. + logFile string + + // When logFile is specified, this limiter makes sure the logFile won't exceeds a certain size. When exceeds, the + // logFile will be cleaned up. If this value is 0, no size limitation will be applied to logFile. + logFileMaxSizeMB uint64 + + // If true, do not add the prefix headers, useful when used with SetOutput + skipHeaders bool + + // If true, do not add the headers to log files + skipLogHeaders bool + + // If true, add the file directory to the header + addDirHeader bool + + // If set, all output will be redirected unconditionally to the provided logr.Logger + logr logr.Logger +} + +// buffer holds a byte Buffer for reuse. The zero value is ready for use. +type buffer struct { + bytes.Buffer + tmp [64]byte // temporary byte array for creating headers. + next *buffer +} + +var logging loggingT + +// setVState sets a consistent state for V logging. +// l.mu is held. +func (l *loggingT) setVState(verbosity Level, filter []modulePat, setFilter bool) { + // Turn verbosity off so V will not fire while we are in transition. + l.verbosity.set(0) + // Ditto for filter length. + atomic.StoreInt32(&l.filterLength, 0) + + // Set the new filters and wipe the pc->Level map if the filter has changed. + if setFilter { + l.vmodule.filter = filter + l.vmap = make(map[uintptr]Level) + } + + // Things are consistent now, so enable filtering and verbosity. + // They are enabled in order opposite to that in V. + atomic.StoreInt32(&l.filterLength, int32(len(filter))) + l.verbosity.set(verbosity) +} + +// getBuffer returns a new, ready-to-use buffer. +func (l *loggingT) getBuffer() *buffer { + l.freeListMu.Lock() + b := l.freeList + if b != nil { + l.freeList = b.next + } + l.freeListMu.Unlock() + if b == nil { + b = new(buffer) + } else { + b.next = nil + b.Reset() + } + return b +} + +// putBuffer returns a buffer to the free list. +func (l *loggingT) putBuffer(b *buffer) { + if b.Len() >= 256 { + // Let big buffers die a natural death. + return + } + l.freeListMu.Lock() + b.next = l.freeList + l.freeList = b + l.freeListMu.Unlock() +} + +var timeNow = time.Now // Stubbed out for testing. + +/* +header formats a log header as defined by the C++ implementation. +It returns a buffer containing the formatted header and the user's file and line number. +The depth specifies how many stack frames above lives the source line to be identified in the log message. + +Log lines have this form: + Lmmdd hh:mm:ss.uuuuuu threadid file:line] msg... +where the fields are defined as follows: + L A single character, representing the log level (eg 'I' for INFO) + mm The month (zero padded; ie May is '05') + dd The day (zero padded) + hh:mm:ss.uuuuuu Time in hours, minutes and fractional seconds + threadid The space-padded thread ID as returned by GetTID() + file The file name + line The line number + msg The user-supplied message +*/ +func (l *loggingT) header(s severity, depth int) (*buffer, string, int) { + _, file, line, ok := runtime.Caller(3 + depth) + if !ok { + file = "???" + line = 1 + } else { + if slash := strings.LastIndex(file, "/"); slash >= 0 { + path := file + file = path[slash+1:] + if l.addDirHeader { + if dirsep := strings.LastIndex(path[:slash], "/"); dirsep >= 0 { + file = path[dirsep+1:] + } + } + } + } + return l.formatHeader(s, file, line), file, line +} + +// formatHeader formats a log header using the provided file name and line number. +func (l *loggingT) formatHeader(s severity, file string, line int) *buffer { + now := timeNow() + if line < 0 { + line = 0 // not a real line number, but acceptable to someDigits + } + if s > fatalLog { + s = infoLog // for safety. + } + buf := l.getBuffer() + if l.skipHeaders { + return buf + } + + // Avoid Fprintf, for speed. The format is so simple that we can do it quickly by hand. + // It's worth about 3X. Fprintf is hard. + _, month, day := now.Date() + hour, minute, second := now.Clock() + // Lmmdd hh:mm:ss.uuuuuu threadid file:line] + buf.tmp[0] = severityChar[s] + buf.twoDigits(1, int(month)) + buf.twoDigits(3, day) + buf.tmp[5] = ' ' + buf.twoDigits(6, hour) + buf.tmp[8] = ':' + buf.twoDigits(9, minute) + buf.tmp[11] = ':' + buf.twoDigits(12, second) + buf.tmp[14] = '.' + buf.nDigits(6, 15, now.Nanosecond()/1000, '0') + buf.tmp[21] = ' ' + buf.nDigits(7, 22, pid, ' ') // TODO: should be TID + buf.tmp[29] = ' ' + buf.Write(buf.tmp[:30]) + buf.WriteString(file) + buf.tmp[0] = ':' + n := buf.someDigits(1, line) + buf.tmp[n+1] = ']' + buf.tmp[n+2] = ' ' + buf.Write(buf.tmp[:n+3]) + return buf +} + +// Some custom tiny helper functions to print the log header efficiently. + +const digits = "0123456789" + +// twoDigits formats a zero-prefixed two-digit integer at buf.tmp[i]. +func (buf *buffer) twoDigits(i, d int) { + buf.tmp[i+1] = digits[d%10] + d /= 10 + buf.tmp[i] = digits[d%10] +} + +// nDigits formats an n-digit integer at buf.tmp[i], +// padding with pad on the left. +// It assumes d >= 0. +func (buf *buffer) nDigits(n, i, d int, pad byte) { + j := n - 1 + for ; j >= 0 && d > 0; j-- { + buf.tmp[i+j] = digits[d%10] + d /= 10 + } + for ; j >= 0; j-- { + buf.tmp[i+j] = pad + } +} + +// someDigits formats a zero-prefixed variable-width integer at buf.tmp[i]. +func (buf *buffer) someDigits(i, d int) int { + // Print into the top, then copy down. We know there's space for at least + // a 10-digit number. + j := len(buf.tmp) + for { + j-- + buf.tmp[j] = digits[d%10] + d /= 10 + if d == 0 { + break + } + } + return copy(buf.tmp[i:], buf.tmp[j:]) +} + +func (l *loggingT) println(s severity, logr logr.Logger, args ...interface{}) { + buf, file, line := l.header(s, 0) + // if logr is set, we clear the generated header as we rely on the backing + // logr implementation to print headers + if logr != nil { + l.putBuffer(buf) + buf = l.getBuffer() + } + fmt.Fprintln(buf, args...) + l.output(s, logr, buf, file, line, false) +} + +func (l *loggingT) print(s severity, logr logr.Logger, args ...interface{}) { + l.printDepth(s, logr, 1, args...) +} + +func (l *loggingT) printDepth(s severity, logr logr.Logger, depth int, args ...interface{}) { + buf, file, line := l.header(s, depth) + // if logr is set, we clear the generated header as we rely on the backing + // logr implementation to print headers + if logr != nil { + l.putBuffer(buf) + buf = l.getBuffer() + } + fmt.Fprint(buf, args...) + if buf.Bytes()[buf.Len()-1] != '\n' { + buf.WriteByte('\n') + } + l.output(s, logr, buf, file, line, false) +} + +func (l *loggingT) printf(s severity, logr logr.Logger, format string, args ...interface{}) { + buf, file, line := l.header(s, 0) + // if logr is set, we clear the generated header as we rely on the backing + // logr implementation to print headers + if logr != nil { + l.putBuffer(buf) + buf = l.getBuffer() + } + fmt.Fprintf(buf, format, args...) + if buf.Bytes()[buf.Len()-1] != '\n' { + buf.WriteByte('\n') + } + l.output(s, logr, buf, file, line, false) +} + +// printWithFileLine behaves like print but uses the provided file and line number. If +// alsoLogToStderr is true, the log message always appears on standard error; it +// will also appear in the log file unless --logtostderr is set. +func (l *loggingT) printWithFileLine(s severity, logr logr.Logger, file string, line int, alsoToStderr bool, args ...interface{}) { + buf := l.formatHeader(s, file, line) + // if logr is set, we clear the generated header as we rely on the backing + // logr implementation to print headers + if logr != nil { + l.putBuffer(buf) + buf = l.getBuffer() + } + fmt.Fprint(buf, args...) + if buf.Bytes()[buf.Len()-1] != '\n' { + buf.WriteByte('\n') + } + l.output(s, logr, buf, file, line, alsoToStderr) +} + +// if loggr is specified, will call loggr.Error, otherwise output with logging module. +func (l *loggingT) errorS(err error, loggr logr.Logger, msg string, keysAndValues ...interface{}) { + if loggr != nil { + loggr.Error(err, msg, keysAndValues) + return + } + l.printS(err, msg, keysAndValues...) +} + +// if loggr is specified, will call loggr.Info, otherwise output with logging module. +func (l *loggingT) infoS(loggr logr.Logger, msg string, keysAndValues ...interface{}) { + if loggr != nil { + loggr.Info(msg, keysAndValues) + return + } + l.printS(nil, msg, keysAndValues...) +} + +// printS is called from infoS and errorS if loggr is not specified. +// if err arguments is specified, will output to errorLog severity +func (l *loggingT) printS(err error, msg string, keysAndValues ...interface{}) { + b := &bytes.Buffer{} + b.WriteString(fmt.Sprintf("%q", msg)) + if err != nil { + b.WriteByte(' ') + b.WriteString(fmt.Sprintf("err=%q", err.Error())) + } + kvListFormat(b, keysAndValues...) + var s severity + if err == nil { + s = infoLog + } else { + s = errorLog + } + l.printDepth(s, logging.logr, 2, b) +} + +const missingValue = "(MISSING)" + +func kvListFormat(b *bytes.Buffer, keysAndValues ...interface{}) { + for i := 0; i < len(keysAndValues); i += 2 { + var v interface{} + k := keysAndValues[i] + if i+1 < len(keysAndValues) { + v = keysAndValues[i+1] + } else { + v = missingValue + } + b.WriteByte(' ') + + switch v.(type) { + case string, error: + b.WriteString(fmt.Sprintf("%s=%q", k, v)) + default: + if _, ok := v.(fmt.Stringer); ok { + b.WriteString(fmt.Sprintf("%s=%q", k, v)) + } else { + b.WriteString(fmt.Sprintf("%s=%+v", k, v)) + } + } + } +} + +// redirectBuffer is used to set an alternate destination for the logs +type redirectBuffer struct { + w io.Writer +} + +func (rb *redirectBuffer) Sync() error { + return nil +} + +func (rb *redirectBuffer) Flush() error { + return nil +} + +func (rb *redirectBuffer) Write(bytes []byte) (n int, err error) { + return rb.w.Write(bytes) +} + +// SetLogger will set the backing logr implementation for klog. +// If set, all log lines will be suppressed from the regular Output, and +// redirected to the logr implementation. +// All log lines include the 'severity', 'file' and 'line' values attached as +// structured logging values. +// Use as: +// ... +// klog.SetLogger(zapr.NewLogger(zapLog)) +func SetLogger(logr logr.Logger) { + logging.logr = logr +} + +// SetOutput sets the output destination for all severities +func SetOutput(w io.Writer) { + logging.mu.Lock() + defer logging.mu.Unlock() + for s := fatalLog; s >= infoLog; s-- { + rb := &redirectBuffer{ + w: w, + } + logging.file[s] = rb + } +} + +// SetOutputBySeverity sets the output destination for specific severity +func SetOutputBySeverity(name string, w io.Writer) { + logging.mu.Lock() + defer logging.mu.Unlock() + sev, ok := severityByName(name) + if !ok { + panic(fmt.Sprintf("SetOutputBySeverity(%q): unrecognized severity name", name)) + } + rb := &redirectBuffer{ + w: w, + } + logging.file[sev] = rb +} + +// LogToStderr sets whether to log exclusively to stderr, bypassing outputs +func LogToStderr(stderr bool) { + logging.mu.Lock() + defer logging.mu.Unlock() + + logging.toStderr = stderr +} + +// output writes the data to the log files and releases the buffer. +func (l *loggingT) output(s severity, log logr.Logger, buf *buffer, file string, line int, alsoToStderr bool) { + l.mu.Lock() + if l.traceLocation.isSet() { + if l.traceLocation.match(file, line) { + buf.Write(stacks(false)) + } + } + data := buf.Bytes() + if log != nil { + // TODO: set 'severity' and caller information as structured log info + // keysAndValues := []interface{}{"severity", severityName[s], "file", file, "line", line} + if s == errorLog { + l.logr.Error(nil, string(data)) + } else { + log.Info(string(data)) + } + } else if l.toStderr { + os.Stderr.Write(data) + } else { + if alsoToStderr || l.alsoToStderr || s >= l.stderrThreshold.get() { + os.Stderr.Write(data) + } + + if logging.logFile != "" { + // Since we are using a single log file, all of the items in l.file array + // will point to the same file, so just use one of them to write data. + if l.file[infoLog] == nil { + if err := l.createFiles(infoLog); err != nil { + os.Stderr.Write(data) // Make sure the message appears somewhere. + l.exit(err) + } + } + l.file[infoLog].Write(data) + } else { + if l.file[s] == nil { + if err := l.createFiles(s); err != nil { + os.Stderr.Write(data) // Make sure the message appears somewhere. + l.exit(err) + } + } + + switch s { + case fatalLog: + l.file[fatalLog].Write(data) + fallthrough + case errorLog: + l.file[errorLog].Write(data) + fallthrough + case warningLog: + l.file[warningLog].Write(data) + fallthrough + case infoLog: + l.file[infoLog].Write(data) + } + } + } + if s == fatalLog { + // If we got here via Exit rather than Fatal, print no stacks. + if atomic.LoadUint32(&fatalNoStacks) > 0 { + l.mu.Unlock() + timeoutFlush(10 * time.Second) + os.Exit(1) + } + // Dump all goroutine stacks before exiting. + trace := stacks(true) + // Write the stack trace for all goroutines to the stderr. + if l.toStderr || l.alsoToStderr || s >= l.stderrThreshold.get() || alsoToStderr { + os.Stderr.Write(trace) + } + // Write the stack trace for all goroutines to the files. + logExitFunc = func(error) {} // If we get a write error, we'll still exit below. + for log := fatalLog; log >= infoLog; log-- { + if f := l.file[log]; f != nil { // Can be nil if -logtostderr is set. + f.Write(trace) + } + } + l.mu.Unlock() + timeoutFlush(10 * time.Second) + os.Exit(255) // C++ uses -1, which is silly because it's anded with 255 anyway. + } + l.putBuffer(buf) + l.mu.Unlock() + if stats := severityStats[s]; stats != nil { + atomic.AddInt64(&stats.lines, 1) + atomic.AddInt64(&stats.bytes, int64(len(data))) + } +} + +// timeoutFlush calls Flush and returns when it completes or after timeout +// elapses, whichever happens first. This is needed because the hooks invoked +// by Flush may deadlock when klog.Fatal is called from a hook that holds +// a lock. +func timeoutFlush(timeout time.Duration) { + done := make(chan bool, 1) + go func() { + Flush() // calls logging.lockAndFlushAll() + done <- true + }() + select { + case <-done: + case <-time.After(timeout): + fmt.Fprintln(os.Stderr, "klog: Flush took longer than", timeout) + } +} + +// stacks is a wrapper for runtime.Stack that attempts to recover the data for all goroutines. +func stacks(all bool) []byte { + // We don't know how big the traces are, so grow a few times if they don't fit. Start large, though. + n := 10000 + if all { + n = 100000 + } + var trace []byte + for i := 0; i < 5; i++ { + trace = make([]byte, n) + nbytes := runtime.Stack(trace, all) + if nbytes < len(trace) { + return trace[:nbytes] + } + n *= 2 + } + return trace +} + +// logExitFunc provides a simple mechanism to override the default behavior +// of exiting on error. Used in testing and to guarantee we reach a required exit +// for fatal logs. Instead, exit could be a function rather than a method but that +// would make its use clumsier. +var logExitFunc func(error) + +// exit is called if there is trouble creating or writing log files. +// It flushes the logs and exits the program; there's no point in hanging around. +// l.mu is held. +func (l *loggingT) exit(err error) { + fmt.Fprintf(os.Stderr, "log: exiting because of error: %s\n", err) + // If logExitFunc is set, we do that instead of exiting. + if logExitFunc != nil { + logExitFunc(err) + return + } + l.flushAll() + os.Exit(2) +} + +// syncBuffer joins a bufio.Writer to its underlying file, providing access to the +// file's Sync method and providing a wrapper for the Write method that provides log +// file rotation. There are conflicting methods, so the file cannot be embedded. +// l.mu is held for all its methods. +type syncBuffer struct { + logger *loggingT + *bufio.Writer + file *os.File + sev severity + nbytes uint64 // The number of bytes written to this file + maxbytes uint64 // The max number of bytes this syncBuffer.file can hold before cleaning up. +} + +func (sb *syncBuffer) Sync() error { + return sb.file.Sync() +} + +// CalculateMaxSize returns the real max size in bytes after considering the default max size and the flag options. +func CalculateMaxSize() uint64 { + if logging.logFile != "" { + if logging.logFileMaxSizeMB == 0 { + // If logFileMaxSizeMB is zero, we don't have limitations on the log size. + return math.MaxUint64 + } + // Flag logFileMaxSizeMB is in MB for user convenience. + return logging.logFileMaxSizeMB * 1024 * 1024 + } + // If "log_file" flag is not specified, the target file (sb.file) will be cleaned up when reaches a fixed size. + return MaxSize +} + +func (sb *syncBuffer) Write(p []byte) (n int, err error) { + if sb.nbytes+uint64(len(p)) >= sb.maxbytes { + if err := sb.rotateFile(time.Now(), false); err != nil { + sb.logger.exit(err) + } + } + n, err = sb.Writer.Write(p) + sb.nbytes += uint64(n) + if err != nil { + sb.logger.exit(err) + } + return +} + +// rotateFile closes the syncBuffer's file and starts a new one. +// The startup argument indicates whether this is the initial startup of klog. +// If startup is true, existing files are opened for appending instead of truncated. +func (sb *syncBuffer) rotateFile(now time.Time, startup bool) error { + if sb.file != nil { + sb.Flush() + sb.file.Close() + } + var err error + sb.file, _, err = create(severityName[sb.sev], now, startup) + sb.nbytes = 0 + if err != nil { + return err + } + + sb.Writer = bufio.NewWriterSize(sb.file, bufferSize) + + if sb.logger.skipLogHeaders { + return nil + } + + // Write header. + var buf bytes.Buffer + fmt.Fprintf(&buf, "Log file created at: %s\n", now.Format("2006/01/02 15:04:05")) + fmt.Fprintf(&buf, "Running on machine: %s\n", host) + fmt.Fprintf(&buf, "Binary: Built with %s %s for %s/%s\n", runtime.Compiler, runtime.Version(), runtime.GOOS, runtime.GOARCH) + fmt.Fprintf(&buf, "Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg\n") + n, err := sb.file.Write(buf.Bytes()) + sb.nbytes += uint64(n) + return err +} + +// bufferSize sizes the buffer associated with each log file. It's large +// so that log records can accumulate without the logging thread blocking +// on disk I/O. The flushDaemon will block instead. +const bufferSize = 256 * 1024 + +// createFiles creates all the log files for severity from sev down to infoLog. +// l.mu is held. +func (l *loggingT) createFiles(sev severity) error { + now := time.Now() + // Files are created in decreasing severity order, so as soon as we find one + // has already been created, we can stop. + for s := sev; s >= infoLog && l.file[s] == nil; s-- { + sb := &syncBuffer{ + logger: l, + sev: s, + maxbytes: CalculateMaxSize(), + } + if err := sb.rotateFile(now, true); err != nil { + return err + } + l.file[s] = sb + } + return nil +} + +const flushInterval = 5 * time.Second + +// flushDaemon periodically flushes the log file buffers. +func (l *loggingT) flushDaemon() { + for range time.NewTicker(flushInterval).C { + l.lockAndFlushAll() + } +} + +// lockAndFlushAll is like flushAll but locks l.mu first. +func (l *loggingT) lockAndFlushAll() { + l.mu.Lock() + l.flushAll() + l.mu.Unlock() +} + +// flushAll flushes all the logs and attempts to "sync" their data to disk. +// l.mu is held. +func (l *loggingT) flushAll() { + // Flush from fatal down, in case there's trouble flushing. + for s := fatalLog; s >= infoLog; s-- { + file := l.file[s] + if file != nil { + file.Flush() // ignore error + file.Sync() // ignore error + } + } +} + +// CopyStandardLogTo arranges for messages written to the Go "log" package's +// default logs to also appear in the Google logs for the named and lower +// severities. Subsequent changes to the standard log's default output location +// or format may break this behavior. +// +// Valid names are "INFO", "WARNING", "ERROR", and "FATAL". If the name is not +// recognized, CopyStandardLogTo panics. +func CopyStandardLogTo(name string) { + sev, ok := severityByName(name) + if !ok { + panic(fmt.Sprintf("log.CopyStandardLogTo(%q): unrecognized severity name", name)) + } + // Set a log format that captures the user's file and line: + // d.go:23: message + stdLog.SetFlags(stdLog.Lshortfile) + stdLog.SetOutput(logBridge(sev)) +} + +// logBridge provides the Write method that enables CopyStandardLogTo to connect +// Go's standard logs to the logs provided by this package. +type logBridge severity + +// Write parses the standard logging line and passes its components to the +// logger for severity(lb). +func (lb logBridge) Write(b []byte) (n int, err error) { + var ( + file = "???" + line = 1 + text string + ) + // Split "d.go:23: message" into "d.go", "23", and "message". + if parts := bytes.SplitN(b, []byte{':'}, 3); len(parts) != 3 || len(parts[0]) < 1 || len(parts[2]) < 1 { + text = fmt.Sprintf("bad log format: %s", b) + } else { + file = string(parts[0]) + text = string(parts[2][1:]) // skip leading space + line, err = strconv.Atoi(string(parts[1])) + if err != nil { + text = fmt.Sprintf("bad line number: %s", b) + line = 1 + } + } + // printWithFileLine with alsoToStderr=true, so standard log messages + // always appear on standard error. + logging.printWithFileLine(severity(lb), logging.logr, file, line, true, text) + return len(b), nil +} + +// setV computes and remembers the V level for a given PC +// when vmodule is enabled. +// File pattern matching takes the basename of the file, stripped +// of its .go suffix, and uses filepath.Match, which is a little more +// general than the *? matching used in C++. +// l.mu is held. +func (l *loggingT) setV(pc uintptr) Level { + fn := runtime.FuncForPC(pc) + file, _ := fn.FileLine(pc) + // The file is something like /a/b/c/d.go. We want just the d. + if strings.HasSuffix(file, ".go") { + file = file[:len(file)-3] + } + if slash := strings.LastIndex(file, "/"); slash >= 0 { + file = file[slash+1:] + } + for _, filter := range l.vmodule.filter { + if filter.match(file) { + l.vmap[pc] = filter.level + return filter.level + } + } + l.vmap[pc] = 0 + return 0 +} + +// Verbose is a boolean type that implements Infof (like Printf) etc. +// See the documentation of V for more information. +type Verbose struct { + enabled bool + logr logr.Logger +} + +func newVerbose(level Level, b bool) Verbose { + if logging.logr == nil { + return Verbose{b, nil} + } + return Verbose{b, logging.logr.V(int(level))} +} + +// V reports whether verbosity at the call site is at least the requested level. +// The returned value is a struct of type Verbose, which implements Info, Infoln +// and Infof. These methods will write to the Info log if called. +// Thus, one may write either +// if glog.V(2).Enabled() { klog.Info("log this") } +// or +// klog.V(2).Info("log this") +// The second form is shorter but the first is cheaper if logging is off because it does +// not evaluate its arguments. +// +// Whether an individual call to V generates a log record depends on the setting of +// the -v and -vmodule flags; both are off by default. The V call will log if its level +// is less than or equal to the value of the -v flag, or alternatively if its level is +// less than or equal to the value of the -vmodule pattern matching the source file +// containing the call. +func V(level Level) Verbose { + // This function tries hard to be cheap unless there's work to do. + // The fast path is two atomic loads and compares. + + // Here is a cheap but safe test to see if V logging is enabled globally. + if logging.verbosity.get() >= level { + return newVerbose(level, true) + } + + // It's off globally but it vmodule may still be set. + // Here is another cheap but safe test to see if vmodule is enabled. + if atomic.LoadInt32(&logging.filterLength) > 0 { + // Now we need a proper lock to use the logging structure. The pcs field + // is shared so we must lock before accessing it. This is fairly expensive, + // but if V logging is enabled we're slow anyway. + logging.mu.Lock() + defer logging.mu.Unlock() + if runtime.Callers(2, logging.pcs[:]) == 0 { + return newVerbose(level, false) + } + v, ok := logging.vmap[logging.pcs[0]] + if !ok { + v = logging.setV(logging.pcs[0]) + } + return newVerbose(level, v >= level) + } + return newVerbose(level, false) +} + +// Enabled will return true if this log level is enabled, guarded by the value +// of v. +// See the documentation of V for usage. +func (v Verbose) Enabled() bool { + return v.enabled +} + +// Info is equivalent to the global Info function, guarded by the value of v. +// See the documentation of V for usage. +func (v Verbose) Info(args ...interface{}) { + if v.enabled { + logging.print(infoLog, v.logr, args...) + } +} + +// Infoln is equivalent to the global Infoln function, guarded by the value of v. +// See the documentation of V for usage. +func (v Verbose) Infoln(args ...interface{}) { + if v.enabled { + logging.println(infoLog, v.logr, args...) + } +} + +// Infof is equivalent to the global Infof function, guarded by the value of v. +// See the documentation of V for usage. +func (v Verbose) Infof(format string, args ...interface{}) { + if v.enabled { + logging.printf(infoLog, v.logr, format, args...) + } +} + +// InfoS is equivalent to the global InfoS function, guarded by the value of v. +// See the documentation of V for usage. +func (v Verbose) InfoS(msg string, keysAndValues ...interface{}) { + if v.enabled { + logging.infoS(v.logr, msg, keysAndValues...) + } +} + +// Error is equivalent to the global Error function, guarded by the value of v. +// See the documentation of V for usage. +func (v Verbose) Error(err error, msg string, args ...interface{}) { + if v.enabled { + logging.errorS(err, v.logr, msg, args...) + } +} + +// Info logs to the INFO log. +// Arguments are handled in the manner of fmt.Print; a newline is appended if missing. +func Info(args ...interface{}) { + logging.print(infoLog, logging.logr, args...) +} + +// InfoDepth acts as Info but uses depth to determine which call frame to log. +// InfoDepth(0, "msg") is the same as Info("msg"). +func InfoDepth(depth int, args ...interface{}) { + logging.printDepth(infoLog, logging.logr, depth, args...) +} + +// Infoln logs to the INFO log. +// Arguments are handled in the manner of fmt.Println; a newline is always appended. +func Infoln(args ...interface{}) { + logging.println(infoLog, logging.logr, args...) +} + +// Infof logs to the INFO log. +// Arguments are handled in the manner of fmt.Printf; a newline is appended if missing. +func Infof(format string, args ...interface{}) { + logging.printf(infoLog, logging.logr, format, args...) +} + +// InfoS structured logs to the INFO log. +// The msg argument used to add constant description to the log line. +// The key/value pairs would be join by "=" ; a newline is always appended. +// +// Basic examples: +// >> klog.InfoS("Pod status updated", "pod", "kubedns", "status", "ready") +// output: +// >> I1025 00:15:15.525108 1 controller_utils.go:116] "Pod status updated" pod="kubedns" status="ready" +func InfoS(msg string, keysAndValues ...interface{}) { + logging.infoS(logging.logr, msg, keysAndValues...) +} + +// Warning logs to the WARNING and INFO logs. +// Arguments are handled in the manner of fmt.Print; a newline is appended if missing. +func Warning(args ...interface{}) { + logging.print(warningLog, logging.logr, args...) +} + +// WarningDepth acts as Warning but uses depth to determine which call frame to log. +// WarningDepth(0, "msg") is the same as Warning("msg"). +func WarningDepth(depth int, args ...interface{}) { + logging.printDepth(warningLog, logging.logr, depth, args...) +} + +// Warningln logs to the WARNING and INFO logs. +// Arguments are handled in the manner of fmt.Println; a newline is always appended. +func Warningln(args ...interface{}) { + logging.println(warningLog, logging.logr, args...) +} + +// Warningf logs to the WARNING and INFO logs. +// Arguments are handled in the manner of fmt.Printf; a newline is appended if missing. +func Warningf(format string, args ...interface{}) { + logging.printf(warningLog, logging.logr, format, args...) +} + +// Error logs to the ERROR, WARNING, and INFO logs. +// Arguments are handled in the manner of fmt.Print; a newline is appended if missing. +func Error(args ...interface{}) { + logging.print(errorLog, logging.logr, args...) +} + +// ErrorDepth acts as Error but uses depth to determine which call frame to log. +// ErrorDepth(0, "msg") is the same as Error("msg"). +func ErrorDepth(depth int, args ...interface{}) { + logging.printDepth(errorLog, logging.logr, depth, args...) +} + +// Errorln logs to the ERROR, WARNING, and INFO logs. +// Arguments are handled in the manner of fmt.Println; a newline is always appended. +func Errorln(args ...interface{}) { + logging.println(errorLog, logging.logr, args...) +} + +// Errorf logs to the ERROR, WARNING, and INFO logs. +// Arguments are handled in the manner of fmt.Printf; a newline is appended if missing. +func Errorf(format string, args ...interface{}) { + logging.printf(errorLog, logging.logr, format, args...) +} + +// ErrorS structured logs to the ERROR, WARNING, and INFO logs. +// the err argument used as "err" field of log line. +// The msg argument used to add constant description to the log line. +// The key/value pairs would be join by "=" ; a newline is always appended. +// +// Basic examples: +// >> klog.ErrorS(err, "Failed to update pod status") +// output: +// >> E1025 00:15:15.525108 1 controller_utils.go:114] "Failed to update pod status" err="timeout" +func ErrorS(err error, msg string, keysAndValues ...interface{}) { + logging.errorS(err, logging.logr, msg, keysAndValues...) +} + +// Fatal logs to the FATAL, ERROR, WARNING, and INFO logs, +// including a stack trace of all running goroutines, then calls os.Exit(255). +// Arguments are handled in the manner of fmt.Print; a newline is appended if missing. +func Fatal(args ...interface{}) { + logging.print(fatalLog, logging.logr, args...) +} + +// FatalDepth acts as Fatal but uses depth to determine which call frame to log. +// FatalDepth(0, "msg") is the same as Fatal("msg"). +func FatalDepth(depth int, args ...interface{}) { + logging.printDepth(fatalLog, logging.logr, depth, args...) +} + +// Fatalln logs to the FATAL, ERROR, WARNING, and INFO logs, +// including a stack trace of all running goroutines, then calls os.Exit(255). +// Arguments are handled in the manner of fmt.Println; a newline is always appended. +func Fatalln(args ...interface{}) { + logging.println(fatalLog, logging.logr, args...) +} + +// Fatalf logs to the FATAL, ERROR, WARNING, and INFO logs, +// including a stack trace of all running goroutines, then calls os.Exit(255). +// Arguments are handled in the manner of fmt.Printf; a newline is appended if missing. +func Fatalf(format string, args ...interface{}) { + logging.printf(fatalLog, logging.logr, format, args...) +} + +// fatalNoStacks is non-zero if we are to exit without dumping goroutine stacks. +// It allows Exit and relatives to use the Fatal logs. +var fatalNoStacks uint32 + +// Exit logs to the FATAL, ERROR, WARNING, and INFO logs, then calls os.Exit(1). +// Arguments are handled in the manner of fmt.Print; a newline is appended if missing. +func Exit(args ...interface{}) { + atomic.StoreUint32(&fatalNoStacks, 1) + logging.print(fatalLog, logging.logr, args...) +} + +// ExitDepth acts as Exit but uses depth to determine which call frame to log. +// ExitDepth(0, "msg") is the same as Exit("msg"). +func ExitDepth(depth int, args ...interface{}) { + atomic.StoreUint32(&fatalNoStacks, 1) + logging.printDepth(fatalLog, logging.logr, depth, args...) +} + +// Exitln logs to the FATAL, ERROR, WARNING, and INFO logs, then calls os.Exit(1). +func Exitln(args ...interface{}) { + atomic.StoreUint32(&fatalNoStacks, 1) + logging.println(fatalLog, logging.logr, args...) +} + +// Exitf logs to the FATAL, ERROR, WARNING, and INFO logs, then calls os.Exit(1). +// Arguments are handled in the manner of fmt.Printf; a newline is appended if missing. +func Exitf(format string, args ...interface{}) { + atomic.StoreUint32(&fatalNoStacks, 1) + logging.printf(fatalLog, logging.logr, format, args...) +} + +// ObjectRef references a kubernetes object +type ObjectRef struct { + Name string `json:"name"` + Namespace string `json:"namespace,omitempty"` +} + +func (ref ObjectRef) String() string { + if ref.Namespace != "" { + return fmt.Sprintf("%s/%s", ref.Namespace, ref.Name) + } + return ref.Name +} + +// KMetadata is a subset of the kubernetes k8s.io/apimachinery/pkg/apis/meta/v1.Object interface +// this interface may expand in the future, but will always be a subset of the +// kubernetes k8s.io/apimachinery/pkg/apis/meta/v1.Object interface +type KMetadata interface { + GetName() string + GetNamespace() string +} + +// KObj returns ObjectRef from ObjectMeta +func KObj(obj KMetadata) ObjectRef { + return ObjectRef{ + Name: obj.GetName(), + Namespace: obj.GetNamespace(), + } +} + +// KRef returns ObjectRef from name and namespace +func KRef(namespace, name string) ObjectRef { + return ObjectRef{ + Name: name, + Namespace: namespace, + } +} diff --git a/vendor/k8s.io/klog/v2/klog_file.go b/vendor/k8s.io/klog/v2/klog_file.go new file mode 100644 index 00000000..de830d92 --- /dev/null +++ b/vendor/k8s.io/klog/v2/klog_file.go @@ -0,0 +1,164 @@ +// Go support for leveled logs, analogous to https://code.google.com/p/google-glog/ +// +// Copyright 2013 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// File I/O for logs. + +package klog + +import ( + "errors" + "fmt" + "os" + "os/user" + "path/filepath" + "runtime" + "strings" + "sync" + "time" +) + +// MaxSize is the maximum size of a log file in bytes. +var MaxSize uint64 = 1024 * 1024 * 1800 + +// logDirs lists the candidate directories for new log files. +var logDirs []string + +func createLogDirs() { + if logging.logDir != "" { + logDirs = append(logDirs, logging.logDir) + } + logDirs = append(logDirs, os.TempDir()) +} + +var ( + pid = os.Getpid() + program = filepath.Base(os.Args[0]) + host = "unknownhost" + userName = "unknownuser" + userNameOnce sync.Once +) + +func init() { + if h, err := os.Hostname(); err == nil { + host = shortHostname(h) + } +} + +func getUserName() string { + userNameOnce.Do(func() { + // On Windows, the Go 'user' package requires netapi32.dll. + // This affects Windows Nano Server: + // https://github.com/golang/go/issues/21867 + // Fallback to using environment variables. + if runtime.GOOS == "windows" { + u := os.Getenv("USERNAME") + if len(u) == 0 { + return + } + // Sanitize the USERNAME since it may contain filepath separators. + u = strings.Replace(u, `\`, "_", -1) + + // user.Current().Username normally produces something like 'USERDOMAIN\USERNAME' + d := os.Getenv("USERDOMAIN") + if len(d) != 0 { + userName = d + "_" + u + } else { + userName = u + } + } else { + current, err := user.Current() + if err == nil { + userName = current.Username + } + } + }) + + return userName +} + +// shortHostname returns its argument, truncating at the first period. +// For instance, given "www.google.com" it returns "www". +func shortHostname(hostname string) string { + if i := strings.Index(hostname, "."); i >= 0 { + return hostname[:i] + } + return hostname +} + +// logName returns a new log file name containing tag, with start time t, and +// the name for the symlink for tag. +func logName(tag string, t time.Time) (name, link string) { + name = fmt.Sprintf("%s.%s.%s.log.%s.%04d%02d%02d-%02d%02d%02d.%d", + program, + host, + getUserName(), + tag, + t.Year(), + t.Month(), + t.Day(), + t.Hour(), + t.Minute(), + t.Second(), + pid) + return name, program + "." + tag +} + +var onceLogDirs sync.Once + +// create creates a new log file and returns the file and its filename, which +// contains tag ("INFO", "FATAL", etc.) and t. If the file is created +// successfully, create also attempts to update the symlink for that tag, ignoring +// errors. +// The startup argument indicates whether this is the initial startup of klog. +// If startup is true, existing files are opened for appending instead of truncated. +func create(tag string, t time.Time, startup bool) (f *os.File, filename string, err error) { + if logging.logFile != "" { + f, err := openOrCreate(logging.logFile, startup) + if err == nil { + return f, logging.logFile, nil + } + return nil, "", fmt.Errorf("log: unable to create log: %v", err) + } + onceLogDirs.Do(createLogDirs) + if len(logDirs) == 0 { + return nil, "", errors.New("log: no log dirs") + } + name, link := logName(tag, t) + var lastErr error + for _, dir := range logDirs { + fname := filepath.Join(dir, name) + f, err := openOrCreate(fname, startup) + if err == nil { + symlink := filepath.Join(dir, link) + os.Remove(symlink) // ignore err + os.Symlink(name, symlink) // ignore err + return f, fname, nil + } + lastErr = err + } + return nil, "", fmt.Errorf("log: cannot create log: %v", lastErr) +} + +// The startup argument indicates whether this is the initial startup of klog. +// If startup is true, existing files are opened for appending instead of truncated. +func openOrCreate(name string, startup bool) (*os.File, error) { + if startup { + f, err := os.OpenFile(name, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0666) + return f, err + } + f, err := os.Create(name) + return f, err +} diff --git a/vendor/k8s.io/kube-openapi/pkg/util/proto/document.go b/vendor/k8s.io/kube-openapi/pkg/util/proto/document.go index 5eb957af..6df0df38 100644 --- a/vendor/k8s.io/kube-openapi/pkg/util/proto/document.go +++ b/vendor/k8s.io/kube-openapi/pkg/util/proto/document.go @@ -21,7 +21,7 @@ import ( "sort" "strings" - "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" "gopkg.in/yaml.v2" ) diff --git a/vendor/k8s.io/kube-openapi/pkg/util/proto/validation/types.go b/vendor/k8s.io/kube-openapi/pkg/util/proto/validation/types.go index 6a9f68c0..e66342a7 100644 --- a/vendor/k8s.io/kube-openapi/pkg/util/proto/validation/types.go +++ b/vendor/k8s.io/kube-openapi/pkg/util/proto/validation/types.go @@ -210,7 +210,7 @@ func (item *primitiveItem) VisitPrimitive(schema *proto.Primitive) { } case proto.Number: switch item.Kind { - case proto.Number: + case proto.Integer, proto.Number: return } case proto.String: diff --git a/vendor/github.com/gophercloud/gophercloud/LICENSE b/vendor/k8s.io/kubectl/LICENSE similarity index 89% rename from vendor/github.com/gophercloud/gophercloud/LICENSE rename to vendor/k8s.io/kubectl/LICENSE index fbbbc9e4..8dada3ed 100644 --- a/vendor/github.com/gophercloud/gophercloud/LICENSE +++ b/vendor/k8s.io/kubectl/LICENSE @@ -1,19 +1,4 @@ -Copyright 2012-2013 Rackspace, Inc. - -Licensed under the Apache License, Version 2.0 (the "License"); you may not use -this file except in compliance with the License. You may obtain a copy of the -License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software distributed -under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR -CONDITIONS OF ANY KIND, either express or implied. See the License for the -specific language governing permissions and limitations under the License. - ------- - - Apache License + Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ @@ -189,3 +174,28 @@ specific language governing permissions and limitations under the License. of your accepting any such warranty or additional liability. END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "{}" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright {yyyy} {name of copyright owner} + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/k8s.io/kubectl/pkg/util/openapi/OWNERS b/vendor/k8s.io/kubectl/pkg/util/openapi/OWNERS new file mode 100644 index 00000000..99dabed0 --- /dev/null +++ b/vendor/k8s.io/kubectl/pkg/util/openapi/OWNERS @@ -0,0 +1,6 @@ +# See the OWNERS docs at https://go.k8s.io/owners + +approvers: +- apelisse +reviewers: +- apelisse diff --git a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/doc.go b/vendor/k8s.io/kubectl/pkg/util/openapi/doc.go similarity index 92% rename from vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/doc.go rename to vendor/k8s.io/kubectl/pkg/util/openapi/doc.go index 7fff1fa8..08194d58 100644 --- a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/doc.go +++ b/vendor/k8s.io/kubectl/pkg/util/openapi/doc.go @@ -18,4 +18,4 @@ limitations under the License. // from a Kubernetes server and then indexing the type definitions. // The openapi spec contains the object model definitions and extensions metadata // such as the patchStrategy and patchMergeKey for creating patches. -package openapi // k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi +package openapi // k8s.io/kubectl/pkg/util/openapi diff --git a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/extensions.go b/vendor/k8s.io/kubectl/pkg/util/openapi/extensions.go similarity index 92% rename from vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/extensions.go rename to vendor/k8s.io/kubectl/pkg/util/openapi/extensions.go index 40eaacea..f1b5cdd4 100644 --- a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/extensions.go +++ b/vendor/k8s.io/kubectl/pkg/util/openapi/extensions.go @@ -18,6 +18,7 @@ package openapi import "github.com/go-openapi/spec" +// PrintColumnsKey is the key that defines which columns should be printed const PrintColumnsKey = "x-kubernetes-print-columns" // GetPrintColumns looks for the open API extension for the display columns. diff --git a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/openapi.go b/vendor/k8s.io/kubectl/pkg/util/openapi/openapi.go similarity index 96% rename from vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/openapi.go rename to vendor/k8s.io/kubectl/pkg/util/openapi/openapi.go index 80164862..b5d01db1 100644 --- a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/openapi.go +++ b/vendor/k8s.io/kubectl/pkg/util/openapi/openapi.go @@ -17,7 +17,7 @@ limitations under the License. package openapi import ( - openapi_v2 "github.com/googleapis/gnostic/OpenAPIv2" + openapi_v2 "github.com/googleapis/gnostic/openapiv2" "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/kube-openapi/pkg/util/proto" @@ -44,6 +44,7 @@ type document struct { var _ Resources = &document{} +// NewOpenAPIData creates a new `Resources` out of the openapi document func NewOpenAPIData(doc *openapi_v2.Document) (Resources, error) { models, err := proto.NewOpenAPIData(doc) if err != nil { diff --git a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/openapi_getter.go b/vendor/k8s.io/kubectl/pkg/util/openapi/openapi_getter.go similarity index 100% rename from vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/openapi_getter.go rename to vendor/k8s.io/kubectl/pkg/util/openapi/openapi_getter.go diff --git a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/BUILD b/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/BUILD deleted file mode 100644 index 5958bec3..00000000 --- a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/BUILD +++ /dev/null @@ -1,72 +0,0 @@ -package(default_visibility = ["//visibility:public"]) - -load( - "@io_bazel_rules_go//go:def.bzl", - "go_library", - "go_test", -) - -go_library( - name = "go_default_library", - srcs = [ - "doc.go", - "dryrun.go", - "extensions.go", - "openapi.go", - "openapi_getter.go", - ], - importpath = "k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi", - deps = [ - "//staging/src/k8s.io/apimachinery/pkg/runtime/schema:go_default_library", - "//staging/src/k8s.io/client-go/discovery:go_default_library", - "//vendor/github.com/go-openapi/spec:go_default_library", - "//vendor/github.com/googleapis/gnostic/OpenAPIv2:go_default_library", - "//vendor/gopkg.in/yaml.v2:go_default_library", - "//vendor/k8s.io/kube-openapi/pkg/util/proto:go_default_library", - ], -) - -go_test( - name = "go_default_test", - size = "small", - srcs = [ - "dryrun_test.go", - "openapi_getter_test.go", - "openapi_suite_test.go", - "openapi_test.go", - ], - data = ["//api/openapi-spec:swagger-spec"], - embed = [":go_default_library"], - deps = [ - "//staging/src/k8s.io/apimachinery/pkg/runtime/schema:go_default_library", - "//vendor/github.com/googleapis/gnostic/OpenAPIv2:go_default_library", - "//vendor/github.com/onsi/ginkgo:go_default_library", - "//vendor/github.com/onsi/ginkgo/config:go_default_library", - "//vendor/github.com/onsi/ginkgo/types:go_default_library", - "//vendor/github.com/onsi/gomega:go_default_library", - "//vendor/k8s.io/kube-openapi/pkg/util/proto:go_default_library", - "//vendor/k8s.io/kube-openapi/pkg/util/proto/testing:go_default_library", - ], -) - -filegroup( - name = "package-srcs", - srcs = glob(["**"]), - tags = ["automanaged"], - visibility = ["//visibility:private"], -) - -filegroup( - name = "all-srcs", - srcs = [ - ":package-srcs", - "//pkg/kubectl/cmd/util/openapi/testing:all-srcs", - "//pkg/kubectl/cmd/util/openapi/validation:all-srcs", - ], - tags = ["automanaged"], -) - -filegroup( - name = "testdata", - srcs = glob(["testdata/*"]), -) diff --git a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/OWNERS b/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/OWNERS deleted file mode 100644 index 7eb2d06f..00000000 --- a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/OWNERS +++ /dev/null @@ -1,4 +0,0 @@ -approvers: -- apelisse -reviewers: -- apelisse diff --git a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/dryrun.go b/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/dryrun.go deleted file mode 100644 index 33cf9e9e..00000000 --- a/vendor/k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi/dryrun.go +++ /dev/null @@ -1,65 +0,0 @@ -/* -Copyright 2017 The Kubernetes Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -*/ - -package openapi - -import ( - "errors" - - openapi_v2 "github.com/googleapis/gnostic/OpenAPIv2" - yaml "gopkg.in/yaml.v2" - "k8s.io/apimachinery/pkg/runtime/schema" -) - -func hasGVKExtension(extensions []*openapi_v2.NamedAny, gvk schema.GroupVersionKind) bool { - for _, extension := range extensions { - if extension.GetValue().GetYaml() == "" || - extension.GetName() != "x-kubernetes-group-version-kind" { - continue - } - var value map[string]string - err := yaml.Unmarshal([]byte(extension.GetValue().GetYaml()), &value) - if err != nil { - continue - } - - if value["group"] == gvk.Group && value["kind"] == gvk.Kind && value["version"] == gvk.Version { - return true - } - return false - } - return false -} - -// SupportsDryRun is a method that let's us look in the OpenAPI if the -// specific group-version-kind supports the dryRun query parameter for -// the PATCH end-point. -func SupportsDryRun(doc *openapi_v2.Document, gvk schema.GroupVersionKind) (bool, error) { - for _, path := range doc.GetPaths().GetPath() { - // Is this describing the gvk we're looking for? - if !hasGVKExtension(path.GetValue().GetPatch().GetVendorExtension(), gvk) { - continue - } - for _, param := range path.GetValue().GetPatch().GetParameters() { - if param.GetParameter().GetNonBodyParameter().GetQueryParameterSubSchema().GetName() == "dryRun" { - return true, nil - } - } - return false, nil - } - - return false, errors.New("couldn't find GVK in openapi") -} diff --git a/vendor/modules.txt b/vendor/modules.txt index d2067403..785f29d3 100644 --- a/vendor/modules.txt +++ b/vendor/modules.txt @@ -1,4 +1,4 @@ -# cloud.google.com/go v0.38.0 +# cloud.google.com/go v0.51.0 cloud.google.com/go/compute/metadata # github.com/Azure/go-autorest/autorest v0.10.0 github.com/Azure/go-autorest/autorest @@ -68,13 +68,15 @@ github.com/docker/go-units github.com/docker/libtrust # github.com/elazarl/go-bindata-assetfs v1.0.1-0.20180223160309-38087fe4dafb github.com/elazarl/go-bindata-assetfs -# github.com/evanphx/json-patch v4.5.0+incompatible +# github.com/evanphx/json-patch v4.9.0+incompatible github.com/evanphx/json-patch # github.com/genuinetools/reg v0.16.1 github.com/genuinetools/reg/registry github.com/genuinetools/reg/repoutils # github.com/ghodss/yaml v1.0.0 github.com/ghodss/yaml +# github.com/go-logr/logr v0.2.0 +github.com/go-logr/logr # github.com/go-openapi/jsonpointer v0.19.3 github.com/go-openapi/jsonpointer # github.com/go-openapi/jsonreference v0.19.3 @@ -86,7 +88,7 @@ github.com/go-openapi/swag # github.com/gogo/protobuf v1.3.1 github.com/gogo/protobuf/proto github.com/gogo/protobuf/sortkeys -# github.com/golang/protobuf v1.3.5 +# github.com/golang/protobuf v1.4.3 github.com/golang/protobuf/proto github.com/golang/protobuf/ptypes github.com/golang/protobuf/ptypes/any @@ -98,7 +100,7 @@ github.com/google/go-cmp/cmp/internal/diff github.com/google/go-cmp/cmp/internal/flags github.com/google/go-cmp/cmp/internal/function github.com/google/go-cmp/cmp/internal/value -# github.com/google/go-jsonnet v0.17.0 +# github.com/google/go-jsonnet v0.15.0 github.com/google/go-jsonnet github.com/google/go-jsonnet/ast github.com/google/go-jsonnet/astgen @@ -107,18 +109,11 @@ github.com/google/go-jsonnet/internal/parser github.com/google/go-jsonnet/internal/program # github.com/google/gofuzz v1.1.0 github.com/google/gofuzz -# github.com/googleapis/gnostic v0.4.0 => github.com/googleapis/gnostic v0.4.0 -github.com/googleapis/gnostic/OpenAPIv2 +# github.com/googleapis/gnostic v0.5.3 github.com/googleapis/gnostic/compiler github.com/googleapis/gnostic/extensions -# github.com/gophercloud/gophercloud v0.1.0 -github.com/gophercloud/gophercloud -github.com/gophercloud/gophercloud/openstack -github.com/gophercloud/gophercloud/openstack/identity/v2/tenants -github.com/gophercloud/gophercloud/openstack/identity/v2/tokens -github.com/gophercloud/gophercloud/openstack/identity/v3/tokens -github.com/gophercloud/gophercloud/openstack/utils -github.com/gophercloud/gophercloud/pagination +github.com/googleapis/gnostic/jsonschema +github.com/googleapis/gnostic/openapiv2 # github.com/hpcloud/tail v1.0.0 github.com/hpcloud/tail github.com/hpcloud/tail/ratelimiter @@ -129,9 +124,9 @@ github.com/hpcloud/tail/winfile github.com/imdario/mergo # github.com/inconshreveable/mousetrap v1.0.0 github.com/inconshreveable/mousetrap -# github.com/json-iterator/go v1.1.9 +# github.com/json-iterator/go v1.1.10 github.com/json-iterator/go -# github.com/konsorten/go-windows-terminal-sequences v1.0.2 +# github.com/konsorten/go-windows-terminal-sequences v1.0.3 github.com/konsorten/go-windows-terminal-sequences # github.com/mailru/easyjson v0.7.1 github.com/mailru/easyjson/buffer @@ -188,20 +183,20 @@ github.com/peterhellberg/link github.com/pkg/errors # github.com/pmezard/go-difflib v1.0.0 github.com/pmezard/go-difflib/difflib -# github.com/sergi/go-diff v1.1.0 +# github.com/sergi/go-diff v1.0.0 github.com/sergi/go-diff/diffmatchpatch -# github.com/sirupsen/logrus v1.4.2 +# github.com/sirupsen/logrus v1.6.0 github.com/sirupsen/logrus -# github.com/spf13/cobra v0.0.7 +# github.com/spf13/cobra v1.0.0 github.com/spf13/cobra # github.com/spf13/pflag v1.0.5 github.com/spf13/pflag -# github.com/stretchr/testify v1.4.0 +# github.com/stretchr/testify v1.5.1 github.com/stretchr/testify/assert github.com/stretchr/testify/require -# golang.org/x/crypto v0.0.0-20200406173513-056763e48d71 +# golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9 golang.org/x/crypto/ssh/terminal -# golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e +# golang.org/x/net v0.0.0-20200707034311-ab3426394381 golang.org/x/net/context golang.org/x/net/context/ctxhttp golang.org/x/net/html @@ -217,10 +212,11 @@ golang.org/x/oauth2/google golang.org/x/oauth2/internal golang.org/x/oauth2/jws golang.org/x/oauth2/jwt -# golang.org/x/sys v0.0.0-20200409092240-59c9f1ba88fa +# golang.org/x/sys v0.0.0-20200622214017-ed371f2e16b4 +golang.org/x/sys/internal/unsafeheader golang.org/x/sys/unix golang.org/x/sys/windows -# golang.org/x/text v0.3.2 +# golang.org/x/text v0.3.3 golang.org/x/text/encoding golang.org/x/text/encoding/charmap golang.org/x/text/encoding/htmlindex @@ -255,7 +251,7 @@ google.golang.org/appengine/internal/modules google.golang.org/appengine/internal/remote_api google.golang.org/appengine/internal/urlfetch google.golang.org/appengine/urlfetch -# google.golang.org/genproto v0.0.0-20200409111301-baae70f3302d +# google.golang.org/genproto v0.0.0-20201019141844-1ed22bb0c154 google.golang.org/genproto/googleapis/rpc/status # google.golang.org/grpc v1.28.1 google.golang.org/grpc/codes @@ -264,6 +260,37 @@ google.golang.org/grpc/grpclog google.golang.org/grpc/internal google.golang.org/grpc/internal/grpclog google.golang.org/grpc/status +# google.golang.org/protobuf v1.24.0 +google.golang.org/protobuf/encoding/prototext +google.golang.org/protobuf/encoding/protowire +google.golang.org/protobuf/internal/descfmt +google.golang.org/protobuf/internal/descopts +google.golang.org/protobuf/internal/detrand +google.golang.org/protobuf/internal/encoding/defval +google.golang.org/protobuf/internal/encoding/messageset +google.golang.org/protobuf/internal/encoding/tag +google.golang.org/protobuf/internal/encoding/text +google.golang.org/protobuf/internal/errors +google.golang.org/protobuf/internal/fieldnum +google.golang.org/protobuf/internal/fieldsort +google.golang.org/protobuf/internal/filedesc +google.golang.org/protobuf/internal/filetype +google.golang.org/protobuf/internal/flags +google.golang.org/protobuf/internal/genname +google.golang.org/protobuf/internal/impl +google.golang.org/protobuf/internal/mapsort +google.golang.org/protobuf/internal/pragma +google.golang.org/protobuf/internal/set +google.golang.org/protobuf/internal/strs +google.golang.org/protobuf/internal/version +google.golang.org/protobuf/proto +google.golang.org/protobuf/reflect/protoreflect +google.golang.org/protobuf/reflect/protoregistry +google.golang.org/protobuf/runtime/protoiface +google.golang.org/protobuf/runtime/protoimpl +google.golang.org/protobuf/types/known/anypb +google.golang.org/protobuf/types/known/durationpb +google.golang.org/protobuf/types/known/timestamppb # gopkg.in/fsnotify.v1 v1.4.7 gopkg.in/fsnotify.v1 # gopkg.in/inf.v0 v0.9.1 @@ -272,13 +299,14 @@ gopkg.in/inf.v0 gopkg.in/tomb.v1 # gopkg.in/yaml.v2 v2.2.8 gopkg.in/yaml.v2 -# k8s.io/api v0.17.4 +# gopkg.in/yaml.v3 v3.0.0-20200615113413-eeeca48fe776 +gopkg.in/yaml.v3 +# k8s.io/api v0.19.3 k8s.io/api/admissionregistration/v1 k8s.io/api/admissionregistration/v1beta1 k8s.io/api/apps/v1 k8s.io/api/apps/v1beta1 k8s.io/api/apps/v1beta2 -k8s.io/api/auditregistration/v1alpha1 k8s.io/api/authentication/v1 k8s.io/api/authentication/v1beta1 k8s.io/api/authorization/v1 @@ -289,12 +317,14 @@ k8s.io/api/autoscaling/v2beta2 k8s.io/api/batch/v1 k8s.io/api/batch/v1beta1 k8s.io/api/batch/v2alpha1 +k8s.io/api/certificates/v1 k8s.io/api/certificates/v1beta1 k8s.io/api/coordination/v1 k8s.io/api/coordination/v1beta1 k8s.io/api/core/v1 k8s.io/api/discovery/v1alpha1 k8s.io/api/discovery/v1beta1 +k8s.io/api/events/v1 k8s.io/api/events/v1beta1 k8s.io/api/extensions/v1beta1 k8s.io/api/flowcontrol/v1alpha1 @@ -313,10 +343,10 @@ k8s.io/api/settings/v1alpha1 k8s.io/api/storage/v1 k8s.io/api/storage/v1alpha1 k8s.io/api/storage/v1beta1 -# k8s.io/apiextensions-apiserver v0.17.4 +# k8s.io/apiextensions-apiserver v0.19.3 k8s.io/apiextensions-apiserver/pkg/apis/apiextensions k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1beta1 -# k8s.io/apimachinery v0.17.4 +# k8s.io/apimachinery v0.19.3 k8s.io/apimachinery/pkg/api/equality k8s.io/apimachinery/pkg/api/errors k8s.io/apimachinery/pkg/api/meta @@ -358,7 +388,7 @@ k8s.io/apimachinery/pkg/version k8s.io/apimachinery/pkg/watch k8s.io/apimachinery/third_party/forked/golang/json k8s.io/apimachinery/third_party/forked/golang/reflect -# k8s.io/client-go v0.17.4 +# k8s.io/client-go v0.19.3 k8s.io/client-go/discovery k8s.io/client-go/discovery/fake k8s.io/client-go/dynamic @@ -395,15 +425,20 @@ k8s.io/client-go/util/homedir k8s.io/client-go/util/jsonpath k8s.io/client-go/util/keyutil k8s.io/client-go/util/retry +k8s.io/client-go/util/workqueue # k8s.io/klog v1.0.0 k8s.io/klog -# k8s.io/kube-openapi v0.0.0-20191107075043-30be4d16710a +# k8s.io/klog/v2 v2.2.0 +k8s.io/klog/v2 +# k8s.io/kube-openapi v0.0.0-20200923155610-8b5066479488 k8s.io/kube-openapi/pkg/util/proto k8s.io/kube-openapi/pkg/util/proto/validation -# k8s.io/kubernetes v1.13.4 -k8s.io/kubernetes/pkg/kubectl/cmd/util/openapi -# k8s.io/utils v0.0.0-20200327001022-6496210b90e8 +# k8s.io/kubectl v0.19.3 +k8s.io/kubectl/pkg/util/openapi +# k8s.io/utils v0.0.0-20200729134348-d5654de09c73 k8s.io/utils/integer k8s.io/utils/pointer +# sigs.k8s.io/structured-merge-diff/v4 v4.0.1 +sigs.k8s.io/structured-merge-diff/v4/value # sigs.k8s.io/yaml v1.2.0 sigs.k8s.io/yaml diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/LICENSE b/vendor/sigs.k8s.io/structured-merge-diff/v4/LICENSE new file mode 100644 index 00000000..8dada3ed --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/LICENSE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "{}" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright {yyyy} {name of copyright owner} + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/allocator.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/allocator.go new file mode 100644 index 00000000..f70cd416 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/allocator.go @@ -0,0 +1,203 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +// Allocator provides a value object allocation strategy. +// Value objects can be allocated by passing an allocator to the "Using" +// receiver functions on the value interfaces, e.g. Map.ZipUsing(allocator, ...). +// Value objects returned from "Using" functions should be given back to the allocator +// once longer needed by calling Allocator.Free(Value). +type Allocator interface { + // Free gives the allocator back any value objects returned by the "Using" + // receiver functions on the value interfaces. + // interface{} may be any of: Value, Map, List or Range. + Free(interface{}) + + // The unexported functions are for "Using" receiver functions of the value types + // to request what they need from the allocator. + allocValueUnstructured() *valueUnstructured + allocListUnstructuredRange() *listUnstructuredRange + allocValueReflect() *valueReflect + allocMapReflect() *mapReflect + allocStructReflect() *structReflect + allocListReflect() *listReflect + allocListReflectRange() *listReflectRange +} + +// HeapAllocator simply allocates objects to the heap. It is the default +// allocator used receiver functions on the value interfaces that do not accept +// an allocator and should be used whenever allocating objects that will not +// be given back to an allocator by calling Allocator.Free(Value). +var HeapAllocator = &heapAllocator{} + +type heapAllocator struct{} + +func (p *heapAllocator) allocValueUnstructured() *valueUnstructured { + return &valueUnstructured{} +} + +func (p *heapAllocator) allocListUnstructuredRange() *listUnstructuredRange { + return &listUnstructuredRange{vv: &valueUnstructured{}} +} + +func (p *heapAllocator) allocValueReflect() *valueReflect { + return &valueReflect{} +} + +func (p *heapAllocator) allocStructReflect() *structReflect { + return &structReflect{} +} + +func (p *heapAllocator) allocMapReflect() *mapReflect { + return &mapReflect{} +} + +func (p *heapAllocator) allocListReflect() *listReflect { + return &listReflect{} +} + +func (p *heapAllocator) allocListReflectRange() *listReflectRange { + return &listReflectRange{vr: &valueReflect{}} +} + +func (p *heapAllocator) Free(_ interface{}) {} + +// NewFreelistAllocator creates freelist based allocator. +// This allocator provides fast allocation and freeing of short lived value objects. +// +// The freelists are bounded in size by freelistMaxSize. If more than this amount of value objects is +// allocated at once, the excess will be returned to the heap for garbage collection when freed. +// +// This allocator is unsafe and must not be accessed concurrently by goroutines. +// +// This allocator works well for traversal of value data trees. Typical usage is to acquire +// a freelist at the beginning of the traversal and use it through out +// for all temporary value access. +func NewFreelistAllocator() Allocator { + return &freelistAllocator{ + valueUnstructured: &freelist{new: func() interface{} { + return &valueUnstructured{} + }}, + listUnstructuredRange: &freelist{new: func() interface{} { + return &listUnstructuredRange{vv: &valueUnstructured{}} + }}, + valueReflect: &freelist{new: func() interface{} { + return &valueReflect{} + }}, + mapReflect: &freelist{new: func() interface{} { + return &mapReflect{} + }}, + structReflect: &freelist{new: func() interface{} { + return &structReflect{} + }}, + listReflect: &freelist{new: func() interface{} { + return &listReflect{} + }}, + listReflectRange: &freelist{new: func() interface{} { + return &listReflectRange{vr: &valueReflect{}} + }}, + } +} + +// Bound memory usage of freelists. This prevents the processing of very large lists from leaking memory. +// This limit is large enough for endpoints objects containing 1000 IP address entries. Freed objects +// that don't fit into the freelist are orphaned on the heap to be garbage collected. +const freelistMaxSize = 1000 + +type freelistAllocator struct { + valueUnstructured *freelist + listUnstructuredRange *freelist + valueReflect *freelist + mapReflect *freelist + structReflect *freelist + listReflect *freelist + listReflectRange *freelist +} + +type freelist struct { + list []interface{} + new func() interface{} +} + +func (f *freelist) allocate() interface{} { + var w2 interface{} + if n := len(f.list); n > 0 { + w2, f.list = f.list[n-1], f.list[:n-1] + } else { + w2 = f.new() + } + return w2 +} + +func (f *freelist) free(v interface{}) { + if len(f.list) < freelistMaxSize { + f.list = append(f.list, v) + } +} + +func (w *freelistAllocator) Free(value interface{}) { + switch v := value.(type) { + case *valueUnstructured: + v.Value = nil // don't hold references to unstructured objects + w.valueUnstructured.free(v) + case *listUnstructuredRange: + v.vv.Value = nil // don't hold references to unstructured objects + w.listUnstructuredRange.free(v) + case *valueReflect: + v.ParentMapKey = nil + v.ParentMap = nil + w.valueReflect.free(v) + case *mapReflect: + w.mapReflect.free(v) + case *structReflect: + w.structReflect.free(v) + case *listReflect: + w.listReflect.free(v) + case *listReflectRange: + v.vr.ParentMapKey = nil + v.vr.ParentMap = nil + w.listReflectRange.free(v) + } +} + +func (w *freelistAllocator) allocValueUnstructured() *valueUnstructured { + return w.valueUnstructured.allocate().(*valueUnstructured) +} + +func (w *freelistAllocator) allocListUnstructuredRange() *listUnstructuredRange { + return w.listUnstructuredRange.allocate().(*listUnstructuredRange) +} + +func (w *freelistAllocator) allocValueReflect() *valueReflect { + return w.valueReflect.allocate().(*valueReflect) +} + +func (w *freelistAllocator) allocStructReflect() *structReflect { + return w.structReflect.allocate().(*structReflect) +} + +func (w *freelistAllocator) allocMapReflect() *mapReflect { + return w.mapReflect.allocate().(*mapReflect) +} + +func (w *freelistAllocator) allocListReflect() *listReflect { + return w.listReflect.allocate().(*listReflect) +} + +func (w *freelistAllocator) allocListReflectRange() *listReflectRange { + return w.listReflectRange.allocate().(*listReflectRange) +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/doc.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/doc.go new file mode 100644 index 00000000..84d7f0f3 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/doc.go @@ -0,0 +1,21 @@ +/* +Copyright 2018 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +// Package value defines types for an in-memory representation of yaml or json +// objects, organized for convenient comparison with a schema (as defined by +// the sibling schema package). Functions for reading and writing the objects +// are also provided. +package value diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/fields.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/fields.go new file mode 100644 index 00000000..be3c6724 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/fields.go @@ -0,0 +1,97 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "sort" + "strings" +) + +// Field is an individual key-value pair. +type Field struct { + Name string + Value Value +} + +// FieldList is a list of key-value pairs. Each field is expected to +// have a different name. +type FieldList []Field + +// Sort sorts the field list by Name. +func (f FieldList) Sort() { + if len(f) < 2 { + return + } + if len(f) == 2 { + if f[1].Name < f[0].Name { + f[0], f[1] = f[1], f[0] + } + return + } + sort.SliceStable(f, func(i, j int) bool { + return f[i].Name < f[j].Name + }) +} + +// Less compares two lists lexically. +func (f FieldList) Less(rhs FieldList) bool { + return f.Compare(rhs) == -1 +} + +// Compare compares two lists lexically. The result will be 0 if f==rhs, -1 +// if f < rhs, and +1 if f > rhs. +func (f FieldList) Compare(rhs FieldList) int { + i := 0 + for { + if i >= len(f) && i >= len(rhs) { + // Maps are the same length and all items are equal. + return 0 + } + if i >= len(f) { + // F is shorter. + return -1 + } + if i >= len(rhs) { + // RHS is shorter. + return 1 + } + if c := strings.Compare(f[i].Name, rhs[i].Name); c != 0 { + return c + } + if c := Compare(f[i].Value, rhs[i].Value); c != 0 { + return c + } + // The items are equal; continue. + i++ + } +} + +// Equals returns true if the two fieldslist are equals, false otherwise. +func (f FieldList) Equals(rhs FieldList) bool { + if len(f) != len(rhs) { + return false + } + for i := range f { + if f[i].Name != rhs[i].Name { + return false + } + if !Equals(f[i].Value, rhs[i].Value) { + return false + } + } + return true +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/jsontagutil.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/jsontagutil.go new file mode 100644 index 00000000..d4adb8fc --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/jsontagutil.go @@ -0,0 +1,91 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "fmt" + "reflect" + "strings" +) + +// TODO: This implements the same functionality as https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/runtime/converter.go#L236 +// but is based on the highly efficient approach from https://golang.org/src/encoding/json/encode.go + +func lookupJsonTags(f reflect.StructField) (name string, omit bool, inline bool, omitempty bool) { + tag := f.Tag.Get("json") + if tag == "-" { + return "", true, false, false + } + name, opts := parseTag(tag) + if name == "" { + name = f.Name + } + return name, false, opts.Contains("inline"), opts.Contains("omitempty") +} + +func isZero(v reflect.Value) bool { + switch v.Kind() { + case reflect.Array, reflect.Map, reflect.Slice, reflect.String: + return v.Len() == 0 + case reflect.Bool: + return !v.Bool() + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + return v.Int() == 0 + case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: + return v.Uint() == 0 + case reflect.Float32, reflect.Float64: + return v.Float() == 0 + case reflect.Interface, reflect.Ptr: + return v.IsNil() + case reflect.Chan, reflect.Func: + panic(fmt.Sprintf("unsupported type: %v", v.Type())) + } + return false +} + +type tagOptions string + +// parseTag splits a struct field's json tag into its name and +// comma-separated options. +func parseTag(tag string) (string, tagOptions) { + if idx := strings.Index(tag, ","); idx != -1 { + return tag[:idx], tagOptions(tag[idx+1:]) + } + return tag, tagOptions("") +} + +// Contains reports whether a comma-separated list of options +// contains a particular substr flag. substr must be surrounded by a +// string boundary or commas. +func (o tagOptions) Contains(optionName string) bool { + if len(o) == 0 { + return false + } + s := string(o) + for s != "" { + var next string + i := strings.Index(s, ",") + if i >= 0 { + s, next = s[:i], s[i+1:] + } + if s == optionName { + return true + } + s = next + } + return false +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/list.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/list.go new file mode 100644 index 00000000..0748f18e --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/list.go @@ -0,0 +1,139 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +// List represents a list object. +type List interface { + // Length returns how many items can be found in the map. + Length() int + // At returns the item at the given position in the map. It will + // panic if the index is out of range. + At(int) Value + // AtUsing uses the provided allocator and returns the item at the given + // position in the map. It will panic if the index is out of range. + // The returned Value should be given back to the Allocator when no longer needed + // by calling Allocator.Free(Value). + AtUsing(Allocator, int) Value + // Range returns a ListRange for iterating over the items in the list. + Range() ListRange + // RangeUsing uses the provided allocator and returns a ListRange for + // iterating over the items in the list. + // The returned Range should be given back to the Allocator when no longer needed + // by calling Allocator.Free(Value). + RangeUsing(Allocator) ListRange + // Equals compares the two lists, and return true if they are the same, false otherwise. + // Implementations can use ListEquals as a general implementation for this methods. + Equals(List) bool + // EqualsUsing uses the provided allocator and compares the two lists, and return true if + // they are the same, false otherwise. Implementations can use ListEqualsUsing as a general + // implementation for this methods. + EqualsUsing(Allocator, List) bool +} + +// ListRange represents a single iteration across the items of a list. +type ListRange interface { + // Next increments to the next item in the range, if there is one, and returns true, or returns false if there are no more items. + Next() bool + // Item returns the index and value of the current item in the range. or panics if there is no current item. + // For efficiency, Item may reuse the values returned by previous Item calls. Callers should be careful avoid holding + // pointers to the value returned by Item() that escape the iteration loop since they become invalid once either + // Item() or Allocator.Free() is called. + Item() (index int, value Value) +} + +var EmptyRange = &emptyRange{} + +type emptyRange struct{} + +func (_ *emptyRange) Next() bool { + return false +} + +func (_ *emptyRange) Item() (index int, value Value) { + panic("Item called on empty ListRange") +} + +// ListEquals compares two lists lexically. +// WARN: This is a naive implementation, calling lhs.Equals(rhs) is typically the most efficient. +func ListEquals(lhs, rhs List) bool { + return ListEqualsUsing(HeapAllocator, lhs, rhs) +} + +// ListEqualsUsing uses the provided allocator and compares two lists lexically. +// WARN: This is a naive implementation, calling lhs.EqualsUsing(allocator, rhs) is typically the most efficient. +func ListEqualsUsing(a Allocator, lhs, rhs List) bool { + if lhs.Length() != rhs.Length() { + return false + } + + lhsRange := lhs.RangeUsing(a) + defer a.Free(lhsRange) + rhsRange := rhs.RangeUsing(a) + defer a.Free(rhsRange) + + for lhsRange.Next() && rhsRange.Next() { + _, lv := lhsRange.Item() + _, rv := rhsRange.Item() + if !EqualsUsing(a, lv, rv) { + return false + } + } + return true +} + +// ListLess compares two lists lexically. +func ListLess(lhs, rhs List) bool { + return ListCompare(lhs, rhs) == -1 +} + +// ListCompare compares two lists lexically. The result will be 0 if l==rhs, -1 +// if l < rhs, and +1 if l > rhs. +func ListCompare(lhs, rhs List) int { + return ListCompareUsing(HeapAllocator, lhs, rhs) +} + +// ListCompareUsing uses the provided allocator and compares two lists lexically. The result will be 0 if l==rhs, -1 +// if l < rhs, and +1 if l > rhs. +func ListCompareUsing(a Allocator, lhs, rhs List) int { + lhsRange := lhs.RangeUsing(a) + defer a.Free(lhsRange) + rhsRange := rhs.RangeUsing(a) + defer a.Free(rhsRange) + + for { + lhsOk := lhsRange.Next() + rhsOk := rhsRange.Next() + if !lhsOk && !rhsOk { + // Lists are the same length and all items are equal. + return 0 + } + if !lhsOk { + // LHS is shorter. + return -1 + } + if !rhsOk { + // RHS is shorter. + return 1 + } + _, lv := lhsRange.Item() + _, rv := rhsRange.Item() + if c := CompareUsing(a, lv, rv); c != 0 { + return c + } + // The items are equal; continue. + } +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/listreflect.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/listreflect.go new file mode 100644 index 00000000..197d4c92 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/listreflect.go @@ -0,0 +1,98 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "reflect" +) + +type listReflect struct { + Value reflect.Value +} + +func (r listReflect) Length() int { + val := r.Value + return val.Len() +} + +func (r listReflect) At(i int) Value { + val := r.Value + return mustWrapValueReflect(val.Index(i), nil, nil) +} + +func (r listReflect) AtUsing(a Allocator, i int) Value { + val := r.Value + return a.allocValueReflect().mustReuse(val.Index(i), nil, nil, nil) +} + +func (r listReflect) Unstructured() interface{} { + l := r.Length() + result := make([]interface{}, l) + for i := 0; i < l; i++ { + result[i] = r.At(i).Unstructured() + } + return result +} + +func (r listReflect) Range() ListRange { + return r.RangeUsing(HeapAllocator) +} + +func (r listReflect) RangeUsing(a Allocator) ListRange { + length := r.Value.Len() + if length == 0 { + return EmptyRange + } + rr := a.allocListReflectRange() + rr.list = r.Value + rr.i = -1 + rr.entry = TypeReflectEntryOf(r.Value.Type().Elem()) + return rr +} + +func (r listReflect) Equals(other List) bool { + return r.EqualsUsing(HeapAllocator, other) +} +func (r listReflect) EqualsUsing(a Allocator, other List) bool { + if otherReflectList, ok := other.(*listReflect); ok { + return reflect.DeepEqual(r.Value.Interface(), otherReflectList.Value.Interface()) + } + return ListEqualsUsing(a, &r, other) +} + +type listReflectRange struct { + list reflect.Value + vr *valueReflect + i int + entry *TypeReflectCacheEntry +} + +func (r *listReflectRange) Next() bool { + r.i += 1 + return r.i < r.list.Len() +} + +func (r *listReflectRange) Item() (index int, value Value) { + if r.i < 0 { + panic("Item() called before first calling Next()") + } + if r.i >= r.list.Len() { + panic("Item() called on ListRange with no more items") + } + v := r.list.Index(r.i) + return r.i, r.vr.mustReuse(v, r.entry, nil, nil) +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/listunstructured.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/listunstructured.go new file mode 100644 index 00000000..64cd8e7c --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/listunstructured.go @@ -0,0 +1,74 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +type listUnstructured []interface{} + +func (l listUnstructured) Length() int { + return len(l) +} + +func (l listUnstructured) At(i int) Value { + return NewValueInterface(l[i]) +} + +func (l listUnstructured) AtUsing(a Allocator, i int) Value { + return a.allocValueUnstructured().reuse(l[i]) +} + +func (l listUnstructured) Equals(other List) bool { + return l.EqualsUsing(HeapAllocator, other) +} + +func (l listUnstructured) EqualsUsing(a Allocator, other List) bool { + return ListEqualsUsing(a, &l, other) +} + +func (l listUnstructured) Range() ListRange { + return l.RangeUsing(HeapAllocator) +} + +func (l listUnstructured) RangeUsing(a Allocator) ListRange { + if len(l) == 0 { + return EmptyRange + } + r := a.allocListUnstructuredRange() + r.list = l + r.i = -1 + return r +} + +type listUnstructuredRange struct { + list listUnstructured + vv *valueUnstructured + i int +} + +func (r *listUnstructuredRange) Next() bool { + r.i += 1 + return r.i < len(r.list) +} + +func (r *listUnstructuredRange) Item() (index int, value Value) { + if r.i < 0 { + panic("Item() called before first calling Next()") + } + if r.i >= len(r.list) { + panic("Item() called on ListRange with no more items") + } + return r.i, r.vv.reuse(r.list[r.i]) +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/map.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/map.go new file mode 100644 index 00000000..168b9fa0 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/map.go @@ -0,0 +1,270 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "sort" +) + +// Map represents a Map or go structure. +type Map interface { + // Set changes or set the value of the given key. + Set(key string, val Value) + // Get returns the value for the given key, if present, or (nil, false) otherwise. + Get(key string) (Value, bool) + // GetUsing uses the provided allocator and returns the value for the given key, + // if present, or (nil, false) otherwise. + // The returned Value should be given back to the Allocator when no longer needed + // by calling Allocator.Free(Value). + GetUsing(a Allocator, key string) (Value, bool) + // Has returns true if the key is present, or false otherwise. + Has(key string) bool + // Delete removes the key from the map. + Delete(key string) + // Equals compares the two maps, and return true if they are the same, false otherwise. + // Implementations can use MapEquals as a general implementation for this methods. + Equals(other Map) bool + // EqualsUsing uses the provided allocator and compares the two maps, and return true if + // they are the same, false otherwise. Implementations can use MapEqualsUsing as a general + // implementation for this methods. + EqualsUsing(a Allocator, other Map) bool + // Iterate runs the given function for each key/value in the + // map. Returning false in the closure prematurely stops the + // iteration. + Iterate(func(key string, value Value) bool) bool + // IterateUsing uses the provided allocator and runs the given function for each key/value + // in the map. Returning false in the closure prematurely stops the iteration. + IterateUsing(Allocator, func(key string, value Value) bool) bool + // Length returns the number of items in the map. + Length() int + // Empty returns true if the map is empty. + Empty() bool + // Zip iterates over the entries of two maps together. If both maps contain a value for a given key, fn is called + // with the values from both maps, otherwise it is called with the value of the map that contains the key and nil + // for the map that does not contain the key. Returning false in the closure prematurely stops the iteration. + Zip(other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool + // ZipUsing uses the provided allocator and iterates over the entries of two maps together. If both maps + // contain a value for a given key, fn is called with the values from both maps, otherwise it is called with + // the value of the map that contains the key and nil for the map that does not contain the key. Returning + // false in the closure prematurely stops the iteration. + ZipUsing(a Allocator, other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool +} + +// MapTraverseOrder defines the map traversal ordering available. +type MapTraverseOrder int + +const ( + // Unordered indicates that the map traversal has no ordering requirement. + Unordered = iota + // LexicalKeyOrder indicates that the map traversal is ordered by key, lexically. + LexicalKeyOrder +) + +// MapZip iterates over the entries of two maps together. If both maps contain a value for a given key, fn is called +// with the values from both maps, otherwise it is called with the value of the map that contains the key and nil +// for the other map. Returning false in the closure prematurely stops the iteration. +func MapZip(lhs, rhs Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + return MapZipUsing(HeapAllocator, lhs, rhs, order, fn) +} + +// MapZipUsing uses the provided allocator and iterates over the entries of two maps together. If both maps +// contain a value for a given key, fn is called with the values from both maps, otherwise it is called with +// the value of the map that contains the key and nil for the other map. Returning false in the closure +// prematurely stops the iteration. +func MapZipUsing(a Allocator, lhs, rhs Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + if lhs != nil { + return lhs.ZipUsing(a, rhs, order, fn) + } + if rhs != nil { + return rhs.ZipUsing(a, lhs, order, func(key string, rhs, lhs Value) bool { // arg positions of lhs and rhs deliberately swapped + return fn(key, lhs, rhs) + }) + } + return true +} + +// defaultMapZip provides a default implementation of Zip for implementations that do not need to provide +// their own optimized implementation. +func defaultMapZip(a Allocator, lhs, rhs Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + switch order { + case Unordered: + return unorderedMapZip(a, lhs, rhs, fn) + case LexicalKeyOrder: + return lexicalKeyOrderedMapZip(a, lhs, rhs, fn) + default: + panic("Unsupported map order") + } +} + +func unorderedMapZip(a Allocator, lhs, rhs Map, fn func(key string, lhs, rhs Value) bool) bool { + if (lhs == nil || lhs.Empty()) && (rhs == nil || rhs.Empty()) { + return true + } + + if lhs != nil { + ok := lhs.IterateUsing(a, func(key string, lhsValue Value) bool { + var rhsValue Value + if rhs != nil { + if item, ok := rhs.GetUsing(a, key); ok { + rhsValue = item + defer a.Free(rhsValue) + } + } + return fn(key, lhsValue, rhsValue) + }) + if !ok { + return false + } + } + if rhs != nil { + return rhs.IterateUsing(a, func(key string, rhsValue Value) bool { + if lhs == nil || !lhs.Has(key) { + return fn(key, nil, rhsValue) + } + return true + }) + } + return true +} + +func lexicalKeyOrderedMapZip(a Allocator, lhs, rhs Map, fn func(key string, lhs, rhs Value) bool) bool { + var lhsLength, rhsLength int + var orderedLength int // rough estimate of length of union of map keys + if lhs != nil { + lhsLength = lhs.Length() + orderedLength = lhsLength + } + if rhs != nil { + rhsLength = rhs.Length() + if rhsLength > orderedLength { + orderedLength = rhsLength + } + } + if lhsLength == 0 && rhsLength == 0 { + return true + } + + ordered := make([]string, 0, orderedLength) + if lhs != nil { + lhs.IterateUsing(a, func(key string, _ Value) bool { + ordered = append(ordered, key) + return true + }) + } + if rhs != nil { + rhs.IterateUsing(a, func(key string, _ Value) bool { + if lhs == nil || !lhs.Has(key) { + ordered = append(ordered, key) + } + return true + }) + } + sort.Strings(ordered) + for _, key := range ordered { + var litem, ritem Value + if lhs != nil { + litem, _ = lhs.GetUsing(a, key) + } + if rhs != nil { + ritem, _ = rhs.GetUsing(a, key) + } + ok := fn(key, litem, ritem) + if litem != nil { + a.Free(litem) + } + if ritem != nil { + a.Free(ritem) + } + if !ok { + return false + } + } + return true +} + +// MapLess compares two maps lexically. +func MapLess(lhs, rhs Map) bool { + return MapCompare(lhs, rhs) == -1 +} + +// MapCompare compares two maps lexically. +func MapCompare(lhs, rhs Map) int { + return MapCompareUsing(HeapAllocator, lhs, rhs) +} + +// MapCompareUsing uses the provided allocator and compares two maps lexically. +func MapCompareUsing(a Allocator, lhs, rhs Map) int { + c := 0 + var llength, rlength int + if lhs != nil { + llength = lhs.Length() + } + if rhs != nil { + rlength = rhs.Length() + } + if llength == 0 && rlength == 0 { + return 0 + } + i := 0 + MapZipUsing(a, lhs, rhs, LexicalKeyOrder, func(key string, lhs, rhs Value) bool { + switch { + case i == llength: + c = -1 + case i == rlength: + c = 1 + case lhs == nil: + c = 1 + case rhs == nil: + c = -1 + default: + c = CompareUsing(a, lhs, rhs) + } + i++ + return c == 0 + }) + return c +} + +// MapEquals returns true if lhs == rhs, false otherwise. This function +// acts on generic types and should not be used by callers, but can help +// implement Map.Equals. +// WARN: This is a naive implementation, calling lhs.Equals(rhs) is typically the most efficient. +func MapEquals(lhs, rhs Map) bool { + return MapEqualsUsing(HeapAllocator, lhs, rhs) +} + +// MapEqualsUsing uses the provided allocator and returns true if lhs == rhs, +// false otherwise. This function acts on generic types and should not be used +// by callers, but can help implement Map.Equals. +// WARN: This is a naive implementation, calling lhs.EqualsUsing(allocator, rhs) is typically the most efficient. +func MapEqualsUsing(a Allocator, lhs, rhs Map) bool { + if lhs == nil && rhs == nil { + return true + } + if lhs == nil || rhs == nil { + return false + } + if lhs.Length() != rhs.Length() { + return false + } + return MapZipUsing(a, lhs, rhs, Unordered, func(key string, lhs, rhs Value) bool { + if lhs == nil || rhs == nil { + return false + } + return EqualsUsing(a, lhs, rhs) + }) +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapreflect.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapreflect.go new file mode 100644 index 00000000..dc8b8c72 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapreflect.go @@ -0,0 +1,209 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "reflect" +) + +type mapReflect struct { + valueReflect +} + +func (r mapReflect) Length() int { + val := r.Value + return val.Len() +} + +func (r mapReflect) Empty() bool { + val := r.Value + return val.Len() == 0 +} + +func (r mapReflect) Get(key string) (Value, bool) { + return r.GetUsing(HeapAllocator, key) +} + +func (r mapReflect) GetUsing(a Allocator, key string) (Value, bool) { + k, v, ok := r.get(key) + if !ok { + return nil, false + } + return a.allocValueReflect().mustReuse(v, nil, &r.Value, &k), true +} + +func (r mapReflect) get(k string) (key, value reflect.Value, ok bool) { + mapKey := r.toMapKey(k) + val := r.Value.MapIndex(mapKey) + return mapKey, val, val.IsValid() && val != reflect.Value{} +} + +func (r mapReflect) Has(key string) bool { + var val reflect.Value + val = r.Value.MapIndex(r.toMapKey(key)) + if !val.IsValid() { + return false + } + return val != reflect.Value{} +} + +func (r mapReflect) Set(key string, val Value) { + r.Value.SetMapIndex(r.toMapKey(key), reflect.ValueOf(val.Unstructured())) +} + +func (r mapReflect) Delete(key string) { + val := r.Value + val.SetMapIndex(r.toMapKey(key), reflect.Value{}) +} + +// TODO: Do we need to support types that implement json.Marshaler and are used as string keys? +func (r mapReflect) toMapKey(key string) reflect.Value { + val := r.Value + return reflect.ValueOf(key).Convert(val.Type().Key()) +} + +func (r mapReflect) Iterate(fn func(string, Value) bool) bool { + return r.IterateUsing(HeapAllocator, fn) +} + +func (r mapReflect) IterateUsing(a Allocator, fn func(string, Value) bool) bool { + if r.Value.Len() == 0 { + return true + } + v := a.allocValueReflect() + defer a.Free(v) + return eachMapEntry(r.Value, func(e *TypeReflectCacheEntry, key reflect.Value, value reflect.Value) bool { + return fn(key.String(), v.mustReuse(value, e, &r.Value, &key)) + }) +} + +func eachMapEntry(val reflect.Value, fn func(*TypeReflectCacheEntry, reflect.Value, reflect.Value) bool) bool { + iter := val.MapRange() + entry := TypeReflectEntryOf(val.Type().Elem()) + for iter.Next() { + next := iter.Value() + if !next.IsValid() { + continue + } + if !fn(entry, iter.Key(), next) { + return false + } + } + return true +} + +func (r mapReflect) Unstructured() interface{} { + result := make(map[string]interface{}, r.Length()) + r.Iterate(func(s string, value Value) bool { + result[s] = value.Unstructured() + return true + }) + return result +} + +func (r mapReflect) Equals(m Map) bool { + return r.EqualsUsing(HeapAllocator, m) +} + +func (r mapReflect) EqualsUsing(a Allocator, m Map) bool { + lhsLength := r.Length() + rhsLength := m.Length() + if lhsLength != rhsLength { + return false + } + if lhsLength == 0 { + return true + } + vr := a.allocValueReflect() + defer a.Free(vr) + entry := TypeReflectEntryOf(r.Value.Type().Elem()) + return m.Iterate(func(key string, value Value) bool { + _, lhsVal, ok := r.get(key) + if !ok { + return false + } + return Equals(vr.mustReuse(lhsVal, entry, nil, nil), value) + }) +} + +func (r mapReflect) Zip(other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + return r.ZipUsing(HeapAllocator, other, order, fn) +} + +func (r mapReflect) ZipUsing(a Allocator, other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + if otherMapReflect, ok := other.(*mapReflect); ok && order == Unordered { + return r.unorderedReflectZip(a, otherMapReflect, fn) + } + return defaultMapZip(a, &r, other, order, fn) +} + +// unorderedReflectZip provides an optimized unordered zip for mapReflect types. +func (r mapReflect) unorderedReflectZip(a Allocator, other *mapReflect, fn func(key string, lhs, rhs Value) bool) bool { + if r.Empty() && (other == nil || other.Empty()) { + return true + } + + lhs := r.Value + lhsEntry := TypeReflectEntryOf(lhs.Type().Elem()) + + // map lookup via reflection is expensive enough that it is better to keep track of visited keys + visited := map[string]struct{}{} + + vlhs, vrhs := a.allocValueReflect(), a.allocValueReflect() + defer a.Free(vlhs) + defer a.Free(vrhs) + + if other != nil { + rhs := other.Value + rhsEntry := TypeReflectEntryOf(rhs.Type().Elem()) + iter := rhs.MapRange() + + for iter.Next() { + key := iter.Key() + keyString := key.String() + next := iter.Value() + if !next.IsValid() { + continue + } + rhsVal := vrhs.mustReuse(next, rhsEntry, &rhs, &key) + visited[keyString] = struct{}{} + var lhsVal Value + if _, v, ok := r.get(keyString); ok { + lhsVal = vlhs.mustReuse(v, lhsEntry, &lhs, &key) + } + if !fn(keyString, lhsVal, rhsVal) { + return false + } + } + } + + iter := lhs.MapRange() + for iter.Next() { + key := iter.Key() + if _, ok := visited[key.String()]; ok { + continue + } + next := iter.Value() + if !next.IsValid() { + continue + } + if !fn(key.String(), vlhs.mustReuse(next, lhsEntry, &lhs, &key), nil) { + return false + } + } + return true +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapunstructured.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapunstructured.go new file mode 100644 index 00000000..d8e20862 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/mapunstructured.go @@ -0,0 +1,190 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +type mapUnstructuredInterface map[interface{}]interface{} + +func (m mapUnstructuredInterface) Set(key string, val Value) { + m[key] = val.Unstructured() +} + +func (m mapUnstructuredInterface) Get(key string) (Value, bool) { + return m.GetUsing(HeapAllocator, key) +} + +func (m mapUnstructuredInterface) GetUsing(a Allocator, key string) (Value, bool) { + if v, ok := m[key]; !ok { + return nil, false + } else { + return a.allocValueUnstructured().reuse(v), true + } +} + +func (m mapUnstructuredInterface) Has(key string) bool { + _, ok := m[key] + return ok +} + +func (m mapUnstructuredInterface) Delete(key string) { + delete(m, key) +} + +func (m mapUnstructuredInterface) Iterate(fn func(key string, value Value) bool) bool { + return m.IterateUsing(HeapAllocator, fn) +} + +func (m mapUnstructuredInterface) IterateUsing(a Allocator, fn func(key string, value Value) bool) bool { + if len(m) == 0 { + return true + } + vv := a.allocValueUnstructured() + defer a.Free(vv) + for k, v := range m { + if ks, ok := k.(string); !ok { + continue + } else { + if !fn(ks, vv.reuse(v)) { + return false + } + } + } + return true +} + +func (m mapUnstructuredInterface) Length() int { + return len(m) +} + +func (m mapUnstructuredInterface) Empty() bool { + return len(m) == 0 +} + +func (m mapUnstructuredInterface) Equals(other Map) bool { + return m.EqualsUsing(HeapAllocator, other) +} + +func (m mapUnstructuredInterface) EqualsUsing(a Allocator, other Map) bool { + lhsLength := m.Length() + rhsLength := other.Length() + if lhsLength != rhsLength { + return false + } + if lhsLength == 0 { + return true + } + vv := a.allocValueUnstructured() + defer a.Free(vv) + return other.Iterate(func(key string, value Value) bool { + lhsVal, ok := m[key] + if !ok { + return false + } + return Equals(vv.reuse(lhsVal), value) + }) +} + +func (m mapUnstructuredInterface) Zip(other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + return m.ZipUsing(HeapAllocator, other, order, fn) +} + +func (m mapUnstructuredInterface) ZipUsing(a Allocator, other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + return defaultMapZip(a, m, other, order, fn) +} + +type mapUnstructuredString map[string]interface{} + +func (m mapUnstructuredString) Set(key string, val Value) { + m[key] = val.Unstructured() +} + +func (m mapUnstructuredString) Get(key string) (Value, bool) { + return m.GetUsing(HeapAllocator, key) +} +func (m mapUnstructuredString) GetUsing(a Allocator, key string) (Value, bool) { + if v, ok := m[key]; !ok { + return nil, false + } else { + return a.allocValueUnstructured().reuse(v), true + } +} + +func (m mapUnstructuredString) Has(key string) bool { + _, ok := m[key] + return ok +} + +func (m mapUnstructuredString) Delete(key string) { + delete(m, key) +} + +func (m mapUnstructuredString) Iterate(fn func(key string, value Value) bool) bool { + return m.IterateUsing(HeapAllocator, fn) +} + +func (m mapUnstructuredString) IterateUsing(a Allocator, fn func(key string, value Value) bool) bool { + if len(m) == 0 { + return true + } + vv := a.allocValueUnstructured() + defer a.Free(vv) + for k, v := range m { + if !fn(k, vv.reuse(v)) { + return false + } + } + return true +} + +func (m mapUnstructuredString) Length() int { + return len(m) +} + +func (m mapUnstructuredString) Equals(other Map) bool { + return m.EqualsUsing(HeapAllocator, other) +} + +func (m mapUnstructuredString) EqualsUsing(a Allocator, other Map) bool { + lhsLength := m.Length() + rhsLength := other.Length() + if lhsLength != rhsLength { + return false + } + if lhsLength == 0 { + return true + } + vv := a.allocValueUnstructured() + defer a.Free(vv) + return other.Iterate(func(key string, value Value) bool { + lhsVal, ok := m[key] + if !ok { + return false + } + return Equals(vv.reuse(lhsVal), value) + }) +} + +func (m mapUnstructuredString) Zip(other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + return m.ZipUsing(HeapAllocator, other, order, fn) +} + +func (m mapUnstructuredString) ZipUsing(a Allocator, other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + return defaultMapZip(a, m, other, order, fn) +} + +func (m mapUnstructuredString) Empty() bool { + return len(m) == 0 +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/reflectcache.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/reflectcache.go new file mode 100644 index 00000000..49e6dd16 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/reflectcache.go @@ -0,0 +1,463 @@ +/* +Copyright 2020 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "bytes" + "encoding/json" + "fmt" + "reflect" + "sort" + "sync" + "sync/atomic" +) + +// UnstructuredConverter defines how a type can be converted directly to unstructured. +// Types that implement json.Marshaler may also optionally implement this interface to provide a more +// direct and more efficient conversion. All types that choose to implement this interface must still +// implement this same conversion via json.Marshaler. +type UnstructuredConverter interface { + json.Marshaler // require that json.Marshaler is implemented + + // ToUnstructured returns the unstructured representation. + ToUnstructured() interface{} +} + +// TypeReflectCacheEntry keeps data gathered using reflection about how a type is converted to/from unstructured. +type TypeReflectCacheEntry struct { + isJsonMarshaler bool + ptrIsJsonMarshaler bool + isJsonUnmarshaler bool + ptrIsJsonUnmarshaler bool + isStringConvertable bool + ptrIsStringConvertable bool + + structFields map[string]*FieldCacheEntry + orderedStructFields []*FieldCacheEntry +} + +// FieldCacheEntry keeps data gathered using reflection about how the field of a struct is converted to/from +// unstructured. +type FieldCacheEntry struct { + // JsonName returns the name of the field according to the json tags on the struct field. + JsonName string + // isOmitEmpty is true if the field has the json 'omitempty' tag. + isOmitEmpty bool + // fieldPath is a list of field indices (see FieldByIndex) to lookup the value of + // a field in a reflect.Value struct. The field indices in the list form a path used + // to traverse through intermediary 'inline' fields. + fieldPath [][]int + + fieldType reflect.Type + TypeEntry *TypeReflectCacheEntry +} + +func (f *FieldCacheEntry) CanOmit(fieldVal reflect.Value) bool { + return f.isOmitEmpty && (safeIsNil(fieldVal) || isZero(fieldVal)) +} + +// GetUsing returns the field identified by this FieldCacheEntry from the provided struct. +func (f *FieldCacheEntry) GetFrom(structVal reflect.Value) reflect.Value { + // field might be nested within 'inline' structs + for _, elem := range f.fieldPath { + structVal = structVal.FieldByIndex(elem) + } + return structVal +} + +var marshalerType = reflect.TypeOf(new(json.Marshaler)).Elem() +var unmarshalerType = reflect.TypeOf(new(json.Unmarshaler)).Elem() +var unstructuredConvertableType = reflect.TypeOf(new(UnstructuredConverter)).Elem() +var defaultReflectCache = newReflectCache() + +// TypeReflectEntryOf returns the TypeReflectCacheEntry of the provided reflect.Type. +func TypeReflectEntryOf(t reflect.Type) *TypeReflectCacheEntry { + cm := defaultReflectCache.get() + if record, ok := cm[t]; ok { + return record + } + updates := reflectCacheMap{} + result := typeReflectEntryOf(cm, t, updates) + if len(updates) > 0 { + defaultReflectCache.update(updates) + } + return result +} + +// TypeReflectEntryOf returns all updates needed to add provided reflect.Type, and the types its fields transitively +// depend on, to the cache. +func typeReflectEntryOf(cm reflectCacheMap, t reflect.Type, updates reflectCacheMap) *TypeReflectCacheEntry { + if record, ok := cm[t]; ok { + return record + } + if record, ok := updates[t]; ok { + return record + } + typeEntry := &TypeReflectCacheEntry{ + isJsonMarshaler: t.Implements(marshalerType), + ptrIsJsonMarshaler: reflect.PtrTo(t).Implements(marshalerType), + isJsonUnmarshaler: reflect.PtrTo(t).Implements(unmarshalerType), + isStringConvertable: t.Implements(unstructuredConvertableType), + ptrIsStringConvertable: reflect.PtrTo(t).Implements(unstructuredConvertableType), + } + if t.Kind() == reflect.Struct { + fieldEntries := map[string]*FieldCacheEntry{} + buildStructCacheEntry(t, fieldEntries, nil) + typeEntry.structFields = fieldEntries + sortedByJsonName := make([]*FieldCacheEntry, len(fieldEntries)) + i := 0 + for _, entry := range fieldEntries { + sortedByJsonName[i] = entry + i++ + } + sort.Slice(sortedByJsonName, func(i, j int) bool { + return sortedByJsonName[i].JsonName < sortedByJsonName[j].JsonName + }) + typeEntry.orderedStructFields = sortedByJsonName + } + + // cyclic type references are allowed, so we must add the typeEntry to the updates map before resolving + // the field.typeEntry references, or creating them if they are not already in the cache + updates[t] = typeEntry + + for _, field := range typeEntry.structFields { + if field.TypeEntry == nil { + field.TypeEntry = typeReflectEntryOf(cm, field.fieldType, updates) + } + } + return typeEntry +} + +func buildStructCacheEntry(t reflect.Type, infos map[string]*FieldCacheEntry, fieldPath [][]int) { + for i := 0; i < t.NumField(); i++ { + field := t.Field(i) + jsonName, omit, isInline, isOmitempty := lookupJsonTags(field) + if omit { + continue + } + if isInline { + buildStructCacheEntry(field.Type, infos, append(fieldPath, field.Index)) + continue + } + info := &FieldCacheEntry{JsonName: jsonName, isOmitEmpty: isOmitempty, fieldPath: append(fieldPath, field.Index), fieldType: field.Type} + infos[jsonName] = info + } +} + +// Fields returns a map of JSON field name to FieldCacheEntry for structs, or nil for non-structs. +func (e TypeReflectCacheEntry) Fields() map[string]*FieldCacheEntry { + return e.structFields +} + +// Fields returns a map of JSON field name to FieldCacheEntry for structs, or nil for non-structs. +func (e TypeReflectCacheEntry) OrderedFields() []*FieldCacheEntry { + return e.orderedStructFields +} + +// CanConvertToUnstructured returns true if this TypeReflectCacheEntry can convert values of its type to unstructured. +func (e TypeReflectCacheEntry) CanConvertToUnstructured() bool { + return e.isJsonMarshaler || e.ptrIsJsonMarshaler || e.isStringConvertable || e.ptrIsStringConvertable +} + +// ToUnstructured converts the provided value to unstructured and returns it. +func (e TypeReflectCacheEntry) ToUnstructured(sv reflect.Value) (interface{}, error) { + // This is based on https://github.com/kubernetes/kubernetes/blob/82c9e5c814eb7acc6cc0a090c057294d0667ad66/staging/src/k8s.io/apimachinery/pkg/runtime/converter.go#L505 + // and is intended to replace it. + + // Check if the object has a custom string converter and use it if available, since it is much more efficient + // than round tripping through json. + if converter, ok := e.getUnstructuredConverter(sv); ok { + return converter.ToUnstructured(), nil + } + // Check if the object has a custom JSON marshaller/unmarshaller. + if marshaler, ok := e.getJsonMarshaler(sv); ok { + if sv.Kind() == reflect.Ptr && sv.IsNil() { + // We're done - we don't need to store anything. + return nil, nil + } + + data, err := marshaler.MarshalJSON() + if err != nil { + return nil, err + } + switch { + case len(data) == 0: + return nil, fmt.Errorf("error decoding from json: empty value") + + case bytes.Equal(data, nullBytes): + // We're done - we don't need to store anything. + return nil, nil + + case bytes.Equal(data, trueBytes): + return true, nil + + case bytes.Equal(data, falseBytes): + return false, nil + + case data[0] == '"': + var result string + err := unmarshal(data, &result) + if err != nil { + return nil, fmt.Errorf("error decoding string from json: %v", err) + } + return result, nil + + case data[0] == '{': + result := make(map[string]interface{}) + err := unmarshal(data, &result) + if err != nil { + return nil, fmt.Errorf("error decoding object from json: %v", err) + } + return result, nil + + case data[0] == '[': + result := make([]interface{}, 0) + err := unmarshal(data, &result) + if err != nil { + return nil, fmt.Errorf("error decoding array from json: %v", err) + } + return result, nil + + default: + var ( + resultInt int64 + resultFloat float64 + err error + ) + if err = unmarshal(data, &resultInt); err == nil { + return resultInt, nil + } else if err = unmarshal(data, &resultFloat); err == nil { + return resultFloat, nil + } else { + return nil, fmt.Errorf("error decoding number from json: %v", err) + } + } + } + + return nil, fmt.Errorf("provided type cannot be converted: %v", sv.Type()) +} + +// CanConvertFromUnstructured returns true if this TypeReflectCacheEntry can convert objects of the type from unstructured. +func (e TypeReflectCacheEntry) CanConvertFromUnstructured() bool { + return e.isJsonUnmarshaler +} + +// FromUnstructured converts the provided source value from unstructured into the provided destination value. +func (e TypeReflectCacheEntry) FromUnstructured(sv, dv reflect.Value) error { + // TODO: this could be made much more efficient using direct conversions like + // UnstructuredConverter.ToUnstructured provides. + st := dv.Type() + data, err := json.Marshal(sv.Interface()) + if err != nil { + return fmt.Errorf("error encoding %s to json: %v", st.String(), err) + } + if unmarshaler, ok := e.getJsonUnmarshaler(dv); ok { + return unmarshaler.UnmarshalJSON(data) + } + return fmt.Errorf("unable to unmarshal %v into %v", sv.Type(), dv.Type()) +} + +var ( + nullBytes = []byte("null") + trueBytes = []byte("true") + falseBytes = []byte("false") +) + +func (e TypeReflectCacheEntry) getJsonMarshaler(v reflect.Value) (json.Marshaler, bool) { + if e.isJsonMarshaler { + return v.Interface().(json.Marshaler), true + } + if e.ptrIsJsonMarshaler { + // Check pointer receivers if v is not a pointer + if v.Kind() != reflect.Ptr && v.CanAddr() { + v = v.Addr() + return v.Interface().(json.Marshaler), true + } + } + return nil, false +} + +func (e TypeReflectCacheEntry) getJsonUnmarshaler(v reflect.Value) (json.Unmarshaler, bool) { + if !e.isJsonUnmarshaler { + return nil, false + } + return v.Addr().Interface().(json.Unmarshaler), true +} + +func (e TypeReflectCacheEntry) getUnstructuredConverter(v reflect.Value) (UnstructuredConverter, bool) { + if e.isStringConvertable { + return v.Interface().(UnstructuredConverter), true + } + if e.ptrIsStringConvertable { + // Check pointer receivers if v is not a pointer + if v.CanAddr() { + v = v.Addr() + return v.Interface().(UnstructuredConverter), true + } + } + return nil, false +} + +type typeReflectCache struct { + // use an atomic and copy-on-write since there are a fixed (typically very small) number of structs compiled into any + // go program using this cache + value atomic.Value + // mu is held by writers when performing load/modify/store operations on the cache, readers do not need to hold a + // read-lock since the atomic value is always read-only + mu sync.Mutex +} + +func newReflectCache() *typeReflectCache { + cache := &typeReflectCache{} + cache.value.Store(make(reflectCacheMap)) + return cache +} + +type reflectCacheMap map[reflect.Type]*TypeReflectCacheEntry + +// get returns the reflectCacheMap. +func (c *typeReflectCache) get() reflectCacheMap { + return c.value.Load().(reflectCacheMap) +} + +// update merges the provided updates into the cache. +func (c *typeReflectCache) update(updates reflectCacheMap) { + c.mu.Lock() + defer c.mu.Unlock() + + currentCacheMap := c.value.Load().(reflectCacheMap) + + hasNewEntries := false + for t := range updates { + if _, ok := currentCacheMap[t]; !ok { + hasNewEntries = true + break + } + } + if !hasNewEntries { + // Bail if the updates have been set while waiting for lock acquisition. + // This is safe since setting entries is idempotent. + return + } + + newCacheMap := make(reflectCacheMap, len(currentCacheMap)+len(updates)) + for k, v := range currentCacheMap { + newCacheMap[k] = v + } + for t, update := range updates { + newCacheMap[t] = update + } + c.value.Store(newCacheMap) +} + +// Below json Unmarshal is fromk8s.io/apimachinery/pkg/util/json +// to handle number conversions as expected by Kubernetes + +// limit recursive depth to prevent stack overflow errors +const maxDepth = 10000 + +// unmarshal unmarshals the given data +// If v is a *map[string]interface{}, numbers are converted to int64 or float64 +func unmarshal(data []byte, v interface{}) error { + switch v := v.(type) { + case *map[string]interface{}: + // Build a decoder from the given data + decoder := json.NewDecoder(bytes.NewBuffer(data)) + // Preserve numbers, rather than casting to float64 automatically + decoder.UseNumber() + // Run the decode + if err := decoder.Decode(v); err != nil { + return err + } + // If the decode succeeds, post-process the map to convert json.Number objects to int64 or float64 + return convertMapNumbers(*v, 0) + + case *[]interface{}: + // Build a decoder from the given data + decoder := json.NewDecoder(bytes.NewBuffer(data)) + // Preserve numbers, rather than casting to float64 automatically + decoder.UseNumber() + // Run the decode + if err := decoder.Decode(v); err != nil { + return err + } + // If the decode succeeds, post-process the map to convert json.Number objects to int64 or float64 + return convertSliceNumbers(*v, 0) + + default: + return json.Unmarshal(data, v) + } +} + +// convertMapNumbers traverses the map, converting any json.Number values to int64 or float64. +// values which are map[string]interface{} or []interface{} are recursively visited +func convertMapNumbers(m map[string]interface{}, depth int) error { + if depth > maxDepth { + return fmt.Errorf("exceeded max depth of %d", maxDepth) + } + + var err error + for k, v := range m { + switch v := v.(type) { + case json.Number: + m[k], err = convertNumber(v) + case map[string]interface{}: + err = convertMapNumbers(v, depth+1) + case []interface{}: + err = convertSliceNumbers(v, depth+1) + } + if err != nil { + return err + } + } + return nil +} + +// convertSliceNumbers traverses the slice, converting any json.Number values to int64 or float64. +// values which are map[string]interface{} or []interface{} are recursively visited +func convertSliceNumbers(s []interface{}, depth int) error { + if depth > maxDepth { + return fmt.Errorf("exceeded max depth of %d", maxDepth) + } + + var err error + for i, v := range s { + switch v := v.(type) { + case json.Number: + s[i], err = convertNumber(v) + case map[string]interface{}: + err = convertMapNumbers(v, depth+1) + case []interface{}: + err = convertSliceNumbers(v, depth+1) + } + if err != nil { + return err + } + } + return nil +} + +// convertNumber converts a json.Number to an int64 or float64, or returns an error +func convertNumber(n json.Number) (interface{}, error) { + // Attempt to convert to an int64 first + if i, err := n.Int64(); err == nil { + return i, nil + } + // Return a float64 (default json.Decode() behavior) + // An overflow will return an error + return n.Float64() +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/scalar.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/scalar.go new file mode 100644 index 00000000..c78a4c18 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/scalar.go @@ -0,0 +1,50 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +// Compare compares floats. The result will be 0 if lhs==rhs, -1 if f < +// rhs, and +1 if f > rhs. +func FloatCompare(lhs, rhs float64) int { + if lhs > rhs { + return 1 + } else if lhs < rhs { + return -1 + } + return 0 +} + +// IntCompare compares integers. The result will be 0 if i==rhs, -1 if i < +// rhs, and +1 if i > rhs. +func IntCompare(lhs, rhs int64) int { + if lhs > rhs { + return 1 + } else if lhs < rhs { + return -1 + } + return 0 +} + +// Compare compares booleans. The result will be 0 if b==rhs, -1 if b < +// rhs, and +1 if b > rhs. +func BoolCompare(lhs, rhs bool) int { + if lhs == rhs { + return 0 + } else if lhs == false { + return -1 + } + return 1 +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/structreflect.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/structreflect.go new file mode 100644 index 00000000..4a7bb5c6 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/structreflect.go @@ -0,0 +1,208 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "fmt" + "reflect" +) + +type structReflect struct { + valueReflect +} + +func (r structReflect) Length() int { + i := 0 + eachStructField(r.Value, func(_ *TypeReflectCacheEntry, s string, value reflect.Value) bool { + i++ + return true + }) + return i +} + +func (r structReflect) Empty() bool { + return eachStructField(r.Value, func(_ *TypeReflectCacheEntry, s string, value reflect.Value) bool { + return false // exit early if the struct is non-empty + }) +} + +func (r structReflect) Get(key string) (Value, bool) { + return r.GetUsing(HeapAllocator, key) +} + +func (r structReflect) GetUsing(a Allocator, key string) (Value, bool) { + if val, ok := r.findJsonNameField(key); ok { + return a.allocValueReflect().mustReuse(val, nil, nil, nil), true + } + return nil, false +} + +func (r structReflect) Has(key string) bool { + _, ok := r.findJsonNameField(key) + return ok +} + +func (r structReflect) Set(key string, val Value) { + fieldEntry, ok := TypeReflectEntryOf(r.Value.Type()).Fields()[key] + if !ok { + panic(fmt.Sprintf("key %s may not be set on struct %T: field does not exist", key, r.Value.Interface())) + } + oldVal := fieldEntry.GetFrom(r.Value) + newVal := reflect.ValueOf(val.Unstructured()) + r.update(fieldEntry, key, oldVal, newVal) +} + +func (r structReflect) Delete(key string) { + fieldEntry, ok := TypeReflectEntryOf(r.Value.Type()).Fields()[key] + if !ok { + panic(fmt.Sprintf("key %s may not be deleted on struct %T: field does not exist", key, r.Value.Interface())) + } + oldVal := fieldEntry.GetFrom(r.Value) + if oldVal.Kind() != reflect.Ptr && !fieldEntry.isOmitEmpty { + panic(fmt.Sprintf("key %s may not be deleted on struct: %T: value is neither a pointer nor an omitempty field", key, r.Value.Interface())) + } + r.update(fieldEntry, key, oldVal, reflect.Zero(oldVal.Type())) +} + +func (r structReflect) update(fieldEntry *FieldCacheEntry, key string, oldVal, newVal reflect.Value) { + if oldVal.CanSet() { + oldVal.Set(newVal) + return + } + + // map items are not addressable, so if a struct is contained in a map, the only way to modify it is + // to write a replacement fieldEntry into the map. + if r.ParentMap != nil { + if r.ParentMapKey == nil { + panic("ParentMapKey must not be nil if ParentMap is not nil") + } + replacement := reflect.New(r.Value.Type()).Elem() + fieldEntry.GetFrom(replacement).Set(newVal) + r.ParentMap.SetMapIndex(*r.ParentMapKey, replacement) + return + } + + // This should never happen since NewValueReflect ensures that the root object reflected on is a pointer and map + // item replacement is handled above. + panic(fmt.Sprintf("key %s may not be modified on struct: %T: struct is not settable", key, r.Value.Interface())) +} + +func (r structReflect) Iterate(fn func(string, Value) bool) bool { + return r.IterateUsing(HeapAllocator, fn) +} + +func (r structReflect) IterateUsing(a Allocator, fn func(string, Value) bool) bool { + vr := a.allocValueReflect() + defer a.Free(vr) + return eachStructField(r.Value, func(e *TypeReflectCacheEntry, s string, value reflect.Value) bool { + return fn(s, vr.mustReuse(value, e, nil, nil)) + }) +} + +func eachStructField(structVal reflect.Value, fn func(*TypeReflectCacheEntry, string, reflect.Value) bool) bool { + for _, fieldCacheEntry := range TypeReflectEntryOf(structVal.Type()).OrderedFields() { + fieldVal := fieldCacheEntry.GetFrom(structVal) + if fieldCacheEntry.CanOmit(fieldVal) { + // omit it + continue + } + ok := fn(fieldCacheEntry.TypeEntry, fieldCacheEntry.JsonName, fieldVal) + if !ok { + return false + } + } + return true +} + +func (r structReflect) Unstructured() interface{} { + // Use number of struct fields as a cheap way to rough estimate map size + result := make(map[string]interface{}, r.Value.NumField()) + r.Iterate(func(s string, value Value) bool { + result[s] = value.Unstructured() + return true + }) + return result +} + +func (r structReflect) Equals(m Map) bool { + return r.EqualsUsing(HeapAllocator, m) +} + +func (r structReflect) EqualsUsing(a Allocator, m Map) bool { + // MapEquals uses zip and is fairly efficient for structReflect + return MapEqualsUsing(a, &r, m) +} + +func (r structReflect) findJsonNameFieldAndNotEmpty(jsonName string) (reflect.Value, bool) { + structCacheEntry, ok := TypeReflectEntryOf(r.Value.Type()).Fields()[jsonName] + if !ok { + return reflect.Value{}, false + } + fieldVal := structCacheEntry.GetFrom(r.Value) + return fieldVal, !structCacheEntry.CanOmit(fieldVal) +} + +func (r structReflect) findJsonNameField(jsonName string) (val reflect.Value, ok bool) { + structCacheEntry, ok := TypeReflectEntryOf(r.Value.Type()).Fields()[jsonName] + if !ok { + return reflect.Value{}, false + } + fieldVal := structCacheEntry.GetFrom(r.Value) + return fieldVal, !structCacheEntry.CanOmit(fieldVal) +} + +func (r structReflect) Zip(other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + return r.ZipUsing(HeapAllocator, other, order, fn) +} + +func (r structReflect) ZipUsing(a Allocator, other Map, order MapTraverseOrder, fn func(key string, lhs, rhs Value) bool) bool { + if otherStruct, ok := other.(*structReflect); ok && r.Value.Type() == otherStruct.Value.Type() { + lhsvr, rhsvr := a.allocValueReflect(), a.allocValueReflect() + defer a.Free(lhsvr) + defer a.Free(rhsvr) + return r.structZip(otherStruct, lhsvr, rhsvr, fn) + } + return defaultMapZip(a, &r, other, order, fn) +} + +// structZip provides an optimized zip for structReflect types. The zip is always lexical key ordered since there is +// no additional cost to ordering the zip for structured types. +func (r structReflect) structZip(other *structReflect, lhsvr, rhsvr *valueReflect, fn func(key string, lhs, rhs Value) bool) bool { + lhsVal := r.Value + rhsVal := other.Value + + for _, fieldCacheEntry := range TypeReflectEntryOf(lhsVal.Type()).OrderedFields() { + lhsFieldVal := fieldCacheEntry.GetFrom(lhsVal) + rhsFieldVal := fieldCacheEntry.GetFrom(rhsVal) + lhsOmit := fieldCacheEntry.CanOmit(lhsFieldVal) + rhsOmit := fieldCacheEntry.CanOmit(rhsFieldVal) + if lhsOmit && rhsOmit { + continue + } + var lhsVal, rhsVal Value + if !lhsOmit { + lhsVal = lhsvr.mustReuse(lhsFieldVal, fieldCacheEntry.TypeEntry, nil, nil) + } + if !rhsOmit { + rhsVal = rhsvr.mustReuse(rhsFieldVal, fieldCacheEntry.TypeEntry, nil, nil) + } + if !fn(fieldCacheEntry.JsonName, lhsVal, rhsVal) { + return false + } + } + return true +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/value.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/value.go new file mode 100644 index 00000000..ea79e3a0 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/value.go @@ -0,0 +1,347 @@ +/* +Copyright 2018 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "bytes" + "fmt" + "io" + "strings" + + jsoniter "github.com/json-iterator/go" + "gopkg.in/yaml.v2" +) + +var ( + readPool = jsoniter.NewIterator(jsoniter.ConfigCompatibleWithStandardLibrary).Pool() + writePool = jsoniter.NewStream(jsoniter.ConfigCompatibleWithStandardLibrary, nil, 1024).Pool() +) + +// A Value corresponds to an 'atom' in the schema. It should return true +// for at least one of the IsXXX methods below, or the value is +// considered "invalid" +type Value interface { + // IsMap returns true if the Value is a Map, false otherwise. + IsMap() bool + // IsList returns true if the Value is a List, false otherwise. + IsList() bool + // IsBool returns true if the Value is a bool, false otherwise. + IsBool() bool + // IsInt returns true if the Value is a int64, false otherwise. + IsInt() bool + // IsFloat returns true if the Value is a float64, false + // otherwise. + IsFloat() bool + // IsString returns true if the Value is a string, false + // otherwise. + IsString() bool + // IsMap returns true if the Value is null, false otherwise. + IsNull() bool + + // AsMap converts the Value into a Map (or panic if the type + // doesn't allow it). + AsMap() Map + // AsMapUsing uses the provided allocator and converts the Value + // into a Map (or panic if the type doesn't allow it). + AsMapUsing(Allocator) Map + // AsList converts the Value into a List (or panic if the type + // doesn't allow it). + AsList() List + // AsListUsing uses the provided allocator and converts the Value + // into a List (or panic if the type doesn't allow it). + AsListUsing(Allocator) List + // AsBool converts the Value into a bool (or panic if the type + // doesn't allow it). + AsBool() bool + // AsInt converts the Value into an int64 (or panic if the type + // doesn't allow it). + AsInt() int64 + // AsFloat converts the Value into a float64 (or panic if the type + // doesn't allow it). + AsFloat() float64 + // AsString converts the Value into a string (or panic if the type + // doesn't allow it). + AsString() string + + // Unstructured converts the Value into an Unstructured interface{}. + Unstructured() interface{} +} + +// FromJSON is a helper function for reading a JSON document. +func FromJSON(input []byte) (Value, error) { + return FromJSONFast(input) +} + +// FromJSONFast is a helper function for reading a JSON document. +func FromJSONFast(input []byte) (Value, error) { + iter := readPool.BorrowIterator(input) + defer readPool.ReturnIterator(iter) + return ReadJSONIter(iter) +} + +// ToJSON is a helper function for producing a JSon document. +func ToJSON(v Value) ([]byte, error) { + buf := bytes.Buffer{} + stream := writePool.BorrowStream(&buf) + defer writePool.ReturnStream(stream) + WriteJSONStream(v, stream) + b := stream.Buffer() + err := stream.Flush() + // Help jsoniter manage its buffers--without this, the next + // use of the stream is likely to require an allocation. Look + // at the jsoniter stream code to understand why. They were probably + // optimizing for folks using the buffer directly. + stream.SetBuffer(b[:0]) + return buf.Bytes(), err +} + +// ReadJSONIter reads a Value from a JSON iterator. +func ReadJSONIter(iter *jsoniter.Iterator) (Value, error) { + v := iter.Read() + if iter.Error != nil && iter.Error != io.EOF { + return nil, iter.Error + } + return NewValueInterface(v), nil +} + +// WriteJSONStream writes a value into a JSON stream. +func WriteJSONStream(v Value, stream *jsoniter.Stream) { + stream.WriteVal(v.Unstructured()) +} + +// ToYAML marshals a value as YAML. +func ToYAML(v Value) ([]byte, error) { + return yaml.Marshal(v.Unstructured()) +} + +// Equals returns true iff the two values are equal. +func Equals(lhs, rhs Value) bool { + return EqualsUsing(HeapAllocator, lhs, rhs) +} + +// EqualsUsing uses the provided allocator and returns true iff the two values are equal. +func EqualsUsing(a Allocator, lhs, rhs Value) bool { + if lhs.IsFloat() || rhs.IsFloat() { + var lf float64 + if lhs.IsFloat() { + lf = lhs.AsFloat() + } else if lhs.IsInt() { + lf = float64(lhs.AsInt()) + } else { + return false + } + var rf float64 + if rhs.IsFloat() { + rf = rhs.AsFloat() + } else if rhs.IsInt() { + rf = float64(rhs.AsInt()) + } else { + return false + } + return lf == rf + } + if lhs.IsInt() { + if rhs.IsInt() { + return lhs.AsInt() == rhs.AsInt() + } + return false + } else if rhs.IsInt() { + return false + } + if lhs.IsString() { + if rhs.IsString() { + return lhs.AsString() == rhs.AsString() + } + return false + } else if rhs.IsString() { + return false + } + if lhs.IsBool() { + if rhs.IsBool() { + return lhs.AsBool() == rhs.AsBool() + } + return false + } else if rhs.IsBool() { + return false + } + if lhs.IsList() { + if rhs.IsList() { + lhsList := lhs.AsListUsing(a) + defer a.Free(lhsList) + rhsList := rhs.AsListUsing(a) + defer a.Free(rhsList) + return lhsList.EqualsUsing(a, rhsList) + } + return false + } else if rhs.IsList() { + return false + } + if lhs.IsMap() { + if rhs.IsMap() { + lhsList := lhs.AsMapUsing(a) + defer a.Free(lhsList) + rhsList := rhs.AsMapUsing(a) + defer a.Free(rhsList) + return lhsList.EqualsUsing(a, rhsList) + } + return false + } else if rhs.IsMap() { + return false + } + if lhs.IsNull() { + if rhs.IsNull() { + return true + } + return false + } else if rhs.IsNull() { + return false + } + // No field is set, on either objects. + return true +} + +// ToString returns a human-readable representation of the value. +func ToString(v Value) string { + if v.IsNull() { + return "null" + } + switch { + case v.IsFloat(): + return fmt.Sprintf("%v", v.AsFloat()) + case v.IsInt(): + return fmt.Sprintf("%v", v.AsInt()) + case v.IsString(): + return fmt.Sprintf("%q", v.AsString()) + case v.IsBool(): + return fmt.Sprintf("%v", v.AsBool()) + case v.IsList(): + strs := []string{} + list := v.AsList() + for i := 0; i < list.Length(); i++ { + strs = append(strs, ToString(list.At(i))) + } + return "[" + strings.Join(strs, ",") + "]" + case v.IsMap(): + strs := []string{} + v.AsMap().Iterate(func(k string, v Value) bool { + strs = append(strs, fmt.Sprintf("%v=%v", k, ToString(v))) + return true + }) + return strings.Join(strs, "") + } + // No field is set, on either objects. + return "{{undefined}}" +} + +// Less provides a total ordering for Value (so that they can be sorted, even +// if they are of different types). +func Less(lhs, rhs Value) bool { + return Compare(lhs, rhs) == -1 +} + +// Compare provides a total ordering for Value (so that they can be +// sorted, even if they are of different types). The result will be 0 if +// v==rhs, -1 if v < rhs, and +1 if v > rhs. +func Compare(lhs, rhs Value) int { + return CompareUsing(HeapAllocator, lhs, rhs) +} + +// CompareUsing uses the provided allocator and provides a total +// ordering for Value (so that they can be sorted, even if they +// are of different types). The result will be 0 if v==rhs, -1 +// if v < rhs, and +1 if v > rhs. +func CompareUsing(a Allocator, lhs, rhs Value) int { + if lhs.IsFloat() { + if !rhs.IsFloat() { + // Extra: compare floats and ints numerically. + if rhs.IsInt() { + return FloatCompare(lhs.AsFloat(), float64(rhs.AsInt())) + } + return -1 + } + return FloatCompare(lhs.AsFloat(), rhs.AsFloat()) + } else if rhs.IsFloat() { + // Extra: compare floats and ints numerically. + if lhs.IsInt() { + return FloatCompare(float64(lhs.AsInt()), rhs.AsFloat()) + } + return 1 + } + + if lhs.IsInt() { + if !rhs.IsInt() { + return -1 + } + return IntCompare(lhs.AsInt(), rhs.AsInt()) + } else if rhs.IsInt() { + return 1 + } + + if lhs.IsString() { + if !rhs.IsString() { + return -1 + } + return strings.Compare(lhs.AsString(), rhs.AsString()) + } else if rhs.IsString() { + return 1 + } + + if lhs.IsBool() { + if !rhs.IsBool() { + return -1 + } + return BoolCompare(lhs.AsBool(), rhs.AsBool()) + } else if rhs.IsBool() { + return 1 + } + + if lhs.IsList() { + if !rhs.IsList() { + return -1 + } + lhsList := lhs.AsListUsing(a) + defer a.Free(lhsList) + rhsList := rhs.AsListUsing(a) + defer a.Free(rhsList) + return ListCompareUsing(a, lhsList, rhsList) + } else if rhs.IsList() { + return 1 + } + if lhs.IsMap() { + if !rhs.IsMap() { + return -1 + } + lhsMap := lhs.AsMapUsing(a) + defer a.Free(lhsMap) + rhsMap := rhs.AsMapUsing(a) + defer a.Free(rhsMap) + return MapCompareUsing(a, lhsMap, rhsMap) + } else if rhs.IsMap() { + return 1 + } + if lhs.IsNull() { + if !rhs.IsNull() { + return -1 + } + return 0 + } else if rhs.IsNull() { + return 1 + } + + // Invalid Value-- nothing is set. + return 0 +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/valuereflect.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/valuereflect.go new file mode 100644 index 00000000..05e70deb --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/valuereflect.go @@ -0,0 +1,294 @@ +/* +Copyright 2019 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "encoding/base64" + "fmt" + "reflect" +) + +// NewValueReflect creates a Value backed by an "interface{}" type, +// typically an structured object in Kubernetes world that is uses reflection to expose. +// The provided "interface{}" value must be a pointer so that the value can be modified via reflection. +// The provided "interface{}" may contain structs and types that are converted to Values +// by the jsonMarshaler interface. +func NewValueReflect(value interface{}) (Value, error) { + if value == nil { + return NewValueInterface(nil), nil + } + v := reflect.ValueOf(value) + if v.Kind() != reflect.Ptr { + // The root value to reflect on must be a pointer so that map.Set() and map.Delete() operations are possible. + return nil, fmt.Errorf("value provided to NewValueReflect must be a pointer") + } + return wrapValueReflect(v, nil, nil) +} + +// wrapValueReflect wraps the provide reflect.Value as a value. If parent in the data tree is a map, parentMap +// and parentMapKey must be provided so that the returned value may be set and deleted. +func wrapValueReflect(value reflect.Value, parentMap, parentMapKey *reflect.Value) (Value, error) { + val := HeapAllocator.allocValueReflect() + return val.reuse(value, nil, parentMap, parentMapKey) +} + +// wrapValueReflect wraps the provide reflect.Value as a value, and panics if there is an error. If parent in the data +// tree is a map, parentMap and parentMapKey must be provided so that the returned value may be set and deleted. +func mustWrapValueReflect(value reflect.Value, parentMap, parentMapKey *reflect.Value) Value { + v, err := wrapValueReflect(value, parentMap, parentMapKey) + if err != nil { + panic(err) + } + return v +} + +// the value interface doesn't care about the type for value.IsNull, so we can use a constant +var nilType = reflect.TypeOf(&struct{}{}) + +// reuse replaces the value of the valueReflect. If parent in the data tree is a map, parentMap and parentMapKey +// must be provided so that the returned value may be set and deleted. +func (r *valueReflect) reuse(value reflect.Value, cacheEntry *TypeReflectCacheEntry, parentMap, parentMapKey *reflect.Value) (Value, error) { + if cacheEntry == nil { + cacheEntry = TypeReflectEntryOf(value.Type()) + } + if cacheEntry.CanConvertToUnstructured() { + u, err := cacheEntry.ToUnstructured(value) + if err != nil { + return nil, err + } + if u == nil { + value = reflect.Zero(nilType) + } else { + value = reflect.ValueOf(u) + } + } + r.Value = dereference(value) + r.ParentMap = parentMap + r.ParentMapKey = parentMapKey + r.kind = kind(r.Value) + return r, nil +} + +// mustReuse replaces the value of the valueReflect and panics if there is an error. If parent in the data tree is a +// map, parentMap and parentMapKey must be provided so that the returned value may be set and deleted. +func (r *valueReflect) mustReuse(value reflect.Value, cacheEntry *TypeReflectCacheEntry, parentMap, parentMapKey *reflect.Value) Value { + v, err := r.reuse(value, cacheEntry, parentMap, parentMapKey) + if err != nil { + panic(err) + } + return v +} + +func dereference(val reflect.Value) reflect.Value { + kind := val.Kind() + if (kind == reflect.Interface || kind == reflect.Ptr) && !safeIsNil(val) { + return val.Elem() + } + return val +} + +type valueReflect struct { + ParentMap *reflect.Value + ParentMapKey *reflect.Value + Value reflect.Value + kind reflectType +} + +func (r valueReflect) IsMap() bool { + return r.kind == mapType || r.kind == structMapType +} + +func (r valueReflect) IsList() bool { + return r.kind == listType +} + +func (r valueReflect) IsBool() bool { + return r.kind == boolType +} + +func (r valueReflect) IsInt() bool { + return r.kind == intType || r.kind == uintType +} + +func (r valueReflect) IsFloat() bool { + return r.kind == floatType +} + +func (r valueReflect) IsString() bool { + return r.kind == stringType || r.kind == byteStringType +} + +func (r valueReflect) IsNull() bool { + return r.kind == nullType +} + +type reflectType = int + +const ( + mapType = iota + structMapType + listType + intType + uintType + floatType + stringType + byteStringType + boolType + nullType +) + +func kind(v reflect.Value) reflectType { + typ := v.Type() + rk := typ.Kind() + switch rk { + case reflect.Map: + if v.IsNil() { + return nullType + } + return mapType + case reflect.Struct: + return structMapType + case reflect.Int, reflect.Int64, reflect.Int32, reflect.Int16, reflect.Int8: + return intType + case reflect.Uint, reflect.Uint32, reflect.Uint16, reflect.Uint8: + // Uint64 deliberately excluded, see valueUnstructured.Int. + return uintType + case reflect.Float64, reflect.Float32: + return floatType + case reflect.String: + return stringType + case reflect.Bool: + return boolType + case reflect.Slice: + if v.IsNil() { + return nullType + } + elemKind := typ.Elem().Kind() + if elemKind == reflect.Uint8 { + return byteStringType + } + return listType + case reflect.Chan, reflect.Func, reflect.Ptr, reflect.UnsafePointer, reflect.Interface: + if v.IsNil() { + return nullType + } + panic(fmt.Sprintf("unsupported type: %v", v.Type())) + default: + panic(fmt.Sprintf("unsupported type: %v", v.Type())) + } +} + +// TODO find a cleaner way to avoid panics from reflect.IsNil() +func safeIsNil(v reflect.Value) bool { + k := v.Kind() + switch k { + case reflect.Chan, reflect.Func, reflect.Map, reflect.Ptr, reflect.UnsafePointer, reflect.Interface, reflect.Slice: + return v.IsNil() + } + return false +} + +func (r valueReflect) AsMap() Map { + return r.AsMapUsing(HeapAllocator) +} + +func (r valueReflect) AsMapUsing(a Allocator) Map { + switch r.kind { + case structMapType: + v := a.allocStructReflect() + v.valueReflect = r + return v + case mapType: + v := a.allocMapReflect() + v.valueReflect = r + return v + default: + panic("value is not a map or struct") + } +} + +func (r valueReflect) AsList() List { + return r.AsListUsing(HeapAllocator) +} + +func (r valueReflect) AsListUsing(a Allocator) List { + if r.IsList() { + v := a.allocListReflect() + v.Value = r.Value + return v + } + panic("value is not a list") +} + +func (r valueReflect) AsBool() bool { + if r.IsBool() { + return r.Value.Bool() + } + panic("value is not a bool") +} + +func (r valueReflect) AsInt() int64 { + if r.kind == intType { + return r.Value.Int() + } + if r.kind == uintType { + return int64(r.Value.Uint()) + } + + panic("value is not an int") +} + +func (r valueReflect) AsFloat() float64 { + if r.IsFloat() { + return r.Value.Float() + } + panic("value is not a float") +} + +func (r valueReflect) AsString() string { + switch r.kind { + case stringType: + return r.Value.String() + case byteStringType: + return base64.StdEncoding.EncodeToString(r.Value.Bytes()) + } + panic("value is not a string") +} + +func (r valueReflect) Unstructured() interface{} { + val := r.Value + switch { + case r.IsNull(): + return nil + case val.Kind() == reflect.Struct: + return structReflect{r}.Unstructured() + case val.Kind() == reflect.Map: + return mapReflect{valueReflect: r}.Unstructured() + case r.IsList(): + return listReflect{r.Value}.Unstructured() + case r.IsString(): + return r.AsString() + case r.IsInt(): + return r.AsInt() + case r.IsBool(): + return r.AsBool() + case r.IsFloat(): + return r.AsFloat() + default: + panic(fmt.Sprintf("value of type %s is not a supported by value reflector", val.Type())) + } +} diff --git a/vendor/sigs.k8s.io/structured-merge-diff/v4/value/valueunstructured.go b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/valueunstructured.go new file mode 100644 index 00000000..ac5a9262 --- /dev/null +++ b/vendor/sigs.k8s.io/structured-merge-diff/v4/value/valueunstructured.go @@ -0,0 +1,178 @@ +/* +Copyright 2018 The Kubernetes Authors. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. +*/ + +package value + +import ( + "fmt" +) + +// NewValueInterface creates a Value backed by an "interface{}" type, +// typically an unstructured object in Kubernetes world. +// interface{} must be one of: map[string]interface{}, map[interface{}]interface{}, []interface{}, int types, float types, +// string or boolean. Nested interface{} must also be one of these types. +func NewValueInterface(v interface{}) Value { + return Value(HeapAllocator.allocValueUnstructured().reuse(v)) +} + +type valueUnstructured struct { + Value interface{} +} + +// reuse replaces the value of the valueUnstructured. +func (vi *valueUnstructured) reuse(value interface{}) Value { + vi.Value = value + return vi +} + +func (v valueUnstructured) IsMap() bool { + if _, ok := v.Value.(map[string]interface{}); ok { + return true + } + if _, ok := v.Value.(map[interface{}]interface{}); ok { + return true + } + return false +} + +func (v valueUnstructured) AsMap() Map { + return v.AsMapUsing(HeapAllocator) +} + +func (v valueUnstructured) AsMapUsing(_ Allocator) Map { + if v.Value == nil { + panic("invalid nil") + } + switch t := v.Value.(type) { + case map[string]interface{}: + return mapUnstructuredString(t) + case map[interface{}]interface{}: + return mapUnstructuredInterface(t) + } + panic(fmt.Errorf("not a map: %#v", v)) +} + +func (v valueUnstructured) IsList() bool { + if v.Value == nil { + return false + } + _, ok := v.Value.([]interface{}) + return ok +} + +func (v valueUnstructured) AsList() List { + return v.AsListUsing(HeapAllocator) +} + +func (v valueUnstructured) AsListUsing(_ Allocator) List { + return listUnstructured(v.Value.([]interface{})) +} + +func (v valueUnstructured) IsFloat() bool { + if v.Value == nil { + return false + } else if _, ok := v.Value.(float64); ok { + return true + } else if _, ok := v.Value.(float32); ok { + return true + } + return false +} + +func (v valueUnstructured) AsFloat() float64 { + if f, ok := v.Value.(float32); ok { + return float64(f) + } + return v.Value.(float64) +} + +func (v valueUnstructured) IsInt() bool { + if v.Value == nil { + return false + } else if _, ok := v.Value.(int); ok { + return true + } else if _, ok := v.Value.(int8); ok { + return true + } else if _, ok := v.Value.(int16); ok { + return true + } else if _, ok := v.Value.(int32); ok { + return true + } else if _, ok := v.Value.(int64); ok { + return true + } else if _, ok := v.Value.(uint); ok { + return true + } else if _, ok := v.Value.(uint8); ok { + return true + } else if _, ok := v.Value.(uint16); ok { + return true + } else if _, ok := v.Value.(uint32); ok { + return true + } + return false +} + +func (v valueUnstructured) AsInt() int64 { + if i, ok := v.Value.(int); ok { + return int64(i) + } else if i, ok := v.Value.(int8); ok { + return int64(i) + } else if i, ok := v.Value.(int16); ok { + return int64(i) + } else if i, ok := v.Value.(int32); ok { + return int64(i) + } else if i, ok := v.Value.(uint); ok { + return int64(i) + } else if i, ok := v.Value.(uint8); ok { + return int64(i) + } else if i, ok := v.Value.(uint16); ok { + return int64(i) + } else if i, ok := v.Value.(uint32); ok { + return int64(i) + } + return v.Value.(int64) +} + +func (v valueUnstructured) IsString() bool { + if v.Value == nil { + return false + } + _, ok := v.Value.(string) + return ok +} + +func (v valueUnstructured) AsString() string { + return v.Value.(string) +} + +func (v valueUnstructured) IsBool() bool { + if v.Value == nil { + return false + } + _, ok := v.Value.(bool) + return ok +} + +func (v valueUnstructured) AsBool() bool { + return v.Value.(bool) +} + +func (v valueUnstructured) IsNull() bool { + return v.Value == nil +} + +func (v valueUnstructured) Unstructured() interface{} { + return v.Value +} From 6a687a1c40cdd6bb458e39db5a4a758a45a7b0eb Mon Sep 17 00:00:00 2001 From: Serge Bazanski Date: Tue, 3 Nov 2020 19:32:08 +0100 Subject: [PATCH 4/4] integration: update against new k8s api --- integration/delete_test.go | 10 ++-- integration/integration_suite_test.go | 9 ++-- integration/kubeflags_test.go | 7 +-- integration/update_test.go | 66 ++++++++++++++------------- 4 files changed, 50 insertions(+), 42 deletions(-) diff --git a/integration/delete_test.go b/integration/delete_test.go index 001a9fdf..d5220d37 100644 --- a/integration/delete_test.go +++ b/integration/delete_test.go @@ -3,7 +3,9 @@ package integration import ( - "k8s.io/api/core/v1" + "context" + + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime" corev1 "k8s.io/client-go/kubernetes/typed/core/v1" @@ -50,7 +52,7 @@ var _ = Describe("delete", func() { Context("With no existing state", func() { It("should succeed", func() { - Expect(c.ConfigMaps(ns).List(metav1.ListOptions{})). + Expect(c.ConfigMaps(ns).List(context.Background(), metav1.ListOptions{})). To(WithTransform(objNames, BeEmpty())) }) }) @@ -68,14 +70,14 @@ var _ = Describe("delete", func() { toCreate = append(toCreate, baz) for _, cm := range toCreate { - _, err := c.ConfigMaps(ns).Create(cm) + _, err := c.ConfigMaps(ns).Create(context.Background(), cm, metav1.CreateOptions{}) Expect(err).To(Not(HaveOccurred())) } }) It("should delete mentioned objects", func() { Eventually(func() (*v1.ConfigMapList, error) { - return c.ConfigMaps(ns).List(metav1.ListOptions{}) + return c.ConfigMaps(ns).List(context.Background(), metav1.ListOptions{}) }).Should(WithTransform(objNames, ConsistOf("baz"))) }) }) diff --git a/integration/integration_suite_test.go b/integration/integration_suite_test.go index 4dfa32ca..00069fd8 100644 --- a/integration/integration_suite_test.go +++ b/integration/integration_suite_test.go @@ -3,6 +3,7 @@ package integration import ( + "context" "flag" "fmt" "io" @@ -12,7 +13,7 @@ import ( "path/filepath" "testing" - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apimachinery/pkg/runtime" @@ -49,11 +50,13 @@ func clusterConfigOrDie() *rest.Config { func createNsOrDie(c corev1.CoreV1Interface, ns string) string { result, err := c.Namespaces().Create( + context.Background(), &v1.Namespace{ ObjectMeta: metav1.ObjectMeta{ GenerateName: ns, }, - }) + }, + metav1.CreateOptions{}) if err != nil { panic(err.Error()) } @@ -63,7 +66,7 @@ func createNsOrDie(c corev1.CoreV1Interface, ns string) string { } func deleteNsOrDie(c corev1.CoreV1Interface, ns string) { - err := c.Namespaces().Delete(ns, &metav1.DeleteOptions{}) + err := c.Namespaces().Delete(context.Background(), ns, metav1.DeleteOptions{}) if err != nil { panic(err.Error()) } diff --git a/integration/kubeflags_test.go b/integration/kubeflags_test.go index 6156318d..17d71cfb 100644 --- a/integration/kubeflags_test.go +++ b/integration/kubeflags_test.go @@ -3,11 +3,12 @@ package integration import ( + "context" "io/ioutil" "os" "os/exec" - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/runtime" corev1 "k8s.io/client-go/kubernetes/typed/core/v1" @@ -93,7 +94,7 @@ var _ = Describe("flags", func() { It("should update correct namespace", func() { Expect(kubecfgExit).NotTo(HaveOccurred()) - Expect(c.ConfigMaps(ns).Get(testName, metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), testName, metav1.GetOptions{})). NotTo(BeNil()) }) }) @@ -108,7 +109,7 @@ var _ = Describe("flags", func() { It("should update correct namespace", func() { Expect(kubecfgExit).NotTo(HaveOccurred()) - Expect(c.ConfigMaps(ns).Get(testName, metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), testName, metav1.GetOptions{})). NotTo(BeNil()) }) }) diff --git a/integration/update_test.go b/integration/update_test.go index 97ae73da..48e06cc3 100644 --- a/integration/update_test.go +++ b/integration/update_test.go @@ -4,10 +4,11 @@ package integration import ( "bytes" + "context" "fmt" appsv1 "k8s.io/api/apps/v1" - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" "k8s.io/apimachinery/pkg/api/errors" "k8s.io/apimachinery/pkg/api/resource" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" @@ -116,7 +117,7 @@ var _ = Describe("update", func() { Context("With no existing state", func() { It("should produce expected object", func() { - Expect(c.ConfigMaps(ns).Get(cmName, metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), cmName, metav1.GetOptions{})). To(WithTransform(cmData, HaveKeyWithValue("foo", "bar"))) Expect(kubecfgOut.String()). To(ContainSubstring("Creating configmaps %s", cmName)) @@ -127,12 +128,12 @@ var _ = Describe("update", func() { Context("With existing non-kubecfg object", func() { BeforeEach(func() { - _, err := c.ConfigMaps(ns).Create(cm) + _, err := c.ConfigMaps(ns).Create(context.Background(), cm, metav1.CreateOptions{}) Expect(err).NotTo(HaveOccurred()) }) It("should succeed", func() { - Expect(c.ConfigMaps(ns).Get(cmName, metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), cmName, metav1.GetOptions{})). To(WithTransform(cmData, HaveKeyWithValue("foo", "bar"))) // NB: may report "Updating" - that's ok. Expect(kubecfgOut.String()). @@ -145,12 +146,12 @@ var _ = Describe("update", func() { err := runKubecfgWith([]string{"update", "-vv", "-n", ns}, []runtime.Object{cm}) Expect(err).NotTo(HaveOccurred()) - Expect(c.ConfigMaps(ns).Get(cmName, metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), cmName, metav1.GetOptions{})). To(WithTransform(metav1.Object.GetAnnotations, HaveKey(kubecfg.AnnotationOrigObject))) }) It("should not report a change", func() { - Expect(c.ConfigMaps(ns).Get(cmName, metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), cmName, metav1.GetOptions{})). To(WithTransform(cmData, HaveKeyWithValue("foo", "bar"))) // no change -> should report neither Updating nor Creating Expect(kubecfgOut.String()). @@ -172,7 +173,7 @@ var _ = Describe("update", func() { }) It("should update the object", func() { - Expect(c.ConfigMaps(ns).Get(cmName, metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), cmName, metav1.GetOptions{})). To(WithTransform(cmData, HaveKeyWithValue("foo", "bar"))) Expect(kubecfgOut.String()). To(ContainSubstring("Updating configmaps %s", cmName)) @@ -193,14 +194,14 @@ var _ = Describe("update", func() { Expect(err).NotTo(HaveOccurred()) // ... and then modified by another controller/whatever. - o, err := c.ConfigMaps(ns).Patch(cmName, types.MergePatchType, - []byte(`{"metadata": {"annotations": {"addedby": "3rd-party"}}}`)) + o, err := c.ConfigMaps(ns).Patch(context.Background(), cmName, types.MergePatchType, + []byte(`{"metadata": {"annotations": {"addedby": "3rd-party"}}}`), metav1.PatchOptions{}) Expect(err).NotTo(HaveOccurred()) fmt.Fprintf(GinkgoWriter, "patch result is %v\n", o) }) It("should update the object", func() { - cm, err := c.ConfigMaps(ns).Get(cmName, metav1.GetOptions{}) + cm, err := c.ConfigMaps(ns).Get(context.Background(), cmName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) Expect(cm).To(WithTransform(cmData, HaveKeyWithValue("foo", "bar"))) Expect(cm).To(WithTransform(metav1.Object.GetAnnotations, And( @@ -282,11 +283,11 @@ var _ = Describe("update", func() { Version: "v1beta1", Resource: "customresourcedefinitions", }) - _ = rc.Delete("things.random.example.com", &metav1.DeleteOptions{}) + _ = rc.Delete(context.Background(), "things.random.example.com", metav1.DeleteOptions{}) }) It("should update correctly", func() { - Expect(rc.Namespace(ns).Get("thing1", metav1.GetOptions{})). + Expect(rc.Namespace(ns).Get(context.Background(), "thing1", metav1.GetOptions{})). To(WithTransform(metav1.Object.GetAnnotations, HaveKeyWithValue("gen", "one"))) @@ -297,7 +298,7 @@ var _ = Describe("update", func() { Expect(err).NotTo(HaveOccurred()) // Verify result - Expect(rc.Namespace(ns).Get("thing1", metav1.GetOptions{})). + Expect(rc.Namespace(ns).Get(context.Background(), "thing1", metav1.GetOptions{})). To(WithTransform(metav1.Object.GetAnnotations, And( HaveKeyWithValue("gen", "two"), HaveKeyWithValue("unrelated", "baz"), @@ -328,7 +329,7 @@ var _ = Describe("update", func() { }) It("should not change ports on subsequent updates", func() { - svc, err := c.Services(ns).Get(svcName, metav1.GetOptions{}) + svc, err := c.Services(ns).Get(context.Background(), svcName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) Expect(svc).To(WithTransform(metav1.Object.GetAnnotations, HaveKeyWithValue("generation", "one"))) @@ -341,7 +342,7 @@ var _ = Describe("update", func() { Expect(err).NotTo(HaveOccurred()) // Check NodePort hasn't changed - svc, err = c.Services(ns).Get(svcName, metav1.GetOptions{}) + svc, err = c.Services(ns).Get(context.Background(), svcName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) Expect(svc).To(WithTransform(metav1.Object.GetAnnotations, HaveKeyWithValue("generation", "two"))) @@ -384,9 +385,10 @@ var _ = Describe("update", func() { It("should merge correctly into containers[]", func() { // Simulate change by 3rd-party appsc := appsv1client.NewForConfigOrDie(clusterConfigOrDie()) - _, err := appsc.Deployments(ns).Patch(deployName, + _, err := appsc.Deployments(ns).Patch(context.Background(), deployName, types.StrategicMergePatchType, []byte(`{"spec": {"template": {"spec": {"containers": [{"name": "c", "resources": {"limits": {"cpu": "1"}}}]}}}}`), + metav1.PatchOptions{}, ) Expect(err).NotTo(HaveOccurred()) @@ -396,7 +398,7 @@ var _ = Describe("update", func() { Expect(err).NotTo(HaveOccurred()) // Check container.resources was preserved - d, err := appsc.Deployments(ns).Get(deployName, metav1.GetOptions{}) + d, err := appsc.Deployments(ns).Get(context.Background(), deployName, metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) Expect(d).To(WithTransform(metav1.Object.GetAnnotations, HaveKeyWithValue("gen", "two"))) @@ -437,13 +439,13 @@ var _ = Describe("update", func() { }) It("should create objects in the correct namespaces", func() { - Expect(c.ConfigMaps(ns).Get("nons", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), "nons", metav1.GetOptions{})). NotTo(BeNil()) - Expect(c.ConfigMaps(ns).Get("ns1", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), "ns1", metav1.GetOptions{})). NotTo(BeNil()) - Expect(c.ConfigMaps(ns2).Get("ns2", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns2).Get(context.Background(), "ns2", metav1.GetOptions{})). NotTo(BeNil()) }) }) @@ -465,7 +467,7 @@ var _ = Describe("update", func() { JustBeforeEach(func() { for _, obj := range preExist { - _, err := c.ConfigMaps(ns).Create(obj) + _, err := c.ConfigMaps(ns).Create(context.Background(), obj, metav1.CreateOptions{}) Expect(err).NotTo(HaveOccurred()) } @@ -578,7 +580,7 @@ var _ = Describe("update", func() { }) It("should add gctag to new object", func() { - o, err := c.ConfigMaps(ns).Get("new", metav1.GetOptions{}) + o, err := c.ConfigMaps(ns).Get(context.Background(), "new", metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // [gctag-migration]: Remove annotation in phase2 Expect(o.ObjectMeta.Annotations). @@ -588,7 +590,7 @@ var _ = Describe("update", func() { }) It("should keep gctag on existing object", func() { - o, err := c.ConfigMaps(ns).Get("existing", metav1.GetOptions{}) + o, err := c.ConfigMaps(ns).Get(context.Background(), "existing", metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // [gctag-migration]: Remove annotation in phase2 Expect(o.ObjectMeta.Annotations). @@ -599,29 +601,29 @@ var _ = Describe("update", func() { // [gctag-migration]: Pre-migration test. Remove in phase2 It("should add gctag label to pre-migration object", func() { - o, err := c.ConfigMaps(ns).Get("existing-premigration", metav1.GetOptions{}) + o, err := c.ConfigMaps(ns).Get(context.Background(), "existing-premigration", metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) Expect(o.ObjectMeta.Labels). To(HaveKeyWithValue(kubecfg.LabelGcTag, gcTag)) }) It("should delete stale object", func() { - _, err := c.ConfigMaps(ns).Get("existing-stale", metav1.GetOptions{}) + _, err := c.ConfigMaps(ns).Get(context.Background(), "existing-stale", metav1.GetOptions{}) Expect(errors.IsNotFound(err)).To(BeTrue()) }) It("should not delete tagless object", func() { - Expect(c.ConfigMaps(ns).Get("existing-stale-notag", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), "existing-stale-notag", metav1.GetOptions{})). NotTo(BeNil()) }) It("should not delete object with different gc tag", func() { - Expect(c.ConfigMaps(ns).Get("existing-othertag", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), "existing-othertag", metav1.GetOptions{})). NotTo(BeNil()) }) It("should not delete strategy=ignore object", func() { - Expect(c.ConfigMaps(ns).Get("existing-precious", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), "existing-precious", metav1.GetOptions{})). NotTo(BeNil()) }) }) @@ -653,12 +655,12 @@ var _ = Describe("update", func() { }) It("should not delete existing object", func() { - Expect(c.ConfigMaps(ns).Get("existing", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), "existing", metav1.GetOptions{})). NotTo(BeNil()) }) It("should not create new object", func() { - _, err := c.ConfigMaps(ns).Get("new", metav1.GetOptions{}) + _, err := c.ConfigMaps(ns).Get(context.Background(), "new", metav1.GetOptions{}) Expect(errors.IsNotFound(err)).To(BeTrue()) }) }) @@ -690,12 +692,12 @@ var _ = Describe("update", func() { }) It("should not delete existing object", func() { - Expect(c.ConfigMaps(ns).Get("existing", metav1.GetOptions{})). + Expect(c.ConfigMaps(ns).Get(context.Background(), "existing", metav1.GetOptions{})). NotTo(BeNil()) }) It("should add gctag to new object", func() { - o, err := c.ConfigMaps(ns).Get("new", metav1.GetOptions{}) + o, err := c.ConfigMaps(ns).Get(context.Background(), "new", metav1.GetOptions{}) Expect(err).NotTo(HaveOccurred()) // [gctag-migration]: Remove annotation in phase2 Expect(o.ObjectMeta.Annotations).